AU746669B2 - Method and apparatus for segmenting images - Google Patents

Method and apparatus for segmenting images Download PDF

Info

Publication number
AU746669B2
AU746669B2 AU52620/99A AU5262099A AU746669B2 AU 746669 B2 AU746669 B2 AU 746669B2 AU 52620/99 A AU52620/99 A AU 52620/99A AU 5262099 A AU5262099 A AU 5262099A AU 746669 B2 AU746669 B2 AU 746669B2
Authority
AU
Australia
Prior art keywords
image
region
regions
pixels
property
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
AU52620/99A
Other versions
AU5262099A (en
Inventor
Mathieu Hitter
Delphine Anh Dao Le
Alison Joan Lennon
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from AUPP6344A external-priority patent/AUPP634498A0/en
Priority claimed from AUPP6343A external-priority patent/AUPP634398A0/en
Priority claimed from AUPP6342A external-priority patent/AUPP634298A0/en
Priority claimed from AUPP6341A external-priority patent/AUPP634198A0/en
Application filed by Canon Inc filed Critical Canon Inc
Priority to AU52620/99A priority Critical patent/AU746669B2/en
Publication of AU5262099A publication Critical patent/AU5262099A/en
Application granted granted Critical
Publication of AU746669B2 publication Critical patent/AU746669B2/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/189Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding
    • H04N19/192Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the adaptation method, adaptation tool or adaptation type used for the adaptive coding the adaptation method, adaptation tool or adaptation type being iterative or recursive
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/14Coding unit complexity, e.g. amount of activity or edge presence estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/96Tree coding, e.g. quad-tree coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20112Image segmentation details
    • G06T2207/20156Automatic seed setting

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)

Description

METHOD AND APPARATUS FOR SEGMENTING IMAGES Field of Invention The present invention relates to the field of image segmentation. In particular, one aspect of the invention relates to the seeding of an image. Another aspect of the invention relates to segmenting images. A still further aspect of the invention relates to encoding an image having a number of segmented regions as a coded representation.
Background of Invention Image segmentation is an initial step in many image processing tasks such as pattern recognition, image coding and image interpretation. For example, in scene understanding applications, the segmentation process generally provides a labelling process with regions to be classified.
The publication entitled "Digital Image Processing" by Rafael C. Gonzalez and Richard E. Woods, Addison Wesley Publishing Company 1993 discloses on page 461- 15 465, a method for image segmentation. This method utilizes a region splitting and merging procedure. This procedure subdivides an image initially into a set of arbitrary, disjointed regions and then merges and/or splits the regions depending whether the regions satisfy a certain homogeneity criteria. Typically, the homogeneity criterion is based on a threshold value arbitrarily selected by a user. However, this method suffers from the disadvantage that the choice of threshold values is critical for successful image segmentation. Specifically, a particular threshold value may work with one image but not necessarily with others. For example, this method often fails to split regions that must be separated or fails to merge regions that need not be separated. This is a consequence of the information about the uniformity in a region corresponding to an object surface and the discontinuity between regions corresponding to different object surfaces not being easily incorporated into the method.
The publication entitled "Seeded Region Growing" IEEE Trans. Pattern Anal.
Machine Intell., vol. 16 pp. 641-647, 1994 (hereinafter called Adams et al) discloses a method for segmentation of images. The Adams method is based on a region growing principle of selecting a pixel adjacent to a region of pixels, which is most similar to the region of pixels. The method does not rely on the arbitrary selection of homogeneity thresholds, but is controlled by choosing a small number of pixels, called seeds. This CFPL010WFqYPR23-26)477727 [O:\CISRA\IPR\IPR 231477727.doc:BFD seed selection may be either automatic or manual. Once the seeds have been selected, the segmented regions are grown in an iterative fashion. Adams suggest using an automatic converging squares method for seeds selection. Adams uses this method to locate objects of minimum and maximum intensity in biomedical images. Each step of the Adam method involves the addition of one of the neighboring pixels to one of the regions grown from the seeds. A measure 6(x) is defined how different each of the neighboring pixels is from that region. The neighboring pixel having the minimum measure 6(x) is added to the region. Adams et al make use of a sorted list in determining the relevant neighboring pixel to be added. In Adams et al, once a pixel has been added to the list, the 6(x) measure is never updated. However, this method is not successful for images where the number of regions is large and the regions have diverse characteristics. Moreover, the Adams et al method suffers from the disadvantage of slow image segmentation. Whilst, this method is robust and easy-to-use, it also suffers from the disadvantage that the resultant segmented image is stored as a pixel-map representation and as such is memory consuming and not efficient for feature computation.
Summary of the Invention It is an object of the present invention to ameliorate one or more disadvantages of the prior art.
20 According to one aspect of the invention, there is provided a method of segmenting 0 an image, the image comprising a plurality of pixels, wherein said method comprises the s t eps of: allocating a plurality of pixels as seeds; growing regions from said seeds so as to osegment the image into a number of regions, wherein a number of pixels that border said growing regions are considered and that pixel of said number that is most similar in a S 25 property to a region it borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
According to another aspect of the invention, there is provided a method of segmenting an image, the image comprising a plurality of pixels, wherein said method 30 comprises the steps of: distributing a plurality of seeds in areas of said image as a function 6 rty of said pixels within those areas, wherein fewer seeds are allocated to those S477727_amendments.doc 1 .1 -3said seeds so as to segment the image into a number of regions, wherein a number of pixels that border said growing regions are considered and that pixel of said number that is most similar in said property to a region it borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
According to another aspect of the invention, there is provided apparatus for segmenting an image, the image comprising a plurality of pixels, wherein said apparatus comprises: means for allocating a plurality of pixels as seeds; means for growing regions from said seeds so as to segment the image into a number of regions, wherein a number of pixels that border said growing regions are considered and that pixel of said number that is most similar in a property to a region it borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
According to another aspect of the invention, there is provided apparatus for segmenting an image, the image comprising a plurality of pixels, wherein said apparatus comprises: means for distributing a plurality of seeds in areas of said image as a function of a property of said pixels within those areas, wherein fewer seeds are allocated to those areas of the image having pixels homogeneous in said property; and means for growing regions from said seeds so as to segment the image into a number of regions, wherein a 20 number of pixels that border said growing regions are considered and that pixel of said number that is most similar in said property to a region it borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
S* •According to another aspect of the invention, there is provided a computer program S 25 product, including a computer readable medium having recorded thereon a computer program for segmenting an image, the image comprising a plurality of pixels, wherein said computer program product comprises: means for allocating a plurality of pixels as seeds; and means for growing regions from said seeds so as to segment the image into a number of regions, wherein a number of pixels that border said growing regions are 30 considered and that pixel of said number that is most similar in a property to a region it Sappended to that region and the said property of the appended region is updated 477727 amendments.doc and said growing step is repeated until no pixels bordering the growing regions are available.
According to another aspect of the invention, there is provided a computer program product, including a computer readable medium having recorded thereon a computer program for segmenting an image, the image comprising a plurality of pixels, wherein said computer program product comprises: means for distributing a plurality of seeds in areas of said image as a function of a property of said pixels within those areas, wherein fewer seeds are allocated to those areas of the image having pixels homogeneous in said property; and means for growing regions firom said seeds so as to segment the image into a number of regions, wherein a number of pixels that border said growing regions are considered and that pixel of said number that is most similar in said property to a region it borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
Brief Description of the Drawings Embodiments of the invention are described with reference to the drawings, in which: Fig. 1 is a flow chart of a method of segmenting an image in accordance with a preferred embodiment; S 20 Figs. 2A 2B illustrate an example of a segmented image and a corresponding region adjacency graph; 0o o oeo •o o 00 U o• U. U oeo U *o 477727_amendments.doc Fig. 3 is a flow chart of a method of segmenting an image as used in Fig. 1; Fig. 4A is a flow chart of a method of seeding an image as used in Fig. 3; Fig. 4B13, illustrates an example of a seeded image seeded according to Fig. 4A; Fig. 5A is a flow chart of a method of seeded region growing as used in Fig. 3; Fig. 5B illustrates a simplified example of the preferred region growing process of Fig. Fig. 6 is a flow chart of a method of coding a segmented image as used in Fig. 1; Fig. 7A is a flow chart of splitting a segmented image as used in Fig. 6; Figs. 7B to 7E illustrates an example of splitting a segmented image in accordance lo with the preferred splitting process of Fig. 7A; Fig. 8A is a flow chart of merging a segmented image as used in Fig. 7; Figs. 8B and 8C illustrate an example of merging the split segmented images of Fig.
7E in accordance with the preferred merging process of Fig. 8A; Fig. 9 is a flow chart of MRF merging the segmented image as used in Fig. 1; and Fig. 10 is a block diagram of a general purpose computer.
Detailed Description Where reference is made in any one or more of the accompanying drawings to steps g and/or features, which have the same reference numerals, those steps and/or features have 00:" 20 for the purposes of this description the same function(s) and/or operation(s), unless the 00 0: oo0 0contrary intention appears.
00The principles of the preferred method have general applicability to the segmentation of multiple colored or black and white images. The method has been 0 described with reference to a number of specific examples of images and it is not intended 25 that the invention be limited to such specific examples.
Overview of Preferred Method Fig 1 is a flow diagram of a method of segmenting an image in accordance with a preferred embodiment of the invention. The method commences at step 102, where image data is inputted and any necessary parameters are initialized. The image data is a pixelmap representation of an original image. The pixel-map includes a numerical representation of the particular color for each pixel location in a rectangular array. Any 477727new.doc -6numerical representation of color can be used and can be expressed as a sequence of one or more numbers. Preferably, the image data is inputted on a pixel by pixel basis in raster image order, i.e. from left to right across a display device and from top to bottom of the display device in a non-interlaced fashion. Pixel locations at the pixel grid are represented by an array of row and column specifications. Also, other information may be inputted at this stage, such as header information. Such header information can include the size of the image, i.e. array size (i x j).
In the next step 104, the inputted image is initially segmented into a number of regions. In this first stage, over-segmentation is preferred rather than under-segmentation.
Regions, which should not have been separated, can be merged later during the MRF merging step 110. Whereas regions, which should have been separated, cannot be separated later during the MRF merging step 110. The initial segmentation step 104 is not only fast but also accurate and produces a limited number of regions; i.e. the over splitting is moderate. The initial segmented image is outputted as a pixel-map representation. The preferred initial segmentation process is described in more detail in the section herein entitled "1.1 Process of Initial Segmentation After step 104, the processing continues at step 106.
In step 106, the initial segmented image is converted to a coded representation. This coded representation needs less memory and provides faster access for feature S 20 computation (See Table 1) than the pixel-map of the initial segmented image. The preferred embodiment stores the coded segmented image as a list of rectangles or squares instead of the pixel-map representation. The preferred conversion method is described in more detail in the section herein entitled "1.2 Process of Encoding Initial Segmented Image ooo 25 After step 106, the processing continues at step 108. In this step 108, the coded initial segmented image is represented as a region adjacency graph (RAG). The initial RAG is constructed from the initial segmentation result. Given an image that is segmented into a set of N disjoint regions such that R {Ri, 1_ i the RAG is constructed by denoting each region Ri as a node and connecting the regions which share a common boundary. Thus, a neighborhood system is defined on the graph, so that only spatially adjacent regions are neighbors. A clique is a subset of R such that it contains 477727new.doc either a single node or several nodes that are all neighbors of each other, and where C is the set of all the cliques.
Turning now to, Figs. 2A and 2B there is illustrated an example of a segmented image 200 and a corresponding region adjacency graph 202. The segmented regions s 1,2,3,4,5 and 6 of the image 200 are denoted as respective nodes 1,2,3,4,5 and 6 on the region adjacency graph 202. As can be seen, the couplings between the nodes on the graph 202 represent the common boundaries between the regions.
Returning to Fig. 1, the method continues at step 110 where a Markov Random Field (MRF) merging process is commenced. The RAG and coded initial segmented image are inputted to this step 110. This step 110 merges the segmented images in a systematic approach, wherein the number of segmented regions are reduced by taking into account region uniformity of such features as intensity, color texture etc. Discontinuities can also be taken into account by introducing a boundary process. The output of this merging process is a final segmented image. The preferred MRF merging process is 15 described in more detail in the section herein entitled "1.3 Markov Random Field (MRF) •Merging Process" ••:The processing of the image segmentation terminates at step 112, where the final segmented image is outputted.
20 1.1 Initial Segmentation Process Fig. 3 is a flow chart of a method of initially segmenting an image as used in step 104 of Fig. 1. The method commences at step 102, where the image data is inputted.. As mentioned above, the image data is a pixel-map representation of an original image. In °ooo the next step 304, a list of pixels is generated which are to be used as seeds for region see: 25 growing. The automatic selection of an appropriate set of pixels or set of small connected regions, called seeds controls the method of initially segmenting the image. The selection of the set of seeds or small regions is critical for the success of the image segmentation.
This can be done by using a quad-tree approach in order to distribute seeds according to a homogeneity criterion based on the contrast of the pixels. The preferred process for generating these seeds is described in more detail in the next section, herein entitled "1.1.0 Process for Selecting Seeds 477727new.doc In the next step 306, the process takes the generated seeds and grows seeded regions in an iterative fashion. At each iteration, all those pixels of the image that border the growing regions are considered. The pixel that is most similar to a region that it borders is appended to that region. Alternatively, the similarity of a limited number of these pixels may be considered at each iteration, thus speeding up the processing. The preferred process for growing the seeded regions is described in more detail in the section herein entitled "1.1.1 Process for Growing Seeded Regions This process continues until all pixels have been allocated to an associated region, resulting in a segmented image. The output of the seeded region growing is a set of homogeneous regions, wherein the number of regions obtained is equal to the number of seeds. During this step, the regions will continue to grow until they are bounded on all sides by other growing/grown regions.
Also, some regions will grow more at the expense of others. For instance, there will tend to be large regions in the homogeneous areas and small regions in the non-homogeneous areas. Furthermore, the contrast for each region is re-evaluated while the region grows. In S 15 this way, the preferred method is able to segment the image. After completion of this step 306, the processing terminates and returns to step 106.
1.1.0 Process for Selecting Seeds The seed selection process 304 is a simple and fast quad-tree approach, which S 20 distributes the seeds over the image, but it allocates fewer seeds in homogeneous areas of the image. The seed selection process address color data by preferably processing the oooo luminance image, i.e. a grey-scale image. The homogeneity is measured by a simple oo contrast criterion: the difference between the minimum and maximum luminance.
The following pseudocode is illustrative of the method of seeding an image for use a. 25 in Fig 3.
Pseudocode SEED RECTANGLE A rectangle, given by and (width,height) RECTANGLELIST FIFO list of rectangles SEED_LIST List ofpixels (seeds) 477727new.doc -9- CONTRAST Difference between min and max luminance HI MIN SIZE Maximum block size for contrast assessment LO_MIN_SIZE Minimum block size for block splitting HI_MIN_SIZE) HI_DENSITY, LO_DENSITY Densities for pixel spreading HITHRESHOLD, LOTHRESHOLD Contrast thresholds Initialize RECTANGLE_LIST with the rectangle corresponding to the whole image.
while RECTANGLE_LIST is not empty remove first element from RECTANGLE_LIST and keep it in RECTANGLE; assess CONTRAST for area of the image corresponding to RECTANGLE; if CONTRAST LO THRESHOLD S 15 add the pixel corresponding to the center of RECTANGLE to SEED_LIST; Scontinue loop; if RECTANGLE size HIMINSIZE split RECTANGLE into four and add the new rectangles in RECTANGLE LIST; continue loop if CONTRAST HITHRESHOLD spread pixels over RECTANGLE with LODENSITY add them to SEED LIST; continue loop if RECTANGLE size LO MIN SIZE split RECTANGLE into four and add the new rectangles in
RECTANGLE_LIST;
continue loop 477727new.doc spread pixels over RECTANGLE with HIDENSITY; add them to SEEDLIST; endwhile Turning now to Fig. 4, there is shown a flow chart of the last mentioned pseudocode called SEED. The seed selection processing commences at step 402, where rectangle coordinates corresponding to the entire image are stored in a FIFO buffer called RECTANGLE_LIST. After step 402, the processing continues at decision block 404, where a check is made whether the RECTANGLE_LIST is empty. If the decision block returns true then processing is terminated at step 406. Otherwise, the processing continues at step 408, where the first element in RECTANGLE_LIST is removed and stored in the variable rectangle.
In the next step 410, the contrast of the entire image or a sub-image thereof 15 corresponding to the removed rectangle is determined. The contrast is determined by S•calculating the difference between the minimum and maximum luminance values of the pixels in the rectangle. After step 410, the processing continues at decision block 412, :where a check is made whether the determined contrast is less than a predetermined low threshold value called LO THRESHOLD. If the decision block 412 returns true, then the S 20 co-ordinates of central pixel of the rectangle are added to a list called seed_list. In this way, sub-images corresponding to the rectangle which have a low contrast and which are of any size have a center as a seed (see Table If the decision block 412 returns false, the processing continues at decision block 416. In decision block 416, a check is made whether the size of the rectangle is greater than a predetermined constant called 25 HI_MINSIZE. If the decision block 416 returns true, then the processing continues at step 418. In step 418, the rectangle is divided into four sub-rectangles in the manner of a quadtree approach. In this way, sub-images corresponding to rectangles of a large size having medium and high contrast are split (see Table After step 418, the processing continues at step 404, where the four sub-rectangles are added to the FIFO buffer RECTANGLELIST. If however, the decision block 416 returns false, the processing continues at decision block 420.
477727new.doc -11- In the decision block 420, a check is made whether the determined contrast is less than a predetermined high threshold value called HI_THRESHOLD. If the decision block 420 returns true, then processing continues at step 422, where a number of pixels from the rectangle are added to the SEED_LIST as seeds. These newly added seed pixels are evenly distributed throughout the current rectangle in such a manner that there is a low density of such seed pixels in the rectangle. In this way, a low density seeding is achieved for sub-images of a small and medium size (See Table After step 422, the processing continues at step 404. If, however, the decision block 420 returns false then the processing continues at decision block 424.
In the decision block 424, a check is made whether the size of the rectangle is greater than a predetermined minimum size called LOMIN_SIZE. If the decision block returns false, the processing continues at step 426, where a number of pixels from the rectangle are added to the SEED_LIST as seeds. These newly added seed pixels are evenly distributed throughout the current rectangle in such a manner that there is a high 15 density of such seed pixels in the rectangle. In this way, a high density seeding is achieved for sub-images of a small size (See Table If, however, the decision block °424 returns true then the processing continues at step 428. In step 428, the rectangle is divided into four sub-rectangles in the manner of a quadtree approach. In this way, corresponding sub-images of a medium size and high contrast are split (see Table A).
e..
ease 6,06 TABLE A Allocation of seeds as a function of: the contrast of current sub-image, and *fee 25 the size of the rectangle corresponding to the current sub-image as compared to the size of the rectangle corresponding to the entire image Small size Medium size Large size Low contrast Center of rectangle is Center of rectangle Center of rectangle a seed is a seed is a seed Med contrast Low density seeding Low density seeding Split rectangle High contrast High density seeding Split rectangle Split rectangle 477727new.doc -12- Turning now to Table A, it can be seen that the split rectangular regions of the image of any size whose pixels have small variance in luminance (low contrast) are seeded in their center. In addition, split rectangular regions of a small or medium size whose pixels have a medium variance in luminance (medium contrast) are seeded evenly throughout these regions in a low density manner. Furthermore, rectangular regions of a small size whose pixels have a high variance in luminance (high contrast) are seeded evenly throughout the region in a high density manner. On the other hand, rectangular regions of medium size and high contrast are split into four rectangular sub-regions. In addition, rectangular regions of a large size and of a medium or high contrast are also split io into rectangular sub-regions. This splitting continues in a quadtree manner until the split sub-region(s) meets the abovementioned relevant size and contrast requirements for seeding. As will be apparent, these rectangular sub-images can be in some circumstances equilateral rectangular sub-images.
Turning now to Fig. 4B, there is a shown an example of a seeded image 452 seeded i s according to the preferred process 306. For simplicity's sake, the image itself is not shown. Initially, during the seeding process, the original image is inputted and its contrast .0 is determined. As the original image in this example has a medium contrast and is of a large size (as compared to itself), the image is split into four rectangles 454,456,458, and 460. The process then considers each of the these rectangles 454,456,458, and 460. As the 0 00 20 images, in this example, within rectangles 454,456, and 458 are of low contrast, and of a to 0: large size as compared to the original image, the centers of these rectangles are taken as seeds 462. However, as the image, in rectangle 460 is of a high contrast and large size, the rectangle is split further into four sub-rectangles 464,466,468, and 470. The process 600.
0.00 then considers each sub-rectangle 464,466,468, and 470. As rectangle 464 and 466 are oooo 25 both of a high contrast and medium size they are each further split into four subrectangles. As rectangle 468 is of a medium contrast and size, the rectangle is allocated seeds 472 in a low density manner. In addition, as rectangle 470 is of a low contrast and medium size, the center of this rectangle is taken as a seed 474. The seeding processing continues in a similar manner, until all split rectangles have been seeded. In this particular example, the split rectangles are center seeded 476,478,480,482,484, and 486 and the remaining split rectangles are allocated seeds 488, and 490 in a high density manner. At 477727new.doc 13the completion of the seeding process, a list of all the pixel locations of seeds 462,472,474,476,478,480,482,484,486, 488, and 490 is established.
The preferred seeding process is a simple and fast approach which distributes the seeds over the entire image, while allocating fewer seeds in homogeneous areas of the image. Furthermore, there is a high probability at least one seed will be allocated to each homogeneous region of the image.
1.1.1 Process for Growing Seeded Regions The seeded region growing process 306 takes a set of seeds, individual pixels or small groups of connected pixels, generated by step 304, as input. The preferred process 306 grows the seed regions in an iterative fashion. At each iteration, all those pixels that border the growing regions are considered. The pixel, which is most similar to a region that it borders, is appended to that region. In the preferred process, all the regions can be grown simultaneously.
15 The process evolves inductively from the seeds, namely, the initial state of the sets or regions A l
A
2 Each step of the iterative process involves the addition of one .i pixel to one of the above sets. We now consider the state of the sets after m steps.
Let T be the set of all as-yet unallocated pixels which border at least one of the regions.
ST= X jAi N(x) ,Ai 0, Eqn (1) where N(x) is the set of immediate neighbors of the pixel x. For each candidate pixel x, an index i(x) is found, which correspond to the adjacent region where x is most likely to be included and a criterion 6(x) is computed; 6(x) measures how good a candidate x is for region A,.
If, for x T we have that N(x) meets just one of the then we define i(x) e n} to be that index such that N(x) r Ai(x) 0 and define 6(x) to be a measure of how different x is from the region it adjoins. The simplest definition for 6(x) is g(x) mean[g(y)] Eqn (2) yrA(x) 477727new.doc -14where g(x) is the grey value of the image point x adjoining region Ai(x) and g(y) is the grey value of the image pointy within region Ai(x). The extension of this criterion to color images requires the choice of a suitable metric in color space. For example, the absolute value of the Euclidean distance between the color of pixel x and the mean color of region Ai(x) could be used. Alternatively, the segmentation could be performed on the luminance image as per on gray-scale images.
If N(x) meets two or more of the we take i(x) to be a value of i such that N(x) meets A, and 6(x) is minimized.
Then, a candidate pixel z e T is chosen such that ,o min{f6(x)}. Eqn (3) xET and append z to Ai(z).
This completes step m+l. The process is repeated until all pixels have been allocated. The process commences with each A i being just one of the seeds. The equations 2 and 3 ensure that the final segmentation is into regions as homogeneous as s15 possible given the connectivity constraint.
."In the preferred process 306, 6(x) is updated only for a limited number of the candidate pixels at each step of the iteration. Consequently, as the color of the limited number of candidate pixels is always compared with the updated mean color of the .°"neighboring regions, the quality of the segmentation is comparable with Adams et al.
S20 Furthermore, as the process does not consider all the candidate pixels, especially when the i' list is relatively long, the speed of the region growing process can be significantly increased without reducing the quality of the final segmentation. The region growing process 306 uses two ways, either alone or in combination, to avoid scanning the whole candidate pixels list.
The first one is to use a variable step when scanning the candidate pixel list, the value of the step depending on the size of the candidate list: the longer the list, the bigger the step. Another advantage of this method is a better control on the processing time (if there is a linear relation between the size of the candidate list and the step).
The second process consists in skipping a whole part of the list by choosing the first candidate pixel x such that 6(x) is smaller than z being the pixel selected at the 477727new.doc 4..e 4
S..
4. 4 4 4. 0 45 44 0 4 4464 4 4
S
4 4
S
previous step. If such a pixel is found, then the scanning of the list is interrupted prematurely; otherwise, the whole list is scanned to find the candidate pixel with the minimum 8(x) value, and the threshold is tuned to that value.
As successive best candidates often belong to the same region, inserting the new candidates (neighbors of the selected pixel) at the beginning of the list can reduce the computation time. However, they are not considered at the first step after their insertion in order not to introduce a bias in favor of a particular region.
The following pseudocode is illustrative of the preferred method of seeding an image for use in Fig 3.
Pseudo-code REGION GROWING SEED_LIST List of seeds (pixels) CANDIDATE_LIST List of pixels which are neighboring at least one region REGION Array used to store the growing regions i.e. the lists of 15 classified pixels MEAN Array containing the mean gray value or luminance color of the regions DELTA Function measuring the difference between a pixel and a neighboring region 20 MIN Variable used for storing the minimum DELTA CHOSEN_PIX Chosen pixel CHOSEN_REG Index of the chosen region DYNTHRESHOLD Dynamic threshold to allow early selection of a candidate pixel DYN_STEP Dynamic step for the scan loop of CANDIDATE_LIST 25 DYN_START Dynamic starting position for the scan loop Initialize each REGION with the corresponding seed of SEED_LIST and initialize CANDIDATE_LIST with the neighbors of each seed; DYN THRESHOLD 0 DYNSTART 0; while CANDIDATELIST is not empty 477727new.doc -16- Set DYN_STEP depending the size of CANDIDATELIST, e.g. DYN_STEP size of CANDIDATELIST 300 for i DYN START to size of CANDIDATE_LIST, i i+DYNSTEP CURRENT_PIX pixel i in CANDIDATE_LIST if (DELTA(CURRENT_PIX) MIN) MIN DELTA(CURRENT_PIX) CHOSEN PIX PROCES PIX CHOSEN_REG index of the chosen region stop for if MIN DYN_THRESHOLD endfor put each unclassified neighbor of CHOSENPIX in CANDIDATE_LIST and set DYN_START as the number of new pixels in CANDIDATE_LIST; Put CHOSEN_PIX in REGION[CHOSEN_REG]; Update MEAN[CHOSENREG] 15 Remove CHOSEN PIX from CANDIDATE_LIST; DYNTHRESHOLD max (DYN_THRESHOLD, MIN) 00 endwhile Turning now to Fig. 5A, there is shown a flow chart of the last mentioned pseudocode named REGION GROWING for growing the segmented regions. The 20 processing commences at step 502, where the seed list is stored in an array called
S
REGION[]. This array REGION[] is used to store the growing regions, i.e. the lists of classified pixels. Initially, the seeds of the image are denoted as the initial regions for S growing.
In the next step 504, the neighboring pixels of each seed are determined and stored :oo 25 in a list called CANDIDATELIST. In the next step 506, a variable DYN_THRESHOLD is set to zero. This variable stores a dynamic threshold to allow early selection of a candidate pixel in a manner, which will be explained below. After step 506 the processing continues at decision box 508, in which a check is made whether the CANDIDATE_LIST is empty. The CANDIDATELIST will be empty once there are no more pixels neighboring the growing regions. If the decision box 508 returns true then the processing continues at step 510 where the region growing process 306 is completed.
If the decision block 508 returns false then the processing continues at step 512.
477727new.doc -17- In step 512 the variable loop counter i is set to zero, the variable MIN is set to 256, and the variable DYN STEP is set to the current size of the CANDIDATE LIST divided by 300. The variable MIN is used for storing the minimum delta value of the previous iteration of the loop 508-512,..,and 536. The variable DYN_STEP is used for storing a variable step value used for scanning the CANDIDATE-LIST. This variable step value is used for determining the delta values for a limited number of candidates in the CANDIDATE-LIST. Specifically, only those candidates spaced apart by a value equal to the step value will be considered for allocation to the region. After step 512, the processing continues at decision box 514, where a check is made whether the loop lo counter is less than the size of the CANDIDATE LIST.
If the decision block 514 returns false, the processing continues at step 526, which is described below. If, however, the decision box 514 returns true then the region growing process has not considered all the limited number of neighboring pixels. In this situation the processing continues at step 516, where the variable CURRENT_PIX is set to pixel i Is in the CANDIDATELIST. This step 516 sets the next candidate pixel to be considered.
S.It should be noted that this pixel is spaced apart from the previous pixel considered by a distance equal to the value stored in the variable DYN STEP. After step 516, the processing continues at the decision box 518.
In decision box 518, a comparison is made whether the difference between the luminance value of the current candidate pixel and the mean of the neighboring region is less than MIN. If the decision box 518 returns false, then the processing continues at step 520. In step 520 the loop counter i is incremented by the step value stored in DYNSTEP. If the decision box 518, returns true, then the processing continues at step g: 522. In step 522, the MIN variable is now set to the minimum delta value determined for 25 the current pixel. In addition, the variable CHOSENPIX is set to the selected current pixel and the variable CHOSEN_REG is set to the index of the current region. After step 522, the processing continues at step 524.
In decision block 524, a comparison is made whether the current minimum delta value stored in MIN is less than the current value stored in DYN THRESHOLD. If the decision block 524 returns false then the processing continues at step 520, where the loop counter i is incremented by the step value stored in DYN_STEP. Otherwise, if the decision block 524 returns true then the processing continues at step 526. In step 526, 477727new.doc -18each pixel neighboring the current pixel stored in CHOSEN_PIX, and not previously stored in the CANDIDATE_LIST, is now added to the CANDIDATE_LIST. During step 528, the contrast of the region is updated and stored in REGION[CHOSEN_REG]. After step 528, the processing continues at step 530, where the current pixel stored in CHOSEN_PIXD( is added to the region which is stored in REGION[CHOSEN_REG]. At the next step 534, the current pixel stored in CHOSEN_PIX is removed from the candidates in CANDIDATE_LIST. The processing then continues at step 5636, where the variable is reset to the maximum value of the current values stored in MIN and DYN_THRESHOLD. After which, the processing returns to decision block 508. The process terminates when the CANDIDATE_LIST is empty.
The preferred growing process of Fig. 5A continues until all pixels have been allocated to an associated region, resulting in a segmented image. The output of the seeded region growing is a set of homogeneous regions, wherein the number of regions obtained is equal to the number of seeds. The regions will continue to grow until they are s15 bounded on all sides by other growing/grown regions. Also, some regions will grow more at the expense of others. For instance,, there will tend to be large regions in the 0 homogeneous areas and small regions in the non-homogeneous areas. Furthermore, the S: contrast for each region is re-evaluated while the region grows. In this way, the preferred method is able to segment the image.
Fig. 5B illustrates a simplified example of the preferred region growing process.
For simplicities sake, this example shows only one region of the region growing process, :whereas the preferred method allows the simultaneous growing of multiple regions. An initial region 550 consisting of a plurality of pixels (not shown) is surrounded by a 0:00 number of candidate pixels 552 to be added to the region 550. Firstly, the process 25 calculates the mean of the luminance values of the pixels of the region 550. Then the process determines the difference between this mean and the luminance value of a limited number of candidate pixels 552 in turn. The process then determines the minimum difference of these differences and allocates the candidate pixel associated with this minimum difference to the region 550. If however, the luminance difference value of any candidate pixel is less than the minimum difference value in the previous iteration, then the process instead allocates this candidate pixel to the region 550 and then proceeds to 477727new.doc -19the next iteration. In the next iteration, the mean value of the luminance values of the pixels of the grown region 550 is then recalculated and the process continues.
1.2 Process of Encoding Segmented Image Fig 6 is a flow diagram of the method 106 of encoding the segmented image utilized in Fig. 1. The input data to the method 106 is a pixel-map representation of the image which has previously been segmented into regions in step 104, each region being defined as a set of connected pixels. In the method 106 a different name is assigned to each segmented region, which can be done arbitrarily. The method 106 comprises two consecutive steps: splitting 604 each region's shape into non-overlapping rectangular blocks; and merging 606 rectangular blocks which belong to the same region in order to limit their number. The step 606 outputs to step 108 a coded representation of the segmented image. The regions of the segmented image are defined as a list of distinct rectangles. Specifically, each region of any arbitrary shape, is partitioned into a set of non-overlapping blocks which cover its entire area. The preferred method 106 limits the number of blocks used for each region. The coded representation is outputted in the form S.of a set co-ordinates of the vertices of the rectangles together with their associated color o. information. In this way, the storage requirements for the segmented image are minimized. Moreover, since it is faster to iterate through a list of blocks than through a list of pixels, the access to the pixels of the region for feature computation is faster. The preferred process 604 for splitting the segmented image is described in more detail in the section herein entitled "1.2.1 Process of Splitting the Segmented Image". The preferred process 606 for merging the split regions is described in more detail in the section herein entitled "1.2.2 Process for Merging the Split Regions" oo 1.2.1 Process of Splitting the Segmented Image The splitting step is based on a quad-tree splitting approach. The input image, i.e.
the initial block, is divided into 4 rectangular subblocks and each block is recursively split into 4 subblocks until the blocks are included (or mostly included) in a region or reach a minimum size. This implies finding all the regions which are represented in each block and the area of overlapping between the block and those regions. This search can be limited by using the following scheme.
477727new.doc Beforehand, all the region names are stored in a list. The initial list of all the names is associated with the initial block. Then a reduced list of names will be associated with each block, which consists of the names of the regions which can be found in the block; those names are the names of the regions which were previously found in the parent block. Thus, the search of the regions which are present in a given block is not too long.
The following pseudo-code is illustrative of the process 604 of splitting the segmented image into regions as used in Fig. 6.
pseudocode SPLIT lo Variables MIN_SIZE CONSTANT: Minimum size for rectangles minimum height and width).
DELTA CONSTANT: If more than DELTA percent of a rectangle is covered by a single region, then the rectangle is considered to be included in the region (DELTA is a percentage between 0.5 and 1).
NOREGIONS Constant: Number of regions in the image.
RECT A rectangle.
SoNAMELIST A list of region names.
RECT_QUEUE FIFO queue containing rectangles to be processed.
:ii 20 NAMEQUEUE FIFO queue containing the lists of possible region names for each rectangle the names of the regions which can be found in the rectangle) in RECT_QUEUE.
REGION[] OUTPUT: An array of lists of rectangles, of length SNOREGIONS. Each list of rectangles corresponds to the n. 25 rectangles which are included in a specific region of the image.
oooo (The regions are ordered arbitrarily.) Procedure Initialise RECT_QUEUE with the rectangle corresponding to the whole image and NAME_QUEUE with the list of all the region names.
while (RECTQUEUE is not empty) 477727new.doc -21 remove first element from RECTQUEUE and keep it in RECT remove first element from NAMEQUEUE and keep it in NAMELIST if (RECT is bigger than MIN_SIZE) if (there is a region of NAME_LIST which takes part in, at least, DELTA of RECT) add RECT to REGION[index of that dominant region's name] else split RECT, add the four new rectangles into RECTQUEUE, and add the new name list made of the name of the regions which actually were in RECT in NAME_QUEUE.
else find the dominant region in RECT and add RECT to REGION[index of the region's name] Turning now to Fig. 7A, there is shown a flow chart of the last mentioned pseudo code called SPLIT for splitting the segmented image. The processing commences at step 104. In the next step 704, the entire image is stored in a FIFO queue called RECT-queue S 15 for processing. In addition, the names of the regions which can be found in the current Srectangle are stored in a FIFO queue called NAMEQUEUE. After step 706, the S•processing continues at decision box 708.
S••In decision box 708, a check is made whether the FIFO queue called RECTQUEUE is empty. If the decision box returns true, then the processing continues S 20 at step 710 where the splitting process terminates and continues with the merging process 606. If the decision box 708 returns false, the processing continues at step 710. In step 710 a variable called RECT is set to the first rectangle in the RECT_QUEUE.
Subsequently this rectangle is then removed from the queue called RECT_QUEUE. In the next step 712 the first name in the NAMEQUEUE is retrieved and stored in the 25 NAMELIST. The name is then deleted from the NAMEQUEUE Afterwards the processing continues at decision box 714.
In decision box 714, a check is made whether the size of the rectangle is greater than a minimum size MIN-SIZE. If the decision box 714, returns false then the processing continues to step 716. In step 716, the dominant region within the rectangle is determined and this rectangle is added to a list of rectangles and indexed by the dominant regions name. The dominant region is preferably the largest region within the rectangle.
After step 716, the processing returns to decision box 708.
477727new.doc 22- If the decision box 714 returns true, then the processing continues at decision box 718. In decision box 718 a check is made whether there exists a region REGION[i] on the NAMELIST whose intersection with the rectangle is greater than the value DELTA times the size of the rectangle. It is in this way, the process determines the dominant segmefited region within the rectangle. If the decision box 718, returns true then the processing continues at step 720. In step 720, the rectangle is added to the list of rectangles REGION[]. The rectangle is added to the REGION[i] array at that location in the array corresponding to the index of dominant regions name. After step 720 the processing returns to decision block 708.
If the decision box 718 returns false, then the processing continues at step 722. In step 722, the rectangle is split into four sub-rectangles. These four sub-rectangles are then added to the FIFO queue RECT_QUEUE and the names of the regions which were in the previous rectangle are added to the FIFO queue NAMEQUEUE. After step 722 the processing returns to decision block 708.
Fig 7B to 7E illustrate an example of the splitting process 604 used in Fig. 6.
Initially a segmented image 750 is inputted. The segmented image is in the form a pixel map representation having a number of segmented regions 752, 754, 756, and 758 having the same associated color data. In the first step, the whole segmented image is input as the current rectangle. Next a check is made whether the dominant segmented region, that is 20o the largest segmented region in the current rectangle, is greater than 95% the current rectangle's size. In the present example, this check reveals that the segmented regions S.752, 754, 756 and 758 are less than 95% the size of the whole image. In the next step the process divides the segmented image into four rectangles 760,762,764, and 766. The process again makes the same check commencing on rectangle 760. In the present :2 5 example, the check reveals the dominant region 752 is greater than 95% of the current rectangle size and the vertices of this rectangle are stored against this dominant region 752. The same check is made against rectangle 766, which reveals no dominant region and as such the rectangular is further split into four sub-rectangles 768,770,772, and 774.
This processing continues until all of split sub-rectangles are allocated to one of the segmented regions 752,754,756, and 758 of Fig. 7E. In the event, the splitting continues below a predetermined size MIN_SIZE, the most dominant segmented region in this case is allocated the rectangle. Preferably, MIN_SIZE is greater than a lxl pixel block, but not 477727new.doc 23 too large either. After the processing is completed, the segmented regions 752,754,756, and 758 are each represented by vertices of a series of rectangles.
1.2.2 Process for Merging the Split Regions The second step 606 of the encoding process 106 consists in merging adjacent rectangles in each region in order to minimise the number of rectangles representing a region. Two rectangles are considered adjacent if they share a common edge; thus, their reunion is still a rectangle.
The rectangle lists are still stored in the array REGION[] which comes from the 1o splitting step 604. Adjacent rectangles to a specific rectangle are found by scanning the rectangle list and picking the first adjacent rectangle found. The pseudocode for performing the preferred merging is shown below.
*for (i 0 to i NOREGIONS 1) count 0 e:oo while (count number of rectangles in REGION[i]) if (there is an adjacent rectangle of rectangle count, in region i) merge both rectangles, add the result in REGION[i] and remove them from the list S"else count++ 20 Turning now to Fig. 8A, there is shown a flowchart of the last mentioned pseudo code for merging the rectangles. After the splitting step 604, the processing commences at step 804, where a loop counter called i is set to zero. After which processing continues at the decision box 806. In the decision box 806 a check is made whether the loop counter i is less than the constant NO REGIONS-1. If the decision box 806 returns false, 25 the processing terminates at step 808.
If the decision box 806 returns true then the processing continues at step 810 where the variable called COUNT is set to zero, after which the processing continues at decision box 812. In decision box 812, a check is made whether the variable count is less than the number of rectangles in REGION[i]. If the decision box 812 returns false then the processing continues at step 814 where the loop counter i is incremented by one and the processing is returned to decision box 806.
477727new.doc -24- The decision box 812 returns true then the processing continues at the decision box 814. In decision box 814, a check is made whether there exists an rectangle adjacent to the rectangle COUNT in REGION[i]. If the decision box 814 returns false. then the processing continues at step 816 where the variable COUNT is incremented by one.
After step 816 the processing is returned to decision box 812..
If the decision box 814 returns true then the processing continues at step 818. In step 818 both the adjacent rectangles are merged together and the result is listed in the REGION[i] at the end of the list. In addition, both the rectangles are removed from the list of rectangles stored in REGION[i] in step 820. After step 820, the processing returns to decision box 812.
Turning now to 8B, there is shown an example of merging the split segmented images of Fig. 7E in accordance with the preferred merging process of Fig. 8A. The primary step includes merging the adjacent rectangles within each segmented region 752,754,756,and 758. Fig. 8B illustrates the merged adjacent regions. The output for this 15 process is a list of the vertices of rectangles for each of the segmented regions 752, 754, 756, and 758. Thus instead of storing against each segmented region each pixel within the segmented region, the preferred process stores the vertices of a series of rectangles that make up that segmented region.
S: 20 1.3 Markov Random Field (MRF) Merging Process The preferred merging process 110 is based on a Markov Random Field model.
In the preferred Markov Random Field model, X {Xi, 15 i N} denotes a set of :random variables where Xi is associated with region Ri of the segmented image, and A= Xm} is a set of possible labels such that Xi e A for all i.
25 For the MRF segmenter, the labels correspond to region names (m In the initial segmentation obtained from the initial segmentation 104, the image is assumed to be over-segmented. The MRF segmenter iteratively updates the region names so that similar regions get the same name and are eventually merged.
In this model o (Xi xl, X2 XN xN) is one of all instances from the configuration space. The vector X is a Markov random field if: 1. P(o) 0, for all o; 2. P(Xi xi Xj xj, Rj Ri) P(Xi xi Xj xj, Rj e Gi), 477727new.doc where P(o) and I are the joint and conditional probability density functions (pdf) respectively, and Gi is the set of all the nodes in R which are neighbors of Ri. Intuitively, the MRF is a random field with the property that the statistic at a particular node depends only on that of its neighbors.
The pdf of the MRF has a general form, known as the Gibbs distribution: P(co)= Z- Eqn (4) where Z is a normalizing constant and U(o) an energy function.
The function U(o) is obtained by summing some clique functions Vc(o): U(o) Vc(o), Eqn cEC Vc(co) is a clique function which depends only on the variables Xi such that the region Ri belongs to clique c.
The MRF segmenter is preferably based on a region process, i.e. it merges regions based on their similarity. Since there should exist a discontinuity on the common boundary between two different segmented regions, the MRF segmenter could also use a 15 boundary process.
The clique functions are defined based on the constraint that a segmented region should be uniform in a set of features such as intensity, color, texture, etc. A list of features that can be used is shown in Table B.
*"k In this MRF model, F, denotes the k feature for region Ri and M the number of features. Cliques which contain two regions are only considered and the clique functions are defined as: SVc(O) ij max Fik FF Eqn (6) IR,Rj}eC where r)ij is a binary variable which has the value 1 if Xi and Xj have the same region label, and the value 0 otherwise.
All the features are normalized using the mean and standard deviation computed for all the regions of the initial RAG, wherein the mean for each feature is zero and the STD is one. This allows the merging method to consider only the most discriminant feature for each region pair, as shown in equation 477727new.doc -26- Table B. Description of the features measured for a region. Cov 3x3 covariance matrix; lumx normalized luminance at pixel x; Nx set of 4 neighbors of pixel x.
Feature Definition r Mean R (R+G+B) g Mean G (R+G+B) b Mean B (R+G+B) Cl Cov[0,0] C2 Cov[0,1] C3 Cov[0,2] C4 Cov[1,1] C6 Cov[1,2] MLum Mean luminance SDLum Standard deviation of luminance r-b r-g saturation [max(r,g,b) min(r,g,b)] max(r,g,b) smoothness Ilum lum 1 lum lum area(Ri) xeRi yeN x The segmented image is merged by optimizing of the MRF model. MAP estimators aim at finding the configuration of the random field which maximizes the probability distribution eqn i.e. minimizing the energy function eqn The minimization method used in the preferred merging process is a deterministic relaxation process. The binary clique which has the smallest Vc(co), i.e. the smallest difference in features, is updated first: one of the regions takes the other region's name and both regions are merged. Then, the segmenter keeps updating the clique which has the smallest clique function until Vc(co) is bigger than a predefined energy threshold Te.
Thus, regions having features which are different stay separated.
Once two regions are merged they can't be split again. After each merging, the energy needs to be updated only in the vicinity of the two regions which have just been merged.
477727new.doc -27- Fig 9 is a flow chart of the preferred method of merging the regions of the segmented images as used in Fig. 1. After RAGs of the segmented regions have been built by the step 108, the merging process commences. In step 902, clique functions Vc(co) for all pairs of neighboring segmented regions. In the next step 904, the pair of segmented regions are selected which has the smallest clique function value. The processing continues at decision block 906, where a check is made whether this clique function value is less than a predetermined threshold value. If the decision block 906 returns true, then the processing continues at step 908. In step 908, both selected regions are merged and the features for merged regions are recomputed. In addition, the clique functions of the merged region with neighboring regions is updated. The merging process continues until the decision block 906 determines clique functions are greater than or equal to the predetermined threshold value. The merging process then terminates at step 112, where the final segmented image is outputted.
This merging process allows a good trade offbetween accuracy and speed.
Preferred Embodiment of Apparatus(s) The image segmentation processes are preferably practiced using a conventional general-purpose computer, such as the one shown in Fig. 10, wherein the processes of Figs. 1 to 9 may be implemented as software executing on the computer. In particular, S 20 the steps of the image segmentation method are effected by instructions in the software that are carried out by the computer. The software may be divided into two separate parts; one part for carrying out the image segmentation methods; and another part to o: manage the user interface between the latter and the user. The software may be stored in a computer readable medium, including the storage devices described below, for example.
The software is loaded into the computer from the computer readable medium, and then executed by the computer. A computer readable medium having such software or computer program recorded on it is a computer program product. The use of the computer program product in the computer preferably effects an advantageous apparatus for image segmentation in accordance with the embodiments of the invention.
The computer system 1000 consists of the computer 1002, a video display 1016, and input devices 1018, 1020. In addition, the computer system 1000 can have any of a number of other output devices including line printers, laser printers, plotters, and other reproduction devices connected to the computer 1002. The computer system 1000 can be Sconnected to one or more other computers via a communication interface 1008c using an 477727new.doc -28appropriate communication channel 1030 such as a modem communications path, a computer network, or the like. The computer network may include a local area network (LAN), a wide area network (WAN), an Intranet, and/or the Internet The computer 1002 itself consists of a central processing unit(s) (simply referred to as a processor hereinafter) 1004, a memory 1006 which may include random access memory (RAM) and read-only memory (ROM), input/output (IO) interfaces 1008a, 1008b 1008c, a video interface 1010, and one or more storage devices generally represented by a block 1012 in Fig. 10. The storage device(s) 1012 can consist of one or more of the following: a floppy disc, a hard disc drive, a magneto-optical disc drive, CD- ROM, magnetic tape or any other of a number of non-volatile storage devices well known to those skilled in the art. Each of the components 1004 to 1012 is typically connected to one or more of the other devices via a bus 1014 that in turn can consist of data, address, and control buses.
The video interface 1010 is connected to the video display 1016 and provides video 'i 15 signals from the computer 1002 for display on the video display 1016. User input to operate the computer 1002 can be provided by one or more input devices 1008b. For o *example, an operator can use the keyboard 1018 and/or a pointing device such as the mouse 1020 to provide input to the computer 1002.
The system 1000 is simply provided for illustrative purposes and other configurations can be employed without departing from the scope and spirit of the :invention. Exemplary computers on which the embodiment can be practiced include IBM-PC/ATs or compatibles, one of the Macintosh (TM) family of PCs, Sun Sparcstation or the like. The foregoing are merely exemplary of the types of computers with which the embodiments of the invention may be practiced. Typically, the processes of the embodiments, described hereinafter, are resident as software or a program recorded on a hard disk drive (generally depicted as block 1012 in Fig. 10) as the computer readable medium, and read and controlled using the processor 1004. Intermediate storage of the program and pixel data and any data fetched from the network may be accomplished using the semiconductor memory 1006, possibly in concert with the hard disk drive 1012.
In some instances, the program may be supplied to the user encoded on a CD-ROM or a floppy disk (both generally depicted by block 1012), or alternatively could be read by the user from the network via a modem device connected to the computer, for example.
Still further, the software can also be loaded into the computer system 1000 from other computer readable medium including magnetic tape, a ROM or integrated circuit, a 477727new.doc -29magneto-optical disk, a radio or infra-red transmission channel between the computer and another device, a computer readable card such as a PCMCIA card, and the Intemrnet and Intranets including email transmissions and information recorded on websites and the like. The foregoing are merely exemplary of relevant computer readable mediums. Other computer readable mediums may be practiced without departing from the scope and spirit of the invention.
The methods of image segmentation may alternatively be implemented in dedicated hardware such as one or more integrated circuits performing the functions or sub functions of the image segmentation. Such dedicated hardware may include graphic processors, digital signal processors, or one or more microprocessors and associated memories.
The foregoing only describes a small number of embodiments of the present invention, however, modifications and/or changes can be made thereto without departing from the scope and spirit of the invention. The present embodiments are, therefore, to be 15 considered in all respects to be illustrative and not restrictive.
In the context of this specification, the word "comprising" means "including principally but not necessarily solely" or "having" or "including" and not "consisting only of'.
Variations of the word comprising, such as "comprise" and "comprises" have corresponding meanings.
•0o o..o oooo 9 9000 00 0 0 477727new.doc

Claims (23)

1. A method of segmenting an image, the image comprising a plurality of pixels, wherein said method comprises the steps of: allocating a plurality of pixels as seeds; growing regions from said seeds so as to segment the image into a number of regions, wherein a number of pixels that border said growing regions are considered and that pixel of said number that is most similar in a property to a region it borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
2. A method as claimed in claim 1, wherein said property of said pixels is luminance.
3. A method as claimed in claim 1 or 2, wherein said growing step comprises the sub- steps of: generating a list of pixels that border the growing regions; scanning a said number of said list of pixels in a predetermined manner; determining a value, for each said scanned pixel, indicative of the similarity of a said property of said scanned pixel and the corresponding said property of a growing oooo 20 region that said scanned pixel borders; selecting a pixel that has a minimum said value; appending said selected pixel to said region it borders; updating the corresponding said property of the appended region; and repeating the sub-steps of the growing step until the image is segmented. S
4. A method as claimed in any one of the preceding claims, wherein said selection step oooo selects the first scanned pixel having a said value below a minimum threshold.
5. A method as claimed in any one of the preceding claims, wherein said selection step 30 selects a pixel having a said value which is the mimimum of all said scanned pixels. 477727_amendments.doc -31
6. A method as claimed in any one of the preceding claims, wherein said property is the grey-value of said pixels.
7. A method as claimed in claim 3, wherein said value is determined in accordance with the following: S(x) g(x) mean[g(y)], where g(x) is the grey value of the image point x adjoining region Ai(x) and g(y) is the grey value of the image point y within region Ai(x).
8. A method as claimed in claim 3, wherein said value is determined in accordance with a metric in color space.
9. A method as claimed in any one of the preceding claims, wherein the method further comprises a step of merging said grown regions which have similarilities. A method as claimed in claim 9, wherein said merging step comprises the sub-steps of: determining for each pair of neighboring grown regions a clique function value representative of the similarity of said property of said pair of neighboring grown regions; O 20 selecting the pair of grown regionswith the smallest clique function value; merging both said selected regions to produce a merged region and updating the merged region's clique functions with neighboring regions, if said smallest clique S• •function value is less than a predetermined threshold; repeating the sub-steps of the merging step until said smallest clique function value 25 is greater than or equal to said threshold.
11. A method of segmenting an image, the image comprising a plurality of pixels, wherein said method comprises the steps of: *o distributing a plurality of seeds in areas of said image as a function of a property of 3 o•ixels within those areas, wherein fewer seeds are allocated to those areas of the image ving pixels homogeneous in said property; and 477727_amendments.doc -32- growing regions from said seeds so as to segment the image into a number of regions, wherein a number of pixels that border said growing regions are considered and that pixel of said number that is most similar in said property to a region it borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
12. A method as claimed in claim 11, wherein said distribution step comprises the sub- steps of: dividing the image into a plurality of areas; allocating, for each divided area, one or more seeds as a function of said property within the divided area and the size of the divided area as compared to the size of the image; and storing the locations of each allocated seed.
13. A method as claimed in claim 12, wherein said allocation step comprises the sub- steps of: selecting the image or a previously divided area of the image as the current area; seeding the center of said current area when the difference between the maximum and minimum values of said property of the pixels within the current area is less than a S 20 first predetermined threshold; subdividing said current area when said size of said current area is greater than a 9. second predetermined threshold and said difference is greater than or equal to said first •predetermined threshold; •uniformly seeding said current area in a low density manner when said size of said area is less than or equal to said second predetermined threshold and said difference is greater than or equal to said first predetermined threshold and less than a third too: predetermined threshold; 4 •subdividing said current area when said size of said area is greater than a fourth predetermined threshold and less than or equal to said second predetermined threshold 30 and said difference is greater than or equal to said third predetermined threshold; uniformly seeding said current area in a high density manner when said size of said 477727_amendments.doc -33- area is less than or equal to a fourth predetermined threshold and said difference is greater than or equal to said third predetermined threshold; and repeating the sub-steps of the allocation step until all of said divided areas are seeded.
14. A method as claimed in claim 13, wherein said subdividing steps comprise subdividing said current area into four areas. A method as claimed in any one of the preceding claims 11 to 14, wherein said property of said pixels is luminance.
16. A method as claimed in any one of the preceding claims 11 to 14, wherein said growing step comprises the sub-steps of: generating a list of pixels that border the growing regions; scanning a said number of said list of pixels in a predetermined manner; determining a value, for each said scanned pixel, indicative of the similarity of a said property of said scanned pixel and the corresponding said property of a growing o° region that said scanned pixel borders; selecting a pixel that has a minimum said value; S. 20 appending said selected pixel to said region it borders; •"updating the corresponding said property of the appended region; and oo• repeating the sub-steps of the growing step until the image is segmented. t' o o: so.* ooo•: 17. A method as claimed in claim 16, wherein said selection step selects the first 25 scanned pixel having a said value below a minimum threshold.
18. A method as claimed in claim 16, wherein said selection step selects a pixel having a said value which is the mimimum of all said scanned pixels.
19. A method as claimed in any one of the preceding claims 11 to 18, wherein the A4 l d further comprises a step of merging said grown regions which have similarilities. 477727_amendments.doc -34- A method as claimed in any one of the preceding claims 11 to 19, wherein the method further comprises the steps of: splitting said image into a plurality of rectangular areas, wherein each rectangular area comprises a said region or part thereof forming a dominant portion of the rectangular area; merging said rectangular areas which have a common said dominant portion and share a common edge; and outputting the vertices of the merged rectangular areas as a representation of the segmented image.
21. A method as claimed in claim 19, wherein said merging step comprises the sub- steps of: determining for each pair of neighboring grown regions a clique function value representative of the similarity of said property of said pair of neighboring grown regions; S 15 selecting the pair of grown regions with the smallest clique function value; merging both said selected regions to produce a merged region and updating the merged region's clique functions with neighboring regions, if said smallest clique K °function value is less than a predetermined threshold; repeating the sub-steps of the merging step until said smallest clique function value 2o is greater than or equal to said threshold. 9o .9 22. Apparatus for segmenting an image, the image comprising a plurality of pixels, wherein said apparatus comprises: means for allocating a plurality of pixels as seeds; 25 means for growing regions from said seeds so as to segment the image into a number of regions, wherein a number of pixels that border said growing regions are considered and that pixel of said number that is most similar in a property to a region it borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available. 477727_amendments.doc
23. Apparatus for segmenting an image, the image comprising a plurality of pixels, wherein said apparatus comprises: means for distributing a plurality of seeds in areas of said image as a function of a property of said pixels within those areas, wherein fewer seeds are allocated to those areas of the image having pixels homogeneous in said property; and means for growing regions from said seeds so as to segment the image into a number of regions, wherein a number of pixels that border said growing regions are considered and that pixel of said number that is most similar in said property to a region it borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
24. A computer program product, including a computer readable medium having recorded thereon a computer program for segmenting an image, the image comprising a plurality of pixels, wherein said computer program product comprises: means for allocating a plurality of pixels as seeds; and means for growing regions from said seeds so as to segment the image into a go• i: o• number of regions, wherein a number of pixels that border said growing regions are considered and that pixel of said number that is most similar in a property to a region it *i 20 borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
25. A computer program product, including a computer readable medium having S 25 recorded thereon a computer program for segmenting an image, the image comprising a plurality of pixels, wherein said computer program product comprises: means for distributing a plurality of seeds in areas of said image as a function of a property of said pixels within those areas, wherein fewer seeds are allocated to those areas of the image having pixels homogeneous in said property; and means for growing regions from said seeds so as to segment the image into a e J4er of regions, wherein a number of pixels that border said growing regions are $cons1i and that pixel of said number that is most similar in said property to a region it 477727 amendments.doc -36- borders is appended to that region and the said property of the appended region is updated and said growing step is repeated until no pixels bordering the growing regions are available.
26. A method of segmenting an image, the method substantially as described herein with reference to Fig. 5A of the accompanying drawings.
27. Apparatus for segmenting an image, the apparatus substantially as described herein with reference to Fig. 5A of the accompanying drawings.
28. A computer program product, including a computer readable medium having recorded thereon a computer program for segmenting an image, the computer program product substantially as described herein with reference to Fig. 5A of the accompanying drawings. DATED this twenty-first Day of February, 2002 i Canon Kabushiki Kaisha Patent Attorneys for the Applicant SPRUSON FERGUSON oo S.* *go oo *o oo* 477727_amendments.doc
AU52620/99A 1998-10-02 1999-10-01 Method and apparatus for segmenting images Ceased AU746669B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU52620/99A AU746669B2 (en) 1998-10-02 1999-10-01 Method and apparatus for segmenting images

Applications Claiming Priority (9)

Application Number Priority Date Filing Date Title
AUPP6344A AUPP634498A0 (en) 1998-10-02 1998-10-02 Method and apparatus for encoding segmented images
AUPP6344 1998-10-02
AUPP6343A AUPP634398A0 (en) 1998-10-02 1998-10-02 Method and apparatus for seeding an image
AUPP6342A AUPP634298A0 (en) 1998-10-02 1998-10-02 Method and apparatus for segmenting images
AUPP6342 1998-10-02
AUPP6341A AUPP634198A0 (en) 1998-10-02 1998-10-02 Method and apparatus for segmenting images
AUPP6341 1998-10-02
AUPP6343 1998-10-02
AU52620/99A AU746669B2 (en) 1998-10-02 1999-10-01 Method and apparatus for segmenting images

Publications (2)

Publication Number Publication Date
AU5262099A AU5262099A (en) 2000-04-06
AU746669B2 true AU746669B2 (en) 2002-05-02

Family

ID=27507020

Family Applications (1)

Application Number Title Priority Date Filing Date
AU52620/99A Ceased AU746669B2 (en) 1998-10-02 1999-10-01 Method and apparatus for segmenting images

Country Status (1)

Country Link
AU (1) AU746669B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1771008A2 (en) 2005-09-28 2007-04-04 Novell, Inc. Adaptive method and system for encoding digital images for the internet

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110782463B (en) * 2019-10-31 2022-06-24 京东方科技集团股份有限公司 Method and device for determining division mode, display method and equipment and storage medium
CN112069953B (en) * 2020-08-26 2021-10-29 江苏农林职业技术学院 Automatic identification method and device for rice seedling growth period
CN115239746B (en) * 2022-09-23 2022-12-06 成都国星宇航科技股份有限公司 Object-oriented remote sensing image segmentation method, device, equipment and medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5572258A (en) * 1993-09-28 1996-11-05 Nec Corporation Motion compensation estimating device and method achieving improved motion compensation
US5748761A (en) * 1995-04-08 1998-05-05 Daewoo Electronics Co., Ltd. Method for segmenting and estimating a moving object motion

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5572258A (en) * 1993-09-28 1996-11-05 Nec Corporation Motion compensation estimating device and method achieving improved motion compensation
US5748761A (en) * 1995-04-08 1998-05-05 Daewoo Electronics Co., Ltd. Method for segmenting and estimating a moving object motion

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1771008A2 (en) 2005-09-28 2007-04-04 Novell, Inc. Adaptive method and system for encoding digital images for the internet
EP1771008A3 (en) * 2005-09-28 2009-04-01 Novell, Inc. Adaptive method and system for encoding digital images for the internet
US7602984B2 (en) 2005-09-28 2009-10-13 Novell, Inc. Adaptive method and system for encoding digital images for the internet
US7826616B2 (en) 2005-09-28 2010-11-02 Novell, Inc. Adaptive method and system for encoding digital images for the internet
EP2458872A1 (en) 2005-09-28 2012-05-30 Oracle International Corporation Adaptive method and system for encoding digital images for the internet

Also Published As

Publication number Publication date
AU5262099A (en) 2000-04-06

Similar Documents

Publication Publication Date Title
US7031517B1 (en) Method and apparatus for segmenting images
US6766037B1 (en) Segmenting moving objects and determining their motion
JP5008572B2 (en) Image processing method, image processing apparatus, and computer-readable medium
US6993185B2 (en) Method of texture-based color document segmentation
US5280547A (en) Dense aggregative hierarhical techniques for data analysis
JP3748172B2 (en) Image processing device
CN100469099C (en) System and method for connecting regions of image data having similar characteristics
US8542923B2 (en) Live coherent image selection
US6728400B1 (en) Apparatus, method, and storage medium for setting an extraction area in an image
JP2001101426A (en) Digital image processing method
CN109345536B (en) Image super-pixel segmentation method and device
US7366348B2 (en) Segmentation of images using the watershed method
JP4149464B2 (en) Image processing device
AU746669B2 (en) Method and apparatus for segmenting images
US6608929B1 (en) Image segmentation apparatus, method thereof, and recording medium storing processing program
US5231676A (en) Hierarchical operations on border attribute data for image regions
JP4756436B2 (en) Pattern recognition apparatus, pattern recognition method, and pattern recognition program
AU735807B2 (en) Segmenting moving objects and determining their motion
JP2005302056A (en) Pattern extracting apparatus
Kovalevsky et al. Image segmentation and connected components
JPH10255036A (en) Dynamic binarizing method for gray level image
Ngọc et al. A minimum barrier distance for multivariate images with applications
Nickels et al. Textured image segmentation: Returning multiple solutions
AU742875B2 (en) Method and apparatus for generating a distance image of an original image
AU763919B2 (en) Tracking objects from video sequences

Legal Events

Date Code Title Description
FGA Letters patent sealed or granted (standard patent)