WO1999017250A1 - Image comparing system - Google Patents

Image comparing system Download PDF

Info

Publication number
WO1999017250A1
WO1999017250A1 PCT/US1998/020581 US9820581W WO9917250A1 WO 1999017250 A1 WO1999017250 A1 WO 1999017250A1 US 9820581 W US9820581 W US 9820581W WO 9917250 A1 WO9917250 A1 WO 9917250A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
measure
candidate
blob
blobs
Prior art date
Application number
PCT/US1998/020581
Other languages
French (fr)
Inventor
Paul Wensley
Iain Huxley
Original Assignee
Island Graphics Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Island Graphics Corporation filed Critical Island Graphics Corporation
Priority to AU96762/98A priority Critical patent/AU9676298A/en
Publication of WO1999017250A1 publication Critical patent/WO1999017250A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/5838Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features

Definitions

  • the present invention relates generally to image processing techniques for comparing images, and in particular to a method and system for extracting from an image database a set of images that are close in appearance to a target image.
  • the general problem to be solved is that of retrieving from a large and diverse database of images all of the images that share certain properties. Attempts have been made to solve this problem by assigning to each image a set of keywords at the time it is inserted into the database. Images are then judged to be similar if they are tagged with the same keywords.
  • the problem with this method is that it is impossible to encapsulate in a few words everything about the image that might be used as a basis for judging image similarity.
  • a picture of a car on a beach may be tagged with the key words "car” and "beach”, but probably will not be tagged with such terms as,”brown pebbly beach” or “beach next to lake with blue green water” or “beach on the left; lake on the right”.
  • actual image comparison is often based on just these non-verbal attributes on an image, e.g., on what the image is like instead of how the image would be described in words.
  • the advent of databases of digital images on computers makes it possible to compare images on the basis of their actual visual attributes (colors, textures, shapes, etc.). This permits image search by example; the operator of a computer image search system selects a given target image and then requests the computer system to find all images in the database which resemble the example
  • the present invention provides a method and system for quickly comparing a target image with candidate images in a database, and for extracting those images in the database that best match the target.
  • the invention associates with each image a set of image statistics characterizing the image.
  • the invention is similar to prior-art keyword-tagging search schemes in the sense that a set of characteristics is assigned to each image.
  • the selection of this set of characteristics is based on algorithmic examination and decomposition of the image by a computer program and is not subject to human idiosyncrasies or errors. When a target image is selected or inputted, the same decomposition is done to it.
  • Each blob is a cohesive area of the original image (roughly uniform in color, or bounded by a distinct edge) which can be transformed into an exactly uniform-in-color region in the decomposed image.
  • Each cohesive region or blob is characterized by a limited set of numerical parameters (e.g. x and y extent, center of gravity, color, shape, texture, etc.).
  • the set of blobs in the image, along with the characterizing statistics of each blob, constitute the characterizing statistics for the image.
  • An image-similarity score is calculated for any pair of images based on a comparison of the image statistics of the two images.
  • the computation of an image- similarity score between two images typically comprises the three steps of (a) placing the blobs of the two images in one-to-one correspondence, (b) computing a similarity score for each pair of blobs, and then (c) obtaining an overall similarity score for the two images as a function of the similarity scores of the paired blobs in the two images.
  • the user is able to modify aspects of the image-comparison algorithm by varying the weights assigned to the parameters (e.g., size, color, position, etc.) used in generating an image-similarity score.
  • a computer-implemented method for selecting from a computer database of candidate images one or more images which closely match a target image.
  • the method typically includes the steps of extending the image database by computing, for each candidate image, image-characterizing statistics and adding the statistics to the database; computing image-characterizing statistics for the target image; computing, for each candidate image, a measure of its similarity to the target image, wherein the measure is computed as a function of the image-characterizing statistics of the target image and of the candidate image; and displaying at least a portion of one or more of the candidate images having the best image-similarity measures.
  • the image processing system typically includes a memory for storing a plurality of candidate images and image- characterizing statistics associated with each candidate image; and input means for inputting a target image for comparison with the candidate images.
  • the system also typically includes a microprocessor coupled to the memory and the input means, wherein the microprocessor computes image-characterizing statistics for the target image, and wherein for each candidate image the microprocessor determines a measure of the similarity of the candidate image to the target image, wherein the similarity measure is computed as a function of the image- characterizing statistics of the target image and the image-characterizing statistics of the candidate image; and a display for displaying at least a portion of one or more of the candidate images having the best image-similarity measures.
  • Figure 1 illustrates an exemplary image processing system for extracting images from a database "by example" according to an embodiment of the present invention
  • Figure 2 is a flowchart showing the process of analyzing images to be stored in the database
  • Figure 3 is a flowchart showing the process of obtaining a target image, generating statistics for it, comparing it with images stored in the database and displaying the result;
  • Figure 4 illustrates a lion cub image and an owl image and accompanying statistics after reduction of the images to blobs
  • Figure 5 shows the computer display a user might see after seeking a set of twenty candidate images matching the lion cub image
  • Figure 6 illustrates the image match controls in an embodiment of the invention
  • Figure 7 is a flowchart showing the process of comparing the target image with images stored in the database
  • Figure 8 is a flowchart showing the process of generating match scores for blob pairs.
  • Image processing system 70 includes a computer system 71 comprising a microprocessor 72 and a memory 74.
  • Microprocessor 72 performs the image processing and memory 74 stores computer code for processing images.
  • Computer system 71 is any type of computer, such as a PC, a Macintosh, laptop, mainframe or the like.
  • Imaging system 70 also includes a scanner 80 for scanning images directly.
  • Computer system 71 is coupled to momtor 76 for displaying a graphical user interface as well as images.
  • Computer system 71 is also coupled to various interface devices such as internal or external memory mouse and a keyboard (not shown).
  • Printer 78 allows for the printing of any images as required by the user.
  • Cable 82 provides the ability to transfer images to and from another computer device via e-mail, the Internet, direct access or the like.
  • Figure 2 is a flowchart showing the process of analyzing images to store their characteristic data according to an embodiment of the present invention. The process breaks down into the following general steps as shown in Figure 2:
  • an image is provided to the imaging system.
  • the image is provided by selecting an image from an existing collection of images stored in a memory.
  • an image could be provided to the imaging system using a color scanner, digital camera, paint program, or the like.
  • the image is resized to a standard size while maintaining the aspect ratio.
  • the scale factor is stored for later comparisons based on size.
  • the image is reduced to a maximum 64-by-64 pixel resolution. Other resolutions may be used as desired. There is a tradeoff between the speed and the accuracy of the image comparison process. Smaller resolutions provide for increased speed with some loss of accuracy. Maintenance of the aspect ratio means that if the original image is non-square, then the longer axis of the reduced image will have the designated size (e.g., 64 pixels) and the shorter axis will be proportionately smaller.
  • detail is removed from the reduced-size image.
  • the image is blurred using a 10-pixel radius Gaussian blur filter.
  • a median filter may also be used to blur the image.
  • an edge-preserving blur is used to reduce detail.
  • One embodiment uses a Sigma filter as an edge-preserving blur; each pixel is replaced by a mean value of all pixels (a) which are within a given distance of the target pixel, and (b) whose color differs from the color of the target pixel by less than a specified amount.
  • the blurred, reduced-size image is decomposed into a set of cohesive regions or "blobs".
  • blobs of identical color are generated by reducing the number of different colors in the image to a small number. This is done according to one embodiment by using resampling techniques developed originally for computer video displays. Early computer video displays had a small palette of distinct displayable colors. In some such displays the number of displayable colors was limited to a value such as 64 or 256, but each color in this limited palette could be chosen at run time from millions of candidate colors. Hence, technologies were developed to reduce the set of perhaps millions of distinct colors in an image to a representative set of, say, 256 of these colors.
  • One embodiment of this invention uses one such image resampling technique, a median-cut algorithm, as is described, for example, by James D. Foley, van Dam, Feiner and Hughes, in Computer Graphics, Principles and Practice, Addison- Wesley, 1995, at p. 600, the disclosure of which is hereby incorporated by reference.
  • a median-cut algorithm as is described, for example, by James D. Foley, van Dam, Feiner and Hughes, in Computer Graphics, Principles and Practice, Addison- Wesley, 1995, at p. 600, the disclosure of which is hereby incorporated by reference.
  • the number of colors should be less than about 10 to speed the subsequent image-match algorithm. Note that these colors can be completely different across images.
  • the image is now divided into a set of areas of solid color, i.e., blobs. These blobs are catalogued using a flood-fill type algorithm, as is well-known in the art and which is described in Foley, van Dam, Feiner and Hughes, op.
  • An alternative embodiment for the blob-generation process of step 40 employs an adaptive color seed-fill algorithm, thus eliminating the need for image resampling.
  • the image is scanned pixel by pixel, left to right, top to bottom.
  • the first pixel in the image, at the top left, is taken to be the first pixel of the first blob.
  • the second pixel scanned is added to the first blob if it is sufficiently similar in color to the first pixel. Otherwise, it becomes the first pixel of a second blob.
  • a pixel scanned subsequently is added to the blob enclosing one of its adjacent already-scanned neighbor pixels if its color is sufficiently similar to the color of the adjacent pixel. Otherwise it becomes the first pixel in a new blob.
  • This algorithm is a variant of a seed-fill algorithm as is well-known in the art and as is described in Foley, van Dam, Feiner and Hughes, op. cit., pp. 979-980.
  • This algorithm varies from standard seed- fill algorithms in its adaptive property. Standard seed-fill algorithms cease adding pixels to an area when a pixel is encountered that fails a fixed test; e.g., the test might be that the pixel have a color not too different from black.
  • the seed fill algorithm used in this embodiment is adaptive in the sense that the test for inclusion of a pixel into the blob enclosing a neighbor pixel depends on the color of the neighbor pixel.
  • the test for inclusion or exclusion of pixels adapts itself depending on the color of the target pixel.
  • an original-image area of gradually-changing color e.g., a vignette, gradient, or ramp
  • the blob-generating algorithm may be parsed by the blob-generating algorithm as a single blob.
  • the entire reduced-size image has been partitioned into a set of blobs; every pixel in the reduced-size image has been assigned to a blob.
  • step 45 is entered. Step 45 is used to ascertain whether a pre-specified set of criteria concerning the total number of blobs in the reduced-size image has been achieved. If so, flow of control passes to the blob- analysis step 50. If not, steps 30 and 40 are repeated, but with the parameters of the detail- removal and blob-generation algorithms modified so that on the subsequent pass through steps 30 and 40 the image will be decomposed into a smaller total number of blobs.
  • the adaptive color seed-fill algorithm may be used to generate blobs, then on each iteration through step 40 it may be programmed to be more liberal and less discriminating in the criteria it applies when deciding whether or not to add a given image pixel to an existing blob.
  • the system is programmed to cycle through steps 30, 40 and 45 until the predetermined goal has been reached, or until a predetermined maximum number of cycles have been taken. Control then passes to step 50. On each iteration through steps 30 and 40, the total number of blobs in the decomposed image declines.
  • the number of blobs is preferably reduced to ten blobs.
  • any number of blobs can be used.
  • one such halting criterion is that the largest p blobs (for example, the largest 10 blobs) occupy an area equal to a pre-specified proportion (e.g., 75%) of the reduced-size image.
  • a pre-specified proportion e.g. 75% of the reduced-size image.
  • Figure 4 shows two images after they have been reduced to 64-by-64 pixel resolution, ten-blob images.
  • Image 200 of Figure 4 is the blob image of the original lion cub image 300 shown in Figure 5 (and of thumbnail image 305 of Figure 5).
  • Figure 4 is the blob image of the great horned owl thumbnail image 310 shown in Figure 5.
  • the characteristics e.g., color, size, center of gravity, moment of inertia, texture, etc.
  • step 50 is combined with step 40; i.e., the image statistics are in fact generated as the blobs are being generated.
  • step 60 The numerical view of each image created in step 50 is stored (usually, but not necessarily, in a database) in step 60.
  • Figure 4 shows statistics for the four largest (amongst ten generated) of the blobs in the lion cub image 200 and the owl image 210, and, in addition, other statistics generated after matching the two images.
  • Column 0 headed “Match” enumerates the matches between the largest four blobs of the image, in order, with the best match shown first.
  • Column 1, headed “Blob” shows which blobs are matched in each Match. The first two entries in the "Blob” column as shown are zero and zero, indicating that the match is between blob 0 of image 0, background area 202 of lion cub image 200, and blob 0 of image 1, background area 212 of owl image 210.
  • the next column headed "ValA” shows an overall match score for the two blobs.
  • the next column headed “Val” shows a normalized match score, ValA divided by an Area measure, for the two blobs.
  • the next column headed "Area” shows the areas in pixels of the two blobs. Subsequent columns show the statistics summarized below (in each case the statistic characterizes a blob): X: the X position of the center of gravity;
  • Y the Y position of the center of gravity
  • H the hue (a color measure)
  • S the saturation (a color measure)
  • V the value (a color measure)
  • Xe the X extent, in pixels;
  • Ye the Y extent, in pixels; Mo: the moment of inertia; Ra: the minimum radius;
  • the image statistics illustrated in Figure 4 exemplify one embodiment. Other embodiments will vary.
  • the shared goal in the various embodiments is to include statistics measuring for each blob its size (Area in the example), location (X and Y in the example), color (H, S and V) in the example, and shape (Area, Xe, Ye, Mo, Ra, An and Sk in the example).
  • Other embodiments add to this list a set of measures of the textures of blobs.
  • FIG. 3 is a flowchart showing the process of analyzing and comparing the target image with a collection of stored images. The matching process br ⁇ aks down into the following general steps as shown in Figure 3 :
  • the target image is provided by selecting an image from a pre-existing collection of images.
  • the target image is provided to the imaging system using a color scanner, digital camera, paint program, or the like.
  • the target image is provided, the target image is subjected in step 115 of Figure 3 to the same sequence of image statistics generation operations as were applied to database images in step 15 of Figure 2.
  • the numerical results of the statistic-generating step 115 are cached in computer memory for later comparison to the same statistics generated for images in the database, which statistics were stored in the database at step 60 of Figure 2.
  • step 160 the specific requirements of the image processing system operator are obtained.
  • the user has control in determining which search parameters are most important (e.g., whether color and/or location are the most important parameters when searching for matches).
  • a set of sliders such as are shown in Figure 6, is presented to the user to permit setting of the importance of various factors to be used in the comparison of the target image with candidate images. These factors include, for example: 1.
  • the maximum number of candidate image matches to display e.g., the
  • the maximum number of blobs to compare (e.g., the "Max Blobs to Compare" slider 405 in Figure 6).
  • a measure of the importance of color in the match e.g., the "Color Weight” slider 420 in Figure 6).
  • step 170 once the statistics for the target image have been determined, the given target image is compared with all stored candidate images and a match score is generated for each pair of the form (target image, candidate image) .
  • Figure 7 is a flow chart displaying the details of the "Compare with Images in Database" step 170 of Figure 3.
  • An image match score for each pair of images is generated from the similarity scores of the paired blobs from the two images. Consequently, before an image match score is generated it is necessary to place all or some of the blobs from the two images into one-to-one correspondence.
  • the correspondence is such that similar blobs from each image are paired with each other.
  • 10 blobs are generated for each image, and four blobs from each image are placed in one-to-one correspondence with each other.
  • the general rule is that if p blobs are generated for each image, then n blobs from each image, n ⁇ p, are placed in one-to-one correspondence with each other.
  • the former number p is the number of generated blobs in the image and the latter number n is the number of significant blobs in the image.
  • the process of placing the significant blobs in one to one correspondence is shown as step 510 of Figure 7.
  • step 510 the n significant blobs are placed in one to one correspondence. This requires as input a set of measures of the similarity of blob pairs. These measures are generated at step 500 of Figure 7, the original step in operation 170.
  • match scores are developed for pairs of blobs.
  • match scores are generated for all p-by-p pairs of generated blobs, with each pair consisting of one generated blob from the target image and one generated blob from the candidate image.
  • the set of n significant blobs (n ⁇ p) to be placed in one-to-one correspondence is then chosen on the basis of these match scores: if the best (largest) match score matches blob i of the target image to blob j of the candidate image, then blob i from the target is one of the n significant blobs, as is blob j from the candidate.
  • Target blob i and candidate blob j are then placed in one-to-one correspondence.
  • n significant blobs to be matched from each image are chosen a priori to be the n largest blobs in the image and blob match scores are generated for only the n-by-n pairs of these blobs.
  • the matching of blobs at step 510 is done on the basis of these n-by-n match scores: if the best match score matches blob i of the target image to blob j of the candidate image, then target blob i and candidate blob j are then placed in one-to-one correspondence. This process is repeated until all n significant blobs from the target have been placed in one-to-one correspondence with all n significant blobs from the candidate.
  • Figure 8 shows details of step 500, the process of generating match scores for pairs of blobs.
  • similarity scores are generated for each separate statistical component - that is, for each of the several measures which collectively measure the area, location, color, shape and texture of a blob.
  • an overall blob match score is generated from the individual component similarity scores.
  • the individual component similarity scores share the same bounds (from 0 to 1, or from 0 to 100), and the overall blob match score is a measure of the mean of the individual component scores, either the arithmetic mean or the geometric mean or some other measure with the property of a mean.
  • the latter mean similarity score is weighted by the mean areas of the blobs being compared, so as to give a larger similarity score to paired large blobs.
  • step 170 of Figure 3 which comprises steps 500, 510 and 520 of Figure 7.
  • step 520 the resulting set of candidate-image match scores is passed to the final step 180 of Figure 3.
  • the system is programmed to display the candidate images in the database identified as having the best matches with the target image (e.g., the top 20, or the top 50, etc.) given the user's desired input requirements (i.e., parameter settings). If the results are not to the user's liking, the user is able to modify the input parameters and search again.
  • Figure 5 shows one set of displayed results
  • Figure 4 shows the associated image and match statistics for one match.
  • the goal was to match the lion cub image 300 with images in the database.
  • the system returned the 20 best matches.
  • the target lion cub image 300 is itself a member of the database
  • the best match is between the lion cub image and itself, as shown by the thumbnail lion cub image 305.
  • the best non- trivial match is the second best overall match, between the target lion cub image 300 and the great horned owl image 310.
  • the present invention provides a simple, efficient solution to the problem of extracting images from a data base "by example.” While the above is a complete description of the preferred embodiments of the invention, various alternatives, modifications and equivalents may be used.
  • the target image is not a photographic image but an image painted by the user using computer graphic painting software.
  • the user who wants to find a lion cub in the image database first paints an image of the lion cub and then looks for matches to the painted image.
  • the search for matching candidate images can be iterated as the painting progresses; a rough draft of the lion cub painting will yield a first set of matches. As detail is added other sets of matches will be found.
  • Texture matching techniques It is often useful to look for image matches based on image textures, such as the textures in fabrics, in grass, in sand, or in the bark of trees. Texture matching techniques
  • Texture-based comparisons are introduced into this invention in the following manner. Once the image has been decomposed into blobs, each such blob can be used as an index back into the original image.
  • blob 202 is the body of the lion cub; blob 212 is the body of the owl. Areas of the original images 300 and 310 of Figure 5 corresponding to each such blob are found, and texture measures are computed over the indicated areas of the original images. The resulting texture measures are added to the set of blob-characterizing statistics, and a texture similarity score is computed for each blob pair.
  • a texture comparison from, on the one hand, the bodies of the owl 305, serval 315, second lion cub 320 and puma 325 to, on the other hand, the body of the target lion cub 300 will reveal the greater similarities of the fur-to-fur texture comparison between cat-cat pairs than the fur-to-feather comparison between the cat-owl pair.
  • Another variant embodiment modifies the image similarity score algorithm and then cycles through the image-comparison step 170 of Figure 3, culling the set of candidates to a smaller number on each pass.
  • the very first comparisons between a target image and the set of candidate images may be a simple and fast culling operation using a relatively small set of image statistics over a small number of blobs per image, and basing image comparisons on relatively simple measures of differences between blobs.
  • Such a first- pass culling operation can be used to reduce the number of candidate images from, for example about 1,000,000 to about 100,000.
  • a slightly more-sophisticated set of tests is then used to reduce the set of candidate images to about 10,000, and so on, until a manageable number of candidate images, for example about 20, remain.
  • the candidate images are then the ones that match best to all n target images, as measured by the mean of all n matches, or the maximum or minimum of the n matches, or some compound of such measures.
  • Comparison can be based on difference measures just as well as on similarity measures, because difference measures can be constructed as inverses of similarity measures. Comparison can also be based on propinquity measures, since two sets of numbers can be said to be similar to the extent that they are close to each other. Comparison can also be based on distance measures just as well as on propinquity measures, since distance measures can be constructed as inverses of propinquity measures.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Library & Information Science (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

A method and system for quickly comparing a target image (110) with candidate images in a database, and for extracting those images in the database that best match the target image. The method uses a fundamental comparison technique (170) based on the decomposition of the images into 'blobs'. A given image is modified so as to reduce detail. Cohesive regions of the reduced-detail image are transformed into uniform-color blobs. Statistics are generated (15) for each such blob, characterizing, for example, its area, color, location and shape, and also, optionally, measures of the texture of the corresponding area in the original image. An image-similarity score is computed for any pair of images from the blob-specific image statistics. The image-similarity measure is computed by placing the blobs of the target image in one-to-one correspondence (510) with blobs of the candidate image, generating blob-similarity scores (520) over these paired blobs from the pre-computed blob-specific statistics of the images, and generating an overall image similarity score (600) as a function of the blob-similarity scores.

Description

IMAGE COMPARING SYSTEM
BACKGROUND OF THE INVENTION The present invention relates generally to image processing techniques for comparing images, and in particular to a method and system for extracting from an image database a set of images that are close in appearance to a target image.
The general problem to be solved is that of retrieving from a large and diverse database of images all of the images that share certain properties. Attempts have been made to solve this problem by assigning to each image a set of keywords at the time it is inserted into the database. Images are then judged to be similar if they are tagged with the same keywords. The problem with this method is that it is impossible to encapsulate in a few words everything about the image that might be used as a basis for judging image similarity. For example, a picture of a car on a beach may be tagged with the key words "car" and "beach", but probably will not be tagged with such terms as,"brown pebbly beach" or "beach next to lake with blue green water" or "beach on the left; lake on the right". People see a lot of things they do not commonly put into words. However, actual image comparison is often based on just these non-verbal attributes on an image, e.g., on what the image is like instead of how the image would be described in words. The advent of databases of digital images on computers makes it possible to compare images on the basis of their actual visual attributes (colors, textures, shapes, etc.). This permits image search by example; the operator of a computer image search system selects a given target image and then requests the computer system to find all images in the database which resemble the example
It is, however, difficult to design a successful search-by-example system. The problem is that a human being in deciding whether or not two images are similar processes the data in an image in a complex manner. Color, shape, texture, etc. are interdependent in the effect they exert on a person's judgment of image similarity. Existing prior-art systems, however, have placed too much emphasis on single sets of features of the images being compared. A particular prior-art technique consists of generating a frequency diagram or histogram of all the colors in the image. Two images are judged to be similar if their color histograms are similar. Such techniques ignore the shapes of objects in the scene, and hence do a poor job of imitating a human's methodology for comparison of images. Other techniques look for image shapes of a specific type, for example, human faces or thumbprints. These methods do analyze objects in the image, but they are limited to the specific task of the identification of specific target objects.
Many prior-art computer techniques also require extensive analysis of candidate images at search time and hence are slow. Accordingly, it is clear that what is needed in the art is an improved methodology for extracting images from a database that is both quicker and more like real human image-matching methodology than is the prior art.
SUMMARY OF THE INVENTION The present invention provides a method and system for quickly comparing a target image with candidate images in a database, and for extracting those images in the database that best match the target. The invention associates with each image a set of image statistics characterizing the image. Hence the invention is similar to prior-art keyword-tagging search schemes in the sense that a set of characteristics is assigned to each image. In the present invention, however, the selection of this set of characteristics is based on algorithmic examination and decomposition of the image by a computer program and is not subject to human idiosyncrasies or errors. When a target image is selected or inputted, the same decomposition is done to it.
The process that associates with each image a set of image statistics makes use of decomposition of the image into a set of "blobs". Each blob is a cohesive area of the original image (roughly uniform in color, or bounded by a distinct edge) which can be transformed into an exactly uniform-in-color region in the decomposed image. Each cohesive region or blob is characterized by a limited set of numerical parameters (e.g. x and y extent, center of gravity, color, shape, texture, etc.). The set of blobs in the image, along with the characterizing statistics of each blob, constitute the characterizing statistics for the image. An image-similarity score is calculated for any pair of images based on a comparison of the image statistics of the two images. The computation of an image- similarity score between two images typically comprises the three steps of (a) placing the blobs of the two images in one-to-one correspondence, (b) computing a similarity score for each pair of blobs, and then (c) obtaining an overall similarity score for the two images as a function of the similarity scores of the paired blobs in the two images. The user is able to modify aspects of the image-comparison algorithm by varying the weights assigned to the parameters (e.g., size, color, position, etc.) used in generating an image-similarity score.
A computer-implemented method is provided for selecting from a computer database of candidate images one or more images which closely match a target image. The method typically includes the steps of extending the image database by computing, for each candidate image, image-characterizing statistics and adding the statistics to the database; computing image-characterizing statistics for the target image; computing, for each candidate image, a measure of its similarity to the target image, wherein the measure is computed as a function of the image-characterizing statistics of the target image and of the candidate image; and displaying at least a portion of one or more of the candidate images having the best image-similarity measures.
An image processing system is also provided. The image processing system typically includes a memory for storing a plurality of candidate images and image- characterizing statistics associated with each candidate image; and input means for inputting a target image for comparison with the candidate images. The system also typically includes a microprocessor coupled to the memory and the input means, wherein the microprocessor computes image-characterizing statistics for the target image, and wherein for each candidate image the microprocessor determines a measure of the similarity of the candidate image to the target image, wherein the similarity measure is computed as a function of the image- characterizing statistics of the target image and the image-characterizing statistics of the candidate image; and a display for displaying at least a portion of one or more of the candidate images having the best image-similarity measures.
Reference to the remaining portions of the specification, including the drawings and claims, will realize other features and advantages of the present invention. Further features and advantages of the present invention, as well as the structure and operation of various embodiments of the present invention, are described in detail below with respect to the accompanying drawings. In the drawings, like reference numbers indicate identical or functionally similar elements.
BRIEF DESCRIPTION OF THE DRAWINGS Figure 1 illustrates an exemplary image processing system for extracting images from a database "by example" according to an embodiment of the present invention; Figure 2 is a flowchart showing the process of analyzing images to be stored in the database; Figure 3 is a flowchart showing the process of obtaining a target image, generating statistics for it, comparing it with images stored in the database and displaying the result;
Figure 4 illustrates a lion cub image and an owl image and accompanying statistics after reduction of the images to blobs;
Figure 5 shows the computer display a user might see after seeking a set of twenty candidate images matching the lion cub image;
Figure 6 illustrates the image match controls in an embodiment of the invention; Figure 7 is a flowchart showing the process of comparing the target image with images stored in the database; and
Figure 8 is a flowchart showing the process of generating match scores for blob pairs.
DESCRIPTION OF THE SPECIFIC EMBODIMENTS Figure 1 illustrates an embodiment of an image processing system for implementing the image processing and comparison techniques of the present invention. Image processing system 70 includes a computer system 71 comprising a microprocessor 72 and a memory 74. Microprocessor 72 performs the image processing and memory 74 stores computer code for processing images. Computer system 71 is any type of computer, such as a PC, a Macintosh, laptop, mainframe or the like. Imaging system 70 also includes a scanner 80 for scanning images directly. Computer system 71 is coupled to momtor 76 for displaying a graphical user interface as well as images. Computer system 71 is also coupled to various interface devices such as internal or external memory
Figure imgf000007_0001
mouse and a keyboard (not shown). Printer 78 allows for the printing of any images as required by the user. Cable 82 provides the ability to transfer images to and from another computer device via e-mail, the Internet, direct access or the like. Figure 2 is a flowchart showing the process of analyzing images to store their characteristic data according to an embodiment of the present invention. The process breaks down into the following general steps as shown in Figure 2:
a. Image insertion; b. create "blobs"; c. analyze "blobs"; and d. store results.
In step 10, an image is provided to the imaging system. In one embodiment, the image is provided by selecting an image from an existing collection of images stored in a memory. Alternatively, an image could be provided to the imaging system using a color scanner, digital camera, paint program, or the like.
Once an image is provided, the sequence of operations outlined by box 15 of Figure 2 will result in the generation of a set of statistics characterizing the image. This sequence of operations is decomposed into the specific steps described below.
At step 20, the image is resized to a standard size while maintaining the aspect ratio. The scale factor is stored for later comparisons based on size. In one embodiment, the image is reduced to a maximum 64-by-64 pixel resolution. Other resolutions may be used as desired. There is a tradeoff between the speed and the accuracy of the image comparison process. Smaller resolutions provide for increased speed with some loss of accuracy. Maintenance of the aspect ratio means that if the original image is non-square, then the longer axis of the reduced image will have the designated size (e.g., 64 pixels) and the shorter axis will be proportionately smaller. In step 30, detail is removed from the reduced-size image. In one embodiment, the image is blurred using a 10-pixel radius Gaussian blur filter. This effectively removes most of the detail of the image while keeping the dominant colors mostly intact. Alternatively or additionally, a median filter may also be used to blur the image. In another embodiment, an edge-preserving blur is used to reduce detail. One embodiment uses a Sigma filter as an edge-preserving blur; each pixel is replaced by a mean value of all pixels (a) which are within a given distance of the target pixel, and (b) whose color differs from the color of the target pixel by less than a specified amount.
In step 40, the blurred, reduced-size image is decomposed into a set of cohesive regions or "blobs". In one embodiment, blobs of identical color are generated by reducing the number of different colors in the image to a small number. This is done according to one embodiment by using resampling techniques developed originally for computer video displays. Early computer video displays had a small palette of distinct displayable colors. In some such displays the number of displayable colors was limited to a value such as 64 or 256, but each color in this limited palette could be chosen at run time from millions of candidate colors. Hence, technologies were developed to reduce the set of perhaps millions of distinct colors in an image to a representative set of, say, 256 of these colors. One embodiment of this invention uses one such image resampling technique, a median-cut algorithm, as is described, for example, by James D. Foley, van Dam, Feiner and Hughes, in Computer Graphics, Principles and Practice, Addison- Wesley, 1995, at p. 600, the disclosure of which is hereby incorporated by reference. Although more colors can be used, in preferred aspects the number of colors should be less than about 10 to speed the subsequent image-match algorithm. Note that these colors can be completely different across images. The image is now divided into a set of areas of solid color, i.e., blobs. These blobs are catalogued using a flood-fill type algorithm, as is well-known in the art and which is described in Foley, van Dam, Feiner and Hughes, op. cit., pp. 979-980.
An alternative embodiment for the blob-generation process of step 40 employs an adaptive color seed-fill algorithm, thus eliminating the need for image resampling. In this embodiment, the image is scanned pixel by pixel, left to right, top to bottom. The first pixel in the image, at the top left, is taken to be the first pixel of the first blob. The second pixel scanned is added to the first blob if it is sufficiently similar in color to the first pixel. Otherwise, it becomes the first pixel of a second blob. A pixel scanned subsequently is added to the blob enclosing one of its adjacent already-scanned neighbor pixels if its color is sufficiently similar to the color of the adjacent pixel. Otherwise it becomes the first pixel in a new blob. This algorithm is a variant of a seed-fill algorithm as is well-known in the art and as is described in Foley, van Dam, Feiner and Hughes, op. cit., pp. 979-980. This algorithm varies from standard seed- fill algorithms in its adaptive property. Standard seed-fill algorithms cease adding pixels to an area when a pixel is encountered that fails a fixed test; e.g., the test might be that the pixel have a color not too different from black. The seed fill algorithm used in this embodiment is adaptive in the sense that the test for inclusion of a pixel into the blob enclosing a neighbor pixel depends on the color of the neighbor pixel. Hence, because the colors of the pixels within a blob vary at this stage, the test for inclusion or exclusion of pixels adapts itself depending on the color of the target pixel. A result is that an original-image area of gradually-changing color (e.g., a vignette, gradient, or ramp) may be parsed by the blob-generating algorithm as a single blob. At the end of the blob generation step 40 the entire reduced-size image has been partitioned into a set of blobs; every pixel in the reduced-size image has been assigned to a blob.
When the blob-generation step 40 is completed, step 45 is entered. Step 45 is used to ascertain whether a pre-specified set of criteria concerning the total number of blobs in the reduced-size image has been achieved. If so, flow of control passes to the blob- analysis step 50. If not, steps 30 and 40 are repeated, but with the parameters of the detail- removal and blob-generation algorithms modified so that on the subsequent pass through steps 30 and 40 the image will be decomposed into a smaller total number of blobs. For example, if the adaptive color seed-fill algorithm is used to generate blobs, then on each iteration through step 40 it may be programmed to be more liberal and less discriminating in the criteria it applies when deciding whether or not to add a given image pixel to an existing blob. The system is programmed to cycle through steps 30, 40 and 45 until the predetermined goal has been reached, or until a predetermined maximum number of cycles have been taken. Control then passes to step 50. On each iteration through steps 30 and 40, the total number of blobs in the decomposed image declines. (Strictly speaking, the number .either declines or stays the same.) The goal of iterating over steps 30 and 40 is, roughly speaking, to reduce the number of blobs to a predetermined maximum number. In one embodiment, the number of blobs is preferably reduced to ten blobs. However, any number of blobs can be used. Additionally, in many embodiments it is efficient to impose a halting criterion that does not refer explicitly to the target number of blobs into which the reduced-size image has been decomposed. For example, one such halting criterion is that the largest p blobs (for example, the largest 10 blobs) occupy an area equal to a pre-specified proportion (e.g., 75%) of the reduced-size image. If the minimal perimeter adaptive fill seed algorithm has been used for blob generation there is no guarantee that each individual blob remaining in the image at the beginning of step 50 will be filled with pixels of identical color. However, it is required that at the beginning of step 50 a mean color for each blob be known. Hence, in this case, there may be an additional step between steps 45 and 50 comprising replacing the colors of pixels within each blob with the average of the colors of all pixels in the blob. Alternatively, the average color of each blob may be computed as the blob is constructed, so that the mean color of the blob is known before step 50 is entered.
Figure 4 shows two images after they have been reduced to 64-by-64 pixel resolution, ten-blob images. Image 200 of Figure 4 is the blob image of the original lion cub image 300 shown in Figure 5 (and of thumbnail image 305 of Figure 5). Image 210 of
Figure 4 is the blob image of the great horned owl thumbnail image 310 shown in Figure 5. In step 50, the characteristics (e.g., color, size, center of gravity, moment of inertia, texture, etc.) of the blobs are determined, and a numerical view of the blobs is created. For efficiency, in one embodiment, step 50 is combined with step 40; i.e., the image statistics are in fact generated as the blobs are being generated.
The numerical view of each image created in step 50 is stored (usually, but not necessarily, in a database) in step 60.
Figure 4 shows statistics for the four largest (amongst ten generated) of the blobs in the lion cub image 200 and the owl image 210, and, in addition, other statistics generated after matching the two images. Column 0, headed "Match" enumerates the matches between the largest four blobs of the image, in order, with the best match shown first. Column 1, headed "Blob" shows which blobs are matched in each Match. The first two entries in the "Blob" column as shown are zero and zero, indicating that the match is between blob 0 of image 0, background area 202 of lion cub image 200, and blob 0 of image 1, background area 212 of owl image 210. The next column headed "ValA" shows an overall match score for the two blobs. The next column headed "Val" shows a normalized match score, ValA divided by an Area measure, for the two blobs. The next column headed "Area" shows the areas in pixels of the two blobs. Subsequent columns show the statistics summarized below (in each case the statistic characterizes a blob): X: the X position of the center of gravity;
Y: the Y position of the center of gravity; H: the hue (a color measure); S: the saturation (a color measure); V: the value (a color measure); Xe: the X extent, in pixels;
Ye: the Y extent, in pixels; Mo: the moment of inertia; Ra: the minimum radius;
An: the angle from the horizontal of the major axis; Sk: the skewness;
The image statistics illustrated in Figure 4 exemplify one embodiment. Other embodiments will vary. The shared goal in the various embodiments is to include statistics measuring for each blob its size (Area in the example), location (X and Y in the example), color (H, S and V) in the example, and shape (Area, Xe, Ye, Mo, Ra, An and Sk in the example). Other embodiments add to this list a set of measures of the textures of blobs.
The above process of image statistics generation, as shown in box 15 of Figure 2, is repeated for each image desired to be stored.
After all information has been created a user inputs a target image desired to be matched with the collection of stored images. The target image is analyzed as above. Figure 3 is a flowchart showing the process of analyzing and comparing the target image with a collection of stored images. The matching process brςaks down into the following general steps as shown in Figure 3 :
a. Generate image statistics; b. obtain user requirements (e.g., color important, position important, etc.); c. compare to stored images; and d. display results of best/closest match(es).
According to one embodiment, in step 110 of Figure 3 the target image is provided by selecting an image from a pre-existing collection of images. Alternatively, the target image is provided to the imaging system using a color scanner, digital camera, paint program, or the like.
Once the target image is provided, the target image is subjected in step 115 of Figure 3 to the same sequence of image statistics generation operations as were applied to database images in step 15 of Figure 2.
At step 150 of Figure 3, the numerical results of the statistic-generating step 115 are cached in computer memory for later comparison to the same statistics generated for images in the database, which statistics were stored in the database at step 60 of Figure 2.
In step 160, the specific requirements of the image processing system operator are obtained. The user has control in determining which search parameters are most important (e.g., whether color and/or location are the most important parameters when searching for matches). A set of sliders, such as are shown in Figure 6, is presented to the user to permit setting of the importance of various factors to be used in the comparison of the target image with candidate images. These factors include, for example: 1. The maximum number of candidate image matches to display (e.g., the
"Max Ids to Return" slider 400 in Figure 6).
2. The maximum number of blobs to compare (e.g., the "Max Blobs to Compare" slider 405 in Figure 6).
3. A measure of the importance of color in the match (e.g., the "Color Weight" slider 420 in Figure 6).
4. A measure of the importance of position iq., the match (e.g., the "Location Weight" slider 415 in Figure 6), affecting how the center of gravity parameter is weighted in the matching computation.
5. Measures of the importance of shape in the match (e.g., the "Area", "Extents", "Inertia", "Radius", "Angle", and "Skew" sliders 410, 425, 430, 435, 440 and 445, respectively, of Figure 6. These affect how the moment of inertia and x and y extents, etc., are used in the match).
In step 170, once the statistics for the target image have been determined, the given target image is compared with all stored candidate images and a match score is generated for each pair of the form (target image, candidate image) .
Figure 7 is a flow chart displaying the details of the "Compare with Images in Database" step 170 of Figure 3. An image match score for each pair of images is generated from the similarity scores of the paired blobs from the two images. Consequently, before an image match score is generated it is necessary to place all or some of the blobs from the two images into one-to-one correspondence. The correspondence is such that similar blobs from each image are paired with each other. In one preferred embodiment 10 blobs are generated for each image, and four blobs from each image are placed in one-to-one correspondence with each other. The general rule is that if p blobs are generated for each image, then n blobs from each image, n < p, are placed in one-to-one correspondence with each other. The former number p is the number of generated blobs in the image and the latter number n is the number of significant blobs in the image. The process of placing the significant blobs in one to one correspondence is shown as step 510 of Figure 7.
In step 510, the n significant blobs are placed in one to one correspondence. This requires as input a set of measures of the similarity of blob pairs. These measures are generated at step 500 of Figure 7, the original step in operation 170.
In Step 500 match scores are developed for pairs of blobs. In one embodiment, match scores are generated for all p-by-p pairs of generated blobs, with each pair consisting of one generated blob from the target image and one generated blob from the candidate image. The set of n significant blobs (n < p) to be placed in one-to-one correspondence is then chosen on the basis of these match scores: if the best (largest) match score matches blob i of the target image to blob j of the candidate image, then blob i from the target is one of the n significant blobs, as is blob j from the candidate. Target blob i and candidate blob j are then placed in one-to-one correspondence. This process is repeated until n blobs from the target have been placed in one-to-one correspondence with n blobs from the candidate. In another embodiment, the n significant blobs to be matched from each image are chosen a priori to be the n largest blobs in the image and blob match scores are generated for only the n-by-n pairs of these blobs. In this latter embodiment the matching of blobs at step 510 is done on the basis of these n-by-n match scores: if the best match score matches blob i of the target image to blob j of the candidate image, then target blob i and candidate blob j are then placed in one-to-one correspondence. This process is repeated until all n significant blobs from the target have been placed in one-to-one correspondence with all n significant blobs from the candidate.
Figure 8 shows details of step 500, the process of generating match scores for pairs of blobs. First, at step 600, for each given pair of blobs, similarity scores are generated for each separate statistical component - that is, for each of the several measures which collectively measure the area, location, color, shape and texture of a blob. At step 610, an overall blob match score is generated from the individual component similarity scores. In some embodiments the individual component similarity scores share the same bounds (from 0 to 1, or from 0 to 100), and the overall blob match score is a measure of the mean of the individual component scores, either the arithmetic mean or the geometric mean or some other measure with the property of a mean. In one embodiment, the latter mean similarity score is weighted by the mean areas of the blobs being compared, so as to give a larger similarity score to paired large blobs. After n significant blobs from the target image have been placed in one-to-one correspondence with n significant blobs from the candidate image at step 510 of Figure 7, the overall image match score for the pair of images is generated at step 520 of Figure 7. The overall image match score is generated as a sum or mean (or other increasing function) of the n match scores for the n paired blobs in the one-to-one correspondence list of step 510. The blob match scores used are the same ones that were generated at step 500 of Figure 7.
The set of all candidate-image match scores i§ computed by the operation of step 170 of Figure 3, which comprises steps 500, 510 and 520 of Figure 7. When step 520 is completed the resulting set of candidate-image match scores is passed to the final step 180 of Figure 3. In step 180, the system is programmed to display the candidate images in the database identified as having the best matches with the target image (e.g., the top 20, or the top 50, etc.) given the user's desired input requirements (i.e., parameter settings). If the results are not to the user's liking, the user is able to modify the input parameters and search again.
Figure 5 shows one set of displayed results, and Figure 4 shows the associated image and match statistics for one match. In the case illustrated by Figure 5 the goal was to match the lion cub image 300 with images in the database. The system returned the 20 best matches. Because the target lion cub image 300 is itself a member of the database, the best match is between the lion cub image and itself, as shown by the thumbnail lion cub image 305. The best non- trivial match is the second best overall match, between the target lion cub image 300 and the great horned owl image 310.
In conclusion, the present invention provides a simple, efficient solution to the problem of extracting images from a data base "by example." While the above is a complete description of the preferred embodiments of the invention, various alternatives, modifications and equivalents may be used.
In one variant embodiment, the target image is not a photographic image but an image painted by the user using computer graphic painting software. In this embodiment the user who wants to find a lion cub in the image database first paints an image of the lion cub and then looks for matches to the painted image. The search for matching candidate images can be iterated as the painting progresses; a rough draft of the lion cub painting will yield a first set of matches. As detail is added other sets of matches will be found.
It is often useful to look for image matches based on image textures, such as the textures in fabrics, in grass, in sand, or in the bark of trees. Texture matching techniques
> are often based on the spectral decomposition, such as can be obtained by a Fourier transformation, of areas of the image; texture matching can also be done by a process known as Wold decomposition, and described in A new Wold ordering for image similarity; Rosalind W. Picard and Fang Liu, Proc. IEEE Conf. on Acoustics, Speech, and Signal Proc, Adelaide, Australia, April 1994, pp. 129-132, the contents of which are hereby incorporated by reference for all purposes. Texture-based comparisons are introduced into this invention in the following manner. Once the image has been decomposed into blobs, each such blob can be used as an index back into the original image. For example, in Figure 4, blob 202 is the body of the lion cub; blob 212 is the body of the owl. Areas of the original images 300 and 310 of Figure 5 corresponding to each such blob are found, and texture measures are computed over the indicated areas of the original images. The resulting texture measures are added to the set of blob-characterizing statistics, and a texture similarity score is computed for each blob pair. Referring to Figure 5, a texture comparison from, on the one hand, the bodies of the owl 305, serval 315, second lion cub 320 and puma 325 to, on the other hand, the body of the target lion cub 300 will reveal the greater similarities of the fur-to-fur texture comparison between cat-cat pairs than the fur-to-feather comparison between the cat-owl pair.
Another variant embodiment modifies the image similarity score algorithm and then cycles through the image-comparison step 170 of Figure 3, culling the set of candidates to a smaller number on each pass. The very first comparisons between a target image and the set of candidate images may be a simple and fast culling operation using a relatively small set of image statistics over a small number of blobs per image, and basing image comparisons on relatively simple measures of differences between blobs. Such a first- pass culling operation can be used to reduce the number of candidate images from, for example about 1,000,000 to about 100,000. A slightly more-sophisticated set of tests is then used to reduce the set of candidate images to about 10,000, and so on, until a manageable number of candidate images, for example about 20, remain.
Another variant of the invention bases the search for candidate images not on a single target image but on n = 2 or more target images. The candidate images are then the ones that match best to all n target images, as measured by the mean of all n matches, or the maximum or minimum of the n matches, or some compound of such measures.
Other variants of the invention use blob-comrjarison measures and image- comparison measures other than similarity measures. Comparison can be based on difference measures just as well as on similarity measures, because difference measures can be constructed as inverses of similarity measures. Comparison can also be based on propinquity measures, since two sets of numbers can be said to be similar to the extent that they are close to each other. Comparison can also be based on distance measures just as well as on propinquity measures, since distance measures can be constructed as inverses of propinquity measures.
In light of the various alternatives, modifications and equivalents to the present invention, the above description should not be taken as limiting the scope of the invention which is defined by the appended claims.

Claims

WHAT IS CLAIMED IS:
1. A computer-implemented method of comparing images comprising the steps of: (a) providing two images to be compared; (b) computing image-characterizing statistics for both images; and (c) computing an overall comparison measure for the two images, said measure derived from the two sets of image-characterizing statistics, one set corresponding to each image.
2. The method of claim 1 wherein the overall comparison measure is selected from the group consisting of a similarity measure, a difference measure, a propinquity measure and a distance measure.
3. The method of claim 1 wherein the operation of computing image- characterizing statistics for an image includes the steps of: (i) decomposing the image into one or more blobs; and (ii) computing blob-characterizing statistics for each blob in the decomposed image.
4. The method of claim 3 wherein the step of decomposing the image into one or more blobs includes the steps of: modifying the image to reduce detail; and identifying blobs of the reduced-detail image.
5. The method of claim 4, wherein the step of modifying the image to reduce detail includes resizing the image.
6. The method of claim 4, wherein the step of modifying the image to reduce detail includes the step of applying to the image one of a Gaussian blur filter, a median filter, and an edge-preserving blur filter.
7. The method of claim 4 wherein the step of identifying blobs includes the step of reducing the number of distinct colors in the reduced-detail image.
8. The method of claim 7 wherein the step of reducing the number of distinct colors in the reduced-detail image includes the step of applying to the reduced-detail image one of a median-cut algorithm, and an adaptive color seed-fill algorithm.
9. The method of claim 4 wherein the steps of modifying the image to reduce detail and identifying blobs are sequentially and adaptively iterated until a halting criterion is satisfied.
10. The method of claim 9 wherein the halting criterion is satisfied when at least a specified proportion of the image is covered by no more than a specified number ofblobs.
11. The method of claim 3 wherein the step of computing image- characterizing statistics from the decomposed image includes the step of computing at least one of an area measure, a location measure, a color measure, a shape measure and a texture measure for each of the one or more blobs.
12. The method of claim 1 wherein one of the two images is provided as a candidate image and the other is provided as a target image, wherein step (c) of computing an overall comparison measure for the two images includes the steps of: (i) applying at least one comparison measure to pairs of sets of one or more image statistics, one set from the candidate image and one set from the target image; and thereafter (ii) combining the comparison measures calculated on the pairs of sets of one or more image statistics into one overall comparison measure between the candidate image and the target image.
13. The method of claim 3, wherein one of the two images is provided as a candidate image and the other is provided as a target image, wherein step (c) of computing an overall comparison measure for the two images includes the steps of: (i) placing a number, n, of the blobs in the target image in one-to-one correspondence with n blobs in the candidate image, the one-to-one correspondence associating similar blobs from each image; (ii) for each blob pair, applying at least one comparison measure to pairs of sets of one or more blob statistics, one set from a blob of the candidate image and one set from a blob of the target image; (iii) combining the comparison measures calculated on the pairs of sets of one or more blob statistic into one overall blob-comparison measure between each blob of the candidate image and its corresponding blob of the target image; and thereafter (iv) combining the blob-comparison measures of the paired blobs into one overall image-comparison measure between the candidate image and the target image.
14. The method of claim 12 or 13, wherein the at least one comparison measure is selected from the group consisting of a similarity measure, a difference measure, a propinquity measure and a distance measure.
15. The method of claim 12 or 13 wherein the operation of combining the comparison measures calculated on the pairs of sets of one or more image statistics into one overall comparison measure between the candidate image and the target image is governed by a set of user-modifiable parameters affecting the relative importance of the pairwise comparisons in the computation of the overall comparison measure.
16. A computer-implemented method of selecting from a computer database of candidate images one or more images that closely match a target image, the method comprising the steps of: (a) computing for each candidate image image-characterizing statistics and storing each candidate image's image-characterizing statistics in a computer memory; (b) computing image-characterizing statistics for the target image; (c) computing for each candidate image an overall comparison measure for the candidate image and the target image using the image-characterizing statistics of the candidate image and of the target image, and storing the overall comparison measure for each candidate image in the computer memory; and (d) using the set of overall comparison measures to identify one or more candidate images which closely match the target image.
17. The method of claim 16 wherein the operation of computing image- characterizing statistics for an image includes the steps of: (i) decomposing the image into one or more blobs; and (ii) computing blob-characterizing statistics for each blob in the decomposed image.
18. The method of claim 17 wherein the step of decomposing the image into one or more blobs includes the steps of: modifying the image to reduce detail; and identifying blobs of the reduced-detail image.
19. The method of claim 16 wherein one of the two images is provided as a candidate image and the other is provided as a target image, and wherein step (c) includes the steps of: (i) applying, for each candidate image, at least one comparison measure to pairs of sets of one or more image statistics, one set from the candidate image and one set from the target image; and thereafter (ii) combining the comparison measures calculated on the pairs of sets of one or more image statistics into the overall comparison measure between the candidate image and the target image.
20. The method of claim 17 wherein one of the two images is provided as a candidate image and the other is provided as a target image, and wherein step (c) includes the steps of: (i) for each candidate image, placing a number, n, of the blobs in the target image in one-to-one correspondence with n blobs in the candidate image, the one-to-one correspondence associating similar blobs from each image; (ii) for each blob pair, applying at least one comparison measure to pairs of sets of one or more blob statistics, one set from a blob of the candidate image and one set from a blob of the target image; (iii) combining the comparison measures calculated on the pairs of sets of one or more blob statistic into one overall blob-comparison measure between each blob of the candidate image and its corresponding blob of the target image; and thereafter (iv) combining the blob-comparison measures of the paired blobs into one overall image-comparison measure between the candidate image and the target image.
21. The method of claim 16 further comprising the step of: displaying at least a portion of each of the one or more candidate images that closely match the target image.
22. The method of claim 16 wherein step (c) is adaptively iterated so that on each iteration the number of candidate images that closely match the target image is reduced by culling out a number of poorly-matching images.
23. The method of claim 22 wherein on each iteration computation of the comparison measures is varied so that image comparisons execute faster on early iterations when there are many candidate images than on later iterations when there are fewer candidate images.
24. An image processing system comprising: a memory for storing a plurality of candidate images and image-characterizing statistics associated with each candidate image; input means for inputting a target image for comparison with the candidate images; a microprocessor coupled to the memory and the input means, wherein the microprocessor computes image-characterizing statistics for the target image, and wherein for each candidate image the microprocessor determines an image-comparison measure for the candidate image and the target image, wherein the image-comparison measure is computed as a function of the image-characterizing statistics of the target image and the image-characterizing statistics of the candidate image; and a display for displaying at least a portion of one or more of the candidate images having the best image-comparison measures.
25. The image processing system of claim 24, wherein when each of the candidate images is stored to the memory, the microprocessor computes the associated image-characterizing statistics and stores the associated image-characterizing statistics to the memory.
26. The image processing system of claim 24, wherein the microprocessor includes means for decomposing the image into one or more blobs, and wherein the microprocessor computes image-characterizing statistics from the decomposed image.
27. The image processing system of claim 26, wherein the image- characterizing statistics for an image include at least one of an area measure, a location measure, a color measure, a shape measure and a texture measure for each of the one or more blobs.
28. The image processing system of claim 24, wherein the microprocessor includes means for modifying an image to reduce detail so as to produce a reduced-detail image and means for identifying blobs in the reduced-detail image.
29. The image processing system of claim 28, wherein the means for modifying the image to reduce detail includes means for resizing the image.
30. The image processing system of claim 28, wherein the means for modifying the image to reduce detail includes means for applying one of a Gaussian blur filter, a median filter, and an edge-preserving blur filter.
31. The image processing system of claim 28, wherein the microprocessor sequentially, adaptively and iteratively modifies the image to reduce detail and identify blobs, until at least a specified proportion of the image is covered by no more than a specified number of blobs.
32. The image processing system of claim 24, wherein the microprocessor determines each image-comparison measure by applying comparison measures to pairs of sets of one or more image statistics, one set from the candidate image and one set from the target image, and combining the comparison measures into one overall image- comparison measure between the candidate image and the target image.
33. The image processing system of claim 26, wherein the microprocessor determines each image-comparison measure by placing a number, n, of the blobs in the target image in one-to-one coπespondence with n blobs in the candidate image, the one- to-one correspondence associating similar blobs from each image, and then for each blob pair, applies comparison measures to pairs of sets of one or more blob statistics, one set from a blob of the candidate image and one set from a blob of the target image, and then combines the comparison measures into one overall blob-comparison measure for each pair of blobs, and then combines the blob-comparison measures into one overall image- comparison measure between the candidate image and the target image.
34. The image processing system of claim 32, wherein the microprocessor adaptively and iteratively determines the image-comparison measure between the candidate image and the target image, so that on each iteration the number of candidate images that closely match the target image is reduced by culling out a number of poorly- matching images.
35. The image processing system of claim 34 wherein on each iteration the microprocessor varies the computation of the comparison measures so that image- comparisons execute faster on early iterations when there are many candidate images than on later iterations when there are fewer candidate images.
36. An image processing system comprising: means for providing a target image and at least one candidate image; and a processor coupled to the providing means, wherein the microprocessor computes image-characterizing statistics for the at least one candidate image and for the target image, and wherein the processor computes an overall comparison measure for the at least one candidate image and the target image, the measure being derived from the image characterizing statistics of the target image and of the at least one candidate image.
37. The image processing system of claim 36, wherein the processor includes a means for decomposing an image into one or more blobs, and a means for computing blob-characterizing statistics for each blob in the decomposed image, wherein the image- characterizing statistics for the target image and the at least one candidate image are computed using the corresponding blob-characterizing statistics.
38. The image processing system of claim 37, wherein the means for decomposing an image includes means for modifying the image to reduce detail and means for identifying blobs in the reduced-detail image.
39. The system of claim 38, wherein the means for modifying the image to reduce detail includes means for resizing the image.
40. The system of claim 38, wherein the means for modifying the image to reduce detail includes means for applying to the image one of a Gaussian blur filter, a median filter, and an edge-preserving blur filter.
41. The system of claim 38, wherein the means for identifying blobs includes means for reducing the number of distinct colors in the reduced-detail image.
42. The system of claim 41, wherein the means for reducing the number of distinct colors in the reduced-detail image includes means for applying to the reduced- detail image one of a median-cut algorithm, and an adaptive color seed-fill algorithm.
43. The system of claim 37, wherein the means for computing image- characterizing statistics from the decomposed image includes means for computing at least one of an area measure, a location measure, a color measure, a shape measure and a texture measure for each of the one or more blobs.
44. The image processing system of claim 36 wherein the processor determines at least one comparison measure for pairs of sets of one or more image statistics, one set from the at least one candidate image and one set from the target image, and wherein the processor combines the comparison measures to determine the overall comparison measure.
45. The image processing system of claim 37, wherein the processor includes: means for placing a number, n, of the blobs in the target image in one-to-one correspondence with n blobs in the at least one candidate image, the one-to-one correspondence associating similar blobs from each image; means for applying, for each blob pair, at least one comparison measure to pairs of sets of one or more blob statistics, one set from a blob of the at least one candidate image and one set from a blob of the target image; means for combining the comparison measures calculated on the pairs of sets of one or more blob statistic into one overall blob-comparison measure between each blob of the at least one candidate image and the corresponding blob of the target image; and means for combining the blob-comparison measures of the paired blobs into an overall image-comparison measure between the at least one candidate image and the target image.
46. The image processing system of claim 44 or 45, wherein the at least one comparison measure is selected from the group consisting of a similarity measure, a difference measure, a propinquity measure and a distance measure.
PCT/US1998/020581 1997-10-01 1998-09-30 Image comparing system WO1999017250A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU96762/98A AU9676298A (en) 1997-10-01 1998-09-30 Image comparing system

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US6058497P 1997-10-01 1997-10-01
US60/060,584 1997-10-01
US16292898A 1998-09-29 1998-09-29
US09/162,928 1998-09-29

Publications (1)

Publication Number Publication Date
WO1999017250A1 true WO1999017250A1 (en) 1999-04-08

Family

ID=26740076

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US1998/020581 WO1999017250A1 (en) 1997-10-01 1998-09-30 Image comparing system

Country Status (2)

Country Link
AU (1) AU9676298A (en)
WO (1) WO1999017250A1 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002099738A1 (en) * 2001-06-07 2002-12-12 C Technologies Ab Method and apparatus for extracting information from a target area within a two-dimensional graphical object in an image
GB2391677A (en) * 1999-07-05 2004-02-11 Mitsubishi Electric Inf Tech Method for representing and searching for an object in an image
EP1496473A2 (en) * 1999-04-29 2005-01-12 Mitsubishi Electric Information Technology Centre Europe B.V. Method and apparatus for representing and searching for colour images
WO2006013384A1 (en) * 2004-08-04 2006-02-09 Forensic-Eye Limited Image comparison system and method
EP1669908A3 (en) * 2004-12-07 2007-04-25 Canon Kabushiki Kaisha Image search apparatus, image search method, program, and storage medium
US20080192116A1 (en) * 2005-03-29 2008-08-14 Sportvu Ltd. Real-Time Objects Tracking and Motion Capture in Sports Events
EP1164506A3 (en) * 2000-05-22 2008-11-26 Eastman Kodak Company Determining sets of materials interesting for a user by analyzing images
US8077927B1 (en) * 2006-11-17 2011-12-13 Corelogic Real Estate Solutions, Llc Updating a database with determined change identifiers

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4942619A (en) * 1986-01-20 1990-07-17 Nikon Corporation Pattern inspecting apparatus
US5093867A (en) * 1987-07-22 1992-03-03 Sony Corporation Candidate article recognition with assignation of reference points and respective relative weights
US5251265A (en) * 1990-10-27 1993-10-05 International Business Machines Corporation Automatic signature verification
US5253306A (en) * 1990-01-12 1993-10-12 Futec Inc. Method of matching patterns and apparatus therefor
US5261008A (en) * 1990-08-07 1993-11-09 Yozan, Inc. Fingerprint verification method
US5432893A (en) * 1992-02-11 1995-07-11 Purdue Research Foundation Sequential scalar quantization of digital color image using mean squared error-minimizing quantizer density function
US5450504A (en) * 1992-05-19 1995-09-12 Calia; James Method for finding a most likely matching of a target facial image in a data base of facial images
US5539841A (en) * 1993-12-17 1996-07-23 Xerox Corporation Method for comparing image sections to determine similarity therebetween
US5583950A (en) * 1992-09-16 1996-12-10 Mikos, Ltd. Method and apparatus for flash correlation

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4942619A (en) * 1986-01-20 1990-07-17 Nikon Corporation Pattern inspecting apparatus
US5093867A (en) * 1987-07-22 1992-03-03 Sony Corporation Candidate article recognition with assignation of reference points and respective relative weights
US5253306A (en) * 1990-01-12 1993-10-12 Futec Inc. Method of matching patterns and apparatus therefor
US5261008A (en) * 1990-08-07 1993-11-09 Yozan, Inc. Fingerprint verification method
US5251265A (en) * 1990-10-27 1993-10-05 International Business Machines Corporation Automatic signature verification
US5432893A (en) * 1992-02-11 1995-07-11 Purdue Research Foundation Sequential scalar quantization of digital color image using mean squared error-minimizing quantizer density function
US5450504A (en) * 1992-05-19 1995-09-12 Calia; James Method for finding a most likely matching of a target facial image in a data base of facial images
US5583950A (en) * 1992-09-16 1996-12-10 Mikos, Ltd. Method and apparatus for flash correlation
US5539841A (en) * 1993-12-17 1996-07-23 Xerox Corporation Method for comparing image sections to determine similarity therebetween

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1496473A2 (en) * 1999-04-29 2005-01-12 Mitsubishi Electric Information Technology Centre Europe B.V. Method and apparatus for representing and searching for colour images
US7636094B2 (en) 1999-04-29 2009-12-22 Mitsubishi Denki Kabushiki Kaisha Method and apparatus for representing and searching for colour images
EP1496473A3 (en) * 1999-04-29 2007-03-07 Mitsubishi Electric Information Technology Centre Europe B.V. Method and apparatus for representing and searching for colour images
US7356203B2 (en) 1999-07-05 2008-04-08 Mitsubishi Denki Kabushiki Kaisha Method, apparatus, computer program, computer system, and computer-readable storage medium for representing and searching for an object in an image
US7483594B2 (en) 1999-07-05 2009-01-27 Mitsubishi Denki Kabushiki Kaisha Method, apparatus, computer program, computer system, and computer-readable storage medium for representing and searching for an object in an image
GB2391677A (en) * 1999-07-05 2004-02-11 Mitsubishi Electric Inf Tech Method for representing and searching for an object in an image
GB2391677B (en) * 1999-07-05 2004-05-12 Mitsubishi Electric Inf Tech Method and apparatus for representing and searching for an object in an image
US7542626B2 (en) 1999-07-05 2009-06-02 Mitsubishi Denki Kabushiki Kaisha Method, apparatus, computer program, computer system, and computer-readable storage medium for representing and searching for an object in an image
US7505637B2 (en) 1999-07-05 2009-03-17 Mitsubishi Denki Kabushiki Kaisha Method, apparatus, computer program, computer system, and computer-readable storage medium for representing and searching for an object in an image
US7492972B2 (en) 1999-07-05 2009-02-17 Mitsubishi Denki Kabushiki Kaisha Method, apparatus, computer program, computer system, and computer-readable storage medium for representing and searching for an object in an image
US6882756B1 (en) 1999-07-05 2005-04-19 Mitsubishi Denki Kabushiki Kaisha Method and device for displaying or searching for object in image and computer-readable storage medium
EP1164506A3 (en) * 2000-05-22 2008-11-26 Eastman Kodak Company Determining sets of materials interesting for a user by analyzing images
WO2002099738A1 (en) * 2001-06-07 2002-12-12 C Technologies Ab Method and apparatus for extracting information from a target area within a two-dimensional graphical object in an image
WO2006013384A1 (en) * 2004-08-04 2006-02-09 Forensic-Eye Limited Image comparison system and method
US7508998B2 (en) 2004-12-07 2009-03-24 Canon Kabushiki Kaisha Image search apparatus, image search method, program, and storage medium
EP1669908A3 (en) * 2004-12-07 2007-04-25 Canon Kabushiki Kaisha Image search apparatus, image search method, program, and storage medium
US20080192116A1 (en) * 2005-03-29 2008-08-14 Sportvu Ltd. Real-Time Objects Tracking and Motion Capture in Sports Events
US8077927B1 (en) * 2006-11-17 2011-12-13 Corelogic Real Estate Solutions, Llc Updating a database with determined change identifiers

Also Published As

Publication number Publication date
AU9676298A (en) 1999-04-23

Similar Documents

Publication Publication Date Title
US11004129B2 (en) Image processing
Ahmad Deep image retrieval using artificial neural network interpolation and indexing based on similarity measurement
US10496695B2 (en) Automated data extraction from scatter plot images
JP4234381B2 (en) Method and computer program product for locating facial features
US7440638B2 (en) Image retrieving system, image classifying system, image retrieving program, image classifying program, image retrieving method and image classifying method
US6674915B1 (en) Descriptors adjustment when using steerable pyramid to extract features for content based search
RU2385494C1 (en) Method for recognition of cell texture image
WO2005066896A2 (en) Detection of sky in digital color images
CN101551823A (en) Comprehensive multi-feature image retrieval method
CN1322471C (en) Comparing patterns
Singh et al. A Machine Learning Model for Content-Based Image Retrieval
Fuertes et al. A scheme of colour image retrieval from databases
KR100323364B1 (en) A method for extracting characterized vectors in color image and a retrieval method using the characterized vectors
CN112651955A (en) Intestinal tract image identification method and terminal device
EP1530159B1 (en) Systems and methods for indexing and retrieving images
WO1999017250A1 (en) Image comparing system
JP3819236B2 (en) Pattern recognition method and computer-readable storage medium storing program for performing pattern recognition
CN115272741A (en) Detection method of slender flexible object, terminal equipment and storage medium
JP2004192121A (en) Image retrieval device, image classification method, image retrieval method, and program
Al-Oraiqat et al. A modified image comparison algorithm using histogram features
Mustapha et al. Compact composite descriptors for content based image retrieval
Jayabalaji et al. Enhancing Content-Based Image Retrieval for Lung Cancer Diagnosis: Leveraging Texture Analysis and Ensemble Models
CN118135241A (en) Peach seed phenotype feature extraction method and device, electronic equipment and storage medium
CN114429519A (en) Method for constructing VR content library based on visual modeling and word meaning analysis
CN116415026A (en) Image retrieval video method, device, equipment and storage medium

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GD GE GH GM HR HU ID IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG US UZ VN YU ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

NENP Non-entry into the national phase

Ref country code: CA

122 Ep: pct application non-entry in european phase