WO2009135151A1 - System for using image alignment to map objects across disparate images - Google Patents

System for using image alignment to map objects across disparate images Download PDF

Info

Publication number
WO2009135151A1
WO2009135151A1 PCT/US2009/042563 US2009042563W WO2009135151A1 WO 2009135151 A1 WO2009135151 A1 WO 2009135151A1 US 2009042563 W US2009042563 W US 2009042563W WO 2009135151 A1 WO2009135151 A1 WO 2009135151A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
images
mapping
pixels
instructions
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/US2009/042563
Other languages
English (en)
French (fr)
Inventor
Ira Wallace
Dan Caligor
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
EyeIC Inc
Original Assignee
EyeIC Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by EyeIC Inc filed Critical EyeIC Inc
Priority to JP2011507690A priority Critical patent/JP2011520190A/ja
Priority to EP09739953.9A priority patent/EP2286370A4/en
Priority to CA2723225A priority patent/CA2723225A1/en
Priority to AU2009242513A priority patent/AU2009242513A1/en
Priority to US13/501,637 priority patent/US20120294537A1/en
Publication of WO2009135151A1 publication Critical patent/WO2009135151A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/14Transformations for image registration, e.g. adjusting or mapping for alignment of images
    • G06T3/147Transformations for image registration, e.g. adjusting or mapping for alignment of images using affine transformations
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • G06V10/245Aligning, centring, orientation detection or correction of the image by locating a pattern; Special marks for positioning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/757Matching configurations of points or features
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/80Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
    • G06V10/806Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of extracted features

Definitions

  • the invention relates to a system and method for mapping still or moving images of different types and/or different views of a scene or object at the same or different points in time such that a specific object or location in the scene may be identified and tracked in the respective images.
  • the system and method may be applied to virtually any current or future image types, both 2-D and 3-D, both single-frame and multi-frame (video).
  • the system and method also may be used in association with any known method of image alignment by applying the same form of transformation to images as applied by that image alignment method.
  • images may be different types (e.g., x-ray, photograph, line drawing, map, satellite image, etc.), similar image types taken from different perspectives (e.g., different camera angle, rotation, focal length or subject-focal plane relationship), similar or different images taken at different points in time, or a combination of all of these.
  • images may be different types (e.g., x-ray, photograph, line drawing, map, satellite image, etc.), similar image types taken from different perspectives (e.g., different camera angle, rotation, focal length or subject-focal plane relationship), similar or different images taken at different points in time, or a combination of all of these.
  • imaging types may be used with such imaging types or other imaging types that capture and present images of 3-D space (e.g., CAT and MRI, which use multiple 2-D slices) or that create 3-D renderings from 2-D images (e.g., stereoscopic slides such as are used in the current ophthalmology image comparison gold standard or "3-D” technologies such as used in entertainment today).
  • the image types may also include video and film, which are composed of individual images (2-D or stereoscopic).
  • a user may (1) estimate using various visual and intuitive techniques, (2) estimate using mathematical techniques, or (3) use computer image morphing techniques to align and overlay the images using, e.g., flicker chronoscopy, which is used in many other disciplines such as engineering and astronomy to identify change or motion.
  • Each of these techniques has important shortcomings, including relatively low accuracy, being slow or time consuming, requiring high levels of skill or specialized knowledge, and being highly prone to error.
  • An improved technique without such shortcomings is desired.
  • the cross-image mapping (CIM) technique of the invention is designed to increase the ease, speed and accuracy of mapping objects across images for a variety of applications. These include - but are not limited to - flicker chronoscopy for medical tracking and diagnostic purposes, cartographic applications, tracking objects across multiple sequential images or video frames, and many others.
  • the CIM technique of the invention makes it possible to locate specific coordinates, objects or features in one image within the context of another.
  • the CIM technique can be applied to any current or future imaging technology or representation, whether 2-D or 3-D, single-frame (still) images or multi-frame (video or other moving image types).
  • the process can be easily automated, and can be applied in a variety of ways described below.
  • the CIM technique of the invention generally employs three broad steps:
  • mapping and/or alignment parameters to identify and highlight the pixel in one image corresponding to the comparable location in another (i.e., identify the pixel that shows the same location relative to some landmark in each image).
  • the method may also include the ability to indicate the accuracy or reliability of mapped pixel locations.
  • This accuracy or reliability assessment may be based on outputs or byproducts of the alignment algorithm(s) or tool(s) employed in the mapping, or on assessment of aligned images after the fact.
  • Such accuracy or reliability measures may be presented in many ways, including but not limited to visual modification of the mapped marking (through modification of line thickness, color, or other attributes) and quantitative or qualitative indicators inside or outside of the image area (e.g., red/yellow/green or indexed metrics).
  • the scope of the invention includes a method, computer system and/or computer readable medium including software that implements a method for mapping images having a common landmark or common reference point (e.g., global positioning system tags, latitude/longitude data, and/or coordinate system data) therein so as to, for example, enable the creation, location and/or mapping of pixels, coordinates, markings, cursors, text and/or annotations across aligned and/or unaligned images.
  • a common landmark or common reference point e.g., global positioning system tags, latitude/longitude data, and/or coordinate system data
  • the computer- implemented method includes selecting at least two images having the common landmark or common reference point, mapping the selected images so as to generate mapping parameters that map a first location on a first image to the corresponding location of the first location on a second image, and identifying at least one pixel on the first image and applying the mapping parameters to at least one pixel on the first image to identify the corresponding pixel or pixels in the second image.
  • the mapping parameters then may be used to locate or reproduce any pixels, coordinates, markings, cursors, text and/or annotations of the first image at the corresponding location of the second image.
  • the two images may be of different image types including: x-ray image, photograph, line drawing, map image, satellite image, CAT image, magnetic resonance image, stereoscopic slides, video, and film.
  • the images also may be taken from different perspectives and/or at different points in time.
  • the images may be aligned using an automated image matching algorithm that aligns the first and second images and generates alignment parameters, or a user may manually align the first and second images by manipulating one or both images until they are aligned.
  • Manual or automatic landmark mapping may also be used to identify the common landmark in the first and second images.
  • associated software may generate mapping parameters based on the locations in the first and second images of the common landmark.
  • the first image may be morphed to the second image whereby the common landmark in each image has the same coordinates.
  • an indication of a degree of accuracy of the alignment and/or mapping of the selected images at respective points in an output image may also be provided.
  • Such indications may include means for visually distinguishing displayed pixels for different degrees of reliability of the alignment and/or mapping of the display pixels at respective points.
  • different colors or line thicknesses may be used in accordance with the degree of reliability of the alignment and/or mapping at the respective points or, alternatively, a numerical value for points on the output image pointed to by a user input device.
  • the mapping may also be extended to pixels on at least one of the images that is outside of an area of overlap of the first and second images.
  • Figure 1 illustrates images of a single, unchanged object where unaligned image A illustrates the object as taken straight on from a specific number of feet and unaligned image B illustrates the same object from a lower vantage, further away, with the camera rotated relative to the horizon, and a different placement of the object in the image.
  • Figure 2 illustrates how image B is modified to correspond to image A.
  • Figure 3 illustrates the mapping parameters for mapping unaligned image B to unaligned image A.
  • Figure 4a illustrates the mapping of a user-drawn circle at a user-defined location from the input image B to the output image (aligned image B or input image A).
  • Figure 4b illustrates the application of alignment parameters (e.g. lines) to the images to indicate shift by mapping "before and after" marks from two or more images onto the marked images or other images from the image set.
  • alignment parameters e.g. lines
  • Figure 5 illustrates two types of images of the same object where common identifying features are provided in each image.
  • Figure 6 illustrates the alignment of the images of Figure 5 using a common feature by modifying one or more of the images to compensate for camera angle, etc. using a manual landmark application or an automated algorithm.
  • Figure 7 illustrates the parameters for mapping from one image in a set to another, based on alignment of the two images (note the parameters are the same as in Figure 3 except that the images are not aligned).
  • Figure 8 illustrates the mapping of a user-entered input marking in image A to image B or aligned image B.
  • Figure 9 illustrates an exemplary computer system for implementing the CIM technique of the invention.
  • Figure 10 illustrates a flow diagram of the CIM software of the invention.
  • Figure 11 illustrates the operation of a sample landmark tagging application in accordance with the invention whereby corresponding landmarks are identified in two images either manually or through automation.
  • Figure 12 illustrates the expression of a given "location" or "reference point” in an image in terms of a common landmark or by using a convention such as the uppermost left-hand pixel in the overlapping area of aligned images.
  • Figure 13 illustrates examples of displaying accuracy or reliability in the comparison of images using the CIM techniques of the invention.
  • Figure 14 illustrates aligned and mapped images in which image A covers a small portion of the area covered by image B, and illustrates a means for identifying coordinates of a landmark in image B relative to image A coordinate system but beyond the area covered by image A.
  • the CIM technique of the invention employs computer-enabled image morphing and alignment or, alternatively, mapping through landmark tagging or other techniques, as the basis of its capabilities. Specifically, two or more images are aligned and/or mapped to each other such that specific landmarks in either image fall in the same spot on the other. It is noted that the alignment may be of only part of each of the images. For example, the images may depict areas with very little common overlap, such as images of adjacent areas. In addition, one image may cover a small area included in a second, larger area covered by the second image.
  • the landmarks or pixels shown in the overlap area though bearing the same relationship to each other in both images and ostensibly representative of the same spot in space, might fall in very different locations within the image relative to the center, corner or edge.
  • This alignment can be achieved by changing one image to match the other or by changing both to match a third, aligned image (in the case of multiple input images or video images, the same principles are applied several times over) or by mapping the images (mapping one image to match the other or by mapping both to match a third, aligned image) to each other without actually changing the images.
  • mapping software such as Photoshop
  • technologies such as the Dual Bootstrap algorithm.
  • the end result is (1) a set of images including two or more unaligned images (unaligned image A, unaligned image B, and so on) and two or more aligned images (aligned image A, aligned image B, and so on), such that the aligned images can be overlaid and only those landmarks that have moved or changed will appear in different pixel locations and/or (2) a set of parameters for mapping one or more image in the set to another such that this mapping could be used to achieve alignment as in (1). It is important to note that the CIM technique is indifferent to the mechanism used for aligning or mapping images and does not purport to accomplish the actual alignment of images.
  • aligning means to transform a first image so that it overlays a second image.
  • image alignment may include the modification of one or more images to the best possible consistency in pixel dimensions (size and shape) and/or location of specified content within the image (e.g. , where only part of images are aligned).
  • mapping means to identify a mathematical relationship that can be used to identify spot or pixel in one image that corresponds to the spot or pixel in another image. It is not necessary to modify either image to establish mapping. For example, mapping is used to create alignment and/or to represent the operations performed to achieve alignment. Mapping parameters are the output of the mapping operation and are used to perform the calculations of pixel locations when performing landmark tagging.
  • this common pixel can be located in each image in any of several ways, typically (but not limited to) relative to a specified landmark or relative to a common reference point or pixel (e.g., uppermost left-hand) in the overlapping portion of the images.
  • an "input image” is the image used to identify the pixel or location used for generating mapping
  • an “output image” is the image upon which the mapped pixel or location is located and/or displayed. Both input and output images may be aligned or unaligned.
  • mark tagging refers to various forms of identifying common landmarks or registration points in unaligned images, either automatically (e.g. through shape recognition) or manually (i.e. user-identified).
  • the CIM technique of the invention first creates formulae for mapping a specific location within any image in the input or aligned image sets to any other in the set.
  • the formulae contain parameters for shifting image centers (or other reference point) up/down and left/right, rotating around a defined center point, stretching one or more axes or edges or dimensions to shift perspective, and so on.
  • These formulae can (1) be captured as part of the output of automated alignment algorithms such as Dual Bootstrap, or (2) be calculated using landmark matching in a landmark tagging or other conventional application.
  • the landmark tagging application will present the user with two or more images, allow the user to "tag" specific, multiple landmarks in each of the images, and use the resulting data to calculate the formulae or parameters that enable a computer program to map any given point in an image to the comparable point in another image within the set.
  • landmark tagging may be achieved through automated processes using shape recognition, color or texture matching, or other current or future techniques.
  • the user selects two (or more) images from the image set for mapping. These may be all aligned images, a mix of unaligned and aligned images, or all unaligned images. These may be a mix of image types, for example drawings and photographs, 2-D video frames and 3-D still or moving renderings, etc. (e.g., CAT, MRI, stereoscopic slides, video, or film).
  • the selected images are displayed by the landmark tagging application in any of several ways (e.g., side by side, or in overlapping tabs).
  • the user may then identify a pixel, feature or location in one of the selected images (the input image), and the CIM application will identify and indicate the corresponding pixel (same object, landmark or location) in the other selected images (output images).
  • the manner of identification can be any of several, including clicking with a mouse or pointing device, drawing shapes, lines or other markings, drawing freehand with an appropriate pointing device, marking hard copies and scanning, or other computer input techniques.
  • Selected pixels or landmarks can be identified with transient indicators, by translating the lines or shapes from the input image into corresponding display in the output image, or by returning coordinates in the output image in terms of pixel location or other coordinate system.
  • the input image can be an unaligned or aligned image, and the output image(s) can also be either unaligned or aligned.
  • two or more images are selected for mapping. These input images may have differences in perspective, camera angle, focal length or magnification, rotation, or position within a frame.
  • Figure 1 illustrative images of a single, unchanged object are shown.
  • input image A the object is shown as taken straight on from a specific distance (e.g., 6 feet), while input image B illustrates the same object from a lower vantage point, further away, with the camera rotated relative to the horizon, and a different placement of the object in the image.
  • the object will have changed shape, size, or position relative to other landmarks.
  • Parameters for aligning and/or mapping the images are calculated. If an automated matching/morphing algorithm such as Dual Bootstrap is used, this process is automated. Alternatively, if manual landmark tagging is used, the user identifies several distinctive landmarks in each image and "tags" them (e.g., see the triangles in Figure 2). Finally, the images may be aligned through manual morphing/stretching (such as in Photoshop transforms). In either case, the best alignment or mapping possible is established. It is noted that, in some cases, some features may not align or map across images but that cross-image mapping may still be desirable.
  • a photograph of a small area of landscape may be mapped to a vector map covering a far larger area (e.g., see Figure 14).
  • Figure 2 shows how the input image B is modified to correspond to input image A.
  • the resultant alignment and/or mapping parameters are recorded and translated into appropriate variables and/or formulae for aligning and/or mapping any two images in the image set.
  • the mapping parameters for mapping input image B to input image A are shown. Typically, these parameters are expressed as a series of mathematical equations.
  • Alignment and/or mapping parameters are applied to align and/or map a location in one image to the equivalent location in another image within the image set.
  • a specific spot along the edge of a shape has been circled by the user, and the CIM application displays the equivalent shape on another image from the set (note that on the output image, the circle is foreshortened due to morphing for alignment).
  • the item tagged may be a specific pixel, a circle or shape, or other form of annotation.
  • Alignment and/or mapping parameters also may be applied to indicate shift by mapping "before and after" marks from two or more images onto the marked images or other images from the image set.
  • two lines are drawn by the user (e.g., tracing the edge of a bone structure in an x-ray), and the two lines are plotted together onto a single image.
  • the image onto which the lines are plotted may be a third image from the same set and that more than two markings and/or more than two images may be used for this technique.
  • the drawing of the lines may be automated using edge detection or other techniques.
  • mapping two images of different types may be used as input images, and a shared element of the two images may be used to calculate mapping parameters. Examples of this form include combining (1) x-rays and photos of tissue in the same spot, (2) photos and line maps or vector maps such as those used in Computer Aided Mapping (CAM) applications used to track water or electrical conduits beneath streets, (3) infrared and standard photographs, or (4) aerial or satellite photographs and assorted forms of a printed or computerized map.
  • CAM Computer Aided Mapping
  • a common feature is used to align and/or map - and if necessary conform through morphing - two or more images.
  • Examples of a common feature include: teeth visible in both a dental x-ray and a dental photograph; buildings visible in photographs and CAM maps; known coordinates in both images, e.g., a confluence of rivers or streets or latitude and longitude.
  • input image A in Figure 5 may represent image type A such as an x-ray of teeth or a vector drawing such as in a CIM map.
  • the illustrated white shape may be an identifying feature such as a tooth or a building.
  • Input image B may represent image type B such as a photo of tissue above an x-ray or an aerial photo of an area in a vector map.
  • the white shape may be an identifying feature such as a tooth or building.
  • Figure 6 illustrates the alignment of the input images using the common feature (e.g., tooth or building) by morphing one or more of the images to compensate for camera angle, etc. using a CIM landmark tagging application, an automated algorithm, or using manual alignment (e.g., moving the images around in Photoshop until they align). In some cases, alignment and/or mapping may be achieved automatically using shapes or other features common to both images (such as teeth in the above example).
  • the parameters for mapping from one image in a set to another are calculated and expressed as a series of mathematical equations as shown in Figures 3 and 7.
  • mapping capability can now be used to identify the location of a landmark or point of interest in one image within the area of another from the set. This is illustrated in Figure 8, where a user-entered input marking in input image A is mapped to output image B using the techniques of the invention. If required, morphing of images may be applied in addition to re-orientation, x,y shift, rotation, and so on.
  • mapping technique of the invention need not be limited to mapping visible markings. It could, for instance, be used to translate cursor location when moving a cursor over one image to the mapped location in another image.
  • FIG. 9 illustrates an exemplary computer system for implementing the CIM technique of the invention.
  • a microprocessor 100 receives two or more user-selected input images 110 and 120 and processes these images for display on display 130, printing on printer 132, and/or storage in electronic storage device 134.
  • Memory 140 stores software including matching algorithm 150 and landmark tagging algorithm 155 that are optionally processed by microprocessor 100 for used in aligning the images and to generate and capture alignment parameters. Matching algorithm 150 and landmark tagging algorithm 155 may be selected from conventional algorithms known by those skilled in the art.
  • CIM software 160 in accordance with the invention is also stored in memory 140 for processing by microprocessor 100.
  • FIG 10 illustrates a flow diagram of the CIM software 160 of the invention.
  • the CIM software 160 enables the user to select two or more images or portions of images at step 200.
  • the selected images are then aligned in step 210 using the automated matching algorithm 150, and alignment parameters (e.g., Figure 3) are generated/captured from the algorithm at step 220.
  • the alignment may also be performed manually by allowing the user to manipulate, reorient and/or stretch one or both images until they are aligned.
  • the mapping would document the manipulation and alignment parameters would be generated at step 220 based on the mapping documentation.
  • landmark tagging also may be used to map images by determining transformations without changing the images at step 230 and generating the mapping parameters generated by the mapping application (e.g., CIM matching algorithm) at step 240.
  • the alignment and/or mapping parameters are used to define formulae for aligning/mapping between all image pairs in a set of images (e.g., unaligned-unaligned, unaligned-aligned, aligned-aligned).
  • a pixel or pixels on any image in an image set (e.g., an input image) is then identified at step 260 and the afore-mentioned formulae are applied thereto to identify the corresponding pixel or pixels in other images in the image set (e.g., output images) at step 270.
  • the pixel location is mapped, any markings, text or other annotations entered on an input image are optionally reproduced on one or more output images, the pixel location is identified and/or displayed, and/or pixel coordinates are returned at step 280.
  • the degree of accuracy or reliability is calculated and/or displayed to user, as described below in connection with Figure 13.
  • FIG 11 illustrates a sample landmark mapping application used in step 230 in accordance with the invention in which the user selects two or more images that are displayed side-by-side, in a tabbed view, or in some other manner.
  • the user selects landmarks such as corners of the same object in the two images and marks each landmark in each image using a mouse or other input device.
  • the selected landmarks are identified as comparable locations in each image (e.g., by entering numbers or using a point-and- click interface).
  • the CIM software 160 uses the corresponding points to calculate the best formulae for translation from one image to another, including x,y shift of the image(s), rotation, and stretching in one or more dimensions.
  • the images need not be actually aligned; rather, the mapping formulae are used to map pixels, coordinates, markings, cursors, text, annotations, etc. from one image to another using the techniques described herein.
  • Additional levels of functionality may easily be added to the CIM software 160.
  • manual tagging or automated edge detection may be used to identify a specific landmark in two images, as well as a reference landmark of known size (e.g., a foreign object introduced into one image to establish a size reference) or location (e.g., the edge of a bone that has not changed).
  • a CIM application or module within another application can calculate distances or percentage changes between two or more images.
  • Additional information about the mapping may be displayed visually or in other ways. For example, statistical measures of image fit may be used to estimate the accuracy and/or reliability of the mapping, and to display this degree of accuracy or "confidence range" through color, line thickness, quantitative displays or other means. Furthermore, such information may be a function of location within an image (e.g., along an edge that has been greatly stretched versus an edge that has not); these differences may be reflected in the display of such additional information either visually on an image (e.g., through line thickness or color of markings) or through representations such as quantitative measures.
  • a specific pixel in an input image may correspond to a larger number of pixels in the output image (for example, a ratio of 1 pixel to four).
  • the line on the output image may be shown as four pixels wide for every pixel of width in the input image.
  • this can be shown with colors, patterns or other visual indicators by, for example, showing less accurate location mappings in red instead of black, or dotted instead of solid lines.
  • the mapped locations might be one fourth the width; in this case, the line can be shown as using one quarter the pixel width, or as green, or as bold or double line.
  • This approach to showing accuracy of mapping can be based on factors other than resolution.
  • descriptive statistics characterizing the accuracy of alignment may be used, including measures derived from comparison of each pixel in an input and output image, measures derived from the number of iterations, processing time or other indications of "work" performed by the alignment algorithm, and so on. Such statistics may be employed as a measure of accuracy or fit.
  • the uniformity of morphing applied can be used. For instance, if an image is stretched on one edge but not on another, the accuracy can be shown as greatest on the portion of the image that has been stretched the least.
  • any indication of accuracy of alignment, reliability of/confidence in an alignment or other qualifying measures may be used as the basis of indicating these confidence levels.
  • Figure 13 illustrates examples of displaying accuracy or reliability as just described.
  • the input image on the left of the figure
  • the mapping of pixels to the output image will be more accurate for the bottom line than the top line.
  • this can be indicated through changes in the thickness of the line (Output A), the color of the line (Output B), attributes of the line (Output C), or by other, similar means.
  • accuracy or reliability also may be indicated using a quantitative or qualitative display linked to the cursor, as in Output D.
  • the cursor is pointed at various locations in the image and a "score" showing accuracy or reliability of the alignment is shown for that location in the image.
  • GPS global positioning system
  • GIS global information system
  • coordinates may be extended beyond the area of overlap in the one or more images.
  • the CIM technique of the invention may be used to infer the location of a pixel or object in image B based on extrapolation of coordinates attached to image A and mapped to image B using the overlapping area.
  • Figure 14 illustrates aligned and mapped images in which image A covers a small portion of the area covered by image B.
  • image A has associated coordinate data (e.g. latitude/longitude) and image B does not.
  • a location in image B outside of the area of overlap with image A is selected as an input location, making image B the input image.
  • the common landmark in the overlap area is at known coordinates in image A.
  • CIM the parameters for mapping the overlapped areas are known and by extension areas that do not overlap are known. This allows one to establish the location of any pixel in image B by (1) applying the image A coordinate data within the overlap area to image B within the overlap area, and (2) extending the mapping beyond the area of overlap to infer the coordinates within the image A coordinate system of a pixel in image B, even if it is outside of the area covered by image A.
  • the output location cannot be shown on image A but can be expressed in the coordinate system applied to image A.
  • CIM can be used to establish mappings outside the area of overlap.
  • MRI, CT, stereoscopic photographs, various forms of 3-D video or other imaging types may all have CIM techniques applied to and between them.
  • CIM techniques applied to and between them.
  • an MRI and CT scan can be mapped using CIM techniques, allowing objects visible in one to be located within the other.
  • the structures that appear to have moved or changed in the respective input images may be located on the input images using the technique of the invention.
  • structures or baselines e.g., jaw bone in dental images
  • the technique may also be used to show corresponding internal and external features in images (e.g., abscesses on x-rays or gum surface in dental x-rays). This technique may also be used to show structures or baselines in successive frames of a video or other moving image source.
  • a frame from a video of a changing perspective may be aligned to a map or satellite image.
  • landmark tagging Once landmark tagging has been established, a given object in the video may be tracked in subsequent frames of the video by applying landmark tagging or other techniques establishing mapping parameters to the subsequent frames of the video.
  • the CIM techniques described herein may be employed within a single moving image source by applying the technique to successive frames.
  • a moving object in a video from a stationary perspective may be identified using landmark tagging or other techniques establishing mapping parameters and then tracked from frame to frame using successive applications of CIM .
  • a stationary object in a video taken from a moving perspective e.g., from an aircraft
  • CIM applications or CIM modules within other applications include:
  • a patient may be exhibiting jaw bone loss, a very common problem.
  • the doctor may compare two or more dental x-rays of the same area of the patient's jaw taken months apart. By marking the bone line in one and using CIM to map this marking to other images, the doctor, patient or other parties can see how much the bone has moved during the period between image captures, thus quantifying both the pace and magnitude of change.
  • the doctor could highlight the bone line along the top of the bottom jaw in each of the two images as well as a baseline (for example the bottom edge of the bottom jaw).
  • the CIM application could then calculate bone loss as a percentage of total bone mass.
  • a reference object could be included in one or more images, and the CIM application could then express bone loss in millimeters.
  • Examples of how this technique might be employed include using the overlay of a CIM map of gas mains and an aerial photo of a city block to pinpoint a location for digging which can be found by workers using landmarks rather than surveying equipment.
  • GPS coordinates associated with one or both images may be used to identify additional images or areas of images contained in databases with which to align.
  • This application can also use various measures of the accuracy and precision of alignment to indicate precision of mapping.
  • the technique may also be used to examine the bones underneath a specific area of inflamed tissue or to locate a specific object visible in one photograph by mapping it against a shared feature in a map or alternate photograph.
  • the input can take a variety of forms.
  • Input mechanisms include (1) drawing lines, shapes or other markings using a mouse, touch-screen or other input device, so they are visible on the input image, (2) drawing lines, shapes or other markings using a mouse, touch-screen or other input device so they are not visible on the input image, (3) entering coordinate data such as latitude/longitude or map grids, such that specific pixels are identified on an input image with such associated coordinates, or (4) entering other types of information associated with specific locations within an image. Examples of other types of information include altitude data on a topographical map or population density in a map or other database.
  • the form of input could be to specify all areas corresponding to a specific altitude or range of altitudes, or to a specific population density or range of population densities.
  • Other means of input either existing or invented in the future, may be used to achieve the same result.
  • the output can take a variety of forms. These include (1) showing lines, shapes or other markings on the output image, (2) returning the pixel location(s) of corresponding pixels in the output image, (3) returning latitude and longitude or other coordinates associated with pixels or specific locations in the output image, and (4) other forms of information associated with specific locations within an image.
  • some input or output methods do not require the display of one or both images to be effective.
  • the location to be mapped may be indicated by inputting appropriate coordinates, or alternatively values such as altitude ranges or population densities even if the input image is not displayed. These locations may then be displayed or otherwise identified or indicated in the output image.
  • these coordinates can be identified or returned, without the output image itself being displayed.
  • the user may then identify a feature or location in one of the selected images (the input image), and the CIM application will identify and indicate the corresponding pixel (same object, landmark or location) in a second selected image (output image).
  • the manner of identification may be any of several, as described above.
  • Selected pixels or landmarks may be identified with transient indicators or by translating the lines or shapes from the input image into corresponding display in the output image, or by returning coordinates or other location indicators in the output image.
  • the input image may be either an aligned or unaligned image, and the output image(s) also may be either an unaligned or aligned image.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Image Processing (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)
  • Image Analysis (AREA)
  • Magnetic Resonance Imaging Apparatus (AREA)
PCT/US2009/042563 2008-05-02 2009-05-01 System for using image alignment to map objects across disparate images Ceased WO2009135151A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
JP2011507690A JP2011520190A (ja) 2008-05-02 2009-05-01 画像アラインメントを使用してオブジェクトを別画像に亘ってマッピングするシステム
EP09739953.9A EP2286370A4 (en) 2008-05-02 2009-05-01 SYSTEM FOR USING IMAGE DEFINITION FOR THE PICTURE OF OBJECTS IN UNEVEN PICTURES
CA2723225A CA2723225A1 (en) 2008-05-02 2009-05-01 System for using image alignment to map objects across disparate images
AU2009242513A AU2009242513A1 (en) 2008-05-02 2009-05-01 System for using image alignment to map objects across disparate images
US13/501,637 US20120294537A1 (en) 2008-05-02 2009-05-01 System for using image alignment to map objects across disparate images

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US4995408P 2008-05-02 2008-05-02
US61/049,954 2008-05-02

Publications (1)

Publication Number Publication Date
WO2009135151A1 true WO2009135151A1 (en) 2009-11-05

Family

ID=41255452

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2009/042563 Ceased WO2009135151A1 (en) 2008-05-02 2009-05-01 System for using image alignment to map objects across disparate images

Country Status (6)

Country Link
US (1) US20120294537A1 (enExample)
EP (1) EP2286370A4 (enExample)
JP (1) JP2011520190A (enExample)
AU (1) AU2009242513A1 (enExample)
CA (1) CA2723225A1 (enExample)
WO (1) WO2009135151A1 (enExample)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2013202286B2 (en) * 2012-03-30 2016-02-25 Samsung Electronics Co., Ltd. A method of preserving tags for edited content
GB2536274A (en) * 2015-03-12 2016-09-14 Mirada Medical Ltd Method and apparatus for assessing image registration
CN117115570A (zh) * 2023-10-25 2023-11-24 成都数联云算科技有限公司 一种基于Canvas的图像标注方法和系统

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8787651B2 (en) * 2010-09-28 2014-07-22 Flagship Biosciences, LLC Methods for feature analysis on consecutive tissue sections
JP6049882B2 (ja) * 2013-07-18 2016-12-21 三菱電機株式会社 目標類識別装置
US20150134361A1 (en) * 2013-11-08 2015-05-14 The Cleveland Clinic Foundation Graphical generation and retrieval of medical records
FR3030091B1 (fr) * 2014-12-12 2018-01-26 Airbus Operations Procede et systeme de detection automatique d'un desalignement en operation d'un capteur de surveillance d'un aeronef.
KR101784063B1 (ko) * 2015-08-03 2017-10-11 국립암센터 펜타입의 의료용 형광 이미지 장치 및 이를 이용한 다중 형광 영상의 정합 시스템
US10491778B2 (en) 2017-09-21 2019-11-26 Honeywell International Inc. Applying features of low-resolution data to corresponding high-resolution data
US10778916B2 (en) 2018-10-24 2020-09-15 Honeywell International Inc. Applying an annotation to an image based on keypoints
KR102195179B1 (ko) * 2019-03-05 2020-12-24 경북대학교 산학협력단 항공사진을 이용한 정사영상 구축방법
CN110986969B (zh) * 2019-11-27 2021-12-28 Oppo广东移动通信有限公司 地图融合方法及装置、设备、存储介质
CN111599007B (zh) * 2020-05-26 2021-05-25 王梅莹 一种基于无人机航拍的智慧城市cim道路贴图方法
KR102498781B1 (ko) * 2021-01-15 2023-02-13 국방과학연구소 영상 내 객체에 대한 장면 그래프를 생성하는 방법 및 이를 이용한 전자 장치
EP4036856B1 (en) 2021-02-02 2025-04-30 Axis AB Updating of annotated points in a digital image
US11816174B2 (en) * 2022-03-29 2023-11-14 Ebay Inc. Enhanced search with morphed images

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050094858A1 (en) * 2003-10-31 2005-05-05 Sirohey Saad A. Method and apparatus for synchronizing corresponding landmarks among a plurality of images
US20070127787A1 (en) * 2005-10-24 2007-06-07 Castleman Kenneth R Face recognition system and method

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2900632B2 (ja) * 1991-04-19 1999-06-02 株式会社日立製作所 ディジタル地図処理装置及びディジタル地図の表示方法
JP3406785B2 (ja) * 1996-09-26 2003-05-12 株式会社東芝 心機能解析支援装置
JP2001160133A (ja) * 1999-12-01 2001-06-12 Hitachi Ltd コンピュータマッピングシステム
US8010180B2 (en) * 2002-03-06 2011-08-30 Mako Surgical Corp. Haptic guidance system and method
WO2004069040A2 (en) * 2003-02-04 2004-08-19 Z-Kat, Inc. Method and apparatus for computer assistance with intramedullary nail procedure
US7574032B2 (en) * 2003-10-31 2009-08-11 General Electric Company Method and apparatus for virtual subtraction of stool from registration and shape based analysis of prone and supine scans of the colon
JP4677199B2 (ja) * 2004-04-14 2011-04-27 株式会社日立メディコ 超音波診断装置
GB0424737D0 (en) * 2004-11-09 2004-12-08 Isis Innovation Method,computer program and signal processing apparatus for determining statistical information of a signal
US8014625B2 (en) * 2004-11-10 2011-09-06 Agfa Healthcare Method of performing measurements on digital images
EP1815432B1 (en) * 2004-11-17 2016-08-03 Koninklijke Philips N.V. Improved elastic image registration functionality
JP4820680B2 (ja) * 2006-04-12 2011-11-24 株式会社東芝 医用画像表示装置
US8218905B2 (en) * 2007-10-12 2012-07-10 Claron Technology Inc. Method, system and software product for providing efficient registration of 3D image data

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050094858A1 (en) * 2003-10-31 2005-05-05 Sirohey Saad A. Method and apparatus for synchronizing corresponding landmarks among a plurality of images
US20070127787A1 (en) * 2005-10-24 2007-06-07 Castleman Kenneth R Face recognition system and method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP2286370A4 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2013202286B2 (en) * 2012-03-30 2016-02-25 Samsung Electronics Co., Ltd. A method of preserving tags for edited content
GB2536274A (en) * 2015-03-12 2016-09-14 Mirada Medical Ltd Method and apparatus for assessing image registration
GB2536274B (en) * 2015-03-12 2019-10-16 Mirada Medical Ltd Method and apparatus for assessing image registration
CN117115570A (zh) * 2023-10-25 2023-11-24 成都数联云算科技有限公司 一种基于Canvas的图像标注方法和系统
CN117115570B (zh) * 2023-10-25 2023-12-29 成都数联云算科技有限公司 一种基于Canvas的图像标注方法和系统

Also Published As

Publication number Publication date
JP2011520190A (ja) 2011-07-14
CA2723225A1 (en) 2009-11-05
AU2009242513A1 (en) 2009-11-05
US20120294537A1 (en) 2012-11-22
EP2286370A1 (en) 2011-02-23
EP2286370A4 (en) 2014-12-10

Similar Documents

Publication Publication Date Title
US20120294537A1 (en) System for using image alignment to map objects across disparate images
CN100538714C (zh) 数字医学图像中自动映射几何对象的方法
US8463071B2 (en) Navigating images using image based geometric alignment and object based controls
Karsch et al. ConstructAide: analyzing and visualizing construction sites through photographs and building models
JP4537557B2 (ja) 情報呈示システム
US7746377B2 (en) Three-dimensional image display apparatus and method
EP2546806B1 (en) Image based rendering for ar - enabling user generation of 3d content
US11900541B2 (en) Method and system of depth determination with closed form solution in model fusion for laparoscopic surgical guidance
CA2703423C (en) Pitch determination systems and methods for aerial roof estimation
US9014438B2 (en) Method and apparatus featuring simple click style interactions according to a clinical task workflow
CA2819166A1 (en) Systems and methods for processing images with edge detection and snap-to feature
EP1709394A1 (en) Transprojection of geometry data
US20100034485A1 (en) Computer vision system and language
JP2015228215A (ja) 位置情報処理方法
CN112001889A (zh) 医学影像处理方法、装置及医学影像显示方法
CN100392677C (zh) 显示截面图像的装置及方法
Sapirstein Photogrammetry as a tool for architectural analysis: the digital architecture project at Olympia
Cucchiara et al. An image analysis approach for automatically re-orienteering CT images for dental implants
JP2005310044A (ja) データ処理装置、データ処理方法、及びデータ処理プログラム
CN116269831A (zh) 基于全息影像的手术辅助系统
Abdelhafiz et al. Automatic texture mapping mega-projects
AU2018217240A1 (en) Pitch determination systems and methods for aerial roof estimation
Morago Multi-modality fusion: registering photographs, videos, and LIDAR range scans
Adcox The Utility of Digital Imaging Technologies for the Virtual Curation and Metric Analysis of Skeletal Remains
CN114943814A (zh) 一种三维扫描辅助面部建模分析系统

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09739953

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2009242513

Country of ref document: AU

WWE Wipo information: entry into national phase

Ref document number: 2723225

Country of ref document: CA

WWE Wipo information: entry into national phase

Ref document number: 2011507690

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2009242513

Country of ref document: AU

Date of ref document: 20090501

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 2009739953

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 13501637

Country of ref document: US