US20040202357A1 - Silhouette image acquisition - Google Patents
Silhouette image acquisition Download PDFInfo
- Publication number
- US20040202357A1 US20040202357A1 US10/413,493 US41349303A US2004202357A1 US 20040202357 A1 US20040202357 A1 US 20040202357A1 US 41349303 A US41349303 A US 41349303A US 2004202357 A1 US2004202357 A1 US 2004202357A1
- Authority
- US
- United States
- Prior art keywords
- image
- desired feature
- sample
- scan
- subimage
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/695—Preprocessing, e.g. image segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10056—Microscopic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20016—Hierarchical, coarse-to-fine, multiscale or multiresolution image processing; Pyramid transform
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20021—Dividing image into blocks, subimages or windows
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30024—Cell structures in vitro; Tissue sections in vitro
Definitions
- the present application describes systems and techniques relating to image data acquisition.
- Anatomic pathology enables the diagnosis of physiological conditions such as cancer, infectious disease and prenatal disorders through analysis of tissue samples.
- a biological specimen from a patient is obtained, imaged, and analyzed.
- the specimen may be “stained” using a material chosen to reveal the presence of one or more molecules.
- the specimen may then be placed on a slide for viewing with a microscope.
- An image of the sample may be obtained using a digital camera.
- the sample may be scanned, and a series of images may be obtained.
- the image data may then be analyzed.
- an image may be analyzed to identify the presence of particular characteristic features of a particular disease, where the particular characteristic features indicate the presence of or predisposition towards the disease.
- a method includes obtaining an image of at least a portion of a scan region, and analyzing the image to determine if the image includes a first desired feature. If the image includes the first desired feature, the method may include electronically dividing the image into a plurality of subimages at the same magnification and determining the presence or absence of a second desired feature in each subimage.
- the first desired feature and the second desired feature may be the same or different.
- the first desired feature may be the presence of at least a portion of a sample in the scan region.
- the method may include analyzing images and/or subimages to determine the presence or absence of other features besides the first and second desired features.
- the image may be acquired at a first magnification, which may comprise a field of view smaller than the scan region.
- the plurality of subimages may each comprise an area that is substantially equal to the field of view of a higher magnification.
- the method may further include acquiring an image of an area of the scan region including the second desired feature at higher magnification. For example, one or more higher magnification images may be obtained, so that each subimage including the second desired feature is imaged at the higher magnification.
- the scan region may be a microscope slide or a portion of a microscope slide.
- the microscope slide may have a sample thereon, or may have more than one sample (e.g., a tissue array).
- the sample may be a biological sample.
- the method may include storing an identifier for the image, as well as data associated with the identifier, where the data indicates the presence or absence of the first desired feature.
- the method may include storing an identifier for the subimages, as well as data associated with the identifier, where the data indicates the presence or absence of the second desired feature.
- a system in one aspect, includes an image acquisition system, which may include a camera and/or microscope.
- a data processor may be coupled with the image acquisition system.
- the data processor may receive image data from the image acquisition system, and may determine the presence or absence of a desired feature in the image. If the data processor determines the presence of the desired feature, the data processor may electronically divide the image data into a plurality of subimage portions. The data processor may determine the presence or absence of the desired feature, another feature, or both in the plurality of subimage portions.
- a method may include obtaining a first focal length at a first position comprising a sample, obtaining a second focal length at a second position comprising a sample furthest from the first position, and obtaining a third focal length at a third position comprising a sample at a position furthest from the first and second positions.
- the method may include generating a focal plane comprising the focal length of the first, second, and third positions.
- FIG. 1 shows a block diagram of a system that may be used to acquire image data.
- FIG. 2 is a flowchart showing a process for acquiring and analyzing image data for a plurality of areas.
- FIG. 3 is a flowchart showing a process of acquiring and analyzing image data for a plurality of subdivisions.
- FIG. 4 is a flowchart showing a process of acquiring and analyzing image data for a region designated as interesting.
- FIGS. 5A and 5B show a system that may be used to acquire image data.
- FIGS. 6A and 6B show a microscope subsystem for acquiring image data.
- FIG. 7 shows a slide including a scan region.
- FIG. 8 shows an expanded view of a scan region.
- FIG. 9 shows another expanded view of a scan region.
- FIG. 10 shows a FIND silhouette.
- FIG. 11 shows a SCAN silhouette
- FIG. 12 illustrates a method of performing a silhouette scan operation.
- the systems and techniques described herein relate to image acquisition.
- the systems and techniques provide for efficient and accurate imaging and analysis of materials such as biological samples.
- the challenges of obtaining and processing images are interconnected.
- at least a portion of the sample may need to be viewed or analyzed at high resolution.
- high resolution images are generally very large data files. If the entire sample were to be imaged and image data stored at high resolution, the data storage and analysis challenge may be greater than the ability of an available data processing system.
- a digital representation of an entire microscope slide obtained using a high power lens may include 100 to 400 separate images. Obtaining, storing, and analyzing such a large number of digital images may be prohibitive in many applications.
- a user views a low resolution image of, for example, a microscope slide.
- the user defines an area of interest on the slide (e.g., the outline of a tissue sample) to be imaged at higher resolution.
- an area of interest on the slide e.g., the outline of a tissue sample
- these systems require substantial user involvement and may also require that the user be trained to recognize areas of interest.
- FIG. 1 shows a system 100 that may be used to acquire image data, according to various implementations.
- a microscope 110 such as an optical microscope, receives a sample holder with a scan region, where the scan regions includes a sample to be imaged.
- a slide input 120 may provide one or more microscope slides including (for example) a tissue sample to microscope 110 .
- Microscope 110 may also be coupled with a slide output 130 to receive slides that have been processed.
- An image acquisition system 140 which includes a digital camera or other image data acquisition element, is coupled with microscope 110 .
- Microscope 110 and image acquisition system 140 may be coupled with one or more data processing systems such as a data processor 150 .
- a display 160 may be coupled with data processor 150 and/or image acquisition system 140 , for providing a visual display of an image and/or image data. Note that the elements of FIG. 1 are shown separately; however, some or all of them may be integrated. For example, slide input 120 and slide output 130 may be integrated with microscope 110 rather than separate.
- FIG. 2 shows a method 200 for acquiring image data using a system such as system 100 of FIG. 1.
- a sample holder with a scan region is provided to a microscope ( 210 ).
- the sample holder may be a microscope slide, while the scan region may be an area of the microscope slide in which a sample (e.g., a tissue specimen) may be found.
- the sample holder may be a microscope slide, while the scan region may include an array of specimens (such as an array of tissue samples).
- Image data for a plurality of areas in the scan region may be acquired ( 220 ).
- Image data for a particular area may be processed to determine if the particular area includes a desired feature ( 230 , 240 ). If the particular area includes a desired feature, the area is designated as “interesting” ( 250 ). If the particular area does not include a desired feature, the area is designated as “non-interesting” ( 260 ).
- the coordinates (e.g., x-y coordinates) of “interesting” areas may be stored for later review and/or processing.
- Method 200 may include processing the image data for each of the plurality of imaged areas, and designating each of the areas as either interesting or non-interesting.
- the plurality of areas may form an array covering the scan region.
- a first silhouette including the interesting areas of the array may be formed.
- the first silhouette forms a low-resolution image of the sample in the scan region. For example, when the sample holder is a microscope slide with a specimen on the scan region of the slide, the scan region may be divided into an array of areas, and each area that includes at least a portion of the specimen will be deemed “interesting” and thus included in the first silhouette.
- Data associated with the first silhouette may be stored in a memory such as a memory included in data processor 150 .
- the memory may store a table including an identifier for each location, where each interesting location is designated as a logical 1, and each non-interesting location is designated as a logical 0.
- the system may scan all the areas; for example, by scanning across each of the rows or columns of the array in turn. In some implementations, the system may scan fewer than all areas. For example, if the specimen is a contiguous tissue sample on a slide, the system may find a portion of the boundary of the tissue sample, and may use this information to scan only some of the areas.
- the areas in the first silhouette may be imaged at higher resolution to obtain desired information about the sample in the scan region. That is, the interesting regions (those that include the desired feature) may be imaged using a higher magnification setting of a microscope, and the subsequent image data may be compressed and stored. However, this may be inefficient (since the interesting areas may include only a small portion of the sample) and may require significant data storage and data analysis. Therefore, in other implementations, the interesting areas of the scan region may be sub-divided and the system may determine which subdivisions include one or more desired features.
- FIG. 3 shows a method 300 that may be used to subdivide the areas of the scan region.
- An interesting area, identified at block 250 may be subdivided into a plurality of subdivisions ( 310 ).
- the image data e.g., a portion of the image data acquired in block 220
- the image data for the particular subdivision may then be processed to determine if the particular subdivision includes a desired feature ( 320 , 330 ). If the particular subdivision includes the desired feature, it may be designated as “interesting” ( 340 ). If the particular subdivision does not include the desired feature, it may be designated as “non-interesting” ( 350 ).
- the coordinates of “interesting” subdivisions can be stored for later review and/or processing at a higher magnification.
- a subdivision also termed “minor-zelle”) area is approximately equal to the field of view of a higher power objective of a microscope.
- each of the subdivisions may be analyzed; in others, less than all subdivisions are analyzed.
- Method 300 may be used to form a second silhouette of higher resolution than the first silhouette. That is, the second silhouette more clearly defines the region in which the sample is located (or in which a different desired feature is present).
- the interesting subdivisions in the second silhouette may be imaged at higher magnification to obtain desired information about the sample in the scan region.
- additional subdivisions may be desired, where each subdivision step generates a higher resolution silhouette than the previous step.
- FIG. 4 shows a method 400 that may be used to image a silhouette, where the areas, subdivisions, etc. of the silhouette are referred to generally as regions. That is, if the silhouette is a first silhouette, the regions correspond to the areas discussed above, if the silhouette is a second silhouette, the regions correspond to the sub-divisions discussed above, if the silhouette is a third silhouette, the regions correspond to sub-sub-divisions, and so on.
- An image of a first region is acquired ( 410 ).
- the image may be processed using one or more image processing algorithms ( 420 ).
- the image may also be analyzed using one or more image analysis algorithms ( 430 ).
- method 400 is repeated for all interesting regions in the silhouette. In others, fewer than all interesting regions may be imaged. For example, if a tissue sample is being analyzed to determine the presence of a particular feature such as the presence of a marker for a particular disease, method 400 may be repeated until the presence of the feature is detected or until all of the regions of the silhouette have been imaged.
- FIGS. 5-12 show exemplary systems and techniques that may be used in some implementations.
- FIGS. 5-12 describe image acquisition and processing in terms of four phases (a FIND phase, a FOCUS phase, a SCAN phase, and a COLLECT phase). Further information may be found in a number of references, such as U.S. Pat. Nos. 6,546,123; 6,518,554; 6,445,817; 6,418,236; 6,404,916; 6,330,349; 6,215,892; and 6,151,405, which are incorporated by reference.
- FIGS. 5A and 5B show a system 500 for image acquisition and processing.
- System 500 includes a microscope subsystem 510 , which is coupled with a controller 512 and a display device 514 .
- System 500 may include, for example, a Chromavision ACISTM system or other appropriate system.
- Controller 512 may be included in a computer such as a desktop, laptop, or host computer, or may be implemented differently (e.g., as a unit integrated with microscope subsystem 510 ). Controller 512 may control system 500 using, for example, image processing algorithms and image analysis algorithms.
- Microscope subsystem 510 includes a barcode reader 516 , a camera 518 , a serial interface 520 , one or more sensors 522 , one or more motors 524 , a light source 526 , a turret 528 , and a data interface 530 .
- Barcode reader 516 may be a standard barcode reader to detect a barcode identifier for the sample.
- a sample barcode may be positioned on a sample slide in an area separate from a scan region.
- Camera 518 may be a digital camera with selectable resolution capabilities.
- camera 518 may be a charge-coupled device (CCD) camera.
- CCD charge-coupled device
- Camera 518 may be mounted on a turret 528 of microscope subsystem 510 , such that its aperture is aligned with the field of view (FOV) of a lens associated with turret 528 .
- An interface such as serial interface 520 may facilitate communication among barcode reader 516 , camera 518 , and controller 512 .
- interface 520 is a serial interface providing a USB coupling between controller 512 and camera 518 .
- Camera 518 may provide a direct video output connect to a video card within control 512 that gathers the image data from camera 518 for processing.
- Sensors 522 may include, for example, position sensors, temperature sensors, light intensity sensors, optical encoders, or other sensors.
- Motors 524 may be servomotors, and may provide motion control of microscope subsystem 510 .
- motors 524 may be used to rotate an appropriate lens within the optical path of the microscope, may be used to adjust focus (e.g., by changing a distance between a sample surface and optical elements), may be used to move an automated X-Y stage (not shown), or may provide different or additional functionality in system 500 .
- Light source 526 may be any suitable light source for appropriately illuminating the FOV of an objective of the microscope such that the creation of a digital image of that FOV is possible.
- Turret 528 may be a motor-driven microscope turret upon which is mounted a set of lenses of varying power that may be rotated into the optical path of the microscope. Turret 528 may also be suitably controlled to provide a desired focus.
- Sensors 522 , motors 524 , light source 526 , and turret 528 may feed electrical inputs of data interface 530 .
- Data interface 530 may be a system driver card to facilitate a data communication link between sensors 522 , motors 524 , light source 526 , turret 528 , and a motion control card within controller 512 .
- System 500 may operate as follows.
- a microscope slide (referred to as a “target slide”) may be fed into an automated X-Y stage of microscope subsystem 510 via an in-feed stage 592 , and subsequently positioned in the FOV of the microscope.
- the identifier of the target microscope slide may be read by barcode reader 516 .
- the target slide may then be scanned at various resolutions and magnifications based upon image-processing algorithms and image analysis algorithms executed by controller 512 .
- image-processing algorithms and image analysis algorithms executed by controller 512 .
- the techniques described below may be used to acquire image data at various magnifications of the microscope.
- the slide may be transferred out of system 500 via an out-feed stage 594 .
- the slide identifier and image data may be transmitted to controller 512 to be stored in one or more memories.
- the motion control system may then move another slide into the FOV of the microscope.
- the process may automatically repeat for each slide that is provided to system 500 .
- the imaging process is initiated when one or more slides are provided to the in-feed stage and continues without user intervention until no more slides are available at the in-feed stage, and as long as no system errors occur.
- a user e.g., a clinician
- FIGS. 6A and 6B show a microscope subsystem 600 that may be used in some implementations.
- a similar microscope subsystem and method of operation is more fully described in pending U.S. Patent Application Serial No. 60/450,824, filed Feb. 27, 2003, which is hereby incorporated by reference.
- FIG. 6A shows a plan view of a microscope subsystem 600 . Shown is slide carrier unloading assembly 634 and unloading platform 636 which in conjunction with slide carrier output hopper 618 function to receive slide carriers which have been analyzed.
- Vibration isolation mounts 640 shown in further detail in FIG. 6B, are provided to isolate the microscope subsystem 600 from mechanical shock and vibration that can occur in a typical laboratory environment. In addition to external sources of vibration, the high-speed operation of the X-Y stage 638 can induce vibration into the microscope subsystem 600 . Such sources of vibration can be isolated from the electro-optical subsystems to avoid any undesirable effects on image quality.
- the isolation mounts 640 comprise a spring 640 a and piston 640 b (see FIG.
- Oculars 620 are shown in FIGS. 6A and 6B, however, their presence is an optional feature. The oculars 620 may be absent without departing from the advantages or functionality of the system.
- FIG. 7 shows a slide 700 (e.g., a conventional glass microscope slide), which may be used some implementations.
- a barcode region 710 including an identifier such as a barcode sticker.
- Slide 700 further includes a scan region 712 surrounded by a non-scan region 714 , where scan region 712 includes a sample 716 to be imaged.
- Sample 716 may be any target specimen, such as a tissue sample resulting from a needle biopsy.
- Non-scan region 714 may include an area along the edge of slide 700 .
- FIG. 8 shows an expanded view of a scan region such as region 712 of slide 700 .
- scan region 712 is partitioned into an array of contiguous segments, referred to as zelles 810 .
- zelles 810 This is not generally a physical partition, but rather a logical segmentation of scan region 712 to be used for imaging scan region 712 .
- the area of the zelles 810 may be determined by the power (i.e., magnification) setting of the microscope. For example, as the magnification is increased, the FOV of the microscope decreases, so that each zelle 810 is smaller for greater magnifications.
- zelles 810 may be defined so that there is some overlap. If there is no overlap in the zelles 810 , inaccuracies in the X-Y stage of the microscope may cause some region of the sample to be missed in the imaging. Overlap of images may not be necessary where the digital spacing between pixels is finer than the optical resolution of an objective lens of the microscope and/or wherein the movement of the X-Y stage has an error of movement less than the optical resolution of the objective lens (see, e.g., U.S. Pat. No. 6,272,235, which is incorporated herein by reference). Although this may be acceptable for some implementations, it may not in others. For example, when the desired sample feature is small, the missed region may lead to undesirable inaccuracy in imaging results.
- zelles 810 may be determined by the FOV of the microscope at generally low magnification and resolution. Therefore, the FIND phase results in a relatively small scan time and amount of stored image data. Further, since low-power lenses have a greater depth of focus than higher-power lenses, re-focusing may be unnecessary. For example, a low-power lens of the microscope of FIG. 5 may be focused at either a best-guess plane or a plane derived from microscope calibration during the FIND phase of operation. Optionally, a separate low-resolution, wide FOV camera may be used during the FIND phase of operation.
- image processing algorithms and image analysis algorithms may be executed by controller 512 to determine any regions of slide 700 having content worthwhile for reexamination at a higher magnification during a subsequent COLLECT phase of operation.
- a zelle 810 found to have specimen content is classified as “interesting” and may be mapped as a logical “1.”
- any zelle 810 found to have no specimen content is classified as “non-interesting” and may be mapped as a logical “0.”
- a silhouette of the specimen i.e. sample 716
- An image table may be generated that represents the low-resolution image of sample 716 that results from the FIND phase of operation.
- Parameters applicable to the particular sample being imaged are provided. The parameters are then used to determine whether a particular zelle 810 is interesting or non-interesting. For example, in a particular sample, a zelle may be interesting if it includes blue stain, red stain, any speck of tissue, a plurality of tissues in a micro-tissue array, or a large volume of tissue. Therefore, one or more parameters that indicate the presence of blue stain, red stain, any speck of tissue, a plurality of tissues in a micro-tissue array, or a large volume of tissue may be provided to analyze the image.
- a statistical algorithm may be used to determine whether a particular zelle 810 is interesting or non-interesting. That is, an image of each zelle 810 is obtained and analyzed using a statistical algorithm to determine if the zelle 810 includes one or more of the applicable parameters. Using the statistical algorithm, zelles 810 that include even a small segment of interest may be identified. At the same time, substantial analysis of empty zelles 810 is avoided.
- FIG. 9 shows another expanded view of scan region 712 of slide 700 .
- one or more zelles 810 may be partitioned into a plurality of minor-zelles 910 .
- Minor zelles 910 may allow sample 716 or one or more features of sample 716 to be more precisely located in scan region 712 .
- some zelles 810 include a large amount of sample 716 , while others include only a small amount.
- minor-zelles 910 areas of zelles 810 not including sample 716 may be identified and disregarded in subsequent actions.
- the minor-zelles 910 may be imaged, and image processing and image analysis algorithms may be used to designate particular minor-zelles 910 as “interesting” or “non-interesting.” Interesting minor-zelles 910 may be mapped as a logical “1,” while non-interesting minor-zelles 910 may be mapped as a logical “0.” One or more image tables may be generated, including information as to whether minor-zelles 910 are uninteresting or interesting. Note that the same and/or different parameters may be used to determine whether a particular minor-zelle is interesting or non-interesting. The image data associated with “non-interesting” regions may be discarded, thereby decreasing the storage space requirements. Image-processing and/or image analysis algorithms may be used. For example, algorithms may be used to determine whether microscope slide 700 contains a valid sample 716 or whether sample 716 meets distribution expectations.
- FIG. 10 shows a FIND silhouette 1000 , which is mapping of minor-zelles 910 designated as either a logical 1 or 0 for those zelles 810 designated as a logical 1.
- the mapping forms a higher resolution silhouette of sample 716 .
- FIG. 11 shows a SCAN silhouette 1100 of scan-zelles 1110 .
- SCAN silhouette 1100 is a “content-is-interesting” silhouette that is captured and analyzed in the SCAN phase of operation.
- FIG. 11 illustrates that the level of “interesting” regions is more finely defined by SCAN silhouette 1100 than by FIND silhouette 1000 .
- SCAN silhouette 1100 of sample 716 can still be thought of as a very coarse demagnified image, in which each scan-zelle 1110 can be thought of as a pixel with a 0 or 1 value.
- An image table may be generated that represents the image of sample 716 that results from the SCAN phase of operation. Subsequently, standard image-processing operations such as erosion, dilation, and hole filling or more complex algorithms such as image segmentation algorithms may be performed.
- criteria other than “interesting” or “non-interesting” may be used to analyze image data.
- knowing the status of each scan-zelle 1110 provides microscope imaging system 500 with knowledge of where to focus, e.g., microscope imaging system 500 can focus on regions known to have tissue. This is required when using a high-power lens with a lesser depth of focus in the SCAN or COLLECT phase of operation as compared with a low-power lens used in the FIND phase of operation.
- scan-zelles 1110 Having classified scan-zelles 1110 as “interesting” or “non-interesting” provides microscope imaging system 500 with the knowledge of what focus algorithm to use.
- An image analysis algorithm can be run to further classify scan-zelles 1110 not only as “interesting,” but also as may designate a scan-zelle 1110 as a good candidate to focus on. Again, an image table may be generated as to whether a specific scan-zelle 1110 is a good candidate to focus upon.
- sequences of operations may occur to modify the “interesting” regions as determined by the requirements of a particular test.
- the result is a table of “interesting” and “non-interesting” scan-zelles 1110 as modified via the specific requirements.
- the sequence of image shots for the SCAN or COLLECT phase of operation is planned and these “interesting” regions are selected for focusing upon; thus, planning can occur for the focus algorithms that are run in the FOCUS phase of operation.
- a first scan-zelle 1110 classified as “interesting for focus point” may be selected as a first focus point.
- a database including information about scan-zelles 1110 may then be searched for a second “interesting for focus point” scan-zelle 1110 of the same classification that is most distant from the first “interesting for focus point” scan-zelle 1110 .
- This second scan-zelle 1110 may then be selected as a second focus point.
- the database may then be searched for a third “interesting for focus point” scan-zelle 1110 of same classification that includes the greatest area between the third and the previous two.
- This third scan-zelle 1110 may be selected as a third focus point.
- the three focus points thus define a focus plane.
- a classification of “interesting for focus points” is used rather than the classification of “interesting.” This may be beneficial in some circumstances, since some scan-zelles 1110 classified as only “interesting” may have gotten that classification from the dilate operation and actually may contain no tissue, and thus may be less desirable candidates for focus points than scan-zelles 1110 containing tissue.
- a table of “interesting for focus points” scan-zelles 1110 algorithms for selecting focus points may be executed.
- the number of focus points is either one of the parameters of the application or can be derived from the tissue sample size and/or the distribution of sample pieces.
- the desired number of focus points is found accordingly during the FOCUS phase of operation, along with some extra focus points in case some are later found to be unusable.
- a FOCUS phase of operation occurs using focus algorithms (a parameter might be set as to which focus algorithm to use).
- the goal is to achieve at least a minimum number of focus points that meet the predetermined criteria, from which a focus plane is constructed. It may be inefficient to stop and focus each image; instead it may be more efficient to compute the Z-axis, drive the microscope lens to that Z-height, capture the image, and unload that image while simultaneously translating the equipment to the next image.
- system 500 can now lay out a series of locations where high-resolution images will be taken during the COLLECT phase of operation.
- the size and shape of scan-zelles 1110 may be set to match the FOV of a high-resolution lens. For example, if the FOV is square and the initial image was a 1 ⁇ lens, a 4 ⁇ lens in which the 4 ⁇ FOVs can be designed to fit in the 1 ⁇ FOV may be used. The resulting scan-zelles 1110 are therefore one-fourth the size of the 1 ⁇ FOV.
- the size and shape of scan-zelles 1110 may not match the FOV of the high-resolution lens.
- One way of handling this situation is that the pattern of high-power FOVs that are to be collected (i.e., “SCAN” FOVs) is predetermined, i.e., the pattern is a tiling of the slide, possibly with overlap (where the percentage of overlap may be a parameter of the program based upon the mechanical stage positional inaccuracy).
- the pattern of “interesting” scan-zelles 1110 may then be projected upon this predetermined pattern of high-power FOVs, using algorithms knowing the dimensions and boundaries of both.
- the high-power FOVs needed to capture all the “interesting” scan-zelles 1110 are thus established and tabulated.
- any high-power FOV intersecting an “interesting” scan-zelles 1110 is flagged as needed.
- Non-intersecting high-power FOVs are flagged as not needed.
- a list is generated of the high-power FOVs of which images are to be taken to capture all “interesting” scan-zelles 1110 . It is possible that only a small portion of these high-power FOVs actually include “interesting” scan-zelles 1110 , but such coverage may be necessary so that no item of interest is missed.
- This method of predetermining the layout of any magnification FOV is used with attributes or combinations of attributes.
- Example attributes include content for focus but not content for scan, content for scan but not content for focus, content for wanting to revisit for higher power, and the like.
- Another way of handling the situation is as follows. Having determined the pattern of “interesting” scan-zelles 1110 or even knowing one of more broad regions that are “interesting,” everything in that known region may be captured at a higher power. An algorithm can be run to determine the minimum number of high-power FOVs necessary to cover a known region of “interesting” scan-zelles 1110 . Additionally, a first high-power FOV may be aligned to a corner of a first “interesting” scan-zelle 1110 ; the pattern of high-power FOVs is then continued until the entire region of “interesting” scan-zelles 1110 is covered. The period of high-power FOVs will differ from the period of “interesting” scan-zelles 1110 . In general, high-power FOVs can be further subdivided into scan-zelles 1110 . The high-power FOVs may be stitched together to reconstruct the picture. This process can be repeated any number of times for any number of levels of magnification.
- Image data compression associated with the SCAN or COLLECT phase of operation may be achieved using block compression algorithms that are applicable to pixel-level compression on the high-resolution images.
- Many data compression techniques may be used.
- An exemplary technique includes the following.
- Each high-power FOV is compressed in two ways: a first pass using a modification of the well-known Classic Huffman Shift compression algorithm (see, for example, I.R.E. September 1952, page 1098, “A Method for the Construction of Minimum-Redundancy Codes” by David A. Huffman, which is herein incorporated by reference) and a second pass using a White Clip compression algorithm.
- the compression algorithms are more efficient in this application because some information has been obtained about the data, i.e., information has been obtained about what is and is not important, and thus parameters may be set accordingly. This knowledge thus allows the compression to be modified for efficiency.
- the Classic Huffman compression algorithm may take advantage of the understanding of the distribution of the colors in the image. Filters may be defined as to what colors are of higher value and which are of less or no value. For example, the non-valuable colors may be set to the same value (and thus require minimal storage space) without compromising picture quality when restored. Whatever color variations exist in the white space, for example, are non-informational; thus, restoring this space as uniformly white is acceptable.
- microscope imaging system 500 may perform a silhouette scan operation for improved image data storage requirements and data acquisition time.
- the imaging system and method described above may allow for an image capture of scan region 712 and a first level of data compression at a low-resolution stage.
- Scan zelles 1110 may be sorted into “non-interesting” vs. “interesting,” and data associated with the “non-interesting” scan-zelles 1110 may be discarded.
- a second level of data compression at the high-resolution stage may be performed, wherein the “non-interesting” space may be set to the same value, while everything of importance is stored losslessly.
- Systems and techniques of the current disclosure may provide a substantial benefit in that they may be performed autonomously. That is, once the process is initiated (e.g., when one or more slides are provided to microscope subsystem 510 ), the process may execute without human intervention or decision-making.
- FIG. 500 An illustration follows of the general operation of microscope imaging system 500 analyzing, for example, sample 716 of a lymph node.
- the lymph node tissue may be stained with a dye that makes a suspected cancer cell turn brown, while the rest of the tissue may be stained with a common blue dye that is used to mark tissue.
- One or more parameters of the test are set accordingly. For example, a parameter may be set to indicate that an area that is blue or brown is “interesting.” Further, another parameter may be set to indicate that an area that is brown is “very interesting.”
- a FIND phase of operation occurs upon microscope slide 700 where a low-resolution image (i.e., FIND silhouette 1000 ) of the entire lymph node is captured. Based upon the predetermined parameters, image-processing algorithms and image analysis algorithms are executed by controller 512 . Anything that is found to be blue or brown is classified as “interesting,” but brown is further classified as “very interesting.”
- a number of focus points and their locations may be determined.
- a focus plane may be derived from the focus points.
- a more precise higher-resolution image is generated (i.e., SCAN silhouette 1100 ) for processing.
- the FOCUS and SCAN phases may be repeated any number of times at progressively higher resolutions to identify “interesting” regions.
- a COLLECT phase of operation is performed, which may use focus cues from the previous phase. Another image scan is done on the “very interesting” scan-zelles 1110 with a high-power lens. This example illustrates that varying degrees of “interestingness” may be defined.
- a set of tables are generated that have effectively compressed scan-zelles 1110 to pixels, each having one or more classifications as, for example, non-interesting, interesting, very interesting, interesting for focus points, interesting for higher power, and the like.
- the image data generated during the SCAN or COLLECT phases of operation may be transmitted to controller 512 .
- Data compression algorithms may be executed on the image data, thereby completing a second level of image data compression. Having completed these two levels of data compression, the image data is stored in memory of controller 512 (or other memory). The image data may subsequently be accessed for future image reconstruction and inspection by a clinician.
- FIG. 12 illustrates a method 1200 of performing a silhouette scan operation for improved image data acquisition time and image data storage.
- the parameters may be used to analyze appropriate zelles 810 and minor-zelles 910 and to determine whether they include anything designated as interesting ( 1210 ).
- parameters may include specimen color, size, features, or distribution and sensitivity for the specific test being run.
- a target slide may be provided to system 500 , and an ID of a target microscope slide 700 may be read by barcode reader 516 so that any image data saved may be linked to the appropriate microscope slide 700 ( 1212 ).
- Scan region 712 of target microscope slide 700 is scanned via microscope subsystem 510 in combination with camera 518 ( 1214 ).
- Scan region 712 is partitioned into an array of contiguous zelles 810 covering the entire area of scan region 712 .
- the area of each zelle 810 may be defined by the power (i.e., magnification) setting of the microscope (for example, the area of each zelle 810 may correspond to the area of the FOV of the setting of the microscope).
- the initial scan of scan region 712 may be performed efficiently.
- Zelles 810 may overlap slightly or abut. Overlap may be useful to ensure that no region is missed due to mechanical inaccuracies in the X-Y stage of microscope subsystem 510 .
- microscope subsystem 510 may be able to search for tissue during the initial scan of scan region 712 without refocusing.
- a low-power lens of the microscope may be focused at a best-guess plane or one derived from microscope calibration.
- a separate low-resolution, wide-FOV camera may be used.
- Images obtained during a scan of scan region 712 may be processed using image-processing algorithms and/or image analysis algorithms to determine any regions of microscope slide 700 having content worthwhile to reexamine at a higher magnification during a subsequent FOCUS, SCAN or COLLECT phase of operation.
- any zelle 810 or minor-zelle 910 found to have specimen content is classified as “interesting” and mapped as a logical “1.”
- any zelle 810 or minor-zelle 910 found to have no specimen content is classified as “non-interesting” and mapped as a logical “0.”
- a silhouette of the specimen i.e., sample 716
- An image table is generated that represents the low-resolution image of sample 716 .
- each zelle 810 or minor-zelle 910 may be subdivided into yet smaller regions using data processing.
- Each zelle 810 or minor-zelle 910 may therefore be further partitioned, forming a SCAN silhouette 600 formed of the plurality of scan-zelles 1110 , as shown in FIG. 11.
- the “interestingness” of these scan-zelles 1110 is assessed via image-processing algorithms and image analysis algorithms. This is done in preparation for a future operation of collecting and saving a series of higher magnification FOVs (i.e., in the COLLECT phase), as only the “interesting” regions are generally examined at a higher magnification.
- scan-zelles 1110 may be defined to match the anticipated FOVs of the higher power images of the COLLECT phase of operation.
- the size of scan-zelles 1110 and whether scan-zelles 1110 match up with the anticipated FOVs of the higher power images are adjustable system parameters that may be set ( 1210 ).
- scan-zelles 1110 found to have specimen content are classified as “interesting” and mapped as a logical “1.”
- scan-zelles 1110 found to have no specimen content are classified as “non-interesting” and mapped as a logical “0.”
- further image compression occurs and a yet more precise silhouette of sample 716 is formed, thereby creating a slightly higher-resolution image, than that generated in the FIND phase.
- This higher resolution image may be processed using image-processing algorithms.
- the system may determine whether another iteration is performed ( 1220 ). The decision may be based on one or more parameters, such as the type of test, expected size of feature, etc. If iteration is desired, the above actions may be repeated ( 1216 et seq.) If iteration is not desired, a sequence of image shots for a COLLECT phase may be planned ( 1222 ). Various sequences of operations may occur to modify the “interesting” regions as determined by the requirements of a particular test. The result is a table of “interesting” and “non-interesting” scan-zelles 1110 as modified via the specific requirements.
- the COLLECT phase may then be performed ( 1224 ). Having established both the pattern of “interesting” scan-zelles 1110 and the focus plane, system 500 now lays out a series of locations where high-resolution images are taken during this COLLECT phase of operation. As a result, with specific knowledge of interesting regions and of diagnostically important items, a high-magnification lens in combination with camera 518 are used to capture an image of all “interesting” scan-zelles 1110 . Image data generated during the COLLECT phase of operation may be transmitted to controller 512 .
- Image data acquired in the SCAN and/or COLLECT phase may be compressed ( 1226 ), as described above.
- a determination may be made as to whether the particular slide 700 with sample 716 is to be examined and analyzed ( 1228 ). If it is, a clinician may recall the image data for the selected microscope slide 700 and controller 512 (or other data processor) may subsequently reconstruct the image by stitching together individual scan-zelles 1110 , and/or high-power FOVs. The image may then be displayed to the clinician for inspection via display device 514 .
- a processor-based system can include a main memory, preferably random access memory (RAM), and can also include a secondary memory.
- the secondary memory can include, for example, a hard disk drive and/or a removable storage drive, representing a floppy disk drive, a magnetic tape drive, an optical disk drive, etc.
- the removable storage drive reads from and/or writes to a removable storage medium.
- Removable storage medium refers to a floppy disk, magnetic tape, optical disk, and the like, which is read by and written to by a removable storage drive.
- the removable storage medium can comprise computer software and/or data.
- the secondary memory may include other similar means for allowing computer programs or other instructions to be loaded into a computer system.
- Such means can include, for example, a removable storage unit and an interface. Examples of such can include a program cartridge and cartridge interface (such as the found in video game devices), a movable memory chip (such as an EPROM or PROM) and associated socket, and other removable storage units and interfaces, which allow software and data to be transferred from the removable storage unit to the computer system.
- the computer system can also include a communications interface.
- Communications interfaces allow software and data to be transferred between computer system and external devices.
- Examples of communications interfaces can include a modem, a network interface (such as, for example, an Ethernet card), a communications port, a PCMCIA slot and card, and the like.
- Software and data transferred via a communications interface are in the form of signals, which can be electronic, electromagnetic, optical or other signals capable of being received by a communications interface. These signals are provided to communications interface via a channel capable of carrying signals and can be implemented using a wireless medium, wire or cable, fiber optics or other communications medium.
- Some examples of a channel can include a phone line, a cellular phone link, an RF link, a network interface, and other communications channels.
- computer program medium and “computer usable medium” are used to refer generally to media such as a removable storage device, a disk capable of installation in a disk drive, and signals on a channel. These computer program products are means for providing software or program instructions to a computer system.
- Computer programs are stored in main memory and/or secondary memory. Computer programs can also be received via a communications interface. Such computer programs, when executed, enable the computer system to perform the features of the invention as discussed herein. In particular, the computer programs, when executed, enable the processor to perform the features of the invention. Accordingly, such computer programs represent controllers of the computer system.
- the software may be stored in, or transmitted via, a computer program product and loaded into a computer system using a removable storage drive, hard drive or communications interface.
- the control logic when executed by the processor, causes the processor to perform the functions of the invention as described herein.
- the elements are implemented primarily in hardware using, for example, hardware components such as PALs, application specific integrated circuits (ASICs) or other hardware components. Implementation of a hardware state machine so as to perform the functions described herein will be apparent to person skilled in the relevant art(s). In yet another embodiment, elements are implanted using a combination of both hardware and software.
- the computer-based methods can be accessed or implemented over the World Wide Web by providing access via a Web Page to the methods of the invention.
- the Web Page is identified by a Universal Resource Locator (URL).
- the URL denotes both the server machine and the particular file or page on that machine.
- a consumer or client computer system interacts with a browser to select a particular URL, which in turn causes the browser to send a request for that URL or page to the server identified in the URL.
- the server responds to the request by retrieving the requested page and transmitting the data for that page back to the requesting client computer system (the client/server interaction is typically performed in accordance with the hypertext transport protocol (“HTTP”)).
- HTTP hypertext transport protocol
- the selected page is then displayed to the user on the client's display screen.
- the client may then cause the server containing a computer program of the invention to launch an application to, for example, perform an analysis according to the invention.
Abstract
Description
- The present application describes systems and techniques relating to image data acquisition.
- Obtaining and analyzing images of materials plays an important role in many areas of technology. For example, the ability to obtain and analyze accurate images efficiently is important in the field of anatomic pathology. Anatomic pathology enables the diagnosis of physiological conditions such as cancer, infectious disease and prenatal disorders through analysis of tissue samples.
- Typically, a biological specimen from a patient is obtained, imaged, and analyzed. The specimen may be “stained” using a material chosen to reveal the presence of one or more molecules. The specimen may then be placed on a slide for viewing with a microscope.
- An image of the sample may be obtained using a digital camera. In some systems, the sample may be scanned, and a series of images may be obtained. The image data may then be analyzed. For example, an image may be analyzed to identify the presence of particular characteristic features of a particular disease, where the particular characteristic features indicate the presence of or predisposition towards the disease.
- Systems and techniques described herein may enable efficient and accurate image acquisition and processing.
- In general, in one aspect, a method includes obtaining an image of at least a portion of a scan region, and analyzing the image to determine if the image includes a first desired feature. If the image includes the first desired feature, the method may include electronically dividing the image into a plurality of subimages at the same magnification and determining the presence or absence of a second desired feature in each subimage. The first desired feature and the second desired feature may be the same or different. For example, the first desired feature may be the presence of at least a portion of a sample in the scan region. The method may include analyzing images and/or subimages to determine the presence or absence of other features besides the first and second desired features.
- The image may be acquired at a first magnification, which may comprise a field of view smaller than the scan region. The plurality of subimages may each comprise an area that is substantially equal to the field of view of a higher magnification. The method may further include acquiring an image of an area of the scan region including the second desired feature at higher magnification. For example, one or more higher magnification images may be obtained, so that each subimage including the second desired feature is imaged at the higher magnification.
- The scan region may be a microscope slide or a portion of a microscope slide. The microscope slide may have a sample thereon, or may have more than one sample (e.g., a tissue array). The sample may be a biological sample.
- The method may include storing an identifier for the image, as well as data associated with the identifier, where the data indicates the presence or absence of the first desired feature. The method may include storing an identifier for the subimages, as well as data associated with the identifier, where the data indicates the presence or absence of the second desired feature.
- In general, in one aspect, a system includes an image acquisition system, which may include a camera and/or microscope. A data processor may be coupled with the image acquisition system. The data processor may receive image data from the image acquisition system, and may determine the presence or absence of a desired feature in the image. If the data processor determines the presence of the desired feature, the data processor may electronically divide the image data into a plurality of subimage portions. The data processor may determine the presence or absence of the desired feature, another feature, or both in the plurality of subimage portions.
- In general, in another aspect, a method may include obtaining a first focal length at a first position comprising a sample, obtaining a second focal length at a second position comprising a sample furthest from the first position, and obtaining a third focal length at a third position comprising a sample at a position furthest from the first and second positions. The method may include generating a focal plane comprising the focal length of the first, second, and third positions.
- Details of one or more implementations are set forth in the accompanying drawings and the description below. Other features and advantages may be apparent from the description and drawings, and from the claims.
- These and other aspects will now be described in detail with reference to the following drawings.
- FIG. 1 shows a block diagram of a system that may be used to acquire image data.
- FIG. 2 is a flowchart showing a process for acquiring and analyzing image data for a plurality of areas.
- FIG. 3 is a flowchart showing a process of acquiring and analyzing image data for a plurality of subdivisions.
- FIG. 4 is a flowchart showing a process of acquiring and analyzing image data for a region designated as interesting.
- FIGS. 5A and 5B show a system that may be used to acquire image data.
- FIGS. 6A and 6B show a microscope subsystem for acquiring image data.
- FIG. 7 shows a slide including a scan region.
- FIG. 8 shows an expanded view of a scan region.
- FIG. 9 shows another expanded view of a scan region.
- FIG. 10 shows a FIND silhouette.
- FIG. 11 shows a SCAN silhouette.
- FIG. 12 illustrates a method of performing a silhouette scan operation.
- Like reference symbols in the various drawings indicate like elements.
- The systems and techniques described herein relate to image acquisition. For example, the systems and techniques provide for efficient and accurate imaging and analysis of materials such as biological samples.
- The challenges of obtaining and processing images are interconnected. In order to accurately determine the characteristics of a sample, at least a portion of the sample may need to be viewed or analyzed at high resolution. However, high resolution images are generally very large data files. If the entire sample were to be imaged and image data stored at high resolution, the data storage and analysis challenge may be greater than the ability of an available data processing system. For example, a digital representation of an entire microscope slide obtained using a high power lens may include 100 to 400 separate images. Obtaining, storing, and analyzing such a large number of digital images may be prohibitive in many applications.
- A number of approaches may be used to meet these challenges. In some systems, a user views a low resolution image of, for example, a microscope slide. The user then defines an area of interest on the slide (e.g., the outline of a tissue sample) to be imaged at higher resolution. Thus, a high resolution image of the entire slide is not necessary. However, these systems require substantial user involvement and may also require that the user be trained to recognize areas of interest.
- Other approaches include using data compression methods to reduce the size of the stored image files and to ease the task of analysis. However, the data compression methods used may decrease the accuracy of the data and, although they may reduce the size of the data files appreciably, may not do so adequately while retaining the desired image quality.
- FIG. 1 shows a
system 100 that may be used to acquire image data, according to various implementations. Amicroscope 110, such as an optical microscope, receives a sample holder with a scan region, where the scan regions includes a sample to be imaged. For example, aslide input 120 may provide one or more microscope slides including (for example) a tissue sample tomicroscope 110.Microscope 110 may also be coupled with aslide output 130 to receive slides that have been processed. - An
image acquisition system 140, which includes a digital camera or other image data acquisition element, is coupled withmicroscope 110.Microscope 110 andimage acquisition system 140 may be coupled with one or more data processing systems such as adata processor 150. Adisplay 160 may be coupled withdata processor 150 and/orimage acquisition system 140, for providing a visual display of an image and/or image data. Note that the elements of FIG. 1 are shown separately; however, some or all of them may be integrated. For example, slideinput 120 andslide output 130 may be integrated withmicroscope 110 rather than separate. - FIG. 2 shows a
method 200 for acquiring image data using a system such assystem 100 of FIG. 1. A sample holder with a scan region is provided to a microscope (210). For example, the sample holder may be a microscope slide, while the scan region may be an area of the microscope slide in which a sample (e.g., a tissue specimen) may be found. In another example, the sample holder may be a microscope slide, while the scan region may include an array of specimens (such as an array of tissue samples). - Image data for a plurality of areas in the scan region may be acquired (220). Image data for a particular area may be processed to determine if the particular area includes a desired feature (230, 240). If the particular area includes a desired feature, the area is designated as “interesting” (250). If the particular area does not include a desired feature, the area is designated as “non-interesting” (260). The coordinates (e.g., x-y coordinates) of “interesting” areas may be stored for later review and/or processing.
Method 200 may include processing the image data for each of the plurality of imaged areas, and designating each of the areas as either interesting or non-interesting. - The plurality of areas may form an array covering the scan region. A first silhouette including the interesting areas of the array may be formed. The first silhouette forms a low-resolution image of the sample in the scan region. For example, when the sample holder is a microscope slide with a specimen on the scan region of the slide, the scan region may be divided into an array of areas, and each area that includes at least a portion of the specimen will be deemed “interesting” and thus included in the first silhouette.
- Data associated with the first silhouette may be stored in a memory such as a memory included in
data processor 150. For example, the memory may store a table including an identifier for each location, where each interesting location is designated as a logical 1, and each non-interesting location is designated as a logical 0. - The system may scan all the areas; for example, by scanning across each of the rows or columns of the array in turn. In some implementations, the system may scan fewer than all areas. For example, if the specimen is a contiguous tissue sample on a slide, the system may find a portion of the boundary of the tissue sample, and may use this information to scan only some of the areas.
- In some implementations, the areas in the first silhouette may be imaged at higher resolution to obtain desired information about the sample in the scan region. That is, the interesting regions (those that include the desired feature) may be imaged using a higher magnification setting of a microscope, and the subsequent image data may be compressed and stored. However, this may be inefficient (since the interesting areas may include only a small portion of the sample) and may require significant data storage and data analysis. Therefore, in other implementations, the interesting areas of the scan region may be sub-divided and the system may determine which subdivisions include one or more desired features. FIG. 3 shows a
method 300 that may be used to subdivide the areas of the scan region. - An interesting area, identified at
block 250, may be subdivided into a plurality of subdivisions (310). The image data (e.g., a portion of the image data acquired in block 220) for the particular subdivision may then be processed to determine if the particular subdivision includes a desired feature (320, 330). If the particular subdivision includes the desired feature, it may be designated as “interesting” (340). If the particular subdivision does not include the desired feature, it may be designated as “non-interesting” (350). The coordinates of “interesting” subdivisions can be stored for later review and/or processing at a higher magnification. In one aspect, a subdivision (also termed “minor-zelle”) area is approximately equal to the field of view of a higher power objective of a microscope. In some implementations, each of the subdivisions may be analyzed; in others, less than all subdivisions are analyzed. -
Method 300 may be used to form a second silhouette of higher resolution than the first silhouette. That is, the second silhouette more clearly defines the region in which the sample is located (or in which a different desired feature is present). As with the first silhouette, in some implementations, the interesting subdivisions in the second silhouette may be imaged at higher magnification to obtain desired information about the sample in the scan region. In some implementations, additional subdivisions may be desired, where each subdivision step generates a higher resolution silhouette than the previous step. - FIG. 4 shows a
method 400 that may be used to image a silhouette, where the areas, subdivisions, etc. of the silhouette are referred to generally as regions. That is, if the silhouette is a first silhouette, the regions correspond to the areas discussed above, if the silhouette is a second silhouette, the regions correspond to the sub-divisions discussed above, if the silhouette is a third silhouette, the regions correspond to sub-sub-divisions, and so on. - An image of a first region is acquired (410). The image may be processed using one or more image processing algorithms (420). The image may also be analyzed using one or more image analysis algorithms (430). In some implementations,
method 400 is repeated for all interesting regions in the silhouette. In others, fewer than all interesting regions may be imaged. For example, if a tissue sample is being analyzed to determine the presence of a particular feature such as the presence of a marker for a particular disease,method 400 may be repeated until the presence of the feature is detected or until all of the regions of the silhouette have been imaged. - FIGS. 5-12 below show exemplary systems and techniques that may be used in some implementations. For explanatory purposes, FIGS. 5-12 describe image acquisition and processing in terms of four phases (a FIND phase, a FOCUS phase, a SCAN phase, and a COLLECT phase). Further information may be found in a number of references, such as U.S. Pat. Nos. 6,546,123; 6,518,554; 6,445,817; 6,418,236; 6,404,916; 6,330,349; 6,215,892; and 6,151,405, which are incorporated by reference.
- FIGS. 5A and 5B show a
system 500 for image acquisition and processing.System 500 includes amicroscope subsystem 510, which is coupled with acontroller 512 and adisplay device 514.System 500 may include, for example, a Chromavision ACIS™ system or other appropriate system.Controller 512 may be included in a computer such as a desktop, laptop, or host computer, or may be implemented differently (e.g., as a unit integrated with microscope subsystem 510).Controller 512 may controlsystem 500 using, for example, image processing algorithms and image analysis algorithms. -
Microscope subsystem 510 includes abarcode reader 516, acamera 518, aserial interface 520, one ormore sensors 522, one ormore motors 524, alight source 526, aturret 528, and adata interface 530.Barcode reader 516 may be a standard barcode reader to detect a barcode identifier for the sample. In some implementations, a sample barcode may be positioned on a sample slide in an area separate from a scan region.Camera 518 may be a digital camera with selectable resolution capabilities. For example,camera 518 may be a charge-coupled device (CCD) camera.Camera 518 may be mounted on aturret 528 ofmicroscope subsystem 510, such that its aperture is aligned with the field of view (FOV) of a lens associated withturret 528. An interface such asserial interface 520 may facilitate communication amongbarcode reader 516,camera 518, andcontroller 512. - In an implementation,
interface 520 is a serial interface providing a USB coupling betweencontroller 512 andcamera 518.Camera 518 may provide a direct video output connect to a video card withincontrol 512 that gathers the image data fromcamera 518 for processing. -
Sensors 522 may include, for example, position sensors, temperature sensors, light intensity sensors, optical encoders, or other sensors.Motors 524 may be servomotors, and may provide motion control ofmicroscope subsystem 510. For example,motors 524 may be used to rotate an appropriate lens within the optical path of the microscope, may be used to adjust focus (e.g., by changing a distance between a sample surface and optical elements), may be used to move an automated X-Y stage (not shown), or may provide different or additional functionality insystem 500. -
Light source 526 may be any suitable light source for appropriately illuminating the FOV of an objective of the microscope such that the creation of a digital image of that FOV is possible.Turret 528 may be a motor-driven microscope turret upon which is mounted a set of lenses of varying power that may be rotated into the optical path of the microscope.Turret 528 may also be suitably controlled to provide a desired focus. -
Sensors 522,motors 524,light source 526, andturret 528 may feed electrical inputs ofdata interface 530.Data interface 530 may be a system driver card to facilitate a data communication link betweensensors 522,motors 524,light source 526,turret 528, and a motion control card withincontroller 512. -
System 500 may operate as follows. A microscope slide (referred to as a “target slide”) may be fed into an automated X-Y stage ofmicroscope subsystem 510 via an in-feed stage 592, and subsequently positioned in the FOV of the microscope. During the transition from the in-feed stage ofsystem 500 to the X-Y stage ofmicroscope subsystem 510, the identifier of the target microscope slide may be read bybarcode reader 516. - The target slide may then be scanned at various resolutions and magnifications based upon image-processing algorithms and image analysis algorithms executed by
controller 512. For example, the techniques described below may be used to acquire image data at various magnifications of the microscope. - After the image scan operation, the slide may be transferred out of
system 500 via an out-feed stage 594. The slide identifier and image data may be transmitted tocontroller 512 to be stored in one or more memories. The motion control system may then move another slide into the FOV of the microscope. - The process may automatically repeat for each slide that is provided to
system 500. In an implementation, the imaging process is initiated when one or more slides are provided to the in-feed stage and continues without user intervention until no more slides are available at the in-feed stage, and as long as no system errors occur. However, a user (e.g., a clinician) may view and/or manipulate the digital image of one or more slides viacontroller 512 anddisplay device 514 for the inspection and analysis of a sample. - FIGS. 6A and 6B show a
microscope subsystem 600 that may be used in some implementations. A similar microscope subsystem and method of operation is more fully described in pending U.S. Patent Application Serial No. 60/450,824, filed Feb. 27, 2003, which is hereby incorporated by reference. - FIG. 6A shows a plan view of a
microscope subsystem 600. Shown is slidecarrier unloading assembly 634 andunloading platform 636 which in conjunction with slidecarrier output hopper 618 function to receive slide carriers which have been analyzed. Vibration isolation mounts 640, shown in further detail in FIG. 6B, are provided to isolate themicroscope subsystem 600 from mechanical shock and vibration that can occur in a typical laboratory environment. In addition to external sources of vibration, the high-speed operation of theX-Y stage 638 can induce vibration into themicroscope subsystem 600. Such sources of vibration can be isolated from the electro-optical subsystems to avoid any undesirable effects on image quality. The isolation mounts 640 comprise aspring 640 a and piston 640 b (see FIG. 6B) submerged in a high viscosity silicon gel which is enclosed in an elastomer membrane bonded to a casing to achieve damping factors on the order of about 17 to 20%. Other dampening devices are known in the art and may be substituted or combined with the dampening device provided herein.Oculars 620 are shown in FIGS. 6A and 6B, however, their presence is an optional feature. Theoculars 620 may be absent without departing from the advantages or functionality of the system. - FIG. 7 shows a slide700 (e.g., a conventional glass microscope slide), which may be used some implementations. Oriented at one end of
slide 700 is abarcode region 710 including an identifier such as a barcode sticker.Slide 700 further includes ascan region 712 surrounded by anon-scan region 714, wherescan region 712 includes asample 716 to be imaged.Sample 716 may be any target specimen, such as a tissue sample resulting from a needle biopsy.Non-scan region 714 may include an area along the edge ofslide 700. - FIG. 8 shows an expanded view of a scan region such as
region 712 ofslide 700. During a FIND phase, scanregion 712 is partitioned into an array of contiguous segments, referred to aszelles 810. Note that this is not generally a physical partition, but rather a logical segmentation ofscan region 712 to be used forimaging scan region 712. The area of thezelles 810 may be determined by the power (i.e., magnification) setting of the microscope. For example, as the magnification is increased, the FOV of the microscope decreases, so that each zelle 810 is smaller for greater magnifications. - In some implementations,
zelles 810 may be defined so that there is some overlap. If there is no overlap in thezelles 810, inaccuracies in the X-Y stage of the microscope may cause some region of the sample to be missed in the imaging. Overlap of images may not be necessary where the digital spacing between pixels is finer than the optical resolution of an objective lens of the microscope and/or wherein the movement of the X-Y stage has an error of movement less than the optical resolution of the objective lens (see, e.g., U.S. Pat. No. 6,272,235, which is incorporated herein by reference). Although this may be acceptable for some implementations, it may not in others. For example, when the desired sample feature is small, the missed region may lead to undesirable inaccuracy in imaging results. - During the FIND phase,
zelles 810 may be determined by the FOV of the microscope at generally low magnification and resolution. Therefore, the FIND phase results in a relatively small scan time and amount of stored image data. Further, since low-power lenses have a greater depth of focus than higher-power lenses, re-focusing may be unnecessary. For example, a low-power lens of the microscope of FIG. 5 may be focused at either a best-guess plane or a plane derived from microscope calibration during the FIND phase of operation. Optionally, a separate low-resolution, wide FOV camera may be used during the FIND phase of operation. - During the FIND phase, image processing algorithms and image analysis algorithms may be executed by
controller 512 to determine any regions ofslide 700 having content worthwhile for reexamination at a higher magnification during a subsequent COLLECT phase of operation. - More specifically, a
zelle 810 found to have specimen content is classified as “interesting” and may be mapped as a logical “1.” By contrast, anyzelle 810 found to have no specimen content is classified as “non-interesting” and may be mapped as a logical “0.” In this manner, a silhouette of the specimen (i.e. sample 716) is formed. The “interesting” zelles thus form a low-resolution image that may be processed using image processing algorithms. An image table may be generated that represents the low-resolution image ofsample 716 that results from the FIND phase of operation. - Parameters applicable to the particular sample being imaged are provided. The parameters are then used to determine whether a
particular zelle 810 is interesting or non-interesting. For example, in a particular sample, a zelle may be interesting if it includes blue stain, red stain, any speck of tissue, a plurality of tissues in a micro-tissue array, or a large volume of tissue. Therefore, one or more parameters that indicate the presence of blue stain, red stain, any speck of tissue, a plurality of tissues in a micro-tissue array, or a large volume of tissue may be provided to analyze the image. - A statistical algorithm may be used to determine whether a
particular zelle 810 is interesting or non-interesting. That is, an image of each zelle 810 is obtained and analyzed using a statistical algorithm to determine if thezelle 810 includes one or more of the applicable parameters. Using the statistical algorithm,zelles 810 that include even a small segment of interest may be identified. At the same time, substantial analysis ofempty zelles 810 is avoided. - FIG. 9 shows another expanded view of
scan region 712 ofslide 700. In order to obtain better information about the location ofsample 716 inscan region 712, one or more zelles 810 may be partitioned into a plurality of minor-zelles 910.Minor zelles 910 may allowsample 716 or one or more features ofsample 716 to be more precisely located inscan region 712. Referring again to FIG. 8, somezelles 810 include a large amount ofsample 716, while others include only a small amount. By defining minor-zelles 910, areas ofzelles 810 not includingsample 716 may be identified and disregarded in subsequent actions. - The minor-
zelles 910 may be imaged, and image processing and image analysis algorithms may be used to designate particular minor-zelles 910 as “interesting” or “non-interesting.” Interesting minor-zelles 910 may be mapped as a logical “1,” while non-interesting minor-zelles 910 may be mapped as a logical “0.” One or more image tables may be generated, including information as to whether minor-zelles 910 are uninteresting or interesting. Note that the same and/or different parameters may be used to determine whether a particular minor-zelle is interesting or non-interesting. The image data associated with “non-interesting” regions may be discarded, thereby decreasing the storage space requirements. Image-processing and/or image analysis algorithms may be used. For example, algorithms may be used to determine whethermicroscope slide 700 contains avalid sample 716 or whethersample 716 meets distribution expectations. - FIG. 10 shows a
FIND silhouette 1000, which is mapping of minor-zelles 910 designated as either a logical 1 or 0 for thosezelles 810 designated as a logical 1. The mapping forms a higher resolution silhouette ofsample 716. - FIG. 11 shows a
SCAN silhouette 1100 of scan-zelles 1110.SCAN silhouette 1100 is a “content-is-interesting” silhouette that is captured and analyzed in the SCAN phase of operation. FIG. 11 illustrates that the level of “interesting” regions is more finely defined bySCAN silhouette 1100 than byFIND silhouette 1000. - As a result of the SCAN phase, further image data associated with “non-interesting” regions may be discarded, thus further optimizing image data storage upon
controller 512. In general,SCAN silhouette 1100 ofsample 716 can still be thought of as a very coarse demagnified image, in which each scan-zelle 1110 can be thought of as a pixel with a 0 or 1 value. An image table may be generated that represents the image ofsample 716 that results from the SCAN phase of operation. Subsequently, standard image-processing operations such as erosion, dilation, and hole filling or more complex algorithms such as image segmentation algorithms may be performed. - Additionally, criteria other than “interesting” or “non-interesting” may be used to analyze image data. For example, in the FOCUS phase of operation, knowing the status of each scan-
zelle 1110 providesmicroscope imaging system 500 with knowledge of where to focus, e.g.,microscope imaging system 500 can focus on regions known to have tissue. This is required when using a high-power lens with a lesser depth of focus in the SCAN or COLLECT phase of operation as compared with a low-power lens used in the FIND phase of operation. Ideally, in the FOCUS phase of operation it is desirable to know of specific features to focus on, as focus algorithms tend to run better on specific features, such as an edge. Having classified scan-zelles 1110 as “interesting” or “non-interesting” providesmicroscope imaging system 500 with the knowledge of what focus algorithm to use. An image analysis algorithm can be run to further classify scan-zelles 1110 not only as “interesting,” but also as may designate a scan-zelle 1110 as a good candidate to focus on. Again, an image table may be generated as to whether a specific scan-zelle 1110 is a good candidate to focus upon. - In general, various sequences of operations may occur to modify the “interesting” regions as determined by the requirements of a particular test. The result is a table of “interesting” and “non-interesting” scan-
zelles 1110 as modified via the specific requirements. Subsequently, the sequence of image shots for the SCAN or COLLECT phase of operation is planned and these “interesting” regions are selected for focusing upon; thus, planning can occur for the focus algorithms that are run in the FOCUS phase of operation. - There are several different ways to determine the placement of focus points during the FOCUS phase of operation. For example, a first scan-
zelle 1110 classified as “interesting for focus point” may be selected as a first focus point. A database including information about scan-zelles 1110 may then be searched for a second “interesting for focus point” scan-zelle 1110 of the same classification that is most distant from the first “interesting for focus point” scan-zelle 1110. This second scan-zelle 1110 may then be selected as a second focus point. Similarly, the database may then be searched for a third “interesting for focus point” scan-zelle 1110 of same classification that includes the greatest area between the third and the previous two. This third scan-zelle 1110 may be selected as a third focus point. The three focus points thus define a focus plane. In this implementation, a classification of “interesting for focus points” is used rather than the classification of “interesting.” This may be beneficial in some circumstances, since some scan-zelles 1110 classified as only “interesting” may have gotten that classification from the dilate operation and actually may contain no tissue, and thus may be less desirable candidates for focus points than scan-zelles 1110 containing tissue. - Using a table of “interesting for focus points” scan-
zelles 1110, algorithms for selecting focus points may be executed. The number of focus points is either one of the parameters of the application or can be derived from the tissue sample size and/or the distribution of sample pieces. Thus, the desired number of focus points is found accordingly during the FOCUS phase of operation, along with some extra focus points in case some are later found to be unusable. In this way, a FOCUS phase of operation occurs using focus algorithms (a parameter might be set as to which focus algorithm to use). The goal is to achieve at least a minimum number of focus points that meet the predetermined criteria, from which a focus plane is constructed. It may be inefficient to stop and focus each image; instead it may be more efficient to compute the Z-axis, drive the microscope lens to that Z-height, capture the image, and unload that image while simultaneously translating the equipment to the next image. - These various classifications, such as “interesting for focus points,” may be helpful for determining what magnification to use during the subsequent imaging of the SCAN or COLLECT phase of operation. More specifically, during the FIND phase of operation, a lens power may be used that is suitable for looking at the general geometry and morphology of
sample 716. However, in the COLLECT phase of operation, some information has already been acquired regarding which portions ofscan region 712 are interesting and which portions may include diagnostically important items. Therefore, a high-magnification lens in combination withcamera 518 may be used to capture an image of all “interesting” scan-zelles 1110. Subsequently, the image data generated during the COLLECT phase of operation may be transmitted tocontroller 512, and may be compressed and stored for subsequent image reconstruction and inspection by a clinician. - Having established both the pattern of “interesting” scan-zelles1110 and the focus plane,
system 500 can now lay out a series of locations where high-resolution images will be taken during the COLLECT phase of operation. This can be done in various ways. In a first exemplary technique, the size and shape of scan-zelles 1110 may be set to match the FOV of a high-resolution lens. For example, if the FOV is square and the initial image was a 1× lens, a 4× lens in which the 4× FOVs can be designed to fit in the 1× FOV may be used. The resulting scan-zelles 1110 are therefore one-fourth the size of the 1× FOV. - In another exemplary technique, the size and shape of scan-
zelles 1110 may not match the FOV of the high-resolution lens. One way of handling this situation is that the pattern of high-power FOVs that are to be collected (i.e., “SCAN” FOVs) is predetermined, i.e., the pattern is a tiling of the slide, possibly with overlap (where the percentage of overlap may be a parameter of the program based upon the mechanical stage positional inaccuracy). The pattern of “interesting” scan-zelles 1110 may then be projected upon this predetermined pattern of high-power FOVs, using algorithms knowing the dimensions and boundaries of both. - The high-power FOVs needed to capture all the “interesting” scan-
zelles 1110 are thus established and tabulated. In other words, any high-power FOV intersecting an “interesting” scan-zelles 1110 is flagged as needed. Non-intersecting high-power FOVs are flagged as not needed. Thus, a list is generated of the high-power FOVs of which images are to be taken to capture all “interesting” scan-zelles 1110. It is possible that only a small portion of these high-power FOVs actually include “interesting” scan-zelles 1110, but such coverage may be necessary so that no item of interest is missed. This method of predetermining the layout of any magnification FOV is used with attributes or combinations of attributes. Example attributes include content for focus but not content for scan, content for scan but not content for focus, content for wanting to revisit for higher power, and the like. - Another way of handling the situation is as follows. Having determined the pattern of “interesting” scan-
zelles 1110 or even knowing one of more broad regions that are “interesting,” everything in that known region may be captured at a higher power. An algorithm can be run to determine the minimum number of high-power FOVs necessary to cover a known region of “interesting” scan-zelles 1110. Additionally, a first high-power FOV may be aligned to a corner of a first “interesting” scan-zelle 1110; the pattern of high-power FOVs is then continued until the entire region of “interesting” scan-zelles 1110 is covered. The period of high-power FOVs will differ from the period of “interesting” scan-zelles 1110. In general, high-power FOVs can be further subdivided into scan-zelles 1110. The high-power FOVs may be stitched together to reconstruct the picture. This process can be repeated any number of times for any number of levels of magnification. - Alternatively, when the “interesting” items are sparse (spread out), the chance of having two adjacent “interesting” scan-
zelles 1110 is low, especially as the magnification increases, i.e., the object is smaller than the FOV. In this case, instead of working in scan-zelles 1110, an “interesting” item is simply located and a high-power FOV is centered upon them, capturing them. In this case, the location of just the “interesting” items to be located is stored withincontroller 512. - Image data compression associated with the SCAN or COLLECT phase of operation may be achieved using block compression algorithms that are applicable to pixel-level compression on the high-resolution images. Many data compression techniques may be used. An exemplary technique includes the following. Each high-power FOV is compressed in two ways: a first pass using a modification of the well-known Classic Huffman Shift compression algorithm (see, for example, I.R.E. September 1952, page 1098, “A Method for the Construction of Minimum-Redundancy Codes” by David A. Huffman, which is herein incorporated by reference) and a second pass using a White Clip compression algorithm.
- The compression algorithms are more efficient in this application because some information has been obtained about the data, i.e., information has been obtained about what is and is not important, and thus parameters may be set accordingly. This knowledge thus allows the compression to be modified for efficiency. For example, the Classic Huffman compression algorithm may take advantage of the understanding of the distribution of the colors in the image. Filters may be defined as to what colors are of higher value and which are of less or no value. For example, the non-valuable colors may be set to the same value (and thus require minimal storage space) without compromising picture quality when restored. Whatever color variations exist in the white space, for example, are non-informational; thus, restoring this space as uniformly white is acceptable.
- In summary,
microscope imaging system 500 may perform a silhouette scan operation for improved image data storage requirements and data acquisition time. The imaging system and method described above may allow for an image capture ofscan region 712 and a first level of data compression at a low-resolution stage.Scan zelles 1110 may be sorted into “non-interesting” vs. “interesting,” and data associated with the “non-interesting” scan-zelles 1110 may be discarded. - Subsequently, a second level of data compression at the high-resolution stage may be performed, wherein the “non-interesting” space may be set to the same value, while everything of importance is stored losslessly. Systems and techniques of the current disclosure may provide a substantial benefit in that they may be performed autonomously. That is, once the process is initiated (e.g., when one or more slides are provided to microscope subsystem510), the process may execute without human intervention or decision-making.
- An illustration follows of the general operation of
microscope imaging system 500 analyzing, for example, sample 716 of a lymph node. The lymph node tissue may be stained with a dye that makes a suspected cancer cell turn brown, while the rest of the tissue may be stained with a common blue dye that is used to mark tissue. One or more parameters of the test are set accordingly. For example, a parameter may be set to indicate that an area that is blue or brown is “interesting.” Further, another parameter may be set to indicate that an area that is brown is “very interesting.” - A FIND phase of operation occurs upon
microscope slide 700 where a low-resolution image (i.e., FIND silhouette 1000) of the entire lymph node is captured. Based upon the predetermined parameters, image-processing algorithms and image analysis algorithms are executed bycontroller 512. Anything that is found to be blue or brown is classified as “interesting,” but brown is further classified as “very interesting.” - In a FOCUS phase of operation, a number of focus points and their locations may be determined. A focus plane may be derived from the focus points. In a SCAN phase of operation, a more precise higher-resolution image is generated (i.e., SCAN silhouette1100) for processing. Optionally, the FOCUS and SCAN phases may be repeated any number of times at progressively higher resolutions to identify “interesting” regions.
- A COLLECT phase of operation is performed, which may use focus cues from the previous phase. Another image scan is done on the “very interesting” scan-
zelles 1110 with a high-power lens. This example illustrates that varying degrees of “interestingness” may be defined. As a result of the FIND phase, the FOCUS phase, the SCAN phase, and the COLLECT phase, a set of tables are generated that have effectively compressed scan-zelles 1110 to pixels, each having one or more classifications as, for example, non-interesting, interesting, very interesting, interesting for focus points, interesting for higher power, and the like. The image data generated during the SCAN or COLLECT phases of operation may be transmitted tocontroller 512. Data compression algorithms may be executed on the image data, thereby completing a second level of image data compression. Having completed these two levels of data compression, the image data is stored in memory of controller 512 (or other memory). The image data may subsequently be accessed for future image reconstruction and inspection by a clinician. - FIG. 12 illustrates a
method 1200 of performing a silhouette scan operation for improved image data acquisition time and image data storage. In order to perform a specific test of one ormore samples 716 on one ormore slides 700, one or more suitable parameters are determined or provided. The parameters may be used to analyzeappropriate zelles 810 and minor-zelles 910 and to determine whether they include anything designated as interesting (1210). Many parameters are possible. For example, parameters may include specimen color, size, features, or distribution and sensitivity for the specific test being run. - A target slide may be provided to
system 500, and an ID of atarget microscope slide 700 may be read bybarcode reader 516 so that any image data saved may be linked to the appropriate microscope slide 700 (1212). -
Scan region 712 oftarget microscope slide 700 is scanned viamicroscope subsystem 510 in combination with camera 518 (1214).Scan region 712 is partitioned into an array ofcontiguous zelles 810 covering the entire area ofscan region 712. The area of each zelle 810 may be defined by the power (i.e., magnification) setting of the microscope (for example, the area of each zelle 810 may correspond to the area of the FOV of the setting of the microscope). Thus, the initial scan ofscan region 712 may be performed efficiently.Zelles 810 may overlap slightly or abut. Overlap may be useful to ensure that no region is missed due to mechanical inaccuracies in the X-Y stage ofmicroscope subsystem 510. - Since a low-power lens has a greater depth of focus than a high-power lens,
microscope subsystem 510 may be able to search for tissue during the initial scan ofscan region 712 without refocusing. For example, a low-power lens of the microscope may be focused at a best-guess plane or one derived from microscope calibration. Optionally, a separate low-resolution, wide-FOV camera may be used. - Images obtained during a scan of
scan region 712 may be processed using image-processing algorithms and/or image analysis algorithms to determine any regions ofmicroscope slide 700 having content worthwhile to reexamine at a higher magnification during a subsequent FOCUS, SCAN or COLLECT phase of operation. In some implementations, anyzelle 810 or minor-zelle 910 found to have specimen content is classified as “interesting” and mapped as a logical “1.” By contrast, anyzelle 810 or minor-zelle 910 found to have no specimen content is classified as “non-interesting” and mapped as a logical “0.” In this manner, a silhouette of the specimen, i.e.,sample 716, is formed, thereby creating what is effectively a low-resolution image that may be processed. An image table is generated that represents the low-resolution image ofsample 716. - Prior to a higher resolution operation, the number and/or placement of focus points may be selected (1216), as described above. A focus plane may be determined from the focus points. A SCAN operation may be performed, and image data may be compressed (1218). Given that statistically, based upon the entire area of a
particular zelle 810 or minor-zelle 910, there might be very little material of interest, each zelle 810 or minor-zelle 910 may be subdivided into yet smaller regions using data processing. Eachzelle 810 or minor-zelle 910 may therefore be further partitioned, forming aSCAN silhouette 600 formed of the plurality of scan-zelles 1110, as shown in FIG. 11. - Like
zelles 810 or minor-zelles 910, the “interestingness” of these scan-zelles 1110 is assessed via image-processing algorithms and image analysis algorithms. This is done in preparation for a future operation of collecting and saving a series of higher magnification FOVs (i.e., in the COLLECT phase), as only the “interesting” regions are generally examined at a higher magnification. - In some implementations, scan-
zelles 1110 may be defined to match the anticipated FOVs of the higher power images of the COLLECT phase of operation. The size of scan-zelles 1110 and whether scan-zelles 1110 match up with the anticipated FOVs of the higher power images are adjustable system parameters that may be set (1210). Using image-processing algorithms and image analysis algorithms that may be executed by controller 512 (or other data processor), scan-zelles 1110 found to have specimen content are classified as “interesting” and mapped as a logical “1.” By contrast, scan-zelles 1110 found to have no specimen content are classified as “non-interesting” and mapped as a logical “0.” In this manner, further image compression occurs and a yet more precise silhouette ofsample 716 is formed, thereby creating a slightly higher-resolution image, than that generated in the FIND phase. This higher resolution image may be processed using image-processing algorithms. - The system may determine whether another iteration is performed (1220). The decision may be based on one or more parameters, such as the type of test, expected size of feature, etc. If iteration is desired, the above actions may be repeated (1216 et seq.) If iteration is not desired, a sequence of image shots for a COLLECT phase may be planned (1222). Various sequences of operations may occur to modify the “interesting” regions as determined by the requirements of a particular test. The result is a table of “interesting” and “non-interesting” scan-
zelles 1110 as modified via the specific requirements. - The COLLECT phase may then be performed (1224). Having established both the pattern of “interesting” scan-zelles 1110 and the focus plane,
system 500 now lays out a series of locations where high-resolution images are taken during this COLLECT phase of operation. As a result, with specific knowledge of interesting regions and of diagnostically important items, a high-magnification lens in combination withcamera 518 are used to capture an image of all “interesting” scan-zelles 1110. Image data generated during the COLLECT phase of operation may be transmitted tocontroller 512. - Image data acquired in the SCAN and/or COLLECT phase may be compressed (1226), as described above. A determination may be made as to whether the
particular slide 700 withsample 716 is to be examined and analyzed (1228). If it is, a clinician may recall the image data for the selectedmicroscope slide 700 and controller 512 (or other data processor) may subsequently reconstruct the image by stitching together individual scan-zelles 1110, and/or high-power FOVs. The image may then be displayed to the clinician for inspection viadisplay device 514. - The various techniques, methods, and aspects of the invention described above can be implemented in part or in whole using computer-based systems and methods. Additionally, computer-based systems and methods can be used to augment or enhance the functionality described above, increase the speed at which the functions can be performed, and provide additional features and aspects as a part of or in addition to those of the invention described elsewhere in this document. Various computer-based systems, methods and implementations in accordance with the above-described technology are presented below.
- A processor-based system can include a main memory, preferably random access memory (RAM), and can also include a secondary memory. The secondary memory can include, for example, a hard disk drive and/or a removable storage drive, representing a floppy disk drive, a magnetic tape drive, an optical disk drive, etc. The removable storage drive reads from and/or writes to a removable storage medium. Removable storage medium refers to a floppy disk, magnetic tape, optical disk, and the like, which is read by and written to by a removable storage drive. As will be appreciated, the removable storage medium can comprise computer software and/or data.
- In alternative embodiments, the secondary memory may include other similar means for allowing computer programs or other instructions to be loaded into a computer system. Such means can include, for example, a removable storage unit and an interface. Examples of such can include a program cartridge and cartridge interface (such as the found in video game devices), a movable memory chip (such as an EPROM or PROM) and associated socket, and other removable storage units and interfaces, which allow software and data to be transferred from the removable storage unit to the computer system.
- The computer system can also include a communications interface. Communications interfaces allow software and data to be transferred between computer system and external devices. Examples of communications interfaces can include a modem, a network interface (such as, for example, an Ethernet card), a communications port, a PCMCIA slot and card, and the like. Software and data transferred via a communications interface are in the form of signals, which can be electronic, electromagnetic, optical or other signals capable of being received by a communications interface. These signals are provided to communications interface via a channel capable of carrying signals and can be implemented using a wireless medium, wire or cable, fiber optics or other communications medium. Some examples of a channel can include a phone line, a cellular phone link, an RF link, a network interface, and other communications channels.
- In this document, the terms “computer program medium” and “computer usable medium” are used to refer generally to media such as a removable storage device, a disk capable of installation in a disk drive, and signals on a channel. These computer program products are means for providing software or program instructions to a computer system.
- Computer programs (also called computer control logic) are stored in main memory and/or secondary memory. Computer programs can also be received via a communications interface. Such computer programs, when executed, enable the computer system to perform the features of the invention as discussed herein. In particular, the computer programs, when executed, enable the processor to perform the features of the invention. Accordingly, such computer programs represent controllers of the computer system.
- In an embodiment where the elements are implemented using software, the software may be stored in, or transmitted via, a computer program product and loaded into a computer system using a removable storage drive, hard drive or communications interface. The control logic (software), when executed by the processor, causes the processor to perform the functions of the invention as described herein.
- In another embodiment, the elements are implemented primarily in hardware using, for example, hardware components such as PALs, application specific integrated circuits (ASICs) or other hardware components. Implementation of a hardware state machine so as to perform the functions described herein will be apparent to person skilled in the relevant art(s). In yet another embodiment, elements are implanted using a combination of both hardware and software.
- In another embodiment, the computer-based methods can be accessed or implemented over the World Wide Web by providing access via a Web Page to the methods of the invention. Accordingly, the Web Page is identified by a Universal Resource Locator (URL). The URL denotes both the server machine and the particular file or page on that machine. In this embodiment, it is envisioned that a consumer or client computer system interacts with a browser to select a particular URL, which in turn causes the browser to send a request for that URL or page to the server identified in the URL. Typically the server responds to the request by retrieving the requested page and transmitting the data for that page back to the requesting client computer system (the client/server interaction is typically performed in accordance with the hypertext transport protocol (“HTTP”)). The selected page is then displayed to the user on the client's display screen. The client may then cause the server containing a computer program of the invention to launch an application to, for example, perform an analysis according to the invention.
- Although only a few implementations have been described in detail above, other modifications are possible. For example, different types of image acquisition, processing, and analysis systems may be used. The logic flows depicted in FIGS. 2-4 and12 do not require the particular order shown, or sequential order, to achieve desirable results. For example, reading a slide identifier may be performed at many different places within the overall process. In certain implementations, multitasking and parallel processing may be preferable.
- Other embodiments may be within the scope of the following claims.
Claims (27)
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/413,493 US20040202357A1 (en) | 2003-04-11 | 2003-04-11 | Silhouette image acquisition |
PCT/US2004/011334 WO2004093004A2 (en) | 2003-04-11 | 2004-04-12 | Silhouette image acquisition |
US12/587,107 US8369591B2 (en) | 2003-04-11 | 2009-10-01 | Silhouette image acquisition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/413,493 US20040202357A1 (en) | 2003-04-11 | 2003-04-11 | Silhouette image acquisition |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/587,107 Continuation US8369591B2 (en) | 2003-04-11 | 2009-10-01 | Silhouette image acquisition |
Publications (1)
Publication Number | Publication Date |
---|---|
US20040202357A1 true US20040202357A1 (en) | 2004-10-14 |
Family
ID=33131414
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/413,493 Abandoned US20040202357A1 (en) | 2003-04-11 | 2003-04-11 | Silhouette image acquisition |
US12/587,107 Expired - Fee Related US8369591B2 (en) | 2003-04-11 | 2009-10-01 | Silhouette image acquisition |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/587,107 Expired - Fee Related US8369591B2 (en) | 2003-04-11 | 2009-10-01 | Silhouette image acquisition |
Country Status (2)
Country | Link |
---|---|
US (2) | US20040202357A1 (en) |
WO (1) | WO2004093004A2 (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040047033A1 (en) * | 2002-09-10 | 2004-03-11 | Olympus Optical Co., Ltd. | Microscopic image capture apparatus and microscopic image capturing method |
US20040260402A1 (en) * | 2003-06-20 | 2004-12-23 | Baldini Steven E. | Method of manufacturing a socket portion of a prosthetic limb |
US20050136509A1 (en) * | 2003-09-10 | 2005-06-23 | Bioimagene, Inc. | Method and system for quantitatively analyzing biological samples |
US20060149162A1 (en) * | 2004-11-29 | 2006-07-06 | Derek Daw | Graphical user interface for tissue biopsy system |
US20070031056A1 (en) * | 2005-08-02 | 2007-02-08 | Perz Cynthia B | System for and method of focusing in automated microscope systems |
US20070280517A1 (en) * | 2005-06-21 | 2007-12-06 | De La Torre-Bueno Jose | Serial section analysis for computer-controlled microscopic imaging |
US20080231973A1 (en) * | 2006-08-04 | 2008-09-25 | Ikonisys, Inc. | Mechanism for a microscope objective changer and filter changer |
US20100141682A1 (en) * | 2003-04-11 | 2010-06-10 | Carl Zeiss Microlmaging Ais, Inc. | Silhouette image acquisition |
US8116543B2 (en) * | 2005-08-02 | 2012-02-14 | Carl Zeiss Microimaging Gmbh | System for and method of intelligently directed segmentation analysis for automated microscope systems |
US20120082366A1 (en) * | 2009-06-03 | 2012-04-05 | Nec Corporation | Pathologic image diagnostic system, pathologic image diagnostic method, and pathologic image diagnostic program |
US20130011028A1 (en) * | 2010-01-04 | 2013-01-10 | Nec Corporation | Image diagnostic method, image diagnostic apparatus, and image diagnostic program |
US8428887B2 (en) | 2003-09-08 | 2013-04-23 | Ventana Medical Systems, Inc. | Method for automated processing of digital images of tissue micro-arrays (TMA) |
CN104335098A (en) * | 2012-05-28 | 2015-02-04 | 索尼公司 | Information processing device, information processing method, program and microscope system |
US20150124078A1 (en) * | 2012-07-04 | 2015-05-07 | Sony Corporation | Information processing apparatus, information processing method, program, and microscope system |
US20170146782A1 (en) * | 2006-09-06 | 2017-05-25 | Leica Microsystems Cms Gmbh | Method and microscopy system for scanning a sample |
DE102016110988A1 (en) * | 2016-06-15 | 2017-12-21 | Sensovation Ag | Method for digitally recording a sample through a microscope |
US10314563B2 (en) | 2014-11-26 | 2019-06-11 | Devicor Medical Products, Inc. | Graphical user interface for biopsy device |
CN113631982A (en) * | 2019-02-04 | 2021-11-09 | 脱其泰有限责任公司 | Automated micro-scanning system and method |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8169495B2 (en) * | 2006-12-01 | 2012-05-01 | Broadcom Corporation | Method and apparatus for dynamic panoramic capturing |
WO2008118886A1 (en) * | 2007-03-23 | 2008-10-02 | Bioimagene, Inc. | Digital microscope slide scanning system and methods |
JP6838912B2 (en) * | 2016-09-29 | 2021-03-03 | キヤノン株式会社 | Image processing equipment, image processing methods and programs |
ES2928577T3 (en) | 2017-09-29 | 2022-11-21 | Leica Biosystems Imaging Inc | 2D and 3D Fixed Z-Scan |
TWI819421B (en) | 2021-12-02 | 2023-10-21 | 倍利科技股份有限公司 | High-speed automatic scanning AI-assisted interpretation system and scanning method |
Citations (94)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3824393A (en) * | 1971-08-25 | 1974-07-16 | American Express Invest | System for differential particle counting |
US4011004A (en) * | 1975-02-06 | 1977-03-08 | Geometric Data Corporation | Pivotable stage for microscope system |
US4196265A (en) * | 1977-06-15 | 1980-04-01 | The Wistar Institute | Method of producing antibodies |
US4210419A (en) * | 1978-02-17 | 1980-07-01 | California Institute Of Technology | Automated quantitative muscle biopsy analysis system |
US4249825A (en) * | 1979-05-14 | 1981-02-10 | The Trustees Of Boston University | Method and apparatus suitable for ocular blood flow analysis |
US4338024A (en) * | 1980-05-02 | 1982-07-06 | International Remote Imaging Systems, Inc. | Flow analyzer and system for analysis of fluids with particles |
US4393466A (en) * | 1980-09-12 | 1983-07-12 | International Remote Imaging Systems | Method of analyzing particles in a dilute fluid sample |
US4513438A (en) * | 1982-04-15 | 1985-04-23 | Coulter Electronics, Inc. | Automated microscopy system and method for locating and re-locating objects in an image |
US4612614A (en) * | 1980-09-12 | 1986-09-16 | International Remote Imaging Systems, Inc. | Method of analyzing particles in a fluid sample |
US4656594A (en) * | 1985-05-06 | 1987-04-07 | National Biomedical Research Foundation | Operator-interactive automated chromosome analysis system producing a karyotype |
US4673973A (en) * | 1985-02-04 | 1987-06-16 | National Biomedical Research Foundation | Split-image, multi-power microscopic image display system and method |
US4741043A (en) * | 1985-11-04 | 1988-04-26 | Cell Analysis Systems, Inc. | Method of and an apparatus for image analyses of biological specimens |
US4945220A (en) * | 1988-11-16 | 1990-07-31 | Prometrix Corporation | Autofocusing system for microscope having contrast detection means |
US4991223A (en) * | 1988-06-30 | 1991-02-05 | American Innovision, Inc. | Apparatus and method for recognizing image features using color elements |
US5003165A (en) * | 1989-05-25 | 1991-03-26 | International Remote Imaging Systems, Inc. | Method and an apparatus for selecting the best focal position from a plurality of focal positions |
US5008185A (en) * | 1985-11-04 | 1991-04-16 | Cell Analysis Systems, Inc. | Methods and apparatus for the quantitation of nuclear proteins |
US5014198A (en) * | 1987-01-13 | 1991-05-07 | Kabushiki Kaisha Toshiba | Image information compression apparatus |
US5016173A (en) * | 1989-04-13 | 1991-05-14 | Vanguard Imaging Ltd. | Apparatus and method for monitoring visually accessible surfaces of the body |
US5085325A (en) * | 1988-03-08 | 1992-02-04 | Simco/Ramic Corporation | Color sorting system and method |
US5087965A (en) * | 1989-06-28 | 1992-02-11 | American Innovision, Inc. | Recognition of image colors using arbitrary shapes in color space |
US5123055A (en) * | 1989-08-10 | 1992-06-16 | International Remote Imaging Systems, Inc. | Method and an apparatus for differentiating a sample of biological cells |
US5202931A (en) * | 1987-10-06 | 1993-04-13 | Cell Analysis Systems, Inc. | Methods and apparatus for the quantitation of nuclear protein |
US5231580A (en) * | 1991-04-01 | 1993-07-27 | The United States Of America As Represented By The Secretary Of The Department Of Health And Human Services | Automated method and apparatus for determining characteristics of nerve fibers |
US5233684A (en) * | 1990-06-26 | 1993-08-03 | Digital Equipment Corporation | Method and apparatus for mapping a digital color image from a first color space to a second color space |
US5287272A (en) * | 1988-04-08 | 1994-02-15 | Neuromedical Systems, Inc. | Automated cytological specimen classification system and method |
US5317140A (en) * | 1992-11-24 | 1994-05-31 | Dunthorn David I | Diffusion-assisted position location particularly for visual pen detection |
US5321545A (en) * | 1988-10-21 | 1994-06-14 | Biocom S.A. | Microscope stage for rapid and indexed analysis of filters and other media carrying multiple samples, and a method of analyzing such samples using said stage |
US5333207A (en) * | 1990-11-07 | 1994-07-26 | Neuromedical Systems, Inc. | Inspection apparatus and method with inspection auditing for images presented on a display |
US5338924A (en) * | 1992-08-11 | 1994-08-16 | Lasa Industries, Inc. | Apparatus and method for automatic focusing of light using a fringe plate |
US5409007A (en) * | 1993-11-26 | 1995-04-25 | General Electric Company | Filter to reduce speckle artifact in ultrasound imaging |
US5428690A (en) * | 1991-09-23 | 1995-06-27 | Becton Dickinson And Company | Method and apparatus for automated assay of biological specimens |
US5432871A (en) * | 1993-08-04 | 1995-07-11 | Universal Systems & Technology, Inc. | Systems and methods for interactive image data acquisition and compression |
US5449622A (en) * | 1993-02-16 | 1995-09-12 | Yabe; Ryohei | Method and apparatus for analyzing stained particles |
US5481401A (en) * | 1991-05-16 | 1996-01-02 | Olympus Optical Co., Ltd. | Ultraviolet microscope |
US5499097A (en) * | 1994-09-19 | 1996-03-12 | Neopath, Inc. | Method and apparatus for checking automated optical system performance repeatability |
US5515172A (en) * | 1993-07-19 | 1996-05-07 | Xerox Corporation | Apparatus and method for enhanced color to color conversion |
US5526258A (en) * | 1990-10-10 | 1996-06-11 | Cell Analysis System, Inc. | Method and apparatus for automated analysis of biological specimens |
US5533628A (en) * | 1992-03-06 | 1996-07-09 | Agri Tech Incorporated | Method and apparatus for sorting objects by color including stable color transformation |
US5602941A (en) * | 1993-05-21 | 1997-02-11 | Digital Equipment Corporation | Input modification system for multilevel dithering |
US5619032A (en) * | 1995-01-18 | 1997-04-08 | International Remote Imaging Systems, Inc. | Method and apparatus for automatically selecting the best focal position from a plurality of focal positions for a focusing apparatus |
US5621532A (en) * | 1994-12-08 | 1997-04-15 | Nikon Corporation | Laser scanning microscope utilizing detection of a far-field diffraction pattern with 2-dimensional detection |
US5625705A (en) * | 1994-06-03 | 1997-04-29 | Neuromedical Systems, Inc. | Intensity texture based classification system and method |
US5625709A (en) * | 1994-12-23 | 1997-04-29 | International Remote Imaging Systems, Inc. | Method and apparatus for identifying characteristics of an object in a field of view |
US5627907A (en) * | 1994-12-01 | 1997-05-06 | University Of Pittsburgh | Computerized detection of masses and microcalcifications in digital mammograms |
US5632402A (en) * | 1992-12-21 | 1997-05-27 | Graphic Packaging Corporation | Carton blank and method for forming it |
US5646677A (en) * | 1995-02-23 | 1997-07-08 | Motorola, Inc. | Method and apparatus for interactively viewing wide-angle images from terrestrial, space, and underwater viewpoints |
US5647025A (en) * | 1994-09-20 | 1997-07-08 | Neopath, Inc. | Automatic focusing of biomedical specimens apparatus |
US5671288A (en) * | 1995-05-31 | 1997-09-23 | Neopath, Inc. | Method and apparatus for assessing slide and specimen preparation quality |
US5706093A (en) * | 1995-12-05 | 1998-01-06 | Olympus Optical Co., Ltd. | Color classification apparatus |
US5726009A (en) * | 1989-03-20 | 1998-03-10 | Anticancer, Inc. | Native-state method and system for determining viability and proliferative capacity of tissues in vitro |
US5735387A (en) * | 1995-07-14 | 1998-04-07 | Chiron Diagnostics Corporation | Specimen rack handling system |
US5740270A (en) * | 1988-04-08 | 1998-04-14 | Neuromedical Systems, Inc. | Automated cytological specimen classification system and method |
US5783814A (en) * | 1994-01-18 | 1998-07-21 | Ultrapointe Corporation | Method and apparatus for automatically focusing a microscope |
US5867598A (en) * | 1996-09-26 | 1999-02-02 | Xerox Corporation | Method and apparatus for processing of a JPEG compressed image |
US5877161A (en) * | 1993-08-05 | 1999-03-02 | University Technologies International Inc. | Cyclin D1 negative regulatory activity |
US5880473A (en) * | 1997-07-28 | 1999-03-09 | Applied Imaging, Inc. | Multifluor-fluorescence in-situ hybridization (M-FISH) imaging techniques using multiple multiband filters with image registration |
US5889881A (en) * | 1992-10-14 | 1999-03-30 | Oncometrics Imaging Corp. | Method and apparatus for automatically detecting malignancy-associated changes |
US5888742A (en) * | 1997-10-28 | 1999-03-30 | Incyte Pharmaceuticals, Inc. | Human phospholipid binding proteins |
US5911003A (en) * | 1996-04-26 | 1999-06-08 | Pressco Technology Inc. | Color pattern evaluation system for randomly oriented articles |
US5911327A (en) * | 1996-10-02 | 1999-06-15 | Nippon Steel Corporation | Method of automatically discriminating and separating scraps containing copper from iron scraps |
US6011595A (en) * | 1997-09-19 | 2000-01-04 | Eastman Kodak Company | Method for segmenting a digital image into a foreground region and a key color region |
US6031929A (en) * | 1996-07-18 | 2000-02-29 | University Of Pittsburgh | Image facsimile with real time image segmentation |
US6040139A (en) * | 1995-09-19 | 2000-03-21 | Bova; G. Steven | Laser cell purification system |
US6058322A (en) * | 1997-07-25 | 2000-05-02 | Arch Development Corporation | Methods for improving the accuracy in differential diagnosis on radiologic examinations |
US6072570A (en) * | 1997-07-24 | 2000-06-06 | Innotech | Image quality mapper for progressive eyeglasses |
US6097838A (en) * | 1996-12-19 | 2000-08-01 | Xerox Corporation | Color correction of a compressed image |
US6101265A (en) * | 1996-08-23 | 2000-08-08 | Bacus Research Laboratories, Inc. | Method and apparatus for acquiring and reconstructing magnified specimen images from a computer-controlled microscope |
US6117985A (en) * | 1995-06-16 | 2000-09-12 | Stemcell Technologies Inc. | Antibody compositions for preparing enriched cell preparations |
US6122400A (en) * | 1997-09-26 | 2000-09-19 | Sarnoff Corporation | Compression encoder bit allocation utilizing colormetric-adaptive weighting as in flesh-tone weighting |
US6125194A (en) * | 1996-02-06 | 2000-09-26 | Caelum Research Corporation | Method and system for re-screening nodules in radiological images using multi-resolution processing, neural network, and image processing |
US6169816B1 (en) * | 1997-05-14 | 2001-01-02 | Applied Imaging, Inc. | Identification of objects of interest using multiple illumination schemes and finding overlap of features in corresponding multiple images |
US6215892B1 (en) * | 1995-11-30 | 2001-04-10 | Chromavision Medical Systems, Inc. | Method and apparatus for automated image analysis of biological specimens |
US6215894B1 (en) * | 1999-02-26 | 2001-04-10 | General Scanning, Incorporated | Automatic imaging and analysis of microarray biochips |
US6225636B1 (en) * | 1996-10-25 | 2001-05-01 | Applied Imaging, Inc. | Multifluor-fluorescence in-situ hybridization (M-FISH) imaging techniques using multiple multiband filters with image registration |
US6236031B1 (en) * | 1998-09-04 | 2001-05-22 | Sony Corporation | Optical head, recording and/or reproducing apparatus, and optical disc drive with an auxiliary focus servo system |
US6238892B1 (en) * | 1991-10-25 | 2001-05-29 | N.V. Innogenetics S.A. | Monoclonal antibodies directed against the microtubule-associated protein tau |
US6259826B1 (en) * | 1997-06-12 | 2001-07-10 | Hewlett-Packard Company | Image processing method and device |
US6259807B1 (en) * | 1997-05-14 | 2001-07-10 | Applied Imaging Corp. | Identification of objects of interest using multiple illumination schemes and finding overlap of features in corresponding multiple images |
US6272235B1 (en) * | 1997-03-03 | 2001-08-07 | Bacus Research Laboratories, Inc. | Method and apparatus for creating a virtual microscope slide |
US6275777B1 (en) * | 1997-10-28 | 2001-08-14 | Olympus Optical Co., Ltd. | Scanning cytometer |
US6281874B1 (en) * | 1998-08-27 | 2001-08-28 | International Business Machines Corporation | Method and system for downloading graphic images on the internet |
US6290907B1 (en) * | 1997-09-11 | 2001-09-18 | Hitachi, Ltd. | Sample handling system |
US6374989B1 (en) * | 1997-11-14 | 2002-04-23 | Bayer Corporation | Conveyor system for clinical test apparatus |
US20020067409A1 (en) * | 2000-12-06 | 2002-06-06 | Bio-View Ltd. | Data acquisition and display system and method |
US6404916B1 (en) * | 1999-08-04 | 2002-06-11 | Chromavision Medical Systems, Inc. | Method and apparatus for applying color thresholds in light microscopy |
US6404906B2 (en) * | 1997-03-03 | 2002-06-11 | Bacus Research Laboratories,Inc. | Method and apparatus for acquiring and reconstructing magnified specimen images from a computer-controlled microscope |
US6415058B2 (en) * | 1998-10-27 | 2002-07-02 | Hewlett-Packard Company | System for compression of digital images comprising low detail areas |
US6418236B1 (en) * | 1999-06-24 | 2002-07-09 | Chromavision Medical Systems, Inc. | Histological reconstruction and automated image analysis |
US6514722B2 (en) * | 1997-03-27 | 2003-02-04 | Oncosis | Method and apparatus for selectively targeting specific cells within a cell population |
US6518554B1 (en) * | 2000-05-24 | 2003-02-11 | Chromavision Medical Systems, Inc. | Reverse focusing methods and systems |
US20030124589A1 (en) * | 2001-10-12 | 2003-07-03 | Vysis, Inc. | Imaging microarrays |
US6674896B1 (en) * | 1999-07-13 | 2004-01-06 | Chromavision Medical Systems, Inc. | Device for applying accurate color thresholds in real time |
US6697509B2 (en) * | 2001-10-04 | 2004-02-24 | Chromavision Medical Systems, Inc. | Method and apparatus for scoring the uptake of markers in cells |
US6718053B1 (en) * | 1996-11-27 | 2004-04-06 | Chromavision Medical Systems, Inc. | Method and apparatus for automated image analysis of biological specimens |
Family Cites Families (76)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE2126393A1 (en) | 1970-11-23 | 1971-12-02 | Talbot J | Method and device for the analysis of geometric properties of the components of a system |
US4207554A (en) | 1972-08-04 | 1980-06-10 | Med-El Inc. | Method and apparatus for automated classification and analysis of cells |
US4125828A (en) | 1972-08-04 | 1978-11-14 | Med-El Inc. | Method and apparatus for automated classification and analysis of cells |
US3851972A (en) * | 1973-10-18 | 1974-12-03 | Coulter Electronics | Automatic method and system for analysis and review of a plurality of stored slides |
US4037965A (en) | 1976-03-23 | 1977-07-26 | Leeds & Northrup Company | Method and optical means for determining dimensional characteristics of the particle distribution in a collection of particles |
US4097845A (en) | 1976-11-01 | 1978-06-27 | Rush-Presbyterian-St. Luke's Medical Center | Method of and an apparatus for automatic classification of red blood cells |
DE2903625A1 (en) | 1978-02-03 | 1979-08-09 | Rush Presbyterian St Luke | DEVICE FOR AUTOMATIC BLOOD ANALYSIS |
JPS5661650A (en) | 1979-10-24 | 1981-05-27 | Omron Tateisi Electronics Co | Analyzing device of cell |
US4453266A (en) | 1980-04-21 | 1984-06-05 | Rush-Presbyterian-St. Luke's Medical Center | Method and apparatus for measuring mean cell volume of red blood cells |
US4549204A (en) | 1981-11-26 | 1985-10-22 | The Secretary Of State For Defence In Her Britannic Majesty's Government Of The United Kingdom Of Great Britain And Northern Ireland | Diffraction limited imaging systems |
US4538299A (en) | 1981-12-04 | 1985-08-27 | International Remote Imaging Systems, Inc. | Method and apparatus for locating the boundary of an object |
DE3340647A1 (en) | 1983-11-10 | 1985-05-23 | Will Wetzlar Gmbh, 6330 Wetzlar | Method for focusing a microscope and microscope for carrying out the method |
SE455736B (en) | 1984-03-15 | 1988-08-01 | Sarastro Ab | PROCEDURE KIT AND MICROPHOTOMETRATION AND ADDITIONAL IMAGE COMPOSITION |
US4700298A (en) | 1984-09-14 | 1987-10-13 | Branko Palcic | Dynamic microscope image processing scanner |
DE3610165A1 (en) | 1985-03-27 | 1986-10-02 | Olympus Optical Co., Ltd., Tokio/Tokyo | OPTICAL SCAN MICROSCOPE |
GB8509493D0 (en) | 1985-04-12 | 1985-05-15 | Plessey Co Plc | Scanning microscopes |
JPH0713707B2 (en) | 1985-04-26 | 1995-02-15 | オリンパス光学工業株式会社 | Scanning optical microscope |
US4697594A (en) | 1985-08-21 | 1987-10-06 | North American Philips Corporation | Displaying a single parameter image |
JPH0718975B2 (en) | 1986-06-20 | 1995-03-06 | オリンパス光学工業株式会社 | Scanning optical microscope |
US4771469A (en) | 1986-06-30 | 1988-09-13 | Honeywell Inc. | Means and method of representing an object shape by hierarchical boundary decomposition |
JPH0774856B2 (en) | 1986-10-16 | 1995-08-09 | オリンパス光学工業株式会社 | Automatic focus adjustment method |
JPS63164151A (en) | 1986-12-26 | 1988-07-07 | Fuji Photo Film Co Ltd | Electron microscope image output method |
JPS63306413A (en) | 1987-06-09 | 1988-12-14 | Olympus Optical Co Ltd | Scanning type optical microscope |
US5121436A (en) | 1987-08-14 | 1992-06-09 | International Remote Imaging Systems, Inc. | Method and apparatus for generating a plurality of parameters of an object in a field of view |
US4845552A (en) | 1987-08-20 | 1989-07-04 | Bruno Jaggi | Quantitative light microscope using a solid state detector in the primary image plane |
JPH01132932A (en) | 1987-11-18 | 1989-05-25 | Omron Tateisi Electron Co | Signal beam detecting optical system of flow type particle analyser |
JP2686274B2 (en) | 1988-03-24 | 1997-12-08 | 東亜医用電子株式会社 | Cell image processing method and apparatus |
JPH0289016A (en) | 1988-09-26 | 1990-03-29 | Nikon Corp | Scanning type microscope |
US5068909A (en) | 1989-05-18 | 1991-11-26 | Applied Imaging Corporation | Method and apparatus for generating quantifiable video displays |
JP2724502B2 (en) | 1989-06-19 | 1998-03-09 | 東京エレクトロン株式会社 | Scanning microscope equipment |
US5065447A (en) | 1989-07-05 | 1991-11-12 | Iterated Systems, Inc. | Method and apparatus for processing digital data |
US5268966A (en) | 1989-08-10 | 1993-12-07 | International Remote Imaging Systems, Inc. | Method of differentiating particles based upon a dynamically changing threshold |
US5313532A (en) | 1990-01-23 | 1994-05-17 | Massachusetts Institute Of Technology | Recognition of patterns in images |
US5042950A (en) | 1990-05-22 | 1991-08-27 | The United States Of America As Represented By The United States Department Of Energy | Apparatus and method for laser beam diagnosis |
US5555889A (en) | 1990-06-20 | 1996-09-17 | Cedars-Sinai Medical Center | Methods for detecting propensity fibrillation |
US5257182B1 (en) * | 1991-01-29 | 1996-05-07 | Neuromedical Systems Inc | Morphological classification system and method |
US5784162A (en) | 1993-08-18 | 1998-07-21 | Applied Spectral Imaging Ltd. | Spectral bio-imaging methods for biological research, medical diagnostics and therapy |
CA2077781A1 (en) | 1991-09-23 | 1993-03-24 | James W. Bacus | Method and apparatus for automated assay of biological specimens |
US5375177A (en) | 1991-09-27 | 1994-12-20 | E. I. Du Pont De Nemours And Company | Method of identifying and characterizing a valid object by color |
US5463470A (en) | 1991-10-09 | 1995-10-31 | Fuji Photo Film Co., Ltd. | Methods of collecting photometric image data and determining light exposure by extracting feature image data from an original image |
US5309221A (en) | 1991-12-31 | 1994-05-03 | Corning Incorporated | Measurement of fiber diameters with high precision |
US5635402A (en) | 1992-03-05 | 1997-06-03 | Alfano; Robert R. | Technique for determining whether a cell is malignant as opposed to non-malignant using extrinsic fluorescence spectroscopy |
ES2104767T3 (en) | 1992-04-30 | 1997-10-16 | Becton Dickinson Co | PROCEDURE AND DEVICE TO CONTROL THE COORDINATED DISPLACEMENTS OF A PLATFORM. |
US5384862A (en) | 1992-05-29 | 1995-01-24 | Cimpiter Corporation | Radiographic image evaluation apparatus and method |
EP0814334B1 (en) | 1992-07-31 | 2004-09-29 | Fuji Photo Film Co., Ltd. | Method and apparatus for obtaining three-dimensional information of samples |
US5254845A (en) | 1992-08-18 | 1993-10-19 | Applied Imaging International Limited | Automatic focusing system using focal length-changing element |
EP0611231B1 (en) | 1993-02-12 | 1999-09-29 | Eastman Kodak Company | Method for cross-device color calibration and enhancement using explicit constraints |
US5521984A (en) * | 1993-06-10 | 1996-05-28 | Verification Technologies, Inc. | System for registration, identification and verification of items utilizing unique intrinsic features |
IL106691A (en) | 1993-08-13 | 1998-02-08 | Sophis View Tech Ltd | System and method for diagnosis of living tissue diseases |
US5352613A (en) | 1993-10-07 | 1994-10-04 | Tafas Triantafillos P | Cytological screening method |
JP3039594B2 (en) | 1993-10-08 | 2000-05-08 | 株式会社日立製作所 | Staining reagent and method of use |
US5469353A (en) | 1993-11-26 | 1995-11-21 | Access Radiology Corp. | Radiological image interpretation apparatus and method |
JP3134660B2 (en) | 1994-04-14 | 2001-02-13 | 松下電器産業株式会社 | Color conversion method and color conversion device |
US5966465A (en) | 1994-09-21 | 1999-10-12 | Ricoh Corporation | Compression/decompression using reversible embedded wavelets |
US5846749A (en) | 1994-10-12 | 1998-12-08 | The Regents Of The University Of California | Quantitative measurement of tissue protein identified by immunohistochemistry and standardized protein determination |
DE69417899T2 (en) | 1994-11-17 | 1999-11-04 | Chemunex Maisons Alfort | Device and method for recognizing and counting rare mammalian cells |
US5848177A (en) | 1994-12-29 | 1998-12-08 | Board Of Trustees Operating Michigan State University | Method and system for detection of biological materials using fractal dimensions |
US5623415A (en) | 1995-02-16 | 1997-04-22 | Smithkline Beecham Corporation | Automated sampling and testing of biological materials |
US5701172A (en) | 1995-06-07 | 1997-12-23 | Gas Research Institute | Optical flowmeter |
US5690892A (en) | 1995-09-15 | 1997-11-25 | Accumed, Inc. | Cassette for use with automated specimen handling system |
US6151405A (en) | 1996-11-27 | 2000-11-21 | Chromavision Medical Systems, Inc. | System and method for cellular specimen grading |
US6330349B1 (en) | 1995-11-30 | 2001-12-11 | Chromavision Medical Systems, Inc. | Automated method for image analysis of residual protein |
US6005964A (en) | 1996-01-24 | 1999-12-21 | The Board Of Trustees Of The University Of Illinois | Automatic machine vision microscope slide inspection system and method |
US6006191A (en) | 1996-05-13 | 1999-12-21 | Dirienzo; Andrew L. | Remote access medical image exchange system and methods of operation therefor |
US6141602A (en) | 1997-09-25 | 2000-10-31 | Hitachi, Ltd. | Specimen processing system |
DE19819813C2 (en) | 1998-05-04 | 2000-11-02 | Olympus Diagnostica Gmbh | Use of a laboratory primary sample distributor for archiving |
US6313452B1 (en) | 1998-06-10 | 2001-11-06 | Sarnoff Corporation | Microscopy system utilizing a plurality of images for enhanced image processing capabilities |
WO2000062247A1 (en) | 1999-04-13 | 2000-10-19 | Chromavision Medical Systems, Inc. | Histological reconstruction and automated image analysis |
WO2001022350A1 (en) | 1999-07-13 | 2001-03-29 | Chromavision Medical Systems, Inc. | Apparatus for counting color transitions and areas in real time camera images |
US6841388B2 (en) | 2000-12-05 | 2005-01-11 | Vysis, Inc. | Method and system for diagnosing pathology in biological samples by detection of infrared spectral markers |
US6466690C1 (en) | 2000-12-19 | 2008-11-18 | Bacus Res Lab Inc | Method and apparatus for processing an image of a tissue sample microarray |
US7581027B2 (en) * | 2001-06-27 | 2009-08-25 | Ricoh Co., Ltd. | JPEG 2000 for efficent imaging in a client/server environment |
EP1415187A4 (en) | 2001-08-06 | 2008-03-26 | Bioview Ltd | Image focusing in fluorescent imaging |
US7383134B2 (en) | 2002-01-15 | 2008-06-03 | Piper James R | Method and/or system for analyzing biological samples using a computer system |
US20040202357A1 (en) * | 2003-04-11 | 2004-10-14 | Perz Cynthia B. | Silhouette image acquisition |
US7711174B2 (en) * | 2004-05-13 | 2010-05-04 | The Charles Stark Draper Laboratory, Inc. | Methods and systems for imaging cells |
-
2003
- 2003-04-11 US US10/413,493 patent/US20040202357A1/en not_active Abandoned
-
2004
- 2004-04-12 WO PCT/US2004/011334 patent/WO2004093004A2/en active Application Filing
-
2009
- 2009-10-01 US US12/587,107 patent/US8369591B2/en not_active Expired - Fee Related
Patent Citations (101)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3824393A (en) * | 1971-08-25 | 1974-07-16 | American Express Invest | System for differential particle counting |
US4011004A (en) * | 1975-02-06 | 1977-03-08 | Geometric Data Corporation | Pivotable stage for microscope system |
US4196265A (en) * | 1977-06-15 | 1980-04-01 | The Wistar Institute | Method of producing antibodies |
US4210419A (en) * | 1978-02-17 | 1980-07-01 | California Institute Of Technology | Automated quantitative muscle biopsy analysis system |
US4249825A (en) * | 1979-05-14 | 1981-02-10 | The Trustees Of Boston University | Method and apparatus suitable for ocular blood flow analysis |
US4338024A (en) * | 1980-05-02 | 1982-07-06 | International Remote Imaging Systems, Inc. | Flow analyzer and system for analysis of fluids with particles |
US4393466A (en) * | 1980-09-12 | 1983-07-12 | International Remote Imaging Systems | Method of analyzing particles in a dilute fluid sample |
US4612614A (en) * | 1980-09-12 | 1986-09-16 | International Remote Imaging Systems, Inc. | Method of analyzing particles in a fluid sample |
US4513438A (en) * | 1982-04-15 | 1985-04-23 | Coulter Electronics, Inc. | Automated microscopy system and method for locating and re-locating objects in an image |
US4673973A (en) * | 1985-02-04 | 1987-06-16 | National Biomedical Research Foundation | Split-image, multi-power microscopic image display system and method |
US4656594A (en) * | 1985-05-06 | 1987-04-07 | National Biomedical Research Foundation | Operator-interactive automated chromosome analysis system producing a karyotype |
US4741043A (en) * | 1985-11-04 | 1988-04-26 | Cell Analysis Systems, Inc. | Method of and an apparatus for image analyses of biological specimens |
US5008185A (en) * | 1985-11-04 | 1991-04-16 | Cell Analysis Systems, Inc. | Methods and apparatus for the quantitation of nuclear proteins |
US5018209A (en) * | 1985-11-04 | 1991-05-21 | Cell Analysis Systems, Inc. | Analysis method and apparatus for biological specimens |
US4741043B1 (en) * | 1985-11-04 | 1994-08-09 | Cell Analysis Systems Inc | Method of and apparatus for image analyses of biological specimens |
US5014198A (en) * | 1987-01-13 | 1991-05-07 | Kabushiki Kaisha Toshiba | Image information compression apparatus |
US5202931A (en) * | 1987-10-06 | 1993-04-13 | Cell Analysis Systems, Inc. | Methods and apparatus for the quantitation of nuclear protein |
US5085325A (en) * | 1988-03-08 | 1992-02-04 | Simco/Ramic Corporation | Color sorting system and method |
US5740270A (en) * | 1988-04-08 | 1998-04-14 | Neuromedical Systems, Inc. | Automated cytological specimen classification system and method |
US5287272B1 (en) * | 1988-04-08 | 1996-08-27 | Neuromedical Systems Inc | Automated cytological specimen classification system and method |
US5287272A (en) * | 1988-04-08 | 1994-02-15 | Neuromedical Systems, Inc. | Automated cytological specimen classification system and method |
US4991223A (en) * | 1988-06-30 | 1991-02-05 | American Innovision, Inc. | Apparatus and method for recognizing image features using color elements |
US5321545A (en) * | 1988-10-21 | 1994-06-14 | Biocom S.A. | Microscope stage for rapid and indexed analysis of filters and other media carrying multiple samples, and a method of analyzing such samples using said stage |
US4945220A (en) * | 1988-11-16 | 1990-07-31 | Prometrix Corporation | Autofocusing system for microscope having contrast detection means |
US5726009A (en) * | 1989-03-20 | 1998-03-10 | Anticancer, Inc. | Native-state method and system for determining viability and proliferative capacity of tissues in vitro |
US5016173A (en) * | 1989-04-13 | 1991-05-14 | Vanguard Imaging Ltd. | Apparatus and method for monitoring visually accessible surfaces of the body |
US5003165A (en) * | 1989-05-25 | 1991-03-26 | International Remote Imaging Systems, Inc. | Method and an apparatus for selecting the best focal position from a plurality of focal positions |
US5087965A (en) * | 1989-06-28 | 1992-02-11 | American Innovision, Inc. | Recognition of image colors using arbitrary shapes in color space |
US5123055A (en) * | 1989-08-10 | 1992-06-16 | International Remote Imaging Systems, Inc. | Method and an apparatus for differentiating a sample of biological cells |
US5233684A (en) * | 1990-06-26 | 1993-08-03 | Digital Equipment Corporation | Method and apparatus for mapping a digital color image from a first color space to a second color space |
US5526258A (en) * | 1990-10-10 | 1996-06-11 | Cell Analysis System, Inc. | Method and apparatus for automated analysis of biological specimens |
US5333207A (en) * | 1990-11-07 | 1994-07-26 | Neuromedical Systems, Inc. | Inspection apparatus and method with inspection auditing for images presented on a display |
US5231580A (en) * | 1991-04-01 | 1993-07-27 | The United States Of America As Represented By The Secretary Of The Department Of Health And Human Services | Automated method and apparatus for determining characteristics of nerve fibers |
US5481401A (en) * | 1991-05-16 | 1996-01-02 | Olympus Optical Co., Ltd. | Ultraviolet microscope |
US5428690A (en) * | 1991-09-23 | 1995-06-27 | Becton Dickinson And Company | Method and apparatus for automated assay of biological specimens |
US6238892B1 (en) * | 1991-10-25 | 2001-05-29 | N.V. Innogenetics S.A. | Monoclonal antibodies directed against the microtubule-associated protein tau |
US5799105A (en) * | 1992-03-06 | 1998-08-25 | Agri-Tech, Inc. | Method for calibrating a color sorting apparatus |
US5533628A (en) * | 1992-03-06 | 1996-07-09 | Agri Tech Incorporated | Method and apparatus for sorting objects by color including stable color transformation |
US5338924A (en) * | 1992-08-11 | 1994-08-16 | Lasa Industries, Inc. | Apparatus and method for automatic focusing of light using a fringe plate |
US5889881A (en) * | 1992-10-14 | 1999-03-30 | Oncometrics Imaging Corp. | Method and apparatus for automatically detecting malignancy-associated changes |
US5317140A (en) * | 1992-11-24 | 1994-05-31 | Dunthorn David I | Diffusion-assisted position location particularly for visual pen detection |
US5632402A (en) * | 1992-12-21 | 1997-05-27 | Graphic Packaging Corporation | Carton blank and method for forming it |
US5449622A (en) * | 1993-02-16 | 1995-09-12 | Yabe; Ryohei | Method and apparatus for analyzing stained particles |
US5602941A (en) * | 1993-05-21 | 1997-02-11 | Digital Equipment Corporation | Input modification system for multilevel dithering |
US5515172A (en) * | 1993-07-19 | 1996-05-07 | Xerox Corporation | Apparatus and method for enhanced color to color conversion |
US5432871A (en) * | 1993-08-04 | 1995-07-11 | Universal Systems & Technology, Inc. | Systems and methods for interactive image data acquisition and compression |
US5877161A (en) * | 1993-08-05 | 1999-03-02 | University Technologies International Inc. | Cyclin D1 negative regulatory activity |
US5409007A (en) * | 1993-11-26 | 1995-04-25 | General Electric Company | Filter to reduce speckle artifact in ultrasound imaging |
US5783814A (en) * | 1994-01-18 | 1998-07-21 | Ultrapointe Corporation | Method and apparatus for automatically focusing a microscope |
US5625705A (en) * | 1994-06-03 | 1997-04-29 | Neuromedical Systems, Inc. | Intensity texture based classification system and method |
US5499097A (en) * | 1994-09-19 | 1996-03-12 | Neopath, Inc. | Method and apparatus for checking automated optical system performance repeatability |
US5647025A (en) * | 1994-09-20 | 1997-07-08 | Neopath, Inc. | Automatic focusing of biomedical specimens apparatus |
US5627907A (en) * | 1994-12-01 | 1997-05-06 | University Of Pittsburgh | Computerized detection of masses and microcalcifications in digital mammograms |
US5621532A (en) * | 1994-12-08 | 1997-04-15 | Nikon Corporation | Laser scanning microscope utilizing detection of a far-field diffraction pattern with 2-dimensional detection |
US5625709A (en) * | 1994-12-23 | 1997-04-29 | International Remote Imaging Systems, Inc. | Method and apparatus for identifying characteristics of an object in a field of view |
US5619032A (en) * | 1995-01-18 | 1997-04-08 | International Remote Imaging Systems, Inc. | Method and apparatus for automatically selecting the best focal position from a plurality of focal positions for a focusing apparatus |
US5646677A (en) * | 1995-02-23 | 1997-07-08 | Motorola, Inc. | Method and apparatus for interactively viewing wide-angle images from terrestrial, space, and underwater viewpoints |
US5671288A (en) * | 1995-05-31 | 1997-09-23 | Neopath, Inc. | Method and apparatus for assessing slide and specimen preparation quality |
US6117985A (en) * | 1995-06-16 | 2000-09-12 | Stemcell Technologies Inc. | Antibody compositions for preparing enriched cell preparations |
US5735387A (en) * | 1995-07-14 | 1998-04-07 | Chiron Diagnostics Corporation | Specimen rack handling system |
US6040139A (en) * | 1995-09-19 | 2000-03-21 | Bova; G. Steven | Laser cell purification system |
US6553135B1 (en) * | 1995-11-30 | 2003-04-22 | Chromavision Medical Systems, Inc. | Method and apparatus for automated image analysis of biological specimens |
US6546123B1 (en) * | 1995-11-30 | 2003-04-08 | Chromavision Medical Systems, Inc. | Automated detection of objects in a biological sample |
US6215892B1 (en) * | 1995-11-30 | 2001-04-10 | Chromavision Medical Systems, Inc. | Method and apparatus for automated image analysis of biological specimens |
US5706093A (en) * | 1995-12-05 | 1998-01-06 | Olympus Optical Co., Ltd. | Color classification apparatus |
US6125194A (en) * | 1996-02-06 | 2000-09-26 | Caelum Research Corporation | Method and system for re-screening nodules in radiological images using multi-resolution processing, neural network, and image processing |
US5911003A (en) * | 1996-04-26 | 1999-06-08 | Pressco Technology Inc. | Color pattern evaluation system for randomly oriented articles |
US6031929A (en) * | 1996-07-18 | 2000-02-29 | University Of Pittsburgh | Image facsimile with real time image segmentation |
US6101265A (en) * | 1996-08-23 | 2000-08-08 | Bacus Research Laboratories, Inc. | Method and apparatus for acquiring and reconstructing magnified specimen images from a computer-controlled microscope |
US6226392B1 (en) * | 1996-08-23 | 2001-05-01 | Bacus Research Laboratories, Inc. | Method and apparatus for acquiring and reconstructing magnified specimen images from a computer-controlled microscope |
US5867598A (en) * | 1996-09-26 | 1999-02-02 | Xerox Corporation | Method and apparatus for processing of a JPEG compressed image |
US5911327A (en) * | 1996-10-02 | 1999-06-15 | Nippon Steel Corporation | Method of automatically discriminating and separating scraps containing copper from iron scraps |
US6225636B1 (en) * | 1996-10-25 | 2001-05-01 | Applied Imaging, Inc. | Multifluor-fluorescence in-situ hybridization (M-FISH) imaging techniques using multiple multiband filters with image registration |
US6718053B1 (en) * | 1996-11-27 | 2004-04-06 | Chromavision Medical Systems, Inc. | Method and apparatus for automated image analysis of biological specimens |
US6097838A (en) * | 1996-12-19 | 2000-08-01 | Xerox Corporation | Color correction of a compressed image |
US6404906B2 (en) * | 1997-03-03 | 2002-06-11 | Bacus Research Laboratories,Inc. | Method and apparatus for acquiring and reconstructing magnified specimen images from a computer-controlled microscope |
US6272235B1 (en) * | 1997-03-03 | 2001-08-07 | Bacus Research Laboratories, Inc. | Method and apparatus for creating a virtual microscope slide |
US6514722B2 (en) * | 1997-03-27 | 2003-02-04 | Oncosis | Method and apparatus for selectively targeting specific cells within a cell population |
US6169816B1 (en) * | 1997-05-14 | 2001-01-02 | Applied Imaging, Inc. | Identification of objects of interest using multiple illumination schemes and finding overlap of features in corresponding multiple images |
US6259807B1 (en) * | 1997-05-14 | 2001-07-10 | Applied Imaging Corp. | Identification of objects of interest using multiple illumination schemes and finding overlap of features in corresponding multiple images |
US6259826B1 (en) * | 1997-06-12 | 2001-07-10 | Hewlett-Packard Company | Image processing method and device |
US6072570A (en) * | 1997-07-24 | 2000-06-06 | Innotech | Image quality mapper for progressive eyeglasses |
US6058322A (en) * | 1997-07-25 | 2000-05-02 | Arch Development Corporation | Methods for improving the accuracy in differential diagnosis on radiologic examinations |
US5880473A (en) * | 1997-07-28 | 1999-03-09 | Applied Imaging, Inc. | Multifluor-fluorescence in-situ hybridization (M-FISH) imaging techniques using multiple multiband filters with image registration |
US6290907B1 (en) * | 1997-09-11 | 2001-09-18 | Hitachi, Ltd. | Sample handling system |
US6011595A (en) * | 1997-09-19 | 2000-01-04 | Eastman Kodak Company | Method for segmenting a digital image into a foreground region and a key color region |
US6122400A (en) * | 1997-09-26 | 2000-09-19 | Sarnoff Corporation | Compression encoder bit allocation utilizing colormetric-adaptive weighting as in flesh-tone weighting |
US5888742A (en) * | 1997-10-28 | 1999-03-30 | Incyte Pharmaceuticals, Inc. | Human phospholipid binding proteins |
US6275777B1 (en) * | 1997-10-28 | 2001-08-14 | Olympus Optical Co., Ltd. | Scanning cytometer |
US6374989B1 (en) * | 1997-11-14 | 2002-04-23 | Bayer Corporation | Conveyor system for clinical test apparatus |
US6281874B1 (en) * | 1998-08-27 | 2001-08-28 | International Business Machines Corporation | Method and system for downloading graphic images on the internet |
US6236031B1 (en) * | 1998-09-04 | 2001-05-22 | Sony Corporation | Optical head, recording and/or reproducing apparatus, and optical disc drive with an auxiliary focus servo system |
US6415058B2 (en) * | 1998-10-27 | 2002-07-02 | Hewlett-Packard Company | System for compression of digital images comprising low detail areas |
US6215894B1 (en) * | 1999-02-26 | 2001-04-10 | General Scanning, Incorporated | Automatic imaging and analysis of microarray biochips |
US6418236B1 (en) * | 1999-06-24 | 2002-07-09 | Chromavision Medical Systems, Inc. | Histological reconstruction and automated image analysis |
US6674896B1 (en) * | 1999-07-13 | 2004-01-06 | Chromavision Medical Systems, Inc. | Device for applying accurate color thresholds in real time |
US6404916B1 (en) * | 1999-08-04 | 2002-06-11 | Chromavision Medical Systems, Inc. | Method and apparatus for applying color thresholds in light microscopy |
US6518554B1 (en) * | 2000-05-24 | 2003-02-11 | Chromavision Medical Systems, Inc. | Reverse focusing methods and systems |
US20020067409A1 (en) * | 2000-12-06 | 2002-06-06 | Bio-View Ltd. | Data acquisition and display system and method |
US6697509B2 (en) * | 2001-10-04 | 2004-02-24 | Chromavision Medical Systems, Inc. | Method and apparatus for scoring the uptake of markers in cells |
US20030124589A1 (en) * | 2001-10-12 | 2003-07-03 | Vysis, Inc. | Imaging microarrays |
Cited By (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7016109B2 (en) * | 2002-09-10 | 2006-03-21 | Olympus Optical Co., Ltd. | Microscopic image capture apparatus and microscopic image capturing method |
US20040047033A1 (en) * | 2002-09-10 | 2004-03-11 | Olympus Optical Co., Ltd. | Microscopic image capture apparatus and microscopic image capturing method |
US7248403B2 (en) | 2002-09-10 | 2007-07-24 | Olympus Corporation | Microscopic image capture apparatus and microscopic image capturing method |
US20050190437A1 (en) * | 2002-09-10 | 2005-09-01 | Olympus Optical Co., Ltd | Microscopic image capture apparatus and microscopic image capturing method |
US8369591B2 (en) * | 2003-04-11 | 2013-02-05 | Carl Zeiss Microimaging Gmbh | Silhouette image acquisition |
US20100141682A1 (en) * | 2003-04-11 | 2010-06-10 | Carl Zeiss Microlmaging Ais, Inc. | Silhouette image acquisition |
US20040260402A1 (en) * | 2003-06-20 | 2004-12-23 | Baldini Steven E. | Method of manufacturing a socket portion of a prosthetic limb |
US8428887B2 (en) | 2003-09-08 | 2013-04-23 | Ventana Medical Systems, Inc. | Method for automated processing of digital images of tissue micro-arrays (TMA) |
US20050266395A1 (en) * | 2003-09-10 | 2005-12-01 | Bioimagene, Inc. | Method and system for morphology based mitosis identification and classification of digital images |
US8515683B2 (en) | 2003-09-10 | 2013-08-20 | Ventana Medical Systems, Inc. | Method and system for automated detection of immunohistochemical (IHC) patterns |
US20050136509A1 (en) * | 2003-09-10 | 2005-06-23 | Bioimagene, Inc. | Method and system for quantitatively analyzing biological samples |
US7941275B2 (en) | 2003-09-10 | 2011-05-10 | Ventana Medical Systems, Inc. | Method and system for automated detection of immunohistochemical (IHC) patterns |
US7979212B2 (en) | 2003-09-10 | 2011-07-12 | Ventana Medical Systems, Inc. | Method and system for morphology based mitosis identification and classification of digital images |
US20060149162A1 (en) * | 2004-11-29 | 2006-07-06 | Derek Daw | Graphical user interface for tissue biopsy system |
US10687733B2 (en) | 2004-11-29 | 2020-06-23 | Senorx, Inc. | Graphical user interface for tissue biopsy system |
US8795195B2 (en) * | 2004-11-29 | 2014-08-05 | Senorx, Inc. | Graphical user interface for tissue biopsy system |
US7873193B2 (en) | 2005-06-21 | 2011-01-18 | Carl Zeiss Microimaging Gmbh | Serial section analysis for computer-controlled microscopic imaging |
US20070280517A1 (en) * | 2005-06-21 | 2007-12-06 | De La Torre-Bueno Jose | Serial section analysis for computer-controlled microscopic imaging |
US20100265323A1 (en) * | 2005-08-02 | 2010-10-21 | Perz Cynthia B | System for and method of focusing in automated microscope systems |
US8000511B2 (en) | 2005-08-02 | 2011-08-16 | Carl Zeiss Microimaging Gmbh | System for and method of focusing in automated microscope systems |
US8116543B2 (en) * | 2005-08-02 | 2012-02-14 | Carl Zeiss Microimaging Gmbh | System for and method of intelligently directed segmentation analysis for automated microscope systems |
US20070031056A1 (en) * | 2005-08-02 | 2007-02-08 | Perz Cynthia B | System for and method of focusing in automated microscope systems |
US20080231973A1 (en) * | 2006-08-04 | 2008-09-25 | Ikonisys, Inc. | Mechanism for a microscope objective changer and filter changer |
US20170146782A1 (en) * | 2006-09-06 | 2017-05-25 | Leica Microsystems Cms Gmbh | Method and microscopy system for scanning a sample |
US10481374B2 (en) * | 2006-09-06 | 2019-11-19 | Leica Microsystems Cms Gmbh | Method and microscopy system for scanning a sample |
US20120082366A1 (en) * | 2009-06-03 | 2012-04-05 | Nec Corporation | Pathologic image diagnostic system, pathologic image diagnostic method, and pathologic image diagnostic program |
US8594412B2 (en) * | 2009-06-03 | 2013-11-26 | Nec Corporation | Pathologic image diagnostic system, pathologic image diagnostic method, and pathologic image diagnostic program |
US9014443B2 (en) * | 2010-01-04 | 2015-04-21 | Nec Corporation | Image diagnostic method, image diagnostic apparatus, and image diagnostic program |
US20130011028A1 (en) * | 2010-01-04 | 2013-01-10 | Nec Corporation | Image diagnostic method, image diagnostic apparatus, and image diagnostic program |
US20150116479A1 (en) * | 2012-05-28 | 2015-04-30 | Sony Corporation | Information processing device, information processing method, program, and microscope system |
EP2857884A4 (en) * | 2012-05-28 | 2016-01-20 | Sony Corp | Information processing device, information processing method, program and microscope system |
CN104335098A (en) * | 2012-05-28 | 2015-02-04 | 索尼公司 | Information processing device, information processing method, program and microscope system |
US11307399B2 (en) * | 2012-05-28 | 2022-04-19 | Sony Corporation | Information processing device, information processing method, program, and microscope system |
US20150124078A1 (en) * | 2012-07-04 | 2015-05-07 | Sony Corporation | Information processing apparatus, information processing method, program, and microscope system |
US10955655B2 (en) * | 2012-07-04 | 2021-03-23 | Sony Corporation | Stitching images based on presence of foreign matter |
US10314563B2 (en) | 2014-11-26 | 2019-06-11 | Devicor Medical Products, Inc. | Graphical user interface for biopsy device |
DE102016110988A1 (en) * | 2016-06-15 | 2017-12-21 | Sensovation Ag | Method for digitally recording a sample through a microscope |
CN113631982A (en) * | 2019-02-04 | 2021-11-09 | 脱其泰有限责任公司 | Automated micro-scanning system and method |
Also Published As
Publication number | Publication date |
---|---|
US8369591B2 (en) | 2013-02-05 |
WO2004093004A2 (en) | 2004-10-28 |
WO2004093004A3 (en) | 2006-01-19 |
US20100141682A1 (en) | 2010-06-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8369591B2 (en) | Silhouette image acquisition | |
US8116543B2 (en) | System for and method of intelligently directed segmentation analysis for automated microscope systems | |
US11927738B2 (en) | Computational microscopy based-system and method for automated imaging and analysis of pathology specimens | |
US8000511B2 (en) | System for and method of focusing in automated microscope systems | |
US7873193B2 (en) | Serial section analysis for computer-controlled microscopic imaging | |
US8582924B2 (en) | Data structure of an image storage and retrieval system | |
CN108982500B (en) | Intelligent auxiliary cervical fluid-based cytology reading method and system | |
US8350905B2 (en) | Microscope system, image generating method, and program for practicing the same | |
US8625930B2 (en) | Digital microscope slide scanning system and methods | |
US20170140528A1 (en) | Automated histological diagnosis of bacterial infection using image analysis | |
US20090196526A1 (en) | Microscopy system having automatic and interactive modes for forming a magnified mosaic image and associated method | |
US7627153B2 (en) | Repositioning inaccuracies in an automated imaging system | |
US20060038144A1 (en) | Method and apparatus for providing optimal images of a microscope specimen | |
US11828927B2 (en) | Accelerating digital microscopy scans using empty/dirty area detection | |
EP1631817B1 (en) | System for organizing multiple objects of interest in field of interest | |
JP2014228755A (en) | Microscope system, image generation method and program | |
US20110115896A1 (en) | High-speed and large-scale microscope imaging | |
US7653260B2 (en) | System and method of registering field of view | |
CN111527438A (en) | Shock rescanning system | |
CN112903675A (en) | Sample analyzer and cell image processing method for sample analyzer | |
CN113758856A (en) | Method for relocating cells on smear and cell image analysis device | |
Fan | Methods for rapid and high quality acquisition of whole slide images |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: CHROMAVISION MEDICAL SYSTEMS, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PERZ, CYNTHIA B.;TORRE-BUENO, JOSE DE LA;ELLIS, ROBERT T.;AND OTHERS;REEL/FRAME:014127/0163;SIGNING DATES FROM 20030527 TO 20030528 |
|
AS | Assignment |
Owner name: CLARIENT INC., CALIFORNIA Free format text: MERGER;ASSIGNOR:CHROMAVISION MEDICAL SYSTEMS, INC.;REEL/FRAME:017240/0641 Effective date: 20050315 |
|
AS | Assignment |
Owner name: CARL ZEISS MICROIMAGING AIS, INC., NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CLARIENT, INC.;REEL/FRAME:020072/0662 Effective date: 20071016 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |