WO2014164757A1 - Imaging blood cells - Google Patents
Imaging blood cells Download PDFInfo
- Publication number
- WO2014164757A1 WO2014164757A1 PCT/US2014/023401 US2014023401W WO2014164757A1 WO 2014164757 A1 WO2014164757 A1 WO 2014164757A1 US 2014023401 W US2014023401 W US 2014023401W WO 2014164757 A1 WO2014164757 A1 WO 2014164757A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- images
- wbc
- objects
- classification
- processor
- Prior art date
Links
- 238000003384 imaging method Methods 0.000 title claims abstract description 40
- 210000000601 blood cell Anatomy 0.000 title abstract description 7
- 210000000265 leukocyte Anatomy 0.000 claims abstract description 135
- 238000000034 method Methods 0.000 claims abstract description 63
- 239000000758 substrate Substances 0.000 claims abstract description 23
- 239000012472 biological sample Substances 0.000 claims abstract description 11
- 210000004027 cell Anatomy 0.000 claims description 66
- 210000003743 erythrocyte Anatomy 0.000 claims description 47
- 210000004698 lymphocyte Anatomy 0.000 claims description 39
- 210000000440 neutrophil Anatomy 0.000 claims description 21
- 239000000463 material Substances 0.000 claims description 16
- 210000001616 monocyte Anatomy 0.000 claims description 16
- 210000004940 nucleus Anatomy 0.000 claims description 14
- 210000003714 granulocyte Anatomy 0.000 claims description 13
- 230000001086 cytosolic effect Effects 0.000 claims description 12
- 230000003287 optical effect Effects 0.000 claims description 12
- 238000005286 illumination Methods 0.000 claims description 11
- 230000011218 segmentation Effects 0.000 claims description 10
- 239000013598 vector Substances 0.000 claims description 10
- 210000003855 cell nucleus Anatomy 0.000 claims description 9
- 238000004458 analytical method Methods 0.000 claims description 7
- 210000003979 eosinophil Anatomy 0.000 claims description 7
- 210000003651 basophil Anatomy 0.000 claims description 6
- 230000035800 maturation Effects 0.000 claims description 4
- 230000037361 pathway Effects 0.000 claims description 4
- 238000004590 computer program Methods 0.000 abstract description 8
- 238000012545 processing Methods 0.000 description 23
- 210000002459 blastocyst Anatomy 0.000 description 14
- 230000015654 memory Effects 0.000 description 12
- 239000003086 colorant Substances 0.000 description 8
- 210000000805 cytoplasm Anatomy 0.000 description 7
- 230000008569 process Effects 0.000 description 6
- 210000004369 blood Anatomy 0.000 description 5
- 239000008280 blood Substances 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 3
- 239000000470 constituent Substances 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 102000001554 Hemoglobins Human genes 0.000 description 2
- 108010054147 Hemoglobins Proteins 0.000 description 2
- 206010033546 Pallor Diseases 0.000 description 2
- 238000013459 approach Methods 0.000 description 2
- 238000004820 blood count Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000012512 characterization method Methods 0.000 description 2
- 239000000428 dust Substances 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 210000004206 promonocyte Anatomy 0.000 description 2
- 210000001995 reticulocyte Anatomy 0.000 description 2
- UIIMBOGNXHQVGW-UHFFFAOYSA-M Sodium bicarbonate Chemical compound [Na+].OC([O-])=O UIIMBOGNXHQVGW-UHFFFAOYSA-M 0.000 description 1
- 230000008649 adaptation response Effects 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 230000002776 aggregation Effects 0.000 description 1
- 238000004220 aggregation Methods 0.000 description 1
- 239000000356 contaminant Substances 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- -1 debris Substances 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 229910003460 diamond Inorganic materials 0.000 description 1
- 239000010432 diamond Substances 0.000 description 1
- 230000004069 differentiation Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 239000012634 fragment Substances 0.000 description 1
- 238000009499 grossing Methods 0.000 description 1
- 238000007654 immersion Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 210000002751 lymph Anatomy 0.000 description 1
- 230000000873 masking effect Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000000386 microscopy Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000000523 sample Substances 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B21/00—Microscopes
- G02B21/36—Microscopes arranged for photographic purposes or projection purposes or digital imaging or video purposes including associated control and data processing arrangements
- G02B21/365—Control or image processing arrangements for digital or video microscopes
- G02B21/367—Control or image processing arrangements for digital or video microscopes providing an output produced by processing a plurality of individual source images, e.g. image tiling, montage, composite images, depth sectioning, image comparison
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N15/00—Investigating characteristics of particles; Investigating permeability, pore-volume or surface-area of porous materials
- G01N15/10—Investigating individual particles
- G01N15/14—Optical investigation techniques, e.g. flow cytometry
- G01N15/1429—Signal processing
- G01N15/1433—Signal processing using image recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
- G06F18/24155—Bayesian classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/136—Segmentation; Edge detection involving thresholding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/693—Acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/695—Preprocessing, e.g. image segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/698—Matching; Classification
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N15/00—Investigating characteristics of particles; Investigating permeability, pore-volume or surface-area of porous materials
- G01N15/01—Investigating characteristics of particles; Investigating permeability, pore-volume or surface-area of porous materials specially adapted for biological cells, e.g. blood cells
- G01N2015/016—White blood cells
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10056—Microscopic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20016—Hierarchical, coarse-to-fine, multiscale or multiresolution image processing; Pyramid transform
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30024—Cell structures in vitro; Tissue sections in vitro
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30242—Counting objects in image
Definitions
- the invention relates to imaging blood cells, e.g., using electronic imaging systems.
- Imaging and scanning applications acquire images in an automated fashion. Interpreting such images may require identifying one or more objects within the images. While such identification may be performed manually, the process is time consuming and may be prone to errors.
- Biological specimens can be imaged at multiple individual wavelengths and a set of images can be acquired. Characteristics of the biological specimens can be determined by processing one or more of the images. For instance, processing of low-magnification images can include counting of red blood cells (RBCs), locating and counting blood constituents (e.g., white blood cells (WBCs) and nucleated red blood cells (nRBCs)), and counting platelets,. For instance, processing of high-magnification images can include characterizing features of RBCs, calculating a WBC differential, and classifying WBCs.
- RBCs red blood cells
- WBCs white blood cells
- nRBCs nucleated red blood cells
- a method for locating cells in a biological sample on a substrate includes acquiring, by an image acquisition device, a plurality of images of a first location on the substrate. Each image of the plurality of images is acquired at a different illumination wavelength. The method also includes identifying, by a processor, locations of cytoplasmic material in the plurality of images, and identifying, by a processor, locations of cell nuclei in the plurality of images. The method further includes identifying, by a processor, locations of cells in the plurality of images based on the locations of cytoplasmic material and the locations of cell nuclei. The method also includes assigning, by a processor, cell types to at least some of the cells in the plurality of images, and counting at least one cell type in the plurality of images.
- the disclosure features a method of determining a white blood cell (WBC) differential of a biological sample on a substrate.
- the method includes acquiring, by an image acquisition device, a plurality of images of a first location on the substrate, including acquiring each image of the plurality of images at a different illumination wavelength, and identifying, by a processor, locations of WBC nuclei in the plurality of images. Identifying the locations of WBC includes generating a first nuclear mask based on a first assumption about the WBCs in the plurality of images, generating a second nuclear mask based on a second assumption about the WBCs in the plurality of images, and intersecting the first nuclear mask and the second nuclear mask to generate a nuclear segmentation.
- the method also includes measuring one or more features of the WBCs based on the nuclear segmentation, and using the measured features to determine the WBC differential.
- the disclosure features a method of classifying white blood cells (WBCs) in a biological sample on a substrate.
- the method includes acquiring, by an image acquisition device, a plurality of images of a first location on the substrate, and classifying, by a processor, objects in the plurality of images into WBC classification groups.
- the method also includes identifying, by a processor, objects from at least some classification groups, as unclassified objects, and displaying, on a user interface, the unclassified objects and at least some of the classified objects.
- the disclosure features computer readable storage devices that are encoded with computer readable instructions that cause one or more processors implement the methods outlined above.
- the disclosure features an imaging system that includes an image acquisition device configured to acquire a plurality of images of a first location on a substrate. Each image of the plurality of images can be acquired at a different illumination wavelength.
- the systems also include one or more processors, connected to the image acquisition device, wherein the processor is configured to implement the methods outlined above.
- Implementations of the above aspects can include one or more of the following.
- the plurality of images can be acquired using a 10X objective lens.
- the at least one cell type can include at least one of white blood cells and nucleated red blood cells.
- Acquiring the plurality of images can include acquiring the plurality of images using a 50X objective lens.
- the first assumption can include an assumption that the WBCs are granulocytes.
- the second assumption can include an assumption that the WBCs are agranular.
- the features can include one or more of an area of a WBC, a shape of a WBC, a color of a WBC, and optical density of a WBC, and a texture of a WBC.
- At least one of a value for a nuclear complexity of neutrophils and a value for an atypicality of lymphocytes can be determined in the WBC classification groups. At least one of the value for the nuclear complexity and the value for the atypicality can be displayed on the user interface.
- Non-WBC objects can be removed from the images prior to classifying the objects.
- the non-WBC objects can include one or more of platelets, clumps, giant platelets, and micromegakaryocytes.
- the objects in the plurality of images can be classified using a multi-stage classifier. At least one stage of the multi-stage classifier can include a linear discriminant analysis (LDA) based classifier. At least another stage of the multi-stage classifier can include a Bayesian classifier.
- LDA linear discriminant analysis
- the WBC classification groups can be based on WBC maturation pathways.
- the unclassified objects can include one or more of immature cells, blasts, promonos, and prolymphs. Rare, damaged, or out-of- focus cells can be excluded from being classified into the WBC classification groups.
- FIG. 1 is a schematic diagram of an embodiment of an imaging system.
- FIG. 2 is a block diagram showing various modules of an imaging system.
- FIG. 3 is a flow chart of a process for locating and counting white blood cells (WBCs) and nucleated red blood cells.
- WBCs white blood cells
- FIG. 4 is a flow chart of a process for determining a WBC differential.
- FIG. 5 is a flow chart of a process for classifying WBCs.
- FIG. 6 is a flow chart of a process for classifying cells.
- FIG. 7 is an example classification scheme.
- FIG. 8 is a schematic diagram of a computing device.
- Biological specimens can be imaged at multiple individual wavelengths and a set of images can be acquired. Characteristics of the biological specimens can be determined by processing one or more of the images. For instance, processing of low-magnification images can include counting of red blood cells (RBCs), locating and counting blood constituents (e.g., white blood cells (WBCs) and nucleated red blood cells (nRBCs)), and counting platelets,. For instance, processing of high-magnification images can include characterizing features of RBCs, calculating a WBC differential, and classifying WBCs. Calculating a WBC differential can include, for example, counting a number of different types of WBCs in the blood. In some implementations, calculating a WBC differential can include calculating percentages of different types of WBC in the blood. Imaging Systems
- FIG. 1 shows one embodiment of an imaging system 100 that employs the fast auto- focusing methods described herein.
- Imaging system 100 includes imaging hardware 105 that is controlled by a computer 190.
- the computer 190 generally includes a central processing unit 192, a hard drive 194, and random access memory 196.
- a light source 110 illuminates a slide 130 comprising a biological specimen 135.
- the light source 110 includes individual light sources of different colors.
- the light source 110 may include blue, green, yellow, and red light emitting diodes (LEDs). Other types of colored light sources can also be used.
- the light source 110 can be used for low magnification imaging, high
- magnification imaging or both.
- the slide is mounted on a first motorized stage 120 capable of moving in a horizontal plane (parallel to the surface of the slide that is imaged) such that any part of the slide 130 can be positioned under an objective lens 140.
- a second motorized stage 150 moves the objective lens 140 up and down to facilitate focusing on the specimen 135 deposited on slide 130.
- the distance between the slide 130 and the objective lens 140 is referred to as the "focal distance.”
- a reduction in focal distance implies, in this example, moving the objective lens 140 vertically towards the slide 130.
- the focal distance can also be adjusted by moving the slide 130 (for example by moving the first motorized stage 120) vertically towards the objective lens 140.
- both the slide 130 and the objective lens 140 can move to facilitate adjusting the focal distance.
- the axes, with respect to which the first motorized stage 120 is moved in a horizontal plane, are typically referred to at the X and Y axes.
- the vertical axis along which the second motorized stage 150 moves the objective lens 140 is typically referred to as the Z axis.
- the three axes define a coordinate system that the system 100 utilizes to image any (x, y, z) point in space relative to the slide 130.
- FIG. 1 depicts an example of "bright field" microscopy where objects on the slide are visible because they absorb light and are therefore darker in the image produced by the camera.
- the imaging hardware 105 can include one or more additional lenses. Other microscopic modes such as fluorescence, dark-field, or phase contrast can also generate images to which the fast auto-focus methods described herein can be applied.
- the system 100 could acquire in- focus images simply by determining the proper z height corresponding to an in-focus image at three (J , y) locations and then fitting a plane to those (x, y, z) points. The plane equation would then provide an in-focus z height for any other (JC, y) location on the slide.
- the focal distance for a given location may not be accurately determined from a plane fit as described above due to irregularities in the surface of the slide 130 and/or the stage 120. Therefore, in general, the focal distance may need to be adjusted slightly for each image that is acquired at a different (x, y) location on the slide.
- camera 160 sends images to the computer 190 for processing. If the images are processed quickly enough, focal distance data from one image location may be used to adjust the focal distance at the next location for capturing an image. This allows the system to adjust to variations in focal distances corresponding to different locations and, in turn, produce more accurately focused images for display. For example, if one region of the slide 130 is slightly thicker than others and ten locations are imaged within that region, the change in thickness can be discovered after the first image is acquired and additional images taken at other locations within the region can be acquired at slightly adjusted focal distances to compensate for the change in thickness using the methods described herein.
- low magnification images can be acquired with a 10X objective, a 0.5X optical coupler, and a 2 megapixel CCD camera.
- high magnification images can be acquired with a 50X objective, a IX optical coupler, and a 0.3 megapixel, 200 frames per second camera.
- a set of black and white images (referred to herein as a "stack" of images) is acquired of the slide at an imaging location.
- Each image of the set acquired using a single wavelength of light (e.g., using each of the blue, green, yellow, and red LEDs of the light source 110).
- color images such as color JPEG images, can be created from the set of black and white images for display purposes.
- objective lenses are not corrected for oil immersion or for the presence of a coverslip.
- Each image of the set is background corrected and aligned with each other image of the set. If more than one image is taken with a certain color at a particular location, one of those images is selected (e.g., the image with the best focus).
- An on-the-fly estimate of the focal plane of the slide can be updated during image acquisition.
- the set of images is processed by one or more imaging modules to yield measurements that can be used to determine blood count, white blood cell (WBC) differential, and other characteristics of the biological specimen.
- WBC white blood
- Illuminating a specimen with different colors of light can result in different information being extracted from the acquired images.
- cells may appear differently under different colored illumination, thereby facilitating easier identification, classification, or differentiation.
- red blood cells absorb significant amounts of blue light due to the presence of hemoglobin, and cell nuclei stained with standard Romanowsky stains absorb yellow light.
- control system 200 image acquisition and processing is controlled by a control system 200.
- the control system 200 is implemented by a single computer 201.
- the control system 200 can be implemented by multiple computers.
- some modules of the control system 200 can be implemented by a first computer and other modules of the control system 200 can be implemented by a second computer.
- a hardware module 202 sends commands to imaging hardware (e.g., controlled by an imaging module 216), which in turn sends pulses to the camera 160 to trigger image acquisition.
- the acquired images are transferred to a computer (e.g., the computer 201).
- a camera module 204 packages the acquired images into stacks.
- a focus module 206 examines each stack of images and estimates at what height the stack is likely to have been acquired, e.g., based on the relative focus scores of the images of the stack. The focus module 206 reports these height estimates to a plane fit module 208. Before each location is imaged, the hardware module 202 asks the plane fit module 208 for a focal plane estimate at that location.
- This mechanism of updating the focal plane estimate with image acquisition can be used by the imaging system 100 or the imaging module 216 or both to adapt to irregularities on the surface of a slide. In some instances, there may be a lag in the adaptive response to slide irregularities, e.g., due to pipelining of the focal plane estimates.
- a registration module 210 and a pick registration frames module 212 can be used to match coordinate systems between low magnification images and high magnification images. Thus, for instance, any object found during low magnification imaging can be precisely located and imaged during high magnification imaging.
- An alignment module 214 provides a capability to align images of a stack.
- the raw images of a stack do not align with each other, for instance, due to slide movement, slightly off-center LED dies, optical effects of the lens, or other factors.
- images acquired under different illumination colors may each have a slightly different magnification if the lens is not perfectly color-corrected.
- the scale of the four colors may be the same but the images may still be misaligned.
- an algorithm can be used that selects a set of high-contrast tie points (e.g., cells) in a target image, locates the same tie points (e.g., cells) in a second image, and transforms the second image to align the second image with the target image.
- an algorithm can try all offsets within a limited range of pixels and select the offset for each color that maximizes the R correlation with the image of a target color (e.g., with a green image).
- Images, results of analysis of the images, or both can be displayed on a user interface, such as a graphical user interface (GUI) 218. Images, results of analysis of the images, or both can be logged by a logging module 220 and stored in a file, database 222, or other storage location.
- GUI graphical user interface
- a calibration can be performed when the imaging system 100 is initialized, e.g., with no slide in place.
- Calibration can include blank image acquisition and initial shutter adjustment.
- a focus search can be performed to determine a tilt of the slide, e.g., using a fast adjustment algorithm that takes only a fraction of a second.
- processing of low-magnification images can include counting of red blood cells (RBCs), locating and counting blood constituents (e.g., white blood cells (WBCs) and nucleated red blood cells (nRBCs)), counting platelets, and other types of processing.
- processing of high-magnification images can include characterizing features of RBCs, calculating a WBC differential, classifying WBCs, and other types of processing.
- image processing is executed by an image processing module 250 that is implemented by the computer 201 implementing the control system 200.
- the image processing module 250 is implemented by a different computer that does not implement any module of the control system 200.
- low magnification images can be processed to count RBCs.
- RBCs can have various sizes (e.g., RBCs can be large, small, or fragmented). RBCs can also have various appearances (e.g., RBCs can have large or small pallor or unusual shapes).
- a filter (referred to herein as a "spot filter") is applied that transforms each RBC, regardless of its size or pallor, into a smaller, dark spot.
- the transformed spots can be easier to count than images of original, varied RBCs.
- the spot filter can be applied only to the blue image of a stack of images, because white cells are nearly invisible in blue images.
- the filtered images can be processed to count the RBCs, including masking dust, debris, and other contaminants; smoothing the image; determining an average area of a RBC, counting the spots in the image; rejecting artifacts (e.g., stain debris); and adjusting the count based on the "dust area.”
- low magnification images can be processed to locate WBCs and nPvBCs.
- WBCs can be located during low magnification imaging, and in some examples can be revisited during high magnification imaging.
- a low magnification image is processed by locating blue spots of approximately the right size to be a nucleus or nuclear fragment (300). Regions of the image that are the right color to be cytoplasm are located (302).
- a distance transform is performed on the nucleus images, constrained to within the cytoplasm boundaries (304).
- a watershed transform with a size limit is performed (306) to divide touching cytoplasms.
- Cell types can then be assigned (308), e.g., using a ten part classifier that identifies five normal WBC types, platelets, nRBCs, double WBCs, damaged WBCs, and artifacts (e.g., stain debris).
- a cell count is performed (310) and locations of interest for high magnification can be selected (312).
- low magnification images can be processed to count platelets.
- Platelets can be located and counted using the filter counting approach described above for RBCs, but applying the spot filter to the yellow and blue images and skipping areas that are under objects already identified as WBCs.
- a classifier can be used to differentiate platelets from artifacts to determine the final count of platelets.
- High magnification imaging provides the ability to examine cells more closely than in low magnification imaging.
- a high magnification of a WBC can be used to classify the WBC.
- a high magnification image of RBCs or platelets can be used to characterize the RBCs or platelets.
- high magnification images can be processed to characterize features of RBCs, such as mean cell volume (MCA 7 ), hemoglobin content (MCH), reticulocyte percent, or other features, such as other elements of a complete blood count (CBC).
- MCA 7 mean cell volume
- MH hemoglobin content
- CBC complete blood count
- a large number of RBCs can be imaged under high magnification (e.g., at least about 10,000 RBCs).
- high magnification images of single, non-overlapping RBCs can be selected for
- MCV and MCH can be calculated from a weighted sum of nine features: the area of a RBC, an integrated optical density of the RBC for each of the four imaging colors, and a volume measurement of the RBC for each of the four imaging colors.
- the reticulocyte percentage can be determined by analyzing the level of blue content within each of the imaged RBCs.
- Other elements of a CBC can be determined from the RBC count
- the MCH (determined from low magnification imaging, e.g., as described above), the MCH, and the MCV of a sample.
- Additional features of high magnification images of RBCs can be used to determine the potential presence of inclusions in the RBCs, to assess RBC shapes, or to perform other characterizations of the RBCs.
- Specific RBCs with potential inclusions, unusual shapes, or both, can be tagged for presentation to a user. For instance, RBCs with potential inclusions can be displayed in an RBC gallery. For instance, RBCs with unusual shapes can be assigned values corresponding to their shapes, allowing the RBCs in the gallery to be sorted by shape.
- high magnification images can be processed to determine a WBC differential.
- An accurate WBC differential can be determined based on segmentation for WBC nuclei and for the cytoplasm.
- WBC nuclei generally stain the same color or similar colors and thus can be relatively easy to recognize.
- Cytoplasm can stain many different colors, and in some cases can be almost transparent.
- WBCs often aggregate; to obtain an accurate WBC differential, the WBCs in an aggregate can be counted individually.
- the cytoplasmic material is located (400). Aggregated WBCs are separated (402). The nuclei of the WBCs are segmented (404). To account for any variations in the stained color of the WBC nuclei, an adaptive threshold can be chosen (406) within each cytoplasm mask based on the yellow, green, and blue images. Two nuclear masks are created (410), one mask of the nucleus segmented as if it were a granulocyte and the other mask of the nucleus segmented as if it were agranular. The two nuclear masks are intersected (412) to produce a nuclear segmentation that is well suited to both granulocytes and agranular nuclei.
- Features are measured on the intersected cell masks (414). Some features can be measured only on the cytoplasm or only on the nucleus region of a WBC; other features can be measured on the entire cell. Area and shape features can be determined based upon just the nuclear and cytoplasmic masks. Color features can be determined based on the grey levels of individual pixels within the nuclear or cytoplasmic mask in different color images. Optical density and texture features can be determined based on the four different color images, within the nuclear or cytoplasmic masks. The measured features can be used to determine the WBC differential (416).
- high magnification images can be processed to classify WBCs.
- a multi-stage "tree" classifier can be used.
- the classifier can have mostly linear discriminants (LDA).
- LDA linear discriminants
- the classifier can also include a few non-LDA nodes, e.g., where the non-LDA nodes can be helpful for separating populations without a normal distribution.
- at least one of the non-LDA nodes can implement a Bayesian
- the classifier sorts out any non-WBC objects in the images that may be similar to WBCs, such as clumps of platelets, giant platelets, micromegakaryocytes, or other non-WBC objects (500).
- the remaining objects are sorted into a rough five-way classification (502), e.g., along the biological lines of WBC maturation pathways. Any damaged or out-of- focus cells from the fringes of each sorted group are removed from the classification (504).
- Each of the five groups is processed to extract potential unclassified cells, such as immature cells, blasts, promonos, prolymphs, or other unclassified cells, and to correct mistakes in the classification of the normal types (506).
- the classification scheme can be designed to minimize the number of false positive unclassified cells. For instance, each of the groups of unclassified cells (e.g., immature granulocytes, blasts, promonos, prolymphs) that were extracted is processed by a secondary classifier (508) to check for false positives. The cells that remain unclassified are grouped together into a single group (510).
- a secondary classifier e.g., immature granulocytes, blasts, promonos, prolymphs
- a value indicating the "nuclear complexity" of neutrophils can be determined.
- a value indicating the "atypicality” of lymphocytes can also be determined (512).
- the results of the classification can be displayed on a viewing station (514), such as a user interface of a computing device.
- the group of unclassified cells can be displayed on the top line of a WBC tab on the user interface.
- the unclassified cells can be displayed, e.g., on a front page of a WBC gallery, along with the equivalent of a hundred cell differential by choosing hundred cells at random from the normal WBC classification categories. Additional WBCs can be displayed responsive to instructions from a user.
- the nuclear complexity and atypicality values can also be displayed or made available to the user.
- a user can classify neutrophils as "banded” or “segmented” based on their nuclear complexity value and can sort neutrophils as banded or segmented based on this classification.
- a user can also classify lymphocytes as "atypical” based on their atypicality value and can sort lymphocytes as typical or atypical based on this classification.
- a multi-stage classifier can be used to classify cells, such as white blood cells (WBCs), and other objects, according to one or more features extracted from images of the biological sample.
- the multi-stage classifier can perform multiple stages of classification.
- different types of classifiers are used at the different stages. For example, one stage of the classifier can use a linear discriminant analysis (LDA) based classifier and another stage can use a non-LDA classifier, such as a Bayesian classifier.
- LDA linear discriminant analysis
- the multi-stage classifier can have a complex topology that can combine topologies from two or more other classifiers, such as tree classifiers and ladder classifiers, to enable efficient, accurate classification of WBCs, or blood cells in general.
- the classifier may classify objects into cell types, such as lymphocytes, monocytes, neutrophils, basophils, and eosinophils.
- the classifier may classify objects into specific other types of objects, such as various specific types of immature cells, or may classify these types of objects into a general category (e.g., an unclassified category).
- Each stage of the multi-stage classifier can classify objects into a plurality of classifications.
- the classification of an object at a particular stage may confirm the classification of the object from an earlier stage (for instance, a lymphocyte may remain classified as a lymphocyte).
- the classification of an object at a particular stage may reclassify the object into another category into which the object had not been classified during an earlier stage (for instance, in a first stage, the object may have been classified as a lymphocyte rather than a neutrophil; in a second stage, the object may be reclassified as a neutrophil).
- the classification of an object at a particular stage may classify the object into a category that was not available at an earlier stage (for instance, a lymphocyte may be classified as a blastocyst, which was not available for classification in earlier stages).
- Objects can be classified based on feature vectors representing one or more features of the objects, such as cell area, cell shape, cell color, cell optical density, cell texture, and other features of the cells.
- the feature vector used for classification can be substantially same for each stage of classification.
- a different feature vector can be used for some or all stages of classification. For instance, feature vectors for later classification stages may have more features than feature vectors for earlier classification stages, enabling more accurate classification, or classification into a different set of classes.
- one or more images of a substrate on which the biological sample is disposed are acquired (600).
- a plurality of features of the objects is extracted from the one or more images (602), e.g., using one or more of the techniques described above.
- a first classification is performed (604) and a second classification is performed (606). Each classification is based on a corresponding feature vector composed of one or more of the extracted features of the respective object. If the results of the second classification are the same as the results of the first classification (608), the classification of the object is maintained (610). If the results of the second classification differ from the results of the first classification (608), the classification of the object is changed (612). For instance, changing a previous classification of the object can include reclassifying the object into a category that was available during a previous classification. Changing a previous classification can include classifying the object into a category that was not available during a previous classification.
- an example topology of a multi-stage classifier 700 is depicted.
- the classifier depicted in this example classifies objects in multiple stages according to an LDA classification (depicted by circular nodes such as node 702) and non-LDA (e.g., Bayesian) classification (depicted by diamond nodes such as node 748).
- LDA linear discriminant analysis
- non-LDA e.g., Bayesian
- f-platelets (referred to as “fplate”) are removed from the set of "all” objects for classification.
- s- platelets s- platelets
- the f-platelets can be classified into platelets (referred to as "pit"), out-of-focus (“oof) material, damaged (“dam") material, or junk (node 706).
- the s-platelets can be reclassified as "all” or classified into platelets, oof material, damaged material, or junk (node 708).
- the objects classified as platelets at nodes 706, 708 are further classified at nodes 710, 712.
- the objects reclassified as "all” are returned to the set of "all” objects for further classification.
- the set of "all” objects can be further classified at node 714 into five cell types: lymphocytes ("lymph”), monocytes ("mono"), neutrophils ("neut”), basophils (“baso”), or eosinophils "(eo”).
- lymphocytes lymphocytes
- monocytes monocytes
- neutrophils neutrophils
- basophils basophils
- eosinophils eos
- the objects in each cell type are further classified. Some classifications can result in the classification of an object being maintained (e.g., a lymphocyte may remain classified as a lymphocyte). Some classifications can result in an object being reclassified as a different cell type (e.g., a lymphocyte may be reclassified as a neutrophil). Some classifications can result in an object being classified into a new classification (e.g., a lymphocyte may be classified as a blastocyst). In the example shown, the following classifications can occur:
- Lymphocytes (node 726) can remain classified as lymphocytes, can be
- neutrophils reclassified as neutrophils, or can be classified as atypical lymphocytes ("alymph”), blastocysts ("blast”), immature granulocytes (“ig”), or rare cells (“rare).
- alymph atypical lymphocytes
- blast blastocysts
- ig immature granulocytes
- rare cells rare cells
- Monocytes can remain classified as monocytes, can be reclassified as neutrophils, or can be classified as atypical lymphocytes, blastocysts, immature granulocytes, or rare cells.
- Neutrophils can remain classified as neutrophils, can be reclassified as lymphocytes or eosinophils, or can be classified as atypical lymphocytes or immature granulocytes. Node 730 classifies objects that were classified as neutrophils by nodes 720, 726, and 728.
- Basophils can remain classified as basophils, can be reclassified as neutrophils, or can be classified as blastocysts or immature granulocytes.
- Eosinophils (node 734) can remain classified as eosinophils or can be
- Rare cells can remain classified as rare cells or can be reclassified as lymphocytes, monocytes, alymphocytes, or blastocysts.
- Node 736 classifies objects that were classified as rare cells by nodes 726 and 728.
- Immature granulocytes can remain classified as immature granulocytes or can be reclassified as neutrophils, lymphocytes, or monocytes.
- Node 738 classifies objects that were classified as immature granulocytes by nodes 726, 728, 730, 732.
- Blastocysts can remain classified as blastocysts or can be reclassified as lymphocytes, monocytes, or atypical lymphocytes.
- Node 740 classifies objects that were classified as blastocysts by nodes 726, 728, 732.
- Atypical lymphocytes (node 742) can remain classified as atypical lymphocytes or can be classified as atypical monocytes ("amono").
- Node 742 classifies objects that were classified as atypical lymphocytes by nodes 726, 728, 730, 736.
- Atypical lymphocytes undergo a further classification and can remain classified as atypical lymphocytes, can be reclassified as
- lymphocytes or can be designated as unclassified ("unclass").
- Node 744 classifies objects that were classified as atypical lymphocytes by node 742.
- Atypical monocytes node 746) can be reclassified as monocytes, atypical lymphocytes, or blastocysts.
- Node 746 classifies objects that were classified as atypical monocytes by node 742.
- Lymphocytes (node 748) can be classified by a non-LDA classifier, such as a Bayesian classifier, as lymphocytes or prolymphocytes ("prolymph").
- Node 748 classifies objects that were classified as lymphocytes by nodes 726, 730, 736, 738, 740, 744.
- Blastocysts can be classified by a non-LDA classifier as blastocysts or junk. Node 750 classifies objects that were classified as blastocysts by nodes 736, 740, 746.
- Monocytes can be classified by a non-LDA classifier as monocytes or promonocytes ("promono").
- Node 752 classifies objects that were classified as monocytes by nodes 728, 736, 738, 740, 746.
- Node 754 classifies objects as either unclassified or junk. Node 754 classifies objects that were classified as prolymphocytes from node 748, rare cells from node 736, blastocysts from node 750, promonocytes from node 752, immature granulocytes from node 738, and unclassified objects from node 744.
- the classifier results in the classification of objects as follows:
- This multi-stage classifier can provide a scheme that enables efficient and accurate classification of objects.
- FIG. 6 is a schematic diagram of a computer system 600 that can be used to control the operations described in association with any of the computer-implemented methods described herein, according to one implementation.
- the system 600 includes a processor 610, a memory 620, a storage device 630, and an input/output device 640. Each of the components 610, 620, 630, and 640 are interconnected using a system bus 650.
- the processor 610 is capable of processing instructions for execution within the system 600. In one implementation, the processor 610 is a single-threaded processor. In another implementation, the processor 610 is a multi-threaded processor.
- the processor 610 is capable of processing instructions stored in the memory 620 or on the storage device 630 to display graphical information for a user interface on the input/output device 640.
- the memory 620 stores information within the system 600. In some embodiments,
- the memory 620 is a computer-readable medium.
- the memory 620 can include volatile memory and/or non- volatile memory.
- the storage device 630 is capable of providing mass storage for the system 600.
- the storage device 630 can include any non-transitory tangible media configured to store computer readable instructions.
- the storage device 630 is a computer-readable medium.
- the storage device 630 may be a floppy disk device, a hard disk device, an optical disk device, or a tape device.
- the input/output device 640 provides input/output operations for the system 600.
- the input/output device 640 includes a keyboard and/or pointing device.
- the input/output device 640 includes a display unit for displaying graphical user interfaces.
- the features described can be implemented in digital electronic circuitry, or in computer hardware, firmware, or in combinations of them.
- the features can be implemented in digital electronic circuitry, or in computer hardware, firmware, or in combinations of them.
- the features can be implemented in digital electronic circuitry, or in computer hardware, firmware, or in combinations of them.
- the features can be implemented in digital electronic circuitry, or in computer hardware, firmware, or in combinations of them.
- the features can be implemented in digital electronic circuitry, or in computer hardware, firmware, or in combinations of them.
- a computer program includes a set of instructions that can be used, directly or indirectly, in a computer to perform a certain activity or bring about a certain result.
- a computer program can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a standalone program or as a module, component, subroutine, or other unit suitable for use in a computing environment.
- a publish/subscribe messaging pattern can be used in implementing the methods and systems described herein.
- the system includes several hardware and software modules that communicate only via a messaging module. Each module can be configured to perform a specific function.
- the system can include one or more of a hardware module, a camera module, and a focus module.
- the hardware module can send commands to the imaging hardware implementing the fast auto-focus, which in turn triggers a camera to acquire images.
- a camera module can receive images from the camera and determine camera parameters such as shutter time or focus. Images can also be buffered in the computer's memory before being processed by the camera module. When performing the initial search for the tilt of the slide, the camera module can also send a message interrupting the hardware module when it has seen enough images to determine the proper shutter time or focus.
- the system can also include a focus module that can be implemented as software, hardware or a combination of software and hardware.
- the focus module examines all the frames in a stack and estimates how far the stack is from the ideal or ideal focal distance.
- the focus module can also be responsible for assigning a focus score to each frame in a stack of images.
- Suitable processors for the execution of a program of instructions include, by way of example, both general and special purpose microprocessors, and the sole processor or one of multiple processors of any kind of computer.
- a processor will receive instructions and data from a read-only memory or a random access memory or both.
- Computers include a processor for executing instructions and one or more memories for storing instructions and data.
- a computer will also include, or be operatively coupled to communicate with, one or more mass storage devices for storing data files; such devices include magnetic disks, such as internal hard disks and removable disks; magneto- optical disks; and optical disks.
- Storage devices suitable for tangibly embodying computer program instructions and data include all forms of non-volatile memory, including by way of example semiconductor memory devices, such as EPROM, EEPROM, and flash memory devices; magnetic disks such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks.
- the processor and the memory can be supplemented by, or incorporated in, ASICs (application-specific integrated circuits).
- the features can be implemented on a computer having a display device such as a CRT (cathode ray tube) or LCD (liquid crystal display) monitor for displaying information to the user and a keyboard and a pointing device such as a mouse or a trackball by which the user can provide input to the computer.
- a display device such as a CRT (cathode ray tube) or LCD (liquid crystal display) monitor for displaying information to the user and a keyboard and a pointing device such as a mouse or a trackball by which the user can provide input to the computer.
- the computer can have no keyboard, mouse, or monitor attached and can be controlled remotely by another computer
- the features can be implemented in a computer system that includes a back-end component, such as a data server, or that includes a middleware component, such as an application server or an Internet server, or that includes a front-end component, such as a client computer having a graphical user interface or an Internet browser, or any combination of them.
- the components of the system can be connected by any form or medium of digital data communication such as a communication network. Examples of communication networks include, e.g., a LAN, a WAN, and the computers and networks forming the Internet.
- the computer system can include clients and servers.
- a client and server are generally remote from each other and typically interact through a network, such as the described one.
- the relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
- the processor 610 carries out instructions related to a computer program.
- the processor 610 can include hardware such as logic gates, adders, multipliers and counters.
- the processor 610 can further include a separate arithmetic logic unit (ALU) that performs arithmetic and logical operations.
- ALU arithmetic logic unit
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Chemical & Material Sciences (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Analytical Chemistry (AREA)
- Optics & Photonics (AREA)
- Pathology (AREA)
- Data Mining & Analysis (AREA)
- Immunology (AREA)
- Biochemistry (AREA)
- Dispersion Chemistry (AREA)
- Signal Processing (AREA)
- Radiology & Medical Imaging (AREA)
- Medical Informatics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Quality & Reliability (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Artificial Intelligence (AREA)
- Probability & Statistics with Applications (AREA)
- Investigating Or Analysing Biological Materials (AREA)
- Investigating Or Analysing Materials By Optical Means (AREA)
Abstract
Description
Claims
Priority Applications (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2016501227A JP6343660B2 (en) | 2013-03-11 | 2014-03-11 | Imaging of blood cells |
AU2014249011A AU2014249011A1 (en) | 2013-03-11 | 2014-03-11 | Imaging blood cells |
CA2904401A CA2904401A1 (en) | 2013-03-11 | 2014-03-11 | Imaging blood cells |
US14/775,381 US9690976B2 (en) | 2013-03-11 | 2014-03-11 | Imaging blood cells |
EP14719553.1A EP2972550B1 (en) | 2013-03-11 | 2014-03-11 | Method for classifying white blood cells |
CN201480026258.7A CN105190400A (en) | 2013-03-11 | 2014-03-11 | Imaging blood cells |
HK16106816.1A HK1218784A1 (en) | 2013-03-11 | 2016-06-14 | Imaging blood cells |
US15/633,315 US10083342B2 (en) | 2013-03-11 | 2017-06-26 | Imaging blood cells |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361776740P | 2013-03-11 | 2013-03-11 | |
US61/776,740 | 2013-03-11 | ||
US201361794946P | 2013-03-15 | 2013-03-15 | |
US61/794,946 | 2013-03-15 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/775,381 A-371-Of-International US9690976B2 (en) | 2013-03-11 | 2014-03-11 | Imaging blood cells |
US15/633,315 Continuation US10083342B2 (en) | 2013-03-11 | 2017-06-26 | Imaging blood cells |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2014164757A1 true WO2014164757A1 (en) | 2014-10-09 |
WO2014164757A8 WO2014164757A8 (en) | 2015-09-24 |
Family
ID=50555249
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2014/023401 WO2014164757A1 (en) | 2013-03-11 | 2014-03-11 | Imaging blood cells |
Country Status (8)
Country | Link |
---|---|
US (2) | US9690976B2 (en) |
EP (1) | EP2972550B1 (en) |
JP (2) | JP6343660B2 (en) |
CN (1) | CN105190400A (en) |
AU (1) | AU2014249011A1 (en) |
CA (1) | CA2904401A1 (en) |
HK (1) | HK1218784A1 (en) |
WO (1) | WO2014164757A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017223412A1 (en) * | 2016-06-24 | 2017-12-28 | Beckman Coulter, Inc. | Image atlas systems and methods |
EP4358034A1 (en) | 2022-10-18 | 2024-04-24 | F. Hoffmann-La Roche AG | Clinical support system and associated computer-implemented methods |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2972550B1 (en) | 2013-03-11 | 2024-09-25 | Roche Diagnostics Hematology, Inc. | Method for classifying white blood cells |
US10706260B2 (en) * | 2014-06-16 | 2020-07-07 | Siemens Healthcare Diagnostics Inc. | Analyzing digital holographic microscopy data for hematology applications |
US12005441B1 (en) | 2014-11-26 | 2024-06-11 | Medica Corporation | Automated microscopic cell analysis |
US11478789B2 (en) | 2014-11-26 | 2022-10-25 | Medica Corporation | Automated microscopic cell analysis |
US10625259B1 (en) | 2014-11-26 | 2020-04-21 | Medica Corporation | Automated microscopic cell analysis |
US20170328924A1 (en) | 2014-11-26 | 2017-11-16 | Ronald Jones | Automated microscopic cell analysis |
CN110914666A (en) * | 2017-05-19 | 2020-03-24 | 兴盛生物科技股份有限公司 | System and method for counting cells |
CN110892247B (en) | 2017-08-17 | 2023-08-25 | 雅培医护站股份有限公司 | Apparatus, systems, and methods for performing optical and electrochemical assays |
US11047845B1 (en) | 2017-11-15 | 2021-06-29 | Medica Corporation | Control material and methods for cell analyzers |
US11443426B2 (en) * | 2018-01-18 | 2022-09-13 | Nantomics, Llc | Real-time whole slide pathology image cell counting |
SG11202011180UA (en) | 2018-05-31 | 2020-12-30 | Berkeley Lights Inc | Automated detection and characterization of micro-objects in microfluidic devices |
JP7141285B2 (en) * | 2018-09-11 | 2022-09-22 | 花王株式会社 | Pants-type disposable diaper |
JP2021040590A (en) | 2019-09-13 | 2021-03-18 | シスメックス株式会社 | Cell image analysis method, cell image analysis apparatus, program, and cell image analysis system |
JP7530709B2 (en) * | 2019-10-11 | 2024-08-08 | 株式会社島津製作所 | Cell image analysis method and cell analysis device |
WO2021242983A1 (en) * | 2020-05-28 | 2021-12-02 | Leuko Labs, Inc. | A method to detect white blood cells and/or white blood cell subtypes form non-invasive capillary videos |
JPWO2022059300A1 (en) * | 2020-09-15 | 2022-03-24 | ||
CN112924452A (en) * | 2021-01-29 | 2021-06-08 | 西安博锐轶信息科技有限公司 | Blood examination auxiliary system |
WO2024138139A1 (en) * | 2022-12-22 | 2024-06-27 | Beckman Coulter, Inc. | Population based cell classification |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1998052018A1 (en) * | 1997-05-14 | 1998-11-19 | Applied Imaging Corporation | Identification of objects by means of multiple imaging |
US20120263369A1 (en) * | 2011-04-14 | 2012-10-18 | Abbott Point Of Care, Inc. | Method and apparatus for compressing imaging data of whole blood sample analyses |
US20120262703A1 (en) * | 2011-04-15 | 2012-10-18 | Constitution Medical, Inc. | Measuring Volume and Constituents of Cells |
Family Cites Families (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS5829872B2 (en) * | 1976-10-25 | 1983-06-25 | 株式会社日立製作所 | White blood cell identification method |
JPS5819064B2 (en) * | 1976-10-25 | 1983-04-15 | 株式会社日立製作所 | Leukocyte classification method |
JPH0652263B2 (en) * | 1985-12-10 | 1994-07-06 | 株式会社日立製作所 | Cell analyzer |
US6026174A (en) * | 1992-10-14 | 2000-02-15 | Accumed International, Inc. | System and method for automatically detecting malignant cells and cells having malignancy-associated changes |
JPH0720124A (en) * | 1993-07-07 | 1995-01-24 | Omron Corp | Blood cell analyzer |
US5978497A (en) * | 1994-09-20 | 1999-11-02 | Neopath, Inc. | Apparatus for the identification of free-lying cells |
JPH08145988A (en) * | 1994-11-22 | 1996-06-07 | Chiyouki Hayashi | Method for identifying lymphocyte of different type |
ES2174428T3 (en) * | 1997-02-28 | 2002-11-01 | Cytec Tech Corp | Hardened aqueous compositions based on ISOCIANATE WITH LOW CONTENT IN VOLATILE ORGANIC COMPOUNDS. |
US8885913B2 (en) * | 1999-01-25 | 2014-11-11 | Amnis Corporation | Detection of circulating tumor cells using imaging flow cytometry |
US7450229B2 (en) * | 1999-01-25 | 2008-11-11 | Amnis Corporation | Methods for analyzing inter-cellular phenomena |
US6228652B1 (en) * | 1999-02-16 | 2001-05-08 | Coulter International Corp. | Method and apparatus for analyzing cells in a whole blood sample |
WO2004053074A2 (en) * | 2002-12-06 | 2004-06-24 | Science And Technology Corporation @ Unm | Outcome prediction and risk classification in childhood leukemia |
US7327880B2 (en) * | 2004-03-12 | 2008-02-05 | Siemens Medical Solutions Usa, Inc. | Local watershed operators for image segmentation |
JP4982385B2 (en) * | 2005-02-01 | 2012-07-25 | アムニス コーポレイション | Analysis of blood and cells using an imaging flow cytometer |
CN100470587C (en) * | 2007-01-26 | 2009-03-18 | 清华大学 | Method for segmenting abdominal organ in medical image |
NZ562237A (en) * | 2007-10-05 | 2011-02-25 | Pacific Edge Biotechnology Ltd | Proliferation signature and prognosis for gastrointestinal cancer |
WO2012030313A1 (en) * | 2008-04-25 | 2012-03-08 | James Winkelman | Method of determining a complete blood count and a white blood cell differential count |
JP4288323B1 (en) * | 2008-09-13 | 2009-07-01 | 独立行政法人科学技術振興機構 | Microscope device and fluorescence observation method using the same |
JP5438962B2 (en) * | 2008-12-25 | 2014-03-12 | シスメックス株式会社 | Cell image display device |
EP2239675A1 (en) * | 2009-04-07 | 2010-10-13 | BIOCRATES Life Sciences AG | Method for in vitro diagnosing a complex disease |
CN101567084B (en) * | 2009-06-05 | 2011-04-06 | 西安电子科技大学 | Method for detecting picture contour based on combination of level set and watershed |
US20120276555A1 (en) * | 2009-10-21 | 2012-11-01 | Peter Kuhn | Method of Using Non-Rare Cells to Detect Rare Cells |
CN101719278B (en) * | 2009-12-21 | 2012-01-04 | 西安电子科技大学 | Automatic tracking method for video microimage cells based on KHM algorithm |
US10203275B2 (en) * | 2010-08-05 | 2019-02-12 | Abbott Point Of Care, Inc. | Method and apparatus for automated whole blood sample analyses from microscopy images |
FR2971337B1 (en) * | 2011-02-04 | 2013-03-01 | Horiba Abx Sas | DEVICE AND METHOD FOR MULTIPARAMETRIC MEASUREMENTS OF MICROPARTICLES IN A FLUID |
ES2915263T3 (en) * | 2011-06-17 | 2022-06-21 | Roche Diagnostics Hematology Inc | Solution and procedure for the histoprocessing of biological samples |
CN103930762B (en) * | 2011-06-17 | 2018-03-20 | 罗氏血液诊断股份有限公司 | The system and method with checking are shown for sample |
EP2972550B1 (en) | 2013-03-11 | 2024-09-25 | Roche Diagnostics Hematology, Inc. | Method for classifying white blood cells |
-
2014
- 2014-03-11 EP EP14719553.1A patent/EP2972550B1/en active Active
- 2014-03-11 CA CA2904401A patent/CA2904401A1/en not_active Abandoned
- 2014-03-11 AU AU2014249011A patent/AU2014249011A1/en not_active Abandoned
- 2014-03-11 US US14/775,381 patent/US9690976B2/en active Active
- 2014-03-11 WO PCT/US2014/023401 patent/WO2014164757A1/en active Application Filing
- 2014-03-11 CN CN201480026258.7A patent/CN105190400A/en active Pending
- 2014-03-11 JP JP2016501227A patent/JP6343660B2/en active Active
-
2016
- 2016-06-14 HK HK16106816.1A patent/HK1218784A1/en unknown
-
2017
- 2017-06-26 US US15/633,315 patent/US10083342B2/en active Active
- 2017-12-21 JP JP2017245118A patent/JP6580117B2/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1998052018A1 (en) * | 1997-05-14 | 1998-11-19 | Applied Imaging Corporation | Identification of objects by means of multiple imaging |
US20120263369A1 (en) * | 2011-04-14 | 2012-10-18 | Abbott Point Of Care, Inc. | Method and apparatus for compressing imaging data of whole blood sample analyses |
US20120262703A1 (en) * | 2011-04-15 | 2012-10-18 | Constitution Medical, Inc. | Measuring Volume and Constituents of Cells |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2017223412A1 (en) * | 2016-06-24 | 2017-12-28 | Beckman Coulter, Inc. | Image atlas systems and methods |
US11448583B2 (en) | 2016-06-24 | 2022-09-20 | Beckman Coulter, Inc. | Image atlas systems and methods |
EP4092403A1 (en) * | 2016-06-24 | 2022-11-23 | Beckman Coulter Inc. | Image atlas systems and methods |
EP4358034A1 (en) | 2022-10-18 | 2024-04-24 | F. Hoffmann-La Roche AG | Clinical support system and associated computer-implemented methods |
Also Published As
Publication number | Publication date |
---|---|
US9690976B2 (en) | 2017-06-27 |
JP6580117B2 (en) | 2019-09-25 |
CN105190400A (en) | 2015-12-23 |
CA2904401A1 (en) | 2014-10-09 |
US20160026852A1 (en) | 2016-01-28 |
JP2016511419A (en) | 2016-04-14 |
EP2972550B1 (en) | 2024-09-25 |
EP2972550A1 (en) | 2016-01-20 |
JP2018077236A (en) | 2018-05-17 |
AU2014249011A1 (en) | 2015-09-24 |
AU2014249011A8 (en) | 2015-10-15 |
WO2014164757A8 (en) | 2015-09-24 |
HK1218784A1 (en) | 2017-03-10 |
US20180012062A1 (en) | 2018-01-11 |
US10083342B2 (en) | 2018-09-25 |
JP6343660B2 (en) | 2018-06-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10083342B2 (en) | Imaging blood cells | |
US10801944B2 (en) | High accuracy 5-part differential with digital holographic microscopy and untouched leukocytes from peripheral blood | |
EP2936116B1 (en) | System and method for classification of particles in a fluid sample | |
JP5717647B2 (en) | Multinuclear cell classification and micronucleus scoring | |
Vink et al. | An automatic vision‐based malaria diagnosis system | |
AU2009251162B2 (en) | Method for classifying slides using scatter plot distributions | |
Meimban et al. | Blood cells counting using python opencv | |
US20230194407A1 (en) | Method and system for label-free imaging and classification of malaria parasites | |
Nasreen et al. | Counting of RBC using circular hough transform with median filtering | |
CN118648020A (en) | Measuring image quality of blood cell images | |
Alqudah et al. | Automatic Segmentation and Classification of White Blood Cells in Peripheral Blood Samples. | |
Ravkin et al. | Automated microscopy system for detection and genetic characterization of fetal nucleated red blood cells on slides | |
US10482314B2 (en) | Automatic calculation for ploidy classification | |
GB2610426A (en) | System and method for identifying and counting biological species | |
WO2020069738A1 (en) | Automatic calculation for ploidy classification | |
JP2015125098A (en) | Image processing apparatus and program | |
US20240037967A1 (en) | Blood analyser with out-of-focus image plane analysis and related methods | |
Shirazi et al. | A novel method for scanning electron microscope image segmentation and its application to blood cell analysis | |
On et al. | Spatio-temporal pattern recognition of dendritic spines and protein dynamics using live multichannel fluorescence microscopy | |
WO2023118440A1 (en) | Biological fluid analyser with light-setting-based cell classification | |
WO2024138139A1 (en) | Population based cell classification | |
EP4238072A1 (en) | A method for automated determination of platelet count based on microscopic images of peripheral blood smears | |
Prasad et al. | High-throughput identification and classification algorithm for leukemia population statistics | |
Tuxworth et al. | Segmentation of Inter-neurons in Three Dimensional Brain Imagery |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201480026258.7 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 14719553 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2904401 Country of ref document: CA |
|
ENP | Entry into the national phase |
Ref document number: 2016501227 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14775381 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2014249011 Country of ref document: AU Date of ref document: 20140311 Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2014719553 Country of ref document: EP |