WO2020219305A1 - Systems and methods to label structures of interest in tissue slide images - Google Patents

Systems and methods to label structures of interest in tissue slide images Download PDF

Info

Publication number
WO2020219305A1
WO2020219305A1 PCT/US2020/028162 US2020028162W WO2020219305A1 WO 2020219305 A1 WO2020219305 A1 WO 2020219305A1 US 2020028162 W US2020028162 W US 2020028162W WO 2020219305 A1 WO2020219305 A1 WO 2020219305A1
Authority
WO
WIPO (PCT)
Prior art keywords
tissue slide
stain
images
image
tissue
Prior art date
Application number
PCT/US2020/028162
Other languages
French (fr)
Inventor
Jack ZENIEH
Marcel PRASTAWA
Gerardo Fernandez
Original Assignee
Icahn School Of Medicine At Mount Sinai
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Icahn School Of Medicine At Mount Sinai filed Critical Icahn School Of Medicine At Mount Sinai
Priority to US17/605,943 priority Critical patent/US11972621B2/en
Priority to EP20795913.1A priority patent/EP3959654A4/en
Publication of WO2020219305A1 publication Critical patent/WO2020219305A1/en
Priority to US18/617,762 priority patent/US20240233418A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/69Microscopic objects, e.g. biological cells or cellular parts
    • G06V20/698Matching; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24133Distances to prototypes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/40Software arrangements specially adapted for pattern recognition, e.g. user interfaces or toolboxes therefor
    • G06F18/41Interactive pattern learning with a human teacher
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/70Labelling scene content, e.g. deriving syntactic or semantic representations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/24Indexing scheme for image data processing or generation, in general involving graphical user interfaces [GUIs]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30024Cell structures in vitro; Tissue sections in vitro

Definitions

  • Some implementations are generally related to biomedical image processing, and in particular, to systems and methods to label structures of interest in tissue slide images.
  • Some implementations can include a method, computer readable medium, and/or system to label structures of interest in tissue slide images. Some implementations can provide more accurate and more efficient generation of the labels for images compared to some conventional techniques.
  • the labeled images can be used as training data for a machine learning model constructed to identify structures of interest in tissue slide images.
  • Some implementations can include a computer-implemented method to label structures in tissue slide images.
  • the method can include scanning one or more stained tissue slides to generate a plurality of tissue slide images including a first tissue slide image of a first tissue slide stained with a first stain and a second tissue slide image of the first tissue slide stained with a second stain, wherein the first stain and the second stain have a common attribute.
  • the method can also include preprocessing the plurality of tissue slide images, and performing an image registration process on the plurality of tissue slide images, wherein the image registration process is performed on the plurality of tissue slide images using an image of the common attribute in each of the plurality of tissue slide images.
  • the method can further include identifying one or more candidate regions of interest (ROI) within the plurality of tissue slide images, and generating one or more label annotations for structures within the plurality of tissue slide images.
  • ROI candidate regions of interest
  • the first stain includes hematoxylin and eosin
  • the second stain includes an immunohistochemistry (IHC) marker.
  • preprocessing the plurality of tissue slide images includes separating foreground structures from background structures in the plurality of tissue slide images, and unmixing color of the plurality of tissue slide images to separate components of the first stain and the second stain into different channels.
  • the method can also include for the tissue slide images stained with the first stain, separating a hematoxylin channel from an eosin channel, for the tissue slide image stained with the second stain, separating a hematoxylin channel from a DAB channel.
  • the image registration process includes minimizing a mutual information metric between deconvolved hematoxylin channels for images of slides stained with the first stain and images of slides stained with the second stain, where the minimizing is performed in two stages: at a first stage, performing a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms; and at a second stage, performing a high resolution registration of localized fields using affine and B-spline transforms.
  • identifying the one or more candidate regions of interest includes creating a registered image for a same tissue area, wherein the registered image includes the first tissue slide image and the second tissue slide image aligned for the same tissue area.
  • the method can further include providing the tissue slide images and the one or more label annotations for structures within the plurality of tissue slide images to a reviewer system graphical user interface (GUI), and receiving updated labels or updated comments from the reviewer system GUI.
  • GUI graphical user interface
  • the method can also include storing the updated labels or updated comments with corresponding tissue slide images, and providing labeled tissue slide images as training data to a model.
  • Some implementations can include a system comprising one or more processors coupled to a computer readable memory having stored thereon software instructions that, when executed by the one or more processors, cause the one or more processors to perform or control performance of operations.
  • the operations can include scanning one or more stained tissue slides to generate a plurality of tissue slide images including a first tissue slide image of a first tissue slide stained with a first stain and a second tissue slide image of the first tissue slide stained with a second stain, wherein the first stain and the second stain have a common attribute.
  • the operations can also include preprocessing the plurality of tissue slide images, and performing an image registration process on the plurality of tissue slide images, wherein the image registration process is performed on the plurality of tissue slide images using an image of the common attribute in each of the plurality of tissue slide images.
  • the operations can further include identifying one or more candidate regions of interest (ROI) within the plurality of tissue slide images, and generating one or more label annotations for structures within the plurality of tissue slide images.
  • ROI regions of interest
  • the first stain includes hematoxylin and eosin
  • the second stain includes an immunohistochemistry (IHC) marker.
  • the preprocessing the plurality of tissue slide images includes separating foreground structures from background structures in the plurality of tissue slide images, and unmixing color of the plurality of tissue slide images to separate components of the first stain and the second stain into different channels.
  • the operations can further include for the tissue slide images stained with the first stain, separating a
  • hematoxylin channel from an eosin channel, for the tissue slide image stained with the second stain, separating a hematoxylin channel from a DAB channel.
  • the image registration process includes: minimizing a mutual information metric between deconvolved hematoxylin channels for images of slides stained with the first stain and images of slides stained with the second stain, where the minimizing is performed in two stages: at a first stage, performing a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms; and at a second stage, performing a high resolution registration of localized fields using affine and B-spline transforms.
  • identifying the one or more candidate regions of interest includes: creating a registered image for a same tissue area, wherein the registered image includes the first tissue slide image and the second tissue slide image aligned for the same tissue area.
  • the operations can also include providing the tissue slide images and the one or more label annotations for structures within the plurality of tissue slide images to a reviewer system graphical user interface (GUI), and receiving updated labels or updated comments from the reviewer system GUI.
  • GUI graphical user interface
  • the operations can further include storing the updated labels or updated comments with corresponding tissue slide images, and providing labeled tissue slide images as training data to a model.
  • Some implementations can include a non-transitory computer readable medium having software instruction stored thereon that, when executed by a processor, cause the processor to perform or control performance of operations.
  • the operations can include scanning one or more stained tissue slides to generate a plurality of tissue slide images including a first tissue slide image of a first tissue slide stained with a first stain and a second tissue slide image of the first tissue slide stained with a second stain, wherein the first stain and the second stain have a common attribute.
  • the operations can also include preprocessing the plurality of tissue slide images, and performing an image registration process on the plurality of tissue slide images, wherein the image registration process is performed on the plurality of tissue slide images using an image of the common attribute in each of the plurality of tissue slide images.
  • the operations can further include identifying one or more candidate regions of interest (ROI) within the plurality of tissue slide images, and generating one or more label annotations for structures within the plurality of tissue slide images.
  • ROI regions of interest
  • the first stain includes hematoxylin and eosin
  • the second stain includes an immunohistochemistry (IHC) marker.
  • the preprocessing the plurality of tissue slide images includes separating foreground structures from background structures in the plurality of tissue slide images, and unmixing color of the plurality of tissue slide images to separate components of the first stain and the second stain into different channels.
  • the operations can further include for the tissue slide images stained with the first stain, separating a
  • the image registration process includes: minimizing a mutual information metric between deconvolved hematoxylin channels for images of slides stained with the first stain and images of slides stained with the second stain, where the minimizing is performed in two stages: at a first stage, performing a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms; and at a second stage, performing a high resolution registration of localized fields using affine and B-spline transforms.
  • identifying the one or more candidate regions of interest includes: creating a registered image for a same tissue area, wherein the registered image includes the first tissue slide image and the second tissue slide image aligned for the same tissue area.
  • the operations can also include providing the tissue slide images and the one or more label annotations for structures within the plurality of tissue slide images to a reviewer system graphical user interface (GUI), and receiving updated labels or updated comments from the reviewer system GUI.
  • GUI graphical user interface
  • the operations can further include storing the updated labels or updated comments with corresponding tissue slide images, and providing labeled tissue slide images as training data to a model.
  • Some implementations can include a system having one or more hardware processors configured to perform or control performance of the method mentioned above.
  • the one or more hardware processors may include a graphics processing unit (GPU) with a large number of processing units.
  • the hardware processor may include a neural network processor.
  • the system can include an imaging system.
  • Some implementations can include a non-transitory computer readable medium having software instruction stored thereon that, when executed by a processor, cause the processor to perform or control performance of operations according to the method mentioned above.
  • FIG. 1 is a diagram of an example tissue slide image labeling environment in accordance with some implementations.
  • FIG. 2 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
  • FIG. 3 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
  • FIG. 4 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
  • FIG. 5 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
  • FIG. 6 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
  • FIG. 7 is a diagram of an example graphical user interface for tissue slide image labeling review in accordance with some implementations.
  • FIG. 8 is a diagram of an example graphical user interface for tissue slide image labeling review in accordance with some implementations.
  • FIG. 9 is a diagram of an example graphical user interface for tissue slide image labeling review in accordance with some implementations.
  • FIG. 10 is a diagram of an example system configured to perform tissue slide image labeling in accordance with some implementations.
  • a need may exist for a system that makes this process of generating labeled data less labor intensive and more accurate.
  • Some implementations include a method to label structures of interest in tissue slide images.
  • the method can include staining a specimen on a slide with hematoxylin and eosin (H&E).
  • H&E stained slide is then imaged with a scanner (e.g., a Philips Ultrafast scanner).
  • the slide is then destained and stained with at least one antibody of interest that is capable of marking the structures of interest (e.g., using an immunohistochemistry or IHC marker technique).
  • IHC staining is complete, the IHC image is scanned on a scanner.
  • the staining and scanning yield two images: an H&E slide image and an IHC slide image.
  • the image of the H&E slide and the corresponding IHC slide are then supplied to an image preprocessing operation.
  • This operation can involve multiple sub-operations.
  • An example first sub-operation is separating foreground from background.
  • color unmixing can involve separation of the hematoxylin from the eosin.
  • color unmixing can include separation of the hematoxylin from the 3,3 '-Diaminobenzi dine (DAB), or other similar IHC stain components.
  • DAB 3,3 '-Diaminobenzi dine
  • Candidate regions of interest can then be extracted from the slide to create a registered H&E and IHC image for the same tissue area.
  • This region of interest extraction or identification can be done manually by a technician or a pathologist by manually selecting a field from the image.
  • the extraction can also be done automatically by a computer in addition to or as an alternative to manual region of interest extraction.
  • the automatic computerized region of interest identification can be guided multiple ways, such as by utilizing heuristics.
  • An example can include having a pathologist pre-label the areas of tumor on a slide.
  • the computer system can then randomly sample areas inside the tumor and outside the tumor so that a predetermined ratio of tumor to non-tumor regions can be extracted.
  • Additional intelligence can be added to the above process to select regions most likely to contain valuable data. For example, if one were looking for mitotic figures and the user was using a mitotic figure stain such as phosphohistone (PHH3), the system could incorporate information looking for the highest density of PHH3 on the slide. This technique would likely generate the greatest amount of mitotic figures for a given slide image.
  • Another technique can be based on a machine learning model technique. An example would be training a convolutional neural network (CNN) to select candidate ROIs based upon an available training set of selected and not selected ROIs based on human input.
  • CNN convolutional neural network
  • Raw IHC images often do not directly translate into binary labels for objects of interest.
  • IHC images are processed in order to produce object labels. This processing can vary depending on the IHC label used.
  • a segmentation technique such as level set segmentation can be used.
  • an efficient process of human review can be undertaken.
  • the user is able to quickly switch between registered H&E and IHC images for the same area of tissue.
  • Some implementations can also provide for a quick multiclass and multilabel system for each ROI.
  • GUIs graphical user interfaces
  • the user can label an overall triage for the image, including Accept, Reject, or Review. They can also comment on the quality of the ROI and the IHC. In addition, the user can quickly comment if there is no mitosis.
  • the user can enter in freetext comments that allow the user to create new labels on demand.
  • Some implementations can include receiving annotations on automatically generated fields using a web-based interface.
  • an expert can confirm, correct, or add information using one or more of the following labels: a) Abnormal Mitotic; b) Mitotic; c) Borderline; d) Regular Nucleus; e) Look Alike; f) Artifact (e.g., hematocyrin); g) Telophase 1; h) Telophase 2; and/or i) Error (e.g., bad staining, algorithm error/bug).
  • Automatic field annotations can be combined with annotation and processed to generate ground truth and validation data.
  • Some implementations can provide near interactive integration of annotations using on-line/incremental learning.
  • FIG. 1 is a diagram of an example tissue slide image labeling environment 100 in accordance with some implementations.
  • the environment 100 includes one or more slides 102 (e.g., slides prepared with H&E, an IHC marker, or the like) and one or more scanners 104 arranged to scan the slides 102.
  • the scanners 104 are coupled to a labeling system 106, which is also coupled to one or more reviewer systems 108 that can each include a graphical user interface (GUI) 110.
  • GUI graphical user interface
  • the environment 100 also includes one or more labeled images 112 output by the labeling system 106 and provided to a training data store 114 that contains training data for a machine learning model 116 that is trained (or being trained) to perform labeling of structures within tissue slide images.
  • the labeling system 106 performs tissue slide image labeling of structures of interest (e.g., epithelia and/or stroma, etc.) according to FIGS. 2-6 described below.
  • FIG. 2 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
  • Processing begins at 202, where one or more slides (e.g., 102) are stained and scanned.
  • a specimen on a slide is stained with hematoxylin and eosin (H&E).
  • H&E hematoxylin and eosin
  • the H&E slide is then imaged or scanned with a scanner (e.g., 104) such as a Philips Ultrafast scanner.
  • the H&E slide is then destained and IHC stained with at least one antibody of interest that is capable of marking the structures of interest.
  • IHC staining is complete, the IHC slide is scanned on a scanner. Processing continues to 204.
  • the slide images are preprocessed.
  • the H&E slide image and the corresponding IHC slide image are preprocessed (e.g., using the labeling system 106), which can include multiple sub-operations.
  • An example first sub-operation is separating foreground from background. After separating foreground from background, color unmixing is performed to separate into different channels each of the stains can be done. In the H&E slide, this would involve separation of the hematoxylin from the eosin. In the case of the IHC slide, this would involve separation of the hematoxylin from the DAB (or other similar IHC stain components).
  • efficient analysis of large whole slide images can be conducted by isolating relevant tissue from background. This isolation or separation can be accomplished by measuring textural content through derivative filters, such as a Laplacian filter. Analysis can be constrained to regions with Laplacian values larger than the average over the image. Structural signatures can be detected by separating or deconvolving the stains from the color images. Relevant signatures include hematoxylin stain channel images from H&E and IHC for registering the slide pairs, and structure-specific stains in IHC for generating annotations. Processing continues to 206.
  • image registration is performed to align the whole slide H&E images and the IHC images. Separation of images into color components helps enable registration.
  • the hematoxylin channel image can be used to register the images. Multiple registration technologies can be used. In some implementations, mutual information can be used for registration. In some implementations, registration of H&E and IHC images can be performed by maximizing the mutual information metric between the deconvolved hematoxylin channels, and performed in two stages. The first stage can include a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms. The second stage can include a high resolution registration of localized H&E and IHC fields using affine and B-spline transforms, which can be performed at a whole slide level. Processing continues to 208.
  • candidate regions of interest are generated or identified within the slide images.
  • candidate regions can be extracted from the slide images to create a registered H&E and IHC image for the same tissue area.
  • region (or field) extraction can be done manually by a technician or a pathologist manually selecting a field from the image.
  • region extraction can be done automatically by a computer system (e.g., labeling system 106).
  • the computerized region extraction can be guided multiple ways.
  • one technique may include utilizing heuristics.
  • a pathologist could prelabel areas of tumor on a slide.
  • the labeling system would then randomly sample areas inside the tumor and outside the tumor so that a predetermined ratio of tumor/non-tumor regions would be extracted.
  • Additional intelligence or program logic can be added to this process to select regions most likely to contain valuable data. For example, if one were looking for mitotic figures and the user was using a mitotic figure stain such as PHH3, the algorithm could incorporate information looking for the highest density of PHH3 on the slide, which would emphasize generating mitotic figures.
  • Yet another algorithm could be based on a machine learning technique (e.g., using model 116 and training data 114).
  • An example would be training a CNN (e.g., 116) to select regions of interest based upon an available training set (e.g., 114) of selected and not selected regions of interest based on human input. Additional detail about identifying and extracting regions of interest is described below in connection with FIG. 3. Processing continues to 210.
  • a CNN e.g., 116
  • an available training set e.g., 114
  • Additional detail about identifying and extracting regions of interest is described below in connection with FIG. 3. Processing continues to 210.
  • label annotations are automatically generated.
  • raw IHC images often do not directly translate into binary labels for objects of interest.
  • IHC images are processed in order to produce object labels. This processing can vary depending on the IHC label used.
  • a segmentation technique such as level set segmentation can be used to segment the image and provide areas for labeling. Labeling of the segments can be provided by a machine learning model (e.g., 116).
  • the system can automatically generate segmentation masks for the local fields. These segmentation masks can serve as a starting point for manual correction, reducing the workload for human reviewers.
  • the system also registers the corresponding IHC ROI using a higher order model (e.g., B-splines). Human reviewers are presented with the H&E ROI, IHC ROI, and the automatically generated masks when performing the annotation.
  • masks can be generated by applying signal
  • the final mask for epithelium can be generated using the signal enhancement followed by clustering using Gabor feature filters, and random walker segmentation.
  • the automatically labeled images can optionally be provided to a reviewer system (e.g., 108) and displayed within a GUI (e.g., 110) and as shown in FIGS. 7-9 and described below.
  • the GUI can display the image(s), the automatically generated labels and structural identifications within the image and can provide an interface for a reviewer to add, delete or modify labels and/or comments corresponding to the image(s) being displayed.
  • Processing continues to 214.
  • the labeling system can receive updated image labels and comments from the reviewer system GUI. Once the automatically generated masks have been corrected and reviewed by a clinical expert, the system generates image data with the associated labels for building models through the use of machine learning techniques.
  • Cell annotations such as the one for mitotic figures, can be performed in multiple stages.
  • the generated regions of interest are reviewed by non-clinicians for obvious errors, and possible mitotic figures are marked (e.g., as solid dark blue cells).
  • the second stage local image patches are presented at the marked locations to clinical experts for cell categorization.
  • the third stage involves the outlining of the cells marked as mitotic or borderline mitotic.
  • the per-pixel labeling of mitotic figures can be used to sample local patches for machine learning.
  • Boundary annotations such as the one for epithelium, can be performed by presenting a non-clinician with the H&E and IHC images outlined using the technique described above. Non-clinicians could correct these outlines by drawing or erasing boundaries, they can also correct by using a paint filling method. To handle slides with significant variability of image appearance due to the staining process, the system may apply stain normalization to standardize the H&E image samples. Processing continues to 216.
  • the labeled image(s) with updated labels and comments are optionally provided to a machine learning system for training and/or feedback.
  • the model can include a neural network such as a convolutional neural network or fully convolutional neural network.
  • a neural network can include a hierarchical composite of multiple primary processing units called neurons. Contrary to the conventional multiple-layer perceptions (MLPs), where each neuron is directly collected to all neurons in a previous layer, a convolutional neural network (CNN) assumes the input as images and exploits spatially local correlation by enforcing a local connectivity pattern between neurons of adjacent layers.
  • MLPs multiple-layer perceptions
  • CNN convolutional neural network
  • the trained CNN can then be applied in multiple ways.
  • the trained CNN can be used on a given ROI to detect the likely mitotic figure candidates.
  • Blocks 202-218 (as well as any other blocks shown and described herein) can be repeated in whole or in part and performed in the same or different orders. Moreover, certain blocks can be eliminated, modified, combined with other blocks, and/or supplemented with other operations.
  • FIG. 3 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations. Processing begins at 302, where the H&E ROI is read (e.g., as a 2000 x 2000 pixel section). Processing continues to 304.
  • the H&E ROI is read (e.g., as a 2000 x 2000 pixel section). Processing continues to 304.
  • a large IHC ROI is extracted based on the H&E ROI. For example, assuming that IHC and HE whole slides are rigidly aligned in a lower resolution, the position of the H&E ROI (in the highest resolution) can be read and a (4000 x 4000 pixel) IHC ROI extracted from the whole slide. The H&E ROI is roughly located in the center of the IHC ROI now. Processing continues to 306.
  • a common channel e.g., hematoxylin
  • IHC ROI IHC ROI
  • the common channels of H&E and IHC images are rigidly aligned.
  • the IHC hematoxylin channel image (with bigger ROI) is rigidly aligned with the hematoxylin H&E channel image (with smaller ROI).
  • Registration can be performed in a three-level multi-resolution approach (4, 2, and 1 times the actual spacing of the image) initialized by geometrical centered transform. Processing continues to 310.
  • a free-form registration between the IHC ROI and the H&E ROI is performed.
  • a free-form registration between IHC hematoxylin channel image and H&E hematoxylin channel image is initialized by the rigid transformation obtained in 308.
  • a three-level, multi-resolution b-spline based registration can be used for deformation compensation between the IHC and H&E ROIs.
  • a mutual information metric can be used that measures statistical similarity between the information content between images in both rigid and free-form registration.
  • some implementations can include using the efficient Mattes Mutual Information metric that performs binning of image intensities and random sampling of pixel data. (See, D. Mattes, D. R.
  • a smaller IHC ROI e.g., 2000 x 2000
  • a smaller IHC ROI e.g., 2000 x 2000
  • FIG. 4 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
  • Processing begins at 402, where a mask is created (e.g., an ini mask).
  • the mask can provide an estimation of epithelial cells and separate the background from the tissue.
  • the mask can be created from the H&E hematoxylin image and the IHC DAB image using a signal enhancement approach, where known properties (e.g.,“blobbiness” of a cell, smooth contiguous regions for glands, etc.) can be enhanced.
  • the mask can be dilated to give a rough estimation of epithelium and separate the background from the tissue. Processing continues to 404.
  • image features are extracted.
  • image features can be extracted from one or more smaller IHC ROI images and the features can include:
  • H&E channel values in the H&E image e.g., H&E hematoxylin and H&E eosin channels
  • the above features can be computed for the hemo HE, eosin HE, hemo IHC, and/or dab_IHC in one frequency and three different rotation angles (e.g., 0, 45, and/or 135 degrees).
  • a total number of 20 features are computed for each pixel, constructed from different combinations of derivative values in different angular directions and at different spatial scales. These features capture patterns such as peaks and valleys in different directions and different sizes, enabling registration based on cellular boundaries and chromatin content. Processing continues to 406.
  • image pixels are clustered based on image features. For example, pixels in the ROI can be assigned to one of three clusters (e.g., nuclei (or cells), tissue, and background) using k-means clustering approach. Processing continues to 408.
  • three clusters e.g., nuclei (or cells), tissue, and background
  • a mask e.g., an epi-stroma mask
  • the epi-stroma mask can include identification of epithelium and stroma regions within one or more images, which can be useful because areas of cancer are sometimes associated with epithelium.
  • the mask can be generated using a random walker technique, which is a semi-supervised segmentation method that involves a number of pixels with known labels (called seeds), can be used to perform image segmentation. The random walk process starts from these seeds.
  • the clustering process in the previous operation basically divides the pixels to three groups: 1) nuclei (cells); 2) tissue; and 3) background. However, the groups are not labeled.
  • the system first determines the background cluster, which can include the cluster having the least overlap (correlation) with the ini_mask (created at 402). Then, for the remaining two clusters (label_l and label_2) (e.g., nuclei and tissue), the system may consider pixels that lay within the ini mask. Two different masks can be created using the random walker segmentation. In the first mask, an eroded version of label l can be used for seeding and a dilated version of label_2 can be used as unknown background. In the second mask, labels can be switched. From the created output masks, the mask having larger average dab IHC in the overlapping pixels is chosen as the output epi-stroma mask, which can be output as an initial ground truth reference that can be utilized in downstream processing and/or corrected by user in some implementations.
  • Some implementations can include a software application for specialized medical equipment (e.g., slide scanning systems, etc,), desktop computers, laptop computers, and mobile devices (e.g., smartphones, tablet computing devices, etc.) that can be used by physicians and/or other health care or biomedical professionals to perform tissue slide image structure labeling.
  • the software application can also be provided as a web service accessible over a wired or wireless computer network.
  • the method, or portions of the method can be initiated automatically by a device.
  • the method (or portions thereol) can be repeatedly performed or performed based on the occurrence of one or more particular events or conditions.
  • events or conditions can include: obtaining one or more images that have been newly captured by, uploaded to, or otherwise accessible by a device (e.g., an imaging device such as a slide scanner as described herein), a predetermined time period having expired since the last performance of method 200, 300, and/or 400, and/or one or more other events or conditions occurring which can be specified in settings of a device implementing method 200, 300, and/or 400.
  • such conditions can be previously specified by a user in stored custom preferences of the user (accessible by a device or method with user consent).
  • an imaging device e.g., slide scanner
  • an imaging device can capture one or more images and can perform the method 200, 300, and/or 400.
  • an imaging device can send one or more captured images to an external system (e.g., a cloud computing system, a server, a mobile device, etc.) over a network, and the external system can process the images using method 200, 300, and/or 400.
  • an external system e.g., a cloud computing system, a server, a mobile device, etc.
  • FIG. 5 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations. Processing begins at 502, where one or more tissue slides are H&E stained and scanned as described herein. Processing continues to 504.
  • the slides are re-stained with an IHC marker and scanned. Processing continues to 508.
  • the hematoxylin channel image is extracted from IHC. Processing continues to 510.
  • an epi-stroma mask is generated.
  • FIG. 6 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations. Processing begins at 602, where one or more tissue slides are H&E stained and scanned as described herein. Processing continues to 604.
  • the slides are re-stained with an IHC marker and scanned. Processing continues to 608.
  • the hematoxylin channel is extracted from the IHC image. Processing continues to 610.
  • candidate field extraction e.g., region of interest
  • labeling using a mask e.g., a PHH3 mask
  • the PHH3 mask can be created from regions marked brown in IHC that are constrained to be small and cell-like. These can be used as a marker of possible mitotic figures for review by experts (either a human expert or an expert system). Processing continues to 614.
  • the mask e.g., PHH3 mask
  • labels and/or labeled objects can be provided as training data for a machine learning model, e.g., a convolutional neural network (CNN) model.
  • CNN convolutional neural network
  • FIGS. 7-9 include diagrams of example graphical user interfaces (GUIs) that show example modes for navigating between images and performing annotations.
  • GUIs graphical user interfaces
  • FIGS. 7 and 8 are diagrams of a GUI for an overview/browsing mode
  • FIG. 9 is a diagram of a GUI for a mitotic figure annotating mode.
  • User can triage fields as accept, reject, or review using butons on top right pane. Reject is for ROIs with bad quality or artifact like tissue being tom. Review is for future review between different experts. Users can provide input to the GUI, e.g., via keyboard up and down keys or other input, to move backward and forward between different ROIs.
  • the user is able to quickly switch between registered H&E and IHC images for the same area of tissue.
  • H&E for a sample of breast tissue is displayed along with an IHC of PHH3 for detecting mitotic figures.
  • FIG. 8 is a diagram of an example graphical user interface (GUI) for tissue slide image labeling review that shows a mode where the registered IHC image is displayed and where users can access and select images for display (e.g., by clicking the pull down menu on top left pane and selecting the desired image to display).
  • GUI graphical user interface
  • Experts can use the IHC images to confirm if a given set of H&E pixels contain a certain structure of interest because the IHC images are stained for specific targets (e.g., cytokeratin or mitotic in brown).
  • the system also provides for a quick multiclass and multilabel system for each ROI.
  • the user can label an overall triage for the image, including Accept, Reject, or Review. They can also comment on the quality of the ROI and the IHC. In addition, the user can quickly comment if there are no mitosis. Furthermore, for each of the fields, the user can enter in freetext comments that allow the user to create new labels on demand.
  • FIG. 9 is a diagram of an example graphical user interface (GUI) for tissue slide image labeling review in accordance with some implementations.
  • GUI graphical user interface
  • a user can drop annotation objects by picking a classification label on botom left of screen and then click on a cell in the field of view to tag it with different labels, which can be shown as different colored rectangles.
  • the user is able to see in a yellow outline the results of IHC
  • the IHC is of PHH3 which indicates mitotic figures. Therefore, the yellow area on the H&E would indicate where a mitotic figure is.
  • the system permits the user to assign multiple classes to various objects in the image.
  • allowed classes include: a) Abnormal mitotic; b) Mitotic; c) Borderline (mitotic); d) Regular nucleus; d) Look alike; e) Artifact; 1) Telophase 1 (one nucleus of a cell dividing); g) Telophase 2 (the other nucleus of a dividing cell); h) Prophase; and/or g) Error.
  • the user can then assign these subclasses to the detected object. Furthermore, the user may also annotate additional objects that are in the scene that were either not detected in error or are not intended to be detected by the IHC but are of utility, e.g., a regular nucleus, look alike, etc.
  • a rich set of ROI level labels can be assigned by the user for future subclassing. Examples include: a) Marker present - indicating that there is marker present on the ROI; b) Border segmentation issues; c) Registration mismatch - the IHC and the H&E do not register - this may be due to faulty registration or differences in the tissue such as one having a tear in it; d) Lumen staining - a lumen is picking up stain where it should not be; and/or e) Weak epi - the epithelial regions are weakly staining
  • FIG. 10 is a block diagram of an example computing device 1000 which may be used to implement one or more features described herein.
  • computing device 1000 may be used to implement a computer device, e.g., tissue slide labeling system or reviewer system device (e.g., 106 and/or 108 of FIG. 1), and perform appropriate method implementations described herein.
  • Device 1000 can be any suitable computer system, server, or other electronic or hardware device.
  • the device 1000 can be a mainframe computer, desktop computer, workstation, portable computer (including a smart phone or laptop), or medical device.
  • device 1000 includes a processor 1002, a memory 1004, and input/output (I/O) interface 1006, all operatively coupled to each other.
  • I/O input/output
  • Processor 1002 can be one or more processors and/or processing circuits to execute program code and control basic operations of the device 1000.
  • A“processor” includes any suitable hardware and/or software system, mechanism or component that processes data, signals or other information.
  • a processor may include a system with a general-purpose central processing unit (CPU), multiple processing units, dedicated circuitry for achieving functionality, or other systems. Processing need not be limited to a particular geographic location, or have temporal limitations. For example, a processor may perform its functions in“real-time,”“offline,” in a“batch mode,” etc. Portions of processing may be performed at different times and at different locations, by different (or the same) processing systems.
  • a computer may be any processor in communication with a memory.
  • Memory 1004 is typically provided in device 1000 for access by the processor 1002, and may be any suitable processor-readable storage medium, e.g., random access memory (RAM), read-only memory (ROM), electrical erasable read-only memory
  • RAM random access memory
  • ROM read-only memory
  • Memory 1004 can store software operating on the server device 1000 by the processor 1002, including an operating system 1008, one or more applications 1010, e.g., a tissue slide image labeling application and application data 1020.
  • applications 1010 can include instructions that, in response to execution by processor 1002, enable processor 1002 to perform or control performance of the functions described herein, e.g., some or all of the methods of FIGS. 2-6.
  • applications 1010 can include a tissue slide image labeling application, which as described herein can provide automatic labels for tissue slide images.
  • Any software in memory 1004 can alternatively be stored on any other suitable storage location or computer-readable medium.
  • memory 1004 (and/or other connected storage device(s)) can store images, and other instructions and data used in the features described herein.
  • Memory 1004 and any other type of storage can be considered “storage” or “storage devices.”
  • application(s) 1010 can include a machine-learning model (e.g., a two-stage neural network model including a coarse detector and a fine discriminator) that can be trained to detect cell objects in images.
  • training data may include tissue slide scan data, CT image data, etc.
  • the model form or structure may specify connectivity between various nodes and organization of nodes into layers. For example, nodes of a first layer (e.g., input layer) may receive data as input data or application data. Such data can include, for example, one or more pixels. Subsequent intermediate layers may receive as input output of nodes of a previous layer per the connectivity specified in the model form or structure. These layers may also be referred to as hidden layers.
  • a final layer (e.g., output layer) produces an output of the machine-learning application. For example, the output may be an identified and/or labeled candidate object or particular object of interest, etc. depending on the specific trained model.
  • the trained model may include a weight individual nodes and/or connections.
  • a respective weight may be applied to a connection between each pair of nodes that are connected per the model form, e.g., nodes in successive layers of the neural network.
  • the respective weights may be randomly assigned, or initialized to default values.
  • the model may then be trained, e.g., using data, to produce a result, where the training can include adjusting one or more of nodes, node structure, connections, and/or weights.
  • a model can include a loss function representing the difference between a predicted value and an actual label.
  • the model can be trained to minimize the loss function.
  • Training can include supervised, unsupervised, or semi-supervised learning techniques.
  • the training data can include a plurality of inputs (e.g., tissue slide images) and a corresponding expected output for each input (e.g., tissue slide images with objects or structures of interest identified and labeled).
  • values of the weights are automatically adjusted, e.g., in a manner that increases a probability that the model produces the expected output when provided similar input (i.e., reduces the loss function).
  • unsupervised learning models leam relationships between elements in a data set and classify raw data without the benefit of labeled training data.
  • Semi-supervised learning can include a combination of supervised and unsupervised techniques, for example, a small amount of labeled training data and a large amount of unlabeled training data can be provided to a model for learning.
  • I/O interface 1006 can provide functions to enable interfacing the computing device 1000 with other systems and devices. For example, network communication devices, storage devices (e.g., memory and/or database), and input/output devices can communicate via the I/O interface 1006. In some implementations, the I/O interface 1006 can connect to interface devices including input devices (keyboard, pointing device, touchscreen, microphone, camera, scanner, etc.) and/or output devices (display device, speaker devices, printer, motor, etc.). Display device 1014 is one example of an output device that can be used to display images and other data, e.g., one or more tissue slide images, structure of interest identification, and labels provided by an application as described herein. Display device 1014 can be connected to device 1000 via local connections (e.g., display bus) and/or via networked connections and can be any suitable display device, some examples of which are described below.
  • local connections e.g., display bus
  • FIG. 10 shows one block for each of processor 1002, memory 1004, I/O interface 1006, and software blocks 1008 and 1010. These blocks may represent one or more processors or processing circuitries, operating systems, memories, I/O interfaces, applications, and/or software modules.
  • device 1000 may not have all of the components shown and/or may have other elements including other types of elements instead of, or in addition to, those shown herein. While labeling system 106 is described as performing operations as described in some implementations herein, any suitable component or combination of components of system 106 or similar system, or any suitable processor or processors associated with such a system, may perform the operations described.
  • a user device can also implement and/or be used with features described herein.
  • Example user devices can be computer devices including some similar components as the device 1000, e.g., processor(s) 1002, memory 1004, and I/O interface 1006.
  • An operating system, software and applications suitable for the client device can be provided in memory and used by the processor.
  • the I/O interface for a client device can be connected to network communication devices, as well as to input and output devices, e.g., a microphone for capturing sound, a camera for capturing images or video, audio speaker devices for outputting sound, a display device for outputting images or video, or other output devices.
  • a display device 1014 can be connected to (or included in) the device 1000 to display images pre- and post-processing as described herein, where such display device can include any suitable display device, e.g., an LCD, LED, or plasma display screen, CRT, television, monitor, touchscreen, 3-D display screen, projector, or other visual display device.
  • display device can include any suitable display device, e.g., an LCD, LED, or plasma display screen, CRT, television, monitor, touchscreen, 3-D display screen, projector, or other visual display device.
  • Some implementations can provide an audio output device, e.g., voice output or synthesis that speaks text.
  • One or more methods described herein can be implemented by computer program instructions or code, which can be executed on a computer.
  • the code can be implemented by one or more digital processors (e.g., microprocessors or other processing circuitry), and can be stored on a computer program product including a non-transitory computer readable medium (e.g., storage medium), e.g., a magnetic, optical, electromagnetic, or semiconductor storage medium, including
  • the program instructions can also be contained in, and provided as, an electronic signal, for example in the form of software as a service (SaaS) delivered from a server (e.g., a distributed system and/or a cloud computing system).
  • SaaS software as a service
  • a server e.g., a distributed system and/or a cloud computing system
  • one or more methods can be implemented in hardware (logic gates, etc.), or in a combination of hardware and software.
  • Example hardware can be programmable processors (e.g.
  • FPGA field-programmable gate array
  • ASIC application specific integrated circuits
  • One or more methods described herein can be run in a standalone program that can be run on any type of computing device, a program run on a web browser, a mobile application (“app”) run on a mobile computing device (e.g., cell phone, smart phone, tablet computer), wearable device.
  • a client/server architecture can be used, e.g., a mobile computing device (as a client device) sends user input data to a server device and receives from the server the final output data for output (e.g., for display).
  • all computations can be performed within the mobile app (and/or other apps) on the mobile computing device.
  • computations can be split between the mobile computing device and one or more server devices.
  • routines may be integrated or divided into different combinations of systems, devices, and functional blocks. Any suitable programming language and programming techniques may be used to implement the routines of particular implementations. Different programming techniques may be employed, e.g., procedural or object-oriented. The routines may execute on a single processing device or multiple processors. Although the steps, operations, or computations may be presented in a specific order, the order may be changed in different particular implementations. In some implementations, multiple steps or operations shown as sequential in this specification may be performed at the same time.
  • Stain with Harris Hematoxylin e.g., Fisher, cat# SH30-500D, or the like, 7 minutes; wash in tap water- 1 minute.
  • Alcoholic Eosin solution e.g., Sigma, cat# HT110116, or the like
  • Anti -histone H3 is a rabbit polyclonal antibody, specific for Histone 3 phosphorylated on SerinelO and Threonine 11. Phosphorylation of H3, especially at highly conserved SerinelO site is crucial for the onset of mitosis. PHH3 staining is useful to distinguish mitotic figures from other morphological changes that can sometimes mimic mitosis, such as apoptosis or necrosis.
  • the Bond Polymer Refine Detection system used at PRECISE MD lab detects tissue- bound mouse and rabbit IgG and some mouse IgM primary antibodies.
  • the detection kit is intended for staining sections of formalin-fixed, paraffin-embedded tissues on the Bond automated system. It utilizes a novel controlled polymerization technology to prepare polymeric horseradish peroxidase (HRP) linker antibody and avoids the use of streptavidin and biotin, which can cause non-specific background due to the presence of endogenous biotin.
  • HRP horseradish peroxidase
  • the system gives great sensitivity and signal amplification, as well as reduces or eliminates the need of including negative controls with every run.
  • Reagent tray Bond Polymer Refined Detection Kit (Leica); primary antibody, PHH3 (Abeam) diluted 1 :2000 with 2% BSA in PBS buffer (prepared in the laboratory).
  • Antigen Retrieval step ER1, pH6 buffer (Leica, Cat# AR9961) 40 min

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Medical Informatics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Quality & Reliability (AREA)
  • Radiology & Medical Imaging (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Computational Linguistics (AREA)
  • Human Computer Interaction (AREA)
  • Investigating Or Analysing Biological Materials (AREA)
  • Image Analysis (AREA)

Abstract

Systems and methods to label structures of interest in tissue slide images are described.

Description

SYSTEMS AND METHODS TO LABEL STRUCTURES OF INTEREST
IN TISSUE SLIDE IMAGES
BACKGROUND
[0001] Computer vision technology has long been an interest in the field of biomedical imaging. Analysis of tissue samples, in particular, can often be tedious, weakly quantitative, and difficult to reproduce due to the vast quantities of data that need to be analyzed. Recently, machine learning based computer vision systems in other fields have achieved performance levels that suggest that significant improvements can be realized in biomedical imaging. However, these learning based systems often require input data that contains labels designating the items of interest.
[0002] The volume of labeled samples needed to train a machine learning system can be quite large compared to what a human may need in order to be trained. Generating large quantities of labeled data can be daunting for an unassisted human. For example, to generate labels that separate epithelial regions from stromal regions, a user manually traces a mask image to delineate the epithelial regions in order to separate those from the stromal regions.
In addition to being laborious, a manual process is error prone due to challenges in accurately and reproducibly determining the epithelial boundary.
[0003] The background description provided herein is for the purpose of generally presenting the context of the disclosure. Work of the presently named inventors, to the extent it is described in this background section, as well as aspects of the description that may not otherwise qualify as prior art at the time of filing, are neither expressly nor impliedly admitted as prior art against the present disclosure.
SUMMARY
[0004] Some implementations are generally related to biomedical image processing, and in particular, to systems and methods to label structures of interest in tissue slide images.
[0005] Some implementations can include a method, computer readable medium, and/or system to label structures of interest in tissue slide images. Some implementations can provide more accurate and more efficient generation of the labels for images compared to some conventional techniques. In some implementations, the labeled images can be used as training data for a machine learning model constructed to identify structures of interest in tissue slide images.
[0006] Some implementations can include a computer-implemented method to label structures in tissue slide images. The method can include scanning one or more stained tissue slides to generate a plurality of tissue slide images including a first tissue slide image of a first tissue slide stained with a first stain and a second tissue slide image of the first tissue slide stained with a second stain, wherein the first stain and the second stain have a common attribute. The method can also include preprocessing the plurality of tissue slide images, and performing an image registration process on the plurality of tissue slide images, wherein the image registration process is performed on the plurality of tissue slide images using an image of the common attribute in each of the plurality of tissue slide images.
[0007] The method can further include identifying one or more candidate regions of interest (ROI) within the plurality of tissue slide images, and generating one or more label annotations for structures within the plurality of tissue slide images. In some
implementations, the first stain includes hematoxylin and eosin, and the second stain includes an immunohistochemistry (IHC) marker.
[0008] In some implementations, preprocessing the plurality of tissue slide images includes separating foreground structures from background structures in the plurality of tissue slide images, and unmixing color of the plurality of tissue slide images to separate components of the first stain and the second stain into different channels. The method can also include for the tissue slide images stained with the first stain, separating a hematoxylin channel from an eosin channel, for the tissue slide image stained with the second stain, separating a hematoxylin channel from a DAB channel.
[0009] In some implementations, the image registration process includes minimizing a mutual information metric between deconvolved hematoxylin channels for images of slides stained with the first stain and images of slides stained with the second stain, where the minimizing is performed in two stages: at a first stage, performing a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms; and at a second stage, performing a high resolution registration of localized fields using affine and B-spline transforms. In some implementations, identifying the one or more candidate regions of interest includes creating a registered image for a same tissue area, wherein the registered image includes the first tissue slide image and the second tissue slide image aligned for the same tissue area. [0010] The method can further include providing the tissue slide images and the one or more label annotations for structures within the plurality of tissue slide images to a reviewer system graphical user interface (GUI), and receiving updated labels or updated comments from the reviewer system GUI. The method can also include storing the updated labels or updated comments with corresponding tissue slide images, and providing labeled tissue slide images as training data to a model.
[0011] Some implementations can include a system comprising one or more processors coupled to a computer readable memory having stored thereon software instructions that, when executed by the one or more processors, cause the one or more processors to perform or control performance of operations. The operations can include scanning one or more stained tissue slides to generate a plurality of tissue slide images including a first tissue slide image of a first tissue slide stained with a first stain and a second tissue slide image of the first tissue slide stained with a second stain, wherein the first stain and the second stain have a common attribute. The operations can also include preprocessing the plurality of tissue slide images, and performing an image registration process on the plurality of tissue slide images, wherein the image registration process is performed on the plurality of tissue slide images using an image of the common attribute in each of the plurality of tissue slide images.
[0012] The operations can further include identifying one or more candidate regions of interest (ROI) within the plurality of tissue slide images, and generating one or more label annotations for structures within the plurality of tissue slide images. In some
implementations, the first stain includes hematoxylin and eosin, and the second stain includes an immunohistochemistry (IHC) marker.
[0013] In some implementations, the preprocessing the plurality of tissue slide images includes separating foreground structures from background structures in the plurality of tissue slide images, and unmixing color of the plurality of tissue slide images to separate components of the first stain and the second stain into different channels. The operations can further include for the tissue slide images stained with the first stain, separating a
hematoxylin channel from an eosin channel, for the tissue slide image stained with the second stain, separating a hematoxylin channel from a DAB channel.
[0014] In some implementations, the image registration process includes: minimizing a mutual information metric between deconvolved hematoxylin channels for images of slides stained with the first stain and images of slides stained with the second stain, where the minimizing is performed in two stages: at a first stage, performing a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms; and at a second stage, performing a high resolution registration of localized fields using affine and B-spline transforms. In some implementations, identifying the one or more candidate regions of interest includes: creating a registered image for a same tissue area, wherein the registered image includes the first tissue slide image and the second tissue slide image aligned for the same tissue area.
[0015] The operations can also include providing the tissue slide images and the one or more label annotations for structures within the plurality of tissue slide images to a reviewer system graphical user interface (GUI), and receiving updated labels or updated comments from the reviewer system GUI. The operations can further include storing the updated labels or updated comments with corresponding tissue slide images, and providing labeled tissue slide images as training data to a model.
[0016] Some implementations can include a non-transitory computer readable medium having software instruction stored thereon that, when executed by a processor, cause the processor to perform or control performance of operations. The operations can include scanning one or more stained tissue slides to generate a plurality of tissue slide images including a first tissue slide image of a first tissue slide stained with a first stain and a second tissue slide image of the first tissue slide stained with a second stain, wherein the first stain and the second stain have a common attribute. The operations can also include preprocessing the plurality of tissue slide images, and performing an image registration process on the plurality of tissue slide images, wherein the image registration process is performed on the plurality of tissue slide images using an image of the common attribute in each of the plurality of tissue slide images.
[0017] The operations can further include identifying one or more candidate regions of interest (ROI) within the plurality of tissue slide images, and generating one or more label annotations for structures within the plurality of tissue slide images. In some
implementations, the first stain includes hematoxylin and eosin, and the second stain includes an immunohistochemistry (IHC) marker.
[0018] In some implementations, the preprocessing the plurality of tissue slide images includes separating foreground structures from background structures in the plurality of tissue slide images, and unmixing color of the plurality of tissue slide images to separate components of the first stain and the second stain into different channels. The operations can further include for the tissue slide images stained with the first stain, separating a
hematoxylin channel from an eosin channel, for the tissue slide image stained with the second stain, separating a hematoxylin channel from a DAB channel. [0019] In some implementations, the image registration process includes: minimizing a mutual information metric between deconvolved hematoxylin channels for images of slides stained with the first stain and images of slides stained with the second stain, where the minimizing is performed in two stages: at a first stage, performing a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms; and at a second stage, performing a high resolution registration of localized fields using affine and B-spline transforms. In some implementations, identifying the one or more candidate regions of interest includes: creating a registered image for a same tissue area, wherein the registered image includes the first tissue slide image and the second tissue slide image aligned for the same tissue area.
[0020] The operations can also include providing the tissue slide images and the one or more label annotations for structures within the plurality of tissue slide images to a reviewer system graphical user interface (GUI), and receiving updated labels or updated comments from the reviewer system GUI. The operations can further include storing the updated labels or updated comments with corresponding tissue slide images, and providing labeled tissue slide images as training data to a model.
[0021] Some implementations can include a system having one or more hardware processors configured to perform or control performance of the method mentioned above. In some implementations, the one or more hardware processors may include a graphics processing unit (GPU) with a large number of processing units. In some implementations, the hardware processor may include a neural network processor. The system can include an imaging system.
[0022] Some implementations can include a non-transitory computer readable medium having software instruction stored thereon that, when executed by a processor, cause the processor to perform or control performance of operations according to the method mentioned above.
BRIEF DESCRIPTION OF THE DRAWINGS
[0023] FIG. 1 is a diagram of an example tissue slide image labeling environment in accordance with some implementations.
[0024] FIG. 2 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations. [0025] FIG. 3 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
[0026] FIG. 4 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
[0027] FIG. 5 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
[0028] FIG. 6 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations.
[0029] FIG. 7 is a diagram of an example graphical user interface for tissue slide image labeling review in accordance with some implementations.
[0030] FIG. 8 is a diagram of an example graphical user interface for tissue slide image labeling review in accordance with some implementations.
[0031] FIG. 9 is a diagram of an example graphical user interface for tissue slide image labeling review in accordance with some implementations.
[0032] FIG. 10 is a diagram of an example system configured to perform tissue slide image labeling in accordance with some implementations.
DETAILED DESCRIPTION
[0033] A need may exist for a system that makes this process of generating labeled data less labor intensive and more accurate. Some implementations were conceived in light of the above-mentioned problems, needs, and limitations, among other things.
[0034] Some implementations include a method to label structures of interest in tissue slide images. The method can include staining a specimen on a slide with hematoxylin and eosin (H&E). The H&E stained slide is then imaged with a scanner (e.g., a Philips Ultrafast scanner). The slide is then destained and stained with at least one antibody of interest that is capable of marking the structures of interest (e.g., using an immunohistochemistry or IHC marker technique). After IHC staining is complete, the IHC image is scanned on a scanner. The staining and scanning yield two images: an H&E slide image and an IHC slide image.
[0035] The image of the H&E slide and the corresponding IHC slide are then supplied to an image preprocessing operation. This operation can involve multiple sub-operations. An example first sub-operation is separating foreground from background. After separating foreground from background, color unmixing to separate, into different channels, each of the stains can be done. In the H&E slide, color unmixing can involve separation of the hematoxylin from the eosin. In the case of the IHC slide, color unmixing can include separation of the hematoxylin from the 3,3 '-Diaminobenzi dine (DAB), or other similar IHC stain components.
[0036] Separation of images into color components enables the next operation of processing, registration. Since both images share a hematoxylin channel, this channel can be used to register the images. Multiple registration techniques can be used. For example, mutual information can be used for registration.
[0037] Candidate regions of interest (ROI) can then be extracted from the slide to create a registered H&E and IHC image for the same tissue area. This region of interest extraction or identification can be done manually by a technician or a pathologist by manually selecting a field from the image. The extraction can also be done automatically by a computer in addition to or as an alternative to manual region of interest extraction.
[0038] The automatic computerized region of interest identification can be guided multiple ways, such as by utilizing heuristics. An example can include having a pathologist pre-label the areas of tumor on a slide. The computer system can then randomly sample areas inside the tumor and outside the tumor so that a predetermined ratio of tumor to non-tumor regions can be extracted.
[0039] Additional intelligence can be added to the above process to select regions most likely to contain valuable data. For example, if one were looking for mitotic figures and the user was using a mitotic figure stain such as phosphohistone (PHH3), the system could incorporate information looking for the highest density of PHH3 on the slide. This technique would likely generate the greatest amount of mitotic figures for a given slide image. [0040] Another technique can be based on a machine learning model technique. An example would be training a convolutional neural network (CNN) to select candidate ROIs based upon an available training set of selected and not selected ROIs based on human input.
[0041] Raw IHC images often do not directly translate into binary labels for objects of interest. IHC images are processed in order to produce object labels. This processing can vary depending on the IHC label used. For a cytoplasmic stain that is attempting to delineate specific cellular compartments such as epithelium in the case of cytokeratin 18 (CK18), a segmentation technique such as level set segmentation can be used.
[0042] With the aforementioned operations complete, an efficient process of human review can be undertaken. For example, in an implementation, the user is able to quickly switch between registered H&E and IHC images for the same area of tissue. In FIG. 7 and FIG. 8, respectively, an H&E slide image for a sample of breast tissue is displayed along with an IHC slide image of PHH3 for detecting mitotic figures.
[0043] Some implementations can also provide for a quick multiclass and multilabel system for each ROI. For example, in the graphical user interfaces (GUIs) shown in FIG. 7 and FIG. 8, the user can label an overall triage for the image, including Accept, Reject, or Review. They can also comment on the quality of the ROI and the IHC. In addition, the user can quickly comment if there is no mitosis. Furthermore, for fields displayed in the GUI, the user can enter in freetext comments that allow the user to create new labels on demand.
[0044] Some implementations can include receiving annotations on automatically generated fields using a web-based interface. In some implementations, an expert can confirm, correct, or add information using one or more of the following labels: a) Abnormal Mitotic; b) Mitotic; c) Borderline; d) Regular Nucleus; e) Look Alike; f) Artifact (e.g., hematocyrin); g) Telophase 1; h) Telophase 2; and/or i) Error (e.g., bad staining, algorithm error/bug). Automatic field annotations can be combined with annotation and processed to generate ground truth and validation data. Some implementations can provide near interactive integration of annotations using on-line/incremental learning.
[0045] FIG. 1 is a diagram of an example tissue slide image labeling environment 100 in accordance with some implementations. The environment 100 includes one or more slides 102 (e.g., slides prepared with H&E, an IHC marker, or the like) and one or more scanners 104 arranged to scan the slides 102. The scanners 104 are coupled to a labeling system 106, which is also coupled to one or more reviewer systems 108 that can each include a graphical user interface (GUI) 110.
[0046] The environment 100 also includes one or more labeled images 112 output by the labeling system 106 and provided to a training data store 114 that contains training data for a machine learning model 116 that is trained (or being trained) to perform labeling of structures within tissue slide images. In operation, the labeling system 106 performs tissue slide image labeling of structures of interest (e.g., epithelia and/or stroma, etc.) according to FIGS. 2-6 described below.
[0047] FIG. 2 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations. Processing begins at 202, where one or more slides (e.g., 102) are stained and scanned. For example, a specimen on a slide is stained with hematoxylin and eosin (H&E). The H&E slide is then imaged or scanned with a scanner (e.g., 104) such as a Philips Ultrafast scanner. The H&E slide is then destained and IHC stained with at least one antibody of interest that is capable of marking the structures of interest.
After IHC staining is complete, the IHC slide is scanned on a scanner. Processing continues to 204.
[0048] At 204, the slide images are preprocessed. For example, the H&E slide image and the corresponding IHC slide image are preprocessed (e.g., using the labeling system 106), which can include multiple sub-operations. An example first sub-operation is separating foreground from background. After separating foreground from background, color unmixing is performed to separate into different channels each of the stains can be done. In the H&E slide, this would involve separation of the hematoxylin from the eosin. In the case of the IHC slide, this would involve separation of the hematoxylin from the DAB (or other similar IHC stain components).
[0049] In some implementations, efficient analysis of large whole slide images can be conducted by isolating relevant tissue from background. This isolation or separation can be accomplished by measuring textural content through derivative filters, such as a Laplacian filter. Analysis can be constrained to regions with Laplacian values larger than the average over the image. Structural signatures can be detected by separating or deconvolving the stains from the color images. Relevant signatures include hematoxylin stain channel images from H&E and IHC for registering the slide pairs, and structure-specific stains in IHC for generating annotations. Processing continues to 206.
[0050] At 206, image registration is performed to align the whole slide H&E images and the IHC images. Separation of images into color components helps enable registration.
Because both images share a hematoxylin channel, the hematoxylin channel image can be used to register the images. Multiple registration technologies can be used. In some implementations, mutual information can be used for registration. In some implementations, registration of H&E and IHC images can be performed by maximizing the mutual information metric between the deconvolved hematoxylin channels, and performed in two stages. The first stage can include a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms. The second stage can include a high resolution registration of localized H&E and IHC fields using affine and B-spline transforms, which can be performed at a whole slide level. Processing continues to 208.
[0051] At 208, candidate regions of interest are generated or identified within the slide images. For example, candidate regions can be extracted from the slide images to create a registered H&E and IHC image for the same tissue area. In some implementations, region (or field) extraction can be done manually by a technician or a pathologist manually selecting a field from the image. In some implementations, region extraction can be done automatically by a computer system (e.g., labeling system 106).
[0052] The computerized region extraction can be guided multiple ways. For example, one technique may include utilizing heuristics. For example, a pathologist could prelabel areas of tumor on a slide. The labeling system would then randomly sample areas inside the tumor and outside the tumor so that a predetermined ratio of tumor/non-tumor regions would be extracted. Additional intelligence or program logic can be added to this process to select regions most likely to contain valuable data. For example, if one were looking for mitotic figures and the user was using a mitotic figure stain such as PHH3, the algorithm could incorporate information looking for the highest density of PHH3 on the slide, which would emphasize generating mitotic figures. Yet another algorithm could be based on a machine learning technique (e.g., using model 116 and training data 114). An example would be training a CNN (e.g., 116) to select regions of interest based upon an available training set (e.g., 114) of selected and not selected regions of interest based on human input. Additional detail about identifying and extracting regions of interest is described below in connection with FIG. 3. Processing continues to 210.
[0053] At 210, label annotations are automatically generated. For example, raw IHC images often do not directly translate into binary labels for objects of interest. IHC images are processed in order to produce object labels. This processing can vary depending on the IHC label used. For a cytoplasmic stain that is attempting to delineate specific cellular compartments such as epithelium in the case of CK18, a segmentation technique such as level set segmentation can be used to segment the image and provide areas for labeling. Labeling of the segments can be provided by a machine learning model (e.g., 116).
[0054] Once candidate regions of interest are detected in the H&E slide, the system can automatically generate segmentation masks for the local fields. These segmentation masks can serve as a starting point for manual correction, reducing the workload for human reviewers. The system also registers the corresponding IHC ROI using a higher order model (e.g., B-splines). Human reviewers are presented with the H&E ROI, IHC ROI, and the automatically generated masks when performing the annotation.
[0055] In some implementations, masks can be generated by applying signal
enhancement and suppression followed by image thresholding to select desired structures. Mitotic figure masks can be generated by enhancing blobby structures and suppressing ridges using the following equation: s(x) = s(x) + (blob(s(x)) - ridge(s(x)) * epsilon
[0056] Cytokeratin masks can be generated by enhancing regions with high nearby hematoxylin and suppressing ridges using the following equation: s(x) = s(x) + [expand(s(x)) * hematoxylin(x) - ridge(s(x))] * epsilon
[0057] In some implementations, the final mask for epithelium can be generated using the signal enhancement followed by clustering using Gabor feature filters, and random walker segmentation.
[0058] Additional details about a segmentation implementation are described below in connection with FIG. 4. Processing continues to 212. [0059] At 212, the automatically labeled images can optionally be provided to a reviewer system (e.g., 108) and displayed within a GUI (e.g., 110) and as shown in FIGS. 7-9 and described below. The GUI can display the image(s), the automatically generated labels and structural identifications within the image and can provide an interface for a reviewer to add, delete or modify labels and/or comments corresponding to the image(s) being displayed. Processing continues to 214.
[0060] At 214, the labeling system can receive updated image labels and comments from the reviewer system GUI. Once the automatically generated masks have been corrected and reviewed by a clinical expert, the system generates image data with the associated labels for building models through the use of machine learning techniques.
[0061] Cell annotations, such as the one for mitotic figures, can be performed in multiple stages. At the first stage, the generated regions of interest are reviewed by non-clinicians for obvious errors, and possible mitotic figures are marked (e.g., as solid dark blue cells). In the second stage local image patches are presented at the marked locations to clinical experts for cell categorization. The third stage involves the outlining of the cells marked as mitotic or borderline mitotic. The per-pixel labeling of mitotic figures can be used to sample local patches for machine learning.
[0062] Boundary annotations, such as the one for epithelium, can be performed by presenting a non-clinician with the H&E and IHC images outlined using the technique described above. Non-clinicians could correct these outlines by drawing or erasing boundaries, they can also correct by using a paint filling method. To handle slides with significant variability of image appearance due to the staining process, the system may apply stain normalization to standardize the H&E image samples. Processing continues to 216.
[0063] At 216, the updated labels and comments are stored with the corresponding image(s). Processing continues to 218.
[0064] At 218, the labeled image(s) with updated labels and comments are optionally provided to a machine learning system for training and/or feedback. The model can include a neural network such as a convolutional neural network or fully convolutional neural network. A neural network can include a hierarchical composite of multiple primary processing units called neurons. Contrary to the conventional multiple-layer perceptions (MLPs), where each neuron is directly collected to all neurons in a previous layer, a convolutional neural network (CNN) assumes the input as images and exploits spatially local correlation by enforcing a local connectivity pattern between neurons of adjacent layers.
[0065] The trained CNN can then be applied in multiple ways. In a basic trained form, the trained CNN can be used on a given ROI to detect the likely mitotic figure candidates. Also, it is possible to use the trained CNN as a hotspot detector over the whole slide where the most likely candidates are located over the entire slide. Additional training or network modification may be needed for the trained CNN to operate on an entire slide, as the trained data set typically excludes extraneous and non-tumor tissue on the slide.
[0066] Blocks 202-218 (as well as any other blocks shown and described herein) can be repeated in whole or in part and performed in the same or different orders. Moreover, certain blocks can be eliminated, modified, combined with other blocks, and/or supplemented with other operations.
[0067] FIG. 3 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations. Processing begins at 302, where the H&E ROI is read (e.g., as a 2000 x 2000 pixel section). Processing continues to 304.
[0068] At 304, a large IHC ROI is extracted based on the H&E ROI. For example, assuming that IHC and HE whole slides are rigidly aligned in a lower resolution, the position of the H&E ROI (in the highest resolution) can be read and a (4000 x 4000 pixel) IHC ROI extracted from the whole slide. The H&E ROI is roughly located in the center of the IHC ROI now. Processing continues to 306.
[0069] At 306, a common channel (e.g., hematoxylin) is extracted from both the H&E ROI and IHC ROI. Processing continues to 308.
[0070] At 308, the common channels of H&E and IHC images are rigidly aligned. For example, the IHC hematoxylin channel image (with bigger ROI) is rigidly aligned with the hematoxylin H&E channel image (with smaller ROI). Registration can be performed in a three-level multi-resolution approach (4, 2, and 1 times the actual spacing of the image) initialized by geometrical centered transform. Processing continues to 310.
[0071] At 310, a free-form registration between the IHC ROI and the H&E ROI is performed. For example, a free-form registration between IHC hematoxylin channel image and H&E hematoxylin channel image is initialized by the rigid transformation obtained in 308. A three-level, multi-resolution b-spline based registration can be used for deformation compensation between the IHC and H&E ROIs. In some implementations, a mutual information metric can be used that measures statistical similarity between the information content between images in both rigid and free-form registration. In particular, some implementations can include using the efficient Mattes Mutual Information metric that performs binning of image intensities and random sampling of pixel data. (See, D. Mattes, D. R. Haynor, H. Vesselle, T. K. Lewellen, and W. Eu-bank,“PET-CT image registration in the chest using free-form deformations,” IEEE Trans. Med. Imag., vol. 22, no. 1, pp. 120-128, Jan. 2003, which is incorporated herein by reference). Processing continues to 312.
[0072] At 312, after registration, a smaller IHC ROI (e.g., 2000 x 2000) corresponding to the H&E ROI is extracted and saved.
[0073] FIG. 4 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations. Processing begins at 402, where a mask is created (e.g., an ini mask). For example, the mask can provide an estimation of epithelial cells and separate the background from the tissue. For example, the mask can be created from the H&E hematoxylin image and the IHC DAB image using a signal enhancement approach, where known properties (e.g.,“blobbiness” of a cell, smooth contiguous regions for glands, etc.) can be enhanced. The mask can be dilated to give a rough estimation of epithelium and separate the background from the tissue. Processing continues to 404.
[0074] At 404, image features are extracted. For example, image features can be extracted from one or more smaller IHC ROI images and the features can include:
[0075] 1) H&E channel values in the H&E image (e.g., H&E hematoxylin and H&E eosin channels)
[0076] 2) IHC hematoxylin and DAB channels
[0077] 3) Local binary patterns computed for hemo HE, eosin HE, hemo IHC, and dab IHC. (See, T. Ojala, M. Pietikainen, and D. Harwood (1996), "A Comparative Study of Texture Measures with Classification Based on Feature Distributions", Pattern Recognition, vol. 29, pp. 51-59, which is incorporated herein by reference). [0078] 4) Texture features based on Gabor filter banks. (See, Grigorescu, S.E., Petkov, N. and Kruizinga, P., 2002. Comparison of texture features based on Gabor filters. IEEE Transactions on Image processing, 11(10), pp. 1160-1167, which is incorporated herein by reference).
[0079] The above features can be computed for the hemo HE, eosin HE, hemo IHC, and/or dab_IHC in one frequency and three different rotation angles (e.g., 0, 45, and/or 135 degrees). In some implementations, a total number of 20 features are computed for each pixel, constructed from different combinations of derivative values in different angular directions and at different spatial scales. These features capture patterns such as peaks and valleys in different directions and different sizes, enabling registration based on cellular boundaries and chromatin content. Processing continues to 406.
[0080] At 406, image pixels are clustered based on image features. For example, pixels in the ROI can be assigned to one of three clusters (e.g., nuclei (or cells), tissue, and background) using k-means clustering approach. Processing continues to 408.
[0081] At 408, a mask (e.g., an epi-stroma mask) is created. In some implementations, the epi-stroma mask can include identification of epithelium and stroma regions within one or more images, which can be useful because areas of cancer are sometimes associated with epithelium. In some implementations, the mask can be generated using a random walker technique, which is a semi-supervised segmentation method that involves a number of pixels with known labels (called seeds), can be used to perform image segmentation. The random walk process starts from these seeds.
[0082] The clustering process in the previous operation basically divides the pixels to three groups: 1) nuclei (cells); 2) tissue; and 3) background. However, the groups are not labeled. The system first determines the background cluster, which can include the cluster having the least overlap (correlation) with the ini_mask (created at 402). Then, for the remaining two clusters (label_l and label_2) (e.g., nuclei and tissue), the system may consider pixels that lay within the ini mask. Two different masks can be created using the random walker segmentation. In the first mask, an eroded version of label l can be used for seeding and a dilated version of label_2 can be used as unknown background. In the second mask, labels can be switched. From the created output masks, the mask having larger average dab IHC in the overlapping pixels is chosen as the output epi-stroma mask, which can be output as an initial ground truth reference that can be utilized in downstream processing and/or corrected by user in some implementations.
[0083] Some implementations can include a software application for specialized medical equipment (e.g., slide scanning systems, etc,), desktop computers, laptop computers, and mobile devices (e.g., smartphones, tablet computing devices, etc.) that can be used by physicians and/or other health care or biomedical professionals to perform tissue slide image structure labeling. The software application can also be provided as a web service accessible over a wired or wireless computer network.
[0084] In some implementations, the method, or portions of the method, can be initiated automatically by a device. For example, the method (or portions thereol) can be repeatedly performed or performed based on the occurrence of one or more particular events or conditions. For example, such events or conditions can include: obtaining one or more images that have been newly captured by, uploaded to, or otherwise accessible by a device (e.g., an imaging device such as a slide scanner as described herein), a predetermined time period having expired since the last performance of method 200, 300, and/or 400, and/or one or more other events or conditions occurring which can be specified in settings of a device implementing method 200, 300, and/or 400. In some implementations, such conditions can be previously specified by a user in stored custom preferences of the user (accessible by a device or method with user consent). In another example, an imaging device (e.g., slide scanner) or other medical system can capture one or more images and can perform the method 200, 300, and/or 400. In addition, or alternatively, an imaging device can send one or more captured images to an external system (e.g., a cloud computing system, a server, a mobile device, etc.) over a network, and the external system can process the images using method 200, 300, and/or 400.
[0085] FIG. 5 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations. Processing begins at 502, where one or more tissue slides are H&E stained and scanned as described herein. Processing continues to 504.
[0086] At 504, the slide(s) are destained. Processing continues to 506.
[0087] At 506, the slides are re-stained with an IHC marker and scanned. Processing continues to 508. [0088] At 508, the hematoxylin channel image is extracted from IHC. Processing continues to 510.
[0089] At 510, registration between the H&E image and the IHC image using the hematoxylin channel is performed. Processing continues to 512.
[0090] At 512, automated ROI extraction and labeling is performed according to one or more techniques described herein. Processing continues to 514.
[0091] At 514, an epi-stroma mask is generated.
[0092] FIG. 6 is a flowchart of an example method for tissue slide image labeling in accordance with some implementations. Processing begins at 602, where one or more tissue slides are H&E stained and scanned as described herein. Processing continues to 604.
[0093] At 604, the slide(s) are destained. Processing continues to 606.
[0094] At 606, the slides are re-stained with an IHC marker and scanned. Processing continues to 608.
[0095] At 608, the hematoxylin channel is extracted from the IHC image. Processing continues to 610.
[0096] At 610, registration between the H&E image and the IHC image using the hematoxylin channel is performed. Registration can be performed as described herein. Processing continues to 612.
[0097] At 612, candidate field extraction (e.g., region of interest) and labeling using a mask (e.g., a PHH3 mask) is performed. For example, in some implementations, the PHH3 mask can be created from regions marked brown in IHC that are constrained to be small and cell-like. These can be used as a marker of possible mitotic figures for review by experts (either a human expert or an expert system). Processing continues to 614.
[0098] At 614, optionally, updated labels and/or comments associated with the candidate fields and/or labels can be received from a reviewer system GUI. Processing continues to 616. [0099] At 616, optionally, the mask (e.g., PHH3 mask), labels and/or labeled objects can be provided as training data for a machine learning model, e.g., a convolutional neural network (CNN) model.
[0100] FIGS. 7-9 include diagrams of example graphical user interfaces (GUIs) that show example modes for navigating between images and performing annotations. For example, FIGS. 7 and 8 are diagrams of a GUI for an overview/browsing mode and FIG. 9 is a diagram of a GUI for a mitotic figure annotating mode. User can triage fields as accept, reject, or review using butons on top right pane. Reject is for ROIs with bad quality or artifact like tissue being tom. Review is for future review between different experts. Users can provide input to the GUI, e.g., via keyboard up and down keys or other input, to move backward and forward between different ROIs.
[0101] In some implementations, the user is able to quickly switch between registered H&E and IHC images for the same area of tissue. In FIGS. 7 and 8, H&E for a sample of breast tissue is displayed along with an IHC of PHH3 for detecting mitotic figures.
[0102] FIG. 8 is a diagram of an example graphical user interface (GUI) for tissue slide image labeling review that shows a mode where the registered IHC image is displayed and where users can access and select images for display (e.g., by clicking the pull down menu on top left pane and selecting the desired image to display). Experts can use the IHC images to confirm if a given set of H&E pixels contain a certain structure of interest because the IHC images are stained for specific targets (e.g., cytokeratin or mitotic in brown).
[0103] The system also provides for a quick multiclass and multilabel system for each ROI. In FIGS. 7 and 8 the user can label an overall triage for the image, including Accept, Reject, or Review. They can also comment on the quality of the ROI and the IHC. In addition, the user can quickly comment if there are no mitosis. Furthermore, for each of the fields, the user can enter in freetext comments that allow the user to create new labels on demand.
[0104] FIG. 9 is a diagram of an example graphical user interface (GUI) for tissue slide image labeling review in accordance with some implementations. For example, via the GUI shown in FIG. 9, a user can drop annotation objects by picking a classification label on botom left of screen and then click on a cell in the field of view to tag it with different labels, which can be shown as different colored rectangles. [0105] In FIG. 9, the user is able to see in a yellow outline the results of IHC
segmentation overlaid onto the H&E. This allows the user to see, based on IHC
segmentation, the corresponding area on the H&E. In this case, the IHC is of PHH3 which indicates mitotic figures. Therefore, the yellow area on the H&E would indicate where a mitotic figure is.
[0106] Furthermore, the system permits the user to assign multiple classes to various objects in the image. In this example, allowed classes include: a) Abnormal mitotic; b) Mitotic; c) Borderline (mitotic); d) Regular nucleus; d) Look alike; e) Artifact; 1) Telophase 1 (one nucleus of a cell dividing); g) Telophase 2 (the other nucleus of a dividing cell); h) Prophase; and/or g) Error.
[0107] With the IHC segmentation as a guide, the user can then assign these subclasses to the detected object. Furthermore, the user may also annotate additional objects that are in the scene that were either not detected in error or are not intended to be detected by the IHC but are of utility, e.g., a regular nucleus, look alike, etc.
[0108] In another embodiment, such as when ckl8 is used on breast tissue for marking epithelial regions, a rich set of ROI level labels can be assigned by the user for future subclassing. Examples include: a) Marker present - indicating that there is marker present on the ROI; b) Border segmentation issues; c) Registration mismatch - the IHC and the H&E do not register - this may be due to faulty registration or differences in the tissue such as one having a tear in it; d) Lumen staining - a lumen is picking up stain where it should not be; and/or e) Weak epi - the epithelial regions are weakly staining
[0109] FIG. 10 is a block diagram of an example computing device 1000 which may be used to implement one or more features described herein. In one example, computing device 1000 may be used to implement a computer device, e.g., tissue slide labeling system or reviewer system device (e.g., 106 and/or 108 of FIG. 1), and perform appropriate method implementations described herein. Device 1000 can be any suitable computer system, server, or other electronic or hardware device. For example, the device 1000 can be a mainframe computer, desktop computer, workstation, portable computer (including a smart phone or laptop), or medical device. In some implementations, device 1000 includes a processor 1002, a memory 1004, and input/output (I/O) interface 1006, all operatively coupled to each other. [0110] Processor 1002 can be one or more processors and/or processing circuits to execute program code and control basic operations of the device 1000. A“processor” includes any suitable hardware and/or software system, mechanism or component that processes data, signals or other information. A processor may include a system with a general-purpose central processing unit (CPU), multiple processing units, dedicated circuitry for achieving functionality, or other systems. Processing need not be limited to a particular geographic location, or have temporal limitations. For example, a processor may perform its functions in“real-time,”“offline,” in a“batch mode,” etc. Portions of processing may be performed at different times and at different locations, by different (or the same) processing systems. A computer may be any processor in communication with a memory.
[0111] Memory 1004 is typically provided in device 1000 for access by the processor 1002, and may be any suitable processor-readable storage medium, e.g., random access memory (RAM), read-only memory (ROM), electrical erasable read-only memory
(EEPROM), flash memory, etc., suitable for storing instructions for execution by the processor, and located separate from processor 1002 and/or integrated therewith. Memory 1004 can store software operating on the server device 1000 by the processor 1002, including an operating system 1008, one or more applications 1010, e.g., a tissue slide image labeling application and application data 1020. In some implementations, applications 1010 can include instructions that, in response to execution by processor 1002, enable processor 1002 to perform or control performance of the functions described herein, e.g., some or all of the methods of FIGS. 2-6.
[0112] For example, applications 1010 can include a tissue slide image labeling application, which as described herein can provide automatic labels for tissue slide images. Any software in memory 1004 can alternatively be stored on any other suitable storage location or computer-readable medium. In addition, memory 1004 (and/or other connected storage device(s)) can store images, and other instructions and data used in the features described herein. Memory 1004 and any other type of storage (magnetic disk, optical disk, magnetic tape, or other tangible media) can be considered "storage" or "storage devices."
[0113] In various implementations, application(s) 1010 can include a machine-learning model (e.g., a two-stage neural network model including a coarse detector and a fine discriminator) that can be trained to detect cell objects in images. In some implementations, training data may include tissue slide scan data, CT image data, etc. [0114] The model form or structure may specify connectivity between various nodes and organization of nodes into layers. For example, nodes of a first layer (e.g., input layer) may receive data as input data or application data. Such data can include, for example, one or more pixels. Subsequent intermediate layers may receive as input output of nodes of a previous layer per the connectivity specified in the model form or structure. These layers may also be referred to as hidden layers. A final layer (e.g., output layer) produces an output of the machine-learning application. For example, the output may be an identified and/or labeled candidate object or particular object of interest, etc. depending on the specific trained model.
[0115] In some implementations, the trained model may include a weight individual nodes and/or connections. A respective weight may be applied to a connection between each pair of nodes that are connected per the model form, e.g., nodes in successive layers of the neural network. For example, the respective weights may be randomly assigned, or initialized to default values. The model may then be trained, e.g., using data, to produce a result, where the training can include adjusting one or more of nodes, node structure, connections, and/or weights.
[0116] A model can include a loss function representing the difference between a predicted value and an actual label. The model can be trained to minimize the loss function. Training can include supervised, unsupervised, or semi-supervised learning techniques. In supervised learning, the training data can include a plurality of inputs (e.g., tissue slide images) and a corresponding expected output for each input (e.g., tissue slide images with objects or structures of interest identified and labeled). Based on a comparison of the output of the model with the expected output (e.g., computing the loss function), values of the weights are automatically adjusted, e.g., in a manner that increases a probability that the model produces the expected output when provided similar input (i.e., reduces the loss function). In unsupervised learning, models leam relationships between elements in a data set and classify raw data without the benefit of labeled training data. Semi-supervised learning can include a combination of supervised and unsupervised techniques, for example, a small amount of labeled training data and a large amount of unlabeled training data can be provided to a model for learning.
[0117] I/O interface 1006 can provide functions to enable interfacing the computing device 1000 with other systems and devices. For example, network communication devices, storage devices (e.g., memory and/or database), and input/output devices can communicate via the I/O interface 1006. In some implementations, the I/O interface 1006 can connect to interface devices including input devices (keyboard, pointing device, touchscreen, microphone, camera, scanner, etc.) and/or output devices (display device, speaker devices, printer, motor, etc.). Display device 1014 is one example of an output device that can be used to display images and other data, e.g., one or more tissue slide images, structure of interest identification, and labels provided by an application as described herein. Display device 1014 can be connected to device 1000 via local connections (e.g., display bus) and/or via networked connections and can be any suitable display device, some examples of which are described below.
[0118] For ease of illustration, FIG. 10 shows one block for each of processor 1002, memory 1004, I/O interface 1006, and software blocks 1008 and 1010. These blocks may represent one or more processors or processing circuitries, operating systems, memories, I/O interfaces, applications, and/or software modules. In other implementations, device 1000 may not have all of the components shown and/or may have other elements including other types of elements instead of, or in addition to, those shown herein. While labeling system 106 is described as performing operations as described in some implementations herein, any suitable component or combination of components of system 106 or similar system, or any suitable processor or processors associated with such a system, may perform the operations described.
[0119] A user device can also implement and/or be used with features described herein. Example user devices can be computer devices including some similar components as the device 1000, e.g., processor(s) 1002, memory 1004, and I/O interface 1006. An operating system, software and applications suitable for the client device can be provided in memory and used by the processor. The I/O interface for a client device can be connected to network communication devices, as well as to input and output devices, e.g., a microphone for capturing sound, a camera for capturing images or video, audio speaker devices for outputting sound, a display device for outputting images or video, or other output devices. A display device 1014, for example, can be connected to (or included in) the device 1000 to display images pre- and post-processing as described herein, where such display device can include any suitable display device, e.g., an LCD, LED, or plasma display screen, CRT, television, monitor, touchscreen, 3-D display screen, projector, or other visual display device. Some implementations can provide an audio output device, e.g., voice output or synthesis that speaks text.
[0120] One or more methods described herein (e.g., the methods shown in FIGS. 2-6) can be implemented by computer program instructions or code, which can be executed on a computer. For example, the code can be implemented by one or more digital processors (e.g., microprocessors or other processing circuitry), and can be stored on a computer program product including a non-transitory computer readable medium (e.g., storage medium), e.g., a magnetic, optical, electromagnetic, or semiconductor storage medium, including
semiconductor or solid state memory, magnetic tape, a removable computer diskette, a random access memory (RAM), a read-only memory (ROM), flash memory, a rigid magnetic disk, an optical disk, a solid-state memory drive, etc. The program instructions can also be contained in, and provided as, an electronic signal, for example in the form of software as a service (SaaS) delivered from a server (e.g., a distributed system and/or a cloud computing system). Alternatively, one or more methods can be implemented in hardware (logic gates, etc.), or in a combination of hardware and software. Example hardware can be programmable processors (e.g. field-programmable gate array (FPGA), complex programmable logic device), general purpose processors, graphics processing units (or GPUs) application specific integrated circuits (ASICs), and the like. One or more methods can be performed as part of or component of an application running on the system, or as an application or software running in conjunction with other applications and operating system.
[0121] One or more methods described herein can be run in a standalone program that can be run on any type of computing device, a program run on a web browser, a mobile application (“app”) run on a mobile computing device (e.g., cell phone, smart phone, tablet computer), wearable device. In one example, a client/server architecture can be used, e.g., a mobile computing device (as a client device) sends user input data to a server device and receives from the server the final output data for output (e.g., for display). In another example, all computations can be performed within the mobile app (and/or other apps) on the mobile computing device. In another example, computations can be split between the mobile computing device and one or more server devices.
[0122] Although the description has been described with respect to particular implementations thereof, these particular implementations are merely illustrative, and not restrictive. Concepts illustrated in the examples may be applied to other examples and implementations. For example, while breast cancer diagnosis or image studies are described as an example to illustrate an implementation of the disclosed subject matter, the system and methods described herein can be extended to image studies of other cancer sites, such as liver, lung, head, neck, etc. or other types of biomedical imaging for other purposes or for diagnosing other diseases.
[0123] Note that the functional blocks, operations, features, methods, devices, and systems described in the present disclosure may be integrated or divided into different combinations of systems, devices, and functional blocks. Any suitable programming language and programming techniques may be used to implement the routines of particular implementations. Different programming techniques may be employed, e.g., procedural or object-oriented. The routines may execute on a single processing device or multiple processors. Although the steps, operations, or computations may be presented in a specific order, the order may be changed in different particular implementations. In some implementations, multiple steps or operations shown as sequential in this specification may be performed at the same time.
Appendix A - Example Lab Protocols for Preparing Tissue
H & E Staining
[0123] Cut paraffin sections (e.g., 4-5 micron thick) and place on positive (+) charged slides. Label slides, using graphite pencil with accession number (corresponding to block number).
[0124] Bake slides at 65°C for 60 min
[0125] Deparaffmize in 3 changes of xylene, for 3 minutes each.
[0126] Hydrate in descending grades of alcohol as follows: a. 100%, 3 changes, 10 dips in each. b. 95%, 3 changes, 10 dips in each.
[0127] Wash in tap water for 1 minute.
[0128] Stain with Harris Hematoxylin (e.g., Fisher, cat# SH30-500D), or the like, 7 minutes; wash in tap water- 1 minute.
[0129] Place slides in acid-alcohol - 1-2 dips, then wash in tap water for 1 min.
[0130] Blue in Ammonia water for 30-60 sec. Wash in tap water for 1 min.
[0131] Dip slides in 95 % alcohol.
[0132] Place slides in Alcoholic Eosin solution (e.g., Sigma, cat# HT110116, or the like) for 2 minutes.
[0133] Dehydrate as follows: a. 95% dehydrant -3 changes, 10 dips each b. 100% dehydrant -3 changes, 10 dips each [0134] Clear in 3 changes of xylene, 1 minute each
[0135] Mount [0136] Dry slides overnight.
H&E Destaining
[0137] Scan slides on Philips scanner before H&E destaining and IHC procedure
[0138] After slides are scanned, place in clean Xylene solution and keep for 1-2 hrs to remove coverslips. Gently lift coverslips from the slides. Dip in xylene and rehydrate in descending grades of alcohol: a. 100%, 2 changes, 10 dips each b. 95%, 2 changes, 10 dips each
[0139] Water, 3 changes, 1 minute each
[0140] Place in PBS
[0141] Destain H&E slides by proceeding directly to IHC staining procedure on Bond RX instrument. Bond Wash buffer and other reagents used during IHC procedure will remove H&E stain from slides. Immunohistochemistry (IHC) for Anti-Histone H3 (phospho S10 + Ti l) antibody (e.g., Abeam, cat# ab32107), using Bond Polymer Refine Detection System (e.g., Leica, DS9800).
[0142] Anti -histone H3 (PHH3) is a rabbit polyclonal antibody, specific for Histone 3 phosphorylated on SerinelO and Threonine 11. Phosphorylation of H3, especially at highly conserved SerinelO site is crucial for the onset of mitosis. PHH3 staining is useful to distinguish mitotic figures from other morphological changes that can sometimes mimic mitosis, such as apoptosis or necrosis.
[0143] The Bond Polymer Refine Detection system used at PRECISE MD lab detects tissue- bound mouse and rabbit IgG and some mouse IgM primary antibodies. The detection kit is intended for staining sections of formalin-fixed, paraffin-embedded tissues on the Bond automated system. It utilizes a novel controlled polymerization technology to prepare polymeric horseradish peroxidase (HRP) linker antibody and avoids the use of streptavidin and biotin, which can cause non-specific background due to the presence of endogenous biotin. The system gives great sensitivity and signal amplification, as well as reduces or eliminates the need of including negative controls with every run. [0144] Place labeled slides on a slide tray, and cover with cover tiles.
[0145] Reagent tray: Bond Polymer Refined Detection Kit (Leica); primary antibody, PHH3 (Abeam) diluted 1 :2000 with 2% BSA in PBS buffer (prepared in the laboratory).
[0146] Antigen Retrieval step: ER1, pH6 buffer (Leica, Cat# AR9961) 40 min
[0147] Peroxidase block- 5 min
[0148] Primary antibody (Abeam, PHH3, 1:2000 dilution ), 60 minutes
[0149] Post Primary (to enhance the penetration of subsequent polymer reagent) - 8 min
[0150] Polymer (poly-HPR anti mouse/rabbit polymer) - 8 min
[0151] DAB (chromogen) - 10 min
[0152] Hematoxylin- 5 min
[0153] When run is finished, unload slides and dehydrate , clear and cover slip:
a. 95% dehydrant -3 changes, 10 dips each
b. 100% dehydrant -3 changes, 10 dips each
[0154] Xylene - 3 changes, 1 min each
[0155] Mount slides with Cytoseal XYL
[0156] Dry slides overnight
[0157] Scan slides

Claims

CLAIMS What is claimed is:
1. A computer-implemented method to label structures in tissue slide images, the computer- implemented method comprising:
scanning one or more stained tissue slides to generate a plurality of tissue slide images including a first tissue slide image of a first tissue slide stained with a first stain and a second tissue slide image of the first tissue slide stained with a second stain, wherein the first stain and the second stain have a common attribute;
preprocessing the plurality of tissue slide images;
performing an image registration process on the plurality of tissue slide images, wherein the image registration process is performed on the plurality of tissue slide images using an image of the common attribute in each of the plurality of tissue slide images;
identifying one or more candidate regions of interest (ROI) within the plurality of tissue slide images; and
generating one or more label annotations for structures within the plurality of tissue slide images.
2. The computer-implemented method of claim 1, wherein:
the first stain includes hematoxylin and eosin; and
the second stain includes an immunohistochemistry (IHC) marker.
3. The computer-implemented method of claim 1, wherein the preprocessing the plurality of tissue slide images includes:
separating foreground structures from background structures in the plurality of tissue slide images; and
unmixing color of the plurality of tissue slide images to separate components of the first stain and the second stain into different channels.
4. The computer-implemented method of claim 3, further comprising:
for the tissue slide images stained with the first stain, separating a hematoxylin channel from an eosin channel, for the tissue slide image stained with the second stain, separating a hematoxylin channel from a DAB channel.
5. The computer-implemented method of claim 1, wherein the image registration process includes:
minimizing a mutual information metric between deconvolved hematoxylin channels for images of slides stained with the first stain and images of slides stained with the second stain, where the minimizing is performed in two stages:
at a first stage, performing a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms; and
at a second stage, performing a high resolution registration of localized fields using affine and B-spline transforms.
6. The computer-implemented method of claim 1, wherein identifying the one or more candidate regions of interest includes:
creating a registered image for a same tissue area, wherein the registered image includes the first tissue slide image and the second tissue slide image aligned for the same tissue area.
7. The computer-implemented method of claim 1, further comprising:
providing the tissue slide images and the one or more label annotations for structures within the plurality of tissue slide images to a reviewer system graphical user interface (GUI);
receiving updated labels or updated comments from the reviewer system GUI; storing the updated labels or updated comments with corresponding tissue slide images; and
providing labeled tissue slide images as training data to a model.
8. A system, comprising:
one or more processors coupled to a computer readable memory having stored thereon software instructions that, when executed by the one or more processors, cause the one or more processors to perform or control performance of operations including: scanning one or more stained tissue slides to generate a plurality of tissue slide images including a first tissue slide image of a first tissue slide stained with a first stain and a second tissue slide image of the first tissue slide stained with a second stain, wherein the first stain and the second stain have a common attribute; preprocessing the plurality of tissue slide images;
performing an image registration process on the plurality of tissue slide images, wherein the image registration process is performed on the plurality of tissue slide images using an image of the common attribute in each of the plurality of tissue slide images;
identifying one or more candidate regions of interest (ROI) within the plurality of tissue slide images; and
generating one or more label annotations for structures within the plurality of tissue slide images.
9. The system of claim 8, wherein:
the first stain includes hematoxylin and eosin; and
the second stain includes an immunohistochemistry (IHC) marker.
10. The system of claim 8, wherein the preprocessing the plurality of tissue slide images includes:
separating foreground structures from background structures in the plurality of tissue slide images; and
unmixing color of the plurality of tissue slide images to separate components of the first stain and the second stain into different channels.
11. The system of claim 10, wherein the operations further comprise:
for the tissue slide images stained with the first stain, separating a hematoxylin channel from an eosin channel, for the tissue slide image stained with the second stain, separating a hematoxylin channel from a DAB channel.
12. The system of claim 8, wherein the image registration process includes:
minimizing a mutual information metric between deconvolved hematoxylin channels for images of slides stained with the first stain and images of slides stained with the second stain, where the minimizing is performed in two stages:
at a first stage, performing a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms; and
at a second stage, performing a high resolution registration of localized fields using affine and B-spline transforms.
13. The system of claim 8, wherein identifying the one or more candidate regions of interest includes:
creating a registered image for a same tissue area, wherein the registered image includes the first tissue slide image and the second tissue slide image aligned for the same tissue area.
14. The system of claim 8, wherein the operations further comprise:
providing the tissue slide images and the one or more label annotations for structures within the plurality of tissue slide images to a reviewer system graphical user interface (GUI);
receiving updated labels or updated comments from the reviewer system GUI; storing the updated labels or updated comments with corresponding tissue slide images; and
providing labeled tissue slide images as training data to a model.
15. A non-transitory computer readable medium having software instruction stored thereon that, when executed by a processor, cause the processor to perform or control performance of operations including: scanning one or more stained tissue slides to generate a plurality of tissue slide images including a first tissue slide image of a first tissue slide stained with a first stain and a second tissue slide image of the first tissue slide stained with a second stain, wherein the first stain and the second stain have a common attribute;
preprocessing the plurality of tissue slide images;
performing an image registration process on the plurality of tissue slide images, wherein the image registration process is performed on the plurality of tissue slide images using an image of the common attribute in each of the plurality of tissue slide images;
identifying one or more candidate regions of interest (ROI) within the plurality of tissue slide images; and
generating one or more label annotations for structures within the plurality of tissue slide images.
16. The non-transitory computer readable medium of claim 15, wherein:
the first stain includes hematoxylin and eosin; and the second stain includes an immunohistochemistry (IHC) marker.
17. The non-transitory computer readable medium of claim 15, wherein the preprocessing the plurality of tissue slide images includes:
separating foreground structures from background structures in the plurality of tissue slide images; and
unmixing color of the plurality of tissue slide images to separate components of the first stain and the second stain into different channels.
18. The non-transitory computer readable medium of claim 17, wherein the operations further comprise:
for the tissue slide images stained with the first stain, separating a hematoxylin channel from an eosin channel, for the tissue slide image stained with the second stain, separating a hematoxylin channel from a DAB channel.
19. The non-transitory computer readable medium of claim 15, wherein the image registration process includes:
minimizing a mutual information metric between deconvolved hematoxylin channels for images of slides stained with the first stain and images of slides stained with the second stain, where the minimizing is performed in two stages:
at a first stage, performing a coarse, low-resolution registration of hematoxylin channels using global rigid and affine transforms; and
at a second stage, performing a high resolution registration of localized fields using affine and B-spline transforms.
20. The non-transitory computer readable medium of claim 15, wherein identifying the one or more candidate regions of interest includes:
creating a registered image for a same tissue area, wherein the registered image includes the first tissue slide image and the second tissue slide image aligned for the same tissue area.
PCT/US2020/028162 2019-04-24 2020-04-14 Systems and methods to label structures of interest in tissue slide images WO2020219305A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US17/605,943 US11972621B2 (en) 2019-04-24 2020-04-14 Systems and methods to label structures of interest in tissue slide images
EP20795913.1A EP3959654A4 (en) 2019-04-24 2020-04-14 Systems and methods to label structures of interest in tissue slide images
US18/617,762 US20240233418A1 (en) 2019-04-24 2024-03-27 Systems and methods to label structures of interest in tissue slide images

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201962838243P 2019-04-24 2019-04-24
US62/838,243 2019-04-24

Related Child Applications (2)

Application Number Title Priority Date Filing Date
US17/605,943 A-371-Of-International US11972621B2 (en) 2019-04-24 2020-04-14 Systems and methods to label structures of interest in tissue slide images
US18/617,762 Continuation US20240233418A1 (en) 2019-04-24 2024-03-27 Systems and methods to label structures of interest in tissue slide images

Publications (1)

Publication Number Publication Date
WO2020219305A1 true WO2020219305A1 (en) 2020-10-29

Family

ID=72941734

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2020/028162 WO2020219305A1 (en) 2019-04-24 2020-04-14 Systems and methods to label structures of interest in tissue slide images

Country Status (3)

Country Link
US (2) US11972621B2 (en)
EP (1) EP3959654A4 (en)
WO (1) WO2020219305A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022090205A1 (en) * 2020-10-30 2022-05-05 Genfit A method and a system for obtaining a training dataset usable for training a neural network to detect cells, using color deconvolutions and geometrical attributes
WO2022261472A1 (en) * 2021-06-11 2022-12-15 Genentech, Inc. Hierarchical workflow for generating annotated training data for machine learning enabled image segmentation
WO2023203513A1 (en) * 2022-04-22 2023-10-26 Verb Surgical Inc. Annotation system for surgical content

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020219305A1 (en) * 2019-04-24 2020-10-29 Icahn School Of Medicine At Mount Sinai Systems and methods to label structures of interest in tissue slide images
KR102241724B1 (en) * 2020-05-22 2021-04-19 주식회사 루닛 Method and system for refining label information
US12079311B2 (en) * 2021-01-08 2024-09-03 Salesforce, Inc. AI-enhanced data labeling
KR20230023359A (en) * 2021-08-10 2023-02-17 한화테크윈 주식회사 surveillance camera system

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110286654A1 (en) * 2010-05-21 2011-11-24 Siemens Medical Solutions Usa, Inc. Segmentation of Biological Image Data
US20120076390A1 (en) * 2010-09-28 2012-03-29 Flagship Bio Methods for feature analysis on consecutive tissue sections
US20130089249A1 (en) * 2010-06-15 2013-04-11 Koninklijke Philips Electronics N.V. Image processing method in microscopy
US20170091937A1 (en) * 2014-06-10 2017-03-30 Ventana Medical Systems, Inc. Methods and systems for assessing risk of breast cancer recurrence
WO2018115055A1 (en) * 2016-12-22 2018-06-28 Ventana Medical Systems, Inc. Computer scoring based on primary stain and immunohistochemistry images
US10043054B2 (en) * 2014-10-17 2018-08-07 Cireca Theranostics, Llc Methods and systems for classifying biological samples, including optimization of analyses and use of correlation

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111602136A (en) * 2018-01-11 2020-08-28 谷歌有限责任公司 Method for creating histopathology ground truth masks using slide re-staining
WO2020219305A1 (en) * 2019-04-24 2020-10-29 Icahn School Of Medicine At Mount Sinai Systems and methods to label structures of interest in tissue slide images

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110286654A1 (en) * 2010-05-21 2011-11-24 Siemens Medical Solutions Usa, Inc. Segmentation of Biological Image Data
US20130089249A1 (en) * 2010-06-15 2013-04-11 Koninklijke Philips Electronics N.V. Image processing method in microscopy
US20120076390A1 (en) * 2010-09-28 2012-03-29 Flagship Bio Methods for feature analysis on consecutive tissue sections
US20170091937A1 (en) * 2014-06-10 2017-03-30 Ventana Medical Systems, Inc. Methods and systems for assessing risk of breast cancer recurrence
US10043054B2 (en) * 2014-10-17 2018-08-07 Cireca Theranostics, Llc Methods and systems for classifying biological samples, including optimization of analyses and use of correlation
WO2018115055A1 (en) * 2016-12-22 2018-06-28 Ventana Medical Systems, Inc. Computer scoring based on primary stain and immunohistochemistry images

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3959654A4 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022090205A1 (en) * 2020-10-30 2022-05-05 Genfit A method and a system for obtaining a training dataset usable for training a neural network to detect cells, using color deconvolutions and geometrical attributes
WO2022261472A1 (en) * 2021-06-11 2022-12-15 Genentech, Inc. Hierarchical workflow for generating annotated training data for machine learning enabled image segmentation
WO2023203513A1 (en) * 2022-04-22 2023-10-26 Verb Surgical Inc. Annotation system for surgical content

Also Published As

Publication number Publication date
US11972621B2 (en) 2024-04-30
US20220222815A1 (en) 2022-07-14
US20240233418A1 (en) 2024-07-11
EP3959654A4 (en) 2023-05-24
EP3959654A1 (en) 2022-03-02

Similar Documents

Publication Publication Date Title
US11972621B2 (en) Systems and methods to label structures of interest in tissue slide images
US11205266B2 (en) Systems and methods for detection of structures and/or patterns in images
US11449998B2 (en) Processing of histology images with a convolutional neural network to identify tumors
US20220351860A1 (en) Federated learning system for training machine learning algorithms and maintaining patient privacy
Lee et al. Deep learning of histopathology images at the single cell level
JP2021524630A (en) Multi-sample whole slide image processing via multi-resolution registration
Negahbani et al. PathoNet introduced as a deep neural network backend for evaluation of Ki-67 and tumor-infiltrating lymphocytes in breast cancer
US20220351379A1 (en) Non-tumor segmentation to support tumor detection and analysis
Cai et al. Saliency-guided level set model for automatic object segmentation
Xu et al. Using transfer learning on whole slide images to predict tumor mutational burden in bladder cancer patients
Alegro et al. Automating cell detection and classification in human brain fluorescent microscopy images using dictionary learning and sparse coding
US20190080146A1 (en) Systems and methods for automatic generation of training sets for machine interpretation of images
US20240232627A1 (en) Systems and Methods to Train A Cell Object Detector
Salvi et al. cyto‐Knet: An instance segmentation approach for multiple myeloma plasma cells using conditional kernels
EP4275052A1 (en) Quantification of conditions on biomedical images across staining modalities using a multi-task deep learning framework
Alahmari et al. A review of nuclei detection and segmentation on microscopy images using deep learning with applications to unbiased stereology counting
Naji et al. HoLy-Net: Segmentation of histological images of diffuse large B-cell lymphoma
US20230307132A1 (en) Hybrid and accelerated ground-truth generation for duplex arrays
Negahbani et al. PathoNet: Deep learning assisted evaluation of Ki-67 and tumor infiltrating lymphocytes (TILs) as prognostic factors in breast cancer; A large dataset and baseline
Xu et al. Detection and Classification of Breast Cancer Metastates Based on U-Net
Chai Using machine learning techniques on digitised histopathological images to distinguish lipoma (LP) from atypical lipomatous tumours (ALT)
JP2024528609A (en) Pathological image analysis method and system
da Silva NeuronDyn: Live Neurotransmitter Vesicle Movement Dynamics in Living Neurons
Scherr Deep Learning-Based Particle Detection and Instance Segmentation for Microscopy Images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20795913

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2020795913

Country of ref document: EP

Effective date: 20211124