WO2017132674A1 - Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos - Google Patents

Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos Download PDF

Info

Publication number
WO2017132674A1
WO2017132674A1 PCT/US2017/015645 US2017015645W WO2017132674A1 WO 2017132674 A1 WO2017132674 A1 WO 2017132674A1 US 2017015645 W US2017015645 W US 2017015645W WO 2017132674 A1 WO2017132674 A1 WO 2017132674A1
Authority
WO
WIPO (PCT)
Prior art keywords
cytoplasm
features
image
images
pronuclei
Prior art date
Application number
PCT/US2017/015645
Other languages
French (fr)
Inventor
Gerard LETTERIE
Andrew Macdonald
Original Assignee
Letterie Gerard
Andrew Macdonald
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Letterie Gerard, Andrew Macdonald filed Critical Letterie Gerard
Priority to CA3013048A priority Critical patent/CA3013048A1/en
Priority to EP17745088.9A priority patent/EP3408796A4/en
Priority to US16/073,126 priority patent/US10902334B2/en
Priority to JP2018559173A priority patent/JP2019509057A/en
Publication of WO2017132674A1 publication Critical patent/WO2017132674A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/217Validation; Performance evaluation; Active pattern learning techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • G06N5/045Explanation of inference; Explainable artificial intelligence [XAI]; Interpretable artificial intelligence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • G06T7/0014Biomedical image inspection using an image reference approach
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/48Extraction of image or video features by mapping characteristic values of the pattern into a parameter space, e.g. Hough transformation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/776Validation; Performance evaluation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/69Microscopic objects, e.g. biological cells or cellular parts
    • G06V20/698Matching; Classification
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/40ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/70ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10056Microscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20048Transform domain processing
    • G06T2207/20061Hough transform
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30024Cell structures in vitro; Tissue sections in vitro

Definitions

  • oocytes are retrieved and immediately sorted based only on maturity or immaturity, with the former being inseminated. Performance of these oocytes awaits an expensive and time consuming routine of culturing over 5 days.
  • Manual oocyte scoring may provide useful clinical information. However, manual assessment of oocytes and embryos remains standard of care and has not changed significantly since inception of human embryology techniques. Thus, there remains a need for better tools to assess the reproductive potential of oocytes and pronuclear embryos to identify those with a high likelihood of developing into blastocysts and ultimately a live birth.
  • a computer system automatically converts a set of training images of cells (e.g., oocytes or pronuclear embryos) and related outcome metadata into a description document by extracting features from the pixel values of the training images and associating the extracted features with the outcome metadata. Then, based on the description document, the computer system automatically computes a decision model that can be used to predict outcomes of new cells.
  • the training images are obtained using bright-field microscopy.
  • the extracted features may include a boundary of a cytoplasm.
  • the computer system may detect such boundaries in the training images by applying an edge detection algorithm to the images to detect edges and applying a circle detection algorithm to the detected edges to detect an approximate cytoplasm boundary.
  • the computer system may convert color images to grayscale images prior to application of the edge detection algorithm.
  • the computer system may create an annular mask image from the approximate cytoplasm boundary and perform image segmentation on the original source image to identify a connected region as the cytoplasm.
  • Additional extracted features may include one or more of the following: total area of the cytoplasm; aspect ratio of the cytoplasm; convexity of the boundary of the cytoplasm; average image intensity of the cytoplasm; standard deviation of the image intensity of the cytoplasm; smoothness of the cytoplasm; subsampled image intensity or smoothness for grid areas within the cytoplasm; cytoplasmic texture or texture distribution; density; and clustering.
  • the average image intensity of the cytoplasm may be calculated by averaging the corresponding pixel values of the identified cytoplasm in a grayscale image.
  • additional extracted features may include one or more of the following: a boundary of a polar body, measured manually or with image analysis using an edge detection algorithm; an indication as to whether the polar body exists; identification of inner and outer edges of a zona pellucida; identification of perivitelline space; a measurement of smoothness of the edges of the zona pellucida; alignment of principal axes of the cytoplasm and the zona; total area of the polar body, the zona pellucida, or the perivitelline space; aspect ratio of the polar body, the zona pellucida, or the perivitelline space; convexity of a boundary of the polar body, the zona pellucida, or the perivitelline space; average image intensity of the polar body, the zona pellucida, or the perivitelline space; standard deviation of the image intensity of the polar body, the zona pellucida, or the perivitelline space; smoothness of the polar body,
  • additional extracted features may include one or more of the following: identification of the boundaries of the pronuclei, measured manually or with image analysis using an edge detection algorithm; area of the pronuclei, measured individually and/or relative to each other; aspect ratio of the pronuclei, measured individually and/or relative to each other; convexity of a boundary of one or more of the pronuclei; average image intensity of one or more of the pronuclei; standard deviation of the image intensity of one or more of the pronuclei; smoothness of one or more of the pronuclei; and subsampled image intensity or smoothness for grid areas within one or more of the pronuclei.
  • a computer system automatically extracts features from new cell images (which, like the training images, also may be obtained using bright-field microscopy) that describe the new cells and predicts one or more outcomes for the new cells by applying the decision model to the features extracted from the new cell images.
  • the features extracted from the new cell images correspond to features of the training images selected for inclusion in the decision model, and may be calculated for in the same way.
  • the outcomes may include one or more of quality or reproductive potential for a new oocyte, whether abnormalities in chromosome number are detected later in development, whether genetic disorders are detected later in development, and abnormal cell growth outcomes.
  • the quality or reproductive potential for a new oocyte may be expressed as one or more of: a probability that the oocyte will fertilize when exposed to human sperm, a probability that a resulting embryo will reach a given stage of development, or a probability that the resulting embryo will result in a live birth when transferred to a uterus.
  • These functions may be implemented as an add-on to an existing image analysis system or as a separate system.
  • the system may be integrated with a microscopy system for obtaining cell images.
  • FIGURE 1A is an image of a mature oocyte that may be analyzed by an automated image-based cell analysis system according to at least one embodiment of the present disclosure
  • FIGURE IB is an image of abnormal oocytes that may be analyzed by an automated image-based cell analysis system according to at least one embodiment of the present disclosure
  • FIGURE 2 is a flow diagram illustrating work flow in an automated image-based cell analysis system according to at least one embodiment of the present disclosure
  • FIGURE 3 is a flow diagram illustrating an algorithm for identification of a boundary of a cytoplasm using image analysis techniques according to at least one embodiment of the present disclosure
  • FIGURE 4 is a flow diagram illustrating an algorithm for identification of inner and outer edges of a zona pellucida of an oocyte or pronuclear embryo according to at least one embodiment of the present disclosure
  • FIGURE 5 is an image of a pronuclear embryo that may be analyzed by an automated image-based cell analysis system according to at least one embodiment of the present disclosure.
  • FIGURE 6 is a block diagram that illustrates aspects of an illustrative computing device appropriate for use in accordance with embodiments of the present disclosure.
  • the present disclosure is directed to computerized image analysis, content-based image retrieval and cellular feature extraction of cells such as oocytes and pronuclear embryos, and their various intracellular components, and predicting outcomes related to cell quality, cell growth, or development (e.g., developmental changes that occur in comparison between oocytes and pronuclear embryos).
  • the present disclosure includes an algorithm for the study of attributes of oocytes and pronuclear embryos that can be used in an automated analytic software tool to identify oocytes and pronuclear embryos with a high likelihood of implanting, among other possible outcomes related to reproductive potential.
  • the present disclosure describes technology with unique clinical benefits distinct from anything commercially available.
  • Disclosed embodiments are particularly significant in their ability to adapt to new cell science; as new outcomes are studied and linked to particular cell features, disclosed embodiments offer the ability to automatically extract those features from training images and associate with them with outcomes, generate decision models, and apply those decision models to new cell images in order to predict outcomes for the new cells.
  • the present disclosure describes aspects of an automated analysis system that evaluates cells, such as oocytes and pronuclear embryos, in a new way while still being compatible with hardware systems using traditional bright field microscopy.
  • Other computer image analysis systems use time-lapsed dark-field systems, and require separate monitoring hardware and additional personnel for maintenance and monitoring.
  • data analyzing outcomes with such systems are of variable quality and have not consistently demonstrated any improvement in outcomes.
  • embodiments of the present disclosure have several potential marketplace advantages.
  • existing hardware e.g., in embryology labs
  • existing microscopy systems present in most labs may be used, with no additional hardware needed.
  • existing image analysis software can be extended by incorporating embodiments of the present disclosure into an existing image analysis software system or into an integrated system of microscopy hardware an image analysis software.
  • the archive of images and outcomes also can be expanded to enhance precision and reliability of the algorithm.
  • Image training can be performed on training images obtained from multiple sources (e.g., clinics, research institutions) stored in a common data repository.
  • sources e.g., clinics, research institutions
  • machine self-training can provide improvements over time thus leading to significant improvements in predicting outcomes with continued use.
  • Embodiments described herein can be applied, for example, to both oocyte cryopreservation and embryo development to predict reproductive potential in oocytes prior to cryopreservation or after insemination during embryo development.
  • embodiments described herein can provide useful analysis of a single cell, potentially offering greater ease in image acquisition and analysis, as well as improved clinical decision making as it relates to cell quality or development, e.g., prior to oocyte freezing for fertility preservation or oocyte donation and improved selection of which oocyte to inseminate and which pronuclear embryo to continue in culture.
  • Embodiments described herein are useful in studying single cells, particularly in the context of reproductive medicine. After insemination and culturing, as the embryo expands in cell number (e.g., beyond 2 days of observation), the image field becomes more crowded, it becomes difficult to isolate individual cells, and the analytics become more complicated. As single cells, oocyte and pronuclear embryos provide excellent opportunities to extract information using computerized image analysis.
  • the present disclosure describes a quality model learned from existing images and related metadata, such as descriptions of outcomes related to cell quality, cell growth, and development.
  • outcomes include blastocyst formation, ploidy determinations, and clinical pregnancies.
  • Reproductive potential can be assessed for, e.g., oocytes after freezing and pronuclear embryos.
  • new forms of data become available and regulatory approvals are obtained for additional study (which may provide additional descriptions of outcomes, or new classes of outcomes)
  • the dataset of existing images, metadata, and outcomes can be extended and the quality models can be improved.
  • the quality or reproductive potential of an oocyte or pronuclear embryo can be expressed in several different ways, such as the probability that the oocyte will fertilize when exposed to human sperm, the probability that the resulting embryo will reach a given stage of development (e.g., the blastocyst stage), or the probability that the embryo will result in a live birth when transferred to a uterus.
  • probability can be expressed as a percentage chance, or as a high/medium/low chance, of reaching a given state of development.
  • genetic information e.g., the genetic complement
  • abnormal cell growth outcomes can be predicted, using techniques described in the present disclosure.
  • an outcome may be whether abnormalities in chromosome number or genetic disorders are detected later on in development.
  • aspects of the present disclosure provide the ability to gauge oocyte quality to guide decision making regarding oocyte cryopreservation.
  • Oocyte preservation may be performed, for example, in an elective oocyte freezing program, in which it may be helpful to determine how many oocytes are to be frozen to maximize future fertility, or to determine a point of diminishing returns beyond which additional oocyte retrievals do not significantly increase the probability of future fertility.
  • aspects of the present disclosure may be used to avoid performing too many or too few oocyte retrievals.
  • Oocyte preservation also may be performed in an oocyte donation program, where the ability to prospectively assess oocyte potential will guide decision making about how many oocytes to batch for a given donation or whether to discard them and forego fertilization.
  • aspects of the present disclosure improve on the current practice of using a single determination of whether an oocyte is mature or immature and freezing 6 to 8 oocytes per batch. Aspects of the present disclosure may be particularly useful in scenarios where the number of oocytes inseminated is limited, and the ability to identify an oocyte with the greatest reproductive potential becomes more important.
  • FIGURE 1A is an image of a mature oocyte, with the zona pellucida (or "zona") and polar body indicated.
  • attributes of the cytoplasm, zona, and polar body can be automatically identified and analyzed, as described in detail below.
  • the overall method is composed of two stages, each of which is composed of two phases.
  • Stage one the model construction stage, involves constructing a decision model (e.g., a model of quality for human oocytes or pronuclear embryos) from training images of cells (e.g., of oocytes or pronuclear embryos for which particular outcomes are known).
  • Stage two, the employment stage involves employing this model to predict outcomes for new cells being analyzed (e.g., in a clinical setting). For example, in the context of reproductive medicine, the employment stage may be used to determine the quality of an individual embryo.
  • the method may be performed by an automated image-based cell analysis system, as described herein, implemented in a computer system comprising one or more computing devices.
  • the automated image-based cell analysis system may include functionality for performing tasks described with reference to FIGURE 2.
  • the automated image-based cell analysis system may include multiple modules or engines to carry out different groups of tasks associated with different stages or phases, such as a model construction engine and a model employment engine, as described in detail below.
  • modules or engines may be implemented in the same computing device, or in different computing devices.
  • these sections or engines may include multiple submodules to carry out more specific tasks within the groups of tasks, such as particular types of feature extraction within the model construction engine.
  • Phase 1 the first phase of the model construction stage, is the feature extraction phase.
  • the automated image-based cell analysis system obtains a set of training images and related metadata from a data store 210 and converts them (e.g., using a model construction engine) into a description document.
  • the digital images that form the training set comprise pixels arranged in rows and columns. Each pixel is defined by a set of one or more sample values, with the number, meaning, and value range of the respective samples being determined by the format, color space, and bit depth (number of bits per channel) of the images. It will be understood that the techniques described herein can be adapted for analysis of digital images of various formats, resolutions, color spaces, bit depths, etc.
  • the feature extraction phase proceeds by extracting features 212 from the training images that describe the cells (e.g., oocytes or pronuclear embryos) in the images numerically, as described in detail below. In the context of reproductive medicine, this may include extraction of features such as the area measurement of the cytoplasm of each oocyte, or cytoplasmic or nuclear features of each pronuclear embryo in the training images.
  • the extracted features are then associated with outcomes in the description document 214. Further details of the features that can be included in a description document are provided below.
  • the term "document” is used herein to refer to a collection of the described information in any format suitable for further processing as described herein, and need not refer to a traditional document designed for human readability.
  • the data store may be updated as needed, e.g., to provide additional training images and related metadata.
  • one required item in the description document is the description of one or more outcomes of what happened when the oocytes in the respective training images were fertilized.
  • the outcomes associated with training images also referred to herein as "Feature 1 " of the respective training images, are not a function of the image data, but are metadata supplied along with the images.
  • the metadata may include information related to developmental outcomes. With regard to a training image of an oocyte, for example, such metadata may include information about whether the oocyte fertilized and how embryo development proceeded.
  • one or more of the following features 2-18 can be extracted from an oocyte image to create the description document in phase 1, as described below. It will be understood that extraction of some features described below (e.g., cytoplasm features) may be useful for many types of cells. For example, the pronuclear embryo is still a single cell (see FIGURE 5), and its features also can be analyzed as described below: dentification of the boundary of the cytoplasm, measured manually or with image analysis using an edge detection algorithm. An illustrative algorithm 300 is illustrated in FIGURE 3 :
  • step 310 convert image from color to grayscale (if needed).
  • step 320 apply an edge detection algorithm (e.g., Canny Threshold) to the grayscale image to detect edges.
  • an edge detection algorithm e.g., Canny Threshold
  • a circle detection algorithm e.g., Hough Circle detection
  • the circle detection algorithm may use known quantities of image resolution and magnification to control for size of circles to search from, and Gaussian blur on Canny-detected edges to further reduce number of detected circles, if needed, until only a single circle remains.
  • step 340 create an annular mask image from the circle, with inner and outer annulus radii as ratios of circle radius.
  • step 350 perform image segmentation on the original source image to identify a connected region as the cytoplasm.
  • a GrabCut method can be used, with the resulting connected region being the cytoplasm, and the boundary of the connected region being the boundary of the cytoplasm.
  • the annular mask can be used to restrict the search.
  • a mask image may be created containing only pixels from the connected region. This image simplifies computing further cytoplasm features.
  • otal area of the cytoplasm in the image defined by the portion of the image inside the boundary identified in feature 2. (The volume of the cytoplasm can also be estimated based on this information.)
  • spect ratio of the cytoplasm being the ratio of the shortest distance across the cytoplasm to the longest.
  • convexity may be calculated as follows: (a) Compute the convex hull of the mask image generated in feature 2, step (e), above.
  • average intensity may be calculated by averaging the corresponding pixel values of the identified cytoplasm in a grayscale image (in which the value of each pixel is a single sample), or by averaging a particular value for corresponding pixels (such as brightness) when represented in a color space such as HSB (Hue-Saturation-Brightness).
  • HSB Human-Saturation-Brightness
  • moothness of the cytoplasm measured by applying a Gaussian filter to the cytoplasm image, varying the size of the Gaussian filter and recording how that affects the number of features that can be detected in the cytoplasm by, e.g., Canny edge detector, Harris corner detector, or another process which detects features which vary by image smoothness. As the Gaussian smoothing is increased the number of features decreases until no features are detected. The system can track how fast the number of features decreases and how much smoothing is needed to eliminate features entirely, which provides information about the smoothness of the cytoplasm.
  • FIGURE 4 An illustrative algorithm 400 is illustrated in FIGURE 4: (a) At step 410, to identify the outer zona boundary, mask out the region of the image that has been identified as containing the boundary of the cytoplasm in feature 2, above.
  • step 420 to identify the inner zona boundary, mask out parts of the image inside the cytoplasm and outside the outer zona boundary, and then identify the inner zona boundary by applying edge detection and circle detection algorithms as in feature 2.
  • image sharpening can be performed before the edge detection steps so that the zona edges are more readily apparent.
  • the system may calculate axes by finding a minimal bounding rectangle around the mask of the cytoplasm and the zona, with alignment of the sides of the rectangle defining the principal axes.
  • the system may determine alignment by estimating the oocyte perimeter from an interior reference point (e.g., centroid) and iteratively region-growing no further than the oocyte edge, passing coordinates through this centroid along its major and minor axes to the oocyte edge, and measuring coordinate lengths within the estimated oocyte region.
  • an interior reference point e.g., centroid
  • cytoplasmic texture features Other features that may be extracted with image analysis techniques include cytoplasmic texture features, texture distribution, densities, and clustering.
  • FIGURE 5 An illustrative image of a pronuclear embryo is shown in FIGURE 5.
  • any of the additional features 19-23 described below can be extracted for the pronuclei in the center of the cytoplasm, as follows: 19. Identification of the boundaries of the pronuclei, measured manually or with image analysis using an edge detection algorithm.
  • the area or volume of the pronuclei measured individually and relative to each other.
  • the features described herein may be used to detect abnormalities in, for example, texture, inclusion bodies, and shape or size, such as the abnormalities depicted in the oocytes shown in FIGURE IB.
  • other tools such as colorization and color filters, also may be used for feature extraction.
  • the color filter can be applied to a color image of the cell before proceeding with further analysis of the image. This process may be performed, for example, on a color source image prior to conversion of the color source image to grayscale for detection of other features.
  • phase 2 determines which features extracted in phase 1 materially affect predictive ability in phase 4.
  • Features selected for inclusion in the decision model may include, for example, features having a numerical predictive value that exceeds a threshold value.
  • Features not selected for inclusion in the decision model, which do not have significant predictive value can be excluded in phases 3 and 4, which are described in detail below.
  • the model construction engine consumes or reads (e.g., using the model construction engine) the descriptions (e.g., numeric descriptions) in the description document and computes a mathematical optimization to determine which of the features are significant in determining quality of outcomes.
  • the output of phase 1 is the input for phase 2.
  • phase 2 is a model (see “generate decision model” block 220 shown in FIGURE 2) that can be used to predict outcomes. For example, in the field of reproductive medicine, such predicted outcomes may be used to determine quality or reproductive potential given the numeric features for a given oocyte or pronuclear embryo. This can be referred to as constructing a model with labeled training data.
  • the image analysis and attributes that the software detects are correlated with outcome and weighted according to significance.
  • a portion of the training data is held out to use for validating the model, and training is done on the remaining training data using cross- validation to ensure the stability of the model.
  • the model can be constructed using logistic regression, decision trees, or some other method that consumes labeled training data. Sufficient training data is required to achieve a significant result from training. When cross-validation implies that the model is predictive, the held-out portion of the training data is used to validate model quality.
  • Training in phase 2 also may be performed using unsupervised learning such as k-means or neural networks to segment the training data into classes. After classes have been identified, the predominant label from the samples in each class is assigned to each class. The learning phase also may be performed using other machine learning methods.
  • the output of phase 2 is a decision model which can be used to predict outcomes, such as cell quality or reproductive potential.
  • the quality or reproductive potential of an oocyte or pronuclear embryo can be expressed in several different ways, such as the probability that the embryo will be chromosomally normal or euploid, or abnormal or aneuploid, the probability that the oocyte will fertilize when exposed to human sperm, the probability that the resulting embryo will reach a given stage of development (e.g., the blastocyst stage), or the probability the embryo will result in a live birth when transferred to a uterus.
  • probability can be expressed as a percentage chance, or as a high/medium/low chance, of reaching a given state of development.
  • phase 3 the first phase of the employment stage, the automated image-based cell analysis system performs (e.g., using the model employment engine) feature extraction on a new image 230 (e.g., an image of a single oocyte or pronuclear embryo).
  • a new image 230 e.g., an image of a single oocyte or pronuclear embryo.
  • the automated image-based cell analysis system generates features 232 for the new image corresponding to one or more features that were generated for the training images in phase 1, other than feature 1 (outcome), which is not yet known. It is not necessary for all of the features described in phase 1 to be extracted for new images in phase 3. For example, features that have been determined in phase 2 not to have predictive value need not be extracted in phase 3.
  • the automated image-based cell analysis system predicts (e.g., using the model employment engine) one or more outcomes 242 for the cell (e.g., oocyte or pronuclear embryo) in the new image by applying the decision model 240 generated in phase 2 to the features extracted in phase 3. For example, a score between 0 and 1 can be assigned, with a greater score indicating a better predicted outcome. The predicted outcome adds a significant benefit compared to, for example, older methods of selecting oocytes for insemination based solely on maturity. A threshold score can be set, below which the oocyte regardless of maturity may be discarded as unlikely if not impossible to yield a blastocyst and live birth.
  • the features may be weighted together in some combination, or used as decision points in a decision tree application to determine the likely outcomes for the cell (e.g., oocyte or pronuclear embryo) for one or more of the scenarios provided as outcomes in feature 1 of the training set.
  • the cell e.g., oocyte or pronuclear embryo
  • Cells can be analyzed separately, or in combination. For example, the quality of oocytes and pronuclear embryos can be analyzed individually and then combined to form an overall quality score.
  • an automated image-based cell analysis system identifies 5 cellular attributes of oocytes. These attributes were selected as possibly predictive of the likelihood reproductive potential of the oocyte and of blastocyst formation and implantation. A numerical score between 0.0 (lowest) and 1.0 (highest) was calculated to estimate these likelihoods.
  • the algorithm employs Stochastic Gradient Descent implemented in the Python programming language through the SciPy extension. Feature extraction uses a variety of additional techniques in C++, including a Hough Circle detector and a Harris Corner detector.
  • Main outcome measurements include calculation of an Oocyte Score (OS) to identify oocytes with a high reproductive potential and likelihood of blastocyst formation and to compare the OS (Group 1) to selection using standard morphology (Group 2) and to a computerized video monitoring system (Group 3).
  • OS Oocyte Score
  • an automated image-based cell analysis system can receive software updates over a network, to provide updates to the analysis system, the training data, or other data.
  • described techniques and tools may be implemented by any suitable computing device or set of devices.
  • an engine e.g., a software engine working in combination with computer hardware and microscopy systems
  • An engine includes logic (e.g., in the form of computer program code) configured to cause one or more computing devices to perform actions described herein as being associated with the engine.
  • a computing device can be specifically programmed to perform the actions by having installed therein a tangible computer-readable medium having computer-executable instructions stored thereon that, when executed by one or more processors of the computing device, cause the computing device to perform the actions.
  • the particular engines described herein are included for ease of discussion, but many alternatives are possible. For example, actions described herein as associated with two or more engines on multiple devices may be performed by a single engine. As another example, actions described herein as associated with a single engine may be performed by two or more engines on the same device or on multiple devices.
  • a data store contains data as described herein and may be hosted, for example, by a database management system (DBMS) to allow a high level of data throughput between the data store and other components of a described system.
  • the DBMS may also allow the data store to be reliably backed up and to maintain a high level of availability.
  • a data store may be accessed by other system components via a network, such as a private network in the vicinity of the system, a secured transmission channel over the public Internet, a combination of private and public networks, and the like.
  • a data store may include structured data stored as files in a traditional file system. Data stores may reside on computing devices that are part of or separate from components of systems described herein.
  • server devices may include suitable computing devices configured to provide information and/or services described herein.
  • Server devices may include any suitable computing devices, such as dedicated server devices.
  • Server functionality provided by server devices may, in some cases, be provided by software (e.g., virtualized computing instances or application objects) executing on a computing device that is not a dedicated server device.
  • client can be used to refer to a computing device that obtains information and/or accesses services provided by a server over a communication link.
  • a particular device does not necessarily require the presence of a server.
  • a single device may act as a server, a client, or both a server and a client, depending on context and configuration.
  • Actual physical locations of clients and servers are not necessarily important, but the locations can be described as "local” for a client and "remote" for a server to illustrate a common usage scenario in which a client is receiving information provided by a server at a remote location.
  • a peer-to-peer arrangement or other models, can be used.
  • FIGURE 6 is a block diagram that illustrates aspects of an illustrative computing device 600 appropriate for use in accordance with embodiments of the present disclosure.
  • the description below is applicable to servers, personal computers, mobile phones, smart phones, tablet computers, embedded computing devices, and other currently available or yet-to-be-developed devices that may be used in accordance with embodiments of the present disclosure.
  • Computing devices described herein may be integrated with specialized hardware, such as microscopy systems, for obtaining images, or as standalone devices that obtain images for analysis in some other way, such as by receiving images stored remotely in a cloud computing arrangement.
  • the computing device 600 includes at least one processor 602 and a system memory 604 connected by a communication bus 606.
  • the system memory 604 may be volatile or nonvolatile memory, such as read only memory (“ROM”), random access memory (“RAM”), EEPROM, flash memory, or other memory technology.
  • ROM read only memory
  • RAM random access memory
  • EEPROM electrically erasable programmable read-only memory
  • flash memory or other memory technology.
  • system memory 604 typically stores data and/or program modules that are immediately accessible to and/or currently being operated on by the processor 602.
  • the processor 602 may serve as a computational center of the computing device 600 by supporting the execution of instructions.
  • the computing device 600 may include a network interface 610 comprising one or more components for communicating with other devices over a network.
  • Embodiments of the present disclosure may access basic services that utilize the network interface 610 to perform communications using common network protocols.
  • the network interface 610 may also include a wireless network interface configured to communicate via one or more wireless communication protocols, such as WiFi, 2G, 3G, 4G, LTE, WiMAX, Bluetooth, and/or the like.
  • the computing device 600 also includes a storage medium 608.
  • services may be accessed using a computing device that does not include means for persisting data to a local storage medium. Therefore, the storage medium 608 depicted in FIGURE 6 is optional.
  • the storage medium 608 may be volatile or nonvolatile, removable or nonremovable, implemented using any technology capable of storing information such as, but not limited to, a hard drive, solid state drive, CD-ROM, DVD, or other disk storage, magnetic tape, magnetic disk storage, and/or the like.
  • computer-readable medium includes volatile and nonvolatile and removable and nonremovable media implemented in any method or technology capable of storing information, such as computer-readable instructions, data structures, program modules, or other data.
  • system memory 604 and storage medium 608 depicted in FIGURE 6 are examples of computer-readable media.
  • FIGURE 6 does not show some of the typical components of many computing devices.
  • the computing device 600 may include input devices, such as a keyboard, keypad, mouse, trackball, microphone, video camera, touchpad, touchscreen, electronic pen, stylus, and/or the like.
  • Such input devices may be coupled to the computing device 600 by wired or wireless connections including RF, infrared, serial, parallel, Bluetooth, USB, or other suitable connection protocols using wireless or physical connections.
  • input data can be captured by input devices and processed, transmitted, or stored (e.g., for future processing).
  • the processing may include encoding data streams, which can be subsequently decoded for presentation by output devices.
  • Media data can be captured by multimedia input devices and stored by saving media data streams as files on a computer-readable storage medium (e.g., in memory or persistent storage on a client device, server, administrator device, or some other device).
  • Input devices can be separate from and communicatively coupled to computing device 600 (e.g., a client device), or can be integral components of the computing device 600.
  • multiple input devices may be combined into a single, multifunction input device (e.g., a video camera with an integrated microphone).
  • the computing device 600 may also include output devices such as a display, speakers, printer, etc.
  • the output devices may include video output devices such as a display or touchscreen.
  • the output devices also may include audio output devices such as external speakers or earphones.
  • the output devices can be separate from and communicatively coupled to the computing device 600, or can be integral components of the computing device 600.
  • Input functionality and output functionality may be integrated into the same input/output device (e.g., a touchscreen). Any suitable input device, output device, or combined input/output device either currently known or developed in the future may be used with described systems.
  • functionality of computing devices described herein may be implemented in computing logic embodied in hardware or software instructions, which can be written in a programming language, such as C, C++, COBOL, JAVATM, PHP, Perl, Python, Ruby, HTML, CSS, JavaScript, VBScript, ASPX, Microsoft .NETTM languages such as C#, and/or the like.
  • Computing logic may be compiled into executable programs or written in interpreted programming languages.
  • functionality described herein can be implemented as logic modules that can be duplicated to provide greater processing capability, merged with other modules, or divided into sub-modules.
  • the computing logic can be stored in any type of computer-readable medium (e.g., a non-transitory medium such as a memory or storage medium) or computer storage device and be stored on and executed by one or more general-purpose or special-purpose processors, thus creating a special-purpose computing device configured to provide functionality described herein.
  • a computer-readable medium e.g., a non-transitory medium such as a memory or storage medium
  • computer storage device e.g., a non-transitory medium such as a memory or storage medium
  • general-purpose or special-purpose processors e.g., a general-purpose or special-purpose processors
  • modules or subsystems can be separated into additional modules or subsystems or combined into fewer modules or subsystems.
  • modules or subsystems can be omitted or supplemented with other modules or subsystems.
  • functions that are indicated as being performed by a particular device, module, or subsystem may instead be performed by one or more other devices, modules, or subsystems.
  • processing stages in the various techniques can be separated into additional stages or combined into fewer stages.
  • processing stages in the various techniques can be omitted or supplemented with other techniques or processing stages.
  • processing stages that are described as occurring in a particular order can instead occur in a different order.
  • processing stages that are described as being performed in a series of steps may instead be handled in a parallel fashion, with multiple modules or software processes concurrently handling one or more of the illustrated processing stages.
  • processing stages that are indicated as being performed by a particular device or module may instead be performed by one or more other devices or modules.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Public Health (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Databases & Information Systems (AREA)
  • Computing Systems (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Epidemiology (AREA)
  • Primary Health Care (AREA)
  • Radiology & Medical Imaging (AREA)
  • General Engineering & Computer Science (AREA)
  • Pathology (AREA)
  • Quality & Reliability (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Geometry (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Apparatus Associated With Microorganisms And Enzymes (AREA)
  • Image Processing (AREA)
  • Measuring Or Testing Involving Enzymes Or Micro-Organisms (AREA)
  • Image Analysis (AREA)
  • Investigating Or Analysing Biological Materials (AREA)

Abstract

A computer system automatically converts a set of training images of cells (e.g., oocytes or pronuclear embryos) and related outcome metadata into a description document by extracting features (e.g., cytoplasm features) from the pixel values of the training images that describe the cells and associating the extracted features with the outcome metadata. Based on the description document, the system automatically computes a decision model that can be used to predict outcomes of new cells. To predict outcomes of new cells, a computer system automatically extracts features from images that describe the new cells and predicts one or more outcomes by applying the decision model. The features extracted from the images that describe the new cells correspond to features selected for inclusion in the decision model, and are calculated in the same way as the corresponding features extracted from the training images.

Description

AUTOMATED IMAGE ANALYSIS TO ASSESS REPRODUCTIVE POTENTIAL OF HUMAN OOCYTES AND PRONUCLEAR EMBRYOS
CROSS-REFERENCE TO RELATED APPLICATION This application claims the benefit of U.S. Provisional Application No. 62/288,379, filed January 28, 2016, the disclosure of which is hereby incorporated by reference herein.
BACKGROUND
Conventional assessment of embryo quality is performed through static, daily measurements by observers using traditional bright field examination of embryos during 5 days of development. Though a variety of scoring systems can be used to translate these observations into semi-quantifiable scores, selection of the embryo or embryos most likely to implant remains a qualitative exercise. This process is labor and cost intensive, requires considerable training and skill, and may be impacted by factors common to any repetitive technique of observation such as fatigue and inter- and intra-ob server variability.
Conventionally, oocytes are retrieved and immediately sorted based only on maturity or immaturity, with the former being inseminated. Performance of these oocytes awaits an expensive and time consuming routine of culturing over 5 days.
Manual oocyte scoring may provide useful clinical information. However, manual assessment of oocytes and embryos remains standard of care and has not changed significantly since inception of human embryology techniques. Thus, there remains a need for better tools to assess the reproductive potential of oocytes and pronuclear embryos to identify those with a high likelihood of developing into blastocysts and ultimately a live birth.
SUMMARY
This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This summary is not intended to identify key features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
In one aspect, a computer system automatically converts a set of training images of cells (e.g., oocytes or pronuclear embryos) and related outcome metadata into a description document by extracting features from the pixel values of the training images and associating the extracted features with the outcome metadata. Then, based on the description document, the computer system automatically computes a decision model that can be used to predict outcomes of new cells. The training images are obtained using bright-field microscopy.
The extracted features may include a boundary of a cytoplasm. The computer system may detect such boundaries in the training images by applying an edge detection algorithm to the images to detect edges and applying a circle detection algorithm to the detected edges to detect an approximate cytoplasm boundary. The computer system may convert color images to grayscale images prior to application of the edge detection algorithm. After the approximate cytoplasm boundary is detected, the computer system may create an annular mask image from the approximate cytoplasm boundary and perform image segmentation on the original source image to identify a connected region as the cytoplasm.
Additional extracted features may include one or more of the following: total area of the cytoplasm; aspect ratio of the cytoplasm; convexity of the boundary of the cytoplasm; average image intensity of the cytoplasm; standard deviation of the image intensity of the cytoplasm; smoothness of the cytoplasm; subsampled image intensity or smoothness for grid areas within the cytoplasm; cytoplasmic texture or texture distribution; density; and clustering. The average image intensity of the cytoplasm may be calculated by averaging the corresponding pixel values of the identified cytoplasm in a grayscale image.
For oocytes or pronuclear embryos, additional extracted features may include one or more of the following: a boundary of a polar body, measured manually or with image analysis using an edge detection algorithm; an indication as to whether the polar body exists; identification of inner and outer edges of a zona pellucida; identification of perivitelline space; a measurement of smoothness of the edges of the zona pellucida; alignment of principal axes of the cytoplasm and the zona; total area of the polar body, the zona pellucida, or the perivitelline space; aspect ratio of the polar body, the zona pellucida, or the perivitelline space; convexity of a boundary of the polar body, the zona pellucida, or the perivitelline space; average image intensity of the polar body, the zona pellucida, or the perivitelline space; standard deviation of the image intensity of the polar body, the zona pellucida, or the perivitelline space; smoothness of the polar body, the zona pellucida, or the perivitelline space; and subsampled image intensity or smoothness for grid areas within the polar body, the zona pellucida, or the perivitelline space.
For pronuclear embryos, additional extracted features may include one or more of the following: identification of the boundaries of the pronuclei, measured manually or with image analysis using an edge detection algorithm; area of the pronuclei, measured individually and/or relative to each other; aspect ratio of the pronuclei, measured individually and/or relative to each other; convexity of a boundary of one or more of the pronuclei; average image intensity of one or more of the pronuclei; standard deviation of the image intensity of one or more of the pronuclei; smoothness of one or more of the pronuclei; and subsampled image intensity or smoothness for grid areas within one or more of the pronuclei.
In another aspect, a computer system automatically extracts features from new cell images (which, like the training images, also may be obtained using bright-field microscopy) that describe the new cells and predicts one or more outcomes for the new cells by applying the decision model to the features extracted from the new cell images. The features extracted from the new cell images correspond to features of the training images selected for inclusion in the decision model, and may be calculated for in the same way. The outcomes may include one or more of quality or reproductive potential for a new oocyte, whether abnormalities in chromosome number are detected later in development, whether genetic disorders are detected later in development, and abnormal cell growth outcomes. The quality or reproductive potential for a new oocyte may be expressed as one or more of: a probability that the oocyte will fertilize when exposed to human sperm, a probability that a resulting embryo will reach a given stage of development, or a probability that the resulting embryo will result in a live birth when transferred to a uterus.
These functions may be implemented as an add-on to an existing image analysis system or as a separate system. The system may be integrated with a microscopy system for obtaining cell images.
BRIEF DESCRIPTION OF THE DRAWINGS
The foregoing aspects and many of the attendant advantages of this invention will become more readily appreciated as the same become better understood by reference to the following detailed description, when taken in conjunction with the accompanying drawings, wherein: FIGURE 1A is an image of a mature oocyte that may be analyzed by an automated image-based cell analysis system according to at least one embodiment of the present disclosure;
FIGURE IB is an image of abnormal oocytes that may be analyzed by an automated image-based cell analysis system according to at least one embodiment of the present disclosure;
FIGURE 2 is a flow diagram illustrating work flow in an automated image-based cell analysis system according to at least one embodiment of the present disclosure;
FIGURE 3 is a flow diagram illustrating an algorithm for identification of a boundary of a cytoplasm using image analysis techniques according to at least one embodiment of the present disclosure;
FIGURE 4 is a flow diagram illustrating an algorithm for identification of inner and outer edges of a zona pellucida of an oocyte or pronuclear embryo according to at least one embodiment of the present disclosure;
FIGURE 5 is an image of a pronuclear embryo that may be analyzed by an automated image-based cell analysis system according to at least one embodiment of the present disclosure; and
FIGURE 6 is a block diagram that illustrates aspects of an illustrative computing device appropriate for use in accordance with embodiments of the present disclosure.
DETAILED DESCRIPTION
In the following description, numerous specific details are set forth in order to provide a thorough understanding of illustrative embodiments of the present disclosure. It will be apparent to one skilled in the art, however, that many embodiments of the present disclosure may be practiced without some or all of the specific details. In some instances, well-known process steps have not been described in detail in order not to unnecessarily obscure various aspects of the present disclosure. Further, it will be appreciated that embodiments of the present disclosure may employ any combination of features described herein. The illustrative examples provided herein are not intended to be exhaustive or to limit the claimed subject matter to the precise forms disclosed.
The present disclosure is directed to computerized image analysis, content-based image retrieval and cellular feature extraction of cells such as oocytes and pronuclear embryos, and their various intracellular components, and predicting outcomes related to cell quality, cell growth, or development (e.g., developmental changes that occur in comparison between oocytes and pronuclear embryos). The present disclosure includes an algorithm for the study of attributes of oocytes and pronuclear embryos that can be used in an automated analytic software tool to identify oocytes and pronuclear embryos with a high likelihood of implanting, among other possible outcomes related to reproductive potential. The present disclosure describes technology with unique clinical benefits distinct from anything commercially available. Disclosed embodiments are particularly significant in their ability to adapt to new cell science; as new outcomes are studied and linked to particular cell features, disclosed embodiments offer the ability to automatically extract those features from training images and associate with them with outcomes, generate decision models, and apply those decision models to new cell images in order to predict outcomes for the new cells.
Software systems for the interpretation of images are recent additions to the catalogue of programs with clinical application. Systems are now available for pattern and texture extraction in defined regions of interest. Extracted features can be added to catalogues to create archives of normal and abnormal images for comparisons and ultimately improved decision making and outcomes as well as possible cost reductions and improvements in work flow within labs. Video image analysis of embryos using dark field assessments has been described. No software, however, has been tested that evaluates attributes of oocytes or pronuclear embryos in standard bright field formats using the techniques described herein.
The present disclosure describes aspects of an automated analysis system that evaluates cells, such as oocytes and pronuclear embryos, in a new way while still being compatible with hardware systems using traditional bright field microscopy. Other computer image analysis systems use time-lapsed dark-field systems, and require separate monitoring hardware and additional personnel for maintenance and monitoring. In addition, data analyzing outcomes with such systems are of variable quality and have not consistently demonstrated any improvement in outcomes.
As suggested above, in contrast to existing systems, embodiments of the present disclosure have several potential marketplace advantages. As one possible advantage, existing hardware (e.g., in embryology labs) can be leveraged for additional benefit using embodiments of the present disclosure. For example, existing microscopy systems present in most labs may be used, with no additional hardware needed. As another possible advantage, existing image analysis software can be extended by incorporating embodiments of the present disclosure into an existing image analysis software system or into an integrated system of microscopy hardware an image analysis software.
The archive of images and outcomes also can be expanded to enhance precision and reliability of the algorithm. Image training can be performed on training images obtained from multiple sources (e.g., clinics, research institutions) stored in a common data repository. Using artificial intelligence, machine self-training can provide improvements over time thus leading to significant improvements in predicting outcomes with continued use. Embodiments described herein can be applied, for example, to both oocyte cryopreservation and embryo development to predict reproductive potential in oocytes prior to cryopreservation or after insemination during embryo development.
Further, embodiments described herein can provide useful analysis of a single cell, potentially offering greater ease in image acquisition and analysis, as well as improved clinical decision making as it relates to cell quality or development, e.g., prior to oocyte freezing for fertility preservation or oocyte donation and improved selection of which oocyte to inseminate and which pronuclear embryo to continue in culture. Embodiments described herein are useful in studying single cells, particularly in the context of reproductive medicine. After insemination and culturing, as the embryo expands in cell number (e.g., beyond 2 days of observation), the image field becomes more crowded, it becomes difficult to isolate individual cells, and the analytics become more complicated. As single cells, oocyte and pronuclear embryos provide excellent opportunities to extract information using computerized image analysis.
The present disclosure describes a quality model learned from existing images and related metadata, such as descriptions of outcomes related to cell quality, cell growth, and development. In the context of reproductive medicine, such outcomes include blastocyst formation, ploidy determinations, and clinical pregnancies. Reproductive potential can be assessed for, e.g., oocytes after freezing and pronuclear embryos. For example, as new forms of data become available and regulatory approvals are obtained for additional study (which may provide additional descriptions of outcomes, or new classes of outcomes), the dataset of existing images, metadata, and outcomes can be extended and the quality models can be improved.
The quality or reproductive potential of an oocyte or pronuclear embryo can be expressed in several different ways, such as the probability that the oocyte will fertilize when exposed to human sperm, the probability that the resulting embryo will reach a given stage of development (e.g., the blastocyst stage), or the probability that the embryo will result in a live birth when transferred to a uterus. In turn, probability can be expressed as a percentage chance, or as a high/medium/low chance, of reaching a given state of development. In some circumstances, there may also be some probability that genetic information (e.g., the genetic complement) or abnormal cell growth outcomes can be predicted, using techniques described in the present disclosure. For example, as new information becomes available linking genetic information or abnormal cell growth with cell features that can be identified in images using the image analysis techniques described herein, these techniques can correspondingly be used to deduce genetic information or abnormal cell growth outcomes from such images. In one possible scenario, instead of an outcome indicating whether an oocyte was successfully fertilized or how embryo development proceeded, an outcome may be whether abnormalities in chromosome number or genetic disorders are detected later on in development.
Additionally, aspects of the present disclosure provide the ability to gauge oocyte quality to guide decision making regarding oocyte cryopreservation. Oocyte preservation may be performed, for example, in an elective oocyte freezing program, in which it may be helpful to determine how many oocytes are to be frozen to maximize future fertility, or to determine a point of diminishing returns beyond which additional oocyte retrievals do not significantly increase the probability of future fertility. Thus, aspects of the present disclosure may be used to avoid performing too many or too few oocyte retrievals. Oocyte preservation also may be performed in an oocyte donation program, where the ability to prospectively assess oocyte potential will guide decision making about how many oocytes to batch for a given donation or whether to discard them and forego fertilization.
Aspects of the present disclosure improve on the current practice of using a single determination of whether an oocyte is mature or immature and freezing 6 to 8 oocytes per batch. Aspects of the present disclosure may be particularly useful in scenarios where the number of oocytes inseminated is limited, and the ability to identify an oocyte with the greatest reproductive potential becomes more important.
FIGURE 1A is an image of a mature oocyte, with the zona pellucida (or "zona") and polar body indicated. In an illustrative embodiment, attributes of the cytoplasm, zona, and polar body can be automatically identified and analyzed, as described in detail below. In an illustrative embodiment, the overall method is composed of two stages, each of which is composed of two phases. Stage one, the model construction stage, involves constructing a decision model (e.g., a model of quality for human oocytes or pronuclear embryos) from training images of cells (e.g., of oocytes or pronuclear embryos for which particular outcomes are known). Stage two, the employment stage, involves employing this model to predict outcomes for new cells being analyzed (e.g., in a clinical setting). For example, in the context of reproductive medicine, the employment stage may be used to determine the quality of an individual embryo.
The method may be performed by an automated image-based cell analysis system, as described herein, implemented in a computer system comprising one or more computing devices. At a high level the automated image-based cell analysis system may include functionality for performing tasks described with reference to FIGURE 2. The automated image-based cell analysis system may include multiple modules or engines to carry out different groups of tasks associated with different stages or phases, such as a model construction engine and a model employment engine, as described in detail below. Such modules or engines may be implemented in the same computing device, or in different computing devices. In turn, these sections or engines may include multiple submodules to carry out more specific tasks within the groups of tasks, such as particular types of feature extraction within the model construction engine.
Phase 1, the first phase of the model construction stage, is the feature extraction phase. In the example shown in FIGURE 2, in phase 1 the automated image-based cell analysis system obtains a set of training images and related metadata from a data store 210 and converts them (e.g., using a model construction engine) into a description document. Like other digital images, the digital images that form the training set comprise pixels arranged in rows and columns. Each pixel is defined by a set of one or more sample values, with the number, meaning, and value range of the respective samples being determined by the format, color space, and bit depth (number of bits per channel) of the images. It will be understood that the techniques described herein can be adapted for analysis of digital images of various formats, resolutions, color spaces, bit depths, etc. For example, although some techniques are described herein as being performed on grayscale images at a predetermined resolution, various techniques for upscaling or downscaling image resolution, converting color images to grayscale, and the like, can be employed in combination with techniques described herein. The feature extraction phase proceeds by extracting features 212 from the training images that describe the cells (e.g., oocytes or pronuclear embryos) in the images numerically, as described in detail below. In the context of reproductive medicine, this may include extraction of features such as the area measurement of the cytoplasm of each oocyte, or cytoplasmic or nuclear features of each pronuclear embryo in the training images. Although the features described herein focus on two-dimensional features such as two-dimensional boundaries, area within such boundaries, etc., it will be understood that such features may be extended to some three-dimensional features, such as volume, to the extent that such features can be accurately estimated based on information obtained from two-dimensional images.
The extracted features are then associated with outcomes in the description document 214. Further details of the features that can be included in a description document are provided below. (The term "document" is used herein to refer to a collection of the described information in any format suitable for further processing as described herein, and need not refer to a traditional document designed for human readability.) The data store may be updated as needed, e.g., to provide additional training images and related metadata.
In an illustrative embodiment, one required item in the description document is the description of one or more outcomes of what happened when the oocytes in the respective training images were fertilized. The outcomes associated with training images, also referred to herein as "Feature 1 " of the respective training images, are not a function of the image data, but are metadata supplied along with the images. The metadata may include information related to developmental outcomes. With regard to a training image of an oocyte, for example, such metadata may include information about whether the oocyte fertilized and how embryo development proceeded.
In at least one embodiment, one or more of the following features 2-18 can be extracted from an oocyte image to create the description document in phase 1, as described below. It will be understood that extraction of some features described below (e.g., cytoplasm features) may be useful for many types of cells. For example, the pronuclear embryo is still a single cell (see FIGURE 5), and its features also can be analyzed as described below: dentification of the boundary of the cytoplasm, measured manually or with image analysis using an edge detection algorithm. An illustrative algorithm 300 is illustrated in FIGURE 3 :
(a) At step 310, convert image from color to grayscale (if needed).
(b) At step 320, apply an edge detection algorithm (e.g., Canny Threshold) to the grayscale image to detect edges.
(c) At step 330, apply a circle detection algorithm (e.g., Hough Circle detection) to find an approximate cytoplasm boundary. In at least one embodiment, the circle detection algorithm may use known quantities of image resolution and magnification to control for size of circles to search from, and Gaussian blur on Canny-detected edges to further reduce number of detected circles, if needed, until only a single circle remains.
If further processing steps that rely on the detected cytoplasm are to be performed, the following steps may be helpful:
(d) At step 340, create an annular mask image from the circle, with inner and outer annulus radii as ratios of circle radius.
(e) At step 350, perform image segmentation on the original source image to identify a connected region as the cytoplasm. For example, a GrabCut method can be used, with the resulting connected region being the cytoplasm, and the boundary of the connected region being the boundary of the cytoplasm. The annular mask can be used to restrict the search.
A mask image may be created containing only pixels from the connected region. This image simplifies computing further cytoplasm features.
otal area of the cytoplasm in the image, defined by the portion of the image inside the boundary identified in feature 2. (The volume of the cytoplasm can also be estimated based on this information.)
spect ratio of the cytoplasm, being the ratio of the shortest distance across the cytoplasm to the longest.
onvexity of the cytoplasm boundary. This feature may relate to comparative changes in the cytoplasm transitioning from oocyte to pronuclear embryo. As an example, convexity may be calculated as follows: (a) Compute the convex hull of the mask image generated in feature 2, step (e), above.
(b) Calculate the ratio of the area of the mask and area of the convex hull.
verage image intensity of the cytoplasm. For example, average intensity may be calculated by averaging the corresponding pixel values of the identified cytoplasm in a grayscale image (in which the value of each pixel is a single sample), or by averaging a particular value for corresponding pixels (such as brightness) when represented in a color space such as HSB (Hue-Saturation-Brightness).
tandard deviation of the image intensity of the cytoplasm.
moothness of the cytoplasm, measured by applying a Gaussian filter to the cytoplasm image, varying the size of the Gaussian filter and recording how that affects the number of features that can be detected in the cytoplasm by, e.g., Canny edge detector, Harris corner detector, or another process which detects features which vary by image smoothness. As the Gaussian smoothing is increased the number of features decreases until no features are detected. The system can track how fast the number of features decreases and how much smoothing is needed to eliminate features entirely, which provides information about the smoothness of the cytoplasm.
ubsampled examination of the cytoplasm, where the cytoplasm area is divided by a grid, and each of the grid areas is subjected to techniques 6, 7 and 8. The results of these procedures can be characterized by:
a. variation of the outcomes for the individual grid areas, or
b. counts generated when given thresholds are exceeded for a grid area.
Identification of the boundary of the polar body, measured manually or with image analysis using an edge detection algorithm.
Indication based on feature 10, above, as to whether the polar body exists for the oocyte or pronuclear embryo.
For the polar body, features analogous to features 3-9, above.
Identification of the inner and outer edges of the zona pellucida of the oocyte or pronuclear embryo. The inner boundary may be coincident with the boundary of the cytoplasm, or gaps may exist between them (perivitelline space). An illustrative algorithm 400 is illustrated in FIGURE 4: (a) At step 410, to identify the outer zona boundary, mask out the region of the image that has been identified as containing the boundary of the cytoplasm in feature 2, above.
(b) At step 420, to identify the inner zona boundary, mask out parts of the image inside the cytoplasm and outside the outer zona boundary, and then identify the inner zona boundary by applying edge detection and circle detection algorithms as in feature 2.
In addition, image sharpening can be performed before the edge detection steps so that the zona edges are more readily apparent.
14. For the zona pellucida, features analogous to features 3-9, above.
15. A measurement of the smoothness of the boundaries of the zona, describing the size and number of features found on the edges.
16. Alignment of the principal axes of the cytoplasm and the zona. For example, the system may calculate axes by finding a minimal bounding rectangle around the mask of the cytoplasm and the zona, with alignment of the sides of the rectangle defining the principal axes. As another example, the system may determine alignment by estimating the oocyte perimeter from an interior reference point (e.g., centroid) and iteratively region-growing no further than the oocyte edge, passing coordinates through this centroid along its major and minor axes to the oocyte edge, and measuring coordinate lengths within the estimated oocyte region.
17. Area of the perivitelline space between the cytoplasm and the zona. This is the area within the inner zona which is not also within the cytoplasm.
18. For the perivitelline space, features analogous to features 3-9, above.
An illustrative description document for 6 images with outcomes and 5 features associated with those images is shown in Table 1, below. Cytoplasm Intensity Intensity std Canny Harris
Image Outcome
size ratio mean dev 1.000000 13.000000
1 1 0.126841 0.478617 0.0623425 0.187357 0.0688431
2 0 0.108994 0.465757 0.0623978 0.198311 0.0519201
3 0 0.117361 0.455848 0.0678655 0.197325 0.0439613
4 1 0.0966727 0.514367 0.0622328 0.0312051 0.00810862
5 1 0.108249 0.513011 0.0877929 0.164327 0.0360605
6 0 0.117361 0.472042 0.0797936 0.179902 0.0625631
Table 1 : Illustrative description document
Other features that may be extracted with image analysis techniques include cytoplasmic texture features, texture distribution, densities, and clustering.
An illustrative image of a pronuclear embryo is shown in FIGURE 5. For pronuclear embryos, in addition to the applicable features described above, any of the additional features 19-23 described below can be extracted for the pronuclei in the center of the cytoplasm, as follows: 19. Identification of the boundaries of the pronuclei, measured manually or with image analysis using an edge detection algorithm.
20. The area or volume of the pronuclei, measured individually and relative to each other.
21. The aspect ratio of the pronuclei measured individually and relative to each other.
22. For the pronuclei, features analogous to features 5-9, above, and comparisons of those features between the two pronuclei.
23. For each of the features that are common to pronuclei and oocyte analysis, generate a set of comparison features by subtraction or division.
The features described herein may be used to detect abnormalities in, for example, texture, inclusion bodies, and shape or size, such as the abnormalities depicted in the oocytes shown in FIGURE IB. In addition to the image analysis tools described above, other tools, such as colorization and color filters, also may be used for feature extraction. For example, to the extent that a particular color filter is useful for detecting a particular feature, the color filter can be applied to a color image of the cell before proceeding with further analysis of the image. This process may be performed, for example, on a color source image prior to conversion of the color source image to grayscale for detection of other features.
As noted above, it is not required for all of the features described above to be extracted in every case, or for all features to be extracted automatically. In some cases, automatic analysis can be performed on manually measured features (e.g., features measured with the help of a customized graphical user interface) or on combinations of manually measured features and automatically extracted features.
Referring again to FIGURE 2, the second phase of the model construction stage (phase 2), is the learning or training phase. Phase 2 determines which features extracted in phase 1 materially affect predictive ability in phase 4. Features selected for inclusion in the decision model may include, for example, features having a numerical predictive value that exceeds a threshold value. Features not selected for inclusion in the decision model, which do not have significant predictive value, can be excluded in phases 3 and 4, which are described in detail below. In phase 2, the model construction engine consumes or reads (e.g., using the model construction engine) the descriptions (e.g., numeric descriptions) in the description document and computes a mathematical optimization to determine which of the features are significant in determining quality of outcomes. Thus, the output of phase 1 is the input for phase 2. The output of phase 2 is a model (see "generate decision model" block 220 shown in FIGURE 2) that can be used to predict outcomes. For example, in the field of reproductive medicine, such predicted outcomes may be used to determine quality or reproductive potential given the numeric features for a given oocyte or pronuclear embryo. This can be referred to as constructing a model with labeled training data. In at least one embodiment, the image analysis and attributes that the software detects are correlated with outcome and weighted according to significance.
In at least one embodiment, a portion of the training data is held out to use for validating the model, and training is done on the remaining training data using cross- validation to ensure the stability of the model. The model can be constructed using logistic regression, decision trees, or some other method that consumes labeled training data. Sufficient training data is required to achieve a significant result from training. When cross-validation implies that the model is predictive, the held-out portion of the training data is used to validate model quality. Training in phase 2 also may be performed using unsupervised learning such as k-means or neural networks to segment the training data into classes. After classes have been identified, the predominant label from the samples in each class is assigned to each class. The learning phase also may be performed using other machine learning methods.
As noted above, the output of phase 2 is a decision model which can be used to predict outcomes, such as cell quality or reproductive potential. As further noted above, the quality or reproductive potential of an oocyte or pronuclear embryo can be expressed in several different ways, such as the probability that the embryo will be chromosomally normal or euploid, or abnormal or aneuploid, the probability that the oocyte will fertilize when exposed to human sperm, the probability that the resulting embryo will reach a given stage of development (e.g., the blastocyst stage), or the probability the embryo will result in a live birth when transferred to a uterus. In turn, probability can be expressed as a percentage chance, or as a high/medium/low chance, of reaching a given state of development.
In phase 3, the first phase of the employment stage, the automated image-based cell analysis system performs (e.g., using the model employment engine) feature extraction on a new image 230 (e.g., an image of a single oocyte or pronuclear embryo). In this phase, the automated image-based cell analysis system generates features 232 for the new image corresponding to one or more features that were generated for the training images in phase 1, other than feature 1 (outcome), which is not yet known. It is not necessary for all of the features described in phase 1 to be extracted for new images in phase 3. For example, features that have been determined in phase 2 not to have predictive value need not be extracted in phase 3.
Although automatic feature extraction has many benefits, as described herein, it is also possible to use the techniques described herein on one or more manually extracted features. For accurate modeling, features that are extracted manually in phase 1 can also be extracted manually in phase 3.
To ensure that values for the features described herein can be compared between images, for both learning and employment phases, a rule that all images be of the same resolution and magnification can be enforced. To comply with this rule, the original images do not necessarily need to be of the same resolution and magnification. Instead, images can be scaled and cropped as needed, either manually or automatically.
In phase 4, the automated image-based cell analysis system predicts (e.g., using the model employment engine) one or more outcomes 242 for the cell (e.g., oocyte or pronuclear embryo) in the new image by applying the decision model 240 generated in phase 2 to the features extracted in phase 3. For example, a score between 0 and 1 can be assigned, with a greater score indicating a better predicted outcome. The predicted outcome adds a significant benefit compared to, for example, older methods of selecting oocytes for insemination based solely on maturity. A threshold score can be set, below which the oocyte regardless of maturity may be discarded as unlikely if not impossible to yield a blastocyst and live birth. Depending on the design of the decision model, the features may be weighted together in some combination, or used as decision points in a decision tree application to determine the likely outcomes for the cell (e.g., oocyte or pronuclear embryo) for one or more of the scenarios provided as outcomes in feature 1 of the training set.
Cells can be analyzed separately, or in combination. For example, the quality of oocytes and pronuclear embryos can be analyzed individually and then combined to form an overall quality score.
As described above, although many features may be extracted, some subsets may be more predictive than others. The system can use various types of predictive analytics to determine which features or combinations of features are more predictive. In an illustrative embodiment, an automated image-based cell analysis system identifies 5 cellular attributes of oocytes. These attributes were selected as possibly predictive of the likelihood reproductive potential of the oocyte and of blastocyst formation and implantation. A numerical score between 0.0 (lowest) and 1.0 (highest) was calculated to estimate these likelihoods. In this illustrative embodiment, the algorithm employs Stochastic Gradient Descent implemented in the Python programming language through the SciPy extension. Feature extraction uses a variety of additional techniques in C++, including a Hough Circle detector and a Harris Corner detector. Main outcome measurements include calculation of an Oocyte Score (OS) to identify oocytes with a high reproductive potential and likelihood of blastocyst formation and to compare the OS (Group 1) to selection using standard morphology (Group 2) and to a computerized video monitoring system (Group 3). In at least one embodiment, an automated image-based cell analysis system can receive software updates over a network, to provide updates to the analysis system, the training data, or other data.
Illustrative Devices and Operating Environments
Unless otherwise specified in the context of specific examples, described techniques and tools may be implemented by any suitable computing device or set of devices.
In any of the described examples, an engine (e.g., a software engine working in combination with computer hardware and microscopy systems) may be used to perform actions described herein. An engine includes logic (e.g., in the form of computer program code) configured to cause one or more computing devices to perform actions described herein as being associated with the engine. For example, a computing device can be specifically programmed to perform the actions by having installed therein a tangible computer-readable medium having computer-executable instructions stored thereon that, when executed by one or more processors of the computing device, cause the computing device to perform the actions. The particular engines described herein are included for ease of discussion, but many alternatives are possible. For example, actions described herein as associated with two or more engines on multiple devices may be performed by a single engine. As another example, actions described herein as associated with a single engine may be performed by two or more engines on the same device or on multiple devices.
In any of the described examples, a data store contains data as described herein and may be hosted, for example, by a database management system (DBMS) to allow a high level of data throughput between the data store and other components of a described system. The DBMS may also allow the data store to be reliably backed up and to maintain a high level of availability. For example, a data store may be accessed by other system components via a network, such as a private network in the vicinity of the system, a secured transmission channel over the public Internet, a combination of private and public networks, and the like. Instead of or in addition to a DBMS, a data store may include structured data stored as files in a traditional file system. Data stores may reside on computing devices that are part of or separate from components of systems described herein. Separate data stores may be combined into a single data store, or a single data store may be split into two or more separate data stores. Some of the functionality described herein may be implemented in the context of a client-server relationship. In this context, server devices may include suitable computing devices configured to provide information and/or services described herein. Server devices may include any suitable computing devices, such as dedicated server devices. Server functionality provided by server devices may, in some cases, be provided by software (e.g., virtualized computing instances or application objects) executing on a computing device that is not a dedicated server device. The term "client" can be used to refer to a computing device that obtains information and/or accesses services provided by a server over a communication link. However, the designation of a particular device as a client device does not necessarily require the presence of a server. At various times, a single device may act as a server, a client, or both a server and a client, depending on context and configuration. Actual physical locations of clients and servers are not necessarily important, but the locations can be described as "local" for a client and "remote" for a server to illustrate a common usage scenario in which a client is receiving information provided by a server at a remote location. Alternatively, a peer-to-peer arrangement, or other models, can be used.
FIGURE 6 is a block diagram that illustrates aspects of an illustrative computing device 600 appropriate for use in accordance with embodiments of the present disclosure. The description below is applicable to servers, personal computers, mobile phones, smart phones, tablet computers, embedded computing devices, and other currently available or yet-to-be-developed devices that may be used in accordance with embodiments of the present disclosure. Computing devices described herein may be integrated with specialized hardware, such as microscopy systems, for obtaining images, or as standalone devices that obtain images for analysis in some other way, such as by receiving images stored remotely in a cloud computing arrangement.
In its most basic configuration, the computing device 600 includes at least one processor 602 and a system memory 604 connected by a communication bus 606. Depending on the exact configuration and type of device, the system memory 604 may be volatile or nonvolatile memory, such as read only memory ("ROM"), random access memory ("RAM"), EEPROM, flash memory, or other memory technology. Those of ordinary skill in the art and others will recognize that system memory 604 typically stores data and/or program modules that are immediately accessible to and/or currently being operated on by the processor 602. In this regard, the processor 602 may serve as a computational center of the computing device 600 by supporting the execution of instructions.
As further illustrated in FIGURE 6, the computing device 600 may include a network interface 610 comprising one or more components for communicating with other devices over a network. Embodiments of the present disclosure may access basic services that utilize the network interface 610 to perform communications using common network protocols. The network interface 610 may also include a wireless network interface configured to communicate via one or more wireless communication protocols, such as WiFi, 2G, 3G, 4G, LTE, WiMAX, Bluetooth, and/or the like.
In the illustrative embodiment depicted in FIGURE 6, the computing device 600 also includes a storage medium 608. However, services may be accessed using a computing device that does not include means for persisting data to a local storage medium. Therefore, the storage medium 608 depicted in FIGURE 6 is optional. In any event, the storage medium 608 may be volatile or nonvolatile, removable or nonremovable, implemented using any technology capable of storing information such as, but not limited to, a hard drive, solid state drive, CD-ROM, DVD, or other disk storage, magnetic tape, magnetic disk storage, and/or the like.
As used herein, the term "computer-readable medium" includes volatile and nonvolatile and removable and nonremovable media implemented in any method or technology capable of storing information, such as computer-readable instructions, data structures, program modules, or other data. In this regard, the system memory 604 and storage medium 608 depicted in FIGURE 6 are examples of computer-readable media.
For ease of illustration and because it is not important for an understanding of the claimed subject matter, FIGURE 6 does not show some of the typical components of many computing devices. In this regard, the computing device 600 may include input devices, such as a keyboard, keypad, mouse, trackball, microphone, video camera, touchpad, touchscreen, electronic pen, stylus, and/or the like. Such input devices may be coupled to the computing device 600 by wired or wireless connections including RF, infrared, serial, parallel, Bluetooth, USB, or other suitable connection protocols using wireless or physical connections.
In any of the described examples, input data can be captured by input devices and processed, transmitted, or stored (e.g., for future processing). The processing may include encoding data streams, which can be subsequently decoded for presentation by output devices. Media data can be captured by multimedia input devices and stored by saving media data streams as files on a computer-readable storage medium (e.g., in memory or persistent storage on a client device, server, administrator device, or some other device). Input devices can be separate from and communicatively coupled to computing device 600 (e.g., a client device), or can be integral components of the computing device 600. In some embodiments, multiple input devices may be combined into a single, multifunction input device (e.g., a video camera with an integrated microphone). The computing device 600 may also include output devices such as a display, speakers, printer, etc. The output devices may include video output devices such as a display or touchscreen. The output devices also may include audio output devices such as external speakers or earphones. The output devices can be separate from and communicatively coupled to the computing device 600, or can be integral components of the computing device 600. Input functionality and output functionality may be integrated into the same input/output device (e.g., a touchscreen). Any suitable input device, output device, or combined input/output device either currently known or developed in the future may be used with described systems.
In general, functionality of computing devices described herein may be implemented in computing logic embodied in hardware or software instructions, which can be written in a programming language, such as C, C++, COBOL, JAVA™, PHP, Perl, Python, Ruby, HTML, CSS, JavaScript, VBScript, ASPX, Microsoft .NET™ languages such as C#, and/or the like. Computing logic may be compiled into executable programs or written in interpreted programming languages. Generally, functionality described herein can be implemented as logic modules that can be duplicated to provide greater processing capability, merged with other modules, or divided into sub-modules. The computing logic can be stored in any type of computer-readable medium (e.g., a non-transitory medium such as a memory or storage medium) or computer storage device and be stored on and executed by one or more general-purpose or special-purpose processors, thus creating a special-purpose computing device configured to provide functionality described herein.
Extensions and Alternatives
Many alternatives to the systems and devices described herein are possible. For example, individual modules or subsystems can be separated into additional modules or subsystems or combined into fewer modules or subsystems. As another example, modules or subsystems can be omitted or supplemented with other modules or subsystems. As another example, functions that are indicated as being performed by a particular device, module, or subsystem may instead be performed by one or more other devices, modules, or subsystems. Although some examples in the present disclosure include descriptions of devices comprising specific hardware components in specific arrangements, techniques and tools described herein can be modified to accommodate different hardware components, combinations, or arrangements. Further, although some examples in the present disclosure include descriptions of specific usage scenarios, techniques and tools described herein can be modified to accommodate different usage scenarios. Functionality that is described as being implemented in software can instead be implemented in hardware, or vice versa.
Many alternatives to the techniques described herein are possible. For example, processing stages in the various techniques can be separated into additional stages or combined into fewer stages. As another example, processing stages in the various techniques can be omitted or supplemented with other techniques or processing stages. As another example, processing stages that are described as occurring in a particular order can instead occur in a different order. As another example, processing stages that are described as being performed in a series of steps may instead be handled in a parallel fashion, with multiple modules or software processes concurrently handling one or more of the illustrated processing stages. As another example, processing stages that are indicated as being performed by a particular device or module may instead be performed by one or more other devices or modules.
The principles, representative embodiments, and modes of operation of the present disclosure have been described in the foregoing description. However, aspects of the present disclosure which are intended to be protected are not to be construed as limited to the particular embodiments disclosed. Further, the embodiments described herein are to be regarded as illustrative rather than restrictive. It will be appreciated that variations and changes may be made by others, and equivalents employed, without departing from the spirit of the present disclosure. Accordingly, it is expressly intended that all such variations, changes, and equivalents fall within the spirit and scope of the claimed subject matter.

Claims

CLAIMS The embodiments of the invention in which an exclusive property or privilege is claimed are defined as follows:
1. A computer-implemented method comprising:
automatically converting a set of training images of cells and related outcome metadata into a description document by extracting features from the training images that describe the cells and associating the extracted features with the outcome metadata, wherein the training images are obtained using bright-field microscopy;
based on the description document, automatically computing a decision model that can be used to predict outcomes of new cells.
2. The method of Claim 1, wherein the extracted features include a boundary of a cytoplasm.
3. The method of Claim 2, wherein the boundary of a cytoplasm is calculated by:
applying an edge detection algorithm to the images to detect edges; and applying a circle detection algorithm to the detected edges to detect an approximate cytoplasm boundary.
4. The method of Claim 3, wherein at least one of the images is a color image, the method further comprising converting the color image to a grayscale image prior to application of the edge detection algorithm.
5. The method of Claim 3 or Claim 4 further comprising:
creating an annular mask image from the approximate cytoplasm boundary; and performing image segmentation on the original source image to identify a connected region as the cytoplasm.
6. The method of any of Claims 2-5, wherein the extracted features further include one or more of the following:
total area of the cytoplasm;
aspect ratio of the cytoplasm;
convexity of the boundary of the cytoplasm; average image intensity of the cytoplasm;
standard deviation of the image intensity of the cytoplasm;
smoothness of the cytoplasm;
sub sampled image intensity or smoothness for grid areas within the cytoplasm; cytoplasmic texture or texture distribution;
density; and
clustering.
7. The method of Claim 6, wherein the average image intensity of the cytoplasm is calculated by averaging the corresponding pixel values of the identified cytoplasm in a grayscale image.
8. The method of any of Claims 1-7, wherein the cells are oocytes or pronuclear embryos.
9. The method of Claim 8, wherein the extracted features include one or more of the following:
a boundary of a polar body, measured manually or with image analysis using an edge detection algorithm;
an indication as to whether the polar body exists;
identification of inner and outer edges of a zona pellucida;
identification of perivitelline space;
a measurement of smoothness of the edges of the zona pellucida;
alignment of principal axes of the cytoplasm and the zona;
total area of the polar body, the zona pellucida, or the perivitelline space;
aspect ratio of the polar body, the zona pellucida, or the perivitelline space;
convexity of a boundary of the polar body, the zona pellucida, or the perivitelline space;
average image intensity of the polar body, the zona pellucida, or the perivitelline space;
standard deviation of the image intensity of the polar body, the zona pellucida, or the perivitelline space;
smoothness of the polar body, the zona pellucida, or the perivitelline space; and subsampled image intensity or smoothness for grid areas within the polar body, the zona pellucida, or the perivitelline space.
10. The method of Claim 8 or Claim 9, wherein the extracted features include one or more of the following for a pronuclear embryo comprising pronuclei:
identification of the boundaries of the pronuclei, measured manually or with image analysis using an edge detection algorithm;
area of the pronuclei, measured individually and/or relative to each other;
aspect ratio of the pronuclei, measured individually and/or relative to each other; convexity of a boundary of one or more of the pronuclei;
average image intensity of one or more of the pronuclei;
standard deviation of the image intensity of one or more of the pronuclei;
smoothness of one or more of the pronuclei; and
subsampled image intensity or smoothness for grid areas within one or more of the pronuclei.
11. The method of any of Claims 1-10, further comprising:
automatically extracting features from images that describe new cells, wherein the images that describe the new cells are obtained using bright-field microscopy;
predicting one or more outcomes for the new cells by applying the decision model to the features extracted from the images that describe the new cells.
12. The method of Claim 11, wherein the features extracted from the images that describe new cells correspond to features selected for inclusion in the decision model.
13. The method of Claim 11 or Claim 12, wherein the one or more outcomes include one or more of quality or reproductive potential for a new oocyte, whether abnormalities in chromosome number are detected later in development, whether genetic disorders are detected later in development, and abnormal cell growth outcomes.
14. The method of Claim 13, wherein the quality or reproductive potential of the new oocyte is expressed as one or more of: a probability that the oocyte will fertilize when exposed to human sperm, a probability that a resulting embryo will reach a given stage of development, or a probability that the resulting embryo will result in a live birth when transferred to a uterus.
15. A computer-implemented method comprising:
automatically extracting features that describe new cells from new cell images; and
predicting one or more outcomes for the new cells by applying a decision model to the features extracted from the new cell images, wherein the decision model is based on a description document that associates outcome metadata with features extracted from training images, wherein the training images and the new cell images are obtained using bright-field microscopy.
16. The method of Claim 15, wherein the extracted features include a boundary of a cytoplasm.
17. The method of Claim 16, wherein the boundary of a cytoplasm is calculated by:
applying an edge detection algorithm to detect edges; and
applying a circle detection algorithm to the detected edges to detect an approximate cytoplasm boundary.
18. The method of Claim 17, wherein at least one of the images is a color image, the method further comprising converting the color image to a grayscale image prior to application of the edge detection algorithm.
19. The method of Claim 18 further comprising:
creating an annular mask image from the approximate cytoplasm boundary; and performing image segmentation on the original source image to identify a connected region as the cytoplasm.
20. The method of any of Claims 16-19, wherein the extracted features further include one or more of the following:
total area of the cytoplasm;
aspect ratio of the cytoplasm;
convexity of the boundary of the cytoplasm; average image intensity of the cytoplasm;
standard deviation of the image intensity of the cytoplasm;
smoothness of the cytoplasm;
sub sampled image intensity or smoothness for grid areas within the cytoplasm; cytoplasmic texture or texture distribution;
density; and
clustering.
21. The method of Claim 20, wherein the average image intensity of the cytoplasm is calculated by averaging the corresponding pixel values of the identified cytoplasm in a grayscale image.
22. The method of any of Claims 15-21, wherein the new cells are oocytes or pronuclear embryos.
23. The method of Claim 22, wherein the features extracted from the new cell images and the features extracted from the training images include one or more of the following:
a boundary of a polar body, measured manually or with image analysis using an edge detection algorithm;
an indication as to whether the polar body exists;
identification of inner and outer edges of a zona pellucida;
identification of perivitelline space;
a measurement of smoothness of the edges of the zona pellucida;
alignment of principal axes of the cytoplasm and the zona;
total area of the polar body, the zona pellucida, or the perivitelline space;
aspect ratio of the polar body, the zona pellucida, or the perivitelline space;
convexity of the boundary of the polar body, the zona pellucida, or the perivitelline space;
average image intensity of the polar body, the zona pellucida, or the perivitelline space;
standard deviation of the image intensity of the polar body, the zona pellucida, or the perivitelline space;
smoothness of the polar body, the zona pellucida, or the perivitelline space; and subsampled image intensity or smoothness for grid areas within the polar body, the zona pellucida, or the perivitelline space.
24. The method of Claim 22 or Claim 23, wherein the features extracted from the new cell images and the features extracted from the training images include one or more of the following for a pronuclear embryo comprising pronuclei:
identification of the boundaries of the pronuclei, measured manually or with image analysis using an edge detection algorithm;
area of the pronuclei, measured individually and/or relative to each other;
aspect ratio of the pronuclei, measured individually and/or relative to each other; convexity of a boundary of one or more of the pronuclei;
average image intensity of one or more of the pronuclei;
standard deviation of the image intensity of one or more of the pronuclei;
smoothness of one or more of the pronuclei; and
subsampled image intensity or smoothness for grid areas within one or more of the pronuclei.
25. The method of any of Claims 22-24, wherein the one or more outcomes include one or more of quality or reproductive potential for a new oocyte, whether abnormalities in chromosome number are detected later in development, whether genetic disorders are detected later in development, and abnormal cell growth outcomes.
26. The method of Claim 25, wherein the quality or reproductive potential of the new oocyte is expressed as one or more of: a probability that the oocyte will fertilize when exposed to human sperm, a probability that a resulting embryo will reach a given stage of development, or a probability that the resulting embryo will result in a live birth when transferred to a uterus.
27. A computer-readable storage medium having stored thereon computer-executable instructions configured to cause one or more computing devices to perform the method of any of Claims 1-26.
28. A system comprising one or more computing devices programmed to perform the method of any of Claims 1-26.
29. The system of Claim 28 further comprising a microscopy system integrated with the one or more computing devices.
PCT/US2017/015645 2016-01-28 2017-01-30 Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos WO2017132674A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
CA3013048A CA3013048A1 (en) 2016-01-28 2017-01-30 Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos
EP17745088.9A EP3408796A4 (en) 2016-01-28 2017-01-30 Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos
US16/073,126 US10902334B2 (en) 2016-01-28 2017-01-30 Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos
JP2018559173A JP2019509057A (en) 2016-01-28 2017-01-30 Automated image analysis for assessing fertility of human oocytes and pronuclear embryos

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201662288379P 2016-01-28 2016-01-28
US62/288,379 2016-01-28

Publications (1)

Publication Number Publication Date
WO2017132674A1 true WO2017132674A1 (en) 2017-08-03

Family

ID=59398851

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2017/015645 WO2017132674A1 (en) 2016-01-28 2017-01-30 Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos

Country Status (5)

Country Link
US (1) US10902334B2 (en)
EP (1) EP3408796A4 (en)
JP (1) JP2019509057A (en)
CA (1) CA3013048A1 (en)
WO (1) WO2017132674A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2019088243A (en) * 2017-11-15 2019-06-13 大日本印刷株式会社 Image analysis device
WO2020016223A1 (en) * 2018-07-18 2020-01-23 Albert-Ludwigs-Universität Freiburg Method for setting up an apparatus for biological processes and apparatus for biological processes
WO2020058931A1 (en) * 2018-09-20 2020-03-26 Aivf Ltd Image feature detection
EP3644228A1 (en) * 2018-10-23 2020-04-29 The Chinese University Of Hong Kong Method and apparatus for segmenting cellular image
CN111279421A (en) * 2017-09-29 2020-06-12 布里格姆妇女医院 Automated evaluation of human embryos
JP2021523355A (en) * 2018-05-02 2021-09-02 アンドロヴィア ライフサイエンシーズ, エルエルシー Methods and test kits for determining male fertility status

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3456812A4 (en) * 2016-06-21 2019-06-26 Sony Corporation Information processing device, information processing method, and program
JP7020156B2 (en) * 2018-02-06 2022-02-16 オムロン株式会社 Evaluation device, motion control device, evaluation method, and evaluation program
US10552957B2 (en) * 2018-03-24 2020-02-04 Dan Nayot Methods and systems for determining quality of an oocyte
WO2020033391A1 (en) * 2018-08-07 2020-02-13 Cornell University System and method for selecting artificially fertilized embryos
US11832872B2 (en) * 2019-04-01 2023-12-05 Anya L. Getman Resonating probe with optional sensor, emitter, and/or injection capability
WO2020198779A1 (en) 2019-04-04 2020-10-08 Presagen Pty Ltd Method and system for selecting embryos
CN110363218B (en) * 2019-06-06 2023-07-11 张孝东 Noninvasive embryo assessment method and device
EP4026089A4 (en) * 2019-09-06 2022-11-09 The Brigham & Women's Hospital, Inc. Automated evaluation of quality assurance metrics for assisted reproduction procedures
WO2021177142A1 (en) * 2020-03-05 2021-09-10 ネクスジェン株式会社 System, information processor, server, terminal, program or method for supporting assisted reproduction
CN111445511B (en) * 2020-03-24 2022-08-05 杭州东信北邮信息技术有限公司 Method for detecting circle in image
US20230326165A1 (en) 2020-09-11 2023-10-12 The Brigham And Women's Hospital, Inc. Determining locations in reproductive cellular structures
US20220383497A1 (en) * 2021-05-28 2022-12-01 Daniel Needleman Automated analysis and selection of human embryos
WO2022260740A1 (en) 2021-06-10 2022-12-15 Alife Health Inc. Machine learning for optimizing ovarian stimulation
CN115239715B (en) * 2022-09-22 2022-12-09 中南大学 Method, system, equipment and storage medium for predicting development result of blastocyst
US12020782B1 (en) 2023-03-24 2024-06-25 Inti Taiwan, Inc. Oocyte quality analysis system
CN116912255B (en) * 2023-09-14 2023-12-19 济南宝林信息技术有限公司 Follicular region segmentation method for ovarian tissue analysis

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110254943A1 (en) * 2010-04-06 2011-10-20 Institute For Systems Biology Automated analysis of images using bright field microscopy
WO2015189264A1 (en) * 2014-06-10 2015-12-17 Ventana Medical Systems, Inc. Predicting breast cancer recurrence directly from image features computed from digitized immunohistopathology tissue slides

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9285575B2 (en) * 2009-01-26 2016-03-15 President And Fellows Of Harvard College Systems and methods for selective detection and imaging in coherent Raman microscopy by spectral excitation shaping
US9652585B2 (en) * 2010-03-16 2017-05-16 Bluegnome Limited Comparative genomic hybridization array method for preimplantation genetic screening
WO2012016242A2 (en) * 2010-07-30 2012-02-02 Aureon Biosciences, Inc. Systems and methods for segmentation and processing of tissue images and feature extraction from same for treating, diagnosing, or predicting medical conditions
BR102012031953B1 (en) 2012-12-14 2020-04-07 Univ Estadual Paulista Julio De Mesquita Filho Unesp method for determining embryo viability and quality
US9177192B2 (en) 2013-02-28 2015-11-03 Progyny, Inc. Apparatus, method, and system for image-based human embryo cell classification
JP6343935B2 (en) 2014-01-09 2018-06-20 大日本印刷株式会社 Growth information management system and growth information management program
WO2015161045A1 (en) 2014-04-16 2015-10-22 President And Fellows Of Harvard College Non-linear imaging systems and methods for assisted reproductive technologies
CN104462673A (en) 2014-11-25 2015-03-25 柳州市妇幼保健院 Computer assisted embryo selection method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110254943A1 (en) * 2010-04-06 2011-10-20 Institute For Systems Biology Automated analysis of images using bright field microscopy
WO2015189264A1 (en) * 2014-06-10 2015-12-17 Ventana Medical Systems, Inc. Predicting breast cancer recurrence directly from image features computed from digitized immunohistopathology tissue slides

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See also references of EP3408796A4 *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111279421A (en) * 2017-09-29 2020-06-12 布里格姆妇女医院 Automated evaluation of human embryos
CN111279421B (en) * 2017-09-29 2024-04-09 布里格姆妇女医院 Automatic evaluation of human embryos
JP2019088243A (en) * 2017-11-15 2019-06-13 大日本印刷株式会社 Image analysis device
JP7097686B2 (en) 2017-11-15 2022-07-08 大日本印刷株式会社 Image analyzer
JP2021523355A (en) * 2018-05-02 2021-09-02 アンドロヴィア ライフサイエンシーズ, エルエルシー Methods and test kits for determining male fertility status
WO2020016223A1 (en) * 2018-07-18 2020-01-23 Albert-Ludwigs-Universität Freiburg Method for setting up an apparatus for biological processes and apparatus for biological processes
WO2020058931A1 (en) * 2018-09-20 2020-03-26 Aivf Ltd Image feature detection
US11954926B2 (en) 2018-09-20 2024-04-09 Aivf Ltd. Image feature detection
EP3644228A1 (en) * 2018-10-23 2020-04-29 The Chinese University Of Hong Kong Method and apparatus for segmenting cellular image
CN111091568A (en) * 2018-10-23 2020-05-01 香港中文大学 Method and apparatus for segmenting cell images
US11210494B2 (en) 2018-10-23 2021-12-28 The Chinese University Of Hong Kong Method and apparatus for segmenting cellular image
CN111091568B (en) * 2018-10-23 2023-06-02 香港中文大学 Method and apparatus for segmenting cell images

Also Published As

Publication number Publication date
JP2019509057A (en) 2019-04-04
CA3013048A1 (en) 2017-08-03
US10902334B2 (en) 2021-01-26
US20190042958A1 (en) 2019-02-07
EP3408796A1 (en) 2018-12-05
EP3408796A4 (en) 2019-10-16

Similar Documents

Publication Publication Date Title
US10902334B2 (en) Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos
CN106570505B (en) Method and system for analyzing histopathological images
EP3176563B1 (en) Identification device and identification method
US20200126207A1 (en) System and Method Associated with Predicting Segmentation Quality of Objects in Analysis of Copious Image Data
US20220215548A1 (en) Method and device for identifying abnormal cell in to-be-detected sample, and storage medium
CN113658174B (en) Microkernel histology image detection method based on deep learning and image processing algorithm
CN112347977B (en) Automatic detection method, storage medium and device for induced pluripotent stem cells
CN112215801A (en) Pathological image classification method and system based on deep learning and machine learning
US20240029266A1 (en) Characterization platform for scalable, spatially-resolved multispectral analysis of tissue
US20150242676A1 (en) Method for the Supervised Classification of Cells Included in Microscopy Images
Kovalev et al. Deep learning in big image data: Histology image classification for breast cancer diagnosis
CN113837255B (en) Method, apparatus and medium for predicting cell-based antibody karyotype class
US20220092774A1 (en) Method and apparatus for analysis of histopathology image data
CN116415020A (en) Image retrieval method, device, electronic equipment and storage medium
CN114358279A (en) Image recognition network model pruning method, device, equipment and storage medium
Aristoteles et al. Identification of human sperm based on morphology using the you only look once version 4 algorithm
CN114118410A (en) Method, device and storage medium for extracting node feature of graph structure
CN109598712A (en) Quality determining method, device, server and the storage medium of plastic foam cutlery box
US10748288B2 (en) Methods and systems for determining quality of an oocyte
KR20210113539A (en) Method, apparatus and computer program for classifying chromosomes using artificial intelligence
Maboudi et al. Drone-based container crane inspection: concept, challenges and preliminary results
CN115004241B (en) Deep learning based shift invariant loss of image segmentation
López-Tapia et al. A fast pyramidal bayesian model for mitosis detection in whole-slide images
CN117975204B (en) Model training method, defect detection method and related device
CN117830273A (en) HE staining image mitosis automatic detection and counting method based on rapid spatial pyramid pooling

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17745088

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2018559173

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 3013048

Country of ref document: CA

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2017745088

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 2017745088

Country of ref document: EP

Effective date: 20180828