US20230326014A1 - Automated aneuploidy screening using arbitrated ensembles - Google Patents
Automated aneuploidy screening using arbitrated ensembles Download PDFInfo
- Publication number
- US20230326014A1 US20230326014A1 US18/025,475 US202118025475A US2023326014A1 US 20230326014 A1 US20230326014 A1 US 20230326014A1 US 202118025475 A US202118025475 A US 202118025475A US 2023326014 A1 US2023326014 A1 US 2023326014A1
- Authority
- US
- United States
- Prior art keywords
- embryo
- parameter
- clinical parameter
- clinical
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 208000036878 aneuploidy Diseases 0.000 title claims abstract description 43
- 231100001075 aneuploidy Toxicity 0.000 title claims abstract description 38
- 238000012216 screening Methods 0.000 title claims abstract description 20
- 210000001161 mammalian embryo Anatomy 0.000 claims abstract description 142
- 238000013528 artificial neural network Methods 0.000 claims abstract description 56
- 238000000034 method Methods 0.000 claims abstract description 55
- 239000002131 composite material Substances 0.000 claims abstract description 27
- 241000282414 Homo sapiens Species 0.000 claims abstract description 26
- 230000015654 memory Effects 0.000 claims description 28
- 210000000287 oocyte Anatomy 0.000 claims description 26
- 210000002257 embryonic structure Anatomy 0.000 claims description 14
- 238000013527 convolutional neural network Methods 0.000 claims description 10
- 238000012706 support-vector machine Methods 0.000 claims description 10
- 230000000306 recurrent effect Effects 0.000 claims description 9
- 230000004720 fertilization Effects 0.000 claims description 2
- 238000003745 diagnosis Methods 0.000 claims 2
- 230000003054 hormonal effect Effects 0.000 claims 1
- 230000006870 function Effects 0.000 description 19
- 238000012545 processing Methods 0.000 description 15
- 238000012549 training Methods 0.000 description 11
- 230000008569 process Effects 0.000 description 9
- 238000004891 communication Methods 0.000 description 8
- 238000004422 calculation algorithm Methods 0.000 description 7
- 238000010586 diagram Methods 0.000 description 7
- 238000004458 analytical method Methods 0.000 description 6
- 230000003322 aneuploid effect Effects 0.000 description 5
- 238000003066 decision tree Methods 0.000 description 5
- 230000001850 reproductive effect Effects 0.000 description 5
- 238000013459 approach Methods 0.000 description 4
- 210000003850 cellular structure Anatomy 0.000 description 4
- 238000003384 imaging method Methods 0.000 description 4
- 230000003068 static effect Effects 0.000 description 4
- 210000002569 neuron Anatomy 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 210000002459 blastocyst Anatomy 0.000 description 2
- 210000000349 chromosome Anatomy 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 230000036541 health Effects 0.000 description 2
- 230000036512 infertility Effects 0.000 description 2
- 208000000509 infertility Diseases 0.000 description 2
- 231100000535 infertility Toxicity 0.000 description 2
- 238000010801 machine learning Methods 0.000 description 2
- 229920000747 poly(lactic acid) Polymers 0.000 description 2
- 239000004626 polylactic acid Substances 0.000 description 2
- 230000035935 pregnancy Effects 0.000 description 2
- 230000011218 segmentation Effects 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- NIXOWILDQLNWCW-UHFFFAOYSA-N acrylic acid group Chemical group C(C=C)(=O)O NIXOWILDQLNWCW-UHFFFAOYSA-N 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 230000004931 aggregating effect Effects 0.000 description 1
- 230000002776 aggregation Effects 0.000 description 1
- 238000004220 aggregation Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000001574 biopsy Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000007635 classification algorithm Methods 0.000 description 1
- 230000001427 coherent effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000005138 cryopreservation Methods 0.000 description 1
- 230000001186 cumulative effect Effects 0.000 description 1
- 230000009429 distress Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 230000035558 fertility Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000007499 fusion processing Methods 0.000 description 1
- 230000002068 genetic effect Effects 0.000 description 1
- 230000012447 hatching Effects 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000000338 in vitro Methods 0.000 description 1
- 238000002347 injection Methods 0.000 description 1
- 239000007924 injection Substances 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000000877 morphologic effect Effects 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 238000007637 random forest analysis Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000006403 short-term memory Effects 0.000 description 1
- 239000000243 solution Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 238000013526 transfer learning Methods 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/698—Matching; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/70—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30044—Fetus; Embryo
Definitions
- the present invention relates generally to the field of assisted fertility, and more particularly, to automated aneuploidy screening using arbitrated ensembles of predictive models.
- Infertility is an underestimated healthcare problem that affects over forty-eight million couples globally and is a cause of distress, depression, and discrimination.
- assisted reproductive technologies such as in-vitro fertilization (IVF)
- IVVF in-vitro fertilization
- IVF in-vitro fertilization
- IVF remains as an expensive solution, with a cost between $7000 and $20,000 per ART cycle in the US, which is generally not covered by insurance.
- many patients require multiple cycles of IVF to achieve pregnancy.
- Non-invasive selection of the top-quality embryo for transfer is one of the most important factors in achieving successful ART outcomes, yet this critical step remains a significant challenge.
- PGT-A Preimplantation genetic testing for aneuploidy
- a method for fully automated screening for aneuploidy in a human embryo.
- An image of the embryo is obtained at an associated imager and provided to a neural network to generate a first clinical parameter.
- a set of at least one parameter representing one of biometric parameters of one of a patient receiving the embryo, an egg utilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided sperm used to create the embryo, and an egg donor who provided the egg is retrieved, and a second clinical parameter is generated from the set of at least one parameter at a predictive model.
- a composite parameter, representing a likelihood of aneuploidy in the embryo is generated from the first clinical parameter and the second clinical parameter.
- a system for fully automated screening for aneuploidy in a human embryo.
- the system includes a processor and a non-transitory computer readable medium storing instructions executable by the processor.
- the machine executable instructions include an imager interface configured to receive an image of the embryo from an associated imager and a neural network configured to generate a first clinical parameter from the image of the embryo.
- a predictive model is configured to generate a second clinical parameter from a set of at least one parameter representing one of a patient receiving the embryo, an oocyte fertilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided the sperm used to create the embryo, and an egg donor who provided the oocyte from an associated memory.
- An arbitrator is configured to generate a composite parameter, representing a likelihood of aneuploidy in the embryo, from the first clinical parameter and the second clinical parameter.
- a method for fully automated screening for aneuploidy in a human embryo.
- An image of the embryo is obtained at an associated imager and a first clinical parameter is generated from the image of the embryo at a convolutional neural network.
- a set of parameters is retrieved from an associated memory. The set of parameters included each of an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo.
- a composite parameter, representing a likelihood of aneuploidy in the embryo, is generated from the first clinical parameter, the second clinical parameter, and the third clinical parameter.
- FIG. 1 illustrates one example of a system for fully automated screening for aneuploidy in a human embryo
- FIG. 2 illustrates another example of a system for fully automated screening for aneuploidy in a human embryo
- FIG. 3 illustrates an example of a method for screening for aneuploidy in a human embryo
- FIG. 4 illustrates another example of a method for screening for aneuploidy in a human embryo
- FIG. 5 is a schematic block diagram illustrating an exemplary system of hardware components capable of implementing examples of the systems and methods disclosed herein.
- Emulating the skill of highly trained embryologists in efficient embryo assessment in a fully automated system is a major unmet challenge in all of the previous work done in embryo computer-aided assessment.
- Current computer vision methods for embryo assessment are semi-automated, limited to measuring specific parameters providing metrics that require further analysis by embryologists, and require strictly controlled imaging systems.
- Previous attempts in developing systems using machine-learning approaches have required intensive image preprocessing followed by human-directed segmentation of embryo features for classification. Owing to the dependency of machine-learning approaches on image processing and segmentation, such methods suffer from the same limitations as computer vision techniques.
- a deep neural networks pretrained with a large set of images, paired with biometric information about the egg, sperm, and their respective donors, to screen embryos for aneuploidy.
- the systems and methods provided herein allows for automated screening for aneuploidy without any assistance by an embryologist.
- a convolutional neural network is applied to identify the shape, structure, and texture variations between morphologically complex embryos, while one or more other predictive models are employed to evaluate the biometric data.
- the system is resilient to changes in image illumination and quality due to data acquisition using multiple instruments. Further, since the additional predictive models can be trained independently of the neural network, the system is effectively modular, allowing for addition or substitution of models in response to new applications.
- continuous parameter is used herein to distinguish numerical values from categorical values or classes, and should be ready to include both truly continuous data as well as data more traditionally referred to as discrete data.
- an “average” of a set of values is any measure of central tendency. For continuous parameters, this includes any of the median, arithmetic mean, and geometric mean of the values. For categorical parameters, the average is the mode of the set of values.
- a “static observation,” as used herein, is an image or group of images of an embryo that represent a single point in the development of the embryo. Where multiple images are used in a static observation, no discernible change in the structure and appearance of the embryo will have taken place between images. Data collected from a static observation represents morphological data for the embryo.
- a “subset” of a set as used here refers to a set containing some or all elements of the set.
- a “proper subset” of a set contains less than all of the elements of the set.
- an embryo is “euploid” when is has forty-six chromosomes.
- An embryo is “aneuploid” when it has a number of chromosomes that is different from the usual forty-six.
- the “ploidy status” of the embryo is the embryos status as either euploid or aneuploid, represented either categorically or as a probability.
- FIG. 1 illustrates one example of a system 100 for fully automated screening for aneuploidy in a human embryo.
- the system 100 includes a processor 102 and a non-transitory computer readable medium 110 that stores machine executable instructions for assigning a value representing a location of interest within a reproductive cellular structure.
- the machine executable instructions include an imager interface 112 that receives an image of embryo from an associated imager.
- the imaging interface 112 can receive the image from the imager via a bus or network connection and condition the image for analysis at a neural network 114 .
- the neural network 114 can be implemented on a cloud computing system, with the image transmitted to the server containing the neural network 114 via a network interface (not shown).
- the neural network 114 determines, from the image of the reproductive cellular structure, a first clinical parameter representing a ploidy status of the embryo.
- the first clinical parameter is a categorical parameter with a first value representing aneuploidy and a second value representing euploidy.
- the first clinical parameter is a continuous parameter representing a likelihood of either aneuploidy or euploidy of the embryo.
- the neural network 114 includes a plurality of nodes having a plurality of interconnections. Values from the image, for example luminance and/or chrominance values associated with the individual pixels, are provided to a plurality of input nodes. The input nodes each provide these input values to layers of one or more intermediate nodes.
- a given intermediate node receives one or more output values from previous nodes.
- the received values are weighted according to a series of weights established during the training of the classifier.
- An intermediate node translates its received values into a single output according to an activation function at the node.
- the intermediate node can sum the received values and subject the sum to an identify function, a step function, a sigmoid function, a hyperbolic tangent, a rectified linear unit, a leaky rectified linear unit, a parametric rectified linear unit, a Gaussian error linear unit, the softplus function, an exponential linear unit, a scaled exponential linear unit, a Gaussian function, a sigmoid linear unit, a growing cosine unit, the Heaviside function, and the mish function.
- a final layer of nodes provides the confidence values for the output classes of the neural network, with each node having an associated value representing a confidence for one of the associated output classes of the classifier.
- ANN classifiers are fully-connected and feedforward.
- a convolutional neural network includes convolutional layers in which nodes from a previous layer are only connected to a subset of the nodes in the convolutional layer.
- Recurrent neural networks are a class of neural networks in which connections between nodes form a directed graph along a temporal sequence. Unlike a feedforward network, recurrent neural networks can incorporate feedback from states caused by earlier inputs, such that an output of the recurrent neural network for a given input can be a function of not only the input but one or more previous inputs.
- LSTM Long Short-Term Memory
- LSTM Long Short-Term Memory
- neural networks are trained in an adversarial manner in which a generative classifier is configured to generate new instances for a discriminative classifier.
- the generative network learns to map from a latent space to a data distribution of interest, while the discriminative network distinguishes candidates produced by the generator from a true data distribution.
- the generative network's training objective is to increase the error rate of the discriminative network.
- Each network learns iteratively to better model the true data distribution across the classes of interest.
- the neural network 114 is trained on a plurality of labeled images of the appropriate reproductive cellular structure.
- labeled images it is meant that the ploidy status of the embryo within the image is known, for example, via expert annotation, and the clinical parameter associated with ploidy status is provided to the neural network along with the image during the training process.
- the weights associated with the interconnections among nodes in the neural network 114 are iteratively changed until, once the network is changed, an output of the network when presented with a novel, unlabeled image provides a clinical parameter representing the ploidy status of the embryo within the novel image.
- the first clinical parameter can be stored on the non-transitory computer readable medium 110 .
- a predictive model 116 is configured to generate a second clinical parameter from a set of at least one parameter representing one of a patient receiving the embryo, an egg utilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided sperm used to create the embryo, and an egg donor who provided the egg.
- appropriate parameters include an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo.
- the second clinical parameter can be a categorical parameter with a first value representing aneuploidy and a second value representing euploidy or a continuous parameter representing a likelihood of either aneuploidy or euploidy of the embryo.
- the predictive model 116 can make use of additional biometric parameters representing of one of a patient receiving the embryo, a sperm donor who provided sperm used to create the embryo, an egg utilized to produce the human embryo, and an egg donor who provided the egg in determining a quality of the embryo as well as numerical features extracted from the image of the embryo.
- the second clinical parameter can be stored on the non-transitory computer readable medium 110
- the predictive model 116 can utilize one or more pattern recognition algorithms, each of which analyze the extracted features or a subset of the extracted features to classify the patients into one of the plurality of classes and provide this information as a second clinical parameter. Where multiple classification or regression models are used, an arbitration element can be utilized to provide a coherent result from the plurality of models.
- the training process of a given classifier will vary with its implementation, but training generally involves a statistical aggregation of training data into one or more parameters associated with the output class.
- rule-based models such as decision trees, domain knowledge, for example, as provided by one or more human experts, can be used in place of or to supplement training data in selecting rules for classifying a patient using the extracted features. Any of a variety of techniques can be utilized for the classification algorithm, including support vector machines (SVMs), regression models, self-organized maps, fuzzy logic systems, data fusion processes, boosting and bagging methods, rule-based systems, or artificial neural networks.
- SVMs support vector machines
- a support vector machine (SVM) classifier can utilize a plurality of functions, referred to as hyperplanes, to conceptually divide boundaries in the N-dimensional feature space, where each of the N dimensions represents one associated feature of the feature vector.
- the boundaries define a range of feature values associated with each class. Accordingly, an output class and an associated confidence value can be determined for a given input feature vector according to its position in feature space relative to the boundaries.
- the SVM can be implemented via a kernel method using a linear or non-linear kernel.
- the predictive model 116 can also use an artificial neural network classifier as described previously.
- a rule-based classifier applies a set of logical rules to the extracted features to select an output class. Generally, the rules are applied in order, with the logical result at each step influencing the analysis at later steps.
- the specific rules and their sequence can be determined from any or all of training data, analogical reasoning from previous cases, or existing domain knowledge.
- One example of a rule-based classifier is a decision tree algorithm, in which the values of features in a feature set are compared to corresponding threshold in a hierarchical tree structure to select a class for the feature vector.
- a random forest classifier is a modification of the decision tree algorithm using a bootstrap aggregating, or “bagging” approach.
- multiple decision trees are trained on random samples of the training set, and an average (e.g., mean, median, or mode) result across the plurality of decision trees is returned.
- an average e.g., mean, median, or mode
- the result from each tree would be categorical, and thus a modal outcome can be used.
- An arbitrator 118 is configured to generate a composite parameter, representing a likelihood of aneuploidy in the embryo, from the first clinical parameter and the second clinical parameter.
- the composite parameter is an average of a set of clinical parameters that includes the first clinical parameter and the second clinical parameter.
- the composite parameter is determined as a weighted linear combination of a set of clinical parameters that includes the first clinical parameter and the second clinical parameter. In this instance, the weights for each parameter can be determined according to a known accuracy of the neural network 114 , the predictive model 116 and any additional models used to generate the parameters.
- the set of clinical parameters that includes the first clinical parameter and the second clinical parameter can be categorical, and the arbitrator 118 can select the composite parameter according to a voting algorithm. In one example, either a majority or plurality vote can be used.
- the composite parameter can then be stored on the non-transitory computer readable medium 110 and/or displayed to a user at an associated output device (not shown).
- FIG. 2 illustrates another example of a system 200 for fully automated screening for aneuploidy in a human embryo.
- the system 200 includes an imager 202 that acquires one or more images of the embryo.
- the imager 202 can include one or more cameras, capable of producing images in the visible or infrared range, paired with appropriate optics to provide an image of an embryo.
- the imager 202 can be implemented to capture images of an embryo at multiple days of development as part of a time-lapse embryo imaging system.
- the imager 202 can be configured to generate a static observation of the embryo as a set of one or more images.
- the imager 202 includes an attachment for a mobile device that operates with a camera of the mobile device to provide the embryo images.
- the housing for the attachment can be 3-D printed using polylactic acid with dimensions of 82 ⁇ 34 ⁇ 48 mm.
- An acrylic lens can be included in the housing to provided appropriate magnification for the embryo images.
- the imager 202 can be implemented as a stand-alone system with an optical housing that is 3-D printed from polylactic acid and overall dimensions of 62 ⁇ 92 ⁇ 175 mm.
- the housing contains an electronic circuit with a white light-emitting diode, a three-volt battery, and a single pole double-throw switch.
- the embryo sample is transilluminated, with a 10 ⁇ Plan-Achromatic objective lens for image magnification and a complementary metal-oxide-semiconductor (CMOS) image sensor for embryo image data acquisition.
- CMOS complementary metal-oxide-semiconductor
- the CMOS sensor can be connected to a single-board computer to process the captured images.
- the imager 202 can be connected to a mobile device via a wireless connection (e.g., Wi-Fi, Bluetooth, or a similar connection) for data processing and visualization.
- a wireless connection e.g., Wi-Fi, Bluetooth, or a similar connection
- Some or all of the one or more images obtained at the imager 202 are provided to an analysis system 210 comprising a processor 212 , an output device 214 , and a non-transitory computer readable medium 220 storing instructions executable by the processor.
- the instructions are executable to provide an imager interface 222 that receives the image or images of the embryo.
- the imager interface 222 can apply one or more imaging condition techniques, such as cropping and filtering, to better prepare the image for analysis.
- the images are then provided to a neural network 224 that provides a first clinical parameter representing the ploidy status of the embryo.
- the neural network 224 is trained on labeled images of euploid and aneuploid embryos, and provides a categorical output that can assume a first value, representing euploidy, and a second value, representing aneuploidy.
- the neural network 224 can be a convolutional neural network, which is a feed-forward artificial neural network that includes convolutional layers, which effectively apply a convolution to the values at the preceding layer of the network to emphasize various sets of features within an image.
- a convolutional layer each neuron is connected only to a proper subset of the neurons in the preceding layer, referred to as the receptive field of the neuron.
- the convolutional neural network is implemented using the Xception architecture.
- at least one chromatic value (e.g., a value for an RGB color channel, a YCrCb color channel, or a grayscale brightness) associated with each pixel is provided as an initial input to the convolutional neural network.
- the neural network 224 can be implemented as a recurrent neural network.
- a recurrent neural network the connections between nodes in the network are selected to form a directed graph along a sequence, allowing it to exhibit dynamic temporal behavior.
- the neural network 224 is implemented and trained as a discriminative network in a generative adversarial model, in which a generative neural network and the discriminative network provide mutual feedback to one another, such that the generative neural network produces increasingly sophisticated samples for the discriminative network to attempt to classify.
- some or all layers of the neural network can be trained via transfer learning from another system, with only some of the layers trained on the training images of the reproductive cellular structure.
- a final layer of the neural network 224 can be implemented as a softmax layer to provide a classification result.
- Some or all of one or more images can also be provided to a feature extractor 226 that reduces the image data into a feature vector comprising a plurality of values representing the content of the image or images.
- the feature extractor 226 extracts a plurality of features, which can be categorical, discrete, and continuous parameters representing the sensor data. It one example, the feature extractor 226 can utilize latent values from the neural network 224 as features representing the image or images.
- a network interface 228 can retrieve a set of biometric parameters from an associated memory, such as an electronic health records (EHR) database.
- EHR electronic health records
- the biometric parameters can represent one or more of a patient receiving the embryo, an egg utilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided sperm used to create the embryo, and an egg donor who provided the egg from an associated memory.
- the parameters include an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo.
- a first subset of the values in the feature vector and the retrieved biometric parameters can be provided to a first predictive model 230 to generate a second clinical parameter.
- the first predictive model 230 is implemented as a support vector machine trained on labeled data representing euploid and aneuploid embryos, and provides the second clinical parameter as a categorical output that can assume a first value, representing euploidy, and a second value, representing aneuploidy.
- a second subset of the values in the feature vector and the retrieved biometric parameters can be provided to a second predictive model 232 to generate a third clinical parameter.
- the second predictive model 232 is implemented as a fully-connected feedforward neural network trained on labeled data representing euploid and aneuploid embryos, and provides the third clinical parameter as a categorical output that can assume a first value, representing euploidy, and a second value, representing aneuploidy.
- the first, second, and third clinical parameters are provided to a voting system 234 that generates a composite parameter representing the ploidy status of the embryo.
- the composite parameter is generated according to a majority vote among the first, second, and third clinical parameters.
- the composite clinical parameter can be provided to a user at the output device 214 via a user interface 236 .
- the user interface 236 can include appropriate software instructions for receiving the output of the voting system 234 and presenting it at the output device 214 .
- the output device 214 can include a mobile device that communicates wirelessly with the analysis system 210 .
- FIGS. 3 and 4 a method in accordance with various aspects of the present invention will be better appreciated with reference to FIGS. 3 and 4 . While, for purposes of simplicity of explanation, the methods of FIGS. 3 and 4 is shown and described as executing serially, it is to be understood and appreciated that the present invention is not limited by the illustrated order, as some aspects could, in accordance with the present invention, occur in different orders and/or concurrently with other aspects from that shown and described herein. Moreover, not all illustrated features may be required to implement a method in accordance with an aspect the present invention.
- FIG. 3 illustrates an example of a method 300 for screening for aneuploidy in a human embryo.
- an image of the embryo is obtained at an associated imager.
- the image of the embryo to is provided a neural network to generate a first clinical parameter from the image of the embryo at a neural network.
- the first clinical parameter can be either categorical or continuous.
- the neural network can be implemented as any of a convolutional neural network, a recurrent neural network, and a discriminative classifier trained as part of a generative adversarial network.
- a set of at least one parameter representing one of biometric parameters of one of a patient receiving the embryo, an egg utilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided sperm used to create the embryo, and an egg donor who provided the egg is retrieved.
- appropriate parameters include an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo.
- a second clinical parameter is generated from the set of at least one parameter at a predictive model.
- the predictive model is implemented as one of a support vector machine and a fully-connected feedforward neural network.
- a composite parameter representing a likelihood of aneuploidy in the embryo, is generated from the first clinical parameter and the second clinical parameter.
- the composite parameter can be generated, for example, as an average of a set of clinical parameters that includes the first clinical parameter and the second clinical parameter, as a weighted linear combination of the set of clinical parameters, or according to a voting algorithm.
- the composite parameter can be stored in memory and/or displayed to a user at an associated display.
- FIG. 4 illustrates another example of a method 400 for screening for aneuploidy in a human embryo.
- an image of the embryo is obtained.
- the image of the embryo to is provided a neural network to generate a first clinical parameter from the image of the embryo at a neural network.
- the first clinical parameter can be either categorical or continuous.
- the neural network can be implemented as any of a convolutional neural network, a recurrent neural network, and a discriminative classifier trained as part of a generative adversarial network.
- a set of parameters are retrieved from an associated memory.
- the set of parameters include at least an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo.
- a second clinical parameter is generated at a first predictive model from a first subset of the set of parameters.
- the first predictive model is implemented as a support vector machine.
- a third clinical parameter is generated at a second predictive model from a second subset of the set of parameters.
- the second predictive model is implemented as a fully-connected feedforward neural network.
- each of the first subset and the second subset contain all of the set of parameters, such that each of the first predictive model and the second predictive model receive the entire set of parameters.
- the first subset and the second subset are each proper subsets of the set of parameters.
- a composite parameter representing a likelihood of aneuploidy in the embryo, is generated from the first clinical parameter, the second clinical parameter, and the third clinical parameter.
- the composite parameter is generated according to a majority vote among the first clinical parameter, the second clinical parameter, and the third clinical parameter.
- the composite parameter can be stored in memory and/or displayed to a user at an associated display.
- FIG. 5 is a schematic block diagram illustrating an exemplary system 500 of hardware components capable of implementing examples of the systems and methods disclosed in FIGS. 1 - 4 , such as the automated embryo evaluation system illustrated in FIG. 1 .
- the system 500 can include various systems and subsystems.
- the system 500 can be any of personal computer, a laptop computer, a workstation, a computer system, an appliance, an application-specific integrated circuit (ASIC), a server, a server blade center, or a server farm.
- ASIC application-specific integrated circuit
- the system 500 can includes a system bus 502 , a processing unit 504 , a system memory 506 , memory devices 508 and 510 , a communication interface 512 (e.g., a network interface), a communication link 514 , a display 516 (e.g., a video screen), and an input device 518 (e.g., a keyboard and/or a mouse).
- the system bus 502 can be in communication with the processing unit 504 and the system memory 506 .
- the additional memory devices 508 and 510 such as a hard disk drive, server, stand-alone database, or other non-volatile memory, can also be in communication with the system bus 502 .
- the system bus 502 interconnects the processing unit 504 , the memory devices 506 - 510 , the communication interface 512 , the display 516 , and the input device 518 .
- the system bus 502 also interconnects an additional port (not shown), such as a universal serial bus (USB) port.
- USB universal serial bus
- the system 500 could be implemented in a computing cloud.
- features of the system 500 such as the processing unit 504 , the communication interface 512 , and the memory devices 508 and 510 could be representative of a single instance of hardware or multiple instances of hardware with applications executing across the multiple of instances (i.e., distributed) of hardware (e.g., computers, routers, memory, processors, or a combination thereof).
- the system 500 could be implemented on a single dedicated server.
- the processing unit 504 can be a computing device and can include an application-specific integrated circuit (ASIC).
- the processing unit 504 executes a set of instructions to implement the operations of examples disclosed herein.
- the processing unit can include a processing core.
- the additional memory devices 506 , 508 , and 510 can store data, programs, instructions, database queries in text or compiled form, and any other information that can be needed to operate a computer.
- the memories 506 , 508 and 510 can be implemented as computer-readable media (integrated or removable) such as a memory card, disk drive, compact disk (CD), or server accessible over a network.
- the memories 506 , 508 and 510 can comprise text, images, video, and/or audio, portions of which can be available in formats comprehensible to human beings.
- system 500 can access an external data source or query source through the communication interface 512 , which can communicate with the system bus 502 and the communication link 514 .
- the system 500 can be used to implement one or more parts of an embryo evaluation system in accordance with the present invention.
- Computer executable logic for implementing the composite applications testing system resides on one or more of the system memory 506 , and the memory devices 508 , 510 in accordance with certain examples.
- the processing unit 504 executes one or more computer executable instructions originating from the system memory 506 and the memory devices 508 and 510 .
- a computer readable medium can include multiple computer readable media each operatively connected to the processing unit.
- Implementation of the techniques, blocks, steps, and means described above can be done in various ways. For example, these techniques, blocks, steps, and means can be implemented in hardware, software, or a combination thereof.
- the processing units can be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described above, and/or a combination thereof.
- ASICs application specific integrated circuits
- DSPs digital signal processors
- DSPDs digital signal processing devices
- PLDs programmable logic devices
- FPGAs field programmable gate arrays
- processors controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described above, and/or a combination thereof.
- the embodiments can be described as a process which is depicted as a flowchart, a flow diagram, a data flow diagram, a structure diagram, or a block diagram. Although a flowchart can describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations can be re-arranged.
- a process is terminated when its operations are completed, but could have additional steps not included in the figure.
- a process can correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or the main function.
- embodiments can be implemented by hardware, software, scripting languages, firmware, middleware, microcode, hardware description languages, and/or any combination thereof.
- the program code or code segments to perform the necessary tasks can be stored in a machine readable medium such as a storage medium.
- a code segment or machine-executable instruction can represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a script, a class, or any combination of instructions, data structures, and/or program statements.
- a code segment can be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters, and/or memory contents. Information, arguments, parameters, data, etc. can be passed, forwarded, or transmitted via any suitable means including memory sharing, message passing, ticket passing, network transmission, etc.
- the methodologies can be implemented with modules (e.g., procedures, functions, and so on) that perform the functions described herein.
- Any machine-readable medium tangibly embodying instructions can be used in implementing the methodologies described herein.
- software codes can be stored in a memory.
- Memory can be implemented within the processor or external to the processor.
- the term “memory” refers to any type of long term, short term, volatile, nonvolatile, or other storage medium and is not to be limited to any particular type of memory or number of memories, or type of media upon which memory is stored.
- the term “storage medium” can represent one or more memories for storing data, including read only memory (ROM), random access memory (RAM), magnetic RAM, core memory, magnetic disk storage mediums, optical storage mediums, flash memory devices and/or other machine readable mediums for storing information.
- ROM read only memory
- RAM random access memory
- magnetic RAM magnetic RAM
- core memory magnetic disk storage mediums
- optical storage mediums flash memory devices and/or other machine readable mediums for storing information.
- computer readable medium and “machine readable medium” includes, but is not limited to portable or fixed storage devices, optical storage devices, wireless channels, and/or various other storage mediums capable of storing that contain or carry instruction(s) and/or data. It will be appreciated that a “computer readable medium” or “machine readable medium” can include multiple media each operatively connected to a processing unit.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Biomedical Technology (AREA)
- Multimedia (AREA)
- Public Health (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Primary Health Care (AREA)
- Epidemiology (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Pathology (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Quality & Reliability (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Image Analysis (AREA)
- Investigating Or Analysing Biological Materials (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Apparatus Associated With Microorganisms And Enzymes (AREA)
- Measuring Or Testing Involving Enzymes Or Micro-Organisms (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Devices For Executing Special Programs (AREA)
- Image Processing (AREA)
- General Factory Administration (AREA)
Abstract
Description
- The present application claims priority to each of U.S. Provisional Patent Application Ser. No. 63/077,405 filed Sep. 11, 2020 entitled ARTIFICIAL INTELLIGENCE-ENABLED SYSTEM TO AID IN ANEUPLOIDY SCREENING OF PREIMPLANTATION EMBRYOS and U.S. Provisional Patent Application Ser. No. 63/077,398 filed Sep. 11, 2020 entitled ARTIFICIAL INTELLIGENCE-ENABLED SYSTEM FOR ALIGNMENT OF OOCYTES AND PREIMPLANTATION EMBRYOS FOR INTRACYTOPLASMIC SPERM INJECTION (ICSI) AND ASSISTED HATCHING (AH) PROCEDURES. The entire content of each of these applications is incorporated herein by reference in its entirety for all purposes.
- This invention was made with government support under one or more of grant numbers R01A1118502, R01A1138800, and R21HD092828, awarded by the National Institutes of Health. The government has certain rights in the invention.
- The present invention relates generally to the field of assisted fertility, and more particularly, to automated aneuploidy screening using arbitrated ensembles of predictive models.
- Infertility is an underestimated healthcare problem that affects over forty-eight million couples globally and is a cause of distress, depression, and discrimination. Although assisted reproductive technologies (ART) such as in-vitro fertilization (IVF) has alleviated the burden of infertility to an extent, it has been inefficient with an average success rate of approximately twenty-six percent reported in 2015 in the US. IVF remains as an expensive solution, with a cost between $7000 and $20,000 per ART cycle in the US, which is generally not covered by insurance. Further, many patients require multiple cycles of IVF to achieve pregnancy. Non-invasive selection of the top-quality embryo for transfer is one of the most important factors in achieving successful ART outcomes, yet this critical step remains a significant challenge.
- Preimplantation genetic testing for aneuploidy (PGT-A) has become a widely utilized tool for screening embryos for transfer. There are limited studies showing the efficacy of this procedure in women under thirty-five years of age. Conversely, there are some reports demonstrating that younger patients have reduced cumulative pregnancy rates when utilizing PGT-A. These lower outcomes are likely due to inaccurate PGT-A results or due to the invasive procedures involved with embryo biopsy and cryopreservation.
- In accordance with an aspect of the present invention, a method is provided for fully automated screening for aneuploidy in a human embryo. An image of the embryo is obtained at an associated imager and provided to a neural network to generate a first clinical parameter. A set of at least one parameter representing one of biometric parameters of one of a patient receiving the embryo, an egg utilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided sperm used to create the embryo, and an egg donor who provided the egg is retrieved, and a second clinical parameter is generated from the set of at least one parameter at a predictive model. A composite parameter, representing a likelihood of aneuploidy in the embryo, is generated from the first clinical parameter and the second clinical parameter.
- In accordance with another aspect of the present invention, a system is provided for fully automated screening for aneuploidy in a human embryo. The system includes a processor and a non-transitory computer readable medium storing instructions executable by the processor. The machine executable instructions include an imager interface configured to receive an image of the embryo from an associated imager and a neural network configured to generate a first clinical parameter from the image of the embryo. A predictive model is configured to generate a second clinical parameter from a set of at least one parameter representing one of a patient receiving the embryo, an oocyte fertilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided the sperm used to create the embryo, and an egg donor who provided the oocyte from an associated memory. An arbitrator is configured to generate a composite parameter, representing a likelihood of aneuploidy in the embryo, from the first clinical parameter and the second clinical parameter.
- In accordance with yet another aspect of the present invention, a method is provided for fully automated screening for aneuploidy in a human embryo. An image of the embryo is obtained at an associated imager and a first clinical parameter is generated from the image of the embryo at a convolutional neural network. A set of parameters is retrieved from an associated memory. The set of parameters included each of an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo. A second clinical parameter at a first predictive model from a first subset of the set of parameters, and a third clinical parameter at a second predictive model from a second subset of the set of parameters. A composite parameter, representing a likelihood of aneuploidy in the embryo, is generated from the first clinical parameter, the second clinical parameter, and the third clinical parameter.
- The foregoing and other features of the present invention will become apparent to those skilled in the art to which the present invention relates upon reading the following description with reference to the accompanying drawings, in which:
-
FIG. 1 illustrates one example of a system for fully automated screening for aneuploidy in a human embryo; -
FIG. 2 illustrates another example of a system for fully automated screening for aneuploidy in a human embryo; -
FIG. 3 illustrates an example of a method for screening for aneuploidy in a human embryo; -
FIG. 4 illustrates another example of a method for screening for aneuploidy in a human embryo; and -
FIG. 5 is a schematic block diagram illustrating an exemplary system of hardware components capable of implementing examples of the systems and methods disclosed herein. - Emulating the skill of highly trained embryologists in efficient embryo assessment in a fully automated system is a major unmet challenge in all of the previous work done in embryo computer-aided assessment. Current computer vision methods for embryo assessment are semi-automated, limited to measuring specific parameters providing metrics that require further analysis by embryologists, and require strictly controlled imaging systems. Previous attempts in developing systems using machine-learning approaches have required intensive image preprocessing followed by human-directed segmentation of embryo features for classification. Owing to the dependency of machine-learning approaches on image processing and segmentation, such methods suffer from the same limitations as computer vision techniques.
- Here, we overcome this challenge by employing a deep neural networks pretrained with a large set of images, paired with biometric information about the egg, sperm, and their respective donors, to screen embryos for aneuploidy. Unlike prior computer-aided algorithms used for embryo assessment, the systems and methods provided herein allows for automated screening for aneuploidy without any assistance by an embryologist. In one example, a convolutional neural network is applied to identify the shape, structure, and texture variations between morphologically complex embryos, while one or more other predictive models are employed to evaluate the biometric data. The system is resilient to changes in image illumination and quality due to data acquisition using multiple instruments. Further, since the additional predictive models can be trained independently of the neural network, the system is effectively modular, allowing for addition or substitution of models in response to new applications.
- The phrase “continuous parameter” is used herein to distinguish numerical values from categorical values or classes, and should be ready to include both truly continuous data as well as data more traditionally referred to as discrete data.
- As used herein, an “average” of a set of values is any measure of central tendency. For continuous parameters, this includes any of the median, arithmetic mean, and geometric mean of the values. For categorical parameters, the average is the mode of the set of values.
- A “static observation,” as used herein, is an image or group of images of an embryo that represent a single point in the development of the embryo. Where multiple images are used in a static observation, no discernible change in the structure and appearance of the embryo will have taken place between images. Data collected from a static observation represents morphological data for the embryo.
- A “subset” of a set as used here, refers to a set containing some or all elements of the set. A “proper subset” of a set contains less than all of the elements of the set.
- As used herein, an embryo is “euploid” when is has forty-six chromosomes. An embryo is “aneuploid” when it has a number of chromosomes that is different from the usual forty-six. The “ploidy status” of the embryo is the embryos status as either euploid or aneuploid, represented either categorically or as a probability.
-
FIG. 1 illustrates one example of asystem 100 for fully automated screening for aneuploidy in a human embryo. Thesystem 100 includes aprocessor 102 and a non-transitory computerreadable medium 110 that stores machine executable instructions for assigning a value representing a location of interest within a reproductive cellular structure. The machine executable instructions include animager interface 112 that receives an image of embryo from an associated imager. For example, theimaging interface 112 can receive the image from the imager via a bus or network connection and condition the image for analysis at aneural network 114. In one example, theneural network 114 can be implemented on a cloud computing system, with the image transmitted to the server containing theneural network 114 via a network interface (not shown). - The
neural network 114 determines, from the image of the reproductive cellular structure, a first clinical parameter representing a ploidy status of the embryo. In one implementation, the first clinical parameter is a categorical parameter with a first value representing aneuploidy and a second value representing euploidy. In another implementation, the first clinical parameter is a continuous parameter representing a likelihood of either aneuploidy or euploidy of the embryo. Theneural network 114 includes a plurality of nodes having a plurality of interconnections. Values from the image, for example luminance and/or chrominance values associated with the individual pixels, are provided to a plurality of input nodes. The input nodes each provide these input values to layers of one or more intermediate nodes. A given intermediate node receives one or more output values from previous nodes. The received values are weighted according to a series of weights established during the training of the classifier. An intermediate node translates its received values into a single output according to an activation function at the node. For example, the intermediate node can sum the received values and subject the sum to an identify function, a step function, a sigmoid function, a hyperbolic tangent, a rectified linear unit, a leaky rectified linear unit, a parametric rectified linear unit, a Gaussian error linear unit, the softplus function, an exponential linear unit, a scaled exponential linear unit, a Gaussian function, a sigmoid linear unit, a growing cosine unit, the Heaviside function, and the mish function. A final layer of nodes provides the confidence values for the output classes of the neural network, with each node having an associated value representing a confidence for one of the associated output classes of the classifier. - Many ANN classifiers are fully-connected and feedforward. A convolutional neural network, however, includes convolutional layers in which nodes from a previous layer are only connected to a subset of the nodes in the convolutional layer. Recurrent neural networks are a class of neural networks in which connections between nodes form a directed graph along a temporal sequence. Unlike a feedforward network, recurrent neural networks can incorporate feedback from states caused by earlier inputs, such that an output of the recurrent neural network for a given input can be a function of not only the input but one or more previous inputs. As an example, Long Short-Term Memory (LSTM) networks are a modified version of recurrent neural networks, which makes it easier to remember past data in memory. In some examples, neural networks are trained in an adversarial manner in which a generative classifier is configured to generate new instances for a discriminative classifier. Typically, the generative network learns to map from a latent space to a data distribution of interest, while the discriminative network distinguishes candidates produced by the generator from a true data distribution. The generative network's training objective is to increase the error rate of the discriminative network. Each network learns iteratively to better model the true data distribution across the classes of interest.
- The
neural network 114 is trained on a plurality of labeled images of the appropriate reproductive cellular structure. By “labeled images,” it is meant that the ploidy status of the embryo within the image is known, for example, via expert annotation, and the clinical parameter associated with ploidy status is provided to the neural network along with the image during the training process. During training, the weights associated with the interconnections among nodes in theneural network 114 are iteratively changed until, once the network is changed, an output of the network when presented with a novel, unlabeled image provides a clinical parameter representing the ploidy status of the embryo within the novel image. The first clinical parameter can be stored on the non-transitory computerreadable medium 110. - A
predictive model 116 is configured to generate a second clinical parameter from a set of at least one parameter representing one of a patient receiving the embryo, an egg utilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided sperm used to create the embryo, and an egg donor who provided the egg. Examples of appropriate parameters include an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo. Like the first clinical parameter, the second clinical parameter can be a categorical parameter with a first value representing aneuploidy and a second value representing euploidy or a continuous parameter representing a likelihood of either aneuploidy or euploidy of the embryo. It will be appreciated that thepredictive model 116 can make use of additional biometric parameters representing of one of a patient receiving the embryo, a sperm donor who provided sperm used to create the embryo, an egg utilized to produce the human embryo, and an egg donor who provided the egg in determining a quality of the embryo as well as numerical features extracted from the image of the embryo. The second clinical parameter can be stored on the non-transitory computerreadable medium 110 - The
predictive model 116 can utilize one or more pattern recognition algorithms, each of which analyze the extracted features or a subset of the extracted features to classify the patients into one of the plurality of classes and provide this information as a second clinical parameter. Where multiple classification or regression models are used, an arbitration element can be utilized to provide a coherent result from the plurality of models. The training process of a given classifier will vary with its implementation, but training generally involves a statistical aggregation of training data into one or more parameters associated with the output class. For rule-based models, such as decision trees, domain knowledge, for example, as provided by one or more human experts, can be used in place of or to supplement training data in selecting rules for classifying a patient using the extracted features. Any of a variety of techniques can be utilized for the classification algorithm, including support vector machines (SVMs), regression models, self-organized maps, fuzzy logic systems, data fusion processes, boosting and bagging methods, rule-based systems, or artificial neural networks. - For example, a support vector machine (SVM) classifier can utilize a plurality of functions, referred to as hyperplanes, to conceptually divide boundaries in the N-dimensional feature space, where each of the N dimensions represents one associated feature of the feature vector. The boundaries define a range of feature values associated with each class. Accordingly, an output class and an associated confidence value can be determined for a given input feature vector according to its position in feature space relative to the boundaries. In one implementation, the SVM can be implemented via a kernel method using a linear or non-linear kernel. The
predictive model 116 can also use an artificial neural network classifier as described previously. - A rule-based classifier applies a set of logical rules to the extracted features to select an output class. Generally, the rules are applied in order, with the logical result at each step influencing the analysis at later steps. The specific rules and their sequence can be determined from any or all of training data, analogical reasoning from previous cases, or existing domain knowledge. One example of a rule-based classifier is a decision tree algorithm, in which the values of features in a feature set are compared to corresponding threshold in a hierarchical tree structure to select a class for the feature vector. A random forest classifier is a modification of the decision tree algorithm using a bootstrap aggregating, or “bagging” approach. In this approach, multiple decision trees are trained on random samples of the training set, and an average (e.g., mean, median, or mode) result across the plurality of decision trees is returned. For a classification task, the result from each tree would be categorical, and thus a modal outcome can be used.
- An
arbitrator 118 is configured to generate a composite parameter, representing a likelihood of aneuploidy in the embryo, from the first clinical parameter and the second clinical parameter. In one implementation, the composite parameter is an average of a set of clinical parameters that includes the first clinical parameter and the second clinical parameter. In another implementation, the composite parameter is determined as a weighted linear combination of a set of clinical parameters that includes the first clinical parameter and the second clinical parameter. In this instance, the weights for each parameter can be determined according to a known accuracy of theneural network 114, thepredictive model 116 and any additional models used to generate the parameters. In a still further implementation, the set of clinical parameters that includes the first clinical parameter and the second clinical parameter can be categorical, and thearbitrator 118 can select the composite parameter according to a voting algorithm. In one example, either a majority or plurality vote can be used. The composite parameter can then be stored on the non-transitory computerreadable medium 110 and/or displayed to a user at an associated output device (not shown). -
FIG. 2 illustrates another example of asystem 200 for fully automated screening for aneuploidy in a human embryo. Thesystem 200 includes animager 202 that acquires one or more images of the embryo. For example, theimager 202 can include one or more cameras, capable of producing images in the visible or infrared range, paired with appropriate optics to provide an image of an embryo. In one implementation, theimager 202 can be implemented to capture images of an embryo at multiple days of development as part of a time-lapse embryo imaging system. In another implementation, theimager 202 can be configured to generate a static observation of the embryo as a set of one or more images. In one implementation, theimager 202 includes an attachment for a mobile device that operates with a camera of the mobile device to provide the embryo images. The housing for the attachment can be 3-D printed using polylactic acid with dimensions of 82×34×48 mm. An acrylic lens can be included in the housing to provided appropriate magnification for the embryo images. - In another implementation, the
imager 202 can be implemented as a stand-alone system with an optical housing that is 3-D printed from polylactic acid and overall dimensions of 62×92×175 mm. The housing contains an electronic circuit with a white light-emitting diode, a three-volt battery, and a single pole double-throw switch. The embryo sample is transilluminated, with a 10× Plan-Achromatic objective lens for image magnification and a complementary metal-oxide-semiconductor (CMOS) image sensor for embryo image data acquisition. The CMOS sensor can be connected to a single-board computer to process the captured images. Theimager 202 can be connected to a mobile device via a wireless connection (e.g., Wi-Fi, Bluetooth, or a similar connection) for data processing and visualization. - Some or all of the one or more images obtained at the
imager 202 are provided to ananalysis system 210 comprising aprocessor 212, anoutput device 214, and a non-transitory computerreadable medium 220 storing instructions executable by the processor. The instructions are executable to provide animager interface 222 that receives the image or images of the embryo. Theimager interface 222 can apply one or more imaging condition techniques, such as cropping and filtering, to better prepare the image for analysis. The images are then provided to aneural network 224 that provides a first clinical parameter representing the ploidy status of the embryo. In the illustrated implementation, theneural network 224 is trained on labeled images of euploid and aneuploid embryos, and provides a categorical output that can assume a first value, representing euploidy, and a second value, representing aneuploidy. - In one implementation, the
neural network 224 can be a convolutional neural network, which is a feed-forward artificial neural network that includes convolutional layers, which effectively apply a convolution to the values at the preceding layer of the network to emphasize various sets of features within an image. In a convolutional layer, each neuron is connected only to a proper subset of the neurons in the preceding layer, referred to as the receptive field of the neuron. In one example, the convolutional neural network is implemented using the Xception architecture. In one implementation, at least one chromatic value (e.g., a value for an RGB color channel, a YCrCb color channel, or a grayscale brightness) associated with each pixel is provided as an initial input to the convolutional neural network. - In another implementation, the
neural network 224 can be implemented as a recurrent neural network. In a recurrent neural network, the connections between nodes in the network are selected to form a directed graph along a sequence, allowing it to exhibit dynamic temporal behavior. In another implementation, theneural network 224 is implemented and trained as a discriminative network in a generative adversarial model, in which a generative neural network and the discriminative network provide mutual feedback to one another, such that the generative neural network produces increasingly sophisticated samples for the discriminative network to attempt to classify. Regardless of the structure of theneural network 224, some or all layers of the neural network can be trained via transfer learning from another system, with only some of the layers trained on the training images of the reproductive cellular structure. A final layer of theneural network 224 can be implemented as a softmax layer to provide a classification result. - Some or all of one or more images can also be provided to a
feature extractor 226 that reduces the image data into a feature vector comprising a plurality of values representing the content of the image or images. In particular, thefeature extractor 226 extracts a plurality of features, which can be categorical, discrete, and continuous parameters representing the sensor data. It one example, thefeature extractor 226 can utilize latent values from theneural network 224 as features representing the image or images. Anetwork interface 228 can retrieve a set of biometric parameters from an associated memory, such as an electronic health records (EHR) database. The biometric parameters can represent one or more of a patient receiving the embryo, an egg utilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided sperm used to create the embryo, and an egg donor who provided the egg from an associated memory. In the illustrated example, the parameters include an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo. - A first subset of the values in the feature vector and the retrieved biometric parameters can be provided to a first
predictive model 230 to generate a second clinical parameter. In the illustrated implementation, the firstpredictive model 230 is implemented as a support vector machine trained on labeled data representing euploid and aneuploid embryos, and provides the second clinical parameter as a categorical output that can assume a first value, representing euploidy, and a second value, representing aneuploidy. A second subset of the values in the feature vector and the retrieved biometric parameters can be provided to a secondpredictive model 232 to generate a third clinical parameter. In the illustrated implementation, the secondpredictive model 232 is implemented as a fully-connected feedforward neural network trained on labeled data representing euploid and aneuploid embryos, and provides the third clinical parameter as a categorical output that can assume a first value, representing euploidy, and a second value, representing aneuploidy. - The first, second, and third clinical parameters are provided to a
voting system 234 that generates a composite parameter representing the ploidy status of the embryo. In the illustrated implementation, the composite parameter is generated according to a majority vote among the first, second, and third clinical parameters. The composite clinical parameter can be provided to a user at theoutput device 214 via auser interface 236. For example, theuser interface 236 can include appropriate software instructions for receiving the output of thevoting system 234 and presenting it at theoutput device 214. In one implementation, theoutput device 214 can include a mobile device that communicates wirelessly with theanalysis system 210. - In view of the foregoing structural and functional features described above, a method in accordance with various aspects of the present invention will be better appreciated with reference to
FIGS. 3 and 4 . While, for purposes of simplicity of explanation, the methods ofFIGS. 3 and 4 is shown and described as executing serially, it is to be understood and appreciated that the present invention is not limited by the illustrated order, as some aspects could, in accordance with the present invention, occur in different orders and/or concurrently with other aspects from that shown and described herein. Moreover, not all illustrated features may be required to implement a method in accordance with an aspect the present invention. -
FIG. 3 illustrates an example of amethod 300 for screening for aneuploidy in a human embryo. At 302, an image of the embryo is obtained at an associated imager. At 304, the image of the embryo to is provided a neural network to generate a first clinical parameter from the image of the embryo at a neural network. The first clinical parameter can be either categorical or continuous. The neural network can be implemented as any of a convolutional neural network, a recurrent neural network, and a discriminative classifier trained as part of a generative adversarial network. At 306, a set of at least one parameter representing one of biometric parameters of one of a patient receiving the embryo, an egg utilized to produce the human embryo, a sperm used to create the embryo, a sperm donor who provided sperm used to create the embryo, and an egg donor who provided the egg is retrieved. Examples of appropriate parameters include an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo. - At 308, a second clinical parameter is generated from the set of at least one parameter at a predictive model. In one example, the predictive model is implemented as one of a support vector machine and a fully-connected feedforward neural network. At 310, a composite parameter, representing a likelihood of aneuploidy in the embryo, is generated from the first clinical parameter and the second clinical parameter. The composite parameter can be generated, for example, as an average of a set of clinical parameters that includes the first clinical parameter and the second clinical parameter, as a weighted linear combination of the set of clinical parameters, or according to a voting algorithm. The composite parameter can be stored in memory and/or displayed to a user at an associated display.
-
FIG. 4 illustrates another example of amethod 400 for screening for aneuploidy in a human embryo. At 402, an image of the embryo is obtained. At 404, the image of the embryo to is provided a neural network to generate a first clinical parameter from the image of the embryo at a neural network. The first clinical parameter can be either categorical or continuous. The neural network can be implemented as any of a convolutional neural network, a recurrent neural network, and a discriminative classifier trained as part of a generative adversarial network. At 406, a set of parameters are retrieved from an associated memory. The set of parameters include at least an age of the egg donor who provided the oocyte fertilized to produce the embryo, a value representing a quality of the sperm used to create the embryo, and a number of embryos that were normally fertilized from oocytes harvested with the oocyte fertilized to produce the embryo. - At 408, a second clinical parameter is generated at a first predictive model from a first subset of the set of parameters. In one implementation, the first predictive model is implemented as a support vector machine. At 410, a third clinical parameter is generated at a second predictive model from a second subset of the set of parameters. In one implementation, the second predictive model is implemented as a fully-connected feedforward neural network. In one example, each of the first subset and the second subset contain all of the set of parameters, such that each of the first predictive model and the second predictive model receive the entire set of parameters. In another example, the first subset and the second subset are each proper subsets of the set of parameters. At 412, a composite parameter, representing a likelihood of aneuploidy in the embryo, is generated from the first clinical parameter, the second clinical parameter, and the third clinical parameter. In the illustrated example, the composite parameter is generated according to a majority vote among the first clinical parameter, the second clinical parameter, and the third clinical parameter. The composite parameter can be stored in memory and/or displayed to a user at an associated display.
-
FIG. 5 is a schematic block diagram illustrating anexemplary system 500 of hardware components capable of implementing examples of the systems and methods disclosed inFIGS. 1-4 , such as the automated embryo evaluation system illustrated inFIG. 1 . Thesystem 500 can include various systems and subsystems. Thesystem 500 can be any of personal computer, a laptop computer, a workstation, a computer system, an appliance, an application-specific integrated circuit (ASIC), a server, a server blade center, or a server farm. - The
system 500 can includes asystem bus 502, aprocessing unit 504, asystem memory 506,memory devices communication link 514, a display 516 (e.g., a video screen), and an input device 518 (e.g., a keyboard and/or a mouse). Thesystem bus 502 can be in communication with theprocessing unit 504 and thesystem memory 506. Theadditional memory devices system bus 502. Thesystem bus 502 interconnects theprocessing unit 504, the memory devices 506-510, thecommunication interface 512, thedisplay 516, and theinput device 518. In some examples, thesystem bus 502 also interconnects an additional port (not shown), such as a universal serial bus (USB) port. - The
system 500 could be implemented in a computing cloud. In such a situation, features of thesystem 500, such as theprocessing unit 504, thecommunication interface 512, and thememory devices system 500 could be implemented on a single dedicated server. - The
processing unit 504 can be a computing device and can include an application-specific integrated circuit (ASIC). Theprocessing unit 504 executes a set of instructions to implement the operations of examples disclosed herein. The processing unit can include a processing core. - The
additional memory devices memories memories - Additionally or alternatively, the
system 500 can access an external data source or query source through thecommunication interface 512, which can communicate with thesystem bus 502 and thecommunication link 514. - In operation, the
system 500 can be used to implement one or more parts of an embryo evaluation system in accordance with the present invention. Computer executable logic for implementing the composite applications testing system resides on one or more of thesystem memory 506, and thememory devices processing unit 504 executes one or more computer executable instructions originating from thesystem memory 506 and thememory devices - Specific details are given in the above description to provide a thorough understanding of the embodiments. However, it is understood that the embodiments can be practiced without these specific details. For example, circuits can be shown in block diagrams in order not to obscure the embodiments in unnecessary detail. In other instances, well-known circuits, processes, algorithms, structures, and techniques can be shown without unnecessary detail in order to avoid obscuring the embodiments.
- Implementation of the techniques, blocks, steps, and means described above can be done in various ways. For example, these techniques, blocks, steps, and means can be implemented in hardware, software, or a combination thereof. For a hardware implementation, the processing units can be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described above, and/or a combination thereof.
- Also, it is noted that the embodiments can be described as a process which is depicted as a flowchart, a flow diagram, a data flow diagram, a structure diagram, or a block diagram. Although a flowchart can describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations can be re-arranged. A process is terminated when its operations are completed, but could have additional steps not included in the figure. A process can correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or the main function.
- Furthermore, embodiments can be implemented by hardware, software, scripting languages, firmware, middleware, microcode, hardware description languages, and/or any combination thereof. When implemented in software, firmware, middleware, scripting language, and/or microcode, the program code or code segments to perform the necessary tasks can be stored in a machine readable medium such as a storage medium. A code segment or machine-executable instruction can represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a script, a class, or any combination of instructions, data structures, and/or program statements. A code segment can be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters, and/or memory contents. Information, arguments, parameters, data, etc. can be passed, forwarded, or transmitted via any suitable means including memory sharing, message passing, ticket passing, network transmission, etc.
- For a firmware and/or software implementation, the methodologies can be implemented with modules (e.g., procedures, functions, and so on) that perform the functions described herein. Any machine-readable medium tangibly embodying instructions can be used in implementing the methodologies described herein. For example, software codes can be stored in a memory. Memory can be implemented within the processor or external to the processor. As used herein the term “memory” refers to any type of long term, short term, volatile, nonvolatile, or other storage medium and is not to be limited to any particular type of memory or number of memories, or type of media upon which memory is stored.
- Moreover, as disclosed herein, the term “storage medium” can represent one or more memories for storing data, including read only memory (ROM), random access memory (RAM), magnetic RAM, core memory, magnetic disk storage mediums, optical storage mediums, flash memory devices and/or other machine readable mediums for storing information. The terms “computer readable medium” and “machine readable medium” includes, but is not limited to portable or fixed storage devices, optical storage devices, wireless channels, and/or various other storage mediums capable of storing that contain or carry instruction(s) and/or data. It will be appreciated that a “computer readable medium” or “machine readable medium” can include multiple media each operatively connected to a processing unit.
- While the principles of the disclosure have been described above in connection with specific apparatuses and methods, it is to be clearly understood that this description is made only by way of example and not as limitation on the scope of the disclosure.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18/025,475 US20230326014A1 (en) | 2020-09-11 | 2021-09-13 | Automated aneuploidy screening using arbitrated ensembles |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US202063077398P | 2020-09-11 | 2020-09-11 | |
US202063077405P | 2020-09-11 | 2020-09-11 | |
PCT/US2021/050058 WO2022056374A1 (en) | 2020-09-11 | 2021-09-13 | Automated aneuploidy screening using arbitrated ensembles |
US18/025,475 US20230326014A1 (en) | 2020-09-11 | 2021-09-13 | Automated aneuploidy screening using arbitrated ensembles |
Publications (1)
Publication Number | Publication Date |
---|---|
US20230326014A1 true US20230326014A1 (en) | 2023-10-12 |
Family
ID=80630102
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/025,480 Pending US20230326165A1 (en) | 2020-09-11 | 2021-09-13 | Determining locations in reproductive cellular structures |
US18/025,475 Pending US20230326014A1 (en) | 2020-09-11 | 2021-09-13 | Automated aneuploidy screening using arbitrated ensembles |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/025,480 Pending US20230326165A1 (en) | 2020-09-11 | 2021-09-13 | Determining locations in reproductive cellular structures |
Country Status (9)
Country | Link |
---|---|
US (2) | US20230326165A1 (en) |
EP (2) | EP4210564A4 (en) |
JP (2) | JP7523680B2 (en) |
KR (2) | KR20230084176A (en) |
CN (1) | CN116438585A (en) |
AU (3) | AU2021338858A1 (en) |
BR (2) | BR112023004578A2 (en) |
CA (2) | CA3192441A1 (en) |
WO (2) | WO2022056374A1 (en) |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9177192B2 (en) * | 2013-02-28 | 2015-11-03 | Progyny, Inc. | Apparatus, method, and system for image-based human embryo cell classification |
CA3013048A1 (en) | 2016-01-28 | 2017-08-03 | Gerard LETTERIE | Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos |
GB201806999D0 (en) * | 2018-04-30 | 2018-06-13 | Univ Birmingham | Automated oocyte detection and orientation |
EP3805836A4 (en) | 2018-05-25 | 2021-07-28 | Sony Group Corporation | Control device, control method, and program |
WO2020058931A1 (en) | 2018-09-20 | 2020-03-26 | Aivf Ltd | Image feature detection |
US11926809B2 (en) * | 2018-09-28 | 2024-03-12 | Brigham And Women's Hospital, Inc. | Automated evaluation of sperm morphology |
EP3904937A4 (en) * | 2018-12-28 | 2022-09-21 | Olympus Corporation | Microscope system |
WO2020157761A1 (en) * | 2019-01-31 | 2020-08-06 | Amnon Buxboim | Automated evaluation of embryo implantation potential |
US10646156B1 (en) * | 2019-06-14 | 2020-05-12 | Cycle Clarity, LLC | Adaptive image processing in assisted reproductive imaging modalities |
-
2021
- 2021-09-13 US US18/025,480 patent/US20230326165A1/en active Pending
- 2021-09-13 JP JP2023516276A patent/JP7523680B2/en active Active
- 2021-09-13 WO PCT/US2021/050058 patent/WO2022056374A1/en active Application Filing
- 2021-09-13 BR BR112023004578A patent/BR112023004578A2/en unknown
- 2021-09-13 AU AU2021338858A patent/AU2021338858A1/en active Pending
- 2021-09-13 US US18/025,475 patent/US20230326014A1/en active Pending
- 2021-09-13 WO PCT/US2021/050050 patent/WO2022056370A1/en active Application Filing
- 2021-09-13 CN CN202180075757.5A patent/CN116438585A/en active Pending
- 2021-09-13 KR KR1020237012202A patent/KR20230084176A/en unknown
- 2021-09-13 EP EP21867751.6A patent/EP4210564A4/en active Pending
- 2021-09-13 AU AU2021339829A patent/AU2021339829B2/en active Active
- 2021-09-13 EP EP21867754.0A patent/EP4211606A4/en active Pending
- 2021-09-13 CA CA3192441A patent/CA3192441A1/en active Pending
- 2021-09-13 KR KR1020237012203A patent/KR20230075468A/en active Search and Examination
- 2021-09-13 CA CA3192444A patent/CA3192444A1/en active Pending
- 2021-09-13 JP JP2023515551A patent/JP7545575B2/en active Active
- 2021-09-13 BR BR112023004565A patent/BR112023004565A2/en unknown
-
2024
- 2024-09-09 AU AU2024219507A patent/AU2024219507A1/en active Pending
Also Published As
Publication number | Publication date |
---|---|
CN116438585A (en) | 2023-07-14 |
EP4210564A1 (en) | 2023-07-19 |
WO2022056370A1 (en) | 2022-03-17 |
JP2023541165A (en) | 2023-09-28 |
US20230326165A1 (en) | 2023-10-12 |
WO2022056374A1 (en) | 2022-03-17 |
EP4210564A4 (en) | 2024-10-16 |
BR112023004565A2 (en) | 2023-04-04 |
KR20230075468A (en) | 2023-05-31 |
AU2021339829A1 (en) | 2023-05-25 |
EP4211606A4 (en) | 2024-10-09 |
JP7523680B2 (en) | 2024-07-26 |
KR20230084176A (en) | 2023-06-12 |
AU2021338858A1 (en) | 2023-05-18 |
AU2021339829B2 (en) | 2024-06-13 |
JP7545575B2 (en) | 2024-09-04 |
CA3192441A1 (en) | 2022-03-17 |
AU2024219507A1 (en) | 2024-09-26 |
EP4211606A1 (en) | 2023-07-19 |
CA3192444A1 (en) | 2022-03-17 |
BR112023004578A2 (en) | 2023-04-11 |
JP2023541841A (en) | 2023-10-04 |
AU2021339829A9 (en) | 2024-02-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11816753B2 (en) | Automated evaluation of human embryos | |
CN110309856A (en) | Image classification method, the training method of neural network and device | |
US11151426B2 (en) | System and method for clustering products by combining attribute data with image recognition | |
US11710552B2 (en) | Method and system for refining label information | |
US20230419170A1 (en) | System and method for efficient machine learning | |
US12079987B2 (en) | Automated quality assessment of ultra-widefield angiography images | |
Lafraxo et al. | Computer-aided system for bleeding detection in wce images based on cnn-gru network | |
Subasi | Applications of artificial intelligence in medical imaging | |
AU2021339829B2 (en) | Automated aneuploidy screening using arbitrated ensembles | |
BalaKrishna et al. | Autism spectrum disorder detection using machine learning | |
Cao et al. | Alzheimer’s Disease Stage Detection Method Based on Convolutional Neural Network | |
Lotus et al. | Unleashing the Potential of Deep Learning in Diabetic Retinopathy: A Comprehensive Survey | |
Jha et al. | Diabetic Retinopathy Detection | |
Sharma et al. | Plant Disease Detection using Machine Learning | |
Manoranjitham | An artificial intelligence ensemble model for paddy leaf disease diagnosis utilizing deep transfer learning | |
Dastkarvelayati et al. | Explainable AI by Training Introspection | |
Abeud et al. | Deep Learning Model for Fashions and Clothes Automated Classification | |
WO2024026004A1 (en) | Method and system for automatic her2 scoring | |
CN116456892A (en) | Determination of position in germ cell structure | |
Manavalan | Diagnosing Epidermal basal Squamous Cell Carcinoma in High-resolution, and Poorly Labeled Histopathological Imaging | |
RODRÍGUEZ | Cost-sensitive classification based on Bregman divergences |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
AS | Assignment |
Owner name: THE BRIGHAM AND WOMEN'S HOSPITAL, INC., MASSACHUSETTS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SHAFIEE, HADI;KANAKASABAPATHY, MANOJ KUMAR;THIRUMALARAJU, PRUDHVI;REEL/FRAME:067638/0825 Effective date: 20240605 Owner name: THE GENERAL HOSPITAL CORPORATION, MASSACHUSETTS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BORMANN, CHARLES;REEL/FRAME:067638/0799 Effective date: 20240604 |