EP4210564A1 - Determining locations in reproductive cellular structures - Google Patents
Determining locations in reproductive cellular structuresInfo
- Publication number
- EP4210564A1 EP4210564A1 EP21867751.6A EP21867751A EP4210564A1 EP 4210564 A1 EP4210564 A1 EP 4210564A1 EP 21867751 A EP21867751 A EP 21867751A EP 4210564 A1 EP4210564 A1 EP 4210564A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- cellular structure
- location
- clinical parameter
- reproductive cellular
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000001850 reproductive effect Effects 0.000 title claims abstract description 73
- 210000003850 cellular structure Anatomy 0.000 title claims abstract description 70
- 238000000034 method Methods 0.000 claims abstract description 68
- 238000013528 artificial neural network Methods 0.000 claims abstract description 45
- 210000001161 mammalian embryo Anatomy 0.000 claims description 16
- 210000000287 oocyte Anatomy 0.000 claims description 15
- 210000004508 polar body Anatomy 0.000 claims description 12
- 230000000306 recurrent effect Effects 0.000 claims description 9
- 210000001109 blastomere Anatomy 0.000 claims description 8
- 238000002347 injection Methods 0.000 claims description 8
- 239000007924 injection Substances 0.000 claims description 8
- 238000013527 convolutional neural network Methods 0.000 claims description 7
- 230000035558 fertility Effects 0.000 claims description 6
- 230000012447 hatching Effects 0.000 claims description 6
- NIXOWILDQLNWCW-UHFFFAOYSA-N acrylic acid group Chemical group C(C=C)(=O)O NIXOWILDQLNWCW-UHFFFAOYSA-N 0.000 claims description 3
- 230000000295 complement effect Effects 0.000 claims description 2
- 239000004065 semiconductor Substances 0.000 claims description 2
- 230000015654 memory Effects 0.000 description 19
- 230000006870 function Effects 0.000 description 18
- 238000012545 processing Methods 0.000 description 16
- 238000004891 communication Methods 0.000 description 8
- 230000008569 process Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 7
- 210000002459 blastocyst Anatomy 0.000 description 5
- 210000002257 embryonic structure Anatomy 0.000 description 5
- 210000004340 zona pellucida Anatomy 0.000 description 5
- 238000004458 analytical method Methods 0.000 description 4
- 210000004027 cell Anatomy 0.000 description 4
- 238000011161 development Methods 0.000 description 4
- 230000018109 developmental process Effects 0.000 description 4
- 238000012549 training Methods 0.000 description 4
- 238000003776 cleavage reaction Methods 0.000 description 3
- 230000004720 fertilization Effects 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- 210000002569 neuron Anatomy 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000007017 scission Effects 0.000 description 3
- 230000003068 static effect Effects 0.000 description 3
- 230000001413 cellular effect Effects 0.000 description 2
- 230000036512 infertility Effects 0.000 description 2
- 208000000509 infertility Diseases 0.000 description 2
- 231100000535 infertility Toxicity 0.000 description 2
- 229920000747 poly(lactic acid) Polymers 0.000 description 2
- 239000004626 polylactic acid Substances 0.000 description 2
- 230000035935 pregnancy Effects 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 241000282414 Homo sapiens Species 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 208000036878 aneuploidy Diseases 0.000 description 1
- 231100001075 aneuploidy Toxicity 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000001574 biopsy Methods 0.000 description 1
- 238000000339 bright-field microscopy Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000009429 distress Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 239000012530 fluid Substances 0.000 description 1
- 230000002068 genetic effect Effects 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 238000002513 implantation Methods 0.000 description 1
- 238000000338 in vitro Methods 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 230000031864 metaphase Effects 0.000 description 1
- 230000000394 mitotic effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 230000006403 short-term memory Effects 0.000 description 1
- 239000000243 solution Substances 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 238000013526 transfer learning Methods 0.000 description 1
- 210000000143 trophectoderm cell Anatomy 0.000 description 1
- 210000004291 uterus Anatomy 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/698—Matching; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/70—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30044—Fetus; Embryo
Definitions
- the present invention relates generally to the field of assisted fertility, and more particularly, to determining locations in reproductive cellular structures.
- Infertility is an underestimated healthcare problem that affects over forty-eight million couples globally and is a cause of distress, depression, and discrimination.
- IVF assisted reproductive technologies
- ART assisted reproductive technologies
- IVF in-vitro fertilization
- ICM inner cell mass
- a system in accordance with an aspect of the present invention, includes a processor and a non-transitory computer readable medium storing machine executable instructions for assigning a value representing a location of interest within a reproductive cellular structure.
- the machine executable instructions include an imager interface that receives an image of the reproductive cellular structure from an associated imager.
- a neural network determines a clinical parameter representing the location of interest within the reproductive cellular structure from the image of the reproductive cellular structure.
- the clinical parameter is stored at the non- transitory computer readable medium.
- a method for determining a clinical parameter representing the location of interest within a reproductive cellular structure.
- An image of the reproductive cellular structure is obtained and provided to a neural network to generate the clinical parameter.
- the clinical parameter is stored on a non-transitory computer readable medium.
- a method for planning an assisted fertility procedure.
- An image of a reproductive cellular structure is obtained and provided to a neural network to generate a clinical parameter representing a location of interest within the reproductive cellular structure.
- a procedure is performed on the reproductive cellular structure at a location determined according to the clinical parameter.
- FIG. 1 illustrates one example of a system for determining a location of interest on a reproductive cellular structure
- FIG. 2 illustrates another example of a system for determining a location of interest on a reproductive cellular structure
- FIG. 3 illustrates a method for determining a clinical parameter representing the location of interest within a reproductive cellular structure
- FIG. 4 illustrates a method for planning an assisted fertility procedure
- FIG. 5 is a schematic block diagram illustrating an exemplary system of hardware components capable of implementing examples of the systems and methods disclosed herein.
- Intracytoplasmic sperm injection is a procedure that includes alignment of metaphase (Mil) oocytes, selection, and immobilization of sperm, and injection of sperm at a precise location that does not interfere with the mitotic spindle.
- the spindle is located adjacent to the extruded polar body and cannot be visualized using bright field microscopy.
- Assisted hatching is a procedure designed to enable embryo escape from the zona pellucida (ZP).
- AH may increase the chance of pregnancy in older women with repeat IVF failure and in frozen embryo transfer cycles. This procedure is widely used on cleavage stage embryos to facilitate herniation and biopsy of trophectoderm cells for Preimplantation Genetic Testing. Blastomeres can be easily damaged if AH is performed too close to healthy cells.
- a “reproductive cellular structure”, as used herein, is a single cell or multicellular structure involved in an assisted fertility procedure on a mammalian subject. Reproductive cellular structures can include gametes, such as oocytes, as well as fertilized embryos prior to implantation.
- a “static observation,” as used herein, is an image or group of images of a reproductive cellular structure that represent a single point in the development of the reproductive cellular structure. Where multiple images are used in a static observation, no discernible change in the structure and appearance of the reproductive cellular structure will have taken place between images.
- FIG. 1 illustrates one example of a system 100 for determining a location of interest on a reproductive cellular structure.
- the system 100 includes a processor 102 and a non-transitory computer readable medium 110 that stores machine executable instructions for assigning a value representing a location of interest within a reproductive cellular structure.
- the machine executable instructions include an imager interface 112 that receives an image of the reproductive cellular structure from an associated imager.
- the imaging interface 112 can receive the image from the imager via a bus or network connection and condition the image for analysis at a neural network 114.
- the neural network 114 can be implemented on a cloud computing system, with the image transmitted to the server containing the neural network 114 via a network interface (not shown).
- the neural network 1 14 determines, from the image of the reproductive cellular structure, a clinical parameter representing the location of interest within the reproductive cellular structure.
- each possible value of the categorical clinical parameter represents a designated location within a representation of the reproductive cellular structure.
- the representation can be substantially circular, and the various values for the clinical parameter can represent individual sectors of the circle.
- the representation is divided into twelve thirty-degree sectors.
- the representation is divided into an array of tiled polygons, and with each polygon in the array being represented by one of the values for the clinical parameter.
- the reproductive cellular structure is an oocyte and the clinical parameter represents a location of the polar body within the oocyte.
- the reproductive cellular structure is an embryo and the clinical parameter represents a location on the zona pellucida that is furthest from a healthy blastomere.
- the neural network 1 14 includes a plurality of nodes having a plurality of interconnections. Values from the image, for example luminance and/or chrominance values associated with the individual pixels, are provided to a plurality of input nodes. The input nodes each provide these input values to layers of one or more intermediate nodes. A given intermediate node receives one or more output values from previous nodes. The received values are weighted according to a series of weights established during the training of the classifier. An intermediate node translates its received values into a single output according to an activation function at the node.
- the intermediate node can sum the received values and subject the sum to an identify function, a step function, a sigmoid function, a hyperbolic tangent, a rectified linear unit, a leaky rectified linear unit, a parametric rectified linear unit, a Gaussian error linear unit, the softplus function, an exponential linear unit, a scaled exponential linear unit, a Gaussian function, a sigmoid linear unit, a growing cosine unit, the Heaviside function, and the mish function.
- a final layer of nodes provides the confidence values for the output classes of the ANN, with each node having an associated value representing a confidence for one of the associated output classes of the classifier.
- ANN classifiers are fully-connected and feedforward.
- a convolutional neural network includes convolutional layers in which nodes from a previous layer are only connected to a subset of the nodes in the convolutional layer.
- Recurrent neural networks are a class of neural networks in which connections between nodes form a directed graph along a temporal sequence. Unlike a feedforward network, recurrent neural networks can incorporate feedback from states caused by earlier inputs, such that an output of the recurrent neural network for a given input can be a function of not only the input but one or more previous inputs.
- LSTM Long Short-Term Memory
- LSTM Long Short-Term Memory
- the neural network 1 14 is trained on a plurality of labeled images of the appropriate reproductive cellular structure.
- labeled images it is meant that the position of the location of interest within the image is known, for example, via expert annotation, and the clinical parameter associated with the position of the location of interest is provided to the neural network along with the image during the training process.
- the weights associated with the interconnections among nodes in the neural network 114 are iteratively changed until, once the network is changed, an output of the network when presented with a novel, unlabeled image provides a clinical parameter representing the location of the location of interest within the novel image.
- This clinical parameter or a representation of the clinical parameter can be stored on the non-transitory computer readable medium 110 and/or provided to a user via an associated output device.
- FIG. 2 illustrates another example of a system 200 for determining a location of interest on a reproductive cellular structure.
- the system 200 generates a categorical clinical parameter representing a location on the reproductive cellular structure.
- the system 200 includes an imager 202 that acquires an image of the reproductive cellular structure on at least one day of development.
- the imager 202 can include one or more cameras, capable of producing images in the visible or infrared range, paired with appropriate optics to provide an image of a reproductive cellular structure.
- the imager 202 can be implemented to capture images of an embryo at multiple days of development as part of a time-lapse embryo imaging system.
- the imager 202 can be configured to generate a static observation of the reproductive cellular structure as a set of one or more images.
- the imager 202 includes an attachment for a mobile device that operates with a camera of the mobile device to provide the images of the reproductive cellular structure.
- the housing for the attachment can be 3-D printed using polylactic acid with dimensions of 82 x 34 x 48 mm.
- An acrylic lens can be included in the housing to provided appropriate magnification for the embryo images.
- the imager 202 can be implemented as a standalone system with an optical housing that is 3-D printed from polylactic acid and overall dimensions of 62 x 92 x 175 mm.
- the housing contains an electronic circuit with a white light-emitting diode, a three-volt battery, and a single pole double-throw switch.
- the embryo sample is transilluminated, with a 10x Plan-Achromatic objective lens for image magnification and a complementary metal-oxide-semiconductor (CMOS) image sensor for embryo image data acquisition.
- CMOS complementary metal-oxide-semiconductor
- the CMOS sensor can be connected to a single-board computer to process the captured images.
- the imager 202 can be connected to a mobile device via a wireless connection (e.g., Wi-Fi, Bluetooth, or a similar connection) for data processing and visualization.
- a wireless connection e.g., Wi-Fi, Bluetooth, or a similar connection
- the one or more images obtained at the imager 202 are provided to an analysis system 210 comprising a processor 212, an output device 214, and a non- transitory computer readable medium 220 storing instructions executable by the processor.
- the instructions are executable to provide an imager interface 222 that receives the image or images of the reproductive cellular structure.
- the imager interface 222 can apply one or more imaging condition techniques, such as cropping and filtering, to better prepare the image for analysis.
- the images are then provided to a neural network 224 that provides the categorical clinical parameter representing the desired location.
- the neural network 224 can be a convolutional neural network, which is a feed-forward artificial neural network that includes convolutional layers, which effectively apply a convolution to the values at the preceding layer of the network to emphasize various sets of features within an image.
- a convolutional layer each neuron is connected only to a proper subset of the neurons in the preceding layer, referred to as the receptive field of the neuron.
- the convolutional neural network is implemented using the Xception architecture.
- at least one chromatic value (e.g., a value for an RGB color channel, a YCrCb color channel, or a grayscale brightness) associated with each pixel is provided as an initial input to the convolutional neural network.
- the neural network 224 can be implemented as a recurrent neural network.
- a recurrent neural network the connections between nodes in the network are selected to form a directed graph along a sequence, allowing it to exhibit dynamic temporal behavior.
- the neural network 224 is implemented and trained as a discriminative network in a generative adversarial model, in which a generative neural network and the discriminative network provide mutual feedback to one another, such that the generative neural network produces increasingly sophisticated samples for the discriminative network to attempt to classify.
- some or all layers of the neural network can be trained via transfer learning from another system, with only some of the layers trained on the training images of the reproductive cellular structure.
- a final layer of the neural network 224 can be implemented as a softmax layer to provide a classification result.
- the neural network 224 In response to a novel image, the neural network 224 generates a clinical parameter representing the portion of the image containing the location of interest.
- the clinical parameter can be provided to a user at the output device 214 via a user interface 226.
- the user interface 226 can include appropriate software instructions for receiving the output of the neural network 224 and presenting it at the output device 214.
- the output device 214 can include a mobile device that communicates wirelessly with the analysis system 210.
- the clinical parameter can be provided to the user as a representation of the cellular reproductive structure with the section of the cellular reproductive structure highlighted within the representation.
- FIGS. 3 and 4 While, for purposes of simplicity of explanation, the methods of FIGS. 3 and 4is shown and described as executing serially, it is to be understood and appreciated that the present invention is not limited by the illustrated order, as some aspects could, in accordance with the present invention, occur in different orders and/or concurrently with other aspects from that shown and described herein. Moreover, not all illustrated features may be required to implement a method in accordance with an aspect the present invention.
- FIG. 3 illustrates one method 300 for determining a clinical parameter representing the location of interest within a reproductive cellular structure.
- the method 300 can be used to determine the location of a polar body within an oocyte, with the clinical parameter representing a portion of the oocyte containing the polar body.
- the method can be used to determine a location on the zona pellucida that is furthest from a healthy blastomere, with the clinical parameter representing the appropriate location.
- an image of the reproductive cellular structure is obtained. For example, an image can be captured at an appropriate imager and provided to a computer system for image processing.
- the image of the reproductive cellular structure is provided to a neural network to generate the clinical parameter.
- Each possible value of the clinical parameter represents a designated location within a representation of the reproductive cellular structure, such that the clinical parameter indicates the location of interest within the image.
- the neural network can be implemented as any of a convolutional neural network, a recurrent neural network, and a discriminative classifier trained as part of a generative adversarial network.
- the clinical parameter is stored on a non- transitory computer readable medium.
- FIG. 4 illustrates a method 400 for planning an assisted fertility procedure.
- an image of the reproductive cellular structure is obtained.
- an image can be captured at an appropriate imager and provided to a computer system for image processing.
- the image of the reproductive cellular structure is provided to a neural network to generate a clinical parameter representing a location of interest within the reproductive cellular structure.
- Each possible value of the clinical parameter represents a designated location within a representation of the reproductive cellular structure, such that the clinical parameter indicates the location of interest within the image.
- the neural network can be implemented as any of a convolutional neural network, a recurrent neural network, and a discriminative classifier trained as part of a generative adversarial network.
- a procedure can be performed on the reproductive cellular structure at a location determined according to the clinical parameter.
- the method 400 can be used to determine the location of a polar body within an oocyte, with the clinical parameter representing a portion of the oocyte containing the polar body.
- the location of the polar body can be used to determine an appropriate location for intracytoplasmic sperm injection.
- the intracytoplasmic sperm injection can be performed at a location spaced ninety degrees from the polar body.
- the method 400 can be used to determine a location on the zona pellucida that is furthest from a healthy blastomere to avoid damaging healthy blastomeres during laser assisted hatching.
- the procedure can be performed by an automated robotic system based upon the location represented by the clinical parameter.
- FIG. 5 is a schematic block diagram illustrating an exemplary system 500 of hardware components capable of implementing examples of the systems and methods disclosed in FIGS. 1 -4, such as the systems illustrated in FIGS. 1 and 2.
- the system 500 can include various systems and subsystems.
- the system 500 can be any of personal computer, a laptop computer, a workstation, a computer system, an appliance, an application-specific integrated circuit (ASIC), a server, a server blade center, or a server farm.
- ASIC application-specific integrated circuit
- the system 500 can includes a system bus 502, a processing unit 504, a system memory 506, memory devices 508 and 510, a communication interface 512 (e.g., a network interface), a communication link 514, a display 516 (e.g., a video screen), and an input device 518 (e.g., a keyboard and/or a mouse).
- the system bus 502 can be in communication with the processing unit 504 and the system memory 506.
- the additional memory devices 508 and 510 such as a hard disk drive, server, stand-alone database, or other non-volatile memory, can also be in communication with the system bus 502.
- the system bus 502 interconnects the processing unit 504, the memory devices 506-510, the communication interface 512, the display 516, and the input device 518. In some examples, the system bus 502 also interconnects an additional port (not shown), such as a universal serial bus (USB) port.
- an additional port not shown, such as a universal serial bus (USB) port.
- USB universal serial bus
- the system 500 could be implemented in a computing cloud.
- features of the system 500 such as the processing unit 504, the communication interface 512, and the memory devices 508 and 510 could be representative of a single instance of hardware or multiple instances of hardware with applications executing across the multiple of instances (i.e., distributed) of hardware (e.g., computers, routers, memory, processors, or a combination thereof).
- the system 500 could be implemented on a single dedicated server.
- the processing unit 504 can be a computing device and can include an application-specific integrated circuit (ASIC).
- the processing unit 504 executes a set of instructions to implement the operations of examples disclosed herein.
- the processing unit can include a processing core.
- the additional memory devices 506, 508, and 510 can store data, programs, instructions, database queries in text or compiled form, and any other information that can be needed to operate a computer.
- the memories 506, 508 and 510 can be implemented as computer-readable media (integrated or removable) such as a memory card, disk drive, compact disk (CD), or server accessible over a network.
- the memories 506, 508 and 510 can comprise text, images, video, and/or audio, portions of which can be available in formats comprehensible to human beings.
- the system 500 can access an external data source or query source through the communication interface 512, which can communicate with the system bus 502 and the communication link 514.
- the system 500 can be used to implement one or more parts of an embryo evaluation system in accordance with the present invention.
- Computer executable logic for implementing the composite applications testing system resides on one or more of the system memory 506, and the memory devices 508, 510 in accordance with certain examples.
- the processing unit 504 executes one or more computer executable instructions originating from the system memory 506 and the memory devices 508 and 510. It will be appreciated that a computer readable medium can include multiple computer readable media each operatively connected to the processing unit.
- the processing units can be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described above, and/or a combination thereof.
- ASICs application specific integrated circuits
- DSPs digital signal processors
- DSPDs digital signal processing devices
- PLDs programmable logic devices
- FPGAs field programmable gate arrays
- processors controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described above, and/or a combination thereof.
- the embodiments can be described as a process which is depicted as a flowchart, a flow diagram, a data flow diagram, a structure diagram, or a block diagram. Although a flowchart can describe the operations as a sequential process, many of the operations can be performed in parallel or concurrently. In addition, the order of the operations can be re-arranged. A process is terminated when its operations are completed, but could have additional steps not included in the figure. A process can correspond to a method, a function, a procedure, a subroutine, a subprogram, etc. When a process corresponds to a function, its termination corresponds to a return of the function to the calling function or the main function.
- embodiments can be implemented by hardware, software, scripting languages, firmware, middleware, microcode, hardware description languages, and/or any combination thereof.
- the program code or code segments to perform the necessary tasks can be stored in a machine readable medium such as a storage medium.
- a code segment or machine-executable instruction can represent a procedure, a function, a subprogram, a program, a routine, a subroutine, a module, a software package, a script, a class, or any combination of instructions, data structures, and/or program statements.
- a code segment can be coupled to another code segment or a hardware circuit by passing and/or receiving information, data, arguments, parameters, and/or memory contents.
- Information, arguments, parameters, data, etc. can be passed, forwarded, or transmitted via any suitable means including memory sharing, message passing, ticket passing, network transmission, etc.
- the methodologies can be implemented with modules (e.g., procedures, functions, and so on) that perform the functions described herein.
- Any machine-readable medium tangibly embodying instructions can be used in implementing the methodologies described herein.
- software codes can be stored in a memory.
- Memory can be implemented within the processor or external to the processor.
- the term “memory” refers to any type of long term, short term, volatile, nonvolatile, or other storage medium and is not to be limited to any particular type of memory or number of memories, or type of media upon which memory is stored.
- the term “storage medium” can represent one or more memories for storing data, including read only memory (ROM), random access memory (RAM), magnetic RAM, core memory, magnetic disk storage mediums, optical storage mediums, flash memory devices and/or other machine readable mediums for storing information.
- ROM read only memory
- RAM random access memory
- magnetic RAM magnetic RAM
- core memory magnetic disk storage mediums
- optical storage mediums flash memory devices and/or other machine readable mediums for storing information.
- computer readable medium includes, but is not limited to portable or fixed storage devices, optical storage devices, wireless channels, and/or various other storage mediums capable of storing that contain or carry instruction(s) and/or data.
- a “computer readable medium” or “machine readable medium” can include multiple media each operatively connected to a processing unit. In such a case, when it is stated that data is stored at the computer readable medium, it can refer to any of the interconnected media within the system.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Biomedical Technology (AREA)
- Multimedia (AREA)
- Public Health (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Primary Health Care (AREA)
- Epidemiology (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Pathology (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Quality & Reliability (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Image Analysis (AREA)
- Investigating Or Analysing Biological Materials (AREA)
- Apparatus Associated With Microorganisms And Enzymes (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Devices For Executing Special Programs (AREA)
- Measuring Or Testing Involving Enzymes Or Micro-Organisms (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- General Factory Administration (AREA)
- Image Processing (AREA)
Abstract
Description
Claims
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US202063077398P | 2020-09-11 | 2020-09-11 | |
US202063077405P | 2020-09-11 | 2020-09-11 | |
PCT/US2021/050050 WO2022056370A1 (en) | 2020-09-11 | 2021-09-13 | Determining locations in reproductive cellular structures |
Publications (2)
Publication Number | Publication Date |
---|---|
EP4210564A1 true EP4210564A1 (en) | 2023-07-19 |
EP4210564A4 EP4210564A4 (en) | 2024-10-16 |
Family
ID=80630102
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP21867754.0A Pending EP4211606A4 (en) | 2020-09-11 | 2021-09-13 | Automated aneuploidy screening using arbitrated ensembles |
EP21867751.6A Pending EP4210564A4 (en) | 2020-09-11 | 2021-09-13 | Determining locations in reproductive cellular structures |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP21867754.0A Pending EP4211606A4 (en) | 2020-09-11 | 2021-09-13 | Automated aneuploidy screening using arbitrated ensembles |
Country Status (9)
Country | Link |
---|---|
US (2) | US20230326014A1 (en) |
EP (2) | EP4211606A4 (en) |
JP (2) | JP7523680B2 (en) |
KR (2) | KR20230075468A (en) |
CN (1) | CN116438585A (en) |
AU (3) | AU2021338858A1 (en) |
BR (2) | BR112023004565A2 (en) |
CA (2) | CA3192441A1 (en) |
WO (2) | WO2022056374A1 (en) |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014134550A1 (en) | 2013-02-28 | 2014-09-04 | Auxogyn, Inc. | Apparatus, method, and system for image-based human embryo cell classification |
US10902334B2 (en) | 2016-01-28 | 2021-01-26 | Gerard Letterie | Automated image analysis to assess reproductive potential of human oocytes and pronuclear embryos |
GB201806999D0 (en) * | 2018-04-30 | 2018-06-13 | Univ Birmingham | Automated oocyte detection and orientation |
US20210200986A1 (en) | 2018-05-25 | 2021-07-01 | Sony Corporation | Control device, control method, and program |
KR20210078488A (en) | 2018-09-20 | 2021-06-28 | 에이아이브이에프 엘티디 | Image feature detection |
US11926809B2 (en) * | 2018-09-28 | 2024-03-12 | Brigham And Women's Hospital, Inc. | Automated evaluation of sperm morphology |
CN113260894B (en) | 2018-12-28 | 2023-04-04 | 仪景通株式会社 | Microscope system |
WO2020157761A1 (en) * | 2019-01-31 | 2020-08-06 | Amnon Buxboim | Automated evaluation of embryo implantation potential |
US10646156B1 (en) * | 2019-06-14 | 2020-05-12 | Cycle Clarity, LLC | Adaptive image processing in assisted reproductive imaging modalities |
-
2021
- 2021-09-13 US US18/025,475 patent/US20230326014A1/en active Pending
- 2021-09-13 KR KR1020237012203A patent/KR20230075468A/en active Search and Examination
- 2021-09-13 EP EP21867754.0A patent/EP4211606A4/en active Pending
- 2021-09-13 US US18/025,480 patent/US20230326165A1/en active Pending
- 2021-09-13 CN CN202180075757.5A patent/CN116438585A/en active Pending
- 2021-09-13 BR BR112023004565A patent/BR112023004565A2/en unknown
- 2021-09-13 AU AU2021338858A patent/AU2021338858A1/en active Pending
- 2021-09-13 JP JP2023516276A patent/JP7523680B2/en active Active
- 2021-09-13 WO PCT/US2021/050058 patent/WO2022056374A1/en active Application Filing
- 2021-09-13 CA CA3192441A patent/CA3192441A1/en active Pending
- 2021-09-13 AU AU2021339829A patent/AU2021339829B2/en active Active
- 2021-09-13 BR BR112023004578A patent/BR112023004578A2/en unknown
- 2021-09-13 WO PCT/US2021/050050 patent/WO2022056370A1/en active Application Filing
- 2021-09-13 CA CA3192444A patent/CA3192444A1/en active Pending
- 2021-09-13 KR KR1020237012202A patent/KR20230084176A/en unknown
- 2021-09-13 JP JP2023515551A patent/JP7545575B2/en active Active
- 2021-09-13 EP EP21867751.6A patent/EP4210564A4/en active Pending
-
2024
- 2024-09-09 AU AU2024219507A patent/AU2024219507A1/en active Pending
Also Published As
Publication number | Publication date |
---|---|
AU2021339829B2 (en) | 2024-06-13 |
US20230326165A1 (en) | 2023-10-12 |
US20230326014A1 (en) | 2023-10-12 |
BR112023004565A2 (en) | 2023-04-04 |
BR112023004578A2 (en) | 2023-04-11 |
AU2021338858A1 (en) | 2023-05-18 |
JP2023541841A (en) | 2023-10-04 |
EP4211606A1 (en) | 2023-07-19 |
EP4211606A4 (en) | 2024-10-09 |
KR20230084176A (en) | 2023-06-12 |
JP7545575B2 (en) | 2024-09-04 |
CA3192444A1 (en) | 2022-03-17 |
KR20230075468A (en) | 2023-05-31 |
AU2024219507A1 (en) | 2024-09-26 |
WO2022056374A1 (en) | 2022-03-17 |
WO2022056370A1 (en) | 2022-03-17 |
AU2021339829A1 (en) | 2023-05-25 |
CN116438585A (en) | 2023-07-14 |
EP4210564A4 (en) | 2024-10-16 |
CA3192441A1 (en) | 2022-03-17 |
JP2023541165A (en) | 2023-09-28 |
JP7523680B2 (en) | 2024-07-26 |
AU2021339829A9 (en) | 2024-02-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111279421B (en) | Automatic evaluation of human embryos | |
Louis et al. | Review of computer vision application in in vitro fertilization: the application of deep learning-based computer vision technology in the world of IVF | |
AU2024200572A1 (en) | Automated evaluation of quality assurance metrics for assisted reproduction procedures | |
Baručić et al. | Automatic evaluation of human oocyte developmental potential from microscopy images | |
CN111401183A (en) | Artificial intelligence-based cell body monitoring method, system, device and electronic equipment | |
US20230326165A1 (en) | Determining locations in reproductive cellular structures | |
CN116456892A (en) | Determination of position in germ cell structure | |
CN115036021A (en) | Embryo development monitoring method based on space dynamics parameters | |
EP4178485A1 (en) | Predicting embryo implantation probability | |
RU2810125C1 (en) | Automated assessment of quality assurance indicators for assisted reproduction procedures | |
Eswaran et al. | Deep Learning Algorithms for Timelapse Image Sequence-Based Automated Blastocyst Quality Detection | |
US20240037743A1 (en) | Systems and methods for evaluating embryo viability using artificial intelligence | |
WO2022150914A1 (en) | Systems and methods for non-invasive preimplantation embryo genetic screening |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20230411 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
P01 | Opt-out of the competence of the unified patent court (upc) registered |
Effective date: 20230728 |
|
DAV | Request for validation of the european patent (deleted) | ||
DAX | Request for extension of the european patent (deleted) | ||
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R079 Free format text: PREVIOUS MAIN CLASS: A61B0005000000 Ipc: G16H0050200000 |
|
A4 | Supplementary search report drawn up and despatched |
Effective date: 20240917 |
|
RIC1 | Information provided on ipc code assigned before grant |
Ipc: G06V 20/69 20220101ALI20240911BHEP Ipc: G06V 10/82 20220101ALI20240911BHEP Ipc: G06V 10/25 20220101ALI20240911BHEP Ipc: G06T 7/70 20170101ALI20240911BHEP Ipc: G16H 30/40 20180101ALI20240911BHEP Ipc: G16H 50/70 20180101ALI20240911BHEP Ipc: G16H 50/20 20180101AFI20240911BHEP |