WO2019150755A1 - 撮像システム - Google Patents
撮像システム Download PDFInfo
- Publication number
- WO2019150755A1 WO2019150755A1 PCT/JP2018/044628 JP2018044628W WO2019150755A1 WO 2019150755 A1 WO2019150755 A1 WO 2019150755A1 JP 2018044628 W JP2018044628 W JP 2018044628W WO 2019150755 A1 WO2019150755 A1 WO 2019150755A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- cell
- imaging system
- biological object
- feature amount
- cells
- Prior art date
Links
- 238000004458 analytical method Methods 0.000 claims abstract description 22
- 238000003384 imaging method Methods 0.000 claims description 114
- 239000000284 extract Substances 0.000 claims description 6
- 230000002093 peripheral effect Effects 0.000 claims description 4
- CIWBSHSKHKDKBQ-JLAZNSOCSA-N Ascorbic acid Chemical compound OC[C@H](O)[C@H]1OC(=O)C(O)=C1O CIWBSHSKHKDKBQ-JLAZNSOCSA-N 0.000 description 161
- 210000004027 cell Anatomy 0.000 description 156
- 238000000034 method Methods 0.000 description 42
- 238000012545 processing Methods 0.000 description 19
- 206010028980 Neoplasm Diseases 0.000 description 12
- 201000011510 cancer Diseases 0.000 description 12
- 238000010586 diagram Methods 0.000 description 11
- 230000008569 process Effects 0.000 description 10
- 239000002609 medium Substances 0.000 description 7
- 238000004364 calculation method Methods 0.000 description 6
- 230000008859 change Effects 0.000 description 6
- 238000012546 transfer Methods 0.000 description 6
- 238000004113 cell culture Methods 0.000 description 4
- 239000011521 glass Substances 0.000 description 4
- 239000000463 material Substances 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 206010061902 Pancreatic neoplasm Diseases 0.000 description 3
- 230000004308 accommodation Effects 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 239000012535 impurity Substances 0.000 description 3
- 208000015486 malignant pancreatic neoplasm Diseases 0.000 description 3
- 239000011159 matrix material Substances 0.000 description 3
- 201000002528 pancreatic cancer Diseases 0.000 description 3
- 208000008443 pancreatic carcinoma Diseases 0.000 description 3
- 238000000605 extraction Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 239000011347 resin Substances 0.000 description 2
- 229920005989 resin Polymers 0.000 description 2
- 238000010187 selection method Methods 0.000 description 2
- 210000001519 tissue Anatomy 0.000 description 2
- 206010009944 Colon cancer Diseases 0.000 description 1
- 241000252212 Danio rerio Species 0.000 description 1
- 206010014733 Endometrial cancer Diseases 0.000 description 1
- 206010014759 Endometrial neoplasm Diseases 0.000 description 1
- 206010058467 Lung neoplasm malignant Diseases 0.000 description 1
- 241000244206 Nematoda Species 0.000 description 1
- 206010033128 Ovarian cancer Diseases 0.000 description 1
- 206010061535 Ovarian neoplasm Diseases 0.000 description 1
- 230000004931 aggregating effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 210000000601 blood cell Anatomy 0.000 description 1
- 239000006285 cell suspension Substances 0.000 description 1
- 239000003153 chemical reaction reagent Substances 0.000 description 1
- 208000029742 colonic neoplasm Diseases 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 235000013601 eggs Nutrition 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000001963 growth medium Substances 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 201000005202 lung cancer Diseases 0.000 description 1
- 208000020816 lung neoplasm Diseases 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 244000005700 microbiome Species 0.000 description 1
- 210000002220 organoid Anatomy 0.000 description 1
- 238000005192 partition Methods 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 241000894007 species Species 0.000 description 1
- 230000037303 wrinkles Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- C—CHEMISTRY; METALLURGY
- C12—BIOCHEMISTRY; BEER; SPIRITS; WINE; VINEGAR; MICROBIOLOGY; ENZYMOLOGY; MUTATION OR GENETIC ENGINEERING
- C12M—APPARATUS FOR ENZYMOLOGY OR MICROBIOLOGY; APPARATUS FOR CULTURING MICROORGANISMS FOR PRODUCING BIOMASS, FOR GROWING CELLS OR FOR OBTAINING FERMENTATION OR METABOLIC PRODUCTS, i.e. BIOREACTORS OR FERMENTERS
- C12M33/00—Means for introduction, transport, positioning, extraction, harvesting, peeling or sampling of biological material in or from the apparatus
- C12M33/04—Means for introduction, transport, positioning, extraction, harvesting, peeling or sampling of biological material in or from the apparatus by injection or suction, e.g. using pipettes, syringes, needles
- C12M33/06—Means for introduction, transport, positioning, extraction, harvesting, peeling or sampling of biological material in or from the apparatus by injection or suction, e.g. using pipettes, syringes, needles for multiple inoculation or multiple collection of samples
-
- C—CHEMISTRY; METALLURGY
- C12—BIOCHEMISTRY; BEER; SPIRITS; WINE; VINEGAR; MICROBIOLOGY; ENZYMOLOGY; MUTATION OR GENETIC ENGINEERING
- C12M—APPARATUS FOR ENZYMOLOGY OR MICROBIOLOGY; APPARATUS FOR CULTURING MICROORGANISMS FOR PRODUCING BIOMASS, FOR GROWING CELLS OR FOR OBTAINING FERMENTATION OR METABOLIC PRODUCTS, i.e. BIOREACTORS OR FERMENTERS
- C12M41/00—Means for regulation, monitoring, measurement or control, e.g. flow regulation
- C12M41/30—Means for regulation, monitoring, measurement or control, e.g. flow regulation of concentration
- C12M41/36—Means for regulation, monitoring, measurement or control, e.g. flow regulation of concentration of biomass, e.g. colony counters or by turbidity measurements
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N15/00—Investigating characteristics of particles; Investigating permeability, pore-volume or surface-area of porous materials
- G01N15/10—Investigating individual particles
- G01N15/14—Optical investigation techniques, e.g. flow cytometry
- G01N15/1429—Signal processing
- G01N15/1433—Signal processing using image recognition
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N33/00—Investigating or analysing materials by specific methods not covered by groups G01N1/00 - G01N31/00
- G01N33/48—Biological material, e.g. blood, urine; Haemocytometers
- G01N33/483—Physical analysis of biological material
- G01N33/4833—Physical analysis of biological material of solid biological material, e.g. tissue samples, cell cultures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/778—Active pattern-learning, e.g. online learning of image or video features
- G06V10/7784—Active pattern-learning, e.g. online learning of image or video features based on feedback from supervisors
- G06V10/7788—Active pattern-learning, e.g. online learning of image or video features based on feedback from supervisors the supervisor being a human, e.g. interactive learning with a human teacher
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/693—Acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/69—Microscopic objects, e.g. biological cells or cellular parts
- G06V20/695—Preprocessing, e.g. image segmentation
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/18—Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N15/00—Investigating characteristics of particles; Investigating permeability, pore-volume or surface-area of porous materials
- G01N15/10—Investigating individual particles
- G01N2015/1006—Investigating individual particles for cytology
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N15/00—Investigating characteristics of particles; Investigating permeability, pore-volume or surface-area of porous materials
- G01N15/10—Investigating individual particles
- G01N15/14—Optical investigation techniques, e.g. flow cytometry
- G01N2015/1486—Counting the particles
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N15/00—Investigating characteristics of particles; Investigating permeability, pore-volume or surface-area of porous materials
- G01N15/10—Investigating individual particles
- G01N15/14—Optical investigation techniques, e.g. flow cytometry
- G01N2015/1497—Particle shape
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10056—Microscopic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20092—Interactive image processing based on input by user
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30024—Cell structures in vitro; Tissue sections in vitro
Definitions
- the present invention relates to an imaging system including an imaging device that images a biological object such as a cell or a cell mass.
- imaging for selecting cells or cell masses may be performed.
- cells picked up on a plate having a large number of receiving recesses are imaged with an imaging device, desired cells are selected based on the obtained images, and the selected cells are aspirated with a chip and transferred to a microplate.
- an operation of performing is performed (for example, Patent Document 1).
- the above-mentioned manual selection method requires a lot of work for the operator because the operator observes individual cells and makes a pass / fail judgment.
- cell quality determination is likely to vary depending on individual subjectivity, experience level, etc., and it is difficult to obtain a uniform quality determination result.
- the method of setting the selection reference value in advance the quality of the cells can be automatically determined, so labor is reduced.
- An object of the present invention is to provide an imaging system capable of accurately selecting a living body object and reducing labor for selecting the living body object.
- An imaging system includes an imaging device capable of simultaneously imaging a plurality of biological objects, a display unit that displays an image including the plurality of biological objects captured by the imaging device, and an operator
- an input unit that accepts an input related to a selection operation for the plurality of biological objects displayed on the display unit, and an image corresponding to the biological object to which the selection operation is given by the input unit
- An analysis unit that extracts a feature amount of an object and a biological object that has been subjected to the selection operation up to a first time point, based on a feature amount extracted by the analysis unit or a predetermined reference feature amount
- a parameter that defines a range of the biological object to be processed is set. Characterized in that it and a specifying unit configured to specify a recommended biological object to prompt the operator to make selections based on.
- FIG. 1 is a diagram schematically illustrating a configuration example of a cell moving device to which an imaging system according to an embodiment of the present invention is applied.
- FIG. 2A is a top view of a dish provided in the sorting container used in the cell transfer device
- FIG. 2B is a cross-sectional view taken along the line IIB-IIB in FIG.
- FIG. 3 is a diagram schematically showing a flow of cell sorting work using the imaging system according to the present embodiment.
- 4A is a diagram illustrating an example of a cell shape
- FIG. 4B is a graph for explaining an example of a selection criterion based on the cell shape.
- FIG. 5 is a diagram for explaining a selection criterion based on the color of a cell.
- FIG. 1 is a diagram schematically illustrating a configuration example of a cell moving device to which an imaging system according to an embodiment of the present invention is applied.
- FIG. 2A is a top view of a dish provided in the sorting
- FIG. 6 is a diagram for explaining a selection criterion based on a cell pattern.
- FIGS. 7A and 7B are diagrams for explaining selection criteria based on the light intensity of cells.
- FIG. 8 is a block diagram of the imaging system according to the embodiment of the present invention.
- FIG. 9 is a flowchart of a cell sorting operation using the imaging system.
- a living body-derived cell can be typically exemplified.
- cells derived from living organisms include single cells (cells) such as blood cells and single cells, tissue pieces such as Histoculture and CTOS, cell aggregates such as spheroids and organoids, zebrafish, nematodes, and fertilized eggs An individual such as 2D or 3D colony.
- biological objects include tissues, microorganisms, small-sized species, and the like. In the embodiment described below, an example is shown in which the biological object is a cell or a cell aggregate formed by aggregating several to several hundreds of thousands of cells (hereinafter collectively referred to simply as “cell C”). .
- FIG. 1 is a diagram schematically showing an overall configuration of a cell transfer device S to which an imaging system according to an embodiment of the present invention is applied.
- a cell moving device S that moves cells C between two containers (dish 2 and microplate 4) is illustrated.
- the cell transfer device S includes a translucent base 1 having a horizontal placement surface (upper surface), a camera unit 5 (imaging device) disposed on the lower side of the base 1, and an upper side of the base 1. And the head unit 6 disposed in the.
- a sorting container 11 having a dish 2 (plate) is placed at the first placement position P1 of the base 1, and a microplate 4 is placed at the second placement position P2.
- the head unit 6 is equipped with a plurality of heads 61 to which a chip 12 that sucks and discharges cells C is mounted and which can move along the Z direction (vertical direction).
- the camera unit 5 and the head unit 6 are movable in the X direction (horizontal direction) and the direction perpendicular to the paper surface of FIG. 1 (Y direction).
- the dish 2 and the microplate 4 are placed on the upper surface of the base 1 within the movable range of the head unit 6.
- the cell transfer device S sucks the cells C individually from each of the plurality of chips 12 from the dish 2 of the sorting container 11 that holds a large number of cells C, and moves the cells C to the microplate 4.
- cells C are simultaneously discharged from a plurality of chips 12 to the microplate 4 (well 41).
- the cells C held on the dish 2 are imaged by the camera unit 5 and a sorting operation for sorting the high-quality cells C to be moved to the microplate 4 is performed.
- the base 1 is a rectangular flat plate having a predetermined rigidity and part or all of which is made of a translucent material.
- a preferred base 1 is a glass plate.
- the sorting container 11 is a container from which the cells C move, stores the medium L, and holds the cell sorting dish 2 in a state of being immersed in the medium L.
- the dish 2 is a plate for holding the cells C, and has a plurality of holding recesses 3 (sections for storing objects) capable of individually storing and holding the cells C on the upper surface.
- the medium L is not particularly limited as long as it does not deteriorate the properties of the cells C, and can be appropriately selected depending on the type of the cells C.
- the sorting container 11 has a rectangular upper opening 11H on the upper surface side.
- the upper opening 11H is an opening for loading the cells C and picking up the sorted cells C.
- the dish 2 is disposed below the upper opening 11H.
- the sorting container 11 and the dish 2 are made of a translucent resin material or glass. This is because the cells C carried on the dish 2 can be observed by the camera unit 5 arranged below the sorting container 11.
- a plurality of cells C dispersed in a cell culture solution are injected into the sorting container 11 from a dispensing chip (not shown).
- the dispensing tip aspirates the cell culture solution together with the cells C from a container that stores a cell culture solution containing a large amount of cells C, and holds it in the dispensing tip. Thereafter, the dispensing tip is moved to the upper position of the sorting container 11 and accesses the upper surface of the dish 2 through the upper opening 11H. Then, with the tip opening of the dispensing tip immersed in the medium L of the sorting container 11, the cells C held in the dispensing tip are discharged onto the dish 2 together with the cell culture solution.
- the microplate 4 is a container to which the cells C are moved, and has a plurality of wells 41 from which the cells C are discharged.
- the well 41 is a hole with a bottom opened on the upper surface of the microplate 4.
- a single well 41 accommodates a necessary number (usually one) of cells C together with the medium L.
- the microplate 4 is also made of a translucent resin material or glass. This is because the cell C carried in the well 41 can be observed by the camera unit 5 arranged below the microplate 4.
- the camera unit 5 captures an image of the cell C held in the sorting container 11 or the microplate 4 from the lower surface side, and includes a lens unit 51 and a camera body 52.
- the lens unit 51 is an objective lens used in an optical microscope, and includes a lens group that forms an optical image with a predetermined magnification and a lens barrel that houses the lens group.
- the camera body 52 includes an image sensor such as a CCD image sensor.
- the lens unit 51 forms an optical image of the imaging object on the light receiving surface of the imaging element.
- the camera unit 5 is movable in the X direction and the Y direction below the base 1 along a guide rail 5G extending in the left-right direction parallel to the base 1. Further, the lens unit 51 is movable in the Z direction for the focusing operation.
- the head unit 6 is provided for picking up the cells C from the dish 2 and moving them to the microplate 4 and includes a plurality of heads 61 and a head body 62 to which these heads 61 are assembled. At the tip of each head 61, a chip 12 that performs suction (pickup) and discharge of cells C is mounted.
- the head main body 62 holds the head 61 so as to be movable up and down in the + Z and ⁇ Z directions, and is movable in the + X and ⁇ X directions along the guide rail 6G.
- the head main body 62 can also move in the Y direction.
- FIG. 2A is a top view of the dish 2
- FIG. 2B is a cross-sectional view taken along the line IIB-IIB in FIG. 2A.
- the dish 2 includes a dish body 20 and a plurality of holding recesses 3 formed in the dish body 20.
- the dish body 20 is made of a flat plate member having a predetermined thickness, and has an upper surface 21 and a lower surface 22.
- the holding recess 3 has a cell C receiving opening (opening 31) on the upper surface 21 side.
- the dish 2 is immersed in the medium L in the sorting container 11. Specifically, the upper surface 21 of the dish body 20 is immersed in the culture medium L in the sorting container 11, while the lower surface 22 is held in the sorting container 11 in a state of being spaced from the bottom plate of the sorting container 11. (See FIG. 1).
- Each holding recess 3 includes an opening 31, a bottom 32, a cylindrical wall 33, a hole 34, and a boundary 35.
- an example is shown in which square holding recesses 3 are arranged in a matrix in a top view.
- the plurality of holding recesses 3 are arranged in a matrix at a predetermined recess array pitch.
- the opening 31 is a square opening provided on the upper surface 21 and has a size that allows the tip opening t of the sorting chip 12 to enter.
- the bottom 32 is located inside the dish body 20 and near the lower surface 22.
- the bottom 32 is an inclined surface that is gently inclined downward toward the center (the center of the square).
- the cylindrical wall surface 33 is a wall surface extending vertically downward from the opening 31 toward the bottom 32.
- the hole 34 is a through hole that vertically penetrates between the center of the bottom 32 and the lower surface 22.
- the boundary portion 35 is a portion that is located on the upper surface 21 and serves as an opening edge of each holding recess 3, and is a ridge line that partitions the holding recesses 3.
- each holding recess 3 define an accommodation space 3H that accommodates the cells C. It is contemplated that one cell C is generally accommodated in the accommodation space 3H.
- the hole 34 is provided in order to let small cells and impurities other than the desired size escape from the accommodation space 3H. Therefore, the size of the hole 34 is selected so that cells C having a desired size cannot pass through and small cells or impurities other than the desired size can pass therethrough. As a result, the cells C to be sorted are trapped in the holding recess 3, while impurities and the like fall from the hole 34 to the bottom plate of the sorting container 11.
- FIG. 3 is a diagram schematically showing the procedure (A) to (G) of the cell sorting operation using the imaging system according to the present embodiment.
- the lens unit 51 of the camera unit 5 has an angle of view capable of simultaneously imaging a plurality of cells C (living subject). That is, the lens unit 51 has an angle of view capable of imaging the plurality of holding recesses 3 of the dish 2 shown in FIG. 2 by one imaging operation.
- the procedure (A) shows an example in which the holding recesses 3 of a 3 ⁇ 3 (m1 to m3 ⁇ n1 to n3) matrix are imaged by one imaging operation.
- one relatively small cell C is held in the holding recess 3 of m1n2 and m1n3
- one relatively large cell C is held in the holding recess 3 of m2n3, and is held in the holding recess 3 of m2n2, m3n1, and m3n3.
- the characteristic amount of the cell C held in each holding recess 3 is extracted by analyzing the image acquired in the procedure (A).
- the feature amount is, for example, the amount of cells C obtained from the number, area, estimated volume, etc. of the cells C, the color and pattern of the cells C, the light intensity when the cells C are made fluorescent. .
- the analysis results of the cells C in each holding recess 3 are each digitized.
- the analysis result obtained in procedure (B) is compared with the parameters that define the range of cells C to be selected.
- This parameter is a characteristic amount of a cell C that is preferable as a selection target obtained from the analysis result of the past cell C based on the images obtained by the current imaging and the selection result (selection operation up to the first time point). It is a parameter indicating a range, that is, a learning value of a selection criterion in the current state.
- the parameters set based on the reference feature amount predetermined as the cell C preferable as the selection target and the procedure Comparison with the analysis result obtained in B) is performed.
- a recommended cell C (recommended biological object) that prompts the operator to select as a movement target is specified based on the comparison result of the procedure (C).
- the recommended cell C is identified based on whether or not the feature amount of each cell C included in the image obtained from the current imaging (the imaging at the second time point after the first time point) belongs to the above parameter range. Is done.
- the recommended cell C is displayed on the display unit such as a monitor for the operator.
- the illustrated procedure (D) an example is shown in which the cells C (recommended cells C) carried in the holding recesses 3 of m1n2 and m1n3 are recommended to the operator.
- a display for recognizing that the operator is prompted to make a selection (such as adding a symbol or the like that surrounds the cells C with a high-intensity marker) is added.
- the recommended cell C is approved by the operator's manual operation.
- the approval work whether or not the recommended cell C presented from the imaging system side in the procedure (D) may be handled as the cell C to be actually selected as a movement target in light of the operator's own experience. Is an operation for the operator to determine
- m1n2 and m1n3 cells C are recommended cells C. The operator performs an approval operation when it is determined that these are cells to be selected as a movement target, and performs an unapproved operation when it is determined that they should not be selected.
- step (E) the imaging system side determines whether there is a cell C that should not be treated as the recommended cell C in step (D). Specifically, the operator determines whether there is a cell C to be moved among the m2n2, m2n3, m3n1, and m3n3 cells that are not recommended cells C, and such a cell C exists. Performs an additional operation to select this.
- the illustrated procedure (E) shows an example in which the recommended cells C of m1n2 and m1n3 are approved and the cell C of m2n3 is additionally selected (additional approval).
- parameters (learning values) that define the range of cells C to be selected are updated based on the results of the operator's approval and additional selection.
- the recommended cell C presented is approved, but the m2n3 cell C is additionally selected. That is, the recommended cell C presented by the imaging system does not match the cell C actually selected by the operator.
- parameters (learning) are set so that these cells C fall within the category of “cells to be selected”. Value) is modified. Then, when the operations of the above procedures (A) to (E) are performed again for each cell C included in the image obtained from the next imaging (imaging at the third time point after the second time point), Based on the corrected parameter, the recommended cell C is specified.
- a predetermined operation is performed on the cell C approved and additionally approved by the operator (procedure (Ga)).
- a predetermined operation is performed on the cells C that are not subject to approval (procedure (Gb)).
- a typical example of the predetermined operation is picking up and moving the cell C.
- m1n2, m1n3 and m2n3 cells C are picked up.
- a pick-up operation for cells C existing in the holding recesses 3 other than m1n2, m1n3, and m2n3, or an additional operation for cells C is performed.
- the cells are imaged in the state of being accommodated in the holding recesses 3 (plate sections) of the dish 2, the image obtained by the imaging is analyzed, and the cells C are measured in units of individual holding recesses 3. Feature quantities are extracted.
- the feature amount that is easy to be displayed in such an image is the shape of the cell C or the amount of the cell C accommodated in one holding recess 3.
- the shape of the cell C can be evaluated by, for example, the roundness of the cell C as shown in FIG. That is, the shape of the cell C can be evaluated by quantifying the degree to which the shape of the outer contour of the cell C identified from the two-dimensional image approximates the perfect circle R.
- the amount of cells C is estimated from, for example, the number of cells C accommodated in one holding recess 3, the area of the cells C on the two-dimensional image, and the outline of the cells C on the two-dimensional image. It is possible to evaluate from the estimated volume.
- FIG. 4B shows an example in which the shape (roundness) of the cell C and the area of the cell C are extracted as feature quantities, and the recommended cell C is specified from the combination of both.
- the roundness of the cell C often affects the soundness of the cell C.
- the area of the cell C is an index that directly indicates the size of the cell C. Therefore, the cell C1 whose roundness and area of the cell C are within a certain range (a triangular region indicated by the symbol a in the figure; this is determined by the learning value) is recommended (OK), and the cell other than the region a
- the recommended cell C can be specified by performing the process of making C2 non-recommended (NG).
- the estimated volume of the cell C is also an index indicating the size of the cell C.
- the estimated volume was obtained by estimating the three-dimensional shape from the outline of the cell C displayed in one two-dimensional image, and was obtained by changing the focal position with respect to the cell C in the Z direction. What was calculated
- the feature amount of the cell C can be a feature amount based on the color of the cell C.
- the color of the cell C often reflects the health and life / death of the cell C.
- FIG. 5 is a diagram for explaining a selection criterion based on the hue of the cell C.
- the cell C may be evaluated by the hue.
- the average luminance of each cell C is extracted as a feature amount from the captured image, and the recommended average cell C between the samples a and b (which is determined by the learning value) is used as a selection threshold. Can be specified.
- the feature amount of the cell C can be a feature amount based on the pattern of the cell C.
- the pattern of the cell C appearing on the captured image may represent the type of the cell C.
- FIG. 6 shows photographs of different types of cancer cells-1, cancer cells-2, and cancer cells-3. Cancer cell-1 indicates colon cancer, cancer cell-2 indicates endometrial cancer, and cancer cell-3 indicates lung cancer. Thus, the pattern of the cell C observed on the image differs depending on the cancer cell type. Therefore, these patterns can be quantified as feature amounts by linearly differentiating and quantifying these patterns. Then, the recommended cell C is identified by setting an appropriate parameter (which is determined by the learning value) for distinguishing the feature amount based on the pattern of the cell C to be selected from the feature amount based on the pattern of the other cell C. can do.
- an appropriate parameter which is determined by the learning value
- the feature amount of the cell C can be a feature amount based on the light intensity of the region of the cell C on the image.
- the intensity of the light emitted from the cell C, particularly the light emitted from the cell C when it is made fluorescent by reacting with an appropriate reagent may be an evaluation standard for the soundness of the cell C. Therefore, as illustrated in FIG. 7A, the difference between the average luminance ⁇ (light intensity) of the peripheral area AC (generally a medium) of the cell C on the image and the average luminance ⁇ of the cell C is calculated. It can be used as a digitized feature value. Then, the recommended cell C can be identified using an appropriate average brightness between the average brightness ⁇ and ⁇ (which is determined by the learning value) as a selection threshold.
- the cell C and its peripheral area AC are divided into a plurality of blocks BL, and the average luminance is obtained for each block BL, and the average luminance varies significantly. Therefore, it can be determined whether or not the cells C satisfying a predetermined average luminance are present in the image.
- FIG. 8 is a block diagram of the imaging system according to the embodiment of the present invention.
- the imaging system includes a camera unit 5 that captures an imaging object on the imaging optical axis, a control unit 7 that controls the operation of the lens unit 51 and performs predetermined processing based on image information acquired by the camera body 52, A lens drive motor 53 that moves the lens unit 51 up and down, a display unit 74, and an input unit 75 are provided.
- the dish 2 accommodated in the sorting container 11 is shown as the imaging object.
- the control unit 7 also controls the pick-up operation and the movement operation of the cell C by controlling the head motor 63 and the shaft motor 64 in the head unit 6.
- the lens drive motor 53 moves in the vertical direction with a predetermined resolution through a power transmission mechanism (not shown) by rotating in the forward or reverse direction. By this movement, the focal position of the lens unit 51 is adjusted to the cell C carried on the dish 2. As shown by a dotted line in FIG. 8, the sorting container 11 itself or a stage (base 1 on which the sorting container 11 is placed) is not a lens unit 51 but a motor other than the lens driving motor 53. ) May be moved up and down.
- the head motor 63 is a motor serving as a driving source for the operation of raising and lowering the head 61 with respect to the head main body 62 and the operation of generating a suction force and a discharge force at the tip opening t of the chip 12 attached to the head 6.
- the shaft motor 64 is a motor serving as a drive source for moving the head unit 6 (head main body 62) along the guide rail 6G (FIG. 1).
- the control unit 7 is composed of a personal computer, for example, and operates to include a drive control unit 71, an image processing unit 72, and a calculation unit 73 functionally by executing a predetermined program.
- the drive control unit 71 controls the operation of the lens drive motor 53, the head motor 63, and the shaft motor 64. Specifically, the drive control unit 71 gives a control pulse for moving the lens unit 51 in the vertical direction at a predetermined pitch (for example, several tens of ⁇ m pitch) to the lens drive motor 53 for the focusing operation. Although not shown in FIG. 8, the drive control unit 71 also controls the operation of the camera shaft drive motor that moves the camera unit 5 along the guide rail 5G. Further, the drive control unit 71 also controls the mechanical operation of the head unit 6. The drive control unit 71 controls the head motor 63 to control the raising / lowering operation of the head 61 and the operation of generating the suction force or the ejection force at the tip opening t of the chip 12.
- a predetermined pitch for example, several tens of ⁇ m pitch
- the image processing unit 72 performs image processing such as edge detection processing and pattern recognition processing with feature amount extraction on the image data acquired by the camera body 52.
- the image processing unit 72 acquires image data of the dish 2 carrying the cells C, and recognizes the cells C present on the dish 2 (holding recesses 3) by the image processing.
- the calculation unit 73 mainly performs various types of analysis on the cells C on the image specified by the image processing unit 72, and executes processing for specifying the recommended cells C to be presented to the operator.
- the calculation unit 73 functionally includes an analysis unit 731, a specification unit 732, and a mode change unit 733.
- the analyzing unit 731 performs a process of extracting a feature amount of the cell C by analyzing the image of the cell C.
- the feature quantity to be extracted is the shape, number, area, estimated volume, color, pattern, light intensity, etc. of the cell C as described above with reference to FIGS.
- the analysis unit 731 digitizes these feature values.
- the specifying unit 732 analyzes the parameters derived based on the operator's past cell selection results and the cells C included in the image acquired by the current imaging. By comparing with the feature amount extracted by 731, processing for specifying the recommended cell C to be presented to the operator is performed.
- the specifying unit 732 sets a parameter that defines the range of the cells C to be selected based on the feature amount extracted by the analyzing unit 731 for the cells C selected by the operator up to a specific first time point. This parameter is the learning value of the selection criterion for the cell C at the first time point.
- the specifying unit 732 specifies the recommended cell C that prompts the operator to select based on the parameter in the image including the cell C obtained by imaging at the second time point after the first time point. That is, among the cells C included in the image acquired at the second time point, the cell C having the characteristic amount belonging to the parameter range is specified as the recommended cell C.
- the specifying unit 732 selects the recommended cell C previously specified in the selection operation of the cell C in the image obtained by the imaging at the second time point.
- the parameter is corrected based on the feature amount of the selected cell C. That is, the learning value of the selection criterion for the cell C at the first time point is updated.
- the specifying unit 732 executes processing for specifying the recommended cell C based on the corrected parameter in the image of the cell C obtained by imaging at the third time point after the second time point.
- the mode changing unit 733 performs processing for changing the operation mode between a manual operation mode for accepting an operator's approval operation for the recommended cell C and an automatic approval mode for automatically approving the recommended cell C.
- the parameter (learned value) is sequentially updated according to the selection result of the operator as described above.
- the automatic approval mode treats the recommended cell C specified by the specifying unit 732 as if it had received the selection operation without performing such an update anymore and without receiving a selection operation from the operator. is there.
- the mode change unit 733 changes the operation mode from the manual operation mode to the automatic approval mode. Thereby, the selection operation
- the display unit 74 is a display that displays an image captured by the camera unit 5.
- an image of the dish 2 captured by the camera unit 5, that is, an image including a plurality of cells C carried by the plurality of holding recesses 3 is displayed. Further, a predetermined display that can be easily identified by the operator is added to the recommended cells C.
- the input unit 75 includes a keyboard, a mouse, and the like, and receives input related to a selection (approval) operation on the cell C displayed on the display unit 74 from the operator.
- the approval operation and the additional approval operation by the operator described in the procedure (E) of FIG. 3 are accepted from the input unit 75.
- the dish 2 is imaged by the camera unit 5.
- the cell suspension is dispensed in advance in the dish 2, and the camera unit 5 images the state in which the cells C are accommodated in the holding recesses 3 of the dish 2 (step S1 / image at the first time point).
- the drive control unit 71 drives the lens unit 51 and the focusing operation for the cell C on the dish 2 is executed.
- the captured image is displayed on the display unit 74 in a state as shown in, for example, the procedure (A) of FIG.
- step S2 an input of an initial selection operation as to which of the cells C displayed on the display unit 74 is to be moved is received from the operator (step S2).
- the operator there is no parameter (learning value) that defines the range of cells C to be selected.
- the recommended cell C is identified based on a predetermined set value, and the display unit 74 displays the recommended cell C so that the recommended cell C is visible.
- the image processing unit 72 acquires image data of the dish 2 carrying the cells C from the camera body 52, and executes image processing for specifying the cells C included in the image.
- This image processing data is sent to the analysis unit 731 of the calculation unit 73, and the analysis unit 731 has the shape, number, area, estimated volume, and color of the cell C and other cells C selected by the operator.
- a process for obtaining feature quantities such as a pattern and light intensity is executed (step S3).
- specification part 732 refers to the feature-value about the cell C which the operator selected among the feature-value of each cell C calculated
- the selection criterion of the cell C is learned (step S4).
- This parameter is, for example, a lower limit area threshold value and an upper limit area threshold value of the cell C that are preferable as a movement target when an area is selected as the feature amount.
- the lower limit average luminance threshold and the upper limit average luminance threshold are used.
- Such parameters are stored in a memory area (not shown) provided in the control unit 7 (step S5).
- step S6 whether or not the operation mode has been changed from the manual operation mode to the automatic approval mode is confirmed by the mode changing unit 733 (step S6). If learning of the parameters is insufficient, the default manual operation mode is maintained (YES in step S6). On the other hand, when the learning of the parameter is sufficiently advanced, for example, when the parameter stored in step S5 has reached a state in which there is almost no change in the most recent update, the operation mode is changed from the manual operation mode. The automatic approval mode is changed (NO in step S6).
- step S6 When the manual operation mode is maintained (YES in step S6), the cell C carried on the dish 2 is imaged again by the camera unit 5 (step S7 / second time imaging).
- the image data obtained by this imaging is subjected to the cell C specifying process by the image processing unit 72 and the feature amount extracting process of each cell C by the analyzing unit 731.
- the specifying unit 732 specifies recommended cells C that prompt the operator to select among the cells C included in the current captured image with reference to the current parameter (learned value).
- the recommended cell C is displayed on the display part 74 in the aspect which can be visually recognized by the operator (step S8).
- step S6 when the mode change unit 733 has changed the operation mode to the automatic approval mode (NO in step S6), the operator's selection operation is omitted thereafter.
- the cell C carried on the dish 2 is imaged by the camera unit 5 (step S10).
- the specifying unit 732 performs a process of specifying the recommended cell C from the cells C included in the obtained image (step S11).
- the recommended cell C identified in step S11 is automatically selected as the cell C to be moved (step S12).
- step S13 Thereafter, it is confirmed whether or not the imaging of the cell C is continued (step S13).
- imaging YES in step S13
- the process returns to step S10, and the next imaging operation for the dish 2 is executed by the camera unit 5.
- the process ends.
- the recommended cell C is presented to the operator on the display unit 74 in the cell C selection operation performed at or after the second time point. . Accordingly, the operator can quickly select the cell C while referring to the recommended cell C.
- the recommended cell C since the recommended cell C is specified based on parameters based on the selection results of the operator up to the first time point, the recommended cell C has high accuracy. Therefore, the operator can select the cell C accurately and can reduce the labor of the operator's selection work of the cell C.
- the specifying unit 732 extracts a feature amount again (learns the feature amount), and the feature Based on the amount, the parameter serving as a selection criterion for the cell C is corrected. Then, at the third time point, the recommended cell C is specified based on the corrected parameter. Accordingly, the accuracy of specifying the recommended cell C can be further improved in the selection operation at and after the third time point. In other words, in parallel with the selection operation by the operator, it is possible to gradually improve the identification accuracy of the recommended cell C based on the learning effect of the feature amount.
- the specifying unit 732 learns the feature amount based on the operator's selection result, when there are a plurality of operators, the selection result of each operator It is possible to cause the specifying unit 732 to learn feature quantities individually according to the above.
- the operation is received in association with the identification ID of the operator, and learning data is generated for each operator and stored in the memory area of the calculation unit 73. deep. This makes it possible to specify individual recommended cells C based on the selection tendency of each operator, instead of the average recommended cells C for all operators.
- the specifying unit 732 is made to learn the feature amount for each type or individual of the biological object, and the specifying unit 732 specifies the recommended cell C for each type or for each individual.
- the cell C is selected in advance before the selection operation of the cell C from the input unit 75. Enter the entered cancer cell classification. And according to the operator's selection result, learning data are produced
- the recommended cell C may differ depending on the individual.
- the identification ID or the like of the individual who has collected the cell C is input.
- learning data is generated for each individual according to the selection result of the operator and stored in the memory area of the calculation unit 73.
- the specifying unit 732 can specify the recommended cell C corresponding to each individual.
- An imaging system includes an imaging device capable of simultaneously imaging a plurality of biological objects, a display unit that displays an image including the plurality of biological objects captured by the imaging device, and an operator
- an input unit that accepts an input related to a selection operation for the plurality of biological objects displayed on the display unit, and an image corresponding to the biological object to which the selection operation is given by the input unit
- An analysis unit that extracts a feature amount of an object and a biological object that has been subjected to the selection operation up to a first time point, based on a feature amount extracted by the analysis unit or a predetermined reference feature amount
- a parameter that defines a range of the biological object to be processed is set. Characterized in that it and a specifying unit configured to specify a recommended biological object to prompt the operator to make selections based on.
- a recommended biological object is presented to the operator in the selection operation at the input unit executed at or after the second time point.
- the operator can quickly select the biological object while referring to the recommended biological object.
- the recommended biological object is specified based on a parameter based on the selection result or the reference feature amount of the operator up to the first time point, it has high accuracy. Therefore, according to the above imaging system, it is possible to accurately select a living body object and reduce the labor of selecting the living body object.
- the input unit does not match the recommended biological object specified by the specifying unit in the selection operation for the plurality of biological objects in the image obtained by imaging at the second time point.
- the analysis unit analyzes an image corresponding to the biological object received by the input unit, and extracts a feature amount of the biological object at the second time point.
- the specifying unit corrects the parameter based on the feature amount at the second time point, and corrects the image including the plurality of biological objects obtained by imaging at the third time point after the second time point. It is preferable that the recommended biological object is specified based on the parameter that has been set.
- the feature value is extracted again (learning the feature value).
- the parameter is corrected based on the feature amount.
- the recommended biological object is specified based on the corrected parameter. Therefore, in the selection operation at the third time point and thereafter, the accuracy of specifying the recommended biological object can be further improved. In other words, in parallel with the selection operation by the operator, it is possible to gradually improve the identification accuracy of the recommended biological object based on the learning effect of the feature amount.
- a mode change that changes an operation mode so that the recommended biological object specified by the specifying unit is handled as a biological object that has received the selection operation by the input unit without receiving the selection operation. It is desirable to further include a section.
- the selection operation by the operator can be omitted when the mode changing unit changes the operation mode.
- the accuracy of specifying the recommended biological object increases so that the operator no longer needs the intervention of the selection operation.
- the selection operation of the living body object can be completely automated by causing the mode changing unit to change the operation mode.
- the feature amount can be a feature amount based on the shape of the biological object.
- the imaging system may further include a plate including a plurality of compartments that house the living body object, and the imaging device captures a state where the living body object is housed in the section of the plate.
- the feature amount can be a feature amount based on the amount of the biological object accommodated in one of the sections.
- the amount of the biological object can be an amount obtained from at least one of the estimated volumes estimated from the number, area, and contour of the biological object.
- the feature quantity may be a feature quantity based on the color of the biological object, and the feature quantity may be a feature quantity based on the pattern of the biological object.
- the feature amount can be a feature amount based on the light intensity of the region of the biological object on the image.
- the said feature-value is good also as a feature-value based on the light intensity of the area
- the imaging system may include a plate having a plurality of compartments for storing the biological object, and a chip for picking up the biological object from the compartments of the plate, and the picked-up biological object can be moved And a control unit that controls the operation of the pickup by the chip, and the imaging device captures an image of a state in which the biological object is accommodated in a section of the plate, It is preferable that the control unit performs control for causing the chip to pick up the biological object to which the selection operation is given by the input unit or the biological object to which the selection operation is not given.
- the imaging system can be provided with a function of picking up a biological object to which the selection operation is given or not given and moving it to a desired position.
- the specifying unit individually selects the recommended object for each operator based on a learning result of the feature amount according to a selection result of each operator. It is desirable to specify. As a result, it is possible to identify individual recommended objects based on the selection tendency of each operator, not the average recommended objects for all operators.
- the specifying unit is configured for each biological object based on the learning result of the feature amount according to the selection result of the operator for each biological object.
- the recommended object is specified individually. As a result, it is possible to specify the recommended target according to the type of the biological object, not the average recommended object for all the biological objects.
- the specifying unit learns the feature amount according to the selection result of the operator for the biological object of each individual. Based on the above, it is desirable to specify the recommended object individually for each individual. Thereby, even if it is the same kind of living body object, it becomes possible to specify the recommendation object according to each individual.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Chemical & Material Sciences (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Biomedical Technology (AREA)
- Multimedia (AREA)
- Databases & Information Systems (AREA)
- Molecular Biology (AREA)
- Biochemistry (AREA)
- Analytical Chemistry (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Wood Science & Technology (AREA)
- Medical Informatics (AREA)
- Organic Chemistry (AREA)
- Zoology (AREA)
- Signal Processing (AREA)
- Pathology (AREA)
- Immunology (AREA)
- Genetics & Genomics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Sustainable Development (AREA)
- Biotechnology (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Quality & Reliability (AREA)
- Evolutionary Computation (AREA)
- Dispersion Chemistry (AREA)
- Radiology & Medical Imaging (AREA)
- Microbiology (AREA)
- Software Systems (AREA)
- Food Science & Technology (AREA)
- Medicinal Chemistry (AREA)
- Urology & Nephrology (AREA)
- Hematology (AREA)
Abstract
Description
図1は、本発明の実施形態に係る撮像システムが適用される細胞移動装置Sの全体構成を概略的に示す図である。ここでは、細胞Cを2つの容器(ディッシュ2とマイクロプレート4)間で移動させる細胞移動装置Sを例示している。
ディッシュ2の詳細構造を説明する。図2(A)は、ディッシュ2の上面図、図2(B)は、図2(A)のIIB-IIB線断面図である。ディッシュ2は、ディッシュ本体20と、該ディッシュ本体20に形成される複数の保持凹部3とを備えている。ディッシュ本体20は、所定の厚みを有する平板状の部材からなり、上面21と下面22とを有する。保持凹部3は、上面21の側に細胞Cの受け入れ開口(開口部31)を有する。ディッシュ2は、選別容器11内の培地L中に浸漬される。詳しくは、ディッシュ本体20の上面21が選別容器11内の培地L中に浸漬される一方、下面22が選別容器11の底板に対して間隔を置いた状態で、選別容器11内で保持される(図1参照)。
続いて、本実施形態の撮像システムを用いた、細胞Cの選別作業の全体的な流れを説明する。図3は、本実施形態に係る撮像システムを用いた細胞選別作業の手順(A)~(G)を模式的に示す図である。最初の手順(A)では、カメラユニット5によりディッシュ2が撮像されることによって、当該ディッシュ2に担持された細胞Cの画像が取得される。カメラユニット5のレンズ部51は、複数の細胞C(生体対象物)を同時に撮像可能な画角を備えている。すなわち、レンズ部51は、1回の撮像動作によって、図2に示すディッシュ2の複数の保持凹部3を撮像できる画角を有している。
続いて、細胞Cの特徴量を画像から抽出する具体例について説明する。本実施形態では、ディッシュ2の保持凹部3(プレートの区画)に収容された状態で細胞が撮像され、その撮像で得られた画像を解析して、個々の保持凹部3の単位で細胞Cの特徴量が抽出される。このような画像に表出し易い特徴量は、細胞Cの形状や、一つの保持凹部3に収容されている細胞Cの量である。
図8は、本発明の実施形態に係る撮像システムのブロック図である。撮像システムは、撮像光軸上の撮像対象物を撮像するカメラユニット5と、レンズ部51の動作を制御すると共にカメラ本体52により取得された画像情報に基づき所定の処理を行う制御部7と、レンズ部51を上下動させるレンズ駆動モータ53と、表示部74と、入力部75とを備えている。図8では、撮像対象物として、選別容器11に収容されたディッシュ2を示している。なお、制御部7は、ヘッドユニット6内のヘッドモータ63と、軸モータ64とを制御することによって、細胞Cのピックアップ動作及び移動動作も制御する。
続いて、図8に示す本実施形態の撮像システムを用いた細胞選別動作を、図9に示すフローチャートに基づいて説明する。処理が開始されると、カメラユニット5によりディッシュ2が撮像される。ディッシュ2には予め細胞懸濁液が分注されており、カメラユニット5はディッシュ2の保持凹部3に細胞Cが収容された状態を撮像する(ステップS1/第1時点の撮像)。この際、駆動制御部71がレンズ部51を駆動して、ディッシュ2上の細胞Cに対する合焦動作が実行される。撮像された画像は、例えば図3の手順(A)に示すような状態で、表示部74に表示される。
以上説明した本実施形態に係る撮像システムによれば、前記第2時点またはそれ以降に実行される細胞Cの選択操作において、操作者には表示部74で推薦細胞Cが提示されるようになる。これにより、操作者は推薦細胞Cを参照しながら、迅速に細胞Cの選択を行うことができる。しかも、この推薦細胞Cは、前記第1時点までの当該操作者の選択実績を踏まえたパラメータに基づいて特定されたものであるので、的確性の高いものとなる。従って、操作者は、的確な細胞Cの選択が行えると共に、操作者の細胞Cの選択作業の労力を軽減することができる。
なお、上述した具体的実施形態には以下の構成を有する発明が主に含まれている。
Claims (14)
- 複数の生体対象物を同時に撮像可能な撮像装置と、
前記撮像装置により撮像された前記複数の生体対象物を含む画像を表示する表示部と、
操作者より、前記表示部に表示された前記複数の生体対象物に対する選択操作に関する入力を受け付ける入力部と、
前記入力部で選択操作が与えられた生体対象物に対応する画像を解析して、当該生体対象物についての特徴量を抽出する解析部と、
第1時点までに前記選択操作が与えられた生体対象物について前記解析部が抽出した特徴量若しくは予め定められた基準特徴量に基づいて、選択されるべき生体対象物の範囲を定めるパラメータを設定し、前記第1時点よりも後の第2時点の撮像で得られる前記複数の生体対象物を含む画像において、前記パラメータに基づいて操作者に選択を促す推薦生体対象物を特定する特定部と、
を備えることを特徴とする撮像システム。 - 請求項1に記載の撮像システムにおいて、
前記入力部が、前記第2時点の撮像で得られた画像における前記複数の生体対象物に対する選択操作において、前記特定部が特定した前記推薦生体対象物と一致しない生体対象物の選択操作の入力を受け付けた場合、
前記解析部は、前記入力部が受け付けた生体対象物に対応する画像を解析して、前記第2時点における当該生体対象物についての特徴量を抽出し、
前記特定部は、前記第2時点における特徴量に基づいて前記パラメータを修正し、前記第2時点よりも後の第3時点の撮像で得られる前記複数の生体対象物を含む画像において、修正された前記パラメータに基づいて前記推薦生体対象物を特定する、撮像システム。 - 請求項2に記載の撮像システムにおいて、
前記特定部が特定した前記推薦生体対象物を、前記選択操作を受け付けることなく、前記入力部で前記選択操作を受け付けた生体対象物と取り扱うよう動作モードを変更するモード変更部をさらに備える、撮像システム。 - 請求項1~3のいずれか1項に記載の撮像システムにおいて、
前記特徴量は、前記生体対象物の形状に基づく特徴量である、撮像システム。 - 請求項1~3のいずれか1項に記載の撮像システムにおいて、
前記生体対象物を収容する区画を複数備えるプレートを備え、
前記撮像装置は、前記生体対象物が前記プレートの区画に収容された状態を撮像するものであって、
前記特徴量は、一つの前記区画に収容されている前記生体対象物の量に基づく特徴量である、撮像システム。 - 請求項5に記載の撮像システムにおいて、
前記生体対象物の量は、当該生体対象物の個数、面積、輪郭から推定される推定体積のうちの少なくとも一つから求められる量である、撮像システム。 - 請求項1~3のいずれか1項に記載の撮像システムにおいて、
前記特徴量は、前記生体対象物の色に基づく特徴量である、撮像システム。 - 請求項1~3のいずれか1項に記載の撮像システムにおいて、
前記特徴量は、前記生体対象物の模様に基づく特徴量である、撮像システム。 - 請求項1~3のいずれか1項に記載の撮像システムにおいて、
前記特徴量は、画像上における前記生体対象物の領域の光強度に基づく特徴量である、撮像システム。 - 請求項9に記載の撮像システムにおいて、
前記特徴量は、画像上における前記生体対象物の領域及びその周辺領域の光強度に基づく特徴量である、撮像システム。 - 請求項1~10のいずれか1項に記載の撮像システムにおいて、
前記生体対象物を収容する区画を複数備えるプレートと、
前記プレートの区画から前記生体対象物をピックアップするチップを有し、ピックアップされた前記生体対象物を移動させることが可能なヘッドと、
前記チップによる前記ピックアップの動作を制御する制御部と、をさらに備え、
前記撮像装置は、前記生体対象物が前記プレートの区画に収容された状態を撮像するものであって、
前記制御部は、前記入力部で選択操作が与えられた生体対象物、若しくは、前記選択操作が与えられなかった生体対象物を、前記チップにピックアップさせる制御を実行する、撮像システム。 - 請求項1~11のいずれか1項に記載の撮像システムにおいて、
前記特定部は、操作者が複数人存在する場合に、各操作者の選択結果に応じた前記特徴量の学習結果に基づいて、操作者毎に個別に前記推薦対象物を特定する、撮像システム。 - 請求項1~11のいずれか1項に記載の撮像システムにおいて、
前記特定部は、生体対象物の種別が複数存在する場合に、各生体対象物に対する操作者の選択結果に応じた前記特徴量の学習結果に基づいて、生体対象物毎に個別に前記推薦対象物を特定する、撮像システム。 - 請求項1~11のいずれか1項に記載の撮像システムにおいて、
前記特定部は、複数の個体から採取された同種の生体対象物が存在する場合に、各個体の生体対象物に対する操作者の選択結果に応じた前記特徴量の学習結果に基づいて、個体毎に個別に前記推薦対象物を特定する、撮像システム。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/965,237 US11367294B2 (en) | 2018-01-31 | 2018-12-04 | Image capture system |
CN201880087823.9A CN111630149A (zh) | 2018-01-31 | 2018-12-04 | 摄像系统 |
JP2019568898A JP7018078B2 (ja) | 2018-01-31 | 2018-12-04 | 撮像システム |
EP18903403.6A EP3733832A4 (en) | 2018-01-31 | 2018-12-04 | IMAGING SYSTEM |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2018015169 | 2018-01-31 | ||
JP2018-015169 | 2018-01-31 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019150755A1 true WO2019150755A1 (ja) | 2019-08-08 |
Family
ID=67478975
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2018/044628 WO2019150755A1 (ja) | 2018-01-31 | 2018-12-04 | 撮像システム |
Country Status (5)
Country | Link |
---|---|
US (1) | US11367294B2 (ja) |
EP (1) | EP3733832A4 (ja) |
JP (1) | JP7018078B2 (ja) |
CN (1) | CN111630149A (ja) |
WO (1) | WO2019150755A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP7237391B1 (ja) | 2022-01-24 | 2023-03-13 | 大和酸素工業株式会社 | 線虫自動移行装置 |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3739036B1 (en) * | 2018-02-15 | 2023-02-08 | Yamaha Hatsudoki Kabushiki Kaisha | Biological subject transfer device |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2011021391A1 (ja) * | 2009-08-19 | 2011-02-24 | 国立大学法人名古屋大学 | 培養細胞評価装置、インキュベータおよびプログラム |
WO2015087371A1 (ja) | 2013-12-12 | 2015-06-18 | ヤマハ発動機株式会社 | 対象物の移動装置 |
JP2015223175A (ja) * | 2014-05-30 | 2015-12-14 | 富士フイルム株式会社 | 細胞評価装置および方法並びにプログラム |
JP2016154450A (ja) * | 2015-02-23 | 2016-09-01 | 大日本印刷株式会社 | 培養液の培養状態解析システム及び培養状態解析方法、並びに、プログラム |
WO2017110005A1 (ja) * | 2015-12-25 | 2017-06-29 | ヤマハ発動機株式会社 | 対象物のピックアップ方法 |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030215936A1 (en) * | 2000-12-13 | 2003-11-20 | Olli Kallioniemi | High-throughput tissue microarray technology and applications |
US20040085443A1 (en) * | 2000-12-13 | 2004-05-06 | Kallioniemi Olli P | Method and system for processing regions of interest for objects comprising biological material |
US7589330B2 (en) * | 2007-06-25 | 2009-09-15 | Olympus Corporation | Ultraviolet microscope apparatus |
EP2270198A4 (en) * | 2008-03-24 | 2013-03-27 | Nikon Corp | PICTURE ANALYSIS METHOD FOR CELL OBSERVATION, PICTURE PROCESSING PROGRAM AND PICTURE PROCESSING DEVICE |
CN102471744B (zh) * | 2009-07-21 | 2015-06-10 | 国立大学法人京都大学 | 图像处理装置、培养观察装置及图像处理方法 |
JP2017009314A (ja) | 2015-06-17 | 2017-01-12 | 株式会社Screenホールディングス | 教示データの作成支援方法、作成支援装置、プログラムおよびプログラム記録媒体 |
US10845186B2 (en) * | 2016-03-09 | 2020-11-24 | Sony Corporation | Information processing device, information processing method, and information processing system |
-
2018
- 2018-12-04 CN CN201880087823.9A patent/CN111630149A/zh active Pending
- 2018-12-04 JP JP2019568898A patent/JP7018078B2/ja active Active
- 2018-12-04 WO PCT/JP2018/044628 patent/WO2019150755A1/ja active Application Filing
- 2018-12-04 US US16/965,237 patent/US11367294B2/en active Active
- 2018-12-04 EP EP18903403.6A patent/EP3733832A4/en not_active Withdrawn
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2011021391A1 (ja) * | 2009-08-19 | 2011-02-24 | 国立大学法人名古屋大学 | 培養細胞評価装置、インキュベータおよびプログラム |
WO2015087371A1 (ja) | 2013-12-12 | 2015-06-18 | ヤマハ発動機株式会社 | 対象物の移動装置 |
JP2015223175A (ja) * | 2014-05-30 | 2015-12-14 | 富士フイルム株式会社 | 細胞評価装置および方法並びにプログラム |
JP2016154450A (ja) * | 2015-02-23 | 2016-09-01 | 大日本印刷株式会社 | 培養液の培養状態解析システム及び培養状態解析方法、並びに、プログラム |
WO2017110005A1 (ja) * | 2015-12-25 | 2017-06-29 | ヤマハ発動機株式会社 | 対象物のピックアップ方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP3733832A4 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP7237391B1 (ja) | 2022-01-24 | 2023-03-13 | 大和酸素工業株式会社 | 線虫自動移行装置 |
JP2023107338A (ja) * | 2022-01-24 | 2023-08-03 | 大和酸素工業株式会社 | 線虫自動移行装置 |
Also Published As
Publication number | Publication date |
---|---|
EP3733832A1 (en) | 2020-11-04 |
JP7018078B2 (ja) | 2022-02-09 |
EP3733832A4 (en) | 2021-02-24 |
US11367294B2 (en) | 2022-06-21 |
US20200372240A1 (en) | 2020-11-26 |
JPWO2019150755A1 (ja) | 2020-12-10 |
CN111630149A (zh) | 2020-09-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10648897B2 (en) | Method and apparatus for the identification and handling of particles | |
KR101120829B1 (ko) | 자동화된 이미지 처리 프로파일 선택 기능을 가지는생물학적 성장판 스캐너 | |
JP6479676B2 (ja) | 流体サンプル中の粒子の分類のためのシステム及び方法 | |
CN105143850B (zh) | 用于血液样品中的粒子分析的自聚焦系统和方法 | |
EP3375859B1 (en) | Method for constructing classifier, and method for determining life or death of cells using same | |
US20030179916A1 (en) | High-throughput cell identification and isolation method and apparatus | |
US20100135566A1 (en) | Analysis and classification, in particular of biological or biochemical objects, on the basis of time-lapse images, applicable in cytometric time-lapse cell analysis in image-based cytometry | |
WO2019150755A1 (ja) | 撮像システム | |
JP6853880B2 (ja) | 細胞移動方法及び細胞移動装置 | |
JP6947905B2 (ja) | 撮像システム及び生体対象物移動装置 | |
EP3730601A1 (en) | Movement method and movement device for biological subject | |
JP6946868B2 (ja) | 細胞吸引支援装置および細胞吸引支援装置の制御方法 | |
JP6710772B2 (ja) | 細胞移動装置及び細胞移動方法 | |
WO2022145086A1 (ja) | 細胞移動装置 | |
US12026880B2 (en) | Fluorescence image display method and fluorescence image analyzer | |
US20220207725A1 (en) | Fluorescence image display method and fluorescence image analyzer | |
KR20230042059A (ko) | 이벤트 기반 이미징을 위한 방법 및 시스템 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18903403 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2019568898 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 18903403.6 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2018903403 Country of ref document: EP Effective date: 20200730 |