US20180275061A1 - Microscope apparatus - Google Patents

Microscope apparatus Download PDF

Info

Publication number
US20180275061A1
US20180275061A1 US15/991,426 US201815991426A US2018275061A1 US 20180275061 A1 US20180275061 A1 US 20180275061A1 US 201815991426 A US201815991426 A US 201815991426A US 2018275061 A1 US2018275061 A1 US 2018275061A1
Authority
US
United States
Prior art keywords
image
capturing
capturing position
optical system
dimensional
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/991,426
Other languages
English (en)
Inventor
Ichiro Sase
Yasutoshi Kaneko
Tatsuo Fukui
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nikon Corp
Original Assignee
Nikon Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nikon Corp filed Critical Nikon Corp
Assigned to NIKON CORPORATION reassignment NIKON CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: FUKUI, TATSUO, KANEKO, YASUTOSHI, SASE, ICHIRO
Publication of US20180275061A1 publication Critical patent/US20180275061A1/en
Priority to US17/178,957 priority Critical patent/US11906431B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/16Microscopes adapted for ultraviolet illumination ; Fluorescence microscopes
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N21/645Specially adapted constructive features of fluorimeters
    • G01N21/6456Spatial resolved fluorescence measurements; Imaging
    • G01N21/6458Fluorescence microscopy
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/0004Microscopes specially adapted for specific applications
    • G02B21/002Scanning microscopes
    • G02B21/0024Confocal scanning microscopes (CSOMs) or confocal "macroscopes"; Accessories which are not restricted to use with CSOMs, e.g. sample holders
    • G02B21/0052Optical details of the image generation
    • G02B21/0076Optical details of the image generation arrangements using fluorescence or luminescence
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/36Microscopes arranged for photographic purposes or projection purposes or digital imaging or video purposes including associated control and data processing arrangements
    • G02B21/365Control or image processing arrangements for digital or video microscopes
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/36Microscopes arranged for photographic purposes or projection purposes or digital imaging or video purposes including associated control and data processing arrangements
    • G02B21/365Control or image processing arrangements for digital or video microscopes
    • G02B21/367Control or image processing arrangements for digital or video microscopes providing an output produced by processing a plurality of individual source images, e.g. image tiling, montage, composite images, depth sectioning, image comparison
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N5/225
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N2201/00Features of devices classified in G01N21/00
    • G01N2201/06Illumination; Optics
    • G01N2201/061Sources
    • G01N2201/06113Coherent sources; lasers

Definitions

  • the present invention relates to a microscope apparatus, an image-capturing method and a program.
  • Patent Document 1 U.S. Patent Application Publication No. 2008/0182336
  • a first aspect of the present invention provides a microscope apparatus including: an illumination optical system that radiates activation light to activate some of fluorescent materials included in a sample and excitation light to excite at least some of the activated fluorescent materials; an image forming optical system that: has an objective lens and an astigmatic optical system that generates astigmatism to at least part of fluorescence from the fluorescent materials; and forms an image of the fluorescence; an image-capturing unit that captures an image formed by the image forming optical system; a drive unit that moves an image-capturing position in the sample along an optical axis-direction of the objective lens; and a control unit, wherein the control unit causes the image-capturing unit to capture images in a plurality of numbers of frames respectively at a first image-capturing position and at a second image-capturing position different from the first image-capturing position.
  • a second aspect of the present invention provides an image-capturing method performed at a microscope apparatus having: an illumination optical system that radiates activation light to activate some of fluorescent materials included in a sample and excitation light to excite at least some of the activated fluorescent materials; an image forming optical system that: has an objective lens and an astigmatic optical system that generates astigmatism to at least part of fluorescence from the fluorescent materials; and forms an image of the fluorescence; an image-capturing unit that captures an image formed by the image forming optical system; and a drive unit that moves an image-capturing position in the sample along an optical axis-direction of the objective lens, the image-capturing method including: causing the image-capturing unit to capture images in a plurality of numbers of frames respectively at a first image-capturing position and at a second image-capturing position different from the first image-capturing position.
  • a third aspect of the present invention provides a computer program product having computer instructions that: are recorded on a computer readable medium; and are executed by a computer that controls a microscope apparatus having:
  • an illumination optical system that radiates activation light to activate some of fluorescent materials included in a sample and excitation light to excite at least some of the activated fluorescent materials
  • an image forming optical system that: has an objective lens and an astigmatic optical system that generates astigmatism to at least part of fluorescence from the fluorescent materials; and forms an image of the fluorescence;
  • a drive unit that moves an image-capturing position in the sample along an optical axis-direction of the objective lens
  • the computer instructions upon being executed by the computer, enabling the computer to perform operations including:
  • FIG. 1 is a figure showing a microscope apparatus 1 according to the present embodiment.
  • FIG. 2 is a figure showing an image forming optical system 5 and an image sensor 40 .
  • FIG. 3 is a figure showing an optical path of fluorescence from and an image of fluorescence from a position Z 2 in a sample W.
  • FIG. 4 is a figure showing an optical path of fluorescence from and an image of fluorescence from a fluorescent material present at a position Z 1 in the sample W.
  • FIG. 5 is a figure showing an optical path of fluorescence from and an image of fluorescence from a fluorescent material present at a position Z 3 in the sample W.
  • FIG. 6 is a figure showing one example of image processing performed by an image processing unit 7 .
  • FIG. 7 shows one example of a data configuration 45 of three-dimensional distribution information about a centroid Q.
  • FIG. 8 is a schematic diagram of a sequence of actions performed by the microscope apparatus 1 .
  • FIG. 9 is a flowchart showing an image-capturing process performed by the microscope apparatus 1 .
  • FIG. 10 shows one example of a screen on which image-capturing conditions in an image-capturing process are input.
  • FIG. 11 is a schematic diagram showing three-dimensional distribution information about the centroid of fluorescence.
  • FIG. 12 shows one example of a screen 150 on which setting of a reconstructed display image of the sample W is performed.
  • FIG. 13 shows a display image CL 1 in which coloring is performed differently layer-by-layer.
  • FIG. 14 is a schematic diagram showing laps among layers in the thickness direction of a sample.
  • FIG. 15 is a schematic diagram for explaining a display image CL 2 to display the centroid included in a lap Lw.
  • FIG. 16 is a schematic diagram of another sequence of actions performed by the microscope apparatus 1 .
  • FIG. 17 shows another example of a screen on which image-capturing conditions in an image-capturing process are input.
  • FIG. 18 is a schematic diagram showing three-dimensional distribution information about the centroid of fluorescence.
  • FIG. 19 shows a display image CL 3 in which coloring is performed differently layer-by-layer.
  • FIG. 20 shows a display image CL 4 in which coloring is performed differently passing procedure-by-passing procedure.
  • a microscope apparatus is, for example, a microscope apparatus utilizing single-molecule localization microscopy such as STORM or PALM.
  • the microscope apparatus according to the present embodiment can generate a three-dimensional super-resolution image.
  • the microscope apparatus according to an embodiment can be utilized for both: fluorescence observation of a sample labelled with one type of fluorescent material; and fluorescence observation of a sample labelled with two or more types of fluorescent material.
  • the sample may be one including live cells, may be one including cells fixed by using a tissue fixing solution such as a formaldehyde solution or may be tissues or the like.
  • the fluorescent material may be a fluorescent pigment such as a cyanine dye or may be a fluorescent protein.
  • the fluorescent pigment includes a reporter pigment that emits fluorescence upon reception of excitation light in its activated state where it is activated.
  • the fluorescent pigment may in some cases include an activator pigment that bring the reporter pigment into an activated state upon reception of activation light. If the fluorescent pigment does not include an activator pigment, the reporter pigment enters an activated state upon reception of activation light.
  • the fluorescent pigment is, for example: a dye pair obtained through bonding between two types of cyanine dye (examples: a Cy3-Cy5 dye pair (Cy3 and Cy5 are registered trademarks), a Cy2-Cy5 dye pair (Cy2 and Cy5 are registered trademarks), a Cy3-Alexa Fluor647 dye pair (Cy3 and Alexa Fluor are registered trademarks)); or one type of dye (example: Alexa Fluor647 (Alexa Fluor is a registered trademark)).
  • the fluorescent protein may be, for example, PA-GFP, Dronpa or the like.
  • FIG. 1 is a figure showing a microscope apparatus 1 according to the present embodiment.
  • the microscope apparatus 1 includes a stage 2 , a light source apparatus 3 , an illumination optical system 4 , an image forming optical system 5 , an image-capturing unit 6 , an image processing unit 7 and a control apparatus 8 .
  • the control apparatus 8 includes a control unit 42 that performs overall control of each unit of the microscope apparatus 1 .
  • the control apparatus 8 may be a computer that executes procedures mentioned below by reading in a software program.
  • the stage 2 holds a cover glass 51 .
  • the cover glass 51 holds a sample W which is an observation target. More specifically, as shown in FIG. 1 , the cover glass 51 is placed on the stage 2 , and the sample W is placed on the cover glass 51 .
  • the stage 2 may move or may not move in the XY plane.
  • the light source apparatus 3 includes an activation light source 10 a , an excitation light source 10 b , a shutter 11 a and a shutter 11 b .
  • the activation light source 10 a emits activation light L 2 that activates some of fluorescent materials included in the sample W.
  • the fluorescent materials include a reporter pigment, but does not include an activator pigment.
  • the reporter pigment of the fluorescent materials is irradiated with the activation light L 2 to enter the activated state where it can emit fluorescence.
  • the fluorescent materials may include a reporter pigment and an activator pigment, and in this case, the activator pigment brings the reporter pigment into the activated state upon reception of the activation light L 2 .
  • the fluorescent materials may be a fluorescent protein such as PA-GFP or Dronpa, for example.
  • the excitation light source 10 b emits excitation light L 1 that excites at least some of the fluorescent materials activated in the sample W.
  • the fluorescent materials emit fluorescence or become deactivated upon being irradiated with the excitation light L 1 in its activated state.
  • the fluorescent materials enter the activated state again upon being irradiated with the activation light L 2 when they are in a state where they are deactivated (hereinafter, referred to as the deactivated state).
  • the activation light source 10 a and excitation light source 10 b for example include solid-state light sources such as laser light sources, and respectively emit laser light at wavelengths corresponding to the types of fluorescent material.
  • the emission wavelength of the activation light source 10 a and the emission wavelength of the excitation light source 10 b are for example selected from among approximately 405 nm, approximately 457 nm, approximately 488 nm, approximately 532 nm, approximately 561 nm, approximately 640 nm, approximately 647 nm and the like.
  • the emission wavelength of the activation light source 10 a is approximately 405 nm
  • the emission wavelength of the excitation light source 10 b is a wavelength selected from among approximately 488 nm, approximately 561 nm and approximately 647 nm.
  • the shutter 11 a is controlled by the control unit 42 and can switch between the state where the activation light L 2 from the activation light source 10 a is allowed to pass and the state where the activation light L 2 is blocked.
  • the shutter 11 b is controlled by the control unit 42 and can switch between the state where the excitation light L 1 from the excitation light source 10 b is allowed to pass and the state where the excitation light L 1 is blocked.
  • the light source apparatus 3 includes a mirror 12 , a dichroic mirror 13 , an acousto-optic element 14 and a lens 15 .
  • the mirror 12 is provided for example on the exit side of the excitation light source 10 b .
  • the excitation light L 1 from the excitation light source 10 b is reflected on the mirror 12 to enter the dichroic mirror 13 .
  • the dichroic mirror 13 is provided for example on the exit side of the activation light source 10 a .
  • the dichroic mirror 13 has characteristics of transmitting the activation light L 2 and of reflecting the excitation light L 1 .
  • the activation light L 2 transmitted through the dichroic mirror 13 and the excitation light L 1 reflected on the dichroic mirror 13 pass through the same optical path to enter the acousto-optic element 14 .
  • the acousto-optic element 14 is, for example, an acousto-optic filter or the like.
  • the acousto-optic element 14 is controlled by the control unit 42 and can adjust the optical intensity of the activation light L 2 and the optical intensity of the excitation light L 1 , respectively.
  • the acousto-optic element 14 is controlled by the control unit 42 and, for each of the activation light L 2 and the excitation light L 1 , can switch between the light-transmitting state where the activation light L 2 or excitation light L 1 is allowed to pass and the light-blocking state where the activation light L 2 or excitation light L 1 is blocked by the acousto-optic element 14 or where the intensity of the activation light L 2 or excitation light L 1 is lowered by the acousto-optic element 14 .
  • the control unit 42 controls the acousto-optic element 14 such that the activation light L 2 and the excitation light L 1 are simultaneously radiated onto the sample W.
  • the control unit 42 for example controls the acousto-optic element 14 such that the excitation light L 1 is radiated onto the sample W after being irradiated with the activation light L 2 .
  • the lens 15 is, for example, a coupler, and concentrates the activation light L 2 and the excitation light L 1 from the acousto-optic element 14 onto a light-guiding member 16 .
  • the microscope apparatus 1 may not include at least part of the light source apparatus 3 .
  • the light source apparatus 3 is formed into a unit, and may be provided replaceably to (attachably to and detachably from) the microscope apparatus 1 .
  • the light source apparatus 3 may be attached to the microscope apparatus 1 at the time of observation with the microscope apparatus 1 or the like.
  • the illumination optical system 4 radiates the activation light L 2 to activate some of fluorescent materials included in the sample W, and the excitation light L 1 to excite at least some of the activated fluorescent materials.
  • the illumination optical system 4 radiates the activation light L 2 and excitation light L 1 from the light source apparatus 3 onto the sample W.
  • the illumination optical system 4 includes the light-guiding member 16 , a lens 17 , a lens 18 , a filter 19 , a dichroic mirror 20 and an objective lens 21 .
  • the light-guiding member 16 is, for example, an optical fiber, and guides the activation light L 2 and excitation light L 1 to the lens 17 .
  • optical paths from the exit end of the light-guiding member 16 to the sample W are indicated with dotted lines.
  • the lens 17 is, for example, a collimator, and converts the activation light L 2 and excitation light L 1 into collimated light.
  • the lens 18 for example, concentrates the activation light L 2 and excitation light L 1 onto a position on the pupil surface of the objective lens 21 .
  • the filter 19 for example, has characteristics of transmitting the activation light L 2 and excitation light L 1 and of blocking at least part of light at other wavelengths.
  • the dichroic mirror 20 has characteristics of reflecting the activation light L 2 and excitation light L 1 and of transmitting light at wavelengths in the wavelength band of fluorescence emitted by fluorescent materials of the sample W.
  • Light from the filter 19 is reflected on the dichroic mirror 20 to enter the objective lens 21 .
  • the sample W is arranged on the front-side focus surface of the objective lens 21 at the time of observation.
  • the activation light L 2 and excitation light L 1 are radiated onto the sample W by the illumination optical system 4 as mentioned above.
  • the above-mentioned illumination optical system 4 is one example, and can be changed as appropriate. For example, part of the above-mentioned illumination optical system 4 may be omitted.
  • the illumination optical system 4 may include at least part of the light source apparatus 3 .
  • the illumination optical system 4 may include an aperture stop, an irradiation field stop or the like.
  • the image forming optical system 5 forms an image of fluorescence from fluorescent materials included in the sample W.
  • the image forming optical system 5 includes: a first optical system 5 A to form a primary image of fluorescence having exited from the sample W; and a second optical system 5 B to form an image of fluorescence as a secondary image at the image-capturing unit 6 , from the primary image generated at the first optical system 5 A.
  • the first optical system 5 A includes the objective lens 21 , the dichroic mirror 20 , a filter 24 , a lens 25 and an optical path switching member 26 .
  • the second optical system 5 B includes a lens 27 , a lens 28 , and a cylindrical lens 61 as an astigmatic optical system.
  • the image forming optical system 5 shares the objective lens 21 and dichroic mirror 20 with the illumination optical system 4 .
  • optical paths between the sample W and the image-capturing unit 6 are indicated with solid lines.
  • a drive unit 50 moves the objective lens 21 in the optical axis-direction of the objective lens 21 , that is, the Z direction in FIG. 1 .
  • the filter 24 has characteristics of selectively transmitting light at wavelengths in a predetermined wavelength band in light from the sample W.
  • the filter 24 for example, blocks illumination light, external light, stray light or the like reflected on the sample W.
  • the filter 24 is, for example, formed into a unit together with the filter 19 and dichroic mirror 20 , and this filter unit 29 is provided replaceably.
  • the filter unit 29 may be replaced, for example, depending on the wavelength of light to exit from the light source apparatus 3 (examples: the wavelength of the activation light L 2 , the wavelength of the excitation light L 1 ), the wavelength of fluorescence radiated from the sample W, or the like, or a single filter unit that can cope with a plurality of excitation and fluorescence wavelengths may be utilized therefor.
  • the optical path switching member 26 is, for example, a prism, and is provided to an optical path of the image forming optical system 5 insertably and removably.
  • the optical path switching member 26 is, for example, inserted into and removed from the optical path of the image forming optical system 5 by a drive unit (not illustrated) controlled by the control unit 42 .
  • the optical path switching member 26 if inserted into the optical path of the image forming optical system 5 , guides fluorescence from the sample W to an optical path toward the image-capturing unit 6 by internal reflection.
  • the image-capturing unit 6 captures an image formed by the image forming optical system 5 (first observation optical system 5 ).
  • the image-capturing unit 6 includes an image sensor 40 and a control unit 41 .
  • the image sensor 40 is, for example, a CMOS image sensor, but may be a CCD image sensor or the like.
  • the image sensor 40 for example, has a plurality of two-dimensionally arrayed pixels, and has a structure in which a photoelectric converting element such as a photodiode is arranged in each pixel.
  • the image sensor 40 for example, reads out electrical charges accumulated in the photoelectric converting elements by a read-out circuit.
  • the image sensor 40 converts the electrical charges read out into digital data, and outputs data in a digital format in which the positions of pixels and gradation values are associated with each other (example: data about an image).
  • the control unit 41 operates the image sensor 40 based on a control signal input from the control unit 42 of the control apparatus 8 , and outputs data about a captured image to the control apparatus 8 .
  • the control unit 41 outputs, to the control apparatus 8 , an electrical charge accumulation period and an electrical charge read-out period.
  • the above-mentioned image forming optical system 5 is one example, and can be changed as appropriate. For example, part of the above-mentioned image forming optical system 5 may be omitted.
  • the image forming optical system 5 may include an aperture stop, a field stop or the like.
  • the microscope apparatus 1 includes a second observation optical system 30 utilized for setting of an observation range or the like.
  • the second observation optical system 30 includes the objective lens 21 , the dichroic mirror 20 , the filter 24 , the lens 25 , a mirror 31 , a lens 32 , a mirror 33 , a lens 34 , a lens 35 , a mirror 36 , and a lens 37 in this order from the sample W to a viewpoint Vp of the observer.
  • the observation optical system 30 shares the configuration ranging from the objective lens 21 to the lens 25 with the image forming optical system 5 .
  • the second observation optical system 30 forms an intermediate image of the sample W in an optical path between the lens 35 and the lens 37 .
  • the lens 37 is, for example, an eye-piece, and the observer can perform setting of an observation range or the like by observing the intermediate image.
  • the control apparatus 8 performs collective control of each unit of the microscope apparatus 1 .
  • the control apparatus 8 includes the control unit 42 and the image processing unit 7 .
  • the control unit 42 supplies the acousto-optic element 14 with a control signal to switch between the light-transmitting state where light from the light source apparatus 3 is allowed to pass and the light-blocking state where light from the light source apparatus 3 is blocked.
  • the acousto-optic element 14 switches between the light-transmitting state and the light-blocking state.
  • the control unit 42 controls the acousto-optic element 14 to control the period during which the sample W is irradiated with the activation light L 2 and the period during which the sample W is not irradiated with the activation light L 2 .
  • the control unit 42 controls the acousto-optic element 14 to control the period during which the sample W is irradiated with the excitation light L 1 and the period during which the sample W is not irradiated with the excitation light L 1 .
  • the control unit 42 controls the acousto-optic element 14 to control the optical intensity of the activation light L 2 radiated onto the sample W and the optical intensity of the excitation light L 1 radiated onto the sample W.
  • the control unit 42 controls the image-capturing unit 6 to cause the image sensor 40 to execute image-capturing.
  • control unit 41 may supply the acousto-optic element 14 with a control signal to switch between the light-blocking state and the light-transmitting state and control the acousto-optic element 14 based on a signal indicating an electrical charge accumulation period and an electrical charge read-out period (information about image-capturing timing).
  • the control unit 42 acquires data as an image-capturing result from the image-capturing unit 6 .
  • the image processing unit 7 uses the image-capturing result of the image-capturing unit 6 to perform image processing such as obtaining the centroids of individual images.
  • the control unit 42 causes the image-capturing unit 6 to capture images in a plurality of frame periods, and the image processing unit 7 generates a single image using at least some of image-capturing results obtained in the plurality of frame periods.
  • the control apparatus 8 is, for example, connected to a storage apparatus 43 and a display apparatus 44 respectively communicatably.
  • the display apparatus 44 is, for example, a liquid crystal display or the like.
  • the display apparatus 44 displays various types of images such as images indicating various types of setting of the microscope apparatus 1 , images captured by the image-capturing unit 6 or images generated from captured images.
  • the control unit 42 controls the display apparatus 44 to cause the display apparatus 44 to display various types of images.
  • the control unit 42 supplies the display apparatus 44 with data about super-resolution images such as a STORM image or a PALM image generated by the image processing unit 7 , and causes the display apparatus 44 to display the image.
  • the microscope apparatus 1 can display, as a live video, super-resolution images of the sample W, which is an observation target, and so on.
  • the storage apparatus 43 is, for example, a magnetic disk, an optical disk or the like, and stores various types of data such as data about various types of setting of the microscope apparatus 1 , data about results of image-capturing by the image-capturing unit 6 or data about images generated by the image processing unit 7 .
  • the control unit 42 can supply the display apparatus 44 with data about a super-resolution image stored in the storage apparatus 43 and cause the display apparatus 44 to display the super-resolution image.
  • the control unit 42 controls the storage apparatus 43 to cause the storage apparatus 43 to store various types of data.
  • FIG. 2 is a figure showing the image forming optical system 5 and the image sensor 40 .
  • the image forming optical system 5 is an optical system whose optical axis 5 a makes a turn at the optical path switching member 26
  • the image forming optical system 5 is conceptually shown as a straight optical system.
  • illustration of the configuration between the objective lens 21 and the cylindrical lens 61 is omitted.
  • the direction parallel with the optical axis 5 a of the image forming optical system 5 is assumed to be the Z direction, and two directions that are vertical to the Z direction and are vertical to each other are assumed to be the X direction and Y direction.
  • the X direction is, for example, the horizontal direction of the image sensor 40
  • the Y direction is, for example, the vertical direction of the image sensor 40 .
  • the cylindrical lens 61 is an optical member having power (refractive power) in only either one of the vertical direction and the horizontal direction.
  • the cylindrical lens 61 has power in the XZ plane, but does not have power in the YZ plane.
  • the astigmatic optical system may be one using a toroidal lens which has power in both the vertical direction and the horizontal direction which is different in these two directions.
  • FIG. 3 is a figure showing an optical path of fluorescence from and an image of fluorescence from a position Z 2 in the sample W.
  • Fluorescence from a fluorescent material present at the position Z 2 in the sample W enters the cylindrical lens 61 as collimated light. This fluorescence heads to the image sensor 40 while at the same time being concentrated by refraction at the cylindrical lens 61 .
  • the width Wx of an image of fluorescence in the X direction becomes the smallest at a position Z 4 corresponding to the focus in the XZ plane.
  • the width Wx increases as the distance from the position Z 4 increases toward the side opposite to the sample W (+Z side).
  • Fluorescence from a fluorescent material present at the position Z 2 in the sample W enters the cylindrical lens 61 as collimated light. This fluorescence enters the lens 28 almost without being refracted by the cylindrical lens 61 . This fluorescence heads to the image sensor 40 while at the same time being concentrated by refraction at the lens 28 .
  • the width Wy of an image of fluorescence in the Y direction becomes the smallest at a position Z 6 corresponding to the focus in the YZ plane. The width Wy increases as the distance from the position Z 6 increases toward the side on which the sample W is located ( ⁇ Z side).
  • a position Z 5 which is an intermediate position between the position Z 4 and the position Z 6 , the width Wx of the image of fluorescence in the X direction and the width Wy of the image of fluorescence in the X direction become equal to each other, and the image of fluorescence becomes a perfect circle. Accordingly, if the image sensor 40 is arranged at the position Z 5 , a perfect circle image is acquired as an image of fluorescence from a fluorescent material present at the position Z 2 in the sample W.
  • FIG. 4 is a figure showing an optical path of fluorescence from and an image of fluorescence from a fluorescent material present at a position Z 1 in the sample W.
  • the position Z 1 is located on the side opposite to the image sensor 40 relative to the position Z 2 ( ⁇ Z side).
  • the width Wx of an image of fluorescence in the X direction becomes the smallest at a position away from the position Z 4 toward the sample W side ( ⁇ Z side).
  • the width Wy of an image of fluorescence in the Y direction becomes the smallest at the position Z 5 .
  • an ellipse image with its major axis extending in the X direction is acquired as an image of fluorescence from a fluorescent material present at the position Z 1 in the sample W.
  • FIG. 5 is a figure showing an optical path of fluorescence from and an image of fluorescence from a fluorescent material present at a position Z 3 in the sample W.
  • the position Z 3 is located on the side on which the image sensor 40 is located (+Z side) relative to the position Z 2 .
  • the width Wx of an image of fluorescence in the X direction becomes the smallest at the position Z 5 .
  • the width Wy of an image of fluorescence in the Y direction becomes the smallest at a position away from the position Z 6 toward the side opposite to the sample W (+Z side).
  • an ellipse image with its major axis extending in the Y direction is acquired as an image of fluorescence from a fluorescent material present at the position Z 3 in the sample W.
  • the image processing unit 7 performs elliptical Gaussian fitting to identify the width Wx and width Wy, in addition to the centroid of an image of fluorescence. Thereby, the image processing unit 7 can identify the position, in the XY plane, of a fluorescent material corresponding to an image of fluorescence based on the centroid, and additionally identify the Z direction position based on the relationship between the width Wx and the width Wy.
  • FIG. 6 is a figure showing one example of image processing performed by the image processing unit 7 at a predetermined image-capturing position.
  • the image-capturing position refers to the position of the front-side focus of the objective lens 21 in the sample W.
  • reference symbols P 1 to Pn represent image-capturing results obtained in predetermined image-capturing frame periods. Because fluorescent materials activated at a probability corresponding to the intensity of the activation light L 2 emit fluorescence upon being irradiated with the excitation light L 1 , the number of fluorescent materials corresponding to this probability emit fluorescence in one frame, and different fluorescent materials emit fluorescence in different frames.
  • the image processing unit 7 calculates, for each among the captured images P 1 to Pn, the centroid and respective widths in the X and Y directions of an image of fluorescence included in each image.
  • the image processing unit 7 for example, performs elliptical Gaussian fitting on the distribution of pixel values in this region to calculate the XY position of the centroid and the widths Wx, Wy.
  • the image processing unit 7 also identifies the Z direction position of the centroid based on the relationship between the calculated widths Wx, Wy and by referring to a prestored table or the like.
  • the image processing unit 7 uses at least some of a plurality of centroids Q corresponding to a plurality of images of fluorescence for which the centroids Q of images of fluorescence Im are calculated and that are included in a plurality of the captured images P 1 to Pn to generate three-dimensional distribution information about the centroids Q.
  • a three-dimensional super-resolution image SP 1 is generated and displayed.
  • the super-resolution image SP is one obtained through reconstruction of a three-dimensional structure of the sample W.
  • the three-dimensional distribution information obtained at a predetermined image-capturing position is referred to as a layer.
  • the three-dimensional super-resolution image SP 1 is shown as a three-dimensional image that allows a user to specify the viewing direction. Instead of this or in addition to this, an image SP 2 projected onto the XY plane and an image SP 3 projected onto the XZ plane may be displayed.
  • FIG. 7 shows one example of a data configuration 45 of three-dimensional distribution information about centroids Q.
  • XYZ coordinates of centroids Q are associated with the numbers to identify these centroids Q.
  • the brightness B and widths Wx, Wy may be associated.
  • Three-dimensional distribution information in this configuration is stored in the storage apparatus 43 .
  • the microscope apparatus 1 allows reconstruction of a structure in the thickness direction of the sample W (that is, the optical axis-direction 21 a of the objective lens) even if the objective lens 21 is fixed in the optical axis-direction.
  • the range in the thickness direction over which the structure can be reconstructed is inevitably restricted by factors such as the depth of field of the image forming optical system 5 , fluorescence intensity, the S/N ratio of an image sensor or the focal distance of the cylindrical lens 61 , with the center of the range being at the image-capturing position.
  • the range in the thickness direction over which a structure can be reconstructed is about 300 nm.
  • a structure of the sample W is reconstructed over a wide range in the thickness direction by capturing images of fluorescence by the image-capturing unit 6 while at the same time moving the objective lens 21 by the drive unit 50 in the optical axis-direction 21 a of the objective lens 21 .
  • FIG. 8 is a schematic diagram of a sequence of actions performed by the microscope apparatus 1 .
  • FIG. 9 is a flowchart showing an image-capturing process performed by the microscope apparatus 1 .
  • FIG. 10 shows one example of a screen on which image-capturing conditions in an image-capturing process are input.
  • Performing image-capturing multiple times at a predetermined image-capturing position by the image-capturing unit 6 is assumed to be one unitary image-capturing process.
  • a procedure of acquiring images at a plurality of predetermined image-capturing positions is referred to as a passing procedure. Therefore, the passing procedure means, for example, that the objective lens 21 moves from an image-capturing position 1 to an image-capturing position 8 in the direction of the optical axis 5 a of the objective lens 21 to capture images.
  • image-capturing conditions are set in an image-capturing process (S 100 ).
  • the microscope apparatus 1 displays a screen 100 as shown in FIG. 10 on the display apparatus 44 .
  • the control apparatus 8 accepts inputs of the position of the lowermost plane for image-capturing, the position of the uppermost plane for image-capturing, the step size in the Z direction, the number of frames to be acquired at each image-capturing position and the name of a file for storing three-dimensional distribution information.
  • the numbers of frames corresponding to image-capturing positions may be automatically calculated based on a predetermined rule. For example, if the density of dyed cellular tissues is high at lower layers of the cell and decreases at higher layers, it is preferable to increase the number of frames as the distance of the image-capturing position from the cover glass 51 increases.
  • an image-capturing sequence in each passing procedure may be specified. For example, in initial setting, a sequential movement from a lower image-capturing position to an upper image-capturing position as mentioned above is selected. Instead of this image-capturing sequence, image-capturing sequences such as from top to bottom, even-numbered image-capturing positions first and odd-numbered image-capturing positions next, or random may be allowed to be selected.
  • the image-capturing sequence may be specified as Z 1 , Z 8 , Z 2 , Z 7 , Z 3 , Z 6 and Z 5 .
  • this image-capturing sequence it becomes possible to obtain an image with a clear outline of the sample.
  • the control apparatus 8 calculates the image-capturing position, that is, the amount of driving by the drive unit 50 based on the image-capturing conditions (S 102 ). In this case, it is preferable to set the image-capturing positions and/or other image-capturing conditions such that adjacent layers partially overlap in the thickness direction of the sample W.
  • FIG. 8 shows an example of a sequence of actions in which the number of layers is eight and the number of times of passing procedures is one.
  • the illumination optical system 4 radiates the activation light L 2 and excitation light L 1 from the light source 3 onto the sample W 1 (S 104 ), and performs image-capturing by the image-capturing unit 6 (S 106 ) in a predetermined number of frames corresponding to an image-capturing position (S 108 : No).
  • Step S 104 to S 109 are repeated for calculated image-capturing positions (S 110 : No).
  • the control unit 8 judges whether or not the process has reached the final passing procedure (S 111 ). If the process has not reached the final passing procedure (S 111 : No), the process proceeds to the next passing procedure and repeats Steps S 104 to S 110 . If the process has reached the final passing procedure (S 111 : Yes), the image-capturing process ends.
  • the structure of the sample can be reconstructed over a range wider than the range in the thickness direction restricted by the focal distance of a cylindrical lens or the like.
  • a structure of a sample is reconstructed on the basis of a result of image-capturing in the number of frames that is different for each image-capturing position, improvement in the image quality of a reconstructed image corresponding to the shape or properties of the sample can be expected.
  • the image processing unit 7 identifies the centroids of fluorescence in each image-capturing frame, as explained with reference to FIG. 6 .
  • the image processing unit 7 further stores, in the storage apparatus 43 and as a file of a data configuration shown in FIG. 7 , three-dimensional distribution information about the centroids identified in each image-capturing frame included in a unitary image-capturing process for each image-capturing position.
  • FIG. 11 is a schematic diagram showing a plurality of layers identified by image-capturing at a plurality of image-capturing positions.
  • FIG. 11 shows the case where the number of layers is three and the number of times of passing procedures is one.
  • the image processing unit 7 stores, in the storage apparatus 43 , the centroids of fluorescence identified passing procedure-by-passing procedure and layer-by-layer, that is, for each unitary image-capturing process, in association with information identifying passing procedures and layers.
  • the Z position of the centroid of each fluorescent beam in the sample space is identified using the image-capturing position based on the amount of driving by the drive unit 50 and the Z position of the centroid of fluorescence in each layer (the Z position of the centroid calculated from a captured image of fluorescence).
  • the image-capturing position is the position of the front-side focus of the objective lens 21 in the sample as mentioned above, but the position of the front-side focus may be a position relative to the cover glass 51 .
  • FIG. 12 shows one example of a screen 150 on which setting of a reconstructed display image of the sample W is performed.
  • the screen 150 is displayed on the display apparatus 44 , and accepts setting about whether or not to display an image while visually distinguishing passing procedures, layers or both of them.
  • FIG. 12 shows an example in which display colors are used, as a method of visual distinguishing.
  • different ones among three attributes of color may be used for passing procedures and layers; for example, passing procedures may be distinguished by hues, and layers may be distinguished by luminosity.
  • passing procedures may be distinguished by hues, and layers may be distinguished by luminosity.
  • visual distinguishing may be enabled using hatching or shapes, for example, using circles, triangles, x marks or the like.
  • a region 152 on the screen 150 is for selecting whether or not to perform coloring differently layer-by-layer
  • a region 154 on the screen 150 is for selecting whether or not to perform coloring differently passing procedure-by-passing procedure.
  • To the right of the region 152 respective outlined layers are schematically shown with the layers being arrayed from top to bottom, and to the further right of them, the state where coloring is performed differently layer-by-layer if it is selected to perform coloring differently layer-by-layer is schematically shown.
  • a box 156 is for inputting the lower limit of the range to be displayed in the Z direction
  • a box 157 is for inputting the upper limit of the range to be displayed in the Z direction
  • a box 158 is for inputting the lower limit of passing procedures to be displayed
  • a box 159 is for inputting the upper limit of passing procedures to be displayed.
  • Information about the configuration and range of data obtained through an image-capturing process is further displayed on the screen 150 .
  • the data configuration the thickness of a layer, number of layers and number of times of passing procedures are displayed.
  • FIG. 13 shows a display image CL 1 in which coloring is performed differently layer-by-layer using the three-dimensional distribution information shown in FIG. 11 and according to the setting shown in FIG. 12 .
  • colors cannot be used in the figures, white, hatching and black are used instead of colors.
  • the image processing unit 7 accesses the centroids of fluorescence stored in the storage apparatus 43 , and allocates colors based on information identifying layers.
  • the image processing unit 7 allocates white to the centroids of fluorescence in a layer 1 , allocates hatching to the centroids of fluorescence in a layer 2 , and allocates black to the centroids of fluorescence in a layer 3 to generate the display image CL 1 and displays it on the display apparatus 44 .
  • the image processing unit 7 displays an image such that it is possible to visually distinguish which centroid belongs to which one among a plurality of layers, a user can easily recognize the degree of drifts of the objective lens 21 and stage 2 among layers (which appear for example as positional displacement).
  • FIG. 14 is a schematic diagram showing laps among layers in the thickness direction of a sample.
  • a plurality of image-capturing positions are preferably set such that parts of three-dimensional distribution information to be generated as a result overlap in the thickness direction of the sample.
  • FIG. 14 indicates that there is a lap Lw between the layer 1 and the layer 2 .
  • the setting screen 150 in FIG. 12 preferably allows setting such that the centroid included in the lap Lw is displayed, manually or automatically.
  • FIG. 15 is a schematic diagram for explaining a display image CL 2 to display the centroid included in the lap Lw.
  • the image processing unit 7 identifies the centroid having a Z coordinate included in the lap Lw based on the three-dimensional distribution information about each layer.
  • the centroid surrounded by broken lines in the layer 1 and layer 2 is identified as one included in the lap Lw.
  • the image processing unit 7 allocates white to the above-mentioned centroid in the layer 1 and allocates hatching to the above-mentioned centroid in the layer 2 to generate the display image CL 2 and displays it on the display apparatus 44 . Thereby, the degree of drifts of the objective lens 21 and stage 2 among layers can be further easily recognized.
  • FIG. 16 is a schematic diagram of another sequence of actions performed by the microscope apparatus 1 .
  • FIG. 17 shows an example of a screen 110 on which image-capturing conditions in an image-capturing process corresponding to FIG. 16 are input.
  • the microscope apparatus 1 repeats passing procedures multiple times. Specifically, passing procedures are repeated ten times starting at a passing procedure 1 and ending at a passing procedure 10 .
  • the number of times of passing procedures is input by a user using the screen 110 o FIG. 17 .
  • the screen 110 accepts inputs of the position of the lowermost plane for image-capturing, the position of the uppermost plane for image-capturing, the step size in the Z direction, the sequence of image-capturing in each passing procedure and the name of a file for storing three-dimensional distribution information.
  • the screen 110 further accepts inputs of the entire number of frames to be used in each layer.
  • the image-capturing process in FIG. 16 and FIG. 17 may also be executed following the flowchart of FIG. 9 .
  • the entire number of frames input through the screen 110 is divided by the number of times of passing procedures to calculate the number of frames to be acquired at each image-capturing position in each passing procedure, that is, the number of frames in each unitary image-capturing process. For example if the number of frames to be used in all the passing procedures for each image-capturing position is 10000 in total, and the number of times of passing procedures is ten, the number of frames to be obtained at each image-capturing position becomes 1000. As in the example of FIG. 8 to FIG. 10 , the number of frames may be different image-capturing position-by-image-capturing position.
  • the action of moving sequentially from the image-capturing position 1 to the image-capturing position 8 is assumed to be one passing procedure, and this action is repeated ten times.
  • the image-capturing position moves to each image-capturing position a number of times of passing procedures, and frames corresponding to the above-mentioned unitary image-capturing process are captured at respective image-capturing positions.
  • the number of image-capturing frames inevitably increases, if image-capturing is performed with the number of times of passing procedures being set to one, the color of a fluorescent material fades at latter image-capturing timing, and for example, the amounts of a fluorescent material may differ largely between a layer at the first image-capturing position and a layer at the final image-capturing position, in some cases.
  • a plurality of passing procedures can be set, differences in the amounts of fluorescent materials among layers due to decoloration can be reduced, and differences in image quality among layers can be reduced.
  • FIG. 18 is a schematic diagram showing a plurality of layers identified by image-capturing in a plurality of passing procedures and at a plurality of image-capturing positions.
  • FIG. 18 shows the case where the number of layers is three and the number of times of passing procedures is two.
  • FIG. 19 shows a display image CL 3 in which coloring is performed differently layer-by-layer using the passing procedures in FIG. 18 .
  • white, hatching and black are used instead of colors.
  • the image processing unit 7 accesses the XYZ positions of images of fluorescence stored in the storage apparatus 43 , and allocates colors based at least on information identifying layers.
  • the image processing unit 7 allocates white to images of fluorescence of the layer 1 included in all the passing procedures, allocates hatching to images of fluorescence of the layer 2 included in all the passing procedures, and allocates black to images of fluorescence of the layer 3 included in all the passing procedures to generate the display image CL 3 and display it on the display apparatus 44 .
  • colors of the layer 1 to the layer 3 are indicated as a legend and additionally that passing procedure-by-passing procedure distinguishing is not done is indicated with a character string, “All Passing Procedures” or the like.
  • the image processing unit 7 displays an image such that it is possible to visually distinguish which three-dimensional position has been identified based on an image-capturing result of which one among a plurality of layers, a user can easily recognize the degree of drifts (positional displacement) generated between the sample W and the objective lens generated due to differences in image-capturing timing among layers.
  • FIG. 20 shows a display image CL 4 in which coloring is performed differently passing procedure-by-passing procedure using the layers in FIG. 18 .
  • white and black are used instead of colors.
  • the image processing unit 7 accesses the XYZ positions of images of fluorescence stored in the storage apparatus 43 , and allocates colors based at least on information identifying passing procedures. In the example of FIG. 12 , the image processing unit 7 allocates white to images of fluorescence of all the layers included in the passing procedure 1 and allocates black to images of fluorescence of all the layers included in the passing procedure 2 to generate the display image CL 4 and display it on the display apparatus 44 . In such a case, preferably, the image processing unit 7 indicates the color of the passing procedure 1 and the color of the passing procedure 2 as a legend, and additionally that layer-by-layer distinguishing is not done is indicated with a character string, “All Layers” or the like.
  • the image processing unit 7 displays an image such that it is possible to visually distinguish which three-dimensional position has been identified based on an image-capturing result of which one among a plurality of passing procedures, a user can easily recognize the degree of drifts of the sample W among passing procedures.
  • three-dimensional distribution information is shown as an “image”, this is for the purpose of explanation. That is, in other cases than the case where the display apparatus 44 is caused to display it as a “display image”, the image processing unit 7 may not generate an image, but only has to treat it as data like the one shown in FIG. 7 .
  • the drive unit 50 moves the objective lens 21 in the optical axis-direction 5 a to move the image-capturing position.
  • the drive unit 50 may move the stage 2 in the optical axis-direction 5 a of the objective lens 21 to move the image-capturing position.
  • excitation light at a wavelength to excite one of the fluorescent materials may be radiated and in even-numbered passing procedures among the plurality of passing procedures, excitation light at a wavelength to excite another fluorescent material may be radiated.

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Chemical & Material Sciences (AREA)
  • Analytical Chemistry (AREA)
  • Multimedia (AREA)
  • Engineering & Computer Science (AREA)
  • Optics & Photonics (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biochemistry (AREA)
  • General Health & Medical Sciences (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Signal Processing (AREA)
  • Microscoopes, Condenser (AREA)
  • Investigating, Analyzing Materials By Fluorescence Or Luminescence (AREA)
  • Studio Devices (AREA)
US15/991,426 2015-11-27 2018-05-29 Microscope apparatus Abandoned US20180275061A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/178,957 US11906431B2 (en) 2015-11-27 2021-02-18 Microscope apparatus

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2015/083511 WO2017090211A1 (fr) 2015-11-27 2015-11-27 Microscope

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2015/083511 Continuation WO2017090211A1 (fr) 2015-11-27 2015-11-27 Microscope

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/178,957 Division US11906431B2 (en) 2015-11-27 2021-02-18 Microscope apparatus

Publications (1)

Publication Number Publication Date
US20180275061A1 true US20180275061A1 (en) 2018-09-27

Family

ID=58763206

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/991,426 Abandoned US20180275061A1 (en) 2015-11-27 2018-05-29 Microscope apparatus
US17/178,957 Active 2036-10-13 US11906431B2 (en) 2015-11-27 2021-02-18 Microscope apparatus

Family Applications After (1)

Application Number Title Priority Date Filing Date
US17/178,957 Active 2036-10-13 US11906431B2 (en) 2015-11-27 2021-02-18 Microscope apparatus

Country Status (4)

Country Link
US (2) US20180275061A1 (fr)
EP (1) EP3385769A4 (fr)
JP (1) JP6617774B2 (fr)
WO (1) WO2017090211A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11385448B2 (en) * 2018-09-28 2022-07-12 Sysmex Corporation Microscope apparatus

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN211830951U (zh) * 2020-09-23 2020-10-30 常州市瑞泰光电有限公司 影像采集装置

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4468684B2 (ja) 2003-12-05 2010-05-26 オリンパス株式会社 走査型共焦点顕微鏡装置
US7838302B2 (en) 2006-08-07 2010-11-23 President And Fellows Of Harvard College Sub-diffraction limit image resolution and other imaging techniques
JP4996319B2 (ja) 2007-04-26 2012-08-08 オリンパス株式会社 レーザ顕微鏡とその画像表示方法
WO2008155883A1 (fr) 2007-06-15 2008-12-24 Nikon Corporation Dispositif de microscope confocal
US8564792B2 (en) * 2007-12-21 2013-10-22 President And Fellows Of Harvard College Sub-diffraction limit image resolution in three dimensions
DE102009060793A1 (de) * 2009-12-22 2011-07-28 Carl Zeiss Microlmaging GmbH, 07745 Hochauflösendes Mikroskop und Verfahren zur zwei- oder dreidimensionalen Positionsbestimmung von Objekten
WO2011152523A1 (fr) * 2010-06-03 2011-12-08 株式会社ニコン Dispositif de microscope
US9946058B2 (en) * 2010-06-11 2018-04-17 Nikon Corporation Microscope apparatus and observation method
JP5747690B2 (ja) * 2011-07-04 2015-07-15 株式会社ニコン 顕微鏡装置及び画像形成方法
US10001622B2 (en) * 2011-10-25 2018-06-19 Sanford Burnham Medical Research Institute Multifunction autofocus system and method for automated microscopy
DE102012200344A1 (de) * 2012-01-11 2013-07-11 Carl Zeiss Microscopy Gmbh Mikroskopsystem und Verfahren für die 3-D hochauflösende Mikroskopie
EP2966492B1 (fr) * 2012-05-02 2020-10-21 Centre National De La Recherche Scientifique Procédé et appareil de localisation de particules uniques au moyen d'une analyse d'ondelettes
US8818117B2 (en) 2012-07-19 2014-08-26 Sony Corporation Method and apparatus for compressing Z-stack microscopy images
JP6131568B2 (ja) * 2012-10-30 2017-05-24 株式会社ニコン 顕微鏡装置及び画像形成方法
US10429628B2 (en) * 2015-04-23 2019-10-01 The University Of British Columbia Multifocal method and apparatus for stabilization of optical systems
US20180329225A1 (en) * 2015-08-31 2018-11-15 President And Fellows Of Harvard College Pattern Detection at Low Signal-To-Noise Ratio
JP6635125B2 (ja) 2015-11-27 2020-01-22 株式会社ニコン 顕微鏡、観察方法、及び画像処理プログラム

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11385448B2 (en) * 2018-09-28 2022-07-12 Sysmex Corporation Microscope apparatus

Also Published As

Publication number Publication date
WO2017090211A1 (fr) 2017-06-01
EP3385769A1 (fr) 2018-10-10
US20210172877A1 (en) 2021-06-10
JP6617774B2 (ja) 2019-12-11
JPWO2017090211A1 (ja) 2018-09-06
US11906431B2 (en) 2024-02-20
EP3385769A4 (fr) 2019-10-30

Similar Documents

Publication Publication Date Title
US8704196B2 (en) Combination microscopy
JP6375254B2 (ja) 蛍光観察用ユニットおよび蛍光観察装置
US9046489B2 (en) Fluorescence imaging device
US11906431B2 (en) Microscope apparatus
US20190294930A1 (en) Information processing device, image processing device, microscope, information processing method, and information processing program
JP6716201B2 (ja) 光学フィルタシステム、蛍光観察システムおよび蛍光観察を実行する方法
US11042016B2 (en) Multi-Z confocal imaging system
CA2865922A1 (fr) Scanner avec une plage dynamique accrue
CN110114709A (zh) 确定荧光强度的方法和显微镜
US10725278B2 (en) Microscope, observation method, and storage medium
JP2022132488A (ja) 装置、方法、及びプログラム
US8294728B2 (en) Process for generating display images from acquired recorded images, and means for carrying out the process
US20230218142A1 (en) Medical optical system, data processing system, computer program, and non-volatile computer-readable storage medium
JP2019508746A (ja) 源自己蛍光を低減させ、均一性を改良するための散乱を伴う撮像システムおよび方法
US20180017773A1 (en) Microscope device, observation method, and storage medium
US10823675B2 (en) Microscope, observation method, and a storage medium
JP6753454B2 (ja) 顕微鏡装置、観察方法、及び制御プログラム
JP2020042283A (ja) 顕微鏡装置
JP6708289B2 (ja) 顕微鏡、観察方法、及び制御プログラム
US20190162954A1 (en) Optical device, phase plate, and image forming method
JP2017058386A (ja) 顕微鏡装置および観察方法
WO2012144515A1 (fr) Dispositif de microscope et procédé de traitement d'image
JP2019078866A (ja) 顕微鏡システム、観察方法、及び観察プログラム
JP4928330B2 (ja) 蛍光顕微鏡
Li Volumetric Adaptive Optical Multiphoton Microscope for in Vivo Imaging and Photomanipulation

Legal Events

Date Code Title Description
AS Assignment

Owner name: NIKON CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SASE, ICHIRO;KANEKO, YASUTOSHI;FUKUI, TATSUO;REEL/FRAME:045922/0925

Effective date: 20180514

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION