US20170237958A1 - Medical inspection apparatus, such as a microscope or endoscope using pseudocolors - Google Patents

Medical inspection apparatus, such as a microscope or endoscope using pseudocolors Download PDF

Info

Publication number
US20170237958A1
US20170237958A1 US15/424,819 US201715424819A US2017237958A1 US 20170237958 A1 US20170237958 A1 US 20170237958A1 US 201715424819 A US201715424819 A US 201715424819A US 2017237958 A1 US2017237958 A1 US 2017237958A1
Authority
US
United States
Prior art keywords
image
fluorescent
light image
visible
pseudocolor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/424,819
Inventor
George Themelis
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Leica Instruments Singapore Pte Ltd
Original Assignee
Leica Instruments Singapore Pte Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Leica Instruments Singapore Pte Ltd filed Critical Leica Instruments Singapore Pte Ltd
Assigned to LEICA INSTRUMENTS (SINGAPORE) PTE. LTD. reassignment LEICA INSTRUMENTS (SINGAPORE) PTE. LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: THEMELIS, GEORGE
Publication of US20170237958A1 publication Critical patent/US20170237958A1/en
Priority to US17/062,595 priority Critical patent/US11330237B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00002Operational features of endoscopes
    • A61B1/00004Operational features of endoscopes characterised by electronic signal processing
    • A61B1/00009Operational features of endoscopes characterised by electronic signal processing of image signals during a use of endoscope
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/00131Accessories for endoscopes
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/04Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
    • A61B1/043Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances for fluorescence imaging
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B1/00Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
    • A61B1/06Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements
    • A61B1/0661Endoscope light sources
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0059Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
    • A61B5/0071Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence by measuring fluorescence emission
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/742Details of notification to user or communication with user or patient ; user input means using visual displays
    • A61B5/7425Displaying combinations of multiple images regardless of image source, e.g. displaying a reference anatomical image with a live image
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N21/645Specially adapted constructive features of fluorimeters
    • G01N21/6456Spatial resolved fluorescence measurements; Imaging
    • G01N21/6458Fluorescence microscopy
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/0004Microscopes specially adapted for specific applications
    • G02B21/002Scanning microscopes
    • G02B21/0024Confocal scanning microscopes (CSOMs) or confocal "macroscopes"; Accessories which are not restricted to use with CSOMs, e.g. sample holders
    • G02B21/0028Confocal scanning microscopes (CSOMs) or confocal "macroscopes"; Accessories which are not restricted to use with CSOMs, e.g. sample holders specially adapted for specific applications, e.g. for endoscopes, ophthalmoscopes, attachments to conventional microscopes
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/0004Microscopes specially adapted for specific applications
    • G02B21/002Scanning microscopes
    • G02B21/0024Confocal scanning microscopes (CSOMs) or confocal "macroscopes"; Accessories which are not restricted to use with CSOMs, e.g. sample holders
    • G02B21/0052Optical details of the image generation
    • G02B21/0076Optical details of the image generation arrangements using fluorescence or luminescence
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/0004Microscopes specially adapted for specific applications
    • G02B21/002Scanning microscopes
    • G02B21/0024Confocal scanning microscopes (CSOMs) or confocal "macroscopes"; Accessories which are not restricted to use with CSOMs, e.g. sample holders
    • G02B21/008Details of detection or image processing, including general computer control
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B21/00Microscopes
    • G02B21/36Microscopes arranged for photographic purposes or projection purposes or digital imaging or video purposes including associated control and data processing arrangements
    • G02B21/365Control or image processing arrangements for digital or video microscopes
    • G02B21/367Control or image processing arrangements for digital or video microscopes providing an output produced by processing a plurality of individual source images, e.g. image tiling, montage, composite images, depth sectioning, image comparison
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/43Conversion of monochrome picture signals to colour picture signals for colour picture display
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N9/00Details of colour television systems
    • H04N9/64Circuits for processing colour signals
    • H04N9/646Circuits for processing colour signals for image enhancement, e.g. vertical detail restoration, cross-colour elimination, contour correction, chrominance trapping filters
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B2576/00Medical imaging apparatus involving image processing or analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10056Microscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10064Fluorescence image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10068Endoscopic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Definitions

  • the invention relates to a medical inspection apparatus, such as a microscope or endoscope, and to a medical inspection method such as a microscopy or endoscopy.
  • Microscopes and endoscopes using fluorescence of fluorophores to better discern certain types of tissue are known.
  • the present invention aims to improve the quality of the images displayed to a surgeon as compared to the existing microscopes or endoscopes.
  • the microscope or endoscope comprises an image processing unit, the image processing unit comprising a first input section configured to receive visible image data representing a visible-light image of an object, a second input section configured to receive fluorescence image data representing a fluorescent-light image of the object, and an output section configured to output pseudocolor image data representing a pseudocolor image of the object, wherein the image processing unit is adapted to compute a color of an output pixel in the pseudocolor image from at least one pseudocolor, a color of a first input pixel in the visible-light image and an intensity of a second input pixel in the fluorescent-light image.
  • the medical inspection method comprises the steps of acquiring visible image data representing a visible-light image of an object, acquiring fluorescence image data representing a fluorescent-light image of the object, combining the visible-light image data, the fluorescence image data and at least one pseudocolor to obtain output data representing a pseudocolor image, wherein a color of an output pixel in the pseudocolor image depends on at least one pseudocolor, a color of a first input pixel in the visible-light image, and an intensity of a second input pixel in the fluorescent-light image.
  • the object is solved by a non-transitory computer readable medium storing a program, causing a medical visualization apparatus to execute the above-cited method.
  • the solution according to the invention creates a pseudocolor image in which the at least one pseudocolor is blended in a natural-looking way into the visible-light image.
  • the section of the object, which is colored by the fluorescent color is presented in a more natural way.
  • the fluorescent-light image is not simply overlaid onto the visible-light image, but is rather merged with the visible-light image using the at least one pseudocolor to display the fluorescent parts of the object.
  • the invention can be further improved by the following aspects, which are each advantageous on their own, and can be combined arbitrarily.
  • image and data can be used synonymously, as the image represents the totality of the data making up the image.
  • the data of an image may also be grouped into pixels.
  • a “pixel” corresponds to an ordered grouping of the “data”.
  • the images and data are processed and/or stored digitally and are themselves digital.
  • the at least one pseudocolor need not correspond to the color of fluorescence of the fluorophore, but can be chosen by the user of the microscope or endoscope.
  • the pseudocolor is a color in the visible-light range which does not naturally occur in the object to be observed.
  • the medical visualization apparatus is stereoscopic.
  • a separate pseudocolor image recorded from a slightly different point of view is presented to each eye of the observer.
  • two separate pseudocolor images are computed from two different visible-light images and two fluorescent-light images.
  • the same devices and processes may be used as for the single channel of a monocular microscope.
  • the fluorescent-light image may be sampled by a digital NIR camera, the visible-light image by a standard camera.
  • the image processing unit may comprise a memory in which image data are stored.
  • the image processing unit may be part of a general-purpose computer which has been adapted, e.g. by software, to provide a control and processing system for a microscope or endoscope. It is also possible that the image processing unit is a specially-designed electronic device that can be solely used in microscopes or endoscopes.
  • the microscope or endoscope may comprise more than one fluorescent channel.
  • two or more fluorophores may be used which are each excited at different excitation wavelengths and which each emit fluorescent light at different emission wavelengths.
  • Two or more fluorophores can be used to identify different aspects of the object under investigation.
  • one type of fluorophore can be used to mark blood vessels whereas another fluorophore may mark tumourous cells.
  • Another fluorophore may be used to mark e.g. nerve cells.
  • the fluorescence image data may contain at least two different fluorescent emission colors and the processing unit may be configured to assign a different pseudocolor to each of the different fluorescent emission colors.
  • the different types of tissues marked by the different types of fluorophores can be discerned in the pseudocolor image by the different pseudocolors.
  • the different pseudocolors are all blended naturally into the visible-light image due to the consideration of both the color of the first input pixel and the intensity of the second input pixel in the fluorescent-light image.
  • the first and second input pixel should correspond to the same feature of the object.
  • the fluorescent-light image, the visible-light image and the pseudocolor image may have the same size, i.e. comprise the same number of pixel rows and pixel columns. If the visible-light and fluorescent-light images are congruent to each other, i.e. the size, orientation and location of the object is the same in the two images, the first and second input pixel are located in the same row and column in their respective images.
  • the color of the output pixel may be located linearly, in a color space, between the color of the first input pixel and the at least one pseudocolor, the distance between the color of the output pixel and the color of the first input pixel being proportional to the intensity of the second input pixel.
  • Typical color spaces are HSV, HSI, RGB and YMCK without being limited to these color spaces.
  • each color can be represented by a positional vector.
  • computing the color of the output pixel can be done, computationally efficient by vector arithmetics.
  • the image processing unit may be configured to linearly interpolate the RGB values of the first input pixel in the visible-light to the selected pseudocolor value dependent on the intensity of the second input pixel to obtain the color of the output pixel.
  • the color (r o , g o , b o ) of the output pixel may be as follows:
  • the color of the output pixel will correspond to the color of the visible-light image.
  • the processing unit may comprise a threshold adjustment module, which is configured to blank a pixel in the fluorescence image if that pixel has an intensity below a threshold value.
  • the threshold value may be adjustable by a user and stored in a memory section of the image processing unit.
  • pixels in the fluorescent-light image having an intensity lower than f min are set to zero.
  • the threshold filter eliminates dark noise and crosstalk. Using the threshold filter or the threshold adjustment module, respectively, modifies the factor h which becomes
  • the image processing unit may comprise a gamma correction module which is adapted to apply a gamma correction to at least one of the fluorescence image data and the visible image data.
  • the gamma correction is applied to the fluorescent-light data in order to make the fluorescent-light image appear to be a black-and-white fluorescent-light image.
  • the processing unit may comprise a homogenization module, which is configured to compensate at least one of vignetting and inhomogeneous illumination in at least one of the visible image data and the fluorescence image data.
  • the homogenization module may apply a flat-field filter to the visible image data and/or the fluorescence image data.
  • the homogenization filter is adapted to correct the inhomogeneities in illumination and/or vignetting by the imaging optics. After application of the homogenization filter, the resulting intermediate image has the same brightness throughout the whole field of view.
  • the working distance, the focal length of the microscope or endoscope optics, the focus and the illumination intensity settings may all be considered.
  • the homogenization filter may be obtained by calibration, e.g. by using as the object a uniformly colored, e.g. grey or white, target from which a calibration image is obtained.
  • the uniform white target reveals both vignetting and illumination inhomogeneities.
  • the homogenization filter may be determined from the calibration image by measuring the RGB values preferably separately along a spatial profile in the visible-light image and/or fluorescent-light image.
  • the spatial profile may be e.g. a diagonal or any other straight line through the center of the calibration image.
  • the spatial profiles may be fitted with a polynomial and may be rotated around the center of the image to create a 2-dimensional inhomogeneity map which, by inversion, results in the homogenization filter.
  • the homogenization filter may be stored in a memory section of the image processing unit.
  • the homogenization module may comprise different homogenization filters for the visible image data and for the fluorescence image data to account for different optics and illumination in the visible-light image on one hand, and of the fluorescent-light image on the other hand.
  • the visible-light image is recorded by a different camera than the fluorescent-light image, two different viewing angles or alignment errors may occur for the visible-light and the fluorescent-light images.
  • the fluorescent-light image and the visible-light image may thus not exactly be congruent to each other.
  • the visible-light image may be slightly rotated with respect to the fluorescent-light image.
  • the fluorescent-light image may be shifted with respect to the visible-light image.
  • the visible-light image and the fluorescent-light image may be anomorphic with respect to each other.
  • the processing unit comprises a spatial adjustment module, which is adapted to bring the visible-light image and the fluorescent-light image into congruence to each other.
  • the spatial adjustment module ensures that the first input pixel and the second input pixel map the same spot on the object.
  • the spatial adjustment module may be adapted to at least one of crop, rotate, shift and stretch at least one of the visible-light image and the fluorescent-light image.
  • a separate camera for each fluorescent emission color may be used. If a single camera for recording a fluorescent-light image comprising more than one fluorescence emission color is used, it is preferred that a color-sensitive camera or recording method is used to be able to distinguish the different fluorescence bands from each other. Band-pass filters may be used, which band-pass filters only let pass the emission wavelengths of the different fluorophores.
  • modules of the image processing unit are, at least partly, functional units incorporated in software executed by the image processing unit. Cumulatively or alternatively, the modules may also be at least partly represented in hardware.
  • FIG. 1 shows a schematic representation of a medical visualization apparatus according to the invention
  • FIG. 2 shows a schematic representation of an image processing unit of a medical visualization apparatus according to the invention
  • FIG. 3A shows a schematic representation of a homogeneously colored object to calibrate a homogenization filter
  • FIG. 3B shows a schematic representation of a homogenization filter
  • FIG. 4 shows a schematic representation of the effects of image processing on image data
  • FIG. 5 shows a schematic representation of the assignment of pseudocolor in an image
  • FIG. 6 shows a schematic representation of a pseudocolor image resulting from the use of more than one fluorophore.
  • FIG. 1 the medical inspection apparatus 1 is shown as a microscope solely for explanatory purposes.
  • the microscope 1 is used to visually inspect an object 2 , such as tissue of a body of a human or animal e.g. for preparing for surgery or during surgery.
  • the object 2 is illuminated by a lighting subsystem 3 comprising at least one light source 4 .
  • the light 5 from the light source 4 may be transmitted through the object 2 or be reflected by the object 2 .
  • a fluorophore 6 i.e. a fluorescent fluid, solid, or suspension, may be present in the object 2 .
  • the light source may emit light 5 containing energy in a band of wavelength, which excites the fluorescence of the at least one fluorophore 6 .
  • the lighting subsystem 3 may comprise one or more illumination filters 7 through which the light 5 from the at least one light source 4 is directed.
  • the illumination filters 7 may comprise a band-pass filter, which allows light to pass only in the excitation band of the at least one fluorophore and in the visible-light range.
  • the at least one illumination filter 7 may block any light from the light source 4 at those wavelengths, at which the at least one fluorophore emits fluorescent light.
  • the illumination filters may also serve to homogenize the illumination, and may include apertures.
  • the light 8 reflected and/or emitted from the object 2 is received by an optical subsystem 9 , such as a magnifying zoom lens.
  • the light from the optical subsystem 9 is passed to an imaging subsystem 10 , which is adapted to extract visible image data 11 and fluorescence image data 12 in the form of electric signals, from the light 8 reflected and/or emitted from the object 2 and any fluorescent material at or in the object 2 .
  • the visible image data 11 are representative of a visible-light image of the object 2 , i.e. a digital image which corresponds to what can be seen by the eyes of a human observer.
  • the fluorescence image data 12 are representative of a fluorescent-light image.
  • the fluorescent-light image corresponds to a digital image of the object in the emission wavelengths of the at least one fluorophore 6 in the object 2 .
  • both the excitation band and the emission band for the at least one fluorophore is not in the visible light range.
  • both the emission band and the excitation band can be in the near infrared (NIR).
  • Suitable fluorophores may be 5-aminolevulinic acid which, in a metabolism, results in protoporphyrin IX, which is fluorescent, and indocyanine green.
  • the imaging subsystem 10 comprises a dichoroic or polychroic beam splitter 13 which separates the incoming light 8 into visible light 14 and NIR light 15 , the latter containing both the excitation wavelengths reflected by the object 2 and the emitted wavelengths from the at least one fluorophore in the object 2 .
  • the imaging subsystem 10 contains a visible-light imaging assembly 16 and a fluorescent-light imaging assembly 17 in which the visible light 14 and NIR light 15 are treated differently, both optically and on the signal level, until the visible image data 11 and the fluorescence image data 12 are combined in an image processing unit 18 of the microscope 1 to a pseudocolor image, which is represented by output data 19 available at the image processing unit 18 .
  • one or more visible observation filters 20 may be arranged which block all but the visible light. Further, the visible observation filter 20 may comprise an optical homogenization filter for rendering the intensity in the field of view 21 observed by the optical subsystem 8 more homogeneous.
  • the visible light 14 is recorded by a visible-light camera 22 and converted to the visible image data 11 .
  • the NIR light 15 is filtered by a fluorescence observation filter 23 and then directed to a fluorescence camera 24 , which may be an NIR camera.
  • the fluorescence observation filter 23 may be configured as a band-pass filter which blocks all but the light in the emission wavelengths of the at least one fluorophore 6 .
  • the NIR camera 24 records images containing information only in the emission wavelengths.
  • the NIR camera may be a black-and-white camera or may be color-sensitive. The latter is particularly useful if more than one fluorophore used as the excitation wavelengths of the various fluorophores can be discerned by their different color in the fluorescent-light image.
  • the fluorescence observation filter may be a multiple band-pass filter for allowing the different fluorescence wavelengths through.
  • the imaging subsystem 10 may comprise a data interface 25 , which makes the visible image data 11 from the visible-light camera 22 and the fluorescence image data 12 from the fluorescence camera 24 available to other subsystems of the microscope 1 .
  • the imaging subsystem 10 operates in real-time by providing the visible image data 11 and the fluorescence image data 12 with no or almost no delay as compared to the optical image received by the optical subsystem 9 .
  • the data interface 25 of the imaging subsystem 10 may provide the visible image data 11 and the fluorescence image data 12 in a standard data format for a video stream. Further, the data interface 25 of the fluorescent imaging subsystem 10 may be configured to receive control signals 26 e.g. to control camera settings. Furthermore, the imaging subsystem may be configured to change settings of at least one of the visible observation filter 20 and the fluorescence observation filter 23 , if at least one of the visible observation filter 20 and the fluorescence observation filter 23 is adjustable.
  • the microscope 1 may be a stereoscopic microscope.
  • an imaging subsystem 10 may be present for each stereoscopic channel.
  • a control and processing subsystem 27 is connected for one- or bi-directional data transfer to the fluorescent imaging subsystem 10 e.g. to receive in operation the visible image data 11 and the fluorescence image data 12 and to exchange control signals 26 .
  • control and processing subsystem 27 may be configured to control the optical subsystem 9 via control signals 26 and/or the lighting subsystem 3 , also via control signals 26 . If the illumination filters are adjustable, the control and the processing subsystem 27 may be configured to also control the illumination filters 7 .
  • Control and processing subsystem 27 may be a general-purpose computer, such as a personal computer, or a dedicated electronic system which has been specifically adapted to the requirements of the microscope 1 .
  • the data transfer between the various subsystem, assemblies and other devices of the microscope 1 may be facilitated if a digital communication bus is used.
  • the control and processing subsystem 27 may comprise several units that may be realized in hardware and/or software.
  • a controller unit 30 may be used to store, alter, and control the setting of operative parameters of the microscope 1 .
  • the operational parameters may include but not be limited to parameters of the optical subsystem 9 , such as an aperture, focus and focal length, parameters of the lighting subsystem 3 such as illumination filter settings, brightness of the light source, parameters of the fluorescent imaging subsystem 10 , such as camera settings and settings of the observation filters, and parameters of the image processing unit 18 .
  • the controller unit 30 may comprise elements for user interaction which, upon operation, change the operational parameters. Such elements may include a graphical user interface on a screen or a touchscreen, and/or mechanical elements such as sliders, push buttons, switches and/or knobs.
  • the image processing unit 18 comprises a first input section 31 , which is configured to receive the visible image data 11 .
  • a second input section 32 of the image processing unit 18 is configured to receive the fluorescence image data 12 .
  • the output data 19 are provided at an output section 33 of the image processing unit 18 .
  • the output data 19 are available in the form of pseudocolor image data which represent a pseudocolor image of the object 2 .
  • the visible-light image is merged with the fluorescent-light image providing smooth color transitions from the visible-light image to the fluorescent-light image, whereby the fluorescent-light image is assigned and displayed in a pseudocolor.
  • the color of an output pixel in the pseudocolor image is computed by the image processing unit 18 from the at least one pseudocolor, a color of a first input pixel in the visible-light image and an intensity of a second input pixel in the fluorescent-light image. If more than one fluorophore is used, each fluorophore, or its fluorescence emitting waveband respectively, is assigned a different pseudocolor, preferably by the user, or automatically.
  • the microscope 1 may either be provided with or connected to a documentation subsystem 35 for storing both all or selective image data preferably together with the microscope settings. Further, the microscope 1 may comprise a monitoring subsystem 36 comprising preferably several displays, such as an eyepiece display 37 and microscope monitor 38 . The microscope 1 may also be provided with a display interface 39 which is configured to supply video data to an external monitor (not shown).
  • a documentation subsystem 35 for storing both all or selective image data preferably together with the microscope settings.
  • the microscope 1 may comprise a monitoring subsystem 36 comprising preferably several displays, such as an eyepiece display 37 and microscope monitor 38 .
  • the microscope 1 may also be provided with a display interface 39 which is configured to supply video data to an external monitor (not shown).
  • FIG. 2 shows schematically the structure of the image processing unit 18 .
  • the image processing unit 18 comprises a plurality of modules, which perform different image processing functions on the image data 11 , 12 in real time.
  • the modules of the image processing unit 18 may be implemented in hardware and/or software. Different modules which perform the same function may be e.g. be implemented as identical software routines which are fed with different data.
  • the modules may be executed in parallel or sequentially provided that in a sequential execution, the output is still available in real time.
  • the image processing unit 18 may comprise a homogenization module 41 which is configured to compensate at least one of vignetting and inhomogeneous illumination in at least one of the visible image data 11 and the fluorescence image data 12 .
  • the homogenization module may be further configured to do a histogram normalization and optimization of the image data 12 in order to make full use of the contrast range of the image.
  • the homogenization module 41 may comprise a digital homogenization filter 42 which may be different for the visible image data 11 and the fluorescence image data 12 as the distribution of illumination may be different for visible light and light in the excitation band of the at least one fluorophore. Further, the cameras may exhibit different vignetting and distortion characteristics which makes an individual compensation necessary.
  • the homogenization filter 43 may be determined using calibration for example of a homogeneously colored calibration object, such as a white, grey or otherwise uniformly colored plate and stored electronically in the image processing unit 18 or an attached memory.
  • FIG. 3A shows an image of such a homogeneously colored calibration object 44 .
  • the inhomogeneous illumination and the vignetting are clearly visible in the image of the calibration object as the periphery of the field of view 21 is significantly darker than the center.
  • a homogenization filter 42 as shown in FIG. 3B is applied in real time to at least one of the visual-light image and the fluorescent-light image.
  • the homogenization filter 42 results from the RGB values along a spatial profile 44 from the image of the calibration object: For each coordinate in the color space, a separate profile is obtained.
  • the different profiles may be fitted with polynomials. Rotating the polynomial curves around the center of the image creates a two-dimensional map of the inhomogeneities in the respective optical path between the object 2 and the sensor in the respective camera 22 , 24 .
  • the homogenization filter 42 results from inverting the homogeneity map.
  • the image processing unit 18 may comprise a spatial adjustment module 45 which preferably acts only on one of the fluorescence image data 12 and the visible image data 11 , preferably the fluorescent data 12 only, as the fluorescence image data 12 may be less than the visible image data 11 due to a lower color depth.
  • the spatial adjustment module 47 is adapted to at least one of crop, rotate, shift and stretch at least one of the visible-light image and the fluorescent-light image.
  • the purpose of the spatial adjustment module 45 is to bring the visible-light image and the fluorescent-light image into congruence to each other, so that a pixel at a certain location in the visible image corresponds to the same spot on the object 2 as the pixel at the same location in the fluorescent-light image.
  • correlation algorithms and pattern detection algorithms may be executed to match corresponding structures in the visible-light image and the fluorescent-light image and to compute the amount of cropping, shifting, stretching and/or rotating necessary to align the two images to each other.
  • the image processing unit 18 may comprise a gamma correction module 46 which is configured to act on at least one of the visible image data 11 and the fluorescence image data 12 .
  • a gamma correction module 46 which is configured to act on at least one of the visible image data 11 and the fluorescence image data 12 .
  • the image processing unit 18 may further comprise a threshold adjustment module 47 which is preferably configured to act on the fluorescence image data 12 only.
  • the controller unit 30 ( FIG. 1 ) may be configured to allow adjustment of the threshold value by a user.
  • Blanking a pixel comprises one of setting the color of the pixel to a pre-determined color, such as black, setting it to zero and making the pixel transparent.
  • the image processing unit 18 may comprise a pseudo color image generator 48 , which is adapted to merge the visible-light image and the fluorescent-light image to generate the pseudocolor image available at the output section 33 .
  • the function of the pseudocolor image generator 48 is described with reference to a color space, for example an RGB color space.
  • a Cartesian coordinate system is formed by the three component colors red (r), green (g), and blue (b).
  • Other color spaces which may be alternatively used may be the CMYK color space or the HSL or HSV color space.
  • any color is represented by its three components (r, g, b) and thus corresponds to a certain location in the 3-dimensional color space. This location corresponds to a position vector pointing from the origin of the color space to the specific color.
  • the pseudocolor image generator 48 is configured to linearly interpolate the color of an output pixel in the pseudocolor image from the pseudocolor to the color of the first input pixel in the visible-light image depending on the intensity of the second input pixel.
  • the color (r o , g o , b o ) of the output pixel is located linearly between the color (r i , g i , b i ) of the first input pixel in the visible-light image and the at least one pseudocolor (r p , g p , b p ) i.e.
  • the distance between the color (r o , g o , b o ) of the output pixel and the color (r i , g i , b i ) of the first input pixel is computed to be proportional to the intensity f of the second input pixel in the fluorescent-light image.
  • both the first input pixel and the second input pixel correspond to the same spot on the object 2 ( FIG. 1 ).
  • the color (r o , g o , b o ) of the output pixel can be calculated in the pseudocolor image generator 48 as follows:
  • the opaqueness factor a may be adjusted by the user upon interaction with the control and processing subsystem 27 to increase or decrease the transparency of the pseudocolor. If factor a is close to zero, even highly fluorescent parts of the fluorescent-light image will hardly be visible in the pseudocolor image. Increasing factor a will increase visibility of the pseudocolor
  • the upper square represents a (schematic) visible light image 49 with 4 ⁇ 4 first input pixels 50 . Only for explanatory purposes, the sample visible-light image 49 contains only four colors which are identical throughout every column in the visible-light image.
  • the lower square on the left-hand side shows schematically the intensity in a sample 4 ⁇ 4 fluorescent-light image 51 .
  • the fluorescent-light image consists of 4 ⁇ 4 second input pixels 52 . Only for explanatory purposes, the intensity in each row of the fluorescent-light image 51 is constant. The upmost row of second input pixels 52 has zero intensity, whereas the lowest row of second input pixels 52 in the fluorescent-light image 51 has maximum intensity.
  • FIG. 5 the different steps for merging the visible-light image 49 and a fluorescent-light image 51 to obtain a pseudocolor image 53 are shown.
  • a first step 60 the visible-light image 49 and the fluorescent-light image 51 are sampled by the visible-light camera 22 and the fluorescent-light camera 24 , respectively.
  • a second step 61 the respective images 49 , 51 are homogenized using the homogenization module.
  • a third step 62 the homogenized fluorescent-light image 51 is brought into congruence with the visible-light image so that the physical structures in the two images 49 , 51 correspond to each other both in size and location.
  • the spatial adjustment is preferably done before the fluorescent-light image 51 is worked upon by the threshold adjustment module 47 , so that the algorithms for the spatial adjustment have more data available for pattern matching.
  • a threshold-filtering of the fluorescent-light image 51 takes place to blank all second input pixels 52 in the fluorescent-light image 51 which are below the intensity threshold f min .
  • a fifth step 64 the pseudocolor image 53 is computed using the pseudocolor image generator 48 with the above-described linear color interpolation.
  • FIG. 6 shows the generation of a pseudocolor image 53 containing two pseudocolors 70 , 71 .
  • the two pseudocolors result from the use of two fluorofores in the object 2 which emit light at two different emission bands and thus have two different fluorescent colors 72 , 72 .
  • linear interpolation takes place after a pseudocolor 70 , 71 has been assigned to a second input pixel 52 in the fluorescent-light image 51 based on the fluorescent color 72 , 73 of the second input pixel 52 .
  • the linear interpolation in color space takes place between the pseudocolor assigned to the specific pixel 50 , 52 , 54 and the color of the first input pixel 50 as explained above.
  • the optical subsystem 8 comprises fiber optics in the case of the endoscope as compared to a microscope 1 .

Abstract

The invention relates to a medical inspection apparatus (1), such as a microscope or endoscope, and to a medical inspection method such as microscopy or endoscopy. Visible image data (11) representing a visible-light image (49) and fluorescence image data (12) representing a fluorescent-light image (51) and a pseudocolor (70, 71) are merged to give an improved visual rendition of an object (2) which comprises at least one fluorophore (6) to mark special features of the object (2). This is accomplished in that an image processing unit (18) of the microscope (1) or endoscope is configured to compute a color (ro, go, bo) of an output pixel (54) in the pseudocolor image (53) from at least one pseudocolor (rp, gp, bp), a color (ri, gi, bi) of a first input pixel (50) in the visible-light image (49) and an intensity (f) of a second input pixel (52) in the fluorescent-light image (51). In particular, the color (ro, go, bo) may result from a linear interpolation in a color space (RGB) between the pseudocolor and the color of the first input pixel (50) of the visible-light image (49) depending on the intensity (f) of the second input pixel (52) in the fluorescent-light image.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application claims priority of European patent application number 16155625.3 filed Feb. 15, 2016, the entire disclosure of which is incorporated by reference herein.
  • FIELD OF THE INVENTION
  • The invention relates to a medical inspection apparatus, such as a microscope or endoscope, and to a medical inspection method such as a microscopy or endoscopy.
  • BACKGROUND OF THE INVENTION
  • Microscopes and endoscopes using fluorescence of fluorophores to better discern certain types of tissue are known. The present invention aims to improve the quality of the images displayed to a surgeon as compared to the existing microscopes or endoscopes.
  • SUMMARY OF THE INVENTION
  • For a medical inspection apparatus such as a microscope or endoscope, this objective is achieved in that the microscope or endoscope comprises an image processing unit, the image processing unit comprising a first input section configured to receive visible image data representing a visible-light image of an object, a second input section configured to receive fluorescence image data representing a fluorescent-light image of the object, and an output section configured to output pseudocolor image data representing a pseudocolor image of the object, wherein the image processing unit is adapted to compute a color of an output pixel in the pseudocolor image from at least one pseudocolor, a color of a first input pixel in the visible-light image and an intensity of a second input pixel in the fluorescent-light image.
  • The medical inspection method according to the invention comprises the steps of acquiring visible image data representing a visible-light image of an object, acquiring fluorescence image data representing a fluorescent-light image of the object, combining the visible-light image data, the fluorescence image data and at least one pseudocolor to obtain output data representing a pseudocolor image, wherein a color of an output pixel in the pseudocolor image depends on at least one pseudocolor, a color of a first input pixel in the visible-light image, and an intensity of a second input pixel in the fluorescent-light image.
  • Further, the object is solved by a non-transitory computer readable medium storing a program, causing a medical visualization apparatus to execute the above-cited method.
  • The solution according to the invention creates a pseudocolor image in which the at least one pseudocolor is blended in a natural-looking way into the visible-light image. Thus, the section of the object, which is colored by the fluorescent color, is presented in a more natural way. The fluorescent-light image is not simply overlaid onto the visible-light image, but is rather merged with the visible-light image using the at least one pseudocolor to display the fluorescent parts of the object.
  • The invention can be further improved by the following aspects, which are each advantageous on their own, and can be combined arbitrarily.
  • In as far as the terms “image” and “data” or operations upon images or data are mentioned in the following, it is to be understood that the terms “image” and “data” can be used synonymously, as the image represents the totality of the data making up the image. The data of an image may also be grouped into pixels. Thus, a “pixel” corresponds to an ordered grouping of the “data”. The images and data are processed and/or stored digitally and are themselves digital.
  • The at least one pseudocolor need not correspond to the color of fluorescence of the fluorophore, but can be chosen by the user of the microscope or endoscope. Preferably, the pseudocolor is a color in the visible-light range which does not naturally occur in the object to be observed.
  • According to another aspect of the invention, the medical visualization apparatus is stereoscopic. In a stereoscopic microscope or endoscope, a separate pseudocolor image recorded from a slightly different point of view, is presented to each eye of the observer. Thus, in a stereoscopic microscope or endoscope, two separate pseudocolor images are computed from two different visible-light images and two fluorescent-light images. For each of the two stereoscopic channels, the same devices and processes may be used as for the single channel of a monocular microscope.
  • The fluorescent-light image may be sampled by a digital NIR camera, the visible-light image by a standard camera. The image processing unit may comprise a memory in which image data are stored. The image processing unit may be part of a general-purpose computer which has been adapted, e.g. by software, to provide a control and processing system for a microscope or endoscope. It is also possible that the image processing unit is a specially-designed electronic device that can be solely used in microscopes or endoscopes.
  • According to another aspect of the invention, the microscope or endoscope may comprise more than one fluorescent channel. For example, two or more fluorophores may be used which are each excited at different excitation wavelengths and which each emit fluorescent light at different emission wavelengths. Two or more fluorophores can be used to identify different aspects of the object under investigation. For example, one type of fluorophore can be used to mark blood vessels whereas another fluorophore may mark tumourous cells. Another fluorophore may be used to mark e.g. nerve cells.
  • According to one aspect of the invention, the fluorescence image data may contain at least two different fluorescent emission colors and the processing unit may be configured to assign a different pseudocolor to each of the different fluorescent emission colors. Thus, the different types of tissues marked by the different types of fluorophores can be discerned in the pseudocolor image by the different pseudocolors. According to the invention, the different pseudocolors are all blended naturally into the visible-light image due to the consideration of both the color of the first input pixel and the intensity of the second input pixel in the fluorescent-light image.
  • For the computation of the color of the output pixel, the first and second input pixel should correspond to the same feature of the object. For example, the fluorescent-light image, the visible-light image and the pseudocolor image may have the same size, i.e. comprise the same number of pixel rows and pixel columns. If the visible-light and fluorescent-light images are congruent to each other, i.e. the size, orientation and location of the object is the same in the two images, the first and second input pixel are located in the same row and column in their respective images.
  • To improve the blending, the color of the output pixel may be located linearly, in a color space, between the color of the first input pixel and the at least one pseudocolor, the distance between the color of the output pixel and the color of the first input pixel being proportional to the intensity of the second input pixel. Typical color spaces are HSV, HSI, RGB and YMCK without being limited to these color spaces. In the color space, each color can be represented by a positional vector. Thus, computing the color of the output pixel can be done, computationally efficient by vector arithmetics.
  • Using an RGB color space as an example, the image processing unit may be configured to linearly interpolate the RGB values of the first input pixel in the visible-light to the selected pseudocolor value dependent on the intensity of the second input pixel to obtain the color of the output pixel. Assuming that the first input pixel has RGB values (ri, gi, bi), and that the fluorescence intensity is f, an interpolation factor (h=f/fmax, where fmax is the maximum expected fluorescence intensity) can be chosen. If the pseudocolor that has been selected by the user for a particular fluorophore has RGB values (rp, gp, bp), the color (ro, go, bo) of the output pixel may be as follows:

  • r o =r i +h·(r p −r i)

  • g o =g i +h·(g p −g i)

  • b o =b i +h·(b p −b i).
  • Thus, if the fluorescence intensity is f=0, then h=0 and the color of the output pixel will correspond to the color of the visible-light image. The more intense the pixel in the fluorescent-light image is, the closer the color of the output pixel will be to the pseudocolor.
  • In the above equation, the factor h may be replaced with a factor k=a·h, where a is a pseudocolor opaqueness factor which determines how opaque the pseudocolor is in the pseudocolor image. The pseudo-color opaqueness factor a may be selected by the user. Setting the opaqueness factor e.g. to very small values below 1 will lead to a very weak pseudocolor in the pseudocolor image, even if f=fmax.
  • To improve the quality of the pseudocolor image and to avoid noise artifacts, the processing unit may comprise a threshold adjustment module, which is configured to blank a pixel in the fluorescence image if that pixel has an intensity below a threshold value. The threshold value may be adjustable by a user and stored in a memory section of the image processing unit. In operation of the medical visualization apparatus, pixels in the fluorescent-light image having an intensity lower than fmin are set to zero. The threshold filter eliminates dark noise and crosstalk. Using the threshold filter or the threshold adjustment module, respectively, modifies the factor h which becomes

  • h=0 for f<f min, and

  • h=f for f>f min.
  • Furthermore, the image processing unit may comprise a gamma correction module which is adapted to apply a gamma correction to at least one of the fluorescence image data and the visible image data. In particular, the gamma correction is applied to the fluorescent-light data in order to make the fluorescent-light image appear to be a black-and-white fluorescent-light image.
  • According to another aspect of the invention, the processing unit may comprise a homogenization module, which is configured to compensate at least one of vignetting and inhomogeneous illumination in at least one of the visible image data and the fluorescence image data. For example, the homogenization module may apply a flat-field filter to the visible image data and/or the fluorescence image data.
  • Inhomogeneities in illumination and vignetting of the imaging optics result in uneven brightness across the field of view. Typically, the periphery of the field of view is significantly darker in microscopes or endoscopes than the center and therefore only the central part of the field of view is typically usable for digital imaging in microscopy or endoscopy. However, the surgeon can still observe visually, via the eyepieces of the microscope or endoscope, the whole field of view. Thus, there can be a mismatch of the fields of view as used in the pseudocolor image on one hand, and as observed by the surgeon on the other hand. To avoid this mismatch, a digital homogenization filter may be applied to the visible image data and/or the fluorescence image data. The homogenization filter is adapted to correct the inhomogeneities in illumination and/or vignetting by the imaging optics. After application of the homogenization filter, the resulting intermediate image has the same brightness throughout the whole field of view. In the homogenization filter, the working distance, the focal length of the microscope or endoscope optics, the focus and the illumination intensity settings may all be considered.
  • In particular, the homogenization filter may be obtained by calibration, e.g. by using as the object a uniformly colored, e.g. grey or white, target from which a calibration image is obtained. The uniform white target reveals both vignetting and illumination inhomogeneities.
  • The homogenization filter may be determined from the calibration image by measuring the RGB values preferably separately along a spatial profile in the visible-light image and/or fluorescent-light image. The spatial profile may be e.g. a diagonal or any other straight line through the center of the calibration image. The spatial profiles may be fitted with a polynomial and may be rotated around the center of the image to create a 2-dimensional inhomogeneity map which, by inversion, results in the homogenization filter.
  • The homogenization filter may be stored in a memory section of the image processing unit. The homogenization module may comprise different homogenization filters for the visible image data and for the fluorescence image data to account for different optics and illumination in the visible-light image on one hand, and of the fluorescent-light image on the other hand.
  • If the visible-light image is recorded by a different camera than the fluorescent-light image, two different viewing angles or alignment errors may occur for the visible-light and the fluorescent-light images. The fluorescent-light image and the visible-light image may thus not exactly be congruent to each other. For example, the visible-light image may be slightly rotated with respect to the fluorescent-light image. Also, the fluorescent-light image may be shifted with respect to the visible-light image. Finally, the visible-light image and the fluorescent-light image may be anomorphic with respect to each other.
  • Thus, in order to blend the visible-light image and the fluorescent-light image naturally into each other, it is advantageous if the processing unit comprises a spatial adjustment module, which is adapted to bring the visible-light image and the fluorescent-light image into congruence to each other. The spatial adjustment module ensures that the first input pixel and the second input pixel map the same spot on the object. In particular, the spatial adjustment module may be adapted to at least one of crop, rotate, shift and stretch at least one of the visible-light image and the fluorescent-light image.
  • If more than one fluorescence channel is used, a separate camera for each fluorescent emission color may be used. If a single camera for recording a fluorescent-light image comprising more than one fluorescence emission color is used, it is preferred that a color-sensitive camera or recording method is used to be able to distinguish the different fluorescence bands from each other. Band-pass filters may be used, which band-pass filters only let pass the emission wavelengths of the different fluorophores.
  • Although several aspects of the microscope or endoscope have been described above with reference to modules of the image processing unit, it is possible that these modules are, at least partly, functional units incorporated in software executed by the image processing unit. Cumulatively or alternatively, the modules may also be at least partly represented in hardware.
  • In the following, exemplary embodiments of the invention are described with reference to the drawings. As explained above, the combination of features shown in these embodiments can be changed according to the needs of the specific application. For example, if, in a specific application, the technical effect of a particular feature of the embodiment shown is not needed, this feature can be omitted. Conversely, if, for a specific embodiment, the technical effect of one of the above-described features is needed, this feature may be added to the shown embodiment.
  • BRIEF DESCRIPTION OF THE DRAWING VIEWS
  • The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
  • Throughout the drawings, the same reference numerals are used for elements which correspond to each other in at least one of their structure and function.
  • In the drawings:
  • FIG. 1 shows a schematic representation of a medical visualization apparatus according to the invention;
  • FIG. 2 shows a schematic representation of an image processing unit of a medical visualization apparatus according to the invention;
  • FIG. 3A shows a schematic representation of a homogeneously colored object to calibrate a homogenization filter;
  • FIG. 3B shows a schematic representation of a homogenization filter;
  • FIG. 4 shows a schematic representation of the effects of image processing on image data;
  • FIG. 5 shows a schematic representation of the assignment of pseudocolor in an image;
  • FIG. 6 shows a schematic representation of a pseudocolor image resulting from the use of more than one fluorophore.
  • DETAILED DESCRIPTION OF THE INVENTION
  • First, the general structure of a medical inspection apparatus 1 according to the invention is described with reference to FIG. 1. In FIG. 1, the medical inspection apparatus 1 is shown as a microscope solely for explanatory purposes.
  • The microscope 1 is used to visually inspect an object 2, such as tissue of a body of a human or animal e.g. for preparing for surgery or during surgery. For this, the object 2 is illuminated by a lighting subsystem 3 comprising at least one light source 4. The light 5 from the light source 4 may be transmitted through the object 2 or be reflected by the object 2. A fluorophore 6, i.e. a fluorescent fluid, solid, or suspension, may be present in the object 2. The light source may emit light 5 containing energy in a band of wavelength, which excites the fluorescence of the at least one fluorophore 6.
  • The lighting subsystem 3 may comprise one or more illumination filters 7 through which the light 5 from the at least one light source 4 is directed. For example, the illumination filters 7 may comprise a band-pass filter, which allows light to pass only in the excitation band of the at least one fluorophore and in the visible-light range. In particular, the at least one illumination filter 7 may block any light from the light source 4 at those wavelengths, at which the at least one fluorophore emits fluorescent light.
  • Additionally or alternatively, the illumination filters may also serve to homogenize the illumination, and may include apertures.
  • The light 8 reflected and/or emitted from the object 2 is received by an optical subsystem 9, such as a magnifying zoom lens.
  • The light from the optical subsystem 9 is passed to an imaging subsystem 10, which is adapted to extract visible image data 11 and fluorescence image data 12 in the form of electric signals, from the light 8 reflected and/or emitted from the object 2 and any fluorescent material at or in the object 2.
  • The visible image data 11 are representative of a visible-light image of the object 2, i.e. a digital image which corresponds to what can be seen by the eyes of a human observer. The fluorescence image data 12 are representative of a fluorescent-light image. The fluorescent-light image corresponds to a digital image of the object in the emission wavelengths of the at least one fluorophore 6 in the object 2.
  • In order to be able to use the full spectrum of visible light in the visible image data 11, it is preferred that both the excitation band and the emission band for the at least one fluorophore is not in the visible light range. For example, both the emission band and the excitation band can be in the near infrared (NIR). Suitable fluorophores may be 5-aminolevulinic acid which, in a metabolism, results in protoporphyrin IX, which is fluorescent, and indocyanine green.
  • The imaging subsystem 10 comprises a dichoroic or polychroic beam splitter 13 which separates the incoming light 8 into visible light 14 and NIR light 15, the latter containing both the excitation wavelengths reflected by the object 2 and the emitted wavelengths from the at least one fluorophore in the object 2. The imaging subsystem 10 contains a visible-light imaging assembly 16 and a fluorescent-light imaging assembly 17 in which the visible light 14 and NIR light 15 are treated differently, both optically and on the signal level, until the visible image data 11 and the fluorescence image data 12 are combined in an image processing unit 18 of the microscope 1 to a pseudocolor image, which is represented by output data 19 available at the image processing unit 18.
  • In the visible-light imaging assembly 16, one or more visible observation filters 20 may be arranged which block all but the visible light. Further, the visible observation filter 20 may comprise an optical homogenization filter for rendering the intensity in the field of view 21 observed by the optical subsystem 8 more homogeneous. The visible light 14 is recorded by a visible-light camera 22 and converted to the visible image data 11.
  • To obtain the fluorescence image data 12, the NIR light 15 is filtered by a fluorescence observation filter 23 and then directed to a fluorescence camera 24, which may be an NIR camera. The fluorescence observation filter 23 may be configured as a band-pass filter which blocks all but the light in the emission wavelengths of the at least one fluorophore 6. Thus, the NIR camera 24 records images containing information only in the emission wavelengths. The NIR camera may be a black-and-white camera or may be color-sensitive. The latter is particularly useful if more than one fluorophore used as the excitation wavelengths of the various fluorophores can be discerned by their different color in the fluorescent-light image. In this case, the fluorescence observation filter may be a multiple band-pass filter for allowing the different fluorescence wavelengths through.
  • The imaging subsystem 10 may comprise a data interface 25, which makes the visible image data 11 from the visible-light camera 22 and the fluorescence image data 12 from the fluorescence camera 24 available to other subsystems of the microscope 1.
  • The imaging subsystem 10 operates in real-time by providing the visible image data 11 and the fluorescence image data 12 with no or almost no delay as compared to the optical image received by the optical subsystem 9.
  • The data interface 25 of the imaging subsystem 10 may provide the visible image data 11 and the fluorescence image data 12 in a standard data format for a video stream. Further, the data interface 25 of the fluorescent imaging subsystem 10 may be configured to receive control signals 26 e.g. to control camera settings. Furthermore, the imaging subsystem may be configured to change settings of at least one of the visible observation filter 20 and the fluorescence observation filter 23, if at least one of the visible observation filter 20 and the fluorescence observation filter 23 is adjustable.
  • The microscope 1 may be a stereoscopic microscope. In this case, an imaging subsystem 10 may be present for each stereoscopic channel.
  • In the embodiment of FIG. 1, a control and processing subsystem 27 is connected for one- or bi-directional data transfer to the fluorescent imaging subsystem 10 e.g. to receive in operation the visible image data 11 and the fluorescence image data 12 and to exchange control signals 26.
  • Further, the control and processing subsystem 27 may be configured to control the optical subsystem 9 via control signals 26 and/or the lighting subsystem 3, also via control signals 26. If the illumination filters are adjustable, the control and the processing subsystem 27 may be configured to also control the illumination filters 7.
  • Control and processing subsystem 27 may be a general-purpose computer, such as a personal computer, or a dedicated electronic system which has been specifically adapted to the requirements of the microscope 1. The data transfer between the various subsystem, assemblies and other devices of the microscope 1 may be facilitated if a digital communication bus is used.
  • The control and processing subsystem 27 may comprise several units that may be realized in hardware and/or software.
  • For example, a controller unit 30 may be used to store, alter, and control the setting of operative parameters of the microscope 1. The operational parameters may include but not be limited to parameters of the optical subsystem 9, such as an aperture, focus and focal length, parameters of the lighting subsystem 3 such as illumination filter settings, brightness of the light source, parameters of the fluorescent imaging subsystem 10, such as camera settings and settings of the observation filters, and parameters of the image processing unit 18. The controller unit 30 may comprise elements for user interaction which, upon operation, change the operational parameters. Such elements may include a graphical user interface on a screen or a touchscreen, and/or mechanical elements such as sliders, push buttons, switches and/or knobs.
  • The image processing unit 18 comprises a first input section 31, which is configured to receive the visible image data 11. A second input section 32 of the image processing unit 18 is configured to receive the fluorescence image data 12. The output data 19 are provided at an output section 33 of the image processing unit 18.
  • At the output section 33, the output data 19 are available in the form of pseudocolor image data which represent a pseudocolor image of the object 2.
  • In the pseudocolor image, the visible-light image is merged with the fluorescent-light image providing smooth color transitions from the visible-light image to the fluorescent-light image, whereby the fluorescent-light image is assigned and displayed in a pseudocolor. The color of an output pixel in the pseudocolor image is computed by the image processing unit 18 from the at least one pseudocolor, a color of a first input pixel in the visible-light image and an intensity of a second input pixel in the fluorescent-light image. If more than one fluorophore is used, each fluorophore, or its fluorescence emitting waveband respectively, is assigned a different pseudocolor, preferably by the user, or automatically.
  • As can be further seen in FIG. 1, the microscope 1 may either be provided with or connected to a documentation subsystem 35 for storing both all or selective image data preferably together with the microscope settings. Further, the microscope 1 may comprise a monitoring subsystem 36 comprising preferably several displays, such as an eyepiece display 37 and microscope monitor 38. The microscope 1 may also be provided with a display interface 39 which is configured to supply video data to an external monitor (not shown).
  • FIG. 2 shows schematically the structure of the image processing unit 18. The image processing unit 18 comprises a plurality of modules, which perform different image processing functions on the image data 11, 12 in real time. The modules of the image processing unit 18 may be implemented in hardware and/or software. Different modules which perform the same function may be e.g. be implemented as identical software routines which are fed with different data. The modules may be executed in parallel or sequentially provided that in a sequential execution, the output is still available in real time.
  • The image processing unit 18 may comprise a homogenization module 41 which is configured to compensate at least one of vignetting and inhomogeneous illumination in at least one of the visible image data 11 and the fluorescence image data 12. The homogenization module may be further configured to do a histogram normalization and optimization of the image data 12 in order to make full use of the contrast range of the image.
  • The homogenization module 41 may comprise a digital homogenization filter 42 which may be different for the visible image data 11 and the fluorescence image data 12 as the distribution of illumination may be different for visible light and light in the excitation band of the at least one fluorophore. Further, the cameras may exhibit different vignetting and distortion characteristics which makes an individual compensation necessary.
  • The homogenization filter 43 may be determined using calibration for example of a homogeneously colored calibration object, such as a white, grey or otherwise uniformly colored plate and stored electronically in the image processing unit 18 or an attached memory. FIG. 3A shows an image of such a homogeneously colored calibration object 44. The inhomogeneous illumination and the vignetting are clearly visible in the image of the calibration object as the periphery of the field of view 21 is significantly darker than the center. In the homogenization module 41, a homogenization filter 42 as shown in FIG. 3B is applied in real time to at least one of the visual-light image and the fluorescent-light image. The homogenization filter 42 results from the RGB values along a spatial profile 44 from the image of the calibration object: For each coordinate in the color space, a separate profile is obtained. The different profiles may be fitted with polynomials. Rotating the polynomial curves around the center of the image creates a two-dimensional map of the inhomogeneities in the respective optical path between the object 2 and the sensor in the respective camera 22, 24. The homogenization filter 42 results from inverting the homogeneity map.
  • Further, the image processing unit 18 may comprise a spatial adjustment module 45 which preferably acts only on one of the fluorescence image data 12 and the visible image data 11, preferably the fluorescent data 12 only, as the fluorescence image data 12 may be less than the visible image data 11 due to a lower color depth. The spatial adjustment module 47 is adapted to at least one of crop, rotate, shift and stretch at least one of the visible-light image and the fluorescent-light image. The purpose of the spatial adjustment module 45 is to bring the visible-light image and the fluorescent-light image into congruence to each other, so that a pixel at a certain location in the visible image corresponds to the same spot on the object 2 as the pixel at the same location in the fluorescent-light image. In the spatial alignment module 45, correlation algorithms and pattern detection algorithms may be executed to match corresponding structures in the visible-light image and the fluorescent-light image and to compute the amount of cropping, shifting, stretching and/or rotating necessary to align the two images to each other.
  • Further, the image processing unit 18 may comprise a gamma correction module 46 which is configured to act on at least one of the visible image data 11 and the fluorescence image data 12. By using the gamma correction, the images can be adapted to human vision.
  • The image processing unit 18 may further comprise a threshold adjustment module 47 which is preferably configured to act on the fluorescence image data 12 only. The threshold adjustment module 47 is configured to blank a pixel in the fluorescence image data 12 if this pixel has an intensity f below a threshold value fmin: f=f, if f>fmin, and f=0, if f<fmin. The controller unit 30 (FIG. 1) may be configured to allow adjustment of the threshold value by a user.
  • Blanking a pixel comprises one of setting the color of the pixel to a pre-determined color, such as black, setting it to zero and making the pixel transparent.
  • Finally, the image processing unit 18 may comprise a pseudo color image generator 48, which is adapted to merge the visible-light image and the fluorescent-light image to generate the pseudocolor image available at the output section 33.
  • In the following, the function of the pseudocolor image generator 48 is described with reference to a color space, for example an RGB color space. In the RGB color space, a Cartesian coordinate system is formed by the three component colors red (r), green (g), and blue (b). Other color spaces which may be alternatively used may be the CMYK color space or the HSL or HSV color space.
  • In RGB color space, any color is represented by its three components (r, g, b) and thus corresponds to a certain location in the 3-dimensional color space. This location corresponds to a position vector pointing from the origin of the color space to the specific color.
  • The pseudocolor image generator 48 is configured to linearly interpolate the color of an output pixel in the pseudocolor image from the pseudocolor to the color of the first input pixel in the visible-light image depending on the intensity of the second input pixel. Thus, in the color space, the color (ro, go, bo) of the output pixel is located linearly between the color (ri, gi, bi) of the first input pixel in the visible-light image and the at least one pseudocolor (rp, gp, bp) i.e. located on a vector pointing from (ri, gi, bi) to (rp, gp, bp). The distance between the color (ro, go, bo) of the output pixel and the color (ri, gi, bi) of the first input pixel is computed to be proportional to the intensity f of the second input pixel in the fluorescent-light image. Thereby, both the first input pixel and the second input pixel correspond to the same spot on the object 2 (FIG. 1). Using the color space allows to do the linear interpolation using computationally efficient vector arithmetics.
  • In particular, the color (ro, go, bo) of the output pixel can be calculated in the pseudocolor image generator 48 as follows:

  • r o =r i +h·(r p −r i)

  • g o =g i +h·(g p −g i)

  • b o =b i +h·(b p −b i),
  • where the factor h=f/fmax, fmax being the maximum expected fluorescence intensity.
  • Thus, the intensity of the fluorescence in the second input pixel determines the distance between the output color and the input color for any given color component. If the fluorescence intensity f=0, i.e. there is no fluorescence, the color of the output pixel will correspond to the color of the first input pixel in the visible-light image. If the fluorescence in the second output pixel is maximum, f=fmax, the color of the output pixel will correspond to the pseudocolor (rp, gp, bp).
  • In a further variant, an opaqueness factor a may be used in combination with the factor f/fmax to form an alternative factor h=a·(f/fmax). The opaqueness factor a may be adjusted by the user upon interaction with the control and processing subsystem 27 to increase or decrease the transparency of the pseudocolor. If factor a is close to zero, even highly fluorescent parts of the fluorescent-light image will hardly be visible in the pseudocolor image. Increasing factor a will increase visibility of the pseudocolor
  • The process of assigning a color (ro, go, bo) in the output data based on the intensity in the fluorescence image data and the color (ri, gi, bi) of a corresponding pixel in the visible image data is exemplarily shown in FIG. 4, where green is used as pseudo color (rp, gp, bp)=(0,256,0), for example. The upper square represents a (schematic) visible light image 49 with 4×4 first input pixels 50. Only for explanatory purposes, the sample visible-light image 49 contains only four colors which are identical throughout every column in the visible-light image.
  • The lower square on the left-hand side shows schematically the intensity in a sample 4×4 fluorescent-light image 51. The fluorescent-light image consists of 4×4 second input pixels 52. Only for explanatory purposes, the intensity in each row of the fluorescent-light image 51 is constant. The upmost row of second input pixels 52 has zero intensity, whereas the lowest row of second input pixels 52 in the fluorescent-light image 51 has maximum intensity.
  • Using the above linear RGB interpolation scheme, a 4×4 pseudocolor image 53 results. Again, it can be seen that, if the intensity of the second input pixel 52 is zero, the original color in the visible-light image 49 is reproduced in the corresponding output pixel 54 of pseudocolor image 53. If the intensity of the second input pixel 52 is maximum, the color in the pseudocolor image 53 depends on the opaqueness factor a as explained above.
  • In FIG. 5, the different steps for merging the visible-light image 49 and a fluorescent-light image 51 to obtain a pseudocolor image 53 are shown.
  • In a first step 60, the visible-light image 49 and the fluorescent-light image 51 are sampled by the visible-light camera 22 and the fluorescent-light camera 24, respectively. In a second step 61, the respective images 49, 51 are homogenized using the homogenization module.
  • In a third step 62, the homogenized fluorescent-light image 51 is brought into congruence with the visible-light image so that the physical structures in the two images 49, 51 correspond to each other both in size and location. The spatial adjustment is preferably done before the fluorescent-light image 51 is worked upon by the threshold adjustment module 47, so that the algorithms for the spatial adjustment have more data available for pattern matching.
  • In a fourth step 63, a threshold-filtering of the fluorescent-light image 51 takes place to blank all second input pixels 52 in the fluorescent-light image 51 which are below the intensity threshold fmin.
  • In a fifth step 64, the pseudocolor image 53 is computed using the pseudocolor image generator 48 with the above-described linear color interpolation.
  • FIG. 6 shows the generation of a pseudocolor image 53 containing two pseudocolors 70, 71. The two pseudocolors result from the use of two fluorofores in the object 2 which emit light at two different emission bands and thus have two different fluorescent colors 72, 72. In such a case, linear interpolation takes place after a pseudocolor 70, 71 has been assigned to a second input pixel 52 in the fluorescent-light image 51 based on the fluorescent color 72, 73 of the second input pixel 52. After this assignment, the linear interpolation in color space takes place between the pseudocolor assigned to the specific pixel 50, 52, 54 and the color of the first input pixel 50 as explained above.
  • Although the invention has been described above with reference to a microscope, it can be applied also to an endoscope, the only difference being that the optical subsystem 8 comprises fiber optics in the case of the endoscope as compared to a microscope 1.
  • REFERENCE NUMERALS
    • 1 microscope
    • 2 object
    • 3 lighting subsystem
    • 4 light source
    • 5 light from lighting subsystem
    • 6 fluorophore
    • 7 illumination filter
    • 8 reflected or emitted light from object
    • 9 optical subsystem
    • 10 imaging subsystem
    • 11 visible image data
    • 12 fluorescence image data
    • 13 beam splitter
    • 14 visible light
    • 15 NIR light
    • 16 visible-light imaging assembly
    • 17 fluorescent-light imaging assembly
    • 18 image processing unit
    • 19 output data of image processing unit
    • 20 visible observation filter
    • 21 field of view
    • 22 visible-light camera
    • 23 fluorescence observation filter
    • 24 fluorescent-light camera
    • 25 data interface of fluorescent imaging subsystem
    • 26 control signal
    • 27 control and processing subsystem
    • 30 controller unit
    • 31 first input section
    • 32 second input section
    • 33 output section
    • 35 documentation subsystem
    • 36 monitoring subsystem
    • 37 eyepiece display
    • 38 microscope monitor
    • 39 display interface
    • 41 homogenization module
    • 42 homogenization filter
    • 43 calibration object for obtaining homogenization filter
    • 44 spatial profile
    • 45 spatial adjustment module
    • 46 gamma correction module
    • 47 threshold adjustment module
    • 48 pseudocolor image generator
    • 49 visible-light image
    • 50 first input pixel
    • 51 fluorescent-light image
    • 52 second input pixel
    • 53 pseudocolor image
    • 54 output pixel
    • 60 first step
    • 61 second step
    • 62 third step
    • 63 fourth step
    • 64 fifth step
    • 70 first pseudocolor
    • 71 second pseudocolor
    • 72 first fluorescence color
    • 73 second fluorescence color

Claims (19)

What is claimed is:
1. A medical inspection apparatus (1) comprising an image processing unit (18), the image processing unit (18) comprising a first input section (31) configured to receive visible image data (11) representing a visible-light image (49) of an object (2), a second input section (32) configured to receive fluorescence image data (12) representing a fluorescent-light image (51) of the object (2), and an output section (33) configured to output pseudocolor image data (34) representing a pseudocolor image (53) of the object (2), wherein the image processing unit (18) is adapted to compute a color (ro, go, bo) of an output pixel (54) in the pseudocolor image (53) from at least one pseudocolor (rp, gp, bp), a color (ri, gi, bo) of a first input pixel (50) in the visible-light image (49) and an intensity (f) of a second input pixel (52) in the fluorescence-light image (51).
2. The medical inspection apparatus (1) according to claim 1, wherein the fluorescent-light image (51) contains at least two different fluorescent emission colors (72, 73) and wherein the processing unit (18) is configured to assign a different pseudocolor (70, 71) to each of the different fluorescent emission colors.
3. The medical inspection apparatus (1) according to claim 1, wherein the color (ro, go, bo) of the output pixel (54), in a color space (RGB), is located linearly between the color (ri, gi, bi) of the first input pixel (50) and the at least one pseudocolor (rp, gp, bp), the distance between the color of the output pixel (54) and the color of the first input pixel (50) being proportionate to the intensity (f) of the second input pixel (52) of the fluorescent-light image (51).
4. The medical inspection apparatus (1) according to claim 1, wherein the image processing unit (18) comprises a homogenization module (41), the homogenization module being configured to compensate at least one of vignetting and inhomogeneous illumination in at least one of the visible-light image (49) and the fluorescent-light image (51).
5. The medical inspection apparatus (1) according to claim 4, wherein the homogenization module (41) comprises different homogenization filters (42) for the visible-light image (49) and data and for the fluorescent-light image (51).
6. The medical inspection apparatus (1) according to claim 1, wherein the image processing unit (18) comprises a threshold adjustment module (47) configured to blank the second input pixel (52) in the fluorescent-light image (51) if the second input pixel (52) has an intensity (f) below a threshold value (fmin).
7. The medical inspection apparatus (1) according to claim 1, wherein the processing unit (18) comprises a spatial adjustment module (45), the spatial adjustment module being adapted to bring the visible-light image (49) and the fluorescent-light image (51) into congruence to each other.
8. The medical inspection apparatus (1) according to claim 7, wherein the spatial adjustment module (45) is adapted to at least one of crop, rotate, shift and stretch at least one of the visible-light image (49) and the fluorescent-light image (51).
9. The medical inspection apparatus (1) according to claim 1, wherein the medical inspection apparatus is a microscope.
10. The medical inspection apparatus (1) according to claim 1, wherein the medical inspection apparatus is an endoscope.
11. A medical inspection method, comprising the steps of:
acquiring visible image data (11) representing a visible-light image (49) of an object (2);
acquiring fluorescence image data (12) representing a fluorescent-light image (51) of the object;
combining the visible image data (11), the fluorescent light data (12) and at least one pseudocolor (rp, gp, bp) to obtain pseudocolor image data (34) representing a pseudocolor image (53), wherein a color (ro, go, bo) of an output pixel (54) in the pseudocolor image (53) depends on the at least one pseudocolor (rp, gp, bp), a color (ri, gi, bi) of a first input pixel (50) in the visible-light image (49), and an intensity (f) of a second input pixel (52) in the fluorescent-light image (51).
12. The medical inspection method according to claim 11, further comprising the step of assigning different pseudocolors (70, 71) to different fluorescent colors (72, 73) in the fluorescent-light image (51).
13. The medical inspection method according to claim 11, further comprising the step of bringing the visible-light image (49) and the fluorescent-light image (51) into congruence with each other before combining the visible-light image (49) and the fluorescent-light image (51) to obtain the pseudocolor image (53).
14. The medical inspection method according to claim 13, wherein at least one of the visible-light image (49) and the fluorescent-light image (51) is at least one of cropped, shifted, stretched and rotated for bringing the visible-light image (49) and the fluorescent-light image (51) into congruence with each other.
15. The medical inspection method according to claim 11, wherein the second input pixel (52) is blanked in the fluorescent-light image (51) if the intensity of the second input pixel (52) is below a threshold (fmin).
16. The medical inspection method according to claim 11, wherein the color (ro, go, bo) of the output pixel (54) is linearly interpolated between the pseudocolor (rp, gp, bp) and the color (ri, gi, bi) of the first input pixel (50) in a color space (RGB).
17. The medical inspection method according to claim 11, wherein the medical inspection method is a microscopy method.
18. The medical inspection method according to claim 11, wherein the medical inspection method is an endoscopy method.
19. A non-transitory computer readable medium storing a program causing a medical inspection apparatus (1) to execute a medical inspection method comprising the steps of:
acquiring visible image data (11) representing a visible-light image (49) of an object (2);
acquiring fluorescence image data (12) representing a fluorescent-light image (51) of the object;
combining the visible image data (11), the fluorescent light data (12) and at least one pseudocolor (rp, gp, bp) to obtain pseudocolor image data (34) representing a pseudocolor image (53), wherein a color (ro, go, bo) of an output pixel (54) in the pseudocolor image (53) depends on the at least one pseudocolor (rp, gp, bp), a color (ri, gi, bi) of a first input pixel (50) in the visible-light image (49), and an intensity (f) of a second input pixel (52) in the fluorescent-light image (51).
US15/424,819 2016-02-15 2017-02-04 Medical inspection apparatus, such as a microscope or endoscope using pseudocolors Abandoned US20170237958A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/062,595 US11330237B2 (en) 2016-02-15 2020-10-04 Medical inspection apparatus, such as a microscope or endoscope using pseudocolors

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP16155625.3A EP3205254B1 (en) 2016-02-15 2016-02-15 Medical inspection apparatus, such as a microscope or endoscope, using pseudocolors
EP16155625.3 2016-02-15

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/062,595 Continuation US11330237B2 (en) 2016-02-15 2020-10-04 Medical inspection apparatus, such as a microscope or endoscope using pseudocolors

Publications (1)

Publication Number Publication Date
US20170237958A1 true US20170237958A1 (en) 2017-08-17

Family

ID=55450982

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/424,819 Abandoned US20170237958A1 (en) 2016-02-15 2017-02-04 Medical inspection apparatus, such as a microscope or endoscope using pseudocolors
US17/062,595 Active US11330237B2 (en) 2016-02-15 2020-10-04 Medical inspection apparatus, such as a microscope or endoscope using pseudocolors

Family Applications After (1)

Application Number Title Priority Date Filing Date
US17/062,595 Active US11330237B2 (en) 2016-02-15 2020-10-04 Medical inspection apparatus, such as a microscope or endoscope using pseudocolors

Country Status (4)

Country Link
US (2) US20170237958A1 (en)
EP (1) EP3205254B1 (en)
JP (1) JP6456416B2 (en)
CN (1) CN107137053B (en)

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10694117B2 (en) * 2018-06-07 2020-06-23 Curadel, LLC Masking approach for imaging multi-peak fluorophores by an imaging system
CN111345903A (en) * 2018-12-21 2020-06-30 徕卡仪器(新加坡)有限公司 Image processing apparatus, fluorescence observation apparatus, and method of simulating fluorescence observation apparatus
DE102019101773A1 (en) * 2019-01-24 2020-07-30 Carl Zeiss Meditec Ag Microscopy system and method for operating a microscopy system
US20200397302A1 (en) * 2019-06-20 2020-12-24 Ethicon Llc Fluorescence imaging in a light deficient environment
US20200405152A1 (en) * 2018-05-31 2020-12-31 Panasonic I-Pro Sensing Solutions Co., Ltd. Camera device, image processing method, and camera system
EP3907497A1 (en) * 2020-05-08 2021-11-10 Leica Microsystems CMS GmbH Apparatus and method for displaying and/or printing images of a specimen including a fluorophore
US11276148B2 (en) * 2019-06-20 2022-03-15 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed fluorescence imaging system
US11280737B2 (en) * 2019-06-20 2022-03-22 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed fluorescence imaging system
US11288772B2 (en) * 2019-06-20 2022-03-29 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed fluorescence imaging system
CN114266817A (en) * 2022-03-02 2022-04-01 广东欧谱曼迪科技有限公司 Fluorescent depth image synthesis method and device, electronic equipment and storage medium
US11391937B2 (en) 2018-09-18 2022-07-19 Carl Zeiss Meditec Ag Method and device for determining a property of an object
US11398011B2 (en) 2019-06-20 2022-07-26 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed laser mapping imaging system
US11793399B2 (en) 2019-06-20 2023-10-24 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed hyperspectral imaging system
EP4275578A1 (en) * 2022-05-13 2023-11-15 Leica Instruments (Singapore) Pte Ltd Method, processor, and medical fluorescence observation device using a color-dependent color conversion function
EP4275580A1 (en) * 2022-05-13 2023-11-15 Leica Instruments (Singapore) Pte Ltd Method, processor, and medical fluorescence observation device using two color images to record fluorescence
EP4275577A1 (en) * 2022-05-13 2023-11-15 Leica Instruments (Singapore) Pte Ltd Method, processor, and medical fluorescence observation device using two color images and color cameras for fluorescence and white-light
WO2023218087A1 (en) * 2022-05-13 2023-11-16 Leica Instruments (Singapore) Pte. Ltd. Method, processor, and medical fluorescence observation device using two color images to record fluorescence

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3505059A1 (en) * 2017-12-28 2019-07-03 Leica Instruments (Singapore) Pte. Ltd. Apparatus and method for measuring blood flow direction using a fluorophore
EP3540494B1 (en) 2018-03-16 2022-11-23 Leica Instruments (Singapore) Pte. Ltd. Augmented reality surgical microscope and microscopy method
DE102019101777B4 (en) * 2019-01-24 2023-11-02 Carl Zeiss Meditec Ag Microscopy method
EP3690819B1 (en) * 2019-01-31 2023-09-13 Leica Instruments (Singapore) Pte. Ltd. Image processing device and method, and medical observation device comprising such an image processing device, using an intensity or brightness-dependent variation of a pseudocolor pattern characteristic
WO2020181413A1 (en) * 2019-03-08 2020-09-17 北京数字精准医疗科技有限公司 Endoscopic fluorescence system image processing method and apparatus, and storage medium
US11733503B2 (en) 2019-11-14 2023-08-22 Leica Instruments (Singapore) Pte. Ltd. System and a method for generating output image data and a microscope
EP3889886A1 (en) * 2020-04-01 2021-10-06 Leica Instruments (Singapore) Pte. Ltd. Systems, methods and computer programs for a microscope system and for determining a transformation function
CN111449611B (en) * 2020-04-10 2021-04-09 深圳迈瑞生物医疗电子股份有限公司 Endoscope system and imaging method thereof
CN114298956B (en) * 2022-03-09 2022-06-28 广东欧谱曼迪科技有限公司 Image fusion method of dual-fluorescence endoscope, electronic equipment and device
EP4275579A1 (en) * 2022-05-13 2023-11-15 Leica Instruments (Singapore) Pte. Ltd. Method, processor, and medical observation device using two color images and color cameras for fluorescence and white-light
EP4275590A1 (en) * 2022-05-13 2023-11-15 Leica Instruments (Singapore) Pte Ltd Data processing device and computer-implemented method combining two images and an overlay color using a uniform color space

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050159662A1 (en) * 2004-01-21 2005-07-21 Yoshikazu Imanishi Methods for assessing a physiological state of a mammalian retina
US20050203343A1 (en) * 2004-03-05 2005-09-15 Korea Electrotechnology Research Institute Fluorescent endoscope system having improved image detection module
US20070203413A1 (en) * 2003-09-15 2007-08-30 Beth Israel Deaconess Medical Center Medical Imaging Systems
US20080267490A1 (en) * 2007-04-26 2008-10-30 General Electric Company System and method to improve visibility of an object in an imaged subject
US20110177003A1 (en) * 2008-06-11 2011-07-21 Xiaokun Shu Proteins that fluoresce at infrared wavelengths or generate singlet oxygen upon illumination
US20110205382A1 (en) * 2010-02-23 2011-08-25 Apple Inc. Systems and methods for efficiently coding and processing image data
US20110305405A1 (en) * 2010-06-11 2011-12-15 Fujifilm Corporation Method, apparatus, and program for aligning images
US20160061654A1 (en) * 2014-08-26 2016-03-03 General Electric Company Calibration of microscopy systems
US20160231170A1 (en) * 2015-02-09 2016-08-11 Datacolor Holding Ag Method and apparatus for color measurement of non-solid colors
US20170213344A1 (en) * 2016-01-21 2017-07-27 Molecular Devices, Llc System and Method for Segmentation of Three-Dimensional Microscope Images
US20170370843A1 (en) * 2015-01-23 2017-12-28 The Regents Of The University Of California Facilitating real-time visualization of tissue features derived from optical signals

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003111716A (en) * 2001-10-09 2003-04-15 Fuji Photo Film Co Ltd Standard light source, correction coefficient calculating method and device, and fluorescent image forming method and device
JP2003159210A (en) * 2001-11-29 2003-06-03 Fuji Photo Film Co Ltd Method for displaying fluorescent diagnostic image and display unit thereof
US8169468B2 (en) * 2008-04-26 2012-05-01 Intuitive Surgical Operations, Inc. Augmented stereoscopic visualization for a surgical robot
US8706184B2 (en) * 2009-10-07 2014-04-22 Intuitive Surgical Operations, Inc. Methods and apparatus for displaying enhanced imaging data on a clinical image
JP5802364B2 (en) * 2009-11-13 2015-10-28 オリンパス株式会社 Image processing apparatus, electronic apparatus, endoscope system, and program
US9129371B2 (en) * 2010-06-25 2015-09-08 Cireca Theranostics, Llc Method for analyzing biological specimens by spectral imaging
JP2014198144A (en) * 2013-03-29 2014-10-23 ソニー株式会社 Image processing apparatus, image processing method, information processing program, fluorescence observation system, and fluorescence navigation surgery system
EP3037030A4 (en) 2013-08-23 2017-04-19 Olympus Corporation Fluorescence observation apparatus
CN104299212A (en) * 2014-09-28 2015-01-21 安徽中科医药成像技术科技有限公司 Method and system for fusing white-light image and fluorescence image

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070203413A1 (en) * 2003-09-15 2007-08-30 Beth Israel Deaconess Medical Center Medical Imaging Systems
US20050159662A1 (en) * 2004-01-21 2005-07-21 Yoshikazu Imanishi Methods for assessing a physiological state of a mammalian retina
US20050203343A1 (en) * 2004-03-05 2005-09-15 Korea Electrotechnology Research Institute Fluorescent endoscope system having improved image detection module
US20080267490A1 (en) * 2007-04-26 2008-10-30 General Electric Company System and method to improve visibility of an object in an imaged subject
US20110177003A1 (en) * 2008-06-11 2011-07-21 Xiaokun Shu Proteins that fluoresce at infrared wavelengths or generate singlet oxygen upon illumination
US20110205382A1 (en) * 2010-02-23 2011-08-25 Apple Inc. Systems and methods for efficiently coding and processing image data
US20110305405A1 (en) * 2010-06-11 2011-12-15 Fujifilm Corporation Method, apparatus, and program for aligning images
US20160061654A1 (en) * 2014-08-26 2016-03-03 General Electric Company Calibration of microscopy systems
US20170370843A1 (en) * 2015-01-23 2017-12-28 The Regents Of The University Of California Facilitating real-time visualization of tissue features derived from optical signals
US20160231170A1 (en) * 2015-02-09 2016-08-11 Datacolor Holding Ag Method and apparatus for color measurement of non-solid colors
US20170213344A1 (en) * 2016-01-21 2017-07-27 Molecular Devices, Llc System and Method for Segmentation of Three-Dimensional Microscope Images

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200405152A1 (en) * 2018-05-31 2020-12-31 Panasonic I-Pro Sensing Solutions Co., Ltd. Camera device, image processing method, and camera system
US10694117B2 (en) * 2018-06-07 2020-06-23 Curadel, LLC Masking approach for imaging multi-peak fluorophores by an imaging system
US11391937B2 (en) 2018-09-18 2022-07-19 Carl Zeiss Meditec Ag Method and device for determining a property of an object
US11089278B2 (en) 2018-12-21 2021-08-10 Leica Instruments (Singapore) Pte. Ltd. Image-processing device, fluorescence observation device and method for emulating a first type of fluorescence observation device on a second type of fluorescence observation device
CN111345903A (en) * 2018-12-21 2020-06-30 徕卡仪器(新加坡)有限公司 Image processing apparatus, fluorescence observation apparatus, and method of simulating fluorescence observation apparatus
DE102019101773B4 (en) * 2019-01-24 2021-06-17 Carl Zeiss Meditec Ag Microscopy system and method for operating a microscopy system
DE102019101773A1 (en) * 2019-01-24 2020-07-30 Carl Zeiss Meditec Ag Microscopy system and method for operating a microscopy system
US11835461B2 (en) 2019-01-24 2023-12-05 Carl Zeiss Meditec Ag Microscopy system and method for operating a microscopy system
DE102019101773B9 (en) 2019-01-24 2021-11-25 Carl Zeiss Meditec Ag Microscopy system and method for operating a microscopy system
US11360028B2 (en) 2019-06-20 2022-06-14 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed hyperspectral, fluorescence, and laser mapping imaging system
US11288772B2 (en) * 2019-06-20 2022-03-29 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed fluorescence imaging system
US20200397302A1 (en) * 2019-06-20 2020-12-24 Ethicon Llc Fluorescence imaging in a light deficient environment
US11280737B2 (en) * 2019-06-20 2022-03-22 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed fluorescence imaging system
US11276148B2 (en) * 2019-06-20 2022-03-15 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed fluorescence imaging system
US11398011B2 (en) 2019-06-20 2022-07-26 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed laser mapping imaging system
US11727542B2 (en) 2019-06-20 2023-08-15 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed hyperspectral, fluorescence, and laser mapping imaging system
US11793399B2 (en) 2019-06-20 2023-10-24 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed hyperspectral imaging system
EP3907497A1 (en) * 2020-05-08 2021-11-10 Leica Microsystems CMS GmbH Apparatus and method for displaying and/or printing images of a specimen including a fluorophore
CN114266817A (en) * 2022-03-02 2022-04-01 广东欧谱曼迪科技有限公司 Fluorescent depth image synthesis method and device, electronic equipment and storage medium
EP4275580A1 (en) * 2022-05-13 2023-11-15 Leica Instruments (Singapore) Pte Ltd Method, processor, and medical fluorescence observation device using two color images to record fluorescence
EP4275577A1 (en) * 2022-05-13 2023-11-15 Leica Instruments (Singapore) Pte Ltd Method, processor, and medical fluorescence observation device using two color images and color cameras for fluorescence and white-light
WO2023218087A1 (en) * 2022-05-13 2023-11-16 Leica Instruments (Singapore) Pte. Ltd. Method, processor, and medical fluorescence observation device using two color images to record fluorescence
WO2023218082A1 (en) * 2022-05-13 2023-11-16 Leica Instruments (Singapore) Pte. Ltd. Method, processor, and medical fluorescence observation device using a color-dependent color conversion function
WO2023218093A1 (en) * 2022-05-13 2023-11-16 Leica Instruments (Singapore) Pte. Ltd. Method, processor, and medical fluorescence observation device using two color images and color cameras for fluorescence and white-light
EP4275578A1 (en) * 2022-05-13 2023-11-15 Leica Instruments (Singapore) Pte Ltd Method, processor, and medical fluorescence observation device using a color-dependent color conversion function

Also Published As

Publication number Publication date
JP6456416B2 (en) 2019-01-23
CN107137053B (en) 2020-01-14
US11330237B2 (en) 2022-05-10
EP3205254A1 (en) 2017-08-16
EP3205254B1 (en) 2020-11-18
US20210021795A1 (en) 2021-01-21
JP2017148503A (en) 2017-08-31
CN107137053A (en) 2017-09-08

Similar Documents

Publication Publication Date Title
US11330237B2 (en) Medical inspection apparatus, such as a microscope or endoscope using pseudocolors
EP3540494B1 (en) Augmented reality surgical microscope and microscopy method
EP2108306B1 (en) Device and method for endoscopic 3D data creation
US20120056996A1 (en) Special-illumination surgical video stereomicroscope
US9727962B2 (en) System for visualizing tissue in a surgical region
CN111803013A (en) Endoscope imaging method and endoscope imaging system
JPWO2019230878A1 (en) Fluorescence observation device and fluorescence observation method
CN105310633B (en) Medical image-processing apparatus and its method of work and endoscopic system
EP2564755B1 (en) Image processing device and fluoroscopy device
CN105212885B (en) Medical image-processing apparatus and its method of work and endoscopic system
JP6608884B2 (en) Observation device for visual enhancement of observation object and operation method of observation device
CN112040831A (en) Camera device, image processing method, and camera system
US20210251570A1 (en) Surgical video creation system
CN102791179B (en) Fluorescence endoscope device
CN107005683A (en) Endoscope apparatus
CN110087528B (en) Endoscope system and image display device
CN112089403A (en) Multispectral medical multi-path imaging method and system thereof
CN212326346U (en) Endoscope imaging system
JP2021145859A (en) Medical image processing device and medical observation system
JP5592715B2 (en) Image processing apparatus and image processing method
US20220318969A1 (en) Method for providing an image representation by means of a surgical microscope, and surgical microscope
CN111345902B (en) Systems and methods for creating HDR monochromatic images of fluorescing fluorophores
EP4275577A1 (en) Method, processor, and medical fluorescence observation device using two color images and color cameras for fluorescence and white-light
EP4277256A1 (en) Image processor and computer-implemented method for a medical observation device, using a location-dependent color conversion function
CN107661087A (en) Medical imaging apparatus and method for the imaging of photosensitive object such as biological tissue

Legal Events

Date Code Title Description
AS Assignment

Owner name: LEICA INSTRUMENTS (SINGAPORE) PTE. LTD., SINGAPORE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:THEMELIS, GEORGE;REEL/FRAME:041174/0115

Effective date: 20170113

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION