US20200015669A1 - Illumination sources for multicore fiber endoscopes - Google Patents
Illumination sources for multicore fiber endoscopes Download PDFInfo
- Publication number
- US20200015669A1 US20200015669A1 US16/558,226 US201916558226A US2020015669A1 US 20200015669 A1 US20200015669 A1 US 20200015669A1 US 201916558226 A US201916558226 A US 201916558226A US 2020015669 A1 US2020015669 A1 US 2020015669A1
- Authority
- US
- United States
- Prior art keywords
- fiber
- endoscope
- cores
- treatment
- tip
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 239000000835 fiber Substances 0.000 title claims abstract description 278
- 238000005286 illumination Methods 0.000 title claims abstract description 102
- 230000003287 optical effect Effects 0.000 claims abstract description 123
- 238000003384 imaging method Methods 0.000 claims abstract description 90
- 238000011282 treatment Methods 0.000 claims abstract description 45
- 230000005855 radiation Effects 0.000 claims description 21
- 239000011800 void material Substances 0.000 claims description 19
- 239000002105 nanoparticle Substances 0.000 claims description 10
- 238000005253 cladding Methods 0.000 claims description 9
- 238000004891 communication Methods 0.000 claims description 6
- 229920000642 polymer Polymers 0.000 claims description 5
- 238000000034 method Methods 0.000 abstract description 53
- 230000010287 polarization Effects 0.000 description 18
- 239000004575 stone Substances 0.000 description 11
- 238000012545 processing Methods 0.000 description 10
- 238000009826 distribution Methods 0.000 description 9
- 239000000463 material Substances 0.000 description 8
- 230000001965 increasing effect Effects 0.000 description 7
- 230000003595 spectral effect Effects 0.000 description 7
- 238000000576 coating method Methods 0.000 description 6
- 230000002708 enhancing effect Effects 0.000 description 6
- 238000004519 manufacturing process Methods 0.000 description 6
- 230000008901 benefit Effects 0.000 description 5
- 239000011248 coating agent Substances 0.000 description 5
- 238000013461 design Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 5
- 230000008569 process Effects 0.000 description 5
- 238000005070 sampling Methods 0.000 description 5
- 238000001228 spectrum Methods 0.000 description 5
- 208000000913 Kidney Calculi Diseases 0.000 description 4
- 206010029148 Nephrolithiasis Diseases 0.000 description 4
- 239000003365 glass fiber Substances 0.000 description 4
- 238000005259 measurement Methods 0.000 description 4
- 229920005594 polymer fiber Polymers 0.000 description 4
- 238000007493 shaping process Methods 0.000 description 4
- 230000002123 temporal effect Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 3
- 230000006872 improvement Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 239000004926 polymethyl methacrylate Substances 0.000 description 3
- 230000001902 propagating effect Effects 0.000 description 3
- 210000001747 pupil Anatomy 0.000 description 3
- 230000009467 reduction Effects 0.000 description 3
- 239000007787 solid Substances 0.000 description 3
- 239000004793 Polystyrene Substances 0.000 description 2
- 238000010521 absorption reaction Methods 0.000 description 2
- 238000005452 bending Methods 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 2
- 230000000052 comparative effect Effects 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 230000008878 coupling Effects 0.000 description 2
- 238000010168 coupling process Methods 0.000 description 2
- 238000005859 coupling reaction Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 239000006185 dispersion Substances 0.000 description 2
- 230000005670 electromagnetic radiation Effects 0.000 description 2
- 238000007380 fibre production Methods 0.000 description 2
- 239000002657 fibrous material Substances 0.000 description 2
- 239000012530 fluid Substances 0.000 description 2
- 238000010438 heat treatment Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000013532 laser treatment Methods 0.000 description 2
- 239000007788 liquid Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000004806 packaging method and process Methods 0.000 description 2
- 238000012856 packing Methods 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 239000002861 polymer material Substances 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 102000001554 Hemoglobins Human genes 0.000 description 1
- 108010054147 Hemoglobins Proteins 0.000 description 1
- 206010028980 Neoplasm Diseases 0.000 description 1
- 238000009825 accumulation Methods 0.000 description 1
- 210000003484 anatomy Anatomy 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 239000000560 biocompatible material Substances 0.000 description 1
- 238000001574 biopsy Methods 0.000 description 1
- 210000001124 body fluid Anatomy 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 230000001427 coherent effect Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000000205 computational method Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 239000004020 conductor Substances 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 238000001816 cooling Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000006866 deterioration Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 238000012377 drug delivery Methods 0.000 description 1
- 230000005611 electricity Effects 0.000 description 1
- 238000001839 endoscopy Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000001747 exhibiting effect Effects 0.000 description 1
- 210000004905 finger nail Anatomy 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 238000011065 in-situ storage Methods 0.000 description 1
- 238000010348 incorporation Methods 0.000 description 1
- 239000012212 insulator Substances 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 239000002184 metal Substances 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000009022 nonlinear effect Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000006213 oxygenation reaction Methods 0.000 description 1
- 239000004038 photonic crystal Substances 0.000 description 1
- 229920002223 polystyrene Polymers 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 238000002106 pulse oximetry Methods 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 125000006850 spacer group Chemical group 0.000 description 1
- 238000012876 topography Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/06—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements
- A61B1/07—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements using light-conductive means, e.g. optical fibres
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00064—Constructional details of the endoscope body
- A61B1/00071—Insertion part of the endoscope body
- A61B1/0008—Insertion part of the endoscope body characterised by distal tip features
- A61B1/00096—Optical elements
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00163—Optical arrangements
- A61B1/00165—Optical arrangements with light-conductive means, e.g. fibre optics
- A61B1/00167—Details of optical fibre bundles, e.g. shape or fibre distribution
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00163—Optical arrangements
- A61B1/00165—Optical arrangements with light-conductive means, e.g. fibre optics
- A61B1/0017—Details of single optical fibres, e.g. material or cladding
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B23/00—Telescopes, e.g. binoculars; Periscopes; Instruments for viewing the inside of hollow bodies; Viewfinders; Optical aiming or sighting devices
- G02B23/24—Instruments or systems for viewing the inside of hollow bodies, e.g. fibrescopes
- G02B23/2407—Optical details
- G02B23/2423—Optical details of the distal end
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B23/00—Telescopes, e.g. binoculars; Periscopes; Instruments for viewing the inside of hollow bodies; Viewfinders; Optical aiming or sighting devices
- G02B23/24—Instruments or systems for viewing the inside of hollow bodies, e.g. fibrescopes
- G02B23/2407—Optical details
- G02B23/2461—Illumination
- G02B23/2469—Illumination using optical fibres
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B23/00—Telescopes, e.g. binoculars; Periscopes; Instruments for viewing the inside of hollow bodies; Viewfinders; Optical aiming or sighting devices
- G02B23/24—Instruments or systems for viewing the inside of hollow bodies, e.g. fibrescopes
- G02B23/26—Instruments or systems for viewing the inside of hollow bodies, e.g. fibrescopes using light guides
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/0075—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 with means for altering, e.g. increasing, the depth of field or depth of focus
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/28—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 for polarising
- G02B27/283—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00 for polarising used for beam splitting or combining
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B27/00—Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
- G02B27/48—Laser speckle optics
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/02—Optical fibres with cladding with or without a coating
- G02B6/02042—Multicore optical fibres
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/254—Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/296—Synchronisation thereof; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
-
- H04N5/2254—
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B6/00—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings
- G02B6/04—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings formed by bundles of fibres
- G02B6/06—Light guides; Structural details of arrangements comprising light guides and other optical elements, e.g. couplings formed by bundles of fibres the relative position of the fibres being the same at both ends, e.g. for transporting images
-
- H04N2005/2255—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/50—Constructional details
- H04N23/555—Constructional details for picking-up images in sites, inaccessible due to their dimensions or hazardous conditions, e.g. endoscopes or borescopes
Definitions
- the present invention relates to the field of endoscopy, and more particularly, to multicore fiber endoscopes.
- Endoscopes in various configurations allow efficient treatment of a range of medical problems, as well as means for manipulating different situations with limited access. Endoscope operations are challenging in that illumination, detection and treatment are confined to long and narrow operations modes. Fiber optics technology is a central enabler for such techniques, and fiber-based endoscope experience continuous improvements.
- One aspect of the present invention provides an endoscope having a distal tip and a proximal tip, the endoscope comprising at least one multicore fiber module comprising at least one hundred cores distributed at a fill factor smaller than 1/4, an illumination source coupled to the at least one multicore fiber module and configured to deliver illumination thereto, at least one optical element, in optical communication with the cores, at the distal tip, a detector, in optical communication with the cores, at the proximal tip, and a processor configured to receive images from the detector; wherein the endoscope is configured to implement super-resolved imaging by micro scanning over a pitch distance between the cores, and wherein the endoscope is configured to implement three dimensional sensing by handling the cores group-wise with respect to radiation delivered therethrough, and to at least one of: enhance, by configuring the at least one optical element, a field of view of the endoscope beyond a region facing the cores at the distal tip, and enhance, by configuring the at least one optical element, a depth of field of
- FIGS. 1A-1E are high level schematic illustrations of endoscope configurations according to some embodiments of the invention.
- FIGS. 2A-2C are high level schematic illustrations of fiber cross sections having a large number of cores in their electromagnetic propagation region(s), according to some embodiments of the invention.
- FIGS. 2D and 2E are high level schematic illustrations of fiber production by packing fiber modules, according to some embodiments of the invention.
- FIGS. 3A-3C are high level schematic cross section illustrations of fibers having working channels and additional channel positions for treatment or illumination fibers, according to some embodiments of the invention.
- FIG. 3D is a high level schematic illustration of a fiber with an assembled front lens, according to some embodiments of the invention.
- FIGS. 3E-3G are high level schematic illustrations of a defogging mechanism and its effects, according to some embodiments of the invention.
- FIGS. 4A-4D are high level schematic illustrations of hollow endoscope fibers having optical elements at the distal tip which compensate for the central void, according to some embodiments of the invention.
- FIGS. 5A-5C are high level schematic illustrations of optical elements, according to some embodiments of the invention.
- FIGS. 6A and 6B are high level schematic illustration of fiber cross sections with different configurations of the cores, according to some embodiments of the invention.
- FIG. 6C illustrates comparative experimental results of full core and hollow core fibers, according to some embodiments of the invention.
- FIG. 7 is a high level schematic flowchart illustrating a method, according to some embodiments of the invention.
- FIGS. 8A-8E are high level schematic illustrations of experimental imaging results for bundled fibers, according to some embodiments of the invention.
- FIGS. 9A-9D are images that provide examples for performance of the endoscope, according to some embodiments of the invention.
- distal and proximal refer to the ends of the endoscope.
- the end and associated parts of the endoscope which are far from the endoscope's interface (detector or eye) and close to the imaged tissue and to its surroundings is termed the distal end, while the end and associated parts of the endoscope which are close to the endoscope's interface and are remote from the imaged tissue, being typically outside the body is termed the proximal end.
- reflected refers to a change in a direction of an illumination wavefront which impacts one or more imaged object or tissue.
- reflection is understood broadly as any radiation gathered by the fiber, irrespective of the source of the illumination which is reflected by the object(s) and/or tissue(s).
- near field imaging refers to the formation of an image (of imaged objects, tissues and/or their surroundings) at the distal end of the endoscope fiber, typically at the fiber's tip. The imaged is then typically transferred through the fiber to the detector, possibly through proximal optical elements.
- near field imaging may relate to different types of optical systems, including direct imaging without any optical elements between the imaged object or tissue and the fiber tip as well as to imaging through optical element(s) such as lenses.
- far field imaging refers to the formation of a Fourier transform of imaged objects, tissues and/or their surroundings at the distal end of the endoscope fiber (e.g., the distal end of the endoscope fiber is at the aperture or pupil plane of the optical system), typically at the fiber's tip.
- the image of the imaged objects, tissues and/or their surroundings may be formed at the proximal end of the endoscope fiber, typically at the fiber's proximal tip or directly on the detector, possibly through proximal optical elements.
- far field imaging may relate to different types of optical systems.
- far field imaging may be direct in the sense that no optical elements are used between the imaged object or tissue and the distal fiber tip, which delivers radiation entering the fiber along the fiber to the detector at the proximal end of the fiber.
- far field imaging may be carried out with optical elements positioned between the imaged object or tissue and the distal fiber tip, with the distal fiber tip being at least approximately at the Fourier plane (also termed aperture plane and pupil plane in different contexts) of the optical elements.
- the fibers may have hundreds or thousands of cores and possibly incorporate working channel(s) and additional fibers.
- the fiber may be used at different optical configurations to capture images of tissue and objects at the distal tip and to enhance a wide range of optical characteristics of the images such as resolution, field of view, depth of field, wavelength ranges etc.
- Near-field imaging as well as far-field imaging may be implemented in the endoscopes and the respective optical features may be utilized to optimize imaging.
- Optical elements may be used at the distal fiber tip, or the distal fiber tip may be lens-less. Diagnostics and optical treatment feedback loops may be implemented and illumination may be adapted to yield full color images, depth estimation, enhanced field of view and/or depth of field and additional diagnostic data, as disclosed below.
- Certain endoscope embodiments may implement far field imaging (see FIG. 1A below), e.g., have the image formed at the proximal end of the endoscope fiber, while certain endoscope embodiments may implement near field imaging (see FIG. 1B below), e.g., have the image formed at the distal end of the endoscope fiber.
- Both far field and near field implementations may have distal optical elements between the imaged objects or tissues and the distal fiber tip (see FIG. 1C below), or may operate without such distal optical elements (see FIG. 1D below).
- each of the four combinations has different features, advantages and disadvantages as exemplified in Table 1, and may be selected according to specific implementation scenarios. Alternation of the combination may be carried out between applications or in real time, to combine advantages of different configuration types. It is further noted that endoscopes may be designed to have several combinations, e.g., a part of the fiber face (or certain fiber modules) having distal optics for imaging far objects and another part of the fiber face (or other fiber modules) lacking distal optics for microscopic imaging.
- Certain embodiments comprise lens-less embodiments in which the distal fiber tip lacks optical elements.
- Lens-less embodiments may implement either far-field or near-field imaging, and may utilize structural features to enhance optical resolution, apply super-resolution methods and retrieve wavefront information while reducing crosstalk between the cores.
- Endoscope embodiments may have full tip cross sections or have working channel(s) within the imaging fiber characterized by different configurations and uses, integrating additional fibers etc., in which case the cores and optical elements may be configured to overcome the reduction of the field of view due to the incorporation of the working channel.
- the disclosed endoscopes may serve different purposes, e.g., may be designed as a laparoscope or an ureteroscope. It is noted that elements disclosed in the context of some of the embodiments are not necessarily limited to these embodiments but may be implemented within other embodiments as well.
- FIGS. 1A-1E are high level schematic illustrations of endoscope configurations according to some embodiments of the invention.
- Proposed micro endoscope 105 is constructed from large plurality of cores (e.g. one hundred cores or more, hundreds of cores, thousands of cores, in certain embodiments tens or hundreds of thousand cores per fiber or fiber module, reaching over a million cores in certain fiber endoscopes), each responsible for transferring a single or a large number of spatial degrees of freedom out of which at the output, proximal end (the one external to the patient body), a high resolution color image may be constructed.
- cores e.g. one hundred cores or more, hundreds of cores, thousands of cores, in certain embodiments tens or hundreds of thousand cores per fiber or fiber module, reaching over a million cores in certain fiber endoscopes
- Multi-core fiber 100 exhibits a high degree of flexibility in its optical design, as exemplified below, which may be utilized and adapted for specific applications, for example for ureteroscopes with a large working channel and a small external diameter or for laparoscopes with a very high resolution obtained at a small external diameter.
- Endoscope 105 may be configured to carry out far-field imaging, near-field imaging or a combination of far-field imaging and near-field imaging. Irrespectively of the imaging mode, endoscope 105 may be configured to have one or more optical elements 140 at a distal tip 101 of fiber 100 or have no optical elements between tip 101 and imaged tissue(s) or object(s) 70 . Certain embodiments may comprise removable or reconfigurable optical elements 140 at tip 101 and/or optical elements 140 affecting only parts of the surface of distal tip 101 (e.g., sub-group(s) of the cores).
- Certain embodiments comprise endoscopes 105 having a plurality of fibers 100 , grouped together, each having at least one hundred cores distributed at a fill factor smaller than 1/4, or even smaller than 1/9, at least one photonic illumination fiber, and at least one optical element at a distal tip of fibers 100 , which may be configured to enhance a field of view and/or a depth of field of endoscope 105 beyond a region facing a tip of fibers 100 and congruent thereto (see details below).
- Endoscope 105 may be further configured to implement three dimensional sensing by handling the cores group-wise with respect to radiation delivered therethrough (see details below).
- Endoscope 105 may be further configured to super-resolved imaging by micro scanning over a pitch distance between the cores (see details below).
- Endoscope 105 may be configured to comprise a LED (light emitting diode) light source located at distal tip 101 as the illumination source.
- LED light emitting diode
- FIG. 1A schematically illustrates far-field imaging, in which an image 73 (indicating any kind of electromagnetic signal reflected from tissue or object 70 ) is delivered through tip 101 and fiber 100 to yield image 75 on detector 91 .
- Tip 101 may be a Fourier plane (also termed aperture plane or pupil plane) at which the Fourier transform 74 of image 73 enters fiber 100 . It is noted that the Fourier plane may be located anywhere along fiber 100 as well as distally or proximally to fiber 100 , in different embodiments of the invention, and be optically transformed to image 75 on detector 91 .
- Fourier image 74 or derivatives thereof may be measured at detector 91 , and/or manipulated to enhance imaging parameters such as resolution, field of view and depth of focus, as non-limiting examples.
- Optical elements may be introduced distally or proximally to fiber 100 to modify or manipulate the radiation entering tip 101 and the radiation falling on detector 91 , respectively.
- FIG. 1B schematically illustrates near-field imaging, in which image 73 yields image 75 at fiber tip 101 .
- Image 75 is then delivered, possibly through optical elements, to detector 91 through fiber 100 . It is noted that image 75 may be formed within fiber 100 and not necessarily exactly at tip 101 . Image 75 delivered via fiber 100 may be measured at detector 91 , and/or manipulated to enhance imaging parameters such as resolution, field of view and depth of focus, as non-limiting examples.
- Optical elements may be introduced distally or proximally to fiber 100 to modify or manipulate the radiation entering tip 101 and the radiation falling on detector 91 , respectively.
- FIG. 1C schematically illustrates optical configurations having one or more optical element(s) 140 at the distal end of fiber 100 , at proximity to imaged tissue 70 .
- Optical element(s) 140 may be attached to tip 101 or may be somewhat distally removed from tip 101 (e.g., held by spacers at a distance therefrom).
- Each optical element 140 may be in optical communication with a respective core or a respective group of cores.
- illumination 85 is delivered to fiber 100 by an illumination source 160
- reflected illumination e.g., in far-field, in near-field or in an intermediate plane
- a detector 91 e.g., via a beam splitter 90 .
- Proximal optical elements may be set and used to manipulate illumination 85 and the reflected illumination, as symbolized below ( FIG. 1D ) by lenses 84 , 94 respectively.
- One or more processor(s) 170 may be configured to control the illumination and/or process the detected illumination, as well as control illumination and image beams in case there are controllable elements in the optical path.
- FIG. 1D schematically illustrates optical configurations having no optical element(s) (also termed below “lens-less” configurations) at the distal end of fiber 100 , so that fiber tip 101 is used directly to deliver and receive illumination to and from imaged tissue 70 .
- Illumination 85 is delivered to fiber 100 proximally, e.g., via an optical element 84 such as a lens, and reflected illumination is directed to detector 91 via another optical element 94 , e.g., a lens.
- One or more processor(s) 170 may be configured to control the illumination and/or process the detected illumination, as well as control illumination and image beams in case there are controllable elements in the optical path.
- lens-less configurations may be configured to generate image at “contact mode”, e.g., with close proximity of the fiber tip to the examined tissue, to yield microscopic resolution determined by the sizes of the cores.
- proximal optical elements 94 may be variable and be used to adjust the plane and depth of focus of captured images in far-field imaging configurations, especially in lens-less configurations.
- FIG. 1E is a high level schematic block diagram illustrating endoscope configurations according to some embodiments of the invention.
- Various embodiments are illustrated, which may be stand-alone embodiments or be implemented in any combination thereof.
- various embodiments of illumination source 160 and of configurations of processor 170 are presented, which may be used to improve the spatial resolution, in particular when using super resolution algorithms, improve the beam quality and/or enhance the functionality of endoscope 105 with respect to its medical uses and image quality.
- Embodiments illustrated in FIG. 1E may be applied to any embodiment of endoscope 105 described herein.
- illumination source 160 may be configured to deliver illumination 85 through one or more dedicated illumination fiber(s) 102 and/or through multicore fiber 100 .
- illumination fiber(s) 102 may be multimode fiber(s), possibly made of glass fiber, which are associated with multicore fiber 100 , e.g., attached thereto or positioned in a cavity in multicore fiber 100 .
- illumination fiber(s) 102 may be positioned to illuminate tissue 70 in any other spatial relation to multicore fiber 100 , possibly in no mechanical association therewith.
- one or more of illumination fiber(s) 102 may be single mode fibers. It is noted that the spatial relation between illumination fiber(s) 102 and multicore fiber 100 may be configured to have multicore fiber 100 receive radiation (illuminated by illumination fiber(s) 102 ) which is reflected off tissue 70 and/or transmitted through tissue 70 , depending on specific use conditions.
- illumination source 160 may comprise a coherence modulator 162 configured to enable processor 170 to implement algorithms for improving super resolution results 172 .
- coherence modulator 162 may be configured to use a coherence modulation of illumination 85 that reduces speckle patterns by modulating the coherence using Barker codes rather than random prior art modulation.
- using Barker codes may reduce the required number of modulation steps for a given reduction of speckle patterns due to the orthogonality between the Barker codes and other characteristics of their definition. Specific Barker codes may be selected to optimize their application.
- illumination source 160 may comprise multiple narrowband wavelengths 62 (e.g., narrowband spectral ranges around specified wavelengths) which may be used in processors 170 configured to provide diagnostics using one or more wavelength combinations 174 , discussed below in more detail, and/or configured to implement wavelength multiplexing super resolution 176 by changing the ratios between specific wavelengths (e.g., between narrowband red, green and blue sources 62 )—to achieve improved super resolution results 180 .
- narrowband wavelengths 62 e.g., narrowband spectral ranges around specified wavelengths
- processors 170 configured to provide diagnostics using one or more wavelength combinations 174 , discussed below in more detail, and/or configured to implement wavelength multiplexing super resolution 176 by changing the ratios between specific wavelengths (e.g., between narrowband red, green and blue sources 62 )—to achieve improved super resolution results 180 .
- illumination source 160 may comprise one or more photonic crystal fiber (PCF) 164 configured to deliver wideband white spectrum 177 into dedicated illumination fiber(s) 120 and/or multicore fiber 100 , e.g., utilizing supercontinuum effects (bandwidth broadening due to nonlinear effects) to provide white illumination of tissues that is closer to natural wideband illumination than illumination with narrowband red, green and blue sources 62 which are delivered to PCF 164 .
- PCF 164 may be coupled to one or more narrowband sources 62 and designed to have zero dispersion point(s) at the wavelengths of source(s) 62 to yield spectral broadening.
- Using several multiple narrowband wavelengths 62 may provide wideband white spectrum 177 through a combination and merging of the broadened spectra of source(s) 62 .
- Improved white spectrum 177 may be advantageous to provide truer imaging colors be endoscope 105 .
- illumination source 160 may comprise structured light patterned illumination 168 , which may be used in processor 170 configured to provide 3D sensing 178 by analyzing the illuminated patterns on the tissue and/or improved super resolution results 180 by utilizing the parameters of the temporally changing spatially projected patterns of illumination 168 to implement temporal multiplexing super resolution 179 .
- illumination source 160 may comprise one or more laser source(s) 64 (possibly narrowband sources 62 ) in illumination source 160 and at least one beam shaping element 182 at the distal end of multicore fiber 100 which is configured to generate an optimized beam profile 184 to improve illumination 85 .
- beam profile 184 may comprise a uniform illumination distribution in space or a rectangular uniform profile (top hat illumination distribution), which are advantageous with respect to prior art Gaussian illumination distribution with respect to various parameters of the resulting images.
- the coherence of laser source(s) 64 may be used to shape illumination beam 85 efficiently by beam shaping element 182 .
- at least one beam shaping element 182 may be set at the proximal end of multicore fiber 100 .
- illumination source 160 may comprise one or more laser treatment source(s) 66 which are configured to apply a specified treatment 67 by endoscope 105 , e.g., to a tissue.
- treatment 67 may be applied to kidney stones in endoscope 105 designed as an ureteroscope, as described below in more details.
- FIGS. 2A-2C are high level schematic illustrations of fiber cross sections having a large number of cores 115 in their electromagnetic propagation region(s) 110 , according to some embodiments of the invention.
- Fiber(s) 100 may comprise central or eccentric optical cores ( 110 ) and/or may have hollow, central or eccentric region(s) ( 112 ) that may be used for treatment such as energy delivery, suction, illumination, drug delivery etc.
- Illumination means (such as dedicated illumination fiber(s) 102 ), may be integrated in various ways within the multicore fibers 100 .
- Fiber 100 illustrated in FIG. 2A may have any form of cross section, e.g., square as illustrated in a non-limiting manner, round, hexagonal, elliptic etc. While FIG. 2A illustrates a solid cross section of fiber 100 , FIG. 2B illustrates hollow endoscope having a void 112 within fiber 100 that may be used for different purposes as disclosed below (e.g., as a working channel for inserting a tool or carrying out suction, for incorporating additional fibers etc.).
- Fibers 100 may be square, round or have any other form, and void 112 may too have any shape and any position within fiber 100 , void(s) 112 and fiber 100 may have any dimensions (R i , R o , D, W etc.), and voids may also be multiple (e.g., fiber 100 may enclose two or more voids), all designed according to requirements from the endoscope.
- FIG. 2C schematically illustrates multicore fiber 100 with cores 115 grouped into “super core” groups 116 that may be configured to sense wavefronts in lens-less configurations, as explained below.
- Multicore fiber 100 may be made of biocompatible materials in case of medical uses, e.g., polymers such as PMMA (poly-methyl methacrylate) and PS (polystyrene) and may be flexible. Fiber 100 may also be made of non-compatible materials and be flexible or rigid in case of industrial uses. Fiber 100 may be configured to have a flexibility characterized by a Young's modulus smaller than 10 GPa and to be disposable. Fiber 100 may thus be more flexible than glass fiber (having a Young's modulus of about 65 GPa), and may reach PMMA flexibility (Young's modulus between 1.8 and 3.1 GPa) or higher flexibility.
- PMMA poly-methyl methacrylate
- PS polystyrene
- Various embodiments compensate for the reduced transparency of polymer fibers with respect to glass fibers, using means such as fiber materials, configuration of cores and interspaces, number and sizes of cores, material modifications of different fiber parts, control over the number of propagation modes in cores 115 , optical means such as lenses or prisms at either side of fiber 100 and their configuration, design and application of different types of illumination and algorithmic solutions, all of which are exemplified below in a non-limiting manner.
- the following disclosure also addresses ways to control cross talk between cores 115 (e.g., interaction effects between radiation propagating in adjacent cores 115 ) and ways to improve the information content and to enhance treatment-relevant information of the detected images.
- Illumination may comprise coherent light or incoherent light, any spectral pattern (broad or narrow wavelength ranges, continuous or discrete ranged), polarized (in various patterns) or non-polarized light and different ranges in the visual or infrared ranges.
- Material differences between cores, interspaces and outer cladding may comprise different materials, using air cores or air interspaces, and doping any of the fiber regions to influence their refractive indices, as explained in more details below. It is noted that any of the embodiments presented below may be used in any of the other embodiments described herein, as long as they are compatible. Particularly, computational methods optical methods and fiber design considerations described in the context of any embodiment may be applied to other embodiments as well.
- FIGS. 2D and 2E are high level schematic illustrations of fiber production by packing fiber modules, according to some embodiments of the invention.
- Multi-core fibers 100 may be produced using fiber modules or units 117 .
- Each fiber module 117 is itself a multicore fiber, possibly configured to have uniform dimensions. Such embodiments are referred to as bundled fibers, and may bundle any number of fiber modules 117 in any configuration (e.g., 2 ⁇ 2 modules, 3 ⁇ 3 modules etc.).
- Fiber module 117 may have any form, such as square, rectangular, round or elliptic, and may be packed into fibers 100 having a wide range of forms and configurations, Introducing fiber modules 117 having an intermediate dimension between cores or core groups and whole fiber 100 (each module 117 may have e.g.
- rectangular fiber 100 may be assembled from rectangularly arranged square fiber modules 117 , e.g., using a package support 118 A and a respective attachable cover 118 B.
- Fiber modules 117 may simply be mechanically held by package support 118 A and cover 118 B at certain regions along fiber 100 and/or fiber modules 117 may be glued together or otherwise attached at least at certain regions.
- fiber modules 117 A, 117 B may be arranged around void 112 .
- fiber modules 117 A, 117 B may be arranged to differ in their observation angles and/or in optical elements 140 attached at fiber tip 110 (see e.g., below, FIGS. 4A-4D ).
- fiber units 117 A may be configured to cover a field of view in front of void 112 (e.g., be inclined inwards or have respective optical elements) while fiber units 117 B may be configured to cover a field of view laterally beyond tip 101 (e.g., be inclined outwards or have respective optical elements).
- non-limiting inclination angles may be 5-20° inwards and 10-50° outwards.
- Respective packaging or attachment configurations may be applied to fixate fiber modules 117 A, 117 B in their respective positions and angles.
- the annular arrangement of fiber modules 117 A, 117 B may be at the fiber's distal end, while fiber modules 117 A, 117 B may be separated and re-arranged differently at the fiber's proximal end, e.g., into a rectangular form to cover a face of a single rectangular detector.
- FIGS. 3A-3C are high level schematic cross section illustrations of fiber 100 having working channel 112 and channel positions 120 for treatment or illumination fibers 102 , according to some embodiments of the invention.
- Working channel 112 depicted as void 112 within fiber 100 , is surrounded by electromagnetic propagation multicore fiber region 110 .
- Treatment and/or illumination fiber(s) 102 may be integrated into fiber 100 of the endoscope in a way that allows combined imaging and treatment using one fiber, immediate image feedback of the treatment etc. Such combination may be used e.g., as ureteroscope or as any other type of endoscope.
- positioning additional fibers in channels 120 near working channel 112 may be configured to cool down the fibers (e.g., treatment fibers) by the liquids flowing through working channel 112 .
- treatment or illumination fibers 102 may be inserted at indicated positions 120 (e.g., grooves, or channels), e.g., at an inner wall of multi-core imaging region 110 in fluid communication with working channel 112 , e.g., on the periphery of voids 112 ( FIG. 3A , channel diameter e.g., ca. 250 ⁇ m), at an outer wall of multi-core imaging region 110 in fluid communication with the surroundings of fiber 100 , e.g., on the periphery of fiber 100 ( FIG. 3B , channel diameter e.g., ca. 250 ⁇ m), within multi-core imaging region 110 ( FIG. 3C , channel diameter e.g., ca.
- Integration of the treatment or illumination fibers 102 may be carried out before, during or after production of fiber 100 .
- glass treatment or illumination fibers 102 may be inserted into grooves 120 after pulling polymer fiber 100 .
- treatment or illumination fibers may be configured and controlled to operate collectively, simultaneously or sequentially, to achieve a desired illumination and/or treatment.
- the treatment channel may be split into several low power channels 120 to have thinner channels and lower power delivery through each channel.
- Such configuration may enable increasing the mechanical flexibility of the endoscope, which is very important, e.g., in the field of ureteroscopy.
- the usage of hollow channels 120 for inserting the external illumination or treatment fibers provides a device configuration exhibiting self-alignment.
- FIG. 3D is a high level schematic illustration of fiber 100 with an assembled lens 119 , according to some embodiments of the invention.
- a modular construction of fiber 100 may be used to modify some of fiber modules 117 to incorporate features into fiber 100 in a simpler manner than incorporating these features into a uniform fiber.
- fiber modules 117 D may be configured in a modular, building block style manner to form various cross sectional organizations with respect to form and functionality of the endoscope.
- two non-adjacent fiber modules 117 D may be coated with a conductor (e.g., a metal) while the rest of fiber modules 117 C may be uncoated (and insulating).
- a conductor e.g., a metal
- Such configurations may be used to deliver electricity to fiber tip 101 .
- electromagnetic signals or electromagnetic radiation may be delivered via fiber modules 117 D to adjacent tissues or to associated devices or components (e.g., checking equipment or endoscope instrumentation).
- electromagnetic energy may be delivered to distal lens 119 for heating it to prevent fogging upon entry to the body.
- an antenna structure (not shown) may be designed upon lens 119 , which receives electromagnetic radiation to heat lens 119 without using contacts.
- radiofrequency (RF) treatment may be applied to tissue or objects surrounding fiber tip 101 via the conductive coating of fiber modules 117 D.
- FIGS. 3E-3F are additional high level schematic illustrations of a defogging mechanism 121 and its effects, according to some embodiments of the invention.
- FIG. 3E illustrates lens 119 coated by a conductive coating 122 connected to an electric circuit 123 configured to heat lens 119 via coating 122 , to prevent fog and to defog lens 119 when required.
- FIG. 3F exemplifies image deterioration by fog accumulation—the top image (A) taken a short time after the beginning of fog accumulation, the bottom image (B) taken later, with the object, marked by an arrow, barely visible.
- FIG. 3G illustrates the image after defogging—both object and illumination spot are clear again.
- endoscope 105 may be designed to be very thin (e.g., 0.5 mm in diameter) while providing high resolution images, and distal lens 119 may also be ultra-thin.
- the disclosed defogging mechanism provides effective control of the temperature of lens 119 using a small amount of electrical power, to prevent fogging and overcome an important prior art limitation.
- endoscope 105 may be operated in the far field ( FIG. 1A ) or in the near field ( FIG. 1B ) by properly adapting the focal length of the external optics (the one outside the patient's body, e.g., optical elements 84 , 94 ) to the working distance of treated tissue 70 from the distal tip of the endoscope.
- Fiber 100 may be configured to deliver full images even with working channel 112 in the middle of the imaging surface by employing far field imaging, e.g., using imaging lens 94 adapted to have a central blocked aperture.
- obtained images may have a number of pixels that is not related to the number of cores 115 , enhancing the image resolution with respect to near field embodiments.
- certain embodiments comprise using as detector 91 an integral imaging sensor capable of sensing wavefront or the 3D topography of inspected tissue 70 .
- cores 115 may be configured to have a small number of possible spatial modes, resembling the Shack-Hartmann interferometer or a wavefront sensor.
- cores 115 may be grouped into “super-cores” 116 (see FIG. 2C ), each comprising a group of adjacent cores 115 .
- Each “super-core” 116 may be handled as a single wavefront sensing element which delivers information about the wavefront by comparing radiation propagating through individual core members 115 within each “super-core” 116 (or light field sensing. e.g., comparing light directions at different cores operating in near field and multi-mode).
- the grouping of cores 115 into “super-cores” 116 may be uniform across the face of fiber 100 or be variable, some core groups being larger than others, see e.g., the larger central core group in FIG. 2C ).
- the grouping of cores 115 may be changed in time according to imaging performance preferences, based e.g., on an even (or uneven) distribution of cores 115 across fiber 110 . It is noted that in such configurations a tradeoff exists between depth measurements and resolution. A larger number of cores 115 in each “super-core” 116 provides more details about the three dimensional structure of the imaged region by using more detailed wavefronts, while smaller numbers of cores 115 per group 116 and no grouping at all provide higher resolution.
- the grouping of cores 115 may hence be designed or modified according to spatially and temporally changing imaging requirements.
- cores 115 may be handled by processor 170 group-wise with respect to the radiation delivered therethrough, to implement each group 116 as a wavefront sensor.
- the allocation of cores 115 to core groups 116 may be carried out dynamically, e.g., by processor 170 . Additionally, grouping considerations may accompany other considerations regarding imaging performance such as suggested techniques for enhancing resolution and/or depth measurements.
- near field implementations may comprise sensing the light field between the cores (operating in multi-mode), e.g., measuring directional components of the radiation to yield 3D imaging.
- Light field sensing may be carried out groupwise with respect to the core grouping.
- endoscope fiber 100 may comprise multiple cores 115 that are not positioned at equal distances but interspaced unevenly (see FIG. 2A for a schematic illustration).
- Uneven (irregular) distribution of cores 115 e.g., a spatial distribution that does not coincide with the spatial distribution of pixels on detector 91
- the distribution of cores 115 and the interspaces across fiber 100 may be designed to optimize resolution enhancement using algorithmic and optical techniques. Indeed, increasing the distances between cores 115 may provide larger benefits from micro-scanning and application of other super resolution techniques.
- the optical design of fiber tip 101 may be configured to have working channel 112 positioned asymmetrically and not centrally within the cross section of the tip (not concentric to the imaging channel).
- the shape of working channel 112 may be configured to different than circular (e.g., elliptic, elongated, polygonal etc.) in order to better encode the optical transfer function (OTF).
- OTF optical transfer function
- the working channel shape may be configured to improve inversing the OTF and the algorithmic correcting of the image via the image post processing to yield a super resolved image.
- an increased depth of focus may be achieved in lens-less embodiments by selecting the best focal positions that can provide the sharpest contrast per each pixel in the generated image, from images captured at different tip positions with respect to tissue 70 .
- the best focus for each pixel may be selected from a plurality of images captured at different tip positions.
- optical elements 140 may be attached to or produced at distal fiber tip 100 (facing tissue 70 ). Optical elements 140 may be used to enhance imaging in both far-field imaging and near-field imaging. For example, optical elements 140 may be used to control the field of view, increasing it beyond the edges of tip 101 outwards and/or inwards (in case of a designed working channel void 112 ).
- FIGS. 4A-4D are high level schematic illustrations of hollow endoscope fiber 100 having optical elements 140 at distal tip 101 which compensate for the central void, according to some embodiments of the invention.
- void(s) 112 at the cross section of fiber 100 at tip 101
- various solutions are presented below for imaging a void-facing area 72 in addition to (or in place of) region 71 facing cores 115 .
- any type of target 70 may be imaged, e.g., tissue, specific anatomical members, bodily fluids, various stones or obstructions, tumors, foreign bodies etc.
- illumination source 160 of endoscope 105 and at least some of the optical elements may be configured to image at least a part of the area facing void(s) 112 (e.g., void-facing area 72 ) differently than a rest of the region facing tip 101 (e.g., core-facing region 71 ).
- the difference in the imaging may lie in any of polarization, wavelength, wavelength range and/or timing of the illumination. Non-limiting examples are presented in the following.
- FIG. 4A schematically illustrates in a non-limiting manner an annular multicore region 110 (with an inner radius R i and an outer radius R o ) having annularly arranged optical elements 140 . Similar principles may be applied to any geometric configuration of fiber tip 100 , e.g., any form thereof, any position and form of void(s) 112 , etc.
- optical elements 140 may comprise gradient index (GRIN) lenses cut at specified angles and glued at tip 101 of micro endoscope 105 .
- Each cut GRIN 140 may be cut and positioned to face a different direction in order to enhance the fiber's field of view (FOV) to equal the number of GRINs 140 multiplied by the FOV of each GRIN 140 (or, complementarily or alternatively, enhance the depth of field by configuring some of GRINs 140 to deliver radiation from different depths of field).
- the cut of the edge of GRIN lenses 140 may realize a prism coupling light into that specific GRIN from different predefined sectors of the field of view.
- Aspheric lenses may be used as alternative to GRIN lenses as optical elements 140 .
- FIGS. 4B-4D schematically illustrates three possible configurations, according to some embodiments of the invention.
- the large circle schematically represents the periphery of the total FOV of fiber tip 101 , which is the boundary of the imaged region facing the cores ( 71 ), while the small circles represent the fields of view of individual optical elements 140 , 141 , taken in a non-limiting illustrative case to be equal.
- tip FOV region 71 plus void-facing area 72
- optical elements 140 may be covered by equally spaced (in FIG. 4B eight) optical elements 140 each imaging a peripheral region 145
- an additional optical element 141 may be configured to image a central region 146 .
- Void-facing area 72 is thus covered centrally by region 146 and its periphery is covered by regions 145 .
- a larger number (in FIG. 4C twenty one) of optical elements 140 may be configured to have angles covering tip FOV in several concentric circular sets of imaging regions—in the illustrated example twelve peripheral regions 145 , eight intermediate regions 146 and one central region.
- annularly arranged optical elements 140 in FIG. 4D twenty five may be configured to have angles covering the tip FOV in a grid-like manner individual regions 145 partly overlapping and covering tip FOV and possible extending into a larger area. This disclosed method provides high flexibility in adapting fiber tip optical elements 140 to yield a required field of view.
- optical element 140 may comprise an annular lens coupled to an annular prism that directs light from the whole FOV into the annular lens.
- the center of FOV may be imaged using selective illumination. Illumination may be directed to the center of FOV and not to its periphery, and accompanying algorithms may be configured to process the detected signals to derive images of the FOV center (e.g., by processor 170 ).
- illumination having different polarizations may be used for the central FOV (e.g., void-facing area 72 ) and for the periphery of FOV (e.g., cores-facing region 71 ), so that the detected signal is spatially encoded by the difference in polarization, and may be decoded to create images of the whole FOV (see more elaborate explanation below).
- Optical elements 140 may be birefringent to directly differently polarized illumination to different geometric areas.
- void 112 may be eccentric or divided into eccentric voids, leaving rooms for ventral cores to image the center of the FOV directly.
- cores 115 may unequally or non-uniformly spaced within fiber 100 , e.g., such that the positions of cores 115 do not coincide with the uniform spatial sampling matrix of the pixels of detector 91 positioned outside the body.
- the lack of coinciding between the two grids may be utilized to apply geometric super resolving algorithms to improve the quality of the captured image (resembling in a sense the micro-scanning technique).
- fiber 100 may exhibit multicore designs having a low fill factor (the fill factor is the ratio between the core area and the square of the distance between cores, the latter termed pitch).
- the core diameter may range between 0.4-2.5 ⁇ m and the pitch may range between 2-10 ⁇ m to yield a range of low fill factors (1/(pitch/core diameter) 2 ), e.g., fill factors between 1/4 and 1/16.
- the fill factor is low (e.g., below 1/4, below 1/9, e.g., 1/16)
- simple movement of tip 101 of the micro endoscope e.g., movement amplitude may equal at least the pitch, e.g.
- micro-scanning concept enable implementation of the micro-scanning concept to significantly increase the geometric resolution of the device.
- the micro scanning procedure cannot increase the geometric resolution of the image but rather only to perform over-sampling of the image—because the point spread function (PSF) of the sampling pixel/core itself limits as a spatial low pass the obtainable resolution.
- PSF point spread function
- spatial scanning methods and temporal scanning methods according to the present disclosure may be combined and adapted to imaging requirements.
- illumination channel 85 may have time-varying optics which realizes a spatial scanning of the illumination spot.
- the spatial illumination scanning may be used to construct a wide field image having large field of view which is not affected by the working channel positioned in the center of the tip even if the tip is in near field with respect to the inspected tissue.
- processor 170 may be configured to process into images radiation delivered from the imaging region through cores 115 to detector 91 and possibly to implement super-resolution algorithms on the detected radiation.
- inspected tissue 70 may be illuminated by a tunable laser (e.g., as laser source 64 ) as illumination source 160 .
- a set of spatial images of tissue 70 may be captured, each image corresponding to a different wavelength.
- the resulting is hyperspectral image may be used for identification of specific types of tissues (e.g., cancerous tissue) to enhance the imaging.
- fiber endoscope 105 may provide diagnostic possibilities carried out using different wavelengths (in a specified diagnostic wavelength range, such as infrared wavelengths used to measure hemoglobin oxygenation) that are used for specific purposes and not necessarily for the imaging illumination.
- multiple narrowband wavelengths 62 may be used to provide diagnostics with one or more wavelength combinations 174 by processor 170 .
- Such combinations may be achieved by using sources with fixed spectral ranges and/or tunable source(s) to change temporally the spectral composition of illumination 85 .
- diagnostics which may be achieved by wavelength combinations 174 include biopsy (diagnostics of removed tissue) and characterization of biological tissues in situ e.g., by measuring reflectance at different and very specific wavelengths.
- a non-limiting example includes pulse oximetry which may be extracted by measuring a ratio of absorption at wavelengths of 600-750 nm (e.g., at 660 nm) and 850-1000 nm (e.g., at 910 nm), e.g., as two distinct wavelength (ranges) 62 , utilizing the different spectral absorption curves of HbO 2 and Hb.
- wavelengths and wavelength bands may be changed during the procedure, manually or automatically, to adapt to different stages in the procedure and different imaging requirements with respect e.g., to spatial or temporal parameters, encountered site and tissue, etc.
- single wavelength bands may be illuminated and analyzed separately, to enhance the derived information.
- Given wavelength bands may be used to illuminate the target from different directions to yield more detailed spatial information.
- working channel 112 of endoscope 105 configured as an ureteroscope may be used to suck out large kidney stones and attach the stones by suction to tip 101 of the endoscope.
- Treatment laser possibly incorporated in fiber 100 , see FIGS. 3A-3C
- Suction may be applied through working channel 112
- the imaging may be used to provide feedback regarding the efficiency of the suction and the treatment. For example, intensive treatment may tend to overcome the suction and release the attached stone.
- the imaging may be used to detect the development of stone disengagement from fiber tip 101 and to adjust suction and/or applied energy respectively.
- splitting of energy application into several fibers as described above may provide more uniform treatment of the stone that employs lower energy concentration at any one point of the stone. Energy application intensity may be regulated at each of the energy sources to avoid stone disengagement from the suction.
- working channel 112 of the ureteroscope may be used to inject liquid and to slightly change the optical conditions of fiber 100 such that effectively the focal length of lens 140 at tip 101 is changed and focal scanning can be realized to produce the sharpest possible image per each pixel in the image.
- Endoscope 105 may be configured as any type of endoscope and be used to handle any type of bodily stones or other obstructions, for example, by laser treatment source 66 .
- FIGS. 5A-5C are high level schematic illustrations of optical elements 140 , according to some embodiments of the invention.
- a polarizing optical element 150 e.g., a Glan Thompson prism
- Polarizing optical element 150 may be configured to increase FOV by polarization multiplexing beyond the limitations of optical element(s) 140 .
- Different fields of view 130 A, 130 B may be polarization-encoded, folded into endoscope fiber 100 and separated at the output (e.g., using a polarized beam splitter (PBS) 93 before reaching detectors 91 , 92 ).
- Polarization-encoding may be carried out using different linear polarization directions (e.g. with 450 therebetween), circular polarization etc.
- Polarization multiplexing may be used to increase the imaged area either laterally or centrally (see above), depending on the configurations of fiber 100 and the optics.
- Polarization multiplexing may be combined with temporal scanning of the field of view. Polarization multiplexing may be used to enhance three dimensional depth imaging in place or in addition to enlarging the field of view.
- Illumination source 160 for polarization multiplexing may be non-polarized (with separation to polarization component being carried out optically), or polarized and have both components.
- FIGS. 5B and 5C schematically illustrate embodiments for optical elements 140 , 150 at fiber tip 101 , namely an angle deflecting element 150 (e.g., a prism) and an imaging optical element 140 ( FIG. 5B ) and a combined configuration with a faceted GRIN lens 140 ( FIG. 5C ).
- an angle deflecting element 150 e.g., a prism
- an imaging optical element 140 FIG. 5B
- FIGS. 5C schematically illustrate embodiments for optical elements 140 , 150 at fiber tip 101 , namely an angle deflecting element 150 (e.g., a prism) and an imaging optical element 140 ( FIG. 5B ) and a combined configuration with a faceted GRIN lens 140 ( FIG. 5C ).
- certain parts of FOV may be imaged by different optical elements 140 (and respective cores 115 ) to enable optical triangulation, e.g., distance measurement from tip 101 and the tissue region.
- Such embodiments allow to trade-off FOV with depth information and thus dynamically allocate imaging resources (e.g., FOV—Field of View, DOF—Depth of Field) according to situation dependent needs.
- different polarizations may be used by different optical elements 140 imaging the same region, so that using polarization enhances depth information instead or in addition to extending the FOV (as explained above). Dynamic variation of polarization may be used to modify the optical performance of fiber 100 during operation.
- different wavelengths may be used by different optical elements 140 imaging the same region, so that using wavelength multiplexing (e.g., using a tunable laser as explained above) enhances depth information instead or in addition to extending the FOV (as explained above).
- wavelength multiplexing e.g., using a tunable laser as explained above
- Dynamic variation of color allocation may be used to modify the optical performance of fiber 100 during operation.
- multiple laser sources having different wavelengths e.g., with multiple narrowband wavelengths 62
- illumination source 160 e.g., four channels, three of which used to yield color imaging and the forth used to derive image depth information via triangulation computation.
- the wavelength used for the fourth channel may be identical to the wavelength used in one of the other three channels to facilitate or simplify the triangulation computation.
- endoscope 105 may be configured to use at least one non-imaged wavelength range, selected to provide additional depth of field or field of view information.
- polarization, wavelength or spatial multiplexing may be used to image a tissue region from different directions, to enable stereoscopic vision of the tissue region.
- Processor 170 may be configured to derive and provide stereo-imaging.
- illumination 85 may be improved in quality in different respects, such as its white light spectrum 177 and beam profile 184 , as disclosed above.
- endoscope 105 may be configured to provide two or more levels of resolution, allow balancing field of view information and depth of field information, or allow balance between any other image parameters by adapting the illumination and/or the image processing procedure disclosed herein.
- FIGS. 6A and 6B are high level schematic illustration of fiber cross sections with different configurations of the cores, according to some embodiments of the invention.
- FIG. 6C illustrates comparative experimental results of full core and hollow core fibers, according to some embodiments of the invention.
- the configuration of the cores may be designed to reduce crosstalk between cores 115 and to be less affected by its banding.
- crosstalk reduction may be achieved in the fabrication process by generating physical barriers between the cores or by using anti-crosstalk layer(s).
- Core spacing may be selected to reduce crosstalk between adjacent cores 115 below a specified threshold.
- crosstalk may be reduced by spacing the cores (e.g., by at least 4 between cores) and by increasing the refraction index difference between the cores and the cladding.
- the cores may be interspaced by structures such as air holes or doped polymer material (e.g., with incorporated nanoparticles).
- Cores 115 may be hollow, made of polymer material and/or include nanoparticles to control the refractive index. In certain embodiments, contrast may be enhanced by placing the hardware with the external holes array.
- an optical element e.g., optical element 94
- the optical element may comprise an intensity mask having a value of one for all core locations and a value of zero for all cladding locations to make all and only information from the cores to propagate to detector 91 .
- the difference in the refraction index between cores 115 and cladding 113 may be designed to be large enough, and/or intermediate elements 111 may be introduced to reduce interaction between radiation propagating in different cores 115 .
- Core 115 and/or cladding 113 and/or elements 111 may comprise polymer with incorporated nanoparticles. Due to plasmonic resonance of the nanoparticles at specific wavelengths an effective increased refraction index may be obtained for the doped material.
- the specific wavelengths may be selected to be close to wavelength bands (e.g., within a few nm, e.g., ⁇ 5 nm at most) of illumination source 160 (e.g., three or four color lasers 62 and/or 64 ). It is noted that as both the plasmonic resonance and the bandwidth of illumination lasers are narrow, they may be matched to yield an effectively increased refractive index by the nanoparticles at the illumination wavelengths.
- hollow cores through which no light coupling is obtained may be interlaced as intermediate elements 111 between cores 115 (see FIG. 6A ). Hollow cores 111 may be used to reduce the effective refraction index difference between light conducting solid cores 115 and their surrounding medium 113 .
- cores 115 may be hollow ( FIG. 6B ) and be isolated by doped or non-doped solid polymer. Hollow cores 115 (air holes) were shown to very significantly reduce material losses ( FIG. 6C ) and are thus exceptionally advantageous when using polymer fibers 100 which are characterized by relatively large losses compared to glass fibers.
- the main advantage of polymer fibers is their flexibility, enable strong bending which is required under certain endoscope applications (e.g., treating kidney stones as presented above).
- Fiber materials (for cladding 113 and intermediate elements 111 if any) and doping may be selected according to the required refractive indices and mechanical properties of fiber 100 , and may comprise various types of biocompatible (or not biocompatible, e.g., in non-medical uses) polymers, possibly doped with nanoparticles to influence the refractive indices. Either or both illumination wavelength ranges and types of nanoparticles may be selected to optimize the changes in the refractive indices to optimize the radiation transfer through the cores.
- core diameter D 1 , diameter of intermediate elements D 2 and distance between cores L may be configured to achieve specified optical performance parameters.
- FIG. 7 is a high level schematic flowchart illustrating a method 200 , according to some embodiments of the invention.
- Data processing stages and control stages may be implemented by respective processors and algorithms may be implemented by respective computer program product(s) comprising a computer usable medium having computer usable program code tangibly embodied thereon, the computer usable program code configured to carry out at least part of the respective stages.
- Method 200 comprises configuring an endoscope from a fiber with at least several hundred cores (stage 210 ), e.g., having a multi-core imaging region or a multi-core tip configured to deliver reflected illumination along the fiber for an external detector.
- Method 200 may comprise implementing near-field imaging (target imaging at the fiber tip) (stage 212 ) and/or implementing far-field imaging (Fourier plane at the fiber tip) (stage 214 ).
- method 200 may comprise configuring an endoscope from a plurality of fibers, grouped together, each having at least one hundred cores distributed at a fill factor smaller than 1/4, or even below 1/9, and at least one photonic illumination fiber, implementing three dimensional sensing by handling the cores group-wise with respect to radiation delivered therethrough, implementing super-resolved imaging by micro scanning over a pitch distance between the cores, and configuring at least one optical element at a distal tip of the fibers to enhance a field of view and/or a depth of field of the endoscope beyond a region facing a tip of the fibers and congruent thereto.
- Method 200 may comprise at least one of the following stages for reducing losses and/or cross talk between cores: incorporating in the cladding, nanoparticles with plasmonic resonances that are in proximity to illumination (and imaging) wavelengths (stage 220 ); interspacing cores by intermediate elements (possibly incorporating nanoparticles) having a different refractive index than the cores (stage 230 ), e.g., by 0.1; interspacing cores by air holes (stage 235 ) and configuring cores as air holes (stage 240 ), and may comprise reducing crosstalk between adjacent cores by interspacing them (stage 245 ).
- method 200 may further comprise incorporating one or more void(s) in the fiber as working channel(s) for treatment, suction and/or illumination (stage 250 ).
- method 200 may further comprise splitting treatment and/or illumination into several fibers operating collectively (stage 260 ) and/or incorporating additional fibers at the periphery of the fiber or of the void(s) (stage 265 ).
- Method 200 may comprise cooling incorporated fibers through the working channel (stage 267 ).
- method 200 may further comprise controlling treatment and/or suction optically or automatically using optical input during the treatment (stage 270 ), and treating bodily stones by the endoscope, e.g., kidney stones with an ureteroscope configuration (stage 275 ).
- Method 200 may further comprise using lens-less configurations, without any distal optical elements (stage 277 ) and/or using distal optical elements to control the field of view, the depth of field, implement image multiplexing and/or determine imaging parameters (stage 282 ), for example by attaching or producing optical element(s) at the fiber tip (stage 280 ).
- Method 200 may comprise enhancing the field of view and/or the depth of field of the endoscope beyond a region facing the tip of the fibers and congruent thereto (stage 285 ).
- Method 200 may comprise configuring the optical element(s) to image void-facing areas (stage 290 ), for example, using a lens with blocked aperture (stage 292 ); using multiple prisms which optically communicate with the cores (stage 295 ) and configuring the prisms to image void-facing areas (stage 300 ), e.g., associating each prism with one or more cores (stage 305 ); imaging void-facing areas using different polarization, wavelength, wavelength range and/or timing of the illumination (stage 310 ), in the former using birefringent optical elements for polarization multiplexing (stage 315 ).
- method 200 may further comprise implementing super-resolution algorithms (on the detected radiation) to enhance resolution, field of view and/or depth of field (stage 320 ).
- method 200 may further comprise reducing speckle patterns by using Barker codes for optimizing coherence modulation (stage 317 ).
- Method 200 may further comprise deriving 3D data using structured light illumination and processing (stage 319 ) and possibly enhancing super resolution processing using the patterned illumination (stage 322 ) as done, e.g., in time multiplexing super resolving concepts.
- Method 200 may further comprise beam-shaping the illumination at the distal tip (stage 324 ).
- method 200 may further comprise any of: distributing the cores irregularly (with respect to detector pixel order) over the tip cross section (stage 332 ), distributing the cores at a small fill factor (stage 334 ), and implementing micro-scanning of the region facing the tip (stage 336 ).
- method 200 may comprise enhancing images by optimizing pixel focus over different tip positions (stage 338 ), for example by selecting the best focus for each pixel from a plurality of images captured at different tip positions, and composing an enhanced imaged from the pixels at their selected best focus.
- method 200 may comprise handling the cores groupwise, possibly with dynamic allocation of cores to groups, to implement wavefront sensing by each group (stage 340 ).
- Method 200 may comprise implementing light field sensing. e.g., comparing light directions at different cores operating in near field and multi-mode.
- method 200 may further comprise using non-imaged wavelengths to provide additional field of view and/or depth of field information (stage 350 ).
- Method 200 may comprise collecting diagnostic data using, possibly non-imaged, diagnostic wavelength ranges (stage 360 ).
- method 200 may comprise configuring the endoscope as a laparoscope or an ureteroscope (stage 370 ).
- method 200 may further comprise configuring the illumination to have multiple narrowband wavelengths (stage 362 ) and possibly deriving diagnostic data from measurements at different wavelengths (stage 364 ) and/or enhancing super resolution processing using wavelength multiplexing with respect to the multiple narrowband sources (stage 366 ).
- method 200 may further comprise providing wideband white illumination using a PCF with zero dispersion point(s) selected to yield spectral broadening (stage 368 ).
- Method 200 may further comprise producing the fiber from standardized fiber modules (stage 380 ).
- method 200 comprises packaging the fiber modules into desired fiber cross section forms or configurations (stage 382 ).
- Method 200 may comprise modifying the spatial relations of the fiber modules along the fiber (stage 385 ), e.g., to have a circumferencial arrangement of fiber modules at the distal tip and a compact arrangement of fiber modules at the proximal tip of the fiber.
- method 200 may further comprise applying conductive coatings to some fiber modules, with other fiber modules as insulators (stage 387 ), e.g., for delivering electromagnetic energy to the fiber tip via the conductive coating, e.g., for heating the fiber tip (stage 390 ), elements associated with the fiber tip and/or a surroundings of the fiber tip.
- Method 200 may further comprise preventing fog upon and defogging the at least one optical element when required via a heated conductive coating thereof.
- FIGS. 8A-8E are high level schematic illustrations of experimental imaging results for bundled fibers, according to some embodiments of the invention.
- the imaging configuration is illustrated schematically in FIG. 1B .
- the presented results represent raw data, prior to the application of the image processing algorithms described above.
- FIGS. 8A-8C illustrate the imaging of three different targets using a two by two bundled fiber (as evident in the four sub-images, each received from one fiber module, having a side of 450p and ca. 23,000 cores per fiber module).
- the targets are respectively a resolution target, characters (person and doll) and an anatomy model.
- FIG. 8D illustrates imaging by a single multicore fiber, 1.8 mm in diameter having ca. 500,000 cores. Both configurations achieve very high resolution which is unattainable by current fibers.
- FIG. 8E illustrates a result achieved by applying image enhancement algorithms applied on the captured image.
- FIGS. 9A-9D are images that provide examples for performance of endoscope 105 , according to some embodiments of the invention.
- FIG. 9A illustrates an example for the multicore configuration of fiber 100 , in the illustrated non-limiting case, fiber 100 has an external diameter of 0.45 mm and includes more than 80,000 cores as well as an integrated illumination channel. The bottom image is a magnified view of the marked section in the top image.
- FIG. 9B illustrates image examples by endoscope 105 of a fingernail (a), a mouth (b), teeth (c) and a fingertip (d). These images were taken using fiber 100 with external diameter of 0.45 mm that includes more than 80,000 cores, and were processed as disclosed above.
- FIGS. 9A-9D are images that provide examples for performance of endoscope 105 , according to some embodiments of the invention.
- FIG. 9A illustrates an example for the multicore configuration of fiber 100 , in the illustrated non-limiting case, fiber 100 has an external diameter of 0.45 mm and includes more
- FIG. 9C and 9D illustrate examples for processing steps, namely the removal of artifacts and core traces (shown in FIG. 9C ) and the improvement of resolution and magnification (examples for the quality improvement of the results is shown in FIG. 9D ).
- disclosed image processing algorithms provide the cleaned image of FIG. 9D at a resolution of 300,000 pixels.
- the images obtained from multicore fibers 100 are steady and are not influences by fiber bending, outperforming prior art multimode fibers.
- very thin endoscope 105 provides high resolution medical imaging combined with high maneuverability and compliance with many medical situations.
- Certain embodiments of the invention may include features from different embodiments disclosed above, and certain embodiments may incorporate elements from other embodiments disclosed above.
- the disclosure of elements of the invention in the context of a specific embodiment is not to be taken as limiting their used in the specific embodiment alone.
Abstract
Description
- This application is a continuation of U.S. patent application Ser. No. 15/387,805 filed on Dec. 22, 2016, which is a continuation in part of U.S. patent application Ser. No. 14/805,915 filed on Jul. 22, 2015 which is now U.S. Pat. No. 9,661,986 registered on May 30, 2017, which claims priority under 35 U.S.C. § 119 to U.S. Provisional Patent Application No. 62/028,346 filed on Jul. 24, 2014 and to U.S. Provisional Patent Application No. 62/119,832 filed on Feb. 24, 2015, all of which are incorporated herein by reference in their entireties.
- The present invention relates to the field of endoscopy, and more particularly, to multicore fiber endoscopes.
- Endoscopes in various configurations allow efficient treatment of a range of medical problems, as well as means for manipulating different situations with limited access. Endoscope operations are challenging in that illumination, detection and treatment are confined to long and narrow operations modes. Fiber optics technology is a central enabler for such techniques, and fiber-based endoscope experience continuous improvements.
- The following is a simplified summary providing an initial understanding of the invention. The summary does not necessarily identify key elements nor limit the scope of the invention, but merely serves as an introduction to the following description.
- One aspect of the present invention provides an endoscope having a distal tip and a proximal tip, the endoscope comprising at least one multicore fiber module comprising at least one hundred cores distributed at a fill factor smaller than 1/4, an illumination source coupled to the at least one multicore fiber module and configured to deliver illumination thereto, at least one optical element, in optical communication with the cores, at the distal tip, a detector, in optical communication with the cores, at the proximal tip, and a processor configured to receive images from the detector; wherein the endoscope is configured to implement super-resolved imaging by micro scanning over a pitch distance between the cores, and wherein the endoscope is configured to implement three dimensional sensing by handling the cores group-wise with respect to radiation delivered therethrough, and to at least one of: enhance, by configuring the at least one optical element, a field of view of the endoscope beyond a region facing the cores at the distal tip, and enhance, by configuring the at least one optical element, a depth of field of the endoscope beyond a region congruent to the distal tip.
- These, additional, and/or other aspects and/or advantages of the present invention are set forth in the detailed description which follows; possibly inferable from the detailed description; and/or learnable by practice of the present invention.
- For a better understanding of embodiments of the invention and to show how the same may be carried into effect, reference will now be made, purely by way of example, to the accompanying drawings in which like numerals designate corresponding elements or sections throughout.
- In the accompanying drawings:
-
FIGS. 1A-1E are high level schematic illustrations of endoscope configurations according to some embodiments of the invention. -
FIGS. 2A-2C are high level schematic illustrations of fiber cross sections having a large number of cores in their electromagnetic propagation region(s), according to some embodiments of the invention. -
FIGS. 2D and 2E are high level schematic illustrations of fiber production by packing fiber modules, according to some embodiments of the invention. -
FIGS. 3A-3C are high level schematic cross section illustrations of fibers having working channels and additional channel positions for treatment or illumination fibers, according to some embodiments of the invention. -
FIG. 3D is a high level schematic illustration of a fiber with an assembled front lens, according to some embodiments of the invention. -
FIGS. 3E-3G are high level schematic illustrations of a defogging mechanism and its effects, according to some embodiments of the invention. -
FIGS. 4A-4D are high level schematic illustrations of hollow endoscope fibers having optical elements at the distal tip which compensate for the central void, according to some embodiments of the invention. -
FIGS. 5A-5C are high level schematic illustrations of optical elements, according to some embodiments of the invention. -
FIGS. 6A and 6B are high level schematic illustration of fiber cross sections with different configurations of the cores, according to some embodiments of the invention. -
FIG. 6C illustrates comparative experimental results of full core and hollow core fibers, according to some embodiments of the invention. -
FIG. 7 is a high level schematic flowchart illustrating a method, according to some embodiments of the invention. -
FIGS. 8A-8E are high level schematic illustrations of experimental imaging results for bundled fibers, according to some embodiments of the invention. -
FIGS. 9A-9D are images that provide examples for performance of the endoscope, according to some embodiments of the invention. - Prior to the detailed description being set forth, it may be helpful to set forth definitions of certain terms that will be used hereinafter.
- The terms “distal” and “proximal” as used in this application refer to the ends of the endoscope. The end and associated parts of the endoscope which are far from the endoscope's interface (detector or eye) and close to the imaged tissue and to its surroundings is termed the distal end, while the end and associated parts of the endoscope which are close to the endoscope's interface and are remote from the imaged tissue, being typically outside the body is termed the proximal end. The term “reflected” as used in this application refers to a change in a direction of an illumination wavefront which impacts one or more imaged object or tissue. The term “reflection” is understood broadly as any radiation gathered by the fiber, irrespective of the source of the illumination which is reflected by the object(s) and/or tissue(s).
- The term “near field imaging” as used in this application refers to the formation of an image (of imaged objects, tissues and/or their surroundings) at the distal end of the endoscope fiber, typically at the fiber's tip. The imaged is then typically transferred through the fiber to the detector, possibly through proximal optical elements. The term “near field imaging” may relate to different types of optical systems, including direct imaging without any optical elements between the imaged object or tissue and the fiber tip as well as to imaging through optical element(s) such as lenses.
- The term “far field imaging” as used in this application refers to the formation of a Fourier transform of imaged objects, tissues and/or their surroundings at the distal end of the endoscope fiber (e.g., the distal end of the endoscope fiber is at the aperture or pupil plane of the optical system), typically at the fiber's tip. The image of the imaged objects, tissues and/or their surroundings may be formed at the proximal end of the endoscope fiber, typically at the fiber's proximal tip or directly on the detector, possibly through proximal optical elements. The term “far field imaging” may relate to different types of optical systems. In one example, “far field imaging” may be direct in the sense that no optical elements are used between the imaged object or tissue and the distal fiber tip, which delivers radiation entering the fiber along the fiber to the detector at the proximal end of the fiber. In another example, “far field imaging” may be carried out with optical elements positioned between the imaged object or tissue and the distal fiber tip, with the distal fiber tip being at least approximately at the Fourier plane (also termed aperture plane and pupil plane in different contexts) of the optical elements.
- With specific reference now to the drawings in detail, it is stressed that the particulars shown are by way of example and for purposes of illustrative discussion of the preferred embodiments of the present invention only, and are presented in the cause of providing what is believed to be the most useful and readily understood description of the principles and conceptual aspects of the invention. In this regard, no attempt is made to show structural details of the invention in more detail than is necessary for a fundamental understanding of the invention, the description taken with the drawings making apparent to those skilled in the art how the several forms of the invention may be embodied in practice.
- Before at least one embodiment of the invention is explained in detail, it is to be understood that the invention is not limited in its application to the details of construction and the arrangement of the components set forth in the following description or illustrated in the drawings. The invention is applicable to other embodiments or of being practiced or carried out in various ways. Also, it is to be understood that the phraseology and terminology employed herein is for the purpose of description and should not be regarded as limiting.
- Endoscopes, multicore endoscope fibers and configuration and operation methods are provided. The fibers may have hundreds or thousands of cores and possibly incorporate working channel(s) and additional fibers. The fiber may be used at different optical configurations to capture images of tissue and objects at the distal tip and to enhance a wide range of optical characteristics of the images such as resolution, field of view, depth of field, wavelength ranges etc. Near-field imaging as well as far-field imaging may be implemented in the endoscopes and the respective optical features may be utilized to optimize imaging. Optical elements may be used at the distal fiber tip, or the distal fiber tip may be lens-less. Diagnostics and optical treatment feedback loops may be implemented and illumination may be adapted to yield full color images, depth estimation, enhanced field of view and/or depth of field and additional diagnostic data, as disclosed below.
- In the following, various embodiments of multicore endoscope fibers are disclosed. The described embodiments are roughly and not exclusively described in groups relating to the following traits. Certain endoscope embodiments may implement far field imaging (see
FIG. 1A below), e.g., have the image formed at the proximal end of the endoscope fiber, while certain endoscope embodiments may implement near field imaging (seeFIG. 1B below), e.g., have the image formed at the distal end of the endoscope fiber. Both far field and near field implementations, may have distal optical elements between the imaged objects or tissues and the distal fiber tip (seeFIG. 1C below), or may operate without such distal optical elements (seeFIG. 1D below). Each of the four combinations (far field with or without distal optical elements and near field with or without distal optical elements) has different features, advantages and disadvantages as exemplified in Table 1, and may be selected according to specific implementation scenarios. Alternation of the combination may be carried out between applications or in real time, to combine advantages of different configuration types. It is further noted that endoscopes may be designed to have several combinations, e.g., a part of the fiber face (or certain fiber modules) having distal optics for imaging far objects and another part of the fiber face (or other fiber modules) lacking distal optics for microscopic imaging. -
TABLE 1 Characteristics of different embodiments Far-field imaging Near-field imaging Distal fiber tip Fourier plane Image plane With distal Larger region of interest Larger field of view, Image optics multiplexing, larger energetic efficiency Larger field of view which may include working channel without compromising (central) regions in the field of view No distal Wavefront sensing, optical Auto focus and optical optics zooming zooming capabilities Simpler production and fiber management - Certain embodiments comprise lens-less embodiments in which the distal fiber tip lacks optical elements. Lens-less embodiments may implement either far-field or near-field imaging, and may utilize structural features to enhance optical resolution, apply super-resolution methods and retrieve wavefront information while reducing crosstalk between the cores.
- Endoscope embodiments may have full tip cross sections or have working channel(s) within the imaging fiber characterized by different configurations and uses, integrating additional fibers etc., in which case the cores and optical elements may be configured to overcome the reduction of the field of view due to the incorporation of the working channel.
- In the following, various configurations of the large number of cores in the fiber are disclosed, which provide solutions to various issues such as reducing crosstalk between the fibers, overcoming material losses, achieving enhanced resolution by different methods, providing required mechanical characteristics and optimizing the imaging performances of the endoscope fibers. The disclosed endoscopes may serve different purposes, e.g., may be designed as a laparoscope or an ureteroscope. It is noted that elements disclosed in the context of some of the embodiments are not necessarily limited to these embodiments but may be implemented within other embodiments as well.
-
FIGS. 1A-1E are high level schematic illustrations of endoscope configurations according to some embodiments of the invention. Proposedmicro endoscope 105 is constructed from large plurality of cores (e.g. one hundred cores or more, hundreds of cores, thousands of cores, in certain embodiments tens or hundreds of thousand cores per fiber or fiber module, reaching over a million cores in certain fiber endoscopes), each responsible for transferring a single or a large number of spatial degrees of freedom out of which at the output, proximal end (the one external to the patient body), a high resolution color image may be constructed.Multi-core fiber 100 exhibits a high degree of flexibility in its optical design, as exemplified below, which may be utilized and adapted for specific applications, for example for ureteroscopes with a large working channel and a small external diameter or for laparoscopes with a very high resolution obtained at a small external diameter. -
Endoscope 105 may be configured to carry out far-field imaging, near-field imaging or a combination of far-field imaging and near-field imaging. Irrespectively of the imaging mode,endoscope 105 may be configured to have one or moreoptical elements 140 at adistal tip 101 offiber 100 or have no optical elements betweentip 101 and imaged tissue(s) or object(s) 70. Certain embodiments may comprise removable or reconfigurableoptical elements 140 attip 101 and/oroptical elements 140 affecting only parts of the surface of distal tip 101 (e.g., sub-group(s) of the cores). - Certain embodiments comprise
endoscopes 105 having a plurality offibers 100, grouped together, each having at least one hundred cores distributed at a fill factor smaller than 1/4, or even smaller than 1/9, at least one photonic illumination fiber, and at least one optical element at a distal tip offibers 100, which may be configured to enhance a field of view and/or a depth of field ofendoscope 105 beyond a region facing a tip offibers 100 and congruent thereto (see details below).Endoscope 105 may be further configured to implement three dimensional sensing by handling the cores group-wise with respect to radiation delivered therethrough (see details below).Endoscope 105 may be further configured to super-resolved imaging by micro scanning over a pitch distance between the cores (see details below).Endoscope 105 may be configured to comprise a LED (light emitting diode) light source located atdistal tip 101 as the illumination source. -
FIG. 1A schematically illustrates far-field imaging, in which an image 73 (indicating any kind of electromagnetic signal reflected from tissue or object 70) is delivered throughtip 101 andfiber 100 to yieldimage 75 ondetector 91.Tip 101 may be a Fourier plane (also termed aperture plane or pupil plane) at which the Fourier transform 74 ofimage 73 entersfiber 100. It is noted that the Fourier plane may be located anywhere alongfiber 100 as well as distally or proximally tofiber 100, in different embodiments of the invention, and be optically transformed to image 75 ondetector 91. Alternatively or complementary,Fourier image 74 or derivatives thereof may be measured atdetector 91, and/or manipulated to enhance imaging parameters such as resolution, field of view and depth of focus, as non-limiting examples. Optical elements may be introduced distally or proximally tofiber 100 to modify or manipulate theradiation entering tip 101 and the radiation falling ondetector 91, respectively. -
FIG. 1B schematically illustrates near-field imaging, in whichimage 73yields image 75 atfiber tip 101.Image 75 is then delivered, possibly through optical elements, todetector 91 throughfiber 100. It is noted thatimage 75 may be formed withinfiber 100 and not necessarily exactly attip 101.Image 75 delivered viafiber 100 may be measured atdetector 91, and/or manipulated to enhance imaging parameters such as resolution, field of view and depth of focus, as non-limiting examples. Optical elements may be introduced distally or proximally tofiber 100 to modify or manipulate theradiation entering tip 101 and the radiation falling ondetector 91, respectively. -
FIG. 1C schematically illustrates optical configurations having one or more optical element(s) 140 at the distal end offiber 100, at proximity to imagedtissue 70. Optical element(s) 140 may be attached to tip 101 or may be somewhat distally removed from tip 101 (e.g., held by spacers at a distance therefrom). Eachoptical element 140 may be in optical communication with a respective core or a respective group of cores. Proximally,illumination 85 is delivered tofiber 100 by anillumination source 160, and reflected illumination (e.g., in far-field, in near-field or in an intermediate plane) is directed from the cores to adetector 91, e.g., via abeam splitter 90. Proximal optical elements may be set and used to manipulateillumination 85 and the reflected illumination, as symbolized below (FIG. 1D ) bylenses -
FIG. 1D schematically illustrates optical configurations having no optical element(s) (also termed below “lens-less” configurations) at the distal end offiber 100, so thatfiber tip 101 is used directly to deliver and receive illumination to and from imagedtissue 70.Illumination 85 is delivered tofiber 100 proximally, e.g., via anoptical element 84 such as a lens, and reflected illumination is directed todetector 91 via anotheroptical element 94, e.g., a lens. One or more processor(s) 170 may be configured to control the illumination and/or process the detected illumination, as well as control illumination and image beams in case there are controllable elements in the optical path. In certain embodiments, lens-less configurations may be configured to generate image at “contact mode”, e.g., with close proximity of the fiber tip to the examined tissue, to yield microscopic resolution determined by the sizes of the cores. - In certain embodiments, proximal optical elements 94 (and possibly
optical elements 84 too) may be variable and be used to adjust the plane and depth of focus of captured images in far-field imaging configurations, especially in lens-less configurations. -
FIG. 1E is a high level schematic block diagram illustrating endoscope configurations according to some embodiments of the invention. Various embodiments are illustrated, which may be stand-alone embodiments or be implemented in any combination thereof. In particular, various embodiments ofillumination source 160 and of configurations ofprocessor 170 are presented, which may be used to improve the spatial resolution, in particular when using super resolution algorithms, improve the beam quality and/or enhance the functionality ofendoscope 105 with respect to its medical uses and image quality. Embodiments illustrated inFIG. 1E may be applied to any embodiment ofendoscope 105 described herein. It is noted thatillumination source 160 may be configured to deliverillumination 85 through one or more dedicated illumination fiber(s) 102 and/or throughmulticore fiber 100. For example, illumination fiber(s) 102 may be multimode fiber(s), possibly made of glass fiber, which are associated withmulticore fiber 100, e.g., attached thereto or positioned in a cavity inmulticore fiber 100. Alternatively or complementarily, illumination fiber(s) 102 may be positioned to illuminatetissue 70 in any other spatial relation tomulticore fiber 100, possibly in no mechanical association therewith. In certain embodiments, one or more of illumination fiber(s) 102 may be single mode fibers. It is noted that the spatial relation between illumination fiber(s) 102 andmulticore fiber 100 may be configured to havemulticore fiber 100 receive radiation (illuminated by illumination fiber(s) 102) which is reflected offtissue 70 and/or transmitted throughtissue 70, depending on specific use conditions. - In some embodiments,
illumination source 160 may comprise acoherence modulator 162 configured to enableprocessor 170 to implement algorithms for improving super resolution results 172. For example,coherence modulator 162 may be configured to use a coherence modulation ofillumination 85 that reduces speckle patterns by modulating the coherence using Barker codes rather than random prior art modulation. Advantageously, using Barker codes may reduce the required number of modulation steps for a given reduction of speckle patterns due to the orthogonality between the Barker codes and other characteristics of their definition. Specific Barker codes may be selected to optimize their application. - In some embodiments,
illumination source 160 may comprise multiple narrowband wavelengths 62 (e.g., narrowband spectral ranges around specified wavelengths) which may be used inprocessors 170 configured to provide diagnostics using one ormore wavelength combinations 174, discussed below in more detail, and/or configured to implement wavelength multiplexingsuper resolution 176 by changing the ratios between specific wavelengths (e.g., between narrowband red, green and blue sources 62)—to achieve improved super resolution results 180. - In some embodiments,
illumination source 160 may comprise one or more photonic crystal fiber (PCF) 164 configured to deliver widebandwhite spectrum 177 into dedicated illumination fiber(s) 120 and/ormulticore fiber 100, e.g., utilizing supercontinuum effects (bandwidth broadening due to nonlinear effects) to provide white illumination of tissues that is closer to natural wideband illumination than illumination with narrowband red, green andblue sources 62 which are delivered toPCF 164.PCF 164 may be coupled to one or morenarrowband sources 62 and designed to have zero dispersion point(s) at the wavelengths of source(s) 62 to yield spectral broadening. Using several multiplenarrowband wavelengths 62 may provide widebandwhite spectrum 177 through a combination and merging of the broadened spectra of source(s) 62. Improvedwhite spectrum 177 may be advantageous to provide truer imaging colors beendoscope 105. - In some embodiments,
illumination source 160 may comprise structured light patternedillumination 168, which may be used inprocessor 170 configured to provide3D sensing 178 by analyzing the illuminated patterns on the tissue and/or improved super resolution results 180 by utilizing the parameters of the temporally changing spatially projected patterns ofillumination 168 to implement temporal multiplexingsuper resolution 179. - In some embodiments,
illumination source 160 may comprise one or more laser source(s) 64 (possibly narrowband sources 62) inillumination source 160 and at least onebeam shaping element 182 at the distal end ofmulticore fiber 100 which is configured to generate an optimizedbeam profile 184 to improveillumination 85. For example,beam profile 184 may comprise a uniform illumination distribution in space or a rectangular uniform profile (top hat illumination distribution), which are advantageous with respect to prior art Gaussian illumination distribution with respect to various parameters of the resulting images. The coherence of laser source(s) 64 may be used to shapeillumination beam 85 efficiently bybeam shaping element 182. In some embodiments, at least onebeam shaping element 182 may be set at the proximal end ofmulticore fiber 100. - In some embodiments,
illumination source 160 may comprise one or more laser treatment source(s) 66 which are configured to apply a specifiedtreatment 67 byendoscope 105, e.g., to a tissue. Forexample treatment 67 may be applied to kidney stones inendoscope 105 designed as an ureteroscope, as described below in more details. -
FIGS. 2A-2C are high level schematic illustrations of fiber cross sections having a large number ofcores 115 in their electromagnetic propagation region(s) 110, according to some embodiments of the invention. Fiber(s) 100 may comprise central or eccentric optical cores (110) and/or may have hollow, central or eccentric region(s) (112) that may be used for treatment such as energy delivery, suction, illumination, drug delivery etc. Illumination means (such as dedicated illumination fiber(s) 102), may be integrated in various ways within themulticore fibers 100. Selection of near-field or far-field configurations, as well as selection if and whichoptical elements 140 are inserted distally to the tip, may be carried out under consideration of the tradeoffs between the different applications (see e.g., Table 1 and other examples below). For example, considerations concerning production, use, optical characteristics and algorithmic parameters may be balanced differently at different embodiments to optimizeendoscope 105 to a wide range of performance and device requirements. -
Fiber 100 illustrated inFIG. 2A may have any form of cross section, e.g., square as illustrated in a non-limiting manner, round, hexagonal, elliptic etc. WhileFIG. 2A illustrates a solid cross section offiber 100,FIG. 2B illustrates hollow endoscope having a void 112 withinfiber 100 that may be used for different purposes as disclosed below (e.g., as a working channel for inserting a tool or carrying out suction, for incorporating additional fibers etc.).Fibers 100 may be square, round or have any other form, and void 112 may too have any shape and any position withinfiber 100, void(s) 112 andfiber 100 may have any dimensions (Ri, Ro, D, W etc.), and voids may also be multiple (e.g.,fiber 100 may enclose two or more voids), all designed according to requirements from the endoscope.FIG. 2C schematically illustratesmulticore fiber 100 withcores 115 grouped into “super core”groups 116 that may be configured to sense wavefronts in lens-less configurations, as explained below. -
Multicore fiber 100 may be made of biocompatible materials in case of medical uses, e.g., polymers such as PMMA (poly-methyl methacrylate) and PS (polystyrene) and may be flexible.Fiber 100 may also be made of non-compatible materials and be flexible or rigid in case of industrial uses.Fiber 100 may be configured to have a flexibility characterized by a Young's modulus smaller than 10 GPa and to be disposable.Fiber 100 may thus be more flexible than glass fiber (having a Young's modulus of about 65 GPa), and may reach PMMA flexibility (Young's modulus between 1.8 and 3.1 GPa) or higher flexibility. - Various embodiments compensate for the reduced transparency of polymer fibers with respect to glass fibers, using means such as fiber materials, configuration of cores and interspaces, number and sizes of cores, material modifications of different fiber parts, control over the number of propagation modes in
cores 115, optical means such as lenses or prisms at either side offiber 100 and their configuration, design and application of different types of illumination and algorithmic solutions, all of which are exemplified below in a non-limiting manner. The following disclosure also addresses ways to control cross talk between cores 115 (e.g., interaction effects between radiation propagating in adjacent cores 115) and ways to improve the information content and to enhance treatment-relevant information of the detected images. - Illumination may comprise coherent light or incoherent light, any spectral pattern (broad or narrow wavelength ranges, continuous or discrete ranged), polarized (in various patterns) or non-polarized light and different ranges in the visual or infrared ranges. Material differences between cores, interspaces and outer cladding may comprise different materials, using air cores or air interspaces, and doping any of the fiber regions to influence their refractive indices, as explained in more details below. It is noted that any of the embodiments presented below may be used in any of the other embodiments described herein, as long as they are compatible. Particularly, computational methods optical methods and fiber design considerations described in the context of any embodiment may be applied to other embodiments as well.
-
FIGS. 2D and 2E are high level schematic illustrations of fiber production by packing fiber modules, according to some embodiments of the invention.Multi-core fibers 100 may be produced using fiber modules orunits 117. Eachfiber module 117 is itself a multicore fiber, possibly configured to have uniform dimensions. Such embodiments are referred to as bundled fibers, and may bundle any number offiber modules 117 in any configuration (e.g., 2×2 modules, 3×3 modules etc.).Fiber module 117 may have any form, such as square, rectangular, round or elliptic, and may be packed intofibers 100 having a wide range of forms and configurations, Introducingfiber modules 117 having an intermediate dimension between cores or core groups and whole fiber 100 (eachmodule 117 may have e.g. tens, hundreds or thousands of cores) enables simpler production and higher flexibility on formingfiber 100 fromfiber modules 117. For example, as illustrated inFIG. 2D ,rectangular fiber 100 may be assembled from rectangularly arrangedsquare fiber modules 117, e.g., using apackage support 118A and a respectiveattachable cover 118B.Fiber modules 117 may simply be mechanically held bypackage support 118A and cover 118B at certain regions alongfiber 100 and/orfiber modules 117 may be glued together or otherwise attached at least at certain regions. In another example, illustrated inFIG. 2E ,fiber modules void 112. In certain embodiments,fiber modules optical elements 140 attached at fiber tip 110 (see e.g., below,FIGS. 4A-4D ). For example,fiber units 117A may be configured to cover a field of view in front of void 112 (e.g., be inclined inwards or have respective optical elements) whilefiber units 117B may be configured to cover a field of view laterally beyond tip 101 (e.g., be inclined outwards or have respective optical elements). For example, non-limiting inclination angles may be 5-20° inwards and 10-50° outwards. Respective packaging or attachment configurations may be applied to fixatefiber modules fiber modules fiber modules fiber 100, to enhance both the optical sensing at the distal end as well as the detection and processing at the proximal end. -
FIGS. 3A-3C are high level schematic cross section illustrations offiber 100 having workingchannel 112 andchannel positions 120 for treatment orillumination fibers 102, according to some embodiments of the invention. Workingchannel 112, depicted asvoid 112 withinfiber 100, is surrounded by electromagnetic propagationmulticore fiber region 110. Treatment and/or illumination fiber(s) 102 may be integrated intofiber 100 of the endoscope in a way that allows combined imaging and treatment using one fiber, immediate image feedback of the treatment etc. Such combination may be used e.g., as ureteroscope or as any other type of endoscope. In certain embodiments, positioning additional fibers inchannels 120 near workingchannel 112 may be configured to cool down the fibers (e.g., treatment fibers) by the liquids flowing through workingchannel 112. - In the illustrated examples, treatment or
illumination fibers 102 may be inserted at indicated positions 120 (e.g., grooves, or channels), e.g., at an inner wall ofmulti-core imaging region 110 in fluid communication with workingchannel 112, e.g., on the periphery of voids 112 (FIG. 3A , channel diameter e.g., ca. 250 μm), at an outer wall ofmulti-core imaging region 110 in fluid communication with the surroundings offiber 100, e.g., on the periphery of fiber 100 (FIG. 3B , channel diameter e.g., ca. 250 μm), within multi-core imaging region 110 (FIG. 3C , channel diameter e.g., ca. 200 μm), or combinations of these possibilities. Integration of the treatment orillumination fibers 102 may be carried out before, during or after production offiber 100. In certain embodiments, glass treatment orillumination fibers 102 may be inserted intogrooves 120 after pullingpolymer fiber 100. - In certain embodiments, treatment or illumination fibers may be configured and controlled to operate collectively, simultaneously or sequentially, to achieve a desired illumination and/or treatment. For example, the treatment channel may be split into several
low power channels 120 to have thinner channels and lower power delivery through each channel. Such configuration may enable increasing the mechanical flexibility of the endoscope, which is very important, e.g., in the field of ureteroscopy. Furthermore, the usage ofhollow channels 120 for inserting the external illumination or treatment fibers provides a device configuration exhibiting self-alignment. -
FIG. 3D is a high level schematic illustration offiber 100 with an assembledlens 119, according to some embodiments of the invention. A modular construction of fiber 100 (see e.g.,FIGS. 2D, 2E ) may be used to modify some offiber modules 117 to incorporate features intofiber 100 in a simpler manner than incorporating these features into a uniform fiber.fiber modules 117D may be configured in a modular, building block style manner to form various cross sectional organizations with respect to form and functionality of the endoscope. In the illustrated example of certain embodiments, twonon-adjacent fiber modules 117D may be coated with a conductor (e.g., a metal) while the rest offiber modules 117C may be uncoated (and insulating). Such configurations may be used to deliver electricity tofiber tip 101. For example, electromagnetic signals or electromagnetic radiation may be delivered viafiber modules 117D to adjacent tissues or to associated devices or components (e.g., checking equipment or endoscope instrumentation). In the illustrated example, electromagnetic energy may be delivered todistal lens 119 for heating it to prevent fogging upon entry to the body. In certain embodiment, an antenna structure (not shown) may be designed uponlens 119, which receives electromagnetic radiation to heatlens 119 without using contacts. In certain embodiments, radiofrequency (RF) treatment may be applied to tissue or objects surroundingfiber tip 101 via the conductive coating offiber modules 117D. -
FIGS. 3E-3F are additional high level schematic illustrations of adefogging mechanism 121 and its effects, according to some embodiments of the invention.FIG. 3E illustrateslens 119 coated by aconductive coating 122 connected to anelectric circuit 123 configured to heatlens 119 viacoating 122, to prevent fog and to defoglens 119 when required.FIG. 3F exemplifies image deterioration by fog accumulation—the top image (A) taken a short time after the beginning of fog accumulation, the bottom image (B) taken later, with the object, marked by an arrow, barely visible.FIG. 3G illustrates the image after defogging—both object and illumination spot are clear again. It is emphasized that asendoscope 105 may be designed to be very thin (e.g., 0.5 mm in diameter) while providing high resolution images, anddistal lens 119 may also be ultra-thin. The disclosed defogging mechanism provides effective control of the temperature oflens 119 using a small amount of electrical power, to prevent fogging and overcome an important prior art limitation. - In certain embodiments,
endoscope 105 may be operated in the far field (FIG. 1A ) or in the near field (FIG. 1B ) by properly adapting the focal length of the external optics (the one outside the patient's body, e.g.,optical elements 84, 94) to the working distance of treatedtissue 70 from the distal tip of the endoscope.Fiber 100 may be configured to deliver full images even with workingchannel 112 in the middle of the imaging surface by employing far field imaging, e.g., usingimaging lens 94 adapted to have a central blocked aperture. - In far field imaging configurations having
lens-less fiber tip 101, obtained images may have a number of pixels that is not related to the number ofcores 115, enhancing the image resolution with respect to near field embodiments. For example, certain embodiments comprise using asdetector 91 an integral imaging sensor capable of sensing wavefront or the 3D topography of inspectedtissue 70. In such embodiments,cores 115 may be configured to have a small number of possible spatial modes, resembling the Shack-Hartmann interferometer or a wavefront sensor. - In certain embodiments,
cores 115 may be grouped into “super-cores” 116 (seeFIG. 2C ), each comprising a group ofadjacent cores 115. Each “super-core” 116 may be handled as a single wavefront sensing element which delivers information about the wavefront by comparing radiation propagating throughindividual core members 115 within each “super-core” 116 (or light field sensing. e.g., comparing light directions at different cores operating in near field and multi-mode). The grouping ofcores 115 into “super-cores” 116 may be uniform across the face offiber 100 or be variable, some core groups being larger than others, see e.g., the larger central core group inFIG. 2C ). - The grouping of
cores 115 may be changed in time according to imaging performance preferences, based e.g., on an even (or uneven) distribution ofcores 115 acrossfiber 110. It is noted that in such configurations a tradeoff exists between depth measurements and resolution. A larger number ofcores 115 in each “super-core” 116 provides more details about the three dimensional structure of the imaged region by using more detailed wavefronts, while smaller numbers ofcores 115 pergroup 116 and no grouping at all provide higher resolution. The grouping ofcores 115 may hence be designed or modified according to spatially and temporally changing imaging requirements. Complementarily,cores 115 may be handled byprocessor 170 group-wise with respect to the radiation delivered therethrough, to implement eachgroup 116 as a wavefront sensor. The allocation ofcores 115 tocore groups 116 may be carried out dynamically, e.g., byprocessor 170. Additionally, grouping considerations may accompany other considerations regarding imaging performance such as suggested techniques for enhancing resolution and/or depth measurements. - In certain embodiments, near field implementations may comprise sensing the light field between the cores (operating in multi-mode), e.g., measuring directional components of the radiation to yield 3D imaging. Light field sensing may be carried out groupwise with respect to the core grouping.
- In certain embodiments,
endoscope fiber 100 may comprisemultiple cores 115 that are not positioned at equal distances but interspaced unevenly (seeFIG. 2A for a schematic illustration). Uneven (irregular) distribution of cores 115 (e.g., a spatial distribution that does not coincide with the spatial distribution of pixels on detector 91) enables, when working in the far-field conditions, to obtain super resolved images since the sampling ofcores 115 in the aperture plane (Fourier plane) is not uniform and thus the sampling at the aperture plane does not affect the field of view or generate visible limitations in the image plane. The distribution ofcores 115 and the interspaces acrossfiber 100 may be designed to optimize resolution enhancement using algorithmic and optical techniques. Indeed, increasing the distances betweencores 115 may provide larger benefits from micro-scanning and application of other super resolution techniques. - In certain embodiments, the optical design of
fiber tip 101 may be configured to have workingchannel 112 positioned asymmetrically and not centrally within the cross section of the tip (not concentric to the imaging channel). The shape of workingchannel 112 may be configured to different than circular (e.g., elliptic, elongated, polygonal etc.) in order to better encode the optical transfer function (OTF). The working channel shape may be configured to improve inversing the OTF and the algorithmic correcting of the image via the image post processing to yield a super resolved image. - In certain near-field imaging embodiments, an increased depth of focus may be achieved in lens-less embodiments by selecting the best focal positions that can provide the sharpest contrast per each pixel in the generated image, from images captured at different tip positions with respect to
tissue 70. The best focus for each pixel may be selected from a plurality of images captured at different tip positions. - In certain embodiments,
optical elements 140 may be attached to or produced at distal fiber tip 100 (facing tissue 70).Optical elements 140 may be used to enhance imaging in both far-field imaging and near-field imaging. For example,optical elements 140 may be used to control the field of view, increasing it beyond the edges oftip 101 outwards and/or inwards (in case of a designed working channel void 112). -
FIGS. 4A-4D are high level schematic illustrations ofhollow endoscope fiber 100 havingoptical elements 140 atdistal tip 101 which compensate for the central void, according to some embodiments of the invention. In embodiments with void(s) 112 at the cross section offiber 100 attip 101, various solutions are presented below for imaging a void-facingarea 72 in addition to (or in place of)region 71 facingcores 115. It is noted that any type oftarget 70 may be imaged, e.g., tissue, specific anatomical members, bodily fluids, various stones or obstructions, tumors, foreign bodies etc. - In certain embodiments,
illumination source 160 ofendoscope 105 and at least some of the optical elements (e.g., tipoptical elements 140, proximaloptical elements 84, 94) may configured to image at least a part of the area facing void(s) 112 (e.g., void-facing area 72) differently than a rest of the region facing tip 101 (e.g., core-facing region 71). The difference in the imaging may lie in any of polarization, wavelength, wavelength range and/or timing of the illumination. Non-limiting examples are presented in the following. -
Multiple cores 115 may be used to generate a full image, overcoming the lack of cores inhollow region 112 and providing imaging (and illumination) oftissue 70 directly opposite to working channel 112 (void-facing area 72). For example,endoscope 105 may be configured to provide a 90° field of view offiber 100.FIG. 4A schematically illustrates in a non-limiting manner an annular multicore region 110 (with an inner radius Ri and an outer radius Ro) having annularly arrangedoptical elements 140. Similar principles may be applied to any geometric configuration offiber tip 100, e.g., any form thereof, any position and form of void(s) 112, etc. - In certain embodiments,
optical elements 140 may comprise gradient index (GRIN) lenses cut at specified angles and glued attip 101 ofmicro endoscope 105. Eachcut GRIN 140 may be cut and positioned to face a different direction in order to enhance the fiber's field of view (FOV) to equal the number ofGRINs 140 multiplied by the FOV of each GRIN 140 (or, complementarily or alternatively, enhance the depth of field by configuring some ofGRINs 140 to deliver radiation from different depths of field). The cut of the edge ofGRIN lenses 140 may realize a prism coupling light into that specific GRIN from different predefined sectors of the field of view. Aspheric lenses may be used as alternative to GRIN lenses asoptical elements 140. -
FIGS. 4B-4D schematically illustrates three possible configurations, according to some embodiments of the invention. The large circle schematically represents the periphery of the total FOV offiber tip 101, which is the boundary of the imaged region facing the cores (71), while the small circles represent the fields of view of individualoptical elements region 71 plus void-facing area 72) may be covered by equally spaced (inFIG. 4B eight)optical elements 140 each imaging aperipheral region 145, and an additionaloptical element 141 may be configured to image acentral region 146. Void-facingarea 72 is thus covered centrally byregion 146 and its periphery is covered byregions 145. In another example, a larger number (inFIG. 4C twenty one) ofoptical elements 140 may be configured to have angles covering tip FOV in several concentric circular sets of imaging regions—in the illustrated example twelveperipheral regions 145, eightintermediate regions 146 and one central region. In another example, annularly arranged optical elements 140 (inFIG. 4D twenty five) may be configured to have angles covering the tip FOV in a grid-like mannerindividual regions 145 partly overlapping and covering tip FOV and possible extending into a larger area. This disclosed method provides high flexibility in adapting fiber tipoptical elements 140 to yield a required field of view. - In certain embodiments,
optical element 140 may comprise an annular lens coupled to an annular prism that directs light from the whole FOV into the annular lens. - In certain embodiments, possibly without the ring of optical elements described above, the center of FOV may be imaged using selective illumination. Illumination may be directed to the center of FOV and not to its periphery, and accompanying algorithms may be configured to process the detected signals to derive images of the FOV center (e.g., by processor 170).
- In certain embodiments, illumination having different polarizations may be used for the central FOV (e.g., void-facing area 72) and for the periphery of FOV (e.g., cores-facing region 71), so that the detected signal is spatially encoded by the difference in polarization, and may be decoded to create images of the whole FOV (see more elaborate explanation below).
Optical elements 140 may be birefringent to directly differently polarized illumination to different geometric areas. - In certain embodiments, void 112 may be eccentric or divided into eccentric voids, leaving rooms for ventral cores to image the center of the FOV directly.
- In certain embodiments,
cores 115 may unequally or non-uniformly spaced withinfiber 100, e.g., such that the positions ofcores 115 do not coincide with the uniform spatial sampling matrix of the pixels ofdetector 91 positioned outside the body. The lack of coinciding between the two grids may be utilized to apply geometric super resolving algorithms to improve the quality of the captured image (resembling in a sense the micro-scanning technique). - Certain embodiments may implement micro scanning via the spatial core configuration. For example,
fiber 100 may exhibit multicore designs having a low fill factor (the fill factor is the ratio between the core area and the square of the distance between cores, the latter termed pitch). For example, the core diameter may range between 0.4-2.5 μm and the pitch may range between 2-10 μm to yield a range of low fill factors (1/(pitch/core diameter)2), e.g., fill factors between 1/4 and 1/16. When the fill factor is low (e.g., below 1/4, below 1/9, e.g., 1/16), simple movement oftip 101 of the micro endoscope (e.g., movement amplitude may equal at least the pitch, e.g. a few microns) enable implementation of the micro-scanning concept to significantly increase the geometric resolution of the device. (It is noted that in case of imaging with large fill factor the micro scanning procedure cannot increase the geometric resolution of the image but rather only to perform over-sampling of the image—because the point spread function (PSF) of the sampling pixel/core itself limits as a spatial low pass the obtainable resolution.) In certain embodiments, spatial scanning methods and temporal scanning methods according to the present disclosure may be combined and adapted to imaging requirements. - In certain embodiments,
illumination channel 85 may have time-varying optics which realizes a spatial scanning of the illumination spot. The spatial illumination scanning may be used to construct a wide field image having large field of view which is not affected by the working channel positioned in the center of the tip even if the tip is in near field with respect to the inspected tissue. - In any of the embodiments,
processor 170 may be configured to process into images radiation delivered from the imaging region throughcores 115 todetector 91 and possibly to implement super-resolution algorithms on the detected radiation. - In certain embodiments, inspected
tissue 70 may be illuminated by a tunable laser (e.g., as laser source 64) asillumination source 160. A set of spatial images oftissue 70 may be captured, each image corresponding to a different wavelength. The resulting is hyperspectral image may be used for identification of specific types of tissues (e.g., cancerous tissue) to enhance the imaging. Thusfiber endoscope 105 may provide diagnostic possibilities carried out using different wavelengths (in a specified diagnostic wavelength range, such as infrared wavelengths used to measure hemoglobin oxygenation) that are used for specific purposes and not necessarily for the imaging illumination. For example, multiplenarrowband wavelengths 62 may be used to provide diagnostics with one ormore wavelength combinations 174 byprocessor 170. Such combinations may be achieved by using sources with fixed spectral ranges and/or tunable source(s) to change temporally the spectral composition ofillumination 85. Examples for diagnostics which may be achieved bywavelength combinations 174 include biopsy (diagnostics of removed tissue) and characterization of biological tissues in situ e.g., by measuring reflectance at different and very specific wavelengths. A non-limiting example includes pulse oximetry which may be extracted by measuring a ratio of absorption at wavelengths of 600-750 nm (e.g., at 660 nm) and 850-1000 nm (e.g., at 910 nm), e.g., as two distinct wavelength (ranges) 62, utilizing the different spectral absorption curves of HbO2 and Hb. - The selection of wavelengths and wavelength bands may be changed during the procedure, manually or automatically, to adapt to different stages in the procedure and different imaging requirements with respect e.g., to spatial or temporal parameters, encountered site and tissue, etc. In one example, single wavelength bands may be illuminated and analyzed separately, to enhance the derived information. Given wavelength bands may be used to illuminate the target from different directions to yield more detailed spatial information.
- In certain embodiments, working
channel 112 ofendoscope 105 configured as an ureteroscope may be used to suck out large kidney stones and attach the stones by suction to tip 101 of the endoscope. Treatment laser (possibly incorporated infiber 100, seeFIGS. 3A-3C ) may then be used to break the stones while the sucking stabilizes the stones and prevents them from moving around during the medical treatment. Suction may be applied through workingchannel 112, and the imaging may be used to provide feedback regarding the efficiency of the suction and the treatment. For example, intensive treatment may tend to overcome the suction and release the attached stone. The imaging may be used to detect the development of stone disengagement fromfiber tip 101 and to adjust suction and/or applied energy respectively. In this context, splitting of energy application into several fibers as described above may provide more uniform treatment of the stone that employs lower energy concentration at any one point of the stone. Energy application intensity may be regulated at each of the energy sources to avoid stone disengagement from the suction. - In certain embodiments, working
channel 112 of the ureteroscope may be used to inject liquid and to slightly change the optical conditions offiber 100 such that effectively the focal length oflens 140 attip 101 is changed and focal scanning can be realized to produce the sharpest possible image per each pixel in the image. -
Endoscope 105 may be configured as any type of endoscope and be used to handle any type of bodily stones or other obstructions, for example, bylaser treatment source 66. -
FIGS. 5A-5C are high level schematic illustrations ofoptical elements 140, according to some embodiments of the invention. In certain embodiments, a polarizing optical element 150 (e.g., a Glan Thompson prism) may be implemented at the end of fiber 100 (FIG. 5A ) in addition to imaging lens(es) 140 attip 101 of the micro-endoscope (e.g., a GRIN lens, aspheric lenses). Polarizingoptical element 150 may be configured to increase FOV by polarization multiplexing beyond the limitations of optical element(s) 140. Different fields ofview endoscope fiber 100 and separated at the output (e.g., using a polarized beam splitter (PBS) 93 before reachingdetectors 91, 92). Polarization-encoding may be carried out using different linear polarization directions (e.g. with 450 therebetween), circular polarization etc. Polarization multiplexing may be used to increase the imaged area either laterally or centrally (see above), depending on the configurations offiber 100 and the optics. Polarization multiplexing may be combined with temporal scanning of the field of view. Polarization multiplexing may be used to enhance three dimensional depth imaging in place or in addition to enlarging the field of view. Different processing algorithms may be applied to the signals ofdetectors Illumination source 160 for polarization multiplexing may be non-polarized (with separation to polarization component being carried out optically), or polarized and have both components. -
FIGS. 5B and 5C schematically illustrate embodiments foroptical elements fiber tip 101, namely an angle deflecting element 150 (e.g., a prism) and an imaging optical element 140 (FIG. 5B ) and a combined configuration with a faceted GRIN lens 140 (FIG. 5C ). - In certain embodiments, certain parts of FOV may be imaged by different optical elements 140 (and respective cores 115) to enable optical triangulation, e.g., distance measurement from
tip 101 and the tissue region. Such embodiments allow to trade-off FOV with depth information and thus dynamically allocate imaging resources (e.g., FOV—Field of View, DOF—Depth of Field) according to situation dependent needs. In certain embodiments, different polarizations may be used by differentoptical elements 140 imaging the same region, so that using polarization enhances depth information instead or in addition to extending the FOV (as explained above). Dynamic variation of polarization may be used to modify the optical performance offiber 100 during operation. In certain embodiments, different wavelengths may be used by differentoptical elements 140 imaging the same region, so that using wavelength multiplexing (e.g., using a tunable laser as explained above) enhances depth information instead or in addition to extending the FOV (as explained above). - Dynamic variation of color allocation may be used to modify the optical performance of
fiber 100 during operation. For example, multiple laser sources having different wavelengths (e.g., with multiple narrowband wavelengths 62) may be used asillumination source 160, e.g., four channels, three of which used to yield color imaging and the forth used to derive image depth information via triangulation computation. In certain embodiments, the wavelength used for the fourth channel may be identical to the wavelength used in one of the other three channels to facilitate or simplify the triangulation computation. - In certain embodiments,
endoscope 105 may be configured to use at least one non-imaged wavelength range, selected to provide additional depth of field or field of view information. In certain embodiments, polarization, wavelength or spatial multiplexing may be used to image a tissue region from different directions, to enable stereoscopic vision of the tissue region.Processor 170 may be configured to derive and provide stereo-imaging. - Moreover,
illumination 85 may be improved in quality in different respects, such as itswhite light spectrum 177 andbeam profile 184, as disclosed above. - In certain embodiments,
endoscope 105 may be configured to provide two or more levels of resolution, allow balancing field of view information and depth of field information, or allow balance between any other image parameters by adapting the illumination and/or the image processing procedure disclosed herein. -
FIGS. 6A and 6B are high level schematic illustration of fiber cross sections with different configurations of the cores, according to some embodiments of the invention.FIG. 6C illustrates comparative experimental results of full core and hollow core fibers, according to some embodiments of the invention. - The configuration of the cores (dimension, material, interspaces) may be designed to reduce crosstalk between
cores 115 and to be less affected by its banding. For example, crosstalk reduction may be achieved in the fabrication process by generating physical barriers between the cores or by using anti-crosstalk layer(s). Core spacing may be selected to reduce crosstalk betweenadjacent cores 115 below a specified threshold. For example, crosstalk may be reduced by spacing the cores (e.g., by at least 4 between cores) and by increasing the refraction index difference between the cores and the cladding. The cores may be interspaced by structures such as air holes or doped polymer material (e.g., with incorporated nanoparticles).Cores 115 may be hollow, made of polymer material and/or include nanoparticles to control the refractive index. In certain embodiments, contrast may be enhanced by placing the hardware with the external holes array. In certain embodiments, an optical element (e.g., optical element 94) may be added between the output offiber 100 and the imaging system and configured to block the output coming from cladding 113 thus transferring only the information going out fromoptical cores 115. The optical element may comprise an intensity mask having a value of one for all core locations and a value of zero for all cladding locations to make all and only information from the cores to propagate todetector 91. - In certain embodiments, the difference in the refraction index between
cores 115 andcladding 113 may be designed to be large enough, and/orintermediate elements 111 may be introduced to reduce interaction between radiation propagating indifferent cores 115.Core 115 and/orcladding 113 and/orelements 111 may comprise polymer with incorporated nanoparticles. Due to plasmonic resonance of the nanoparticles at specific wavelengths an effective increased refraction index may be obtained for the doped material. The specific wavelengths may be selected to be close to wavelength bands (e.g., within a few nm, e.g., ±5 nm at most) of illumination source 160 (e.g., three or fourcolor lasers 62 and/or 64). It is noted that as both the plasmonic resonance and the bandwidth of illumination lasers are narrow, they may be matched to yield an effectively increased refractive index by the nanoparticles at the illumination wavelengths. - In certain embodiments, hollow cores through which no light coupling is obtained may be interlaced as
intermediate elements 111 between cores 115 (seeFIG. 6A ).Hollow cores 111 may be used to reduce the effective refraction index difference between light conductingsolid cores 115 and their surroundingmedium 113. - In certain embodiments,
cores 115 may be hollow (FIG. 6B ) and be isolated by doped or non-doped solid polymer. Hollow cores 115 (air holes) were shown to very significantly reduce material losses (FIG. 6C ) and are thus exceptionally advantageous when usingpolymer fibers 100 which are characterized by relatively large losses compared to glass fibers. The main advantage of polymer fibers is their flexibility, enable strong bending which is required under certain endoscope applications (e.g., treating kidney stones as presented above). - Fiber materials (for cladding 113 and
intermediate elements 111 if any) and doping may be selected according to the required refractive indices and mechanical properties offiber 100, and may comprise various types of biocompatible (or not biocompatible, e.g., in non-medical uses) polymers, possibly doped with nanoparticles to influence the refractive indices. Either or both illumination wavelength ranges and types of nanoparticles may be selected to optimize the changes in the refractive indices to optimize the radiation transfer through the cores. In any of the embodiments, core diameter D1, diameter of intermediate elements D2 and distance between cores L may be configured to achieve specified optical performance parameters. -
FIG. 7 is a high level schematic flowchart illustrating amethod 200, according to some embodiments of the invention. Data processing stages and control stages may be implemented by respective processors and algorithms may be implemented by respective computer program product(s) comprising a computer usable medium having computer usable program code tangibly embodied thereon, the computer usable program code configured to carry out at least part of the respective stages. -
Method 200 comprises configuring an endoscope from a fiber with at least several hundred cores (stage 210), e.g., having a multi-core imaging region or a multi-core tip configured to deliver reflected illumination along the fiber for an external detector.Method 200 may comprise implementing near-field imaging (target imaging at the fiber tip) (stage 212) and/or implementing far-field imaging (Fourier plane at the fiber tip) (stage 214). - In certain embodiments,
method 200 may comprise configuring an endoscope from a plurality of fibers, grouped together, each having at least one hundred cores distributed at a fill factor smaller than 1/4, or even below 1/9, and at least one photonic illumination fiber, implementing three dimensional sensing by handling the cores group-wise with respect to radiation delivered therethrough, implementing super-resolved imaging by micro scanning over a pitch distance between the cores, and configuring at least one optical element at a distal tip of the fibers to enhance a field of view and/or a depth of field of the endoscope beyond a region facing a tip of the fibers and congruent thereto. -
Method 200 may comprise at least one of the following stages for reducing losses and/or cross talk between cores: incorporating in the cladding, nanoparticles with plasmonic resonances that are in proximity to illumination (and imaging) wavelengths (stage 220); interspacing cores by intermediate elements (possibly incorporating nanoparticles) having a different refractive index than the cores (stage 230), e.g., by 0.1; interspacing cores by air holes (stage 235) and configuring cores as air holes (stage 240), and may comprise reducing crosstalk between adjacent cores by interspacing them (stage 245). - In certain embodiments,
method 200 may further comprise incorporating one or more void(s) in the fiber as working channel(s) for treatment, suction and/or illumination (stage 250). - In certain embodiments,
method 200 may further comprise splitting treatment and/or illumination into several fibers operating collectively (stage 260) and/or incorporating additional fibers at the periphery of the fiber or of the void(s) (stage 265).Method 200 may comprise cooling incorporated fibers through the working channel (stage 267). In certain embodiments,method 200 may further comprise controlling treatment and/or suction optically or automatically using optical input during the treatment (stage 270), and treating bodily stones by the endoscope, e.g., kidney stones with an ureteroscope configuration (stage 275). -
Method 200 may further comprise using lens-less configurations, without any distal optical elements (stage 277) and/or using distal optical elements to control the field of view, the depth of field, implement image multiplexing and/or determine imaging parameters (stage 282), for example by attaching or producing optical element(s) at the fiber tip (stage 280).Method 200 may comprise enhancing the field of view and/or the depth of field of the endoscope beyond a region facing the tip of the fibers and congruent thereto (stage 285).Method 200 may comprise configuring the optical element(s) to image void-facing areas (stage 290), for example, using a lens with blocked aperture (stage 292); using multiple prisms which optically communicate with the cores (stage 295) and configuring the prisms to image void-facing areas (stage 300), e.g., associating each prism with one or more cores (stage 305); imaging void-facing areas using different polarization, wavelength, wavelength range and/or timing of the illumination (stage 310), in the former using birefringent optical elements for polarization multiplexing (stage 315). - In certain embodiments,
method 200 may further comprise implementing super-resolution algorithms (on the detected radiation) to enhance resolution, field of view and/or depth of field (stage 320). - In certain embodiments,
method 200 may further comprise reducing speckle patterns by using Barker codes for optimizing coherence modulation (stage 317).Method 200 may further comprise deriving 3D data using structured light illumination and processing (stage 319) and possibly enhancing super resolution processing using the patterned illumination (stage 322) as done, e.g., in time multiplexing super resolving concepts.Method 200 may further comprise beam-shaping the illumination at the distal tip (stage 324). - In certain embodiments,
method 200 may further comprise any of: distributing the cores irregularly (with respect to detector pixel order) over the tip cross section (stage 332), distributing the cores at a small fill factor (stage 334), and implementing micro-scanning of the region facing the tip (stage 336). In certain embodiments,method 200 may comprise enhancing images by optimizing pixel focus over different tip positions (stage 338), for example by selecting the best focus for each pixel from a plurality of images captured at different tip positions, and composing an enhanced imaged from the pixels at their selected best focus. - In certain embodiments,
method 200 may comprise handling the cores groupwise, possibly with dynamic allocation of cores to groups, to implement wavefront sensing by each group (stage 340).Method 200 may comprise implementing light field sensing. e.g., comparing light directions at different cores operating in near field and multi-mode. - In certain embodiments,
method 200 may further comprise using non-imaged wavelengths to provide additional field of view and/or depth of field information (stage 350).Method 200 may comprise collecting diagnostic data using, possibly non-imaged, diagnostic wavelength ranges (stage 360). In any of the embodiments,method 200 may comprise configuring the endoscope as a laparoscope or an ureteroscope (stage 370). - In certain embodiments,
method 200 may further comprise configuring the illumination to have multiple narrowband wavelengths (stage 362) and possibly deriving diagnostic data from measurements at different wavelengths (stage 364) and/or enhancing super resolution processing using wavelength multiplexing with respect to the multiple narrowband sources (stage 366). In certain embodiments,method 200 may further comprise providing wideband white illumination using a PCF with zero dispersion point(s) selected to yield spectral broadening (stage 368). -
Method 200 may further comprise producing the fiber from standardized fiber modules (stage 380). In certain embodiments,method 200 comprises packaging the fiber modules into desired fiber cross section forms or configurations (stage 382).Method 200 may comprise modifying the spatial relations of the fiber modules along the fiber (stage 385), e.g., to have a circumferencial arrangement of fiber modules at the distal tip and a compact arrangement of fiber modules at the proximal tip of the fiber. - In certain embodiments,
method 200 may further comprise applying conductive coatings to some fiber modules, with other fiber modules as insulators (stage 387), e.g., for delivering electromagnetic energy to the fiber tip via the conductive coating, e.g., for heating the fiber tip (stage 390), elements associated with the fiber tip and/or a surroundings of the fiber tip.Method 200 may further comprise preventing fog upon and defogging the at least one optical element when required via a heated conductive coating thereof. -
FIGS. 8A-8E are high level schematic illustrations of experimental imaging results for bundled fibers, according to some embodiments of the invention. The imaging configuration is illustrated schematically inFIG. 1B . The presented results represent raw data, prior to the application of the image processing algorithms described above.FIGS. 8A-8C illustrate the imaging of three different targets using a two by two bundled fiber (as evident in the four sub-images, each received from one fiber module, having a side of 450p and ca. 23,000 cores per fiber module). The targets are respectively a resolution target, characters (person and doll) and an anatomy model.FIG. 8D illustrates imaging by a single multicore fiber, 1.8 mm in diameter having ca. 500,000 cores. Both configurations achieve very high resolution which is unattainable by current fibers.FIG. 8E illustrates a result achieved by applying image enhancement algorithms applied on the captured image. -
FIGS. 9A-9D are images that provide examples for performance ofendoscope 105, according to some embodiments of the invention.FIG. 9A illustrates an example for the multicore configuration offiber 100, in the illustrated non-limiting case,fiber 100 has an external diameter of 0.45 mm and includes more than 80,000 cores as well as an integrated illumination channel. The bottom image is a magnified view of the marked section in the top image.FIG. 9B illustrates image examples byendoscope 105 of a fingernail (a), a mouth (b), teeth (c) and a fingertip (d). These images were taken usingfiber 100 with external diameter of 0.45 mm that includes more than 80,000 cores, and were processed as disclosed above.FIGS. 9C and 9D illustrate examples for processing steps, namely the removal of artifacts and core traces (shown inFIG. 9C ) and the improvement of resolution and magnification (examples for the quality improvement of the results is shown inFIG. 9D ). For example, disclosed image processing algorithms provide the cleaned image ofFIG. 9D at a resolution of 300,000 pixels. Advantageously, the images obtained frommulticore fibers 100 are steady and are not influences by fiber bending, outperforming prior art multimode fibers. Advantageously, verythin endoscope 105 provides high resolution medical imaging combined with high maneuverability and compliance with many medical situations. - In the above description, an embodiment is an example or implementation of the invention. The various appearances of “one embodiment”, “an embodiment”, “certain embodiments” or “some embodiments” do not necessarily all refer to the same embodiments.
- Although various features of the invention may be described in the context of a single embodiment, the features may also be provided separately or in any suitable combination. Conversely, although the invention may be described herein in the context of separate embodiments for clarity, the invention may also be implemented in a single embodiment.
- Certain embodiments of the invention may include features from different embodiments disclosed above, and certain embodiments may incorporate elements from other embodiments disclosed above. The disclosure of elements of the invention in the context of a specific embodiment is not to be taken as limiting their used in the specific embodiment alone.
- Furthermore, it is to be understood that the invention can be carried out or practiced in various ways and that the invention can be implemented in certain embodiments other than the ones outlined in the description above.
- The invention is not limited to those diagrams or to the corresponding descriptions. For example, flow need not move through each illustrated box or state, or in exactly the same order as illustrated and described.
- Meanings of technical and scientific terms used herein are to be commonly understood as by one of ordinary skill in the art to which the invention belongs, unless otherwise defined.
- While the invention has been described with respect to a limited number of embodiments, these should not be construed as limitations on the scope of the invention, but rather as exemplifications of some of the preferred embodiments. Other possible variations, modifications, and applications are also within the scope of the invention. Accordingly, the scope of the invention should not be limited by what has been described, but by the appended claims and their legal equivalents.
Claims (20)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/558,226 US20200015669A1 (en) | 2014-07-24 | 2019-09-02 | Illumination sources for multicore fiber endoscopes |
US18/383,558 US20240049956A1 (en) | 2014-07-24 | 2023-10-25 | Illumination sources for multicore fiber endoscopes |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462028346P | 2014-07-24 | 2014-07-24 | |
US201562119832P | 2015-02-24 | 2015-02-24 | |
US14/805,915 US9661986B2 (en) | 2014-07-24 | 2015-07-22 | Multicore fiber endoscopes |
US15/387,805 US10398294B2 (en) | 2014-07-24 | 2016-12-22 | Illumination sources for multicore fiber endoscopes |
US16/558,226 US20200015669A1 (en) | 2014-07-24 | 2019-09-02 | Illumination sources for multicore fiber endoscopes |
Related Parent Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/387,805 Continuation-In-Part US10398294B2 (en) | 2014-07-24 | 2016-12-22 | Illumination sources for multicore fiber endoscopes |
US15/387,805 Continuation US10398294B2 (en) | 2014-07-24 | 2016-12-22 | Illumination sources for multicore fiber endoscopes |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/383,558 Continuation US20240049956A1 (en) | 2014-07-24 | 2023-10-25 | Illumination sources for multicore fiber endoscopes |
Publications (1)
Publication Number | Publication Date |
---|---|
US20200015669A1 true US20200015669A1 (en) | 2020-01-16 |
Family
ID=58499109
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/387,805 Active 2035-12-14 US10398294B2 (en) | 2014-07-24 | 2016-12-22 | Illumination sources for multicore fiber endoscopes |
US16/558,226 Abandoned US20200015669A1 (en) | 2014-07-24 | 2019-09-02 | Illumination sources for multicore fiber endoscopes |
US18/383,558 Pending US20240049956A1 (en) | 2014-07-24 | 2023-10-25 | Illumination sources for multicore fiber endoscopes |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/387,805 Active 2035-12-14 US10398294B2 (en) | 2014-07-24 | 2016-12-22 | Illumination sources for multicore fiber endoscopes |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/383,558 Pending US20240049956A1 (en) | 2014-07-24 | 2023-10-25 | Illumination sources for multicore fiber endoscopes |
Country Status (1)
Country | Link |
---|---|
US (3) | US10398294B2 (en) |
Families Citing this family (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10537236B2 (en) * | 2013-01-17 | 2020-01-21 | Stryker Corporation | Anti-fogging device for endoscope |
JP6805937B2 (en) | 2017-03-31 | 2020-12-23 | ブラザー工業株式会社 | Water-based ink for inkjet recording |
JP6805936B2 (en) | 2017-03-31 | 2020-12-23 | ブラザー工業株式会社 | Water-based ink for inkjet recording |
JP7012291B2 (en) * | 2017-06-26 | 2022-01-28 | オリンパス株式会社 | Image processing device, operation method and program of image processing device |
EP3654824A4 (en) | 2017-07-17 | 2021-04-21 | Z Square Ltd. | Enhancing imaging by multicore fiber endoscopes |
US11213191B2 (en) | 2018-01-25 | 2022-01-04 | Canon U.S.A., Inc. | Optical fiber arrangement for endoscope |
JP7297891B2 (en) | 2018-07-19 | 2023-06-26 | アクティブ サージカル, インコーポレイテッド | Systems and Methods for Multimodal Sensing of Depth in Vision Systems for Automated Surgical Robots |
JP7405080B2 (en) * | 2018-08-13 | 2023-12-26 | ソニーグループ株式会社 | Medical system, medical light source device, and operating method of medical light source device |
US11510561B2 (en) * | 2018-08-21 | 2022-11-29 | Verily Life Sciences Llc | Endoscope defogging |
EP3667299B1 (en) * | 2018-12-13 | 2022-11-09 | Imec VZW | Multimodal imaging system |
US10839542B2 (en) * | 2019-01-29 | 2020-11-17 | The Chinese University Of Hong Kong | Systems and methods for 3D laparoscopic surface reconstruction |
WO2020176625A1 (en) * | 2019-02-26 | 2020-09-03 | Optecks, Llc | Colonoscopy system and method |
JP2022526626A (en) | 2019-04-08 | 2022-05-25 | アクティブ サージカル, インコーポレイテッド | Systems and methods for medical imaging |
US11925328B2 (en) | 2019-06-20 | 2024-03-12 | Cilag Gmbh International | Noise aware edge enhancement in a pulsed hyperspectral imaging system |
US11898909B2 (en) | 2019-06-20 | 2024-02-13 | Cilag Gmbh International | Noise aware edge enhancement in a pulsed fluorescence imaging system |
US11389066B2 (en) | 2019-06-20 | 2022-07-19 | Cilag Gmbh International | Noise aware edge enhancement in a pulsed hyperspectral, fluorescence, and laser mapping imaging system |
US11540696B2 (en) | 2019-06-20 | 2023-01-03 | Cilag Gmbh International | Noise aware edge enhancement in a pulsed fluorescence imaging system |
US11471055B2 (en) | 2019-06-20 | 2022-10-18 | Cilag Gmbh International | Noise aware edge enhancement in a pulsed fluorescence imaging system |
CN114599263A (en) | 2019-08-21 | 2022-06-07 | 艾科缇弗外科公司 | System and method for medical imaging |
FR3124878A1 (en) | 2021-07-01 | 2023-01-06 | Universite Grenoble Alpes | Imaging process, endoscope and computer program product |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3889662A (en) * | 1973-05-31 | 1975-06-17 | Olympus Optical Co | Endoscope |
US4656508A (en) * | 1984-06-08 | 1987-04-07 | Olympus Optical Co., Ltd. | Measuring endoscope |
US20080058629A1 (en) * | 2006-08-21 | 2008-03-06 | University Of Washington | Optical fiber scope with both non-resonant illumination and resonant collection/imaging for multiple modes of operation |
US20090002794A1 (en) * | 2007-06-29 | 2009-01-01 | Ethicon Endo-Surgery, Inc. | Receiver aperture broadening for scanned beam imaging |
US20090244260A1 (en) * | 2008-03-31 | 2009-10-01 | Hoya Corporation | Endoscope measuring 3-d profile |
US7704206B2 (en) * | 2002-06-05 | 2010-04-27 | Olympus Corporation | Endoscope that provides selection of each illumination mode of four different illumination modes |
US20140024951A1 (en) * | 2012-07-20 | 2014-01-23 | Intuitive Surgical Operations, Inc. | Annular vision system |
US20150015879A1 (en) * | 2012-03-29 | 2015-01-15 | Ecole Polytechnique Federale De Lausanne (Epfl) | Methods and apparatus for imaging with multimode optical fibers |
Family Cites Families (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6747795B2 (en) | 2000-06-30 | 2004-06-08 | The General Hospital Corporation | Fiber-coupled multiplexed confocal microscope |
US6747781B2 (en) | 2001-06-25 | 2004-06-08 | Silicon Light Machines, Inc. | Method, apparatus, and diffuser for reducing laser speckle |
US7377647B2 (en) | 2001-11-13 | 2008-05-27 | Philadelphia Retina Endowment Fund | Clarifying an image of an object to perform a procedure on the object |
WO2005119328A1 (en) | 2004-06-03 | 2005-12-15 | University Of Strathclyde | Improved scanning microscope source |
JP4892316B2 (en) | 2006-11-06 | 2012-03-07 | 株式会社フジクラ | Multi-core fiber |
US20090137893A1 (en) | 2007-11-27 | 2009-05-28 | University Of Washington | Adding imaging capability to distal tips of medical tools, catheters, and conduits |
JP2009288682A (en) | 2008-05-30 | 2009-12-10 | Olympus Medical Systems Corp | Objective optical system for endoscopes |
GB0812712D0 (en) | 2008-07-10 | 2008-08-20 | Imp Innovations Ltd | Improved endoscope |
JP2010253156A (en) | 2009-04-28 | 2010-11-11 | Fujifilm Corp | Endoscope system, endoscope, and endoscope driving method |
US20110301414A1 (en) | 2010-06-04 | 2011-12-08 | Robert Hotto | Intelligent endoscopy systems and methods |
JP5761674B2 (en) | 2010-06-08 | 2015-08-12 | 株式会社クラレ | Composite optical fiber and method of manufacturing the same |
EP2661211B1 (en) | 2011-01-05 | 2022-03-30 | Bar-Ilan University | Imaging system and method using multicore fiber |
WO2012132750A1 (en) | 2011-03-31 | 2012-10-04 | オリンパスメディカルシステムズ株式会社 | Scanning endoscope |
GB2504970A (en) | 2012-08-15 | 2014-02-19 | Swan Thomas & Co Ltd | Optical device and methods to reduce cross-talk |
WO2016013014A1 (en) * | 2014-07-24 | 2016-01-28 | Z Square Ltd. | Multicore fiber endoscopes |
EP3100669A1 (en) | 2015-06-05 | 2016-12-07 | University of Limerick | A spectroscopic imaging device |
-
2016
- 2016-12-22 US US15/387,805 patent/US10398294B2/en active Active
-
2019
- 2019-09-02 US US16/558,226 patent/US20200015669A1/en not_active Abandoned
-
2023
- 2023-10-25 US US18/383,558 patent/US20240049956A1/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3889662A (en) * | 1973-05-31 | 1975-06-17 | Olympus Optical Co | Endoscope |
US4656508A (en) * | 1984-06-08 | 1987-04-07 | Olympus Optical Co., Ltd. | Measuring endoscope |
US7704206B2 (en) * | 2002-06-05 | 2010-04-27 | Olympus Corporation | Endoscope that provides selection of each illumination mode of four different illumination modes |
US20080058629A1 (en) * | 2006-08-21 | 2008-03-06 | University Of Washington | Optical fiber scope with both non-resonant illumination and resonant collection/imaging for multiple modes of operation |
US20090002794A1 (en) * | 2007-06-29 | 2009-01-01 | Ethicon Endo-Surgery, Inc. | Receiver aperture broadening for scanned beam imaging |
US20090244260A1 (en) * | 2008-03-31 | 2009-10-01 | Hoya Corporation | Endoscope measuring 3-d profile |
US20150015879A1 (en) * | 2012-03-29 | 2015-01-15 | Ecole Polytechnique Federale De Lausanne (Epfl) | Methods and apparatus for imaging with multimode optical fibers |
US20140024951A1 (en) * | 2012-07-20 | 2014-01-23 | Intuitive Surgical Operations, Inc. | Annular vision system |
Also Published As
Publication number | Publication date |
---|---|
US20240049956A1 (en) | 2024-02-15 |
US20170100024A1 (en) | 2017-04-13 |
US10398294B2 (en) | 2019-09-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20240049956A1 (en) | Illumination sources for multicore fiber endoscopes | |
US9661986B2 (en) | Multicore fiber endoscopes | |
US11460685B2 (en) | Systems and methods for three-dimensional imaging | |
JP7107944B2 (en) | Spectrally Encoded Forward View Endoscope and Spectrally Encoded Multiview Endoscope, Probe, and Imager | |
CN106028909B (en) | For colored endoscopic device | |
WO2018116302A1 (en) | Illumination sources for multicore fiber endoscopes | |
US10610087B2 (en) | Apparatus, systems, and methods for biomedical imaging and stimulation | |
US11640027B2 (en) | Enhancing imaging by multicore fiber endoscopes | |
US20150009357A1 (en) | Extended depth of focus for high-resolution optical image scanning | |
US20230110978A1 (en) | Enhanced multicore fiber endoscopes | |
US20190191979A1 (en) | Method and system for imaging internal medium | |
KR20210093245A (en) | Multimode waveguide imaging | |
CN104352216B (en) | Endoscope's illumination spectra selecting arrangement and ultraphotic spectrum endoscopic imaging system | |
Ozbay et al. | Miniature Multiphoton Microscopes for Recording Neural Activity in Freely Moving Animals |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |