US20160235482A1 - Methods and systems for intraoperatively confirming location of tissue structures - Google Patents
Methods and systems for intraoperatively confirming location of tissue structures Download PDFInfo
- Publication number
- US20160235482A1 US20160235482A1 US15/023,120 US201415023120A US2016235482A1 US 20160235482 A1 US20160235482 A1 US 20160235482A1 US 201415023120 A US201415023120 A US 201415023120A US 2016235482 A1 US2016235482 A1 US 2016235482A1
- Authority
- US
- United States
- Prior art keywords
- tissue
- image data
- skeletal structure
- intraoperative
- vicinity
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 117
- 210000003484 anatomy Anatomy 0.000 claims abstract description 53
- 230000001747 exhibiting effect Effects 0.000 claims abstract description 36
- 238000012014 optical coherence tomography Methods 0.000 claims abstract description 25
- 230000008520 organization Effects 0.000 claims abstract description 17
- 230000010287 polarization Effects 0.000 claims abstract description 13
- 210000001519 tissue Anatomy 0.000 claims description 175
- 230000002792 vascular Effects 0.000 claims description 78
- 238000000701 chemical imaging Methods 0.000 claims description 31
- 238000002595 magnetic resonance imaging Methods 0.000 claims description 26
- 210000005166 vasculature Anatomy 0.000 claims description 19
- 210000000988 bone and bone Anatomy 0.000 claims description 16
- 238000002591 computed tomography Methods 0.000 claims description 14
- 210000002435 tendon Anatomy 0.000 claims description 11
- 210000004204 blood vessel Anatomy 0.000 claims description 8
- 210000003041 ligament Anatomy 0.000 claims description 8
- 230000005055 memory storage Effects 0.000 claims description 8
- 210000000845 cartilage Anatomy 0.000 claims description 6
- 210000001525 retina Anatomy 0.000 claims description 6
- 239000002872 contrast media Substances 0.000 claims description 5
- 210000003238 esophagus Anatomy 0.000 claims description 5
- 210000004379 membrane Anatomy 0.000 claims description 5
- 239000012528 membrane Substances 0.000 claims description 5
- 210000003205 muscle Anatomy 0.000 claims description 5
- 210000005036 nerve Anatomy 0.000 claims description 5
- 210000002105 tongue Anatomy 0.000 claims description 5
- 210000000515 tooth Anatomy 0.000 claims description 5
- 210000003437 trachea Anatomy 0.000 claims description 5
- 210000002808 connective tissue Anatomy 0.000 claims description 4
- 229910052688 Gadolinium Inorganic materials 0.000 claims description 3
- UIWYJDYFSGRHKR-UHFFFAOYSA-N gadolinium atom Chemical group [Gd] UIWYJDYFSGRHKR-UHFFFAOYSA-N 0.000 claims description 3
- 238000005286 illumination Methods 0.000 description 49
- 238000001356 surgical procedure Methods 0.000 description 40
- 230000003287 optical effect Effects 0.000 description 31
- 239000000835 fiber Substances 0.000 description 27
- 238000003384 imaging method Methods 0.000 description 24
- 230000007246 mechanism Effects 0.000 description 23
- 210000004556 brain Anatomy 0.000 description 22
- 238000001228 spectrum Methods 0.000 description 22
- 238000013461 design Methods 0.000 description 21
- 230000003595 spectral effect Effects 0.000 description 13
- 239000000126 substance Substances 0.000 description 12
- 238000010521 absorption reaction Methods 0.000 description 10
- 206010028980 Neoplasm Diseases 0.000 description 9
- 210000003128 head Anatomy 0.000 description 8
- 239000000523 sample Substances 0.000 description 8
- 241001269524 Dura Species 0.000 description 7
- 238000002271 resection Methods 0.000 description 7
- 210000004369 blood Anatomy 0.000 description 6
- 239000008280 blood Substances 0.000 description 6
- 238000007428 craniotomy Methods 0.000 description 6
- 230000004069 differentiation Effects 0.000 description 6
- 230000008569 process Effects 0.000 description 6
- 238000004458 analytical method Methods 0.000 description 5
- 230000008901 benefit Effects 0.000 description 4
- 230000000740 bleeding effect Effects 0.000 description 4
- 238000011065 in-situ storage Methods 0.000 description 4
- 108010054147 Hemoglobins Proteins 0.000 description 3
- 102000001554 Hemoglobins Human genes 0.000 description 3
- 238000004891 communication Methods 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 238000010191 image analysis Methods 0.000 description 3
- 230000033001 locomotion Effects 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000013439 planning Methods 0.000 description 3
- WZUVPPKBWHMQCE-UHFFFAOYSA-N Haematoxylin Chemical compound C12=CC(O)=C(O)C=C2CC2(O)C1C1=CC=C(O)C(O)=C1OC2 WZUVPPKBWHMQCE-UHFFFAOYSA-N 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 2
- 210000005013 brain tissue Anatomy 0.000 description 2
- 230000002490 cerebral effect Effects 0.000 description 2
- 238000002059 diagnostic imaging Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000009977 dual effect Effects 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 238000003709 image segmentation Methods 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000003211 malignant effect Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 210000003928 nasal cavity Anatomy 0.000 description 2
- 238000012634 optical imaging Methods 0.000 description 2
- 238000006213 oxygenation reaction Methods 0.000 description 2
- 230000035515 penetration Effects 0.000 description 2
- 238000012628 principal component regression Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 238000004088 simulation Methods 0.000 description 2
- 210000003625 skull Anatomy 0.000 description 2
- 238000004611 spectroscopical analysis Methods 0.000 description 2
- 238000002560 therapeutic procedure Methods 0.000 description 2
- 238000012384 transportation and delivery Methods 0.000 description 2
- 238000002604 ultrasonography Methods 0.000 description 2
- 238000012800 visualization Methods 0.000 description 2
- INGWEZCOABYORO-UHFFFAOYSA-N 2-(furan-2-yl)-7-methyl-1h-1,8-naphthyridin-4-one Chemical compound N=1C2=NC(C)=CC=C2C(O)=CC=1C1=CC=CO1 INGWEZCOABYORO-UHFFFAOYSA-N 0.000 description 1
- 241000894006 Bacteria Species 0.000 description 1
- 208000003174 Brain Neoplasms Diseases 0.000 description 1
- 102000008186 Collagen Human genes 0.000 description 1
- 108010035532 Collagen Proteins 0.000 description 1
- 208000034656 Contusions Diseases 0.000 description 1
- 229910000530 Gallium indium arsenide Inorganic materials 0.000 description 1
- 206010021143 Hypoxia Diseases 0.000 description 1
- 238000012307 MRI technique Methods 0.000 description 1
- 229910000661 Mercury cadmium telluride Inorganic materials 0.000 description 1
- 208000006964 Nevi and Melanomas Diseases 0.000 description 1
- 108010064719 Oxyhemoglobins Proteins 0.000 description 1
- 241000219061 Rheum Species 0.000 description 1
- 208000007536 Thrombosis Diseases 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 210000001361 achilles tendon Anatomy 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- VREFGVBLTWBCJP-UHFFFAOYSA-N alprazolam Chemical compound C12=CC(Cl)=CC=C2N2C(C)=NN=C2CN=C1C1=CC=CC=C1 VREFGVBLTWBCJP-UHFFFAOYSA-N 0.000 description 1
- 230000003466 anti-cipated effect Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000000149 argon plasma sintering Methods 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000001574 biopsy Methods 0.000 description 1
- 230000008499 blood brain barrier function Effects 0.000 description 1
- 210000001218 blood-brain barrier Anatomy 0.000 description 1
- 230000000747 cardiac effect Effects 0.000 description 1
- 210000004027 cell Anatomy 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 210000003850 cellular structure Anatomy 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000012569 chemometric method Methods 0.000 description 1
- 238000004140 cleaning Methods 0.000 description 1
- 229920001436 collagen Polymers 0.000 description 1
- 239000000470 constituent Substances 0.000 description 1
- 210000000877 corpus callosum Anatomy 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000001054 cortical effect Effects 0.000 description 1
- 238000005520 cutting process Methods 0.000 description 1
- 230000006378 damage Effects 0.000 description 1
- 238000007418 data mining Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 108010002255 deoxyhemoglobin Proteins 0.000 description 1
- 238000012938 design process Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000009792 diffusion process Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 230000002500 effect on skin Effects 0.000 description 1
- 238000001839 endoscopy Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- YQGOJNYOYNNSMM-UHFFFAOYSA-N eosin Chemical compound [Na+].OC(=O)C1=CC=CC=C1C1=C2C=C(Br)C(=O)C(Br)=C2OC2=C(Br)C(O)=C(Br)C=C21 YQGOJNYOYNNSMM-UHFFFAOYSA-N 0.000 description 1
- 239000012530 fluid Substances 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000004313 glare Effects 0.000 description 1
- 210000004884 grey matter Anatomy 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 230000000971 hippocampal effect Effects 0.000 description 1
- 230000007954 hypoxia Effects 0.000 description 1
- 238000000338 in vitro Methods 0.000 description 1
- 238000002347 injection Methods 0.000 description 1
- 239000007924 injection Substances 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 230000001788 irregular Effects 0.000 description 1
- 238000012804 iterative process Methods 0.000 description 1
- 230000003902 lesion Effects 0.000 description 1
- 239000007788 liquid Substances 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 210000004072 lung Anatomy 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 201000001441 melanoma Diseases 0.000 description 1
- 244000005700 microbiome Species 0.000 description 1
- 230000001423 neocortical effect Effects 0.000 description 1
- 210000004126 nerve fiber Anatomy 0.000 description 1
- 230000037311 normal skin Effects 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000000399 orthopedic effect Effects 0.000 description 1
- 201000008482 osteoarthritis Diseases 0.000 description 1
- 230000007170 pathology Effects 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000002035 prolonged effect Effects 0.000 description 1
- 239000010453 quartz Substances 0.000 description 1
- 238000002310 reflectometry Methods 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000002207 retinal effect Effects 0.000 description 1
- 230000001953 sensory effect Effects 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 238000007493 shaping process Methods 0.000 description 1
- VYPSYNLAJGMNEJ-UHFFFAOYSA-N silicon dioxide Inorganic materials O=[Si]=O VYPSYNLAJGMNEJ-UHFFFAOYSA-N 0.000 description 1
- 238000010186 staining Methods 0.000 description 1
- 238000013020 steam cleaning Methods 0.000 description 1
- 210000000130 stem cell Anatomy 0.000 description 1
- 238000003860 storage Methods 0.000 description 1
- 210000003478 temporal lobe Anatomy 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 210000003462 vein Anatomy 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00163—Optical arrangements
- A61B1/00194—Optical arrangements adapted for three-dimensional imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00163—Optical arrangements
- A61B1/00195—Optical arrangements with eyepieces
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/002—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor having rod-lens arrangements
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/04—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
- A61B1/042—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances characterised by a proximal camera, e.g. a CCD camera
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/04—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances
- A61B1/043—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor combined with photographic or television appliances for fluorescence imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/06—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements
- A61B1/0646—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor with illuminating arrangements with illumination filters
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/313—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor for introducing through surgical openings, e.g. laparoscopes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/00234—Surgical instruments, devices or methods, e.g. tourniquets for minimally invasive surgery
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/02—Detecting, measuring or recording pulse, heart rate, blood pressure or blood flow; Combined pulse/heart-rate/blood pressure determination; Evaluating a cardiovascular condition not otherwise provided for, e.g. using combinations of techniques provided for in this group with electrocardiography or electroauscultation; Heart catheters for measuring blood pressure
- A61B5/02007—Evaluating blood vessel condition, e.g. elasticity, compliance
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/40—Detecting, measuring or recording for evaluating the nervous system
- A61B5/4058—Detecting, measuring or recording for evaluating the nervous system for evaluating the central nervous system
- A61B5/4064—Evaluating the brain
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computed tomography [CT]
- A61B6/032—Transmission computed tomography [CT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/50—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment specially adapted for specific body parts; specially adapted for specific clinical applications
- A61B6/505—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment specially adapted for specific body parts; specially adapted for specific clinical applications for diagnosis of bone
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J3/00—Spectrometry; Spectrophotometry; Monochromators; Measuring colours
- G01J3/02—Details
- G01J3/0205—Optical elements not provided otherwise, e.g. optical manifolds, diffusers, windows
- G01J3/0229—Optical elements not provided otherwise, e.g. optical manifolds, diffusers, windows using masks, aperture plates, spatial light modulators or spatial filters, e.g. reflective filters
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J3/00—Spectrometry; Spectrophotometry; Monochromators; Measuring colours
- G01J3/28—Investigating the spectrum
- G01J3/2823—Imaging spectrometer
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01J—MEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
- G01J3/00—Spectrometry; Spectrophotometry; Monochromators; Measuring colours
- G01J3/28—Investigating the spectrum
- G01J3/42—Absorption spectrometry; Double beam spectrometry; Flicker spectrometry; Reflection spectrometry
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/107—Visualisation of planned trajectories or target regions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2051—Electromagnetic tracking systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2055—Optical tracking systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/30—Devices for illuminating a surgical field, the devices having an interrelation with other surgical devices or with a surgical procedure
- A61B2090/306—Devices for illuminating a surgical field, the devices having an interrelation with other surgical devices or with a surgical procedure using optical fibres
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/361—Image-producing devices, e.g. surgical cameras
- A61B2090/3616—Magnifying glass
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B2090/364—Correlation of different images or relation of image positions in respect to the body
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/373—Surgical systems with images on a monitor during operation using light, e.g. by using optical scanners
- A61B2090/3735—Optical coherence tomography [OCT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2505/00—Evaluating, monitoring or diagnosing in the context of a particular type of medical care
- A61B2505/05—Surgical care
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2576/00—Medical imaging apparatus involving image processing or analysis
- A61B2576/02—Medical imaging apparatus involving image processing or analysis specially adapted for a particular organ or body part
- A61B2576/026—Medical imaging apparatus involving image processing or analysis specially adapted for a particular organ or body part for the brain
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/30—Surgical robots
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0062—Arrangements for scanning
- A61B5/0066—Optical coherence imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/055—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves involving electronic [EMR] or nuclear [NMR] magnetic resonance, e.g. magnetic resonance imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/40—Detecting, measuring or recording for evaluating the nervous system
- A61B5/4029—Detecting, measuring or recording for evaluating the nervous system for evaluating the peripheral nervous systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4504—Bones
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4514—Cartilage
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4519—Muscles
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4523—Tendons
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4533—Ligaments
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4538—Evaluating a particular part of the muscoloskeletal system or a particular medical condition
- A61B5/4542—Evaluating the mouth, e.g. the jaw
- A61B5/4547—Evaluating teeth
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/45—For evaluating or diagnosing the musculoskeletal system or teeth
- A61B5/4538—Evaluating a particular part of the muscoloskeletal system or a particular medical condition
- A61B5/4542—Evaluating the mouth, e.g. the jaw
- A61B5/4552—Evaluating soft tissue within the mouth, e.g. gums or tongue
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7246—Details of waveform analysis using correlation, e.g. template matching or determination of similarity
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7271—Specific aspects of physiological measurement analysis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/50—Supports for surgical instruments, e.g. articulated arms
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
Definitions
- the present disclosure relates to imaging methods for use in minimally invasive therapy and image guided medical procedures using optical imaging, and more particularly, hyperspectral imaging.
- the optical absorption and scattering properties of biological tissue depend on both the chemical and structural properties of the tissue and the wavelength of the interacting light. How these absorption and scattering properties of tissue change as a function of light can be particularly useful, as it is often unique to chemicals or structures in the tissue (the spectrum of the tissue).
- the absorption features of oxy- and deoxy-hemoglobin can be used to measure the oxygenation of blood and tissue
- the scatter changes caused by difference cellular sizes can be used to detect precancerous and cancerous tissue.
- the field of measuring these changes in optical properties, as a function of light is known as spectroscopy and the device to measure the light at the various wavelengths is known as a spectrometer. Spectroscopy has found a wealth of current and potential applications in medicine.
- hyperspectral image analysis or hyperspectral imaging was applied to search for spectral differences between benign and malignant dermal tissue in routine hematoxylin eosin stained specimens (i.e., normal and abnormal skin, benign nevi and melanomas).
- Systems, methods and devices are provided for intraoperatively confirming location of tissue structures during medical procedures.
- preoperative image data of a patient's skeletal structure in a vicinity of an anatomical part undergoing a medical procedure is acquired.
- tissue intraoperative image data is acquired by scanning a selected region of tissue, in a vicinity of the skeletal structure using Polarization Sensitive-Optical Coherence Tomography (PS-OCT).
- PS-OCT Polarization Sensitive-Optical Coherence Tomography
- Regions of tissue exhibiting structural organization in the vicinity of the skeletal structure are identified from the intraoperative (PS-OCT) image data.
- Geometrically correlating and registering the intraoperative (PS-OCT) image data with the preoperative image data of the skeletal structure in the vicinity of the anatomical part is then performed using a priori known anatomical information about the regions of tissue exhibiting structural information.
- global preoperative image data of tissue in an anatomical part is acquired using contrast based magnetic resonance imaging and identifying. From the image data, a global vascular structure within the tissue is identified. After exposing tissue during a medical procedure in the anatomical part, intraoperative image data is acquired by scanning, using hyperspectral imaging, of a selected local region of the tissue. From this intraoperative hyperspectral image data, a local vascular structure in the selected local region of the tissue is located and identified. The global vascular image data is searched for identifying and locating a portion of the global vascular structure geometrically matching the local vascular structure. Upon identifying and locating matching vascular structure between the two imaging modalities, one or more local vascular structures in the selected local region of the tissue is registered with the global vascular structure within the tissue for confirming location of the one or more local vasculature structures.
- a computer implemented method of intraoperatively confirming location of organized tissue structures in relation to a patient's skeletal structure during a medical procedure comprising:
- PS-OCT Polarization Sensitive-Optical Coherence Tomography
- PS-OCT intraoperative
- the pre-operative image data of the skeletal structure in a vicinity of an anatomical part undergoing a medical procedure may be acquired using any one of computed tomography (CT), magnetic resonance imaging (MRI) and optical coherence tomography (OCT).
- CT computed tomography
- MRI magnetic resonance imaging
- OCT optical coherence tomography
- T1 magnetic resonance imaging T1 MRI
- the tissue exhibiting structural organization includes, ligaments, tendons, muscle, cartilage, connective membranes, nerves, retina, blood vessel walls, some bone structures, trachea, esophagus, tongue, teeth and other connective tissues.
- the a priori known anatomical information about the regions of tissue exhibiting structural information may include attachment points of tissue exhibiting structural information to the skeletal structure relative to landmark positions on the skeletal structure.
- a computer implemented system for intraoperatively confirming location of organized tissue structures in relation to a patient's skeletal structure during a medical procedure comprising:
- PS-OCT Polarization Sensitive-Optical Coherence Tomography
- a computer processor having a memory storage, said Polarization Sensitive-Optical Coherence Tomography being connected to the computer processor, said memory storage having stored therein preoperative image data of a patient's skeletal structure in a vicinity of an anatomical part undergoing a medical procedure, said memory storage having stored therein a priori known anatomical information about the regions of tissue exhibiting structural information;
- said computer processor being programmed with instructions to
- the present disclosure also provides a computer implemented method of intraoperatively confirming location of vasculature structures located below a surface tissue during a medical procedure, comprising:
- a method comprising the steps of:
- FIG. 1 shows an example navigation system to support minimally invasive access port-based surgery.
- FIG. 2 is block diagram illustrating system components of a navigation system.
- FIG. 3 is a flow chart illustrating the processing steps involved in a port-based surgical procedure using a navigation system.
- FIG. 4 is an example embodiment port based brain surgery using a video scope.
- FIG. 5A is an example embodiment of a video scope with camera coupler and illumination optics.
- FIG. 5B is an example embodiment of a fiber bundle used to deliver light from external light source to the video scope.
- FIG. 5C is an example embodiment of a video scope and illumination assembly.
- FIG. 6 illustrates an example imaging optical sub-system of the video scope.
- FIG. 7 illustrates the arrangement of illumination optics and filter wheel for wide field of view arrangement.
- FIG. 8A illustrates the non-uniform illumination obtained at the distal end of port with two illumination sources and a port with reflective surface.
- FIG. 8B illustrates the near-uniform illumination obtained at the distal end of the port with two illumination sources and a port with rough surface.
- FIG. 9 an example embodiment illustrating a standard hyperspectral imaging system.
- FIG. 10 is a flow chart illustrating a method to acquire hyperspectral data and white-light images in a multiplex fashion.
- FIG. 11 is an example embodiment illustrating imaging at specific wavelength bands.
- FIG. 12 shows an example, non-limiting implementation of computer control system.
- the terms, “comprises” and “comprising” are to be construed as being inclusive and open ended, and not exclusive. Specifically, when used in the specification and claims, the terms, “comprises” and “comprising” and variations thereof mean the specified features, steps or components are included. These terms are not to be interpreted to exclude the presence of other features, steps or components.
- exemplary means “serving as an example, instance, or illustration,” and should not be construed as preferred or advantageous over other configurations disclosed herein.
- the terms “about” and “approximately” are meant to cover variations that may exist in the upper and lower limits of the ranges of values, such as variations in properties, parameters, and dimensions. In one non-limiting example, the terms “about” and “approximately” mean plus or minus 10 percent or less.
- Port-based surgery is a minimally invasive surgical technique where a port is introduced to access the surgical region of interest using surgical tools. Unlike other minimally invasive techniques, such as laparoscopic techniques, the port diameter is larger than tool diameter. Hence, the tissue region of interest is visible through the port. Accordingly, exposed tissue in a region of interest at a depth few centimetres below the skin surface, and accessible through a narrow corridor in the port, may be visualized using externally positioned optical systems such as microscopes and video scopes.
- Tissue differentiation may be useful because surgeons do not have a quantitative means of effectively confirming tissues types during a surgical procedure.
- hyperspectral imaging has not been anticipated for intra-operative use in brain surgery because this method has a very limited depth of penetration in tissue and may not be effectively used transcranially.
- the narrow corridor in port-based surgery is often occluded when a vessel is accidentally cut.
- the surgeon may be required to stop his current surgical process (e.g. opening of dura, slight retraction of the sulcus for trans-sulcus navigation of port or resection of tumor tissue) and irrigate the cavity to get a better view of the cavity.
- bleeding also limits the surgeon from quickly identifying the location of bleeding so that the particular vessel wall can be coagulated to terminate bleeding.
- systems and methods are provided for utilizing optical imaging in minimally invasive port based surgical procedures.
- hyperspectral devices and methods are described for performing intraoperative tissue differentiation and analysis during such procedures.
- FIG. 1 shows an example navigation system to support minimally invasive access port-based surgery.
- FIG. 1 illustrates a perspective view of a minimally invasive port based surgical procedure.
- surgeon 101 conducts a minimally invasive port-based surgery on a patient 102 in an operating room (OR) environment.
- a navigation system 200 comprising an equipment tower, cameras, displays and tracked instruments assist the surgeon 101 during his procedure.
- An operator 103 is also present to operate, control and provide assistance for the navigation system 200 .
- FIG. 2 is block diagram illustrating system components of an example navigation system.
- Navigation system 200 in FIG. 2 includes a monitor 211 for displaying a video image, an equipment tower 201 , a mechanical arm 202 , which supports an optical scope 204 .
- Equipment tower 201 is mounted on a frame (i.e., a rack or cart) and may contain a computer, planning software, navigation software, a power supply and software to manage the automated arm and tracked instruments.
- the example embodiment envisions the equipment tower 201 as a single tower configuration with dual displays ( 211 , 205 ), however, other configurations may also exists (i.e., dual tower, single display, etc.).
- equipment tower 201 may also configured with a UPS (universal power supply) to provide for emergency power, in addition to a regular AC adapter power supply.
- UPS universal power supply
- Example embodiment FIG. 2 also envisions equipment tower 201 having recording module 220 that provides real-time recording of the surgical procedure, capturing audio, video, sensory and multi-modal (i.e., CT, MR, US, etc) inputs from different sources. All relevant data is received at equipment tower 201 and stored in memory by recording module 220 .
- the surgical procedure may be automatically recorded at the outset or be controlled by the operator and/or administrator. In other embodiments, the procedure may be automatically recorded (by default), but there may be an option to override or delete the recording after the procedure has been completed.
- the patient's brain is held in place by a head holder 217 and inserted into the head is an access port 206 and introducer 210 .
- the introducer 210 is tracked using a tracking system 213 , which provides position information for the navigation system 200 .
- Tracking system 213 may be a 3D optical tracking stereo camera similar to one made by Northern Digital Imaging (NDI).
- Location data of the mechanical arm 202 and port 206 may be determined by the tracking system 213 by detection of fiducial markers 212 placed on these tools.
- a secondary display 205 may provide output of the tracking system 213 . The output may be shown in axial, sagittal and coronal views as part of a multi-view display.
- an introducer 210 with an atraumatic tip may be positioned within the access port and employed to position the access portion within the head.
- the introducer 210 may include fiducial markers 212 for tracking, as presented in FIG. 2 .
- the fiducial markers 212 may be reflective spheres in the case of optical tracking system or pick-up coils in the case of electromagnetic tracking system.
- the fiducial markers 212 are detected by the tracking system 213 and their respective positions are inferred by the tracking software.
- the introducer 210 may be removed to allow for access to the tissue through the central opening of the access port. However, once introducer 210 is removed, the access port can no longer be tracked. Accordingly, the access port may be indirectly tracked by additional pointing tools configured for identification by the navigation system 200 .
- a guide clamp 218 for holding the access port 206 may be provided.
- Guide clamp 218 can optionally engage and disengage with access port 206 without needing to remove the access port from the patient.
- the access port can slide up and down within the clamp while in the closed position.
- a locking mechanism may be attached to or integrated with the guide clamp, and can optionally be actuated with one hand, as described further below.
- a small articulated arm 219 may be provided with an attachment point to hold guide clamp 218 .
- Articulated arm 219 may have up to six degrees of freedom to position guide clamp 218 .
- Articulated arm 219 may be attached or attachable to a point based on patient head holder 217 , or another suitable patient support, to ensure when locked in place, guide clamp 218 cannot move relative to the patient's head.
- the interface between guide clamp 218 and articulated arm 219 may be flexible, or optionally locked into place. Flexibility is desired so the access port can be moved into various positions within the brain, but still rotate about a fixed point.
- a linkage that can achieve this function is a slender bar or rod.
- the bar or rod When the access port 206 is moved to various positions, the bar or rod will oppose such a bend, and move the access port 206 back to the centered position.
- an optional collar may be attached to the linkage between the articulated arm, and the access port guide, such that when engaged, the linkage becomes rigid.
- FIG. 3 is a flow chart illustrating the processing steps involved in a port-based surgical procedure using a navigation system.
- the first step involves importing the port-based surgical plan (step 302 ).
- a detailed description of the process to create and select a surgical plan is outlined in PCT Patent Application No. PCT/CA2014050272, titled “PLANNING, NAVIGATION AND SIMULATION SYSTEMS AND METHODS FOR MINIMALLY INVASIVE THERAPY”, which is hereby incorporated by reference in its entirety, and which claims priority to U.S. Provisional Patent Application Ser. Nos. 61/800,155 and 61/924,993, which are both hereby incorporated by reference in their entirety.
- An example plan may compose of pre-operative 3D imaging data (i.e., MRI, ultrasound, etc.) and overlaying on it, received inputs (i.e., sulci entry points, target locations, surgical outcome criteria, additional 3D image data information) and displaying one or more trajectory paths based on the calculated score for a projected surgical path.
- received inputs i.e., sulci entry points, target locations, surgical outcome criteria, additional 3D image data information
- displaying one or more trajectory paths based on the calculated score for a projected surgical path.
- the aforementioned surgical plan may be one example; other surgical plans and/or methods may also be envisioned.
- the patient is affixed into position using a head or body holding mechanism.
- the head position is also confirmed with the patient plan using the navigation software (step 304 ).
- the next step is to initiate registration of the patient (step 306 ).
- registration or “image registration” refers to the process of transforming different sets of data into one coordinate system. Data may be multiple photographs, data from different sensors, times, depths, or viewpoints.
- the process of “registration” is used in the present application for medical imaging in which images from different imaging modalities are co-registered. Registration is necessary in order to be able to compare or integrate the data obtained from these different modalities.
- Non-limiting examples include intensity-based methods which compare intensity patterns in images via correlation metrics, while feature-based methods find correspondence between image features such as points, lines, and contours.
- Image registration algorithms may also be classified according to the transformation models they use to relate the target image space to the reference image space. Another classification can be made between single-modality and multi-modality methods.
- Single-modality methods typically register images in the same modality acquired by the same scanner/sensor type, for example, a series of MR images can be co-registered, while multi-modality registration methods are used to register images acquired by different scanner/sensor types, for example in MRI and PET.
- multi-modality registration methods are used in medical imaging of the head/brain as images of a subject are frequently obtained from different scanners. Examples include registration of brain CT/MRI images or PET/CT images for tumor localization, registration of contrast-enhanced CT images against non-contrast-enhanced CT images, and registration of ultrasound and CT.
- the patient is draped (step 310 ).
- draping involves covering the patient and surrounding areas with a sterile barrier to create and maintain a sterile field during the surgical procedure.
- the purpose of draping is to eliminate the passage of microorganisms (i.e., bacteria) between non-sterile and sterile areas.
- step 310 Upon completion of draping (step 310 ), the next steps is to confirm patient engagement points (step 312 ) and then prep and plan craniotomy (step 314 ).
- the next step is to cut craniotomy (step 314 ) where a bone flap is temporarily removed from the skull to access the brain (step 316 ).
- Registration data is updated with the navigation system at this point (step 322 ).
- the next step is to confirm the engagement within craniotomy and the motion range (step 318 ). Once this data is confirmed, the procedure advances to the next step of cutting the dura at the engagement points and identifying the sulcus (step 320 ). Registration data is also updated with the navigation system at this point (step 322 ).
- this registration update can be manipulated to ensure the best match for that region, while ignoring any non-uniform tissue deformation affecting areas outside of the surgical field (of interest). Additionally, by matching overlay representations of tissue with an actual view of the tissue of interest, the particular tissue representation can be matched to the video image, and thus tending to ensure registration of the tissue of interest.
- video of post craniotomy brain i.e. brain exposed
- the video position of exposed vessels can be matched with image segmentation of vessels
- the video position of a lesion or tumor can be matched with image segmentation of tumor
- a video image from endoscopy within a nasal cavity can be matched with bone rendering of bone surface on nasal cavity for endonasal alignment.
- multiple cameras can be used and overlaid with tracked instrument(s) views, and thus allowing multiple views of the data and overlays to be presented at the same time, which can tend to provide even greater confidence in a registration, or correction in more than dimensions/views.
- Cannulation involves inserting a port into the brain, typically along a sulci path as identified in step 320 , along a trajectory plan.
- Cannulation is an iterative process that involves repeating the steps of aligning the port on engagement and setting the planned trajectory (step 332 ) and then cannulating to the target depth (step 334 ) until the complete trajectory plan is executed (step 324 ).
- resection step 326
- Resection is a continual loop including both fine and gross resection (step 336 ).
- the next step involves hyperspectral imaging (step 338 ) which may be performed on either fine or gross resection (step 336 ).
- Hyperspectral imaging step 338 is used as a form of tissue differentiation and may assist surgeons to investigate cancerous stem cells.
- the ability to hyperspectrally image tissue being operated on either as part of an external video scope or as a separate module may provide the ability to perform chemical imaging using the absorption of tissue, the ability to differentiate tissues based on scattering properties, and/or the ability to improve visualization by imaging at wavelengths with reduced absorption or scattering properties.
- step 326 the surgeon then decannulates (step 328 ) by removing the port and any tracking instruments from the brain. Finally, the surgeon closes the dura and completes the craniotomy (step 330 ).
- FIG. 4 illustrates an example port-based brain surgery procedure using a video scope.
- operator 404 typically a surgeon, would align video scope 402 to peer down port 406 .
- Video scope 402 may be attached to an adjustable mechanical arm 410 .
- Port 406 may have a tracking tool 408 attached to it where tracking tool 408 is tracked by a tracking system of a navigation system.
- the video scope 402 is commonly an endoscope or a microscope, these devices introduce optical and ergonomic limitations when the surgical procedure is conducted over a confined space and conducted over a prolonged period such as the case with minimally invasive brain surgery.
- FIG. 5A illustrates the design of a video scope that is composed of a lens assembly 511 (explained later) and two illumination delivery points 512 .
- the lens assembly 511 is terminated at the eyepiece end with a sealed window 501 at the proximal end. Sealed window 501 is typically made of quartz, to help maintain water seal since OR devices must be steam cleanable.
- the eyepiece end also has a camera coupler 505 that provides a standardized mounting point for a camera (not shown).
- the camera may be a standard definition (SD), high definition (HD) or ultra high definition (UHD) camera.
- the camera may be replaced by other imaging technologies such as Optical Coherence Tomography (OCT) or Polarization Sensitive-OCT.
- OCT Optical Coherence Tomography
- Polarization Sensitive-OCT Polarization Sensitive-OCT.
- the distal end of the lens assembly is also sealed with a clear window 513 at the distal end.
- the distal end also supports illumination optics 512 .
- the distal end may be also optionally affixed with a polarizing filter to enable polarization sensitive imaging.
- the PS OCT technique described herein may be used to specifically visualize tissue exhibiting structural organization.
- tissue structures include tendons that are attached to bones.
- Other examples of tissue that exhibit structural organization include ligaments, muscle, cartilage, tissue connective membrane, nerves, retina, blood vessel walls, some bone structures, trachea, esophagus, tongue and teeth.
- PS OCT commonly generates a heat map or pseudo colored image (reference: “Correlation of collagen organization with polarization sensitive imaging of in vitro cartilage: implications for osteoarthritis,” W. Drexler et. al, The Journal of Rheumatology, Vol. 28, No. 6, 1311-1318) where tissue structures with high degree of organization appear highlighted.
- the system can be used in orthopedic surgery to visualize tendons and optionally avoid unintentional damage to this tissue during a procedure.
- These identified regions of tissue exhibiting high level of structural organization e.g. tendons and ligaments that are often located near skeletal structure
- entheses The insertion sites, tendon-bone junctions and ligament-bone junctions, are known as entheses.
- the anatomical locations of entheses are well known and landmarks can be identified on the bone in the vicinity of these attachment points (reference: “Anatomy and biochemistry of enthuses,” Michael Benjamin, Ann Rheum Dis 2000, Vol. 59, Issue 12, pg: 995-999).
- this a priori anatomical information about the position of the tendon or ligament relative to bone structures in the vicinity can be used to register intraopertive PS-OCT image of the tendons or ligaments with pre-operative images obtained using other modalities that accurately image the bone structures.
- the tendon-bone junction in the Achilles tendon enthesis is immediately proximal to the superior tuberosity.
- This region is characterized by a highly irregular interface at the attachment points or junction.
- This characteristic structure of the bone can be used to identify the junction where the tendon attaches to the bone.
- image registration or image fusion The geometric correlation of images that are thus obtained using different modalities, and often at different scales, is known as image registration or image fusion.
- Multi-modal image registration for pre-operative planning and image guided neurosurgical procedures
- Risholm, et. al Neurosurg Clin N Am, 2011, April; 22(2): 197-206
- Image registration of ex-vivo MRI to sparsely sectioned histology of hippocampal and neocortical temporal lobe speciments Goubran et. al, Neurolmage, 83 (2013); 770-781.
- Broad classes of image registration methods for medical images is also described in detail in “A survey of medical image registration,” Maintz et. al, Medical Image Analysis (1998), Vol. 2, No. 1, pp: 1-36.
- the illumination optics is comprised of fiber bundles 507 that are rotatably attached using a pair of connectors 510 .
- the connectors 510 allow the fiber bundles to rotate freely ( 570 in FIG. 5C ) within the connector while maintaining a fixed distance between the lens 509 and tip of the fiber bundle 507 using a loose sleeve 508 . This rotation movement will reduce the strain on the fiber bundle when the video scope is moved on a holding system (not shown) or a mechanical arm 410 as seen in FIG. 4 .
- the rotatable connector 510 also aid in easy cable management when the mechanical arm 410 is moved during a surgical procedure.
- the illumination optics are placed as close as possible to the objective lens.
- One non-limiting example of spacing between the optics is approximately 30 to 35 mm, or 32 to 34 mm, between the center of the lenses 509 where the diameter of lenses 509 is approximately 15 mm.
- This configuration is optimal for illuminating the bottom of a surgical port with maximum intensity when the distal end of the video scope is between 25 cm to 40 cm from the bottom of the surgical port.
- An optical compensator 503 is used to act as a thermal compensator to control the stress on optical components during steam cleaning.
- a holder 506 provides an easy to grasp assembly to hold and manipulate the video scope without introducing mechanical stress on the lens assembly.
- the lens assembly is encased in a sealed barrel 511 to avoid ingression of steam and liquids during normal use and cleaning.
- the rotatable attachment mechanism 510 allows free rotation of the fiber bundles when the camera is moved manually or when mounted to a robotic positioning system. This, in turn, avoids undue stress on the fiber bundles that are susceptible to fracture.
- FIG. 5C illustrates a non-limiting example to realize a functionality that allows the illumination assembly 565 to rotate radially 560 around the video scope barrel 502 .
- the illumination assembly 565 is composed of the two fiber bundles 507 on either side of the video scope, mounting mechanism 508 and lens 509 (as in FIG. 5A ). This allows the surgeon to adjust the radial orientation of the illumination and orient the illumination assembly so that it minimally obstructs the surgeon's view of the surgical space.
- the illumination assembly can be freely rotated without rotating the video scope by securing the video scope to an external positioning mechanism, such as 410 , using a removable clamp 555 and an associated lock 550 .
- the removable clamp's distal end 555 and the associated lock 550 may be mated together using a thread mechanism or any other mechanical fastening mechanism.
- the removal clamp's proximal end (not shown) may be secured to the external positioning mechanism 410 .
- the rotation 560 enabled in this design along with the rotation 570 of the fiber bundles within the connectors enable positioning and orientation of the video scope with minimal interruption of the visible surgical space and minimize strain on the fiber bundles during motion.
- the illumination assembly 565 may be replaced with alternative configurations such as ring lights or single illumination points. Ring lights may be realized through circular arrangement of fiber strands (not shown) from an optical fiber bundle around the circumference of the objective lens. Single illumination points may be realized through removal of one of the two split fiber bundles 507 from the design.
- the illumination assembly preferably receives the light input from an optical source that is located away from the video scope. This reduces the total weight of the external scope and allows for easy manipulation of the video scope by a manual positioning system (not shown) or a mechanical arm 410 .
- the light from the light source is delivered to the video scope through the use of a fiber bundle. Presence of two delivery points represented by illumination optics 512 in FIG. 5A requires the use of a fiber bundle that is split in two. This design of fiber bundle is also known as a Y-cables. An example embodiment of this Y-cable design is illustrated in FIG. 5B . In FIG.
- rotatable connections 508 are provided on the fasteners 510 at the two distal end of the Y cable, providing a mechanism for freely rotating the fiber bundles to avoid fracture of the bundles.
- a strain-relief 527 helps maintain a minimum limit on the bend radius 529 of the bundle between the two distal ends and the Y-junction 531 .
- Y-junction 531 helps reduce bend strain on the fiber bundle 507 .
- Strain-relief 533 similarly aids in reducing bend strain near the connector 535 at the proximal end of the Y-cable.
- Cross sections 525 and 537 illustrate fiber bundles at the two ends of the Y-cable.
- the length of the cable may be at least 40 cm with the Y-junction 531 placed equidistant from the two ends. This dimension provides for placement of light source on a cart or instrumentation tower 201 sufficiently away from the mechanical arm 410 while minimizing light loss due to excessive length of the fiber bundle.
- FIG. 6 illustrates an optical design of the video scope that limits the diameter of the objective lens 600 (front lens).
- This design enables the mounting of illumination optics immediately adjacent to the objective lens so that the illumination beam can be almost collinear to the return path of the light reflected from the tissue.
- the illumination beam and the reflected beam need to be as collinear as possible so that maximum illumination is delivered at the bottom of the access port 406 .
- the optical design is constrained so that the length of the lens assembly is minimized to make the whole video scope 402 minimally intrusive to the surgeon's field of view and facilitate easy access to the surgical space by the surgeon. This constraint is a challenge in conventional optical design conventional optical design techniques maximize zoom by utilizing maximum available physical length of the lens assembly during the design process.
- This optical design of the present disclosure is adapted from a conventional endoscopic system that consists of objective lens 600 , relay lens 602 and eyepiece 604 .
- the zoom parameter of the optical assembly is chosen such that the minimum field of view (corresponding to maximum zoom) is equal to approximately 13 mm. This dimension is the diameter of the surgical port.
- the field of view of 13 mm needs to be achieved at a minimum working distance of 25 cm where the minimum working distance is defined as the distance between the distal end of the video scope ( 402 in FIG. 4 ) and bottom of the surgical port ( 406 in FIG. 4 ).
- a coupler 505 is used to attach a camera at the eyepiece end (marked ‘E’ in FIG. 6 ).
- the optical design of the objective is composed of 1 doublet and 1 singlet; the relay is composed of 1 doublet and 1 singlet and the eyepiece is composed of 2 singlet and 1 doublet. Any manufacturing error is compensated using one optical compensator 503 that is placed between the objective and relay. The length of the optical sub-assembly is minimized through the use of higher power lenses and fewer lens groups.
- the type of surgical procedure determines either a wide-field of view (WFOV) or a narrow field of view (NFOV) video scope.
- WFOV wide-field of view
- NFOV narrow field of view
- a neck surgery may benefit from a WFOV video scope where large area is captured by the video scope; whereas, a port-based brain surgery may benefit from a NFOV video scope.
- two separate designs may be developed such that they share several sub-components and the manufacturing process. Hence, it is economical to manufacture two different designs while sharing number of design elements and assembly procedure.
- Both WFOV and NFOV designs share a similar optical illumination system 512 as seen in FIG. 5A
- the WFOV design can be realized by attaching a camera to the camera coupler 505 . The zoom adjustment of the camera is used to determine the field of view in this case.
- FIG. 7 illustrates an assembly with a non-coaxial illumination source.
- the illumination system 710 is similar in design to that illustrated in FIG. 5A and consists of fiber bundles 704 (only a distal portion of which are shown in the Figure).
- An air-filed opaque tube (also known as optical tube) 702 is used to position the illumination mechanism away from the camera attached to the coupler 505 .
- any required magnification may be provided by the camera lens (not shown but typically attached to the camera coupler) for WFOV application.
- a finite space that is at least 1 mm between the plane 706 of the distal end of the optical tube and the plane of the illumination optics 708 helps isolate the illumination light from directly reaching the camera input.
- the dimensions of the WFOV optics will be such that the illumination will not be nearly coaxial with the path of the reflected light. This is not a limitation in this configuration because WFOV is used to observe a surgical space that is larger that of a port (which is approximately 13 mm). Hence, general illumination is sufficient. Placement of the illumination source close to the camera does improve illumination of the surgical area compared to the use of overhead surgical lights and avoids glare from area outside of the surgical space.
- additional components, 712 and 714 are explained below in the context of hyperspectral imaging.
- the illumination sources placed immediately adjacent to the distal end of the video scope may be employ a light source such as luminance light emitting diodes or Super Luminescent Diodes (SLD's) (not shown). Since the light sources are not coaxial to the reflected light path (the light path incident on the lens and camera assembly), the light sources have to be aimed or steered at the focal plane of interest. Such steering may be achieved using movable fiber bundle mounts 510 as shown in FIG. 5A .
- a light source such as luminance light emitting diodes or Super Luminescent Diodes (SLD's)
- the walls of the port are either partially or fully reflective. This introduces localized regions in the imaged surface that have higher intensity of incident light. Such regions are commonly known as hot-spots. It is desirable to avoid such high intensity regions as these tend to saturate sensors and, hence, limit the dynamic range of the sensors in the camera mechanism.
- Use of post-processing to normalize intensities is less optimal as saturation of sensors results in information loss that cannot be recovered. Presence of high intensity regions can be reduced through the use of surface textures on the port walls that diffuse the light. The impact of using smooth and rough surface texture on the port walls is illustrated in FIGS. 8A and 8B , respectively.
- the reflections resulting from textured walls is referred to as Lambertian reflection.
- the assessment presented in FIGS. 8A and 8B were conducted using ray-tracing tools and the resulting intensity of light at the surface of the tissue (distal end of the port) were visualized using heat-maps or pseudo color where high intensity corresponded to white and low intensity corresponded to black.
- Another approach to uniformly illuminating at the bottom of the port is to model the light rays using a commonly known optical modelling method, such as ray tracing, and establish the optimal orientation of the light sources that minimize hot-spots at the bottom of the surgical port.
- Orientation of the light sources may be modified using a beam steering mechanism, such as the one illustrated in FIG. 5A .
- a robotic positioning system may be used to achieve this steering.
- Port-based imaging is also limited by highly reflective nature of some but not all regions of the brain tissue due to the presence of blood, CSF or other fluids. In the latter case, an initial image could be acquired to identify regions with high intensity reflected light and this information can be used to reposition direction of the light sources in an attempt uniformly distribute the incident light.
- imaging using white light has several challenges in the operating room. Several of these challenges can be overcome by limiting the spectral range of the light that is observed or by judiciously combining selected wavelength bands to visualize human tissue in the operating room.
- FIG. 9 illustrates a video scope that has been adapted to accommodate hyperspectral imaging capabilities.
- tunable light source that is adapted based on the surgical context e.g. selection of illumination spectral region where blood is highly absorptive (to detect blood clots) or transmissive (to avoid excessive light scattering) may be used.
- FIG. 9 illustrates one such system.
- the tunable light source is mainly composed of a broadband light source 1100 , a spectral separation mechanism 1140 , a spectral filtering mechanism 1150 and a mechanism to combine the filtered frequency bands 1170 .
- the combining mechanism consists of a lens and a fiber bundle that mixes all the reflected wavelength bands into one beam that is transmitted through the fiber bundle 507 .
- the light from light source 1100 is passed through a slit 1110 to generate a narrow beam. This light is then collimated using optical elements 1120 and 1130 .
- the collimated beam is then split into its spectral components using a prism (not shown), reflective or transmission grating.
- FIG. 9 illustrates the use of a reflective grating 1140 .
- the spatially separated beam is filtered by selectively reflecting portions of the spatially separated beam. This is achieved using a spatial light modulator, SLM 1150 , such as a Digital Light Processor (Texas Instruments Inc).
- SLM is composed of an array of micro-mirrors that can be electronically activated to act as mirrors or deactivated to acts as opaque surfaces. Hence, specific portions of the spectrum are reflected while other regions are suppressed based on the pattern of activated micro-mirrors.
- the beam that is now composed of selective portions of spectrum are combined using focusing optics 1160 and a combiner 1170 .
- the recombined beam is now composed of only those wavelengths that were selectively reflected by the spatial light modulator, SLM 1150 .
- This light can be used as the illumination source of an imaging system or external scope by transmitting the light via a light pipe 507 to the illuminator connector and lens mechanism 510 attached to the external scope.
- the video scope illustrated in FIG. 9 shows the connection of light pipe 507 to only one of the two illuminator connectors 510 for the sake of simplicity of the illustration. Details of connecting the light pipe to the video scope is further explained in FIG. 5A .
- the reflected light from the tissue 1198 is captured by the external scope that is composed of lens assembly 502 .
- the lens assembly is composed; this light is captured using a high resolution detector 1125 that is usually a charge coupled device, CCD.
- the specific band of wavelengths that are reflected by the SLM are controlled by an SLM controller 1180 that is under the command of a computer 1185 .
- the same computer is used to acquire the image from the detector 1125 .
- the computer can synchronize the illumination of a material 1198 with a specific wavelength band or wavelength bands of light and acquire corresponding reflected light.
- This association of illumination wavelength and acquired image can be used to construct a hyper-spectral image where each image is a 2D or 1D image and the third dimension is an index that corresponds to illumination wavelength band(s). Since the individual micro-mirrors located in an SLM can be operated at a rate as high as 4 kHz, subsequent frames of the field of view can be obtained at different wavelength bands.
- some of the acquired frames can be for employed white-light illumination of the tissue. This is possible by operating the acquisition camera at a frame rate that is sufficiently high to provide smooth video playback, as perceived by a human observer when white light frames are intermittently obtained while collecting hyperspectral image frames.
- the frame rate may be selected to be higher than 20 frames per second, higher than 24 frames per second, or higher than 30 frames per second, in order to support white light video acquisition at such frame rates while obtaining hyperspectral data.
- a white-light image can be acquired every 1/20 th of a second and any additional frame can be allocated for acquisition using specific wavelength bands.
- a white light video feed may then be separately generated and displayed based on the collected white light images. This allows the surgeon to continuous view a white-light image of the surgical area while acquiring any additional images at different wavelength bands in a multiplexed manner.
- the white-light image stream (or video) may be viewed in one display or sub-section of a display and other images acquired using other wavelength bands may be viewed in a second display or second sub-section of the same display.
- the individual wavelength bands can be composed of non-overlapping individual wavelength bands or combination of bands that may overlap.
- at least one of the acquired frame can correspond to illumination 1197 of the subject material 1198 using the entire wavelength band of the light source.
- the entire wavelength band could be also normalized to ensure that all the intensity in the output light emanating from the combiner 1170 is consistent across the entire spectrum. This is known as white balancing.
- the same optical mechanism can be used to acquire hyperspectral images and white-light images that are interspersed among each other in the acquired sequence of images. This embodiment eliminates the need for splitting the acquired beam into separate paths so that one beam is captured by a hyperspectral imaging system while the other beam is captured by a white-light camera.
- the sample being imaged 1198 may be an ex-vivo tissue sample or portion of the brain tissue that may be exposed through a port-based neurosurgical access inserted in the skull.
- the software system used to acquire hyperspectral data and white-light images (or video) in a multiplex fashion is illustrated in FIG. 10 .
- First the range of wavelengths (wave bands) that are of interest are stored in a table (step 1200 ).
- specific wave band for illumination is selected from the table (step 1220 ).
- Each entry in this table is used to look up (step 1230 ) specific micro-mirrors that need to be activated using another table (step 1210 ).
- step 1240 only the micro-mirrors associated with specific wavelength bands are activated (step 1240 ).
- Activation of a micro-mirror turns it into a micro-reflector instead of an opaque surface.
- the sample 1198 in FIG. 9 is illuminated with light ( 1197 in FIG.
- the table (step 1200 ) may also include entries that activate the entire spatial light modulator (SLM).
- SLM acts as a mirror for the entire bandwidth of the light source and the acquired image will correspond to white-light illumination.
- the reflected light from the illuminated sample is acquired (step 1250 ) by the same computer and associated with the specific wavelength band (step 1260 ).
- the type of illumination (white-light versus specific wavelength band) used for each acquired image is interrogated (step 1270 ) in order to appropriately classify the acquired image as part of white-light image (video) or part of the hyperspectral image data set. If the acquired image corresponds to a narrow wavelength band then it is stored as part of the hyperspectral image set (step 1280 ). If the image corresponds to white-light illumination, it is stored as white-light image or a stream of such images may be captured to represent a video stream.
- the resulting image set will be composed of both hyperspectral image sets (step 1280 ) and white-light image sets (step 1290 ), all acquired using the same hardware.
- the video stream needs to be at least 30 frames per second to provide a flicker-free video to the surgeon. If a total of 40 frames are acquired per second, the additional 10 frames may be used to store images corresponding to 10 distinct or overlapping wavelength bands. Hence, if the total frame rate of the acquisition system is n frames per second, n ⁇ 30 frames may be allocated towards n ⁇ 30 wavelength bands in the hyperspectral image data set.
- tunable light source 1110 shown in FIG. 9 may be monochromatic, spanning ultra violet (UV), visible, and/or near infrared (NIR) wavelengths, continuous wave or pulsed that is used to illuminate the tissue using free space or fiber coupled mechanism
- UV ultra violet
- NIR near infrared
- specific wavelength bands may be acquired by filtering the reflected light from a broadband light source using such spectral elements as discrete wavelength filters (on filter wheel or spatial on-chip filters), liquid crystal filters, spectrographs/spectrometers/spectral gratings, spatially varying gratings, fiber-coupled spectrometers.
- spectral elements as discrete wavelength filters (on filter wheel or spatial on-chip filters), liquid crystal filters, spectrographs/spectrometers/spectral gratings, spatially varying gratings, fiber-coupled spectrometers.
- FIG. 7 also illustrates the implementation of discrete filters 712 attached to a rotatable filter wheel 714 ) that may be motorized.
- This filter mechanism is attached at the distal end of the video scope.
- Another alternative to discrete filters at the input to the video scope is a liquid crystal-based tunable wavelength filter (not shown) to pass only a narrow range of wavelengths.
- This filter can be tuned to a number of different wavelengths and operates in a similar manner to the discrete filters as an image is acquired for each wavelength the filter is tuned to.
- diffraction grating based systems that separate input light input its constituent wavelengths may be used in lieu of the camera 1125 shown in FIG. 9 .
- Imaging spectrometer systems rely on scanning the entrance slit of the system across the field to be imaged. Thus the acquisition time is limited by the scanning time.
- the entrance slit of the spectrometer can be either free space or fiber coupled to the optical path. If an array-to-line fiber mapping is utilized it is possible to acquire all spatial and spectral information simultaneously.
- the spectrometer could be alternatively equipped with Spatially Varying Gratings where a specialized diffraction grating that allows for the collection of spectra from all pixels in a single acquisition. The grating is divided into a number of spatial gratings each with a varying direction of diffraction. An image is acquired that captures the diffracted light from each of these grating regions, this image is then reconstructed to form the hyperspectral data set.
- Non-limiting examples of camera 1125 include monochrome video camera with resolution up to high definition (HD) or ultra high definition (UHD). CCD, CMOS, InGaAs, or HgCdTe device.
- HD high definition
- UHD ultra high definition
- confocal hyperspectral imaging system Another aspect of confocal hyperspectral imaging system is that the entire tissue surface does not have to be scanned in a raster pattern. Instead, random spots can be accumulated until a reasonable match is found against pre-defined data classes. This can significantly reduce the data acquisition time associated with hyperspectral imaging.
- the hyperspectral imaging system illuminates the tissue with monochromatic or broadband light, collects light reflected from the tissue, controls the wavelength of the detected light in such a way that a series of images, each recorded at different wavelengths or wavelength ranges, is collected.
- This series of images known as a hyperspectral dataset, is processed to extract tissue's bio-chemical or microstructural metrics and reduced to 2D (spatial).
- This reduced 2D image may be spatially registered and can be overlaid on the external video scope image as well as any other pre- and intra-operative images.
- methods of correlating image data are disclosed in PCT Patent Application No.
- Tissue specific information can range from tissue type identification to inferring pathology associated with a region of the acquired image. Examples of the possible processing methods including the following:
- the spectra if the spectral peaks or features of chemical(s) of interest are known, the spectra and be processed, through either peak or feature detection algorithms, to detected the peaks or features to give an indication of the chemical presence and some indication of the concentration or quality. This useful only if the specific chemicals of interest are known.
- the spectra of specific tissues or tissue states of interest can be acquired and stored in a database, as disclosed in PCT Patent Application No. PCT/CA2014/050269, titled “INTRAMODAL SYNCHRONIZATION OF SURGICAL DATA” and filed on Mar. 14, 2014. Spectra then acquired during the surgery can be compared to the spectra stored in the database for similarity and if sufficiently similar to give an indication of what tissue or tissue type the spectra was acquired from.
- Multivariate/chemometric methods which are a wide grouping of statistical techniques where a method is trained on spectra collected from samples with known states (i.e., spectrum and corresponding chemical level, tissue type, tissue state, etc.), may be used to predict the state of a new sample based on the acquired spectrum.
- states i.e., spectrum and corresponding chemical level, tissue type, tissue state, etc.
- Some of the more commonly used employed techniques include principal component regression (PCR), partial least squares (PLS), and neural networks (NN).
- the aforementioned analysis methods can be implemented in a computer system, and hence the results of the analysis can be obtained in near-real time for appropriate use by a surgeon. This may significantly reduce the need for similar analysis by a pathologist and reduces the wait time associated with obtaining results of such tissue analysis.
- Correlation metrics between newly acquired data and representative data in a knowledge-base (or database or training set) provide the surgeons a means of quantifying tissue types. Such metrics may be a representation of confidence associated with automated inference provided by the software algorithm.
- the ability to selectively view narrow bands of the spectra or reject narrow bands of the spectra may allow the surgeon to reject bright reflections from blood.
- the surgeon may be able to view the interior of the corridor and proceed with surgical resection of tumor even when the corridor is occluded by excessive bleeding. This will reduce the need to constantly irrigate the narrow corridor and hence reduce interruption of the surgical procedure.
- the optical characteristics and chemical composition of blood may be taken advantage of to visualize vasculature located at or immediately below tissue surface.
- a common challenge associated with the dural opening step in cranial surgery is the inability to anticipate the presence of vasculature immediately below the dura.
- Hyperspectral imaging (HSI) with emphasis on red and near infrared portions of the imaging spectrum, can be used to selectively visualize regions with high haemoglobin content. Due to the blood-brain barrier that naturally exists in the human brain, this technique will result in selective imaging of vasculature.
- OCT based imaging may be also used to visualize structures located immediately below the dura since the dura is not more than a millimetre thick. Vasculature is one of the predominant structures located below the dura. Others structures of relevance that can be similarly imaged include the sulcal folds.
- the vasculature that is visualized using the techniques described in the preceding paragraphs is a map of the structure of the blood vessels in the vicinity of the region being imaged. Such regions may include the region in the vicinity of the trajectory along which a port is inserted in port-based cranial surgery.
- This vascular structure in the surgical region will be referred to as the in-situ vascular structure or intra-operative local vasculature.
- This in-situ geometry of the vascular structure may be then be compared with whole-head vascular structure images derived from MRI acquired after the injection of Gadolinium or other contrast medium in the veins. The latter whole-head images are obtained prior to a surgical procedure and referred to as pre-operative vascular structures.
- the pre-operatively vascular structures are imaged relative to the anatomy of the patient. In other words, the location and orientation of this vascular structure is known relative to the anatomy of the patient.
- in-situ (intraoperative) vascular structure may be used to infer the exact location of the in-situ vascular structures relative to the pre-operative vascular structures. This allows the surgeon to confirm or infer their current surgical position relative to the anatomy of the patient.
- the vascular structure can be used as a land-mark or reference frame for navigated surgical procedures.
- Geometric correlation of the local vascular structure with the global vascular structure is feasible because the anatomy and pattern of blood vessels is known (reference: “Cortical blood vessels of the human brain,” Duvernoy et. al., Brain Research Bulletin, Vol. 7, Issue 5, November 1981, pg 519-579 and “A computed tomographic guide to the identification of cerebral vascular territories,” Damasio et. al., Arch Neurol, 1983; 40(3): 138-142.
- registration in general, involves the geometric correlation of small scale image (or 3D structure) with a large scale image (or 3D structure); however, this geometric correlation involves the use of features that are common between the two image sets.
- vasculature by itself as a feature is as disclosed herein is the first instance of this.
- the vasculature can be used as a feature because of its unique structure and this uniqueness is well known (reference: see two papers cited in this paragraph).
- Unique vascular structure is known to exist in several regions of the body, for example the cerebral region, retina and the cardiac regions.
- a local vascular structure in the selected local region of the tissue is identified, and then the global vascular image data is searched for identifying and locating a portion of the global vascular structure geometrically matching the local vascular structure, and upon identifying and locating matching vascular structure, geometrically correlating and registering the local vascular structure in the selected local region of the tissue with the a global vascular structure within the tissue for confirming location of the local vasculature structures.
- the location thus inferred, is invaluable for guiding tools in the local region for surgical procedures. Such surgical procedure is known as navigated surgery.
- embodiments provided herein may employ software to process the 3D dimensional data sets to extract the information of interest, and to reduce the data to a 2D image that can be visualized in conjunction with or overlaid on the surgical image acquired by the external video scope.
- These software methods could include everything from simple spectral peak detection to more sophisticated multivariate, chemometric, and data mining techniques to extract the metric of interest from the acquire spectra.
- the spectrum associated with each pixel may be processed according to such methods.
- hyperspectral imaging is an optical technique and limited penetration (2-3 mm), its use is restricted to superficial tissues or those exposed through corridor surgery.
- the unique spectra of chemicals in tissue provide the potential to use hyperspectral imaging to image chemical content and from this provide useful qualitative or quantitative information to the surgeon to assist in decision making during the surgery.
- Chemical imaging can be used to differentiate between different tissues based on differing chemical composition and associated differing absorption (e.g., white vs grey matter), determine tissue state (e.g., normal vs malignant), and determine tissue status and/or health (e.g., state of oxygenation).
- the difference in spectral scattering properties can, similar to absorption changes, be used to determine the properties of tissue based on changes in cellular structure with tissue type (e.g., fat vs nerve fiber) and state (e.g., changes in nuclear and overall cell size with pre and cancerous states).
- tissue type e.g., fat vs nerve fiber
- state e.g., changes in nuclear and overall cell size with pre and cancerous states.
- images at only selected wavelengths or wavelength ranges to improve the visualization of tissue minima or maxima in absorption or scattering. For example, images at wavelengths where hemoglobin absorption is at a minimum, the absorption due to blood will be significantly reduced thus providing additional light for illumination.
- FIG. 11 illustrates a standard color image (A) of a brain region (Corpus Callosum) that is also captured using four different wavelength bands centered at 400 nm (B), 500 nm (C), 600 nm (D) and 700 nm (E) and a bandwidth of 10 nm each. It is evident that that 400 nm filter band clearly illustrates tissue structures that are otherwise invisible in other wavelength bands.
- FIG. 12 illustrates the key components of the computer system 1185 of FIG. 9 .
- FIG. 12 provides an example, non-limiting implementation of computer control system 425 , which includes one or more processors 430 (for example, a CPU/microprocessor), bus 402 , memory 435 , which may include random access memory (RAM) and/or read only memory (ROM), one or more internal storage devices 440 (e.g. a hard disk drive, compact disk drive or internal flash memory), a power supply 445 , one more communications interfaces 450 , and various input/output devices and/or interfaces 460 such as a user interface for a clinician to provide various inputs, run simulations etc.
- processors 430 for example, a CPU/microprocessor
- bus 402 bus 402
- memory 435 which may include random access memory (RAM) and/or read only memory (ROM), one or more internal storage devices 440 (e.g. a hard disk drive, compact disk drive or internal flash memory), a power supply 445 , one more communications
- bus 402 is depicted as a single connection between all of the components, it will be appreciated that the bus 402 may represent one or more circuits, devices or communication channels which link two or more of the components.
- bus 402 often includes or is a motherboard.
- computer control system 425 may be, or include, a general purpose computer or any other hardware equivalents configured for operation in space.
- Computer control system 425 may also be implemented as one or more physical devices that are coupled to processor 430 through one of more communications channels or interfaces.
- computer control system 425 can be implemented using application specific integrated circuits (ASIC).
- ASIC application specific integrated circuits
- computer control system 425 can be implemented as a combination of hardware and software, where the software is loaded into the processor from the memory or over a network connection.
- a vertical slit or a focal point may be imaged by the video scope using a confocal optical design that is commonly used in a microscope (not shown).
- the spot or slit may be then imaged on a photomultiplier to generate a very sensitive hyper-spectral imaging system.
- the focal point may be swept across the sample surface using a scanning mechanism.
- a commonly used scanning mechanism is a galvanometer mirror system.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Surgery (AREA)
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Medical Informatics (AREA)
- Veterinary Medicine (AREA)
- Public Health (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Molecular Biology (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Pathology (AREA)
- Biophysics (AREA)
- Radiology & Medical Imaging (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Optics & Photonics (AREA)
- Neurology (AREA)
- General Physics & Mathematics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- High Energy & Nuclear Physics (AREA)
- Physiology (AREA)
- Robotics (AREA)
- Orthopedic Medicine & Surgery (AREA)
- Dentistry (AREA)
- Gynecology & Obstetrics (AREA)
- Psychology (AREA)
- Neurosurgery (AREA)
- Vascular Medicine (AREA)
- Cardiology (AREA)
- Pulmonology (AREA)
- Theoretical Computer Science (AREA)
- Endoscopes (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Psychiatry (AREA)
- Signal Processing (AREA)
Abstract
Description
- This application claims priority to International Patent Application No. PCT/CA2014/050268, titled “SURGICAL IMAGING SYSTEM” and filed on Mar. 14, 2014, the entire contents of which is incorporated herein by reference for the purposes of any U.S. national phase application or continuation-by-pass application filed from, and claiming priority, from this International PCT patent application.
- The present disclosure relates to imaging methods for use in minimally invasive therapy and image guided medical procedures using optical imaging, and more particularly, hyperspectral imaging.
- The optical absorption and scattering properties of biological tissue depend on both the chemical and structural properties of the tissue and the wavelength of the interacting light. How these absorption and scattering properties of tissue change as a function of light can be particularly useful, as it is often unique to chemicals or structures in the tissue (the spectrum of the tissue). For example the absorption features of oxy- and deoxy-hemoglobin can be used to measure the oxygenation of blood and tissue, and the scatter changes caused by difference cellular sizes can be used to detect precancerous and cancerous tissue. The field of measuring these changes in optical properties, as a function of light, is known as spectroscopy and the device to measure the light at the various wavelengths is known as a spectrometer. Spectroscopy has found a wealth of current and potential applications in medicine.
- Traditional spectrometers measure the spectrum of light from a single point of a sample. However, the spectrum from multiple spatial points can be combined to form a 3D spatial dataset (sometimes referred to as a hypercube), where the first two dimensions are spatial and the third is wavelength. In other words, each image pixel has an entire spectrum rather than just an intensity or RBG value. This is known as hyperspectral imaging and is a powerful technique as spatially resolved tissue chemical or microstructural properties can imaged, thus providing a more complete understanding of the tissue and may be a useful technique for tissue differentiation. According to a paper by Dicker et al [Differentiation of Normal Skin and Melanoma using High Resolution Hyperspectral Imaging], hyperspectral image analysis (or hyperspectral imaging) was applied to search for spectral differences between benign and malignant dermal tissue in routine hematoxylin eosin stained specimens (i.e., normal and abnormal skin, benign nevi and melanomas). The results revealed that all skin conditions in the initial data sets could be objectively differentiated providing that staining and section thickness was controlled.
- Systems, methods and devices are provided for intraoperatively confirming location of tissue structures during medical procedures.
- In an embodiment, preoperative image data of a patient's skeletal structure in a vicinity of an anatomical part undergoing a medical procedure is acquired. During the procedure, after exposing tissue intraoperative image data is acquired by scanning a selected region of tissue, in a vicinity of the skeletal structure using Polarization Sensitive-Optical Coherence Tomography (PS-OCT). Regions of tissue exhibiting structural organization in the vicinity of the skeletal structure are identified from the intraoperative (PS-OCT) image data. Geometrically correlating and registering the intraoperative (PS-OCT) image data with the preoperative image data of the skeletal structure in the vicinity of the anatomical part is then performed using a priori known anatomical information about the regions of tissue exhibiting structural information.
- In another embodiment, global preoperative image data of tissue in an anatomical part is acquired using contrast based magnetic resonance imaging and identifying. From the image data, a global vascular structure within the tissue is identified. After exposing tissue during a medical procedure in the anatomical part, intraoperative image data is acquired by scanning, using hyperspectral imaging, of a selected local region of the tissue. From this intraoperative hyperspectral image data, a local vascular structure in the selected local region of the tissue is located and identified. The global vascular image data is searched for identifying and locating a portion of the global vascular structure geometrically matching the local vascular structure. Upon identifying and locating matching vascular structure between the two imaging modalities, one or more local vascular structures in the selected local region of the tissue is registered with the global vascular structure within the tissue for confirming location of the one or more local vasculature structures.
- Thus, there is disclosed herein a computer implemented method of intraoperatively confirming location of organized tissue structures in relation to a patient's skeletal structure during a medical procedure, comprising:
- acquiring preoperative image data of a patient's skeletal structure in a vicinity of an anatomical part undergoing a medical procedure;
- after exposing tissue during a medical procedure in the anatomical part, acquiring intraoperative image data by scanning a selected region of tissue, in a vicinity of the skeletal structure in the anatomical part undergoing the medical procedure using Polarization Sensitive-Optical Coherence Tomography (PS-OCT);
- identifying, from the intraoperative (PS-OCT) image data, regions of tissue exhibiting structural organization in the vicinity of the skeletal structure; and
- using a priori known anatomical information about the regions of tissue exhibiting structural information for geometrically correlating and registering the intraoperative (PS-OCT) image data with the preoperative image data of the skeletal structure in the vicinity of the anatomical part.
- The pre-operative image data of the skeletal structure in a vicinity of an anatomical part undergoing a medical procedure may be acquired using any one of computed tomography (CT), magnetic resonance imaging (MRI) and optical coherence tomography (OCT).
- An example of the MRI technique is T1 magnetic resonance imaging (T1 MRI).
- The tissue exhibiting structural organization includes, ligaments, tendons, muscle, cartilage, connective membranes, nerves, retina, blood vessel walls, some bone structures, trachea, esophagus, tongue, teeth and other connective tissues.
- The a priori known anatomical information about the regions of tissue exhibiting structural information may include attachment points of tissue exhibiting structural information to the skeletal structure relative to landmark positions on the skeletal structure.
- Disclosed herein is a method, comprising the steps of:
- a) intraoperatively confirming location of organized tissue structures in relation to a patient's skeletal structure during a medical procedure, by:
-
- acquiring preoperative image data of a patient's skeletal structure in a vicinity of an anatomical part undergoing a medical procedure;
- after exposing tissue during a medical procedure in the anatomical part, acquiring intraoperative image data by scanning a selected region of tissue, in a vicinity of the skeletal structure in the anatomical part undergoing the medical procedure using Polarization Sensitive-Optical Coherence Tomography (PS-OCT);
- identifying, from the intraoperative (PS-OCT) image data, regions of tissue exhibiting structural organization in the vicinity of the skeletal structure; and
- using a priori known anatomical information about the regions of tissue exhibiting structural information for geometrically correlating and registering the intraoperative (PS-OCT) image data with the preoperative image data of the skeletal structure in the vicinity of the anatomical part; and
- b) using the registered intraoperative (PS-OCT) image data with the preoperative image data of the skeletal structure in the vicinity of the anatomical part to plan a surgical trajectory to avoid selected regions of the tissue exhibiting structural information.
- Also disclosed is a computer implemented system for intraoperatively confirming location of organized tissue structures in relation to a patient's skeletal structure during a medical procedure, comprising:
- a Polarization Sensitive-Optical Coherence Tomography (PS-OCT) apparatus configured to scan a selected region of tissue after the tissue is exposed during the medical procedure to acquire intraoperative image data of the selected region of tissue;
- a computer processor having a memory storage, said Polarization Sensitive-Optical Coherence Tomography being connected to the computer processor, said memory storage having stored therein preoperative image data of a patient's skeletal structure in a vicinity of an anatomical part undergoing a medical procedure, said memory storage having stored therein a priori known anatomical information about the regions of tissue exhibiting structural information;
- said computer processor being programmed with instructions to
-
- a) identify, from the intraoperative (PS-OCT) image data, regions of tissue exhibiting structural organization in the vicinity of the skeletal structure; and
- b) use the stored priori known anatomical information about the regions of tissue exhibiting structural information to geometrically correlate and register the intraoperative (PS-OCT) image data with the preoperative image data of the skeletal structure in the vicinity of the anatomical part.
- The present disclosure also provides a computer implemented method of intraoperatively confirming location of vasculature structures located below a surface tissue during a medical procedure, comprising:
- acquiring global preoperative image data of tissue in anatomical part undergoing a medical procedure using contrast based magnetic resonance imaging and identifying, from the image data, a global vascular structure within the tissue;
- after exposing tissue during a medical procedure in the anatomical part, acquiring intraoperative image data by scanning, using hyperspectral imaging, a selected local region of the tissue in the anatomical part undergoing the medical procedure;
- identifying, from the intraoperative hyperspectral image data, a local vascular structure in the selected local region of the tissue; and searching the global vascular image data for identifying and locating a portion of the global vascular structure geometrically matching the local vascular structure, and upon identifying and locating matching vascular structure, geometrically correlating and registering the local vascular structure in the selected local region of the tissue with the a global vascular structure within the tissue for confirming location of the local vasculature structures.
- A method is disclosed, comprising the steps of:
- acquiring global preoperative image data of tissue in anatomical part undergoing a medical procedure using contrast based magnetic resonance imaging and identifying, from the image data, a global vascular structure within the tissue;
- after exposing tissue during a medical procedure in the anatomical part, acquiring intraoperative image data by scanning, using hyperspectral imaging, a selected local region of the tissue in the anatomical part undergoing the medical procedure;
- identifying, from the intraoperative hyperspectral image data, a local vascular structure in the selected local region of the tissue;
- searching the global vascular image data for identifying and locating a portion of the global vascular structure geometrically matching the local vascular structure, and upon identifying and locating matching vascular structure, geometrically correlating and registering the local vascular structure in the selected local region of the tissue with the a global vascular structure within the tissue for confirming location of the local vasculature structures; and
- b) using the registered hyperspectral image data with the preoperative image data of the vascular structure in the tissue of the anatomical part to plan a surgical trajectory to navigate through selected regions of the tissue exhibiting vascular structure.
- A further understanding of the functional and advantageous aspects of the disclosure can be realized by reference to the following detailed description and drawings.
- Embodiments will now be described, by way of example only, with reference to the drawings, in which:
-
FIG. 1 shows an example navigation system to support minimally invasive access port-based surgery. -
FIG. 2 is block diagram illustrating system components of a navigation system. -
FIG. 3 is a flow chart illustrating the processing steps involved in a port-based surgical procedure using a navigation system. -
FIG. 4 is an example embodiment port based brain surgery using a video scope. -
FIG. 5A is an example embodiment of a video scope with camera coupler and illumination optics. -
FIG. 5B is an example embodiment of a fiber bundle used to deliver light from external light source to the video scope. -
FIG. 5C is an example embodiment of a video scope and illumination assembly. -
FIG. 6 illustrates an example imaging optical sub-system of the video scope. -
FIG. 7 illustrates the arrangement of illumination optics and filter wheel for wide field of view arrangement. -
FIG. 8A illustrates the non-uniform illumination obtained at the distal end of port with two illumination sources and a port with reflective surface. -
FIG. 8B illustrates the near-uniform illumination obtained at the distal end of the port with two illumination sources and a port with rough surface. -
FIG. 9 an example embodiment illustrating a standard hyperspectral imaging system. -
FIG. 10 is a flow chart illustrating a method to acquire hyperspectral data and white-light images in a multiplex fashion. -
FIG. 11 is an example embodiment illustrating imaging at specific wavelength bands. -
FIG. 12 shows an example, non-limiting implementation of computer control system. - Various embodiments and aspects of the disclosure will be described with reference to details discussed below. The following description and drawings are illustrative of the disclosure and are not to be construed as limiting the disclosure. Numerous specific details are described to provide a thorough understanding of various embodiments of the present disclosure. However, in certain instances, well-known or conventional details are not described in order to provide a concise discussion of embodiments of the present disclosure.
- As used herein, the terms, “comprises” and “comprising” are to be construed as being inclusive and open ended, and not exclusive. Specifically, when used in the specification and claims, the terms, “comprises” and “comprising” and variations thereof mean the specified features, steps or components are included. These terms are not to be interpreted to exclude the presence of other features, steps or components.
- As used herein, the term “exemplary” means “serving as an example, instance, or illustration,” and should not be construed as preferred or advantageous over other configurations disclosed herein.
- As used herein, the terms “about” and “approximately” are meant to cover variations that may exist in the upper and lower limits of the ranges of values, such as variations in properties, parameters, and dimensions. In one non-limiting example, the terms “about” and “approximately” mean plus or minus 10 percent or less.
- Port-based surgery is a minimally invasive surgical technique where a port is introduced to access the surgical region of interest using surgical tools. Unlike other minimally invasive techniques, such as laparoscopic techniques, the port diameter is larger than tool diameter. Hence, the tissue region of interest is visible through the port. Accordingly, exposed tissue in a region of interest at a depth few centimetres below the skin surface, and accessible through a narrow corridor in the port, may be visualized using externally positioned optical systems such as microscopes and video scopes.
- Current methods of tissue differentiation during port-based surgical procedure involves visual verification using externally placed video scope. Tissue differentiation may be useful because surgeons do not have a quantitative means of effectively confirming tissues types during a surgical procedure. Traditionally, hyperspectral imaging has not been anticipated for intra-operative use in brain surgery because this method has a very limited depth of penetration in tissue and may not be effectively used transcranially.
- Further, the narrow corridor in port-based surgery is often occluded when a vessel is accidentally cut. In these incidents, the surgeon may be required to stop his current surgical process (e.g. opening of dura, slight retraction of the sulcus for trans-sulcus navigation of port or resection of tumor tissue) and irrigate the cavity to get a better view of the cavity. Further, such bleeding also limits the surgeon from quickly identifying the location of bleeding so that the particular vessel wall can be coagulated to terminate bleeding.
- Accordingly, in some aspects of the present disclosure, systems and methods are provided for utilizing optical imaging in minimally invasive port based surgical procedures. In some embodiments, hyperspectral devices and methods are described for performing intraoperative tissue differentiation and analysis during such procedures.
-
FIG. 1 shows an example navigation system to support minimally invasive access port-based surgery.FIG. 1 illustrates a perspective view of a minimally invasive port based surgical procedure. As shown inFIG. 1 ,surgeon 101 conducts a minimally invasive port-based surgery on apatient 102 in an operating room (OR) environment. Anavigation system 200 comprising an equipment tower, cameras, displays and tracked instruments assist thesurgeon 101 during his procedure. Anoperator 103 is also present to operate, control and provide assistance for thenavigation system 200. -
FIG. 2 is block diagram illustrating system components of an example navigation system.Navigation system 200 inFIG. 2 includes amonitor 211 for displaying a video image, anequipment tower 201, amechanical arm 202, which supports anoptical scope 204.Equipment tower 201 is mounted on a frame (i.e., a rack or cart) and may contain a computer, planning software, navigation software, a power supply and software to manage the automated arm and tracked instruments. The example embodiment envisions theequipment tower 201 as a single tower configuration with dual displays (211, 205), however, other configurations may also exists (i.e., dual tower, single display, etc.). Furthermore,equipment tower 201 may also configured with a UPS (universal power supply) to provide for emergency power, in addition to a regular AC adapter power supply. - Example embodiment
FIG. 2 also envisionsequipment tower 201 havingrecording module 220 that provides real-time recording of the surgical procedure, capturing audio, video, sensory and multi-modal (i.e., CT, MR, US, etc) inputs from different sources. All relevant data is received atequipment tower 201 and stored in memory byrecording module 220. The surgical procedure may be automatically recorded at the outset or be controlled by the operator and/or administrator. In other embodiments, the procedure may be automatically recorded (by default), but there may be an option to override or delete the recording after the procedure has been completed. - The patient's brain is held in place by a
head holder 217 and inserted into the head is anaccess port 206 andintroducer 210. Theintroducer 210 is tracked using atracking system 213, which provides position information for thenavigation system 200.Tracking system 213 may be a 3D optical tracking stereo camera similar to one made by Northern Digital Imaging (NDI). Location data of themechanical arm 202 andport 206 may be determined by thetracking system 213 by detection offiducial markers 212 placed on these tools. Asecondary display 205 may provide output of thetracking system 213. The output may be shown in axial, sagittal and coronal views as part of a multi-view display. - Minimally invasive brain surgery using access ports is a recently conceived method of performing surgery on brain tumors previously considered inoperable. In order to introduce an access port into the brain, an
introducer 210 with an atraumatic tip may be positioned within the access port and employed to position the access portion within the head. As noted above, theintroducer 210 may includefiducial markers 212 for tracking, as presented inFIG. 2 . Thefiducial markers 212 may be reflective spheres in the case of optical tracking system or pick-up coils in the case of electromagnetic tracking system. Thefiducial markers 212 are detected by thetracking system 213 and their respective positions are inferred by the tracking software. - Once inserted into the brain, the
introducer 210 may be removed to allow for access to the tissue through the central opening of the access port. However, onceintroducer 210 is removed, the access port can no longer be tracked. Accordingly, the access port may be indirectly tracked by additional pointing tools configured for identification by thenavigation system 200. - In
FIG. 2 , aguide clamp 218 for holding theaccess port 206 may be provided.Guide clamp 218 can optionally engage and disengage withaccess port 206 without needing to remove the access port from the patient. In some embodiments, the access port can slide up and down within the clamp while in the closed position. A locking mechanism may be attached to or integrated with the guide clamp, and can optionally be actuated with one hand, as described further below. - Referring again to
FIG. 2 , a small articulatedarm 219 may be provided with an attachment point to holdguide clamp 218. Articulatedarm 219 may have up to six degrees of freedom to positionguide clamp 218. Articulatedarm 219 may be attached or attachable to a point based onpatient head holder 217, or another suitable patient support, to ensure when locked in place,guide clamp 218 cannot move relative to the patient's head. The interface betweenguide clamp 218 and articulatedarm 219 may be flexible, or optionally locked into place. Flexibility is desired so the access port can be moved into various positions within the brain, but still rotate about a fixed point. - An example of such a linkage that can achieve this function is a slender bar or rod. When the
access port 206 is moved to various positions, the bar or rod will oppose such a bend, and move theaccess port 206 back to the centered position. Furthermore, an optional collar may be attached to the linkage between the articulated arm, and the access port guide, such that when engaged, the linkage becomes rigid. Currently, no such mechanisms exist to enable positioning an access port in such a manner. -
FIG. 3 is a flow chart illustrating the processing steps involved in a port-based surgical procedure using a navigation system. The first step involves importing the port-based surgical plan (step 302). A detailed description of the process to create and select a surgical plan is outlined in PCT Patent Application No. PCT/CA2014050272, titled “PLANNING, NAVIGATION AND SIMULATION SYSTEMS AND METHODS FOR MINIMALLY INVASIVE THERAPY”, which is hereby incorporated by reference in its entirety, and which claims priority to U.S. Provisional Patent Application Ser. Nos. 61/800,155 and 61/924,993, which are both hereby incorporated by reference in their entirety. - An example plan, as outlined above, may compose of pre-operative 3D imaging data (i.e., MRI, ultrasound, etc.) and overlaying on it, received inputs (i.e., sulci entry points, target locations, surgical outcome criteria, additional 3D image data information) and displaying one or more trajectory paths based on the calculated score for a projected surgical path. The aforementioned surgical plan may be one example; other surgical plans and/or methods may also be envisioned.
- Once the plan has been imported into the navigation system (step 302), the patient is affixed into position using a head or body holding mechanism. The head position is also confirmed with the patient plan using the navigation software (step 304).
- Returning to
FIG. 3 , the next step is to initiate registration of the patient (step 306). The phrase “registration” or “image registration” refers to the process of transforming different sets of data into one coordinate system. Data may be multiple photographs, data from different sensors, times, depths, or viewpoints. The process of “registration” is used in the present application for medical imaging in which images from different imaging modalities are co-registered. Registration is necessary in order to be able to compare or integrate the data obtained from these different modalities. - Those skilled in the art will appreciate that there are numerous registration techniques available and one or more of them may be used in the present application. Non-limiting examples include intensity-based methods which compare intensity patterns in images via correlation metrics, while feature-based methods find correspondence between image features such as points, lines, and contours. Image registration algorithms may also be classified according to the transformation models they use to relate the target image space to the reference image space. Another classification can be made between single-modality and multi-modality methods. Single-modality methods typically register images in the same modality acquired by the same scanner/sensor type, for example, a series of MR images can be co-registered, while multi-modality registration methods are used to register images acquired by different scanner/sensor types, for example in MRI and PET. In the present disclosure multi-modality registration methods are used in medical imaging of the head/brain as images of a subject are frequently obtained from different scanners. Examples include registration of brain CT/MRI images or PET/CT images for tumor localization, registration of contrast-enhanced CT images against non-contrast-enhanced CT images, and registration of ultrasound and CT.
- Once registration is confirmed (step 308), the patient is draped (step 310). Typically draping involves covering the patient and surrounding areas with a sterile barrier to create and maintain a sterile field during the surgical procedure. The purpose of draping is to eliminate the passage of microorganisms (i.e., bacteria) between non-sterile and sterile areas.
- Upon completion of draping (step 310), the next steps is to confirm patient engagement points (step 312) and then prep and plan craniotomy (step 314).
- Upon completion of the prep and planning of the craniotomy step (step 312), the next step is to cut craniotomy (step 314) where a bone flap is temporarily removed from the skull to access the brain (step 316). Registration data is updated with the navigation system at this point (step 322).
- The next step is to confirm the engagement within craniotomy and the motion range (step 318). Once this data is confirmed, the procedure advances to the next step of cutting the dura at the engagement points and identifying the sulcus (step 320). Registration data is also updated with the navigation system at this point (step 322).
- In an embodiment, by focusing the camera's gaze on the surgical area of interest, this registration update can be manipulated to ensure the best match for that region, while ignoring any non-uniform tissue deformation affecting areas outside of the surgical field (of interest). Additionally, by matching overlay representations of tissue with an actual view of the tissue of interest, the particular tissue representation can be matched to the video image, and thus tending to ensure registration of the tissue of interest.
- For example, video of post craniotomy brain (i.e. brain exposed) can be matched with an imaged sulcal map; the video position of exposed vessels can be matched with image segmentation of vessels; the video position of a lesion or tumor can be matched with image segmentation of tumor; and/or a video image from endoscopy within a nasal cavity can be matched with bone rendering of bone surface on nasal cavity for endonasal alignment.
- In other embodiments, multiple cameras can be used and overlaid with tracked instrument(s) views, and thus allowing multiple views of the data and overlays to be presented at the same time, which can tend to provide even greater confidence in a registration, or correction in more than dimensions/views.
- Thereafter, the cannulation process is initiated (step 324). Cannulation involves inserting a port into the brain, typically along a sulci path as identified in
step 320, along a trajectory plan. Cannulation is an iterative process that involves repeating the steps of aligning the port on engagement and setting the planned trajectory (step 332) and then cannulating to the target depth (step 334) until the complete trajectory plan is executed (step 324). - Returning to
FIG. 3 , the surgeon then performs resection (step 326) to remove part of the brain and/or tumor of interest. Resection (step 326) is a continual loop including both fine and gross resection (step 336). The next step involves hyperspectral imaging (step 338) which may be performed on either fine or gross resection (step 336). Hyperspectral imaging (step 338) is used as a form of tissue differentiation and may assist surgeons to investigate cancerous stem cells. Further, the ability to hyperspectrally image tissue being operated on either as part of an external video scope or as a separate module may provide the ability to perform chemical imaging using the absorption of tissue, the ability to differentiate tissues based on scattering properties, and/or the ability to improve visualization by imaging at wavelengths with reduced absorption or scattering properties. - Once resection is completed (step 326), the surgeon then decannulates (step 328) by removing the port and any tracking instruments from the brain. Finally, the surgeon closes the dura and completes the craniotomy (step 330).
-
FIG. 4 illustrates an example port-based brain surgery procedure using a video scope. InFIG. 4 ,operator 404, typically a surgeon, would alignvideo scope 402 to peer downport 406.Video scope 402 may be attached to an adjustablemechanical arm 410.Port 406 may have atracking tool 408 attached to it where trackingtool 408 is tracked by a tracking system of a navigation system. - Even though the
video scope 402 is commonly an endoscope or a microscope, these devices introduce optical and ergonomic limitations when the surgical procedure is conducted over a confined space and conducted over a prolonged period such as the case with minimally invasive brain surgery. -
FIG. 5A illustrates the design of a video scope that is composed of a lens assembly 511 (explained later) and two illumination delivery points 512. Thelens assembly 511 is terminated at the eyepiece end with a sealedwindow 501 at the proximal end.Sealed window 501 is typically made of quartz, to help maintain water seal since OR devices must be steam cleanable. The eyepiece end also has acamera coupler 505 that provides a standardized mounting point for a camera (not shown). The camera may be a standard definition (SD), high definition (HD) or ultra high definition (UHD) camera. In another embodiment, the camera may be replaced by other imaging technologies such as Optical Coherence Tomography (OCT) or Polarization Sensitive-OCT. The distal end of the lens assembly is also sealed with aclear window 513 at the distal end. The distal end also supportsillumination optics 512. In an alternate embodiment the distal end may be also optionally affixed with a polarizing filter to enable polarization sensitive imaging. - The PS OCT technique described herein may be used to specifically visualize tissue exhibiting structural organization. Examples of such tissue structures include tendons that are attached to bones. Other examples of tissue that exhibit structural organization include ligaments, muscle, cartilage, tissue connective membrane, nerves, retina, blood vessel walls, some bone structures, trachea, esophagus, tongue and teeth. PS OCT commonly generates a heat map or pseudo colored image (reference: “Correlation of collagen organization with polarization sensitive imaging of in vitro cartilage: implications for osteoarthritis,” W. Drexler et. al, The Journal of Rheumatology, Vol. 28, No. 6, 1311-1318) where tissue structures with high degree of organization appear highlighted. Hence the system can be used in orthopedic surgery to visualize tendons and optionally avoid unintentional damage to this tissue during a procedure. These identified regions of tissue exhibiting high level of structural organization (e.g. tendons and ligaments that are often located near skeletal structure) may be used in conjunction with a priori information, such as known points of attachment of tendons to bones, to geometrically correlate PS OCT images to CT and MR images where bones are easily imaged.
- The insertion sites, tendon-bone junctions and ligament-bone junctions, are known as entheses. The anatomical locations of entheses are well known and landmarks can be identified on the bone in the vicinity of these attachment points (reference: “Anatomy and biochemistry of enthuses,” Michael Benjamin, Ann Rheum Dis 2000, Vol. 59, Issue 12, pg: 995-999). Hence, this a priori anatomical information about the position of the tendon or ligament relative to bone structures in the vicinity can be used to register intraopertive PS-OCT image of the tendons or ligaments with pre-operative images obtained using other modalities that accurately image the bone structures.
- For example the tendon-bone junction in the Achilles tendon enthesis is immediately proximal to the superior tuberosity. This region is characterized by a highly irregular interface at the attachment points or junction. This characteristic structure of the bone can be used to identify the junction where the tendon attaches to the bone. The geometric correlation of images that are thus obtained using different modalities, and often at different scales, is known as image registration or image fusion.
- Common methods for multi-modal image registration mentioned above include those described in “Multi-modal image registration for pre-operative planning and image guided neurosurgical procedures,” Risholm, et. al, Neurosurg Clin N Am, 2011, April; 22(2): 197-206 and “Image registration of ex-vivo MRI to sparsely sectioned histology of hippocampal and neocortical temporal lobe speciments,” Goubran et. al, Neurolmage, 83 (2013); 770-781. Broad classes of image registration methods for medical images is also described in detail in “A survey of medical image registration,” Maintz et. al, Medical Image Analysis (1998), Vol. 2, No. 1, pp: 1-36.
- The illumination optics is comprised of
fiber bundles 507 that are rotatably attached using a pair ofconnectors 510. Theconnectors 510 allow the fiber bundles to rotate freely (570 inFIG. 5C ) within the connector while maintaining a fixed distance between thelens 509 and tip of thefiber bundle 507 using aloose sleeve 508. This rotation movement will reduce the strain on the fiber bundle when the video scope is moved on a holding system (not shown) or amechanical arm 410 as seen inFIG. 4 . Therotatable connector 510 also aid in easy cable management when themechanical arm 410 is moved during a surgical procedure. The illumination optics are placed as close as possible to the objective lens. One non-limiting example of spacing between the optics is approximately 30 to 35 mm, or 32 to 34 mm, between the center of thelenses 509 where the diameter oflenses 509 is approximately 15 mm. This configuration is optimal for illuminating the bottom of a surgical port with maximum intensity when the distal end of the video scope is between 25 cm to 40 cm from the bottom of the surgical port. Anoptical compensator 503 is used to act as a thermal compensator to control the stress on optical components during steam cleaning. Aholder 506 provides an easy to grasp assembly to hold and manipulate the video scope without introducing mechanical stress on the lens assembly. The lens assembly is encased in a sealedbarrel 511 to avoid ingression of steam and liquids during normal use and cleaning. Therotatable attachment mechanism 510 allows free rotation of the fiber bundles when the camera is moved manually or when mounted to a robotic positioning system. This, in turn, avoids undue stress on the fiber bundles that are susceptible to fracture. -
FIG. 5C illustrates a non-limiting example to realize a functionality that allows theillumination assembly 565 to rotate radially 560 around thevideo scope barrel 502. Theillumination assembly 565 is composed of the twofiber bundles 507 on either side of the video scope, mountingmechanism 508 and lens 509 (as inFIG. 5A ). This allows the surgeon to adjust the radial orientation of the illumination and orient the illumination assembly so that it minimally obstructs the surgeon's view of the surgical space. The illumination assembly can be freely rotated without rotating the video scope by securing the video scope to an external positioning mechanism, such as 410, using aremovable clamp 555 and an associatedlock 550. The removable clamp'sdistal end 555 and the associatedlock 550 may be mated together using a thread mechanism or any other mechanical fastening mechanism. The removal clamp's proximal end (not shown) may be secured to theexternal positioning mechanism 410. It should be further noted that therotation 560 enabled in this design along with therotation 570 of the fiber bundles within the connectors enable positioning and orientation of the video scope with minimal interruption of the visible surgical space and minimize strain on the fiber bundles during motion. Finally, theillumination assembly 565 may be replaced with alternative configurations such as ring lights or single illumination points. Ring lights may be realized through circular arrangement of fiber strands (not shown) from an optical fiber bundle around the circumference of the objective lens. Single illumination points may be realized through removal of one of the twosplit fiber bundles 507 from the design. - The illumination assembly preferably receives the light input from an optical source that is located away from the video scope. This reduces the total weight of the external scope and allows for easy manipulation of the video scope by a manual positioning system (not shown) or a
mechanical arm 410. The light from the light source is delivered to the video scope through the use of a fiber bundle. Presence of two delivery points represented byillumination optics 512 inFIG. 5A requires the use of a fiber bundle that is split in two. This design of fiber bundle is also known as a Y-cables. An example embodiment of this Y-cable design is illustrated inFIG. 5B . InFIG. 5B ,rotatable connections 508 are provided on thefasteners 510 at the two distal end of the Y cable, providing a mechanism for freely rotating the fiber bundles to avoid fracture of the bundles. A strain-relief 527 helps maintain a minimum limit on thebend radius 529 of the bundle between the two distal ends and the Y-junction 531. Y-junction 531 helps reduce bend strain on thefiber bundle 507. Strain-relief 533 similarly aids in reducing bend strain near theconnector 535 at the proximal end of the Y-cable.Cross sections junction 531 placed equidistant from the two ends. This dimension provides for placement of light source on a cart orinstrumentation tower 201 sufficiently away from themechanical arm 410 while minimizing light loss due to excessive length of the fiber bundle. -
FIG. 6 illustrates an optical design of the video scope that limits the diameter of the objective lens 600 (front lens). This design enables the mounting of illumination optics immediately adjacent to the objective lens so that the illumination beam can be almost collinear to the return path of the light reflected from the tissue. The illumination beam and the reflected beam need to be as collinear as possible so that maximum illumination is delivered at the bottom of theaccess port 406. Finally, the optical design is constrained so that the length of the lens assembly is minimized to make thewhole video scope 402 minimally intrusive to the surgeon's field of view and facilitate easy access to the surgical space by the surgeon. This constraint is a challenge in conventional optical design conventional optical design techniques maximize zoom by utilizing maximum available physical length of the lens assembly during the design process. This optical design of the present disclosure is adapted from a conventional endoscopic system that consists ofobjective lens 600,relay lens 602 andeyepiece 604. The zoom parameter of the optical assembly is chosen such that the minimum field of view (corresponding to maximum zoom) is equal to approximately 13 mm. This dimension is the diameter of the surgical port. The field of view of 13 mm needs to be achieved at a minimum working distance of 25 cm where the minimum working distance is defined as the distance between the distal end of the video scope (402 inFIG. 4 ) and bottom of the surgical port (406 inFIG. 4 ). As explained in FIG. 5A, acoupler 505 is used to attach a camera at the eyepiece end (marked ‘E’ inFIG. 6 ). The optical design of the objective is composed of 1 doublet and 1 singlet; the relay is composed of 1 doublet and 1 singlet and the eyepiece is composed of 2 singlet and 1 doublet. Any manufacturing error is compensated using oneoptical compensator 503 that is placed between the objective and relay. The length of the optical sub-assembly is minimized through the use of higher power lenses and fewer lens groups. - The type of surgical procedure determines either a wide-field of view (WFOV) or a narrow field of view (NFOV) video scope. For example, a neck surgery may benefit from a WFOV video scope where large area is captured by the video scope; whereas, a port-based brain surgery may benefit from a NFOV video scope. Instead of attempting to address both these design requirements using one device, two separate designs may be developed such that they share several sub-components and the manufacturing process. Hence, it is economical to manufacture two different designs while sharing number of design elements and assembly procedure. Both WFOV and NFOV designs share a similar
optical illumination system 512 as seen inFIG. 5A , The WFOV design can be realized by attaching a camera to thecamera coupler 505. The zoom adjustment of the camera is used to determine the field of view in this case. -
FIG. 7 illustrates an assembly with a non-coaxial illumination source. Theillumination system 710 is similar in design to that illustrated inFIG. 5A and consists of fiber bundles 704 (only a distal portion of which are shown in the Figure). An air-filed opaque tube (also known as optical tube) 702 is used to position the illumination mechanism away from the camera attached to thecoupler 505. It should be noted that any required magnification may be provided by the camera lens (not shown but typically attached to the camera coupler) for WFOV application. A finite space that is at least 1 mm between theplane 706 of the distal end of the optical tube and the plane of theillumination optics 708 helps isolate the illumination light from directly reaching the camera input. It should be further noted that the dimensions of the WFOV optics will be such that the illumination will not be nearly coaxial with the path of the reflected light. This is not a limitation in this configuration because WFOV is used to observe a surgical space that is larger that of a port (which is approximately 13 mm). Hence, general illumination is sufficient. Placement of the illumination source close to the camera does improve illumination of the surgical area compared to the use of overhead surgical lights and avoids glare from area outside of the surgical space. The role of additional components, 712 and 714, are explained below in the context of hyperspectral imaging. - In another embodiment of the video scope, the illumination sources placed immediately adjacent to the distal end of the video scope may be employ a light source such as luminance light emitting diodes or Super Luminescent Diodes (SLD's) (not shown). Since the light sources are not coaxial to the reflected light path (the light path incident on the lens and camera assembly), the light sources have to be aimed or steered at the focal plane of interest. Such steering may be achieved using movable fiber bundle mounts 510 as shown in
FIG. 5A . - Application of such externally positioned illumination sources in port-based imaging introduces several challenges. First, the walls of the port are either partially or fully reflective. This introduces localized regions in the imaged surface that have higher intensity of incident light. Such regions are commonly known as hot-spots. It is desirable to avoid such high intensity regions as these tend to saturate sensors and, hence, limit the dynamic range of the sensors in the camera mechanism. Use of post-processing to normalize intensities is less optimal as saturation of sensors results in information loss that cannot be recovered. Presence of high intensity regions can be reduced through the use of surface textures on the port walls that diffuse the light. The impact of using smooth and rough surface texture on the port walls is illustrated in
FIGS. 8A and 8B , respectively. The reflections resulting from textured walls is referred to as Lambertian reflection. The assessment presented inFIGS. 8A and 8B were conducted using ray-tracing tools and the resulting intensity of light at the surface of the tissue (distal end of the port) were visualized using heat-maps or pseudo color where high intensity corresponded to white and low intensity corresponded to black. - Another approach to uniformly illuminating at the bottom of the port is to model the light rays using a commonly known optical modelling method, such as ray tracing, and establish the optimal orientation of the light sources that minimize hot-spots at the bottom of the surgical port. Orientation of the light sources may be modified using a beam steering mechanism, such as the one illustrated in
FIG. 5A . Alternatively, a robotic positioning system may be used to achieve this steering. - Port-based imaging is also limited by highly reflective nature of some but not all regions of the brain tissue due to the presence of blood, CSF or other fluids. In the latter case, an initial image could be acquired to identify regions with high intensity reflected light and this information can be used to reposition direction of the light sources in an attempt uniformly distribute the incident light. As described above, imaging using white light has several challenges in the operating room. Several of these challenges can be overcome by limiting the spectral range of the light that is observed or by judiciously combining selected wavelength bands to visualize human tissue in the operating room.
-
FIG. 9 illustrates a video scope that has been adapted to accommodate hyperspectral imaging capabilities. In this embodiment, tunable light source that is adapted based on the surgical context e.g. selection of illumination spectral region where blood is highly absorptive (to detect blood clots) or transmissive (to avoid excessive light scattering) may be used. -
FIG. 9 illustrates one such system. The tunable light source is mainly composed of abroadband light source 1100, aspectral separation mechanism 1140, aspectral filtering mechanism 1150 and a mechanism to combine the filteredfrequency bands 1170. The combining mechanism consists of a lens and a fiber bundle that mixes all the reflected wavelength bands into one beam that is transmitted through thefiber bundle 507. The light fromlight source 1100 is passed through aslit 1110 to generate a narrow beam. This light is then collimated usingoptical elements -
FIG. 9 illustrates the use of areflective grating 1140. The spatially separated beam is filtered by selectively reflecting portions of the spatially separated beam. This is achieved using a spatial light modulator,SLM 1150, such as a Digital Light Processor (Texas Instruments Inc). An SLM is composed of an array of micro-mirrors that can be electronically activated to act as mirrors or deactivated to acts as opaque surfaces. Hence, specific portions of the spectrum are reflected while other regions are suppressed based on the pattern of activated micro-mirrors. The beam that is now composed of selective portions of spectrum are combined using focusingoptics 1160 and acombiner 1170. - The recombined beam is now composed of only those wavelengths that were selectively reflected by the spatial light modulator,
SLM 1150. This light can be used as the illumination source of an imaging system or external scope by transmitting the light via alight pipe 507 to the illuminator connector andlens mechanism 510 attached to the external scope. It should be noted that the video scope illustrated inFIG. 9 shows the connection oflight pipe 507 to only one of the twoilluminator connectors 510 for the sake of simplicity of the illustration. Details of connecting the light pipe to the video scope is further explained inFIG. 5A . - The reflected light from the
tissue 1198 is captured by the external scope that is composed oflens assembly 502. As detailed inFIG. 5A , the lens assembly is composed; this light is captured using ahigh resolution detector 1125 that is usually a charge coupled device, CCD. The specific band of wavelengths that are reflected by the SLM are controlled by anSLM controller 1180 that is under the command of acomputer 1185. The same computer is used to acquire the image from thedetector 1125. Hence, the computer can synchronize the illumination of amaterial 1198 with a specific wavelength band or wavelength bands of light and acquire corresponding reflected light. This association of illumination wavelength and acquired image can be used to construct a hyper-spectral image where each image is a 2D or 1D image and the third dimension is an index that corresponds to illumination wavelength band(s). Since the individual micro-mirrors located in an SLM can be operated at a rate as high as 4 kHz, subsequent frames of the field of view can be obtained at different wavelength bands. - Further, some of the acquired frames can be for employed white-light illumination of the tissue. This is possible by operating the acquisition camera at a frame rate that is sufficiently high to provide smooth video playback, as perceived by a human observer when white light frames are intermittently obtained while collecting hyperspectral image frames. For example, in some non-limiting examples, the frame rate may be selected to be higher than 20 frames per second, higher than 24 frames per second, or higher than 30 frames per second, in order to support white light video acquisition at such frame rates while obtaining hyperspectral data. For example, at a camera frame rate higher than 20 fps, a white-light image can be acquired every 1/20th of a second and any additional frame can be allocated for acquisition using specific wavelength bands. A white light video feed may then be separately generated and displayed based on the collected white light images. This allows the surgeon to continuous view a white-light image of the surgical area while acquiring any additional images at different wavelength bands in a multiplexed manner. The white-light image stream (or video) may be viewed in one display or sub-section of a display and other images acquired using other wavelength bands may be viewed in a second display or second sub-section of the same display.
- The individual wavelength bands can be composed of non-overlapping individual wavelength bands or combination of bands that may overlap. Alternatively, at least one of the acquired frame can correspond to
illumination 1197 of thesubject material 1198 using the entire wavelength band of the light source. The entire wavelength band could be also normalized to ensure that all the intensity in the output light emanating from thecombiner 1170 is consistent across the entire spectrum. This is known as white balancing. In summary, the same optical mechanism can be used to acquire hyperspectral images and white-light images that are interspersed among each other in the acquired sequence of images. This embodiment eliminates the need for splitting the acquired beam into separate paths so that one beam is captured by a hyperspectral imaging system while the other beam is captured by a white-light camera. This reduces the design complexity of the optical system and aids in making the system more compact as the spectral shaping part of the system can be separated from the imaging system using a light pipe to channel the output light from the light source. It should be noted that the sample being imaged 1198 may be an ex-vivo tissue sample or portion of the brain tissue that may be exposed through a port-based neurosurgical access inserted in the skull. - The software system used to acquire hyperspectral data and white-light images (or video) in a multiplex fashion is illustrated in
FIG. 10 . First the range of wavelengths (wave bands) that are of interest are stored in a table (step 1200). Then, specific wave band for illumination is selected from the table (step 1220). Each entry in this table is used to look up (step 1230) specific micro-mirrors that need to be activated using another table (step 1210). Hence, only the micro-mirrors associated with specific wavelength bands are activated (step 1240). Activation of a micro-mirror turns it into a micro-reflector instead of an opaque surface. Hence, thesample 1198 inFIG. 9 ) is illuminated with light (1197 inFIG. 9 ) that is composed of specific wavelength bands. The table (step 1200) may also include entries that activate the entire spatial light modulator (SLM). In this case, the SLM acts as a mirror for the entire bandwidth of the light source and the acquired image will correspond to white-light illumination. - Returning to
FIG. 10 , the reflected light from the illuminated sample is acquired (step 1250) by the same computer and associated with the specific wavelength band (step 1260). The type of illumination (white-light versus specific wavelength band) used for each acquired image is interrogated (step 1270) in order to appropriately classify the acquired image as part of white-light image (video) or part of the hyperspectral image data set. If the acquired image corresponds to a narrow wavelength band then it is stored as part of the hyperspectral image set (step 1280). If the image corresponds to white-light illumination, it is stored as white-light image or a stream of such images may be captured to represent a video stream. This acquisition is repeated (step 1215) until all the wavelength bands of interested are sequentially used to illuminate the sample material. Hence, the resulting image set will be composed of both hyperspectral image sets (step 1280) and white-light image sets (step 1290), all acquired using the same hardware. - Ideally, the video stream needs to be at least 30 frames per second to provide a flicker-free video to the surgeon. If a total of 40 frames are acquired per second, the additional 10 frames may be used to store images corresponding to 10 distinct or overlapping wavelength bands. Hence, if the total frame rate of the acquisition system is n frames per second, n−30 frames may be allocated towards n−30 wavelength bands in the hyperspectral image data set.
- An alternative to tunable
light source 1110 shown inFIG. 9 may be monochromatic, spanning ultra violet (UV), visible, and/or near infrared (NIR) wavelengths, continuous wave or pulsed that is used to illuminate the tissue using free space or fiber coupled mechanism - In another embodiment, specific wavelength bands may be acquired by filtering the reflected light from a broadband light source using such spectral elements as discrete wavelength filters (on filter wheel or spatial on-chip filters), liquid crystal filters, spectrographs/spectrometers/spectral gratings, spatially varying gratings, fiber-coupled spectrometers.
-
FIG. 7 also illustrates the implementation ofdiscrete filters 712 attached to a rotatable filter wheel 714) that may be motorized. This filter mechanism is attached at the distal end of the video scope. Another alternative to discrete filters at the input to the video scope is a liquid crystal-based tunable wavelength filter (not shown) to pass only a narrow range of wavelengths. This filter can be tuned to a number of different wavelengths and operates in a similar manner to the discrete filters as an image is acquired for each wavelength the filter is tuned to. In yet another embodiment, diffraction grating based systems that separate input light input its constituent wavelengths may be used in lieu of thecamera 1125 shown inFIG. 9 . Imaging spectrometer systems rely on scanning the entrance slit of the system across the field to be imaged. Thus the acquisition time is limited by the scanning time. The entrance slit of the spectrometer can be either free space or fiber coupled to the optical path. If an array-to-line fiber mapping is utilized it is possible to acquire all spatial and spectral information simultaneously. The spectrometer could be alternatively equipped with Spatially Varying Gratings where a specialized diffraction grating that allows for the collection of spectra from all pixels in a single acquisition. The grating is divided into a number of spatial gratings each with a varying direction of diffraction. An image is acquired that captures the diffracted light from each of these grating regions, this image is then reconstructed to form the hyperspectral data set. - Non-limiting examples of
camera 1125 include monochrome video camera with resolution up to high definition (HD) or ultra high definition (UHD). CCD, CMOS, InGaAs, or HgCdTe device. - Another aspect of confocal hyperspectral imaging system is that the entire tissue surface does not have to be scanned in a raster pattern. Instead, random spots can be accumulated until a reasonable match is found against pre-defined data classes. This can significantly reduce the data acquisition time associated with hyperspectral imaging.
- In some embodiments, the hyperspectral imaging system illuminates the tissue with monochromatic or broadband light, collects light reflected from the tissue, controls the wavelength of the detected light in such a way that a series of images, each recorded at different wavelengths or wavelength ranges, is collected. This series of images, known as a hyperspectral dataset, is processed to extract tissue's bio-chemical or microstructural metrics and reduced to 2D (spatial). This reduced 2D image may be spatially registered and can be overlaid on the external video scope image as well as any other pre- and intra-operative images. For example, methods of correlating image data are disclosed in PCT Patent Application No. PCT/CA2014/050269, titled “INTRAMODAL SYNCHRONIZATION OF SURGICAL DATA” and filed on Mar. 14, 2014, the entire contents of which are incorporated herein by reference for the purposes of the U.S. national phase patent application or continuation by pass application that claims priority from this PCT application. Spatial registration is realized by using navigation markers attached directly on the camera or on structures rigidly and consistently attached to the camera. This provides both location and orientation of the imaging system. This is further explained in the disclosure related to automated guidance of imaging system.
- The
hyperspectral dataset 1280 inFIG. 10 is then processed to extract the tissue specific information and reduce the dimensionality of the data. Tissue specific information can range from tissue type identification to inferring pathology associated with a region of the acquired image. Examples of the possible processing methods including the following: - In one embodiment, if the spectral peaks or features of chemical(s) of interest are known, the spectra and be processed, through either peak or feature detection algorithms, to detected the peaks or features to give an indication of the chemical presence and some indication of the concentration or quality. This useful only if the specific chemicals of interest are known.
- In one embodiment, the spectra of specific tissues or tissue states of interest can be acquired and stored in a database, as disclosed in PCT Patent Application No. PCT/CA2014/050269, titled “INTRAMODAL SYNCHRONIZATION OF SURGICAL DATA” and filed on Mar. 14, 2014. Spectra then acquired during the surgery can be compared to the spectra stored in the database for similarity and if sufficiently similar to give an indication of what tissue or tissue type the spectra was acquired from.
- Multivariate/chemometric methods, which are a wide grouping of statistical techniques where a method is trained on spectra collected from samples with known states (i.e., spectrum and corresponding chemical level, tissue type, tissue state, etc.), may be used to predict the state of a new sample based on the acquired spectrum. Some of the more commonly used employed techniques include principal component regression (PCR), partial least squares (PLS), and neural networks (NN).
- The aforementioned analysis methods can be implemented in a computer system, and hence the results of the analysis can be obtained in near-real time for appropriate use by a surgeon. This may significantly reduce the need for similar analysis by a pathologist and reduces the wait time associated with obtaining results of such tissue analysis. Correlation metrics between newly acquired data and representative data in a knowledge-base (or database or training set) provide the surgeons a means of quantifying tissue types. Such metrics may be a representation of confidence associated with automated inference provided by the software algorithm.
- Finally, the ability to selectively view narrow bands of the spectra or reject narrow bands of the spectra may allow the surgeon to reject bright reflections from blood. Hence, the surgeon may be able to view the interior of the corridor and proceed with surgical resection of tumor even when the corridor is occluded by excessive bleeding. This will reduce the need to constantly irrigate the narrow corridor and hence reduce interruption of the surgical procedure.
- In another embodiment the optical characteristics and chemical composition of blood may be taken advantage of to visualize vasculature located at or immediately below tissue surface. For example, a common challenge associated with the dural opening step in cranial surgery is the inability to anticipate the presence of vasculature immediately below the dura. Hyperspectral imaging (HSI), with emphasis on red and near infrared portions of the imaging spectrum, can be used to selectively visualize regions with high haemoglobin content. Due to the blood-brain barrier that naturally exists in the human brain, this technique will result in selective imaging of vasculature.
- An example method for visualizing vascular structures located below other tissue structures is disclosed in “Characterization of vascular structures and skin bruises using hyperspectral imaging, image analysis and diffusion theory,” L. L. Randeberg et. al., Journal of Biophotonics, No. 1-2, 53-65 (2010). Another application of hyperspectral imaging, such as that enabled by invention presented in this document, is the reliance on reflectivity of hemoglobin that is particularly strong in tumor microvasculature. This is described in detailed in “Hyperspectral imaging of hemoglobin saturation in tumor microvasculature and tumor hypoxia development,” B. S Sorg, et. al., Journal of Biomedical Optics, 10(4), July/August 2005.
- Similarly, OCT based imaging may be also used to visualize structures located immediately below the dura since the dura is not more than a millimetre thick. Vasculature is one of the predominant structures located below the dura. Others structures of relevance that can be similarly imaged include the sulcal folds.
- The vasculature that is visualized using the techniques described in the preceding paragraphs is a map of the structure of the blood vessels in the vicinity of the region being imaged. Such regions may include the region in the vicinity of the trajectory along which a port is inserted in port-based cranial surgery. This vascular structure in the surgical region will be referred to as the in-situ vascular structure or intra-operative local vasculature. This in-situ geometry of the vascular structure may be then be compared with whole-head vascular structure images derived from MRI acquired after the injection of Gadolinium or other contrast medium in the veins. The latter whole-head images are obtained prior to a surgical procedure and referred to as pre-operative vascular structures. The pre-operatively vascular structures are imaged relative to the anatomy of the patient. In other words, the location and orientation of this vascular structure is known relative to the anatomy of the patient.
- This comparison of in-situ (intraoperative) vascular structure with pre-operative vascular structures may be used to infer the exact location of the in-situ vascular structures relative to the pre-operative vascular structures. This allows the surgeon to confirm or infer their current surgical position relative to the anatomy of the patient. Hence, the vascular structure can be used as a land-mark or reference frame for navigated surgical procedures. Although the above illustration is in the context of cranial surgery, it will be appreciated that the method can be extended to any other part of the anatomy where vasculature can be visualized pre-operatively and intra-operatively. Example procedures include, retinal surgery and lung biopsy to mention just a few.
- Geometric correlation of the local vascular structure with the global vascular structure is feasible because the anatomy and pattern of blood vessels is known (reference: “Cortical blood vessels of the human brain,” Duvernoy et. al., Brain Research Bulletin, Vol. 7, Issue 5, November 1981, pg 519-579 and “A computed tomographic guide to the identification of cerebral vascular territories,” Damasio et. al., Arch Neurol, 1983; 40(3): 138-142. Thus while it is known that registration, in general, involves the geometric correlation of small scale image (or 3D structure) with a large scale image (or 3D structure); however, this geometric correlation involves the use of features that are common between the two image sets. The use of the vasculature by itself as a feature is as disclosed herein is the first instance of this. The vasculature can be used as a feature because of its unique structure and this uniqueness is well known (reference: see two papers cited in this paragraph). Unique vascular structure is known to exist in several regions of the body, for example the cerebral region, retina and the cardiac regions.
- Thus, using the intraoperative hyperspectral image data, a local vascular structure in the selected local region of the tissue is identified, and then the global vascular image data is searched for identifying and locating a portion of the global vascular structure geometrically matching the local vascular structure, and upon identifying and locating matching vascular structure, geometrically correlating and registering the local vascular structure in the selected local region of the tissue with the a global vascular structure within the tissue for confirming location of the local vasculature structures. The location, thus inferred, is invaluable for guiding tools in the local region for surgical procedures. Such surgical procedure is known as navigated surgery.
- It is noted that embodiments provided herein may employ software to process the 3D dimensional data sets to extract the information of interest, and to reduce the data to a 2D image that can be visualized in conjunction with or overlaid on the surgical image acquired by the external video scope. These software methods could include everything from simple spectral peak detection to more sophisticated multivariate, chemometric, and data mining techniques to extract the metric of interest from the acquire spectra. The spectrum associated with each pixel may be processed according to such methods.
- As hyperspectral imaging is an optical technique and limited penetration (2-3 mm), its use is restricted to superficial tissues or those exposed through corridor surgery. The unique spectra of chemicals in tissue provide the potential to use hyperspectral imaging to image chemical content and from this provide useful qualitative or quantitative information to the surgeon to assist in decision making during the surgery. Chemical imaging can be used to differentiate between different tissues based on differing chemical composition and associated differing absorption (e.g., white vs grey matter), determine tissue state (e.g., normal vs malignant), and determine tissue status and/or health (e.g., state of oxygenation). The difference in spectral scattering properties can, similar to absorption changes, be used to determine the properties of tissue based on changes in cellular structure with tissue type (e.g., fat vs nerve fiber) and state (e.g., changes in nuclear and overall cell size with pre and cancerous states). Lastly, as the acquired hyperspectral data set contains data acquired at a variety of wavelength, images at only selected wavelengths or wavelength ranges to improve the visualization of tissue (minima or maxima in absorption or scattering). For example, images at wavelengths where hemoglobin absorption is at a minimum, the absorption due to blood will be significantly reduced thus providing additional light for illumination.
- This advantage of imaging at specific wavelength bands is illustrated in
FIG. 11 .FIG. 11 illustrates a standard color image (A) of a brain region (Corpus Callosum) that is also captured using four different wavelength bands centered at 400 nm (B), 500 nm (C), 600 nm (D) and 700 nm (E) and a bandwidth of 10 nm each. It is evident that that 400 nm filter band clearly illustrates tissue structures that are otherwise invisible in other wavelength bands. -
FIG. 12 illustrates the key components of thecomputer system 1185 ofFIG. 9 .FIG. 12 provides an example, non-limiting implementation ofcomputer control system 425, which includes one or more processors 430 (for example, a CPU/microprocessor),bus 402,memory 435, which may include random access memory (RAM) and/or read only memory (ROM), one or more internal storage devices 440 (e.g. a hard disk drive, compact disk drive or internal flash memory), apower supply 445, onemore communications interfaces 450, and various input/output devices and/orinterfaces 460 such as a user interface for a clinician to provide various inputs, run simulations etc. - Although only one of each component is illustrated in
FIG. 12 , any number of each component can be includedcomputer control system 425. For example, a computer typically contains a number of different data storage media. Furthermore, althoughbus 402 is depicted as a single connection between all of the components, it will be appreciated that thebus 402 may represent one or more circuits, devices or communication channels which link two or more of the components. For example, in personal computers,bus 402 often includes or is a motherboard. - In one embodiment,
computer control system 425 may be, or include, a general purpose computer or any other hardware equivalents configured for operation in space.Computer control system 425 may also be implemented as one or more physical devices that are coupled toprocessor 430 through one of more communications channels or interfaces. For example,computer control system 425 can be implemented using application specific integrated circuits (ASIC). Alternatively,computer control system 425 can be implemented as a combination of hardware and software, where the software is loaded into the processor from the memory or over a network connection. - In another example embodiment, a vertical slit or a focal point may be imaged by the video scope using a confocal optical design that is commonly used in a microscope (not shown). The spot or slit may be then imaged on a photomultiplier to generate a very sensitive hyper-spectral imaging system. The focal point may be swept across the sample surface using a scanning mechanism. A commonly used scanning mechanism is a galvanometer mirror system.
- The specific embodiments described above have been shown by way of example, and it should be understood that these embodiments may be susceptible to various modifications and alternative forms. It should be further understood that the claims are not intended to be limited to the particular forms disclosed, but rather to cover all modifications, equivalents, and alternatives falling within the spirit and scope of this disclosure.
- While the Applicant's teachings described herein are in conjunction with various embodiments for illustrative purposes, it is not intended that the applicant's teachings be limited to such embodiments. On the contrary, the applicant's teachings described and illustrated herein encompass various alternatives, modifications, and equivalents, without departing from the embodiments, the general scope of which is defined in the appended claims.
Claims (20)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2016574309A JP2017508579A (en) | 2013-03-15 | 2014-09-15 | Method and system for locating tissue structure during a medical procedure |
US15/023,120 US20160235482A1 (en) | 2013-03-15 | 2014-09-15 | Methods and systems for intraoperatively confirming location of tissue structures |
Applications Claiming Priority (9)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361800911P | 2013-03-15 | 2013-03-15 | |
US201361800695P | 2013-03-15 | 2013-03-15 | |
US201361801530P | 2013-03-15 | 2013-03-15 | |
US201361818223P | 2013-05-01 | 2013-05-01 | |
US201361818280P | 2013-05-01 | 2013-05-01 | |
PCT/CA2014/050268 WO2014139020A1 (en) | 2013-03-15 | 2014-03-14 | Surgical imaging systems |
CAPCT/CA2014/050268 | 2014-03-14 | ||
US15/023,120 US20160235482A1 (en) | 2013-03-15 | 2014-09-15 | Methods and systems for intraoperatively confirming location of tissue structures |
PCT/CA2014/050877 WO2015135058A1 (en) | 2014-03-14 | 2014-09-15 | Methods and systems for intraoperatively confirming location of tissue structures |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CA2014/050877 A-371-Of-International WO2015135058A1 (en) | 2013-03-15 | 2014-09-15 | Methods and systems for intraoperatively confirming location of tissue structures |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/516,496 Continuation US20190388160A1 (en) | 2013-03-15 | 2019-07-19 | Methods and systems for intraoperatively confirming location of tissue structures |
Publications (1)
Publication Number | Publication Date |
---|---|
US20160235482A1 true US20160235482A1 (en) | 2016-08-18 |
Family
ID=51535737
Family Applications (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/398,649 Active US10292771B2 (en) | 2013-03-15 | 2014-03-14 | Surgical imaging systems |
US15/023,120 Abandoned US20160235482A1 (en) | 2013-03-15 | 2014-09-15 | Methods and systems for intraoperatively confirming location of tissue structures |
US16/414,957 Active US10786314B2 (en) | 2013-03-15 | 2019-05-17 | Surgical imaging apparatus and methods |
US16/516,496 Abandoned US20190388160A1 (en) | 2013-03-15 | 2019-07-19 | Methods and systems for intraoperatively confirming location of tissue structures |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/398,649 Active US10292771B2 (en) | 2013-03-15 | 2014-03-14 | Surgical imaging systems |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/414,957 Active US10786314B2 (en) | 2013-03-15 | 2019-05-17 | Surgical imaging apparatus and methods |
US16/516,496 Abandoned US20190388160A1 (en) | 2013-03-15 | 2019-07-19 | Methods and systems for intraoperatively confirming location of tissue structures |
Country Status (12)
Country | Link |
---|---|
US (4) | US10292771B2 (en) |
EP (1) | EP2967349B1 (en) |
JP (2) | JP6527086B2 (en) |
KR (1) | KR102364976B1 (en) |
CN (1) | CN105263398B (en) |
AU (2) | AU2014231342B2 (en) |
BR (1) | BR112015023545B1 (en) |
CA (3) | CA2894133C (en) |
HK (1) | HK1219405A1 (en) |
MY (1) | MY177299A (en) |
SG (1) | SG11201507609UA (en) |
WO (1) | WO2014139020A1 (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20160324583A1 (en) * | 2014-10-14 | 2016-11-10 | Leila KHERADPIR | Patient reference tool |
US20170148180A1 (en) * | 2014-06-25 | 2017-05-25 | Koninklijke Philips N.V. | Imaging device for registration of different imaging modalities |
WO2018049215A1 (en) | 2016-09-09 | 2018-03-15 | Intuitive Surgical Operations, Inc. | Simultaneous white light and hyperspectral light imaging systems |
US20180247153A1 (en) * | 2017-02-28 | 2018-08-30 | Verily Life Sciences Llc | System and method for multiclass classification of images using a programmable light source |
EP3477246A1 (en) * | 2017-10-27 | 2019-05-01 | Nokia Technologies Oy | Apparatus, systems and methods for detecting light |
WO2019152261A1 (en) * | 2018-01-30 | 2019-08-08 | Ingle Manish Eknath | Surgical visualization and recording system |
EP4073748A4 (en) * | 2019-12-13 | 2024-01-17 | ChemImage Corporation | Methods for improved operative surgical report generation using machine learning and devices thereof |
Families Citing this family (181)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10064554B2 (en) | 2011-12-14 | 2018-09-04 | The Trustees Of The University Of Pennsylvania | Fiber optic flow and oxygenation monitoring using diffuse correlation and reflectance |
US11871901B2 (en) | 2012-05-20 | 2024-01-16 | Cilag Gmbh International | Method for situational awareness for surgical network or surgical network connected device capable of adjusting function based on a sensed situation or usage |
CN106028930B (en) | 2014-02-21 | 2021-10-22 | 3D集成公司 | Kit comprising a surgical instrument |
US10149618B1 (en) | 2014-03-12 | 2018-12-11 | The Board Of Regents Of The University Of Texas System | Subdural electrode localization and visualization using parcellated, manipulable cerebral mesh models |
GB2524498A (en) * | 2014-03-24 | 2015-09-30 | Scopis Gmbh | Electromagnetic navigation system for microscopic surgery |
US10070940B2 (en) * | 2014-09-15 | 2018-09-11 | Synaptive Medical (Barbados) Inc. | End effector for a positioning device |
US11504192B2 (en) | 2014-10-30 | 2022-11-22 | Cilag Gmbh International | Method of hub communication with surgical instrument systems |
WO2016092408A1 (en) * | 2014-12-09 | 2016-06-16 | Koninklijke Philips N.V. | Feedback for multi-modality auto-registration |
US9924871B2 (en) * | 2015-03-05 | 2018-03-27 | Synaptive Medical (Barbados) Inc. | Optical coherence tomography system including a planarizing transparent material |
US9706927B2 (en) * | 2015-04-13 | 2017-07-18 | Atoptix, Llc | Mobile reflectance optical spectroscopy device and process of using and assembling mobile reflectance optical spectroscopy device |
EP3145420B8 (en) * | 2015-06-05 | 2020-12-30 | Brain Navi Biotechnology Co., Ltd. | Intraoperative tracking method |
GB2557474B (en) * | 2015-06-19 | 2021-07-14 | Synaptive Medical Inc | A medical imaging system for determining a scan orientation |
US11020144B2 (en) * | 2015-07-21 | 2021-06-01 | 3Dintegrated Aps | Minimally invasive surgery system |
EP3145419B1 (en) | 2015-07-21 | 2019-11-27 | 3dintegrated ApS | Cannula assembly kit, trocar assembly kit and minimally invasive surgery system |
DK178899B1 (en) | 2015-10-09 | 2017-05-08 | 3Dintegrated Aps | A depiction system |
US10058393B2 (en) | 2015-10-21 | 2018-08-28 | P Tech, Llc | Systems and methods for navigation and visualization |
EP3165153A1 (en) * | 2015-11-05 | 2017-05-10 | Deutsches Krebsforschungszentrum Stiftung des Öffentlichen Rechts | System for fluorescence aided surgery |
KR20180071328A (en) * | 2015-11-16 | 2018-06-27 | 씽크 써지컬, 인크. | A method for verifying matching of traced bones |
US10162086B2 (en) | 2016-03-07 | 2018-12-25 | Microsoft Technology Licensing, Llc | Imaging through highly diffusive media with wavefront shaping |
WO2017154005A1 (en) * | 2016-03-10 | 2017-09-14 | Biop - Medical Ltd | Device for diagnosing a tissue |
CN105943161A (en) * | 2016-06-04 | 2016-09-21 | 深圳市前海康启源科技有限公司 | Surgical navigation system and method based on medical robot |
WO2018005750A1 (en) * | 2016-07-01 | 2018-01-04 | Intuitive Surgical Operations, Inc. | Computer-assisted medical systems and methods |
WO2018030969A1 (en) * | 2016-08-12 | 2018-02-15 | Karacor Adem Can | Lens and lighting device attachable to communication devices for use in dental imaging applications |
US10993771B2 (en) * | 2016-09-12 | 2021-05-04 | Synaptive Medical Inc. | Trackable apparatuses and methods |
JP6885564B2 (en) | 2016-11-22 | 2021-06-16 | ユニバーシダッド デ ラス パルマス デ グラン カナリア | Non-invasive detection methods and hyperspectral imaging devices for tumors and / or healthy tissues |
CN107569213A (en) * | 2017-09-04 | 2018-01-12 | 上海市第十人民医院 | A kind of medical auxiliary examination equipment of visible ray and near-infrared dual-waveband imaging |
US11373750B2 (en) | 2017-09-08 | 2022-06-28 | The General Hospital Corporation | Systems and methods for brain hemorrhage classification in medical images using an artificial intelligence network |
CN111246818B (en) * | 2017-09-22 | 2024-02-23 | 直观外科手术操作公司 | Enhancing visible differences between different tissues in computer-assisted teleoperated surgery |
CA2981726C (en) * | 2017-10-06 | 2018-12-04 | Synaptive Medical (Barbados) Inc. | Surgical optical zoom system |
US11564756B2 (en) | 2017-10-30 | 2023-01-31 | Cilag Gmbh International | Method of hub communication with surgical instrument systems |
US11291510B2 (en) | 2017-10-30 | 2022-04-05 | Cilag Gmbh International | Method of hub communication with surgical instrument systems |
US11510741B2 (en) | 2017-10-30 | 2022-11-29 | Cilag Gmbh International | Method for producing a surgical instrument comprising a smart electrical system |
US11793537B2 (en) | 2017-10-30 | 2023-10-24 | Cilag Gmbh International | Surgical instrument comprising an adaptive electrical system |
US11911045B2 (en) | 2017-10-30 | 2024-02-27 | Cllag GmbH International | Method for operating a powered articulating multi-clip applier |
US11311342B2 (en) | 2017-10-30 | 2022-04-26 | Cilag Gmbh International | Method for communicating with surgical instrument systems |
US11406390B2 (en) | 2017-10-30 | 2022-08-09 | Cilag Gmbh International | Clip applier comprising interchangeable clip reloads |
US11229436B2 (en) | 2017-10-30 | 2022-01-25 | Cilag Gmbh International | Surgical system comprising a surgical tool and a surgical hub |
US11801098B2 (en) | 2017-10-30 | 2023-10-31 | Cilag Gmbh International | Method of hub communication with surgical instrument systems |
US11317919B2 (en) | 2017-10-30 | 2022-05-03 | Cilag Gmbh International | Clip applier comprising a clip crimping system |
US11786245B2 (en) | 2017-12-28 | 2023-10-17 | Cilag Gmbh International | Surgical systems with prioritized data transmission capabilities |
US11969142B2 (en) | 2017-12-28 | 2024-04-30 | Cilag Gmbh International | Method of compressing tissue within a stapling device and simultaneously displaying the location of the tissue within the jaws |
US11559308B2 (en) | 2017-12-28 | 2023-01-24 | Cilag Gmbh International | Method for smart energy device infrastructure |
US20190206569A1 (en) | 2017-12-28 | 2019-07-04 | Ethicon Llc | Method of cloud based data analytics for use with the hub |
US11432885B2 (en) | 2017-12-28 | 2022-09-06 | Cilag Gmbh International | Sensing arrangements for robot-assisted surgical platforms |
US11278281B2 (en) | 2017-12-28 | 2022-03-22 | Cilag Gmbh International | Interactive surgical system |
US11540855B2 (en) | 2017-12-28 | 2023-01-03 | Cilag Gmbh International | Controlling activation of an ultrasonic surgical instrument according to the presence of tissue |
US11266468B2 (en) * | 2017-12-28 | 2022-03-08 | Cilag Gmbh International | Cooperative utilization of data derived from secondary sources by intelligent surgical hubs |
US11529187B2 (en) | 2017-12-28 | 2022-12-20 | Cilag Gmbh International | Surgical evacuation sensor arrangements |
US11026751B2 (en) | 2017-12-28 | 2021-06-08 | Cilag Gmbh International | Display of alignment of staple cartridge to prior linear staple line |
US11234756B2 (en) | 2017-12-28 | 2022-02-01 | Cilag Gmbh International | Powered surgical tool with predefined adjustable control algorithm for controlling end effector parameter |
US10849697B2 (en) | 2017-12-28 | 2020-12-01 | Ethicon Llc | Cloud interface for coupled surgical devices |
US11076921B2 (en) | 2017-12-28 | 2021-08-03 | Cilag Gmbh International | Adaptive control program updates for surgical hubs |
US11419667B2 (en) | 2017-12-28 | 2022-08-23 | Cilag Gmbh International | Ultrasonic energy device which varies pressure applied by clamp arm to provide threshold control pressure at a cut progression location |
US11903601B2 (en) | 2017-12-28 | 2024-02-20 | Cilag Gmbh International | Surgical instrument comprising a plurality of drive systems |
US11832899B2 (en) | 2017-12-28 | 2023-12-05 | Cilag Gmbh International | Surgical systems with autonomously adjustable control programs |
US11202570B2 (en) | 2017-12-28 | 2021-12-21 | Cilag Gmbh International | Communication hub and storage device for storing parameters and status of a surgical device to be shared with cloud based analytics systems |
US10892899B2 (en) | 2017-12-28 | 2021-01-12 | Ethicon Llc | Self describing data packets generated at an issuing instrument |
US11424027B2 (en) | 2017-12-28 | 2022-08-23 | Cilag Gmbh International | Method for operating surgical instrument systems |
US11771487B2 (en) | 2017-12-28 | 2023-10-03 | Cilag Gmbh International | Mechanisms for controlling different electromechanical systems of an electrosurgical instrument |
US11179175B2 (en) | 2017-12-28 | 2021-11-23 | Cilag Gmbh International | Controlling an ultrasonic surgical instrument according to tissue location |
US11857152B2 (en) | 2017-12-28 | 2024-01-02 | Cilag Gmbh International | Surgical hub spatial awareness to determine devices in operating theater |
US10966791B2 (en) | 2017-12-28 | 2021-04-06 | Ethicon Llc | Cloud-based medical analytics for medical facility segmented individualization of instrument function |
US10943454B2 (en) | 2017-12-28 | 2021-03-09 | Ethicon Llc | Detection and escalation of security responses of surgical instruments to increasing severity threats |
US11896322B2 (en) | 2017-12-28 | 2024-02-13 | Cilag Gmbh International | Sensing the patient position and contact utilizing the mono-polar return pad electrode to provide situational awareness to the hub |
US11744604B2 (en) | 2017-12-28 | 2023-09-05 | Cilag Gmbh International | Surgical instrument with a hardware-only control circuit |
US11818052B2 (en) | 2017-12-28 | 2023-11-14 | Cilag Gmbh International | Surgical network determination of prioritization of communication, interaction, or processing based on system or device needs |
US11633237B2 (en) | 2017-12-28 | 2023-04-25 | Cilag Gmbh International | Usage and technique analysis of surgeon / staff performance against a baseline to optimize device utilization and performance for both current and future procedures |
US11056244B2 (en) | 2017-12-28 | 2021-07-06 | Cilag Gmbh International | Automated data scaling, alignment, and organizing based on predefined parameters within surgical networks |
US10944728B2 (en) | 2017-12-28 | 2021-03-09 | Ethicon Llc | Interactive surgical systems with encrypted communication capabilities |
US11464535B2 (en) | 2017-12-28 | 2022-10-11 | Cilag Gmbh International | Detection of end effector emersion in liquid |
US11109866B2 (en) | 2017-12-28 | 2021-09-07 | Cilag Gmbh International | Method for circular stapler control algorithm adjustment based on situational awareness |
US11284936B2 (en) | 2017-12-28 | 2022-03-29 | Cilag Gmbh International | Surgical instrument having a flexible electrode |
US11612444B2 (en) | 2017-12-28 | 2023-03-28 | Cilag Gmbh International | Adjustment of a surgical device function based on situational awareness |
US11602393B2 (en) | 2017-12-28 | 2023-03-14 | Cilag Gmbh International | Surgical evacuation sensing and generator control |
US11160605B2 (en) | 2017-12-28 | 2021-11-02 | Cilag Gmbh International | Surgical evacuation sensing and motor control |
US11069012B2 (en) | 2017-12-28 | 2021-07-20 | Cilag Gmbh International | Interactive surgical systems with condition handling of devices and data capabilities |
US11013563B2 (en) | 2017-12-28 | 2021-05-25 | Ethicon Llc | Drive arrangements for robot-assisted surgical platforms |
US11659023B2 (en) | 2017-12-28 | 2023-05-23 | Cilag Gmbh International | Method of hub communication |
US11100631B2 (en) | 2017-12-28 | 2021-08-24 | Cilag Gmbh International | Use of laser light and red-green-blue coloration to determine properties of back scattered light |
US20190201087A1 (en) | 2017-12-28 | 2019-07-04 | Ethicon Llc | Smoke evacuation system including a segmented control circuit for interactive surgical platform |
US12096916B2 (en) | 2017-12-28 | 2024-09-24 | Cilag Gmbh International | Method of sensing particulate from smoke evacuated from a patient, adjusting the pump speed based on the sensed information, and communicating the functional parameters of the system to the hub |
US10758310B2 (en) | 2017-12-28 | 2020-09-01 | Ethicon Llc | Wireless pairing of a surgical device with another device within a sterile surgical field based on the usage and situational awareness of devices |
US11589888B2 (en) | 2017-12-28 | 2023-02-28 | Cilag Gmbh International | Method for controlling smart energy devices |
US11253315B2 (en) | 2017-12-28 | 2022-02-22 | Cilag Gmbh International | Increasing radio frequency to create pad-less monopolar loop |
US11896443B2 (en) | 2017-12-28 | 2024-02-13 | Cilag Gmbh International | Control of a surgical system through a surgical barrier |
US11389164B2 (en) | 2017-12-28 | 2022-07-19 | Cilag Gmbh International | Method of using reinforced flexible circuits with multiple sensors to optimize performance of radio frequency devices |
US11419630B2 (en) | 2017-12-28 | 2022-08-23 | Cilag Gmbh International | Surgical system distributed processing |
US11364075B2 (en) | 2017-12-28 | 2022-06-21 | Cilag Gmbh International | Radio frequency energy device for delivering combined electrical signals |
US11166772B2 (en) | 2017-12-28 | 2021-11-09 | Cilag Gmbh International | Surgical hub coordination of control and communication of operating room devices |
US20190201146A1 (en) | 2017-12-28 | 2019-07-04 | Ethicon Llc | Safety systems for smart powered surgical stapling |
US11291495B2 (en) | 2017-12-28 | 2022-04-05 | Cilag Gmbh International | Interruption of energy due to inadvertent capacitive coupling |
US11051876B2 (en) | 2017-12-28 | 2021-07-06 | Cilag Gmbh International | Surgical evacuation flow paths |
US11423007B2 (en) | 2017-12-28 | 2022-08-23 | Cilag Gmbh International | Adjustment of device control programs based on stratified contextual data in addition to the data |
US11571234B2 (en) | 2017-12-28 | 2023-02-07 | Cilag Gmbh International | Temperature control of ultrasonic end effector and control system therefor |
US11937769B2 (en) | 2017-12-28 | 2024-03-26 | Cilag Gmbh International | Method of hub communication, processing, storage and display |
US11317937B2 (en) | 2018-03-08 | 2022-05-03 | Cilag Gmbh International | Determining the state of an ultrasonic end effector |
US11273001B2 (en) | 2017-12-28 | 2022-03-15 | Cilag Gmbh International | Surgical hub and modular device response adjustment based on situational awareness |
US11969216B2 (en) | 2017-12-28 | 2024-04-30 | Cilag Gmbh International | Surgical network recommendations from real time analysis of procedure variables against a baseline highlighting differences from the optimal solution |
US12062442B2 (en) | 2017-12-28 | 2024-08-13 | Cilag Gmbh International | Method for operating surgical instrument systems |
US10987178B2 (en) | 2017-12-28 | 2021-04-27 | Ethicon Llc | Surgical hub control arrangements |
US11786251B2 (en) | 2017-12-28 | 2023-10-17 | Cilag Gmbh International | Method for adaptive control schemes for surgical network control and interaction |
US11324557B2 (en) | 2017-12-28 | 2022-05-10 | Cilag Gmbh International | Surgical instrument with a sensing array |
US11678881B2 (en) | 2017-12-28 | 2023-06-20 | Cilag Gmbh International | Spatial awareness of surgical hubs in operating rooms |
US11559307B2 (en) | 2017-12-28 | 2023-01-24 | Cilag Gmbh International | Method of robotic hub communication, detection, and control |
US11304745B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Surgical evacuation sensing and display |
US11832840B2 (en) | 2017-12-28 | 2023-12-05 | Cilag Gmbh International | Surgical instrument having a flexible circuit |
US11213359B2 (en) | 2017-12-28 | 2022-01-04 | Cilag Gmbh International | Controllers for robot-assisted surgical platforms |
US11666331B2 (en) | 2017-12-28 | 2023-06-06 | Cilag Gmbh International | Systems for detecting proximity of surgical end effector to cancerous tissue |
US10892995B2 (en) | 2017-12-28 | 2021-01-12 | Ethicon Llc | Surgical network determination of prioritization of communication, interaction, or processing based on system or device needs |
US11998193B2 (en) | 2017-12-28 | 2024-06-04 | Cilag Gmbh International | Method for usage of the shroud as an aspect of sensing or controlling a powered surgical device, and a control algorithm to adjust its default operation |
US11147607B2 (en) | 2017-12-28 | 2021-10-19 | Cilag Gmbh International | Bipolar combination device that automatically adjusts pressure based on energy modality |
US11096693B2 (en) | 2017-12-28 | 2021-08-24 | Cilag Gmbh International | Adjustment of staple height of at least one row of staples based on the sensed tissue thickness or force in closing |
US11045591B2 (en) | 2017-12-28 | 2021-06-29 | Cilag Gmbh International | Dual in-series large and small droplet filters |
US11576677B2 (en) | 2017-12-28 | 2023-02-14 | Cilag Gmbh International | Method of hub communication, processing, display, and cloud analytics |
US11464559B2 (en) | 2017-12-28 | 2022-10-11 | Cilag Gmbh International | Estimating state of ultrasonic end effector and control system therefor |
US11410259B2 (en) | 2017-12-28 | 2022-08-09 | Cilag Gmbh International | Adaptive control program updates for surgical devices |
US11304699B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Method for adaptive control schemes for surgical network control and interaction |
US11179208B2 (en) | 2017-12-28 | 2021-11-23 | Cilag Gmbh International | Cloud-based medical analytics for security and authentication trends and reactive measures |
US11308075B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Surgical network, instrument, and cloud responses based on validation of received dataset and authentication of its source and integrity |
US10932872B2 (en) | 2017-12-28 | 2021-03-02 | Ethicon Llc | Cloud-based medical analytics for linking of local usage trends with the resource acquisition behaviors of larger data set |
US11864728B2 (en) | 2017-12-28 | 2024-01-09 | Cilag Gmbh International | Characterization of tissue irregularities through the use of mono-chromatic light refractivity |
US11304720B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Activation of energy devices |
US11376002B2 (en) | 2017-12-28 | 2022-07-05 | Cilag Gmbh International | Surgical instrument cartridge sensor assemblies |
US11257589B2 (en) | 2017-12-28 | 2022-02-22 | Cilag Gmbh International | Real-time analysis of comprehensive cost of all instrumentation used in surgery utilizing data fluidity to track instruments through stocking and in-house processes |
US11132462B2 (en) | 2017-12-28 | 2021-09-28 | Cilag Gmbh International | Data stripping method to interrogate patient records and create anonymized record |
US11311306B2 (en) | 2017-12-28 | 2022-04-26 | Cilag Gmbh International | Surgical systems for detecting end effector tissue distribution irregularities |
US11446052B2 (en) | 2017-12-28 | 2022-09-20 | Cilag Gmbh International | Variation of radio frequency and ultrasonic power level in cooperation with varying clamp arm pressure to achieve predefined heat flux or power applied to tissue |
US11304763B2 (en) | 2017-12-28 | 2022-04-19 | Cilag Gmbh International | Image capturing of the areas outside the abdomen to improve placement and control of a surgical device in use |
CN108013891B (en) * | 2018-01-26 | 2023-08-04 | 中国工程物理研究院激光聚变研究中心 | X-ray diagnostic device |
US10517158B2 (en) * | 2018-02-09 | 2019-12-24 | Gentex Corporation | Systems and methods for detection and illumination of regions of interest |
US11399858B2 (en) | 2018-03-08 | 2022-08-02 | Cilag Gmbh International | Application of smart blade technology |
US11678927B2 (en) | 2018-03-08 | 2023-06-20 | Cilag Gmbh International | Detection of large vessels during parenchymal dissection using a smart blade |
US11259830B2 (en) | 2018-03-08 | 2022-03-01 | Cilag Gmbh International | Methods for controlling temperature in ultrasonic device |
US11259806B2 (en) | 2018-03-28 | 2022-03-01 | Cilag Gmbh International | Surgical stapling devices with features for blocking advancement of a camming assembly of an incompatible cartridge installed therein |
US11471156B2 (en) | 2018-03-28 | 2022-10-18 | Cilag Gmbh International | Surgical stapling devices with improved rotary driven closure systems |
US11096688B2 (en) | 2018-03-28 | 2021-08-24 | Cilag Gmbh International | Rotary driven firing members with different anvil and channel engagement features |
US11207067B2 (en) | 2018-03-28 | 2021-12-28 | Cilag Gmbh International | Surgical stapling device with separate rotary driven closure and firing systems and firing member that engages both jaws while firing |
US11090047B2 (en) | 2018-03-28 | 2021-08-17 | Cilag Gmbh International | Surgical instrument comprising an adaptive control system |
US11219453B2 (en) | 2018-03-28 | 2022-01-11 | Cilag Gmbh International | Surgical stapling devices with cartridge compatible closure and firing lockout arrangements |
US11278280B2 (en) | 2018-03-28 | 2022-03-22 | Cilag Gmbh International | Surgical instrument comprising a jaw closure lockout |
US11213294B2 (en) | 2018-03-28 | 2022-01-04 | Cilag Gmbh International | Surgical instrument comprising co-operating lockout features |
US10973520B2 (en) | 2018-03-28 | 2021-04-13 | Ethicon Llc | Surgical staple cartridge with firing member driven camming assembly that has an onboard tissue cutting feature |
DE102018206406B3 (en) * | 2018-04-25 | 2019-09-12 | Carl Zeiss Meditec Ag | Microscopy system and method for operating a microscopy system |
WO2019219387A1 (en) | 2018-05-16 | 2019-11-21 | Koninklijke Philips N.V. | Automated tumor identification during surgery using machine-learning |
CN111989023B (en) | 2018-08-31 | 2024-02-02 | Hoya株式会社 | Endoscope system and method for operating same |
CN109528313B (en) * | 2018-10-12 | 2021-04-20 | 合刃科技(武汉)有限公司 | Spectrum three-dimensional positioning and identifying system |
US11744655B2 (en) | 2018-12-04 | 2023-09-05 | Globus Medical, Inc. | Drill guide fixtures, cranial insertion fixtures, and related methods and robotic systems |
US11602402B2 (en) | 2018-12-04 | 2023-03-14 | Globus Medical, Inc. | Drill guide fixtures, cranial insertion fixtures, and related methods and robotic systems |
US11107213B2 (en) * | 2018-12-28 | 2021-08-31 | Biosense Webster (Israel) Ltd. | Correcting medical scans |
KR102235818B1 (en) | 2019-01-31 | 2021-04-02 | 한국기술교육대학교 산학협력단 | Endoscopic trainer |
US11751872B2 (en) | 2019-02-19 | 2023-09-12 | Cilag Gmbh International | Insertable deactivator element for surgical stapler lockouts |
US11317915B2 (en) | 2019-02-19 | 2022-05-03 | Cilag Gmbh International | Universal cartridge based key feature that unlocks multiple lockout arrangements in different surgical staplers |
US11369377B2 (en) | 2019-02-19 | 2022-06-28 | Cilag Gmbh International | Surgical stapling assembly with cartridge based retainer configured to unlock a firing lockout |
US11331100B2 (en) | 2019-02-19 | 2022-05-17 | Cilag Gmbh International | Staple cartridge retainer system with authentication keys |
US11357503B2 (en) | 2019-02-19 | 2022-06-14 | Cilag Gmbh International | Staple cartridge retainers with frangible retention features and methods of using same |
CN110013321A (en) * | 2019-02-20 | 2019-07-16 | 广州乔铁医疗科技有限公司 | Visor outside a kind of robot with OCT function |
CN110013320A (en) * | 2019-02-20 | 2019-07-16 | 广州乔铁医疗科技有限公司 | A kind of laparoscope outer mirror device of Application Optics coherence tomography techniques |
US11382549B2 (en) | 2019-03-22 | 2022-07-12 | Globus Medical, Inc. | System for neuronavigation registration and robotic trajectory guidance, and related methods and devices |
US11546569B2 (en) * | 2019-04-15 | 2023-01-03 | Synaptive Medical Inc. | System and methods for correcting image data of distinct images and generating and stereoscopic three-dimensional images |
KR102232043B1 (en) * | 2019-05-09 | 2021-03-25 | 주식회사 디오 | Method and Apparatus for Creating Model for Separating of Teeth and Alveolar Bone in Image Data |
US11045179B2 (en) | 2019-05-20 | 2021-06-29 | Global Medical Inc | Robot-mounted retractor system |
US11986160B2 (en) * | 2019-06-20 | 2024-05-21 | Cllag GmbH International | Image synchronization without input clock and data transmission clock in a pulsed hyperspectral imaging system |
US11895397B2 (en) | 2019-06-20 | 2024-02-06 | Cilag Gmbh International | Image synchronization without input clock and data transmission clock in a pulsed fluorescence imaging system |
US11172811B2 (en) | 2019-06-20 | 2021-11-16 | Cilag Gmbh International | Image rotation in an endoscopic fluorescence imaging system |
US11892403B2 (en) | 2019-06-20 | 2024-02-06 | Cilag Gmbh International | Image synchronization without input clock and data transmission clock in a pulsed fluorescence imaging system |
US11671691B2 (en) | 2019-06-20 | 2023-06-06 | Cilag Gmbh International | Image rotation in an endoscopic laser mapping imaging system |
US11141052B2 (en) | 2019-06-20 | 2021-10-12 | Cilag Gmbh International | Image rotation in an endoscopic fluorescence imaging system |
US11516387B2 (en) | 2019-06-20 | 2022-11-29 | Cilag Gmbh International | Image synchronization without input clock and data transmission clock in a pulsed hyperspectral, fluorescence, and laser mapping imaging system |
US11134832B2 (en) | 2019-06-20 | 2021-10-05 | Cilag Gmbh International | Image rotation in an endoscopic hyperspectral, fluorescence, and laser mapping imaging system |
USD964564S1 (en) | 2019-06-25 | 2022-09-20 | Cilag Gmbh International | Surgical staple cartridge retainer with a closure system authentication key |
USD952144S1 (en) | 2019-06-25 | 2022-05-17 | Cilag Gmbh International | Surgical staple cartridge retainer with firing system authentication key |
USD950728S1 (en) | 2019-06-25 | 2022-05-03 | Cilag Gmbh International | Surgical staple cartridge |
CN116134298A (en) | 2020-06-03 | 2023-05-16 | 伦敦大学国王学院 | Method and system for joint demosaicing and spectral feature map estimation |
WO2022011538A1 (en) * | 2020-07-14 | 2022-01-20 | Covidien Lp | Systems and methods for positioning access ports |
KR102536369B1 (en) * | 2021-02-26 | 2023-05-26 | 주식회사 인피니트헬스케어 | Artificial intelligence-based gastroscopy diagnosis supporting system and method |
KR102531400B1 (en) * | 2021-03-19 | 2023-05-12 | 주식회사 인피니트헬스케어 | Artificial intelligence-based colonoscopy diagnosis supporting system and method |
WO2023042057A1 (en) * | 2021-09-14 | 2023-03-23 | Atai Labs Private Limited | System and method for detecting and analyzing consumer transactions to provide list of selected objects |
TWI782709B (en) * | 2021-09-16 | 2022-11-01 | 財團法人金屬工業研究發展中心 | Surgical robotic arm control system and surgical robotic arm control method |
WO2023078704A1 (en) * | 2021-11-02 | 2023-05-11 | Sony Group Corporation | Apparatuses, methods and computer program products for imaging a structure at least partially obscured, in a visible wavelength range, by a liquid |
CN115040066B (en) * | 2022-08-12 | 2022-11-08 | 北京心联光电科技有限公司 | Multifunctional fundus scanning method and system |
WO2024086564A1 (en) * | 2022-10-17 | 2024-04-25 | Monogram Orthopaedics Inc. | Markerless tracking with spectral imaging camera(s) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090018437A1 (en) * | 2007-07-12 | 2009-01-15 | Cooke T Derek | Radiographic imaging method and apparatus |
US20100256504A1 (en) * | 2007-09-25 | 2010-10-07 | Perception Raisonnement Action En Medecine | Methods and apparatus for assisting cartilage diagnostic and therapeutic procedures |
US20120077743A1 (en) * | 2004-05-25 | 2012-03-29 | Rueger David C | Methods of treating cartilage defects |
Family Cites Families (54)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5452723A (en) | 1992-07-24 | 1995-09-26 | Massachusetts Institute Of Technology | Calibrated spectrographic imaging |
US5311859A (en) * | 1992-09-11 | 1994-05-17 | Welch Allyn, Inc. | Add-on video camera arrangement for optical laparoscope |
US5315342A (en) | 1992-12-28 | 1994-05-24 | Eastman Kodak Company | Automatic focus and indirect illumination camera system |
JPH07184929A (en) * | 1993-12-27 | 1995-07-25 | Olympus Optical Co Ltd | Surgical instrument |
US5823943A (en) | 1994-08-02 | 1998-10-20 | Olympus Optical Co., Ltd | Light source device for endoscopes |
US5795295A (en) * | 1996-06-25 | 1998-08-18 | Carl Zeiss, Inc. | OCT-assisted surgical microscope with multi-coordinate manipulator |
WO1998038907A1 (en) | 1997-03-06 | 1998-09-11 | Massachusetts Institute Of Technology | Instrument for optically scanning of living tissue |
CA2374040C (en) * | 1999-07-02 | 2010-10-19 | Hypermed Imaging, Inc. | Integrated imaging apparatus |
DE19930408A1 (en) * | 1999-07-02 | 2001-01-04 | Zeiss Carl Fa | An optical coherence computer tomography (OCT) system for surgical support combines pre-operation tissue position data and CT data for use in navigation assistance for the operating surgeon |
JP4469044B2 (en) | 2000-01-07 | 2010-05-26 | 株式会社ニデック | Ophthalmic equipment |
EP1301118B1 (en) | 2000-07-14 | 2006-09-06 | Xillix Technologies Corp. | Compact fluorescence endoscopy video system |
DE10116859C2 (en) | 2001-04-04 | 2003-10-09 | Wolf Gmbh Richard | Tissue diagnostic imaging device |
US6899675B2 (en) | 2002-01-15 | 2005-05-31 | Xillix Technologies Corp. | Fluorescence endoscopy video systems with no moving parts in the camera |
DE102004020663A1 (en) | 2004-04-24 | 2005-11-10 | Carl Zeiss Meditec Ag | Device for lighting organic objects |
US7394053B2 (en) | 2004-09-09 | 2008-07-01 | Beth Israel Deaconess Medical Center, Inc. | Systems and methods for multi-modal imaging having a spatial relationship in three dimensions between first and second image data |
US8556806B2 (en) * | 2004-09-24 | 2013-10-15 | Vivid Medical, Inc. | Wavelength multiplexing endoscope |
US9265523B2 (en) | 2011-10-24 | 2016-02-23 | Nico Corporation | Surgical access system with navigation element and method of using same |
CA2631564A1 (en) * | 2004-11-29 | 2006-06-01 | Hypermed, Inc. | Medical hyperspectral imaging for evaluation of tissue and tumor |
CN101237817B (en) * | 2004-12-28 | 2013-01-23 | 超级医药成像有限公司 | Hyperspectral/multispectral imaging in determination, assessment and monitoring of systemic physiology and shock |
CA2947613C (en) * | 2005-04-04 | 2019-11-05 | Hypermed Imaging, Inc. | Hyperspectral imaging in diabetes and peripheral vascular disease |
EP3505052A1 (en) * | 2005-04-25 | 2019-07-03 | University of Massachusetts | Systems and methods for correcting optical reflectance measurements |
WO2007041383A2 (en) | 2005-09-30 | 2007-04-12 | Purdue Research Foundation | Endoscopic imaging device |
US7945077B2 (en) * | 2005-11-30 | 2011-05-17 | Lawrence Livermore National Security, Llc | Hyperspectral microscope for in vivo imaging of microstructures and cells in tissues |
DE102005058480A1 (en) * | 2005-12-07 | 2007-06-14 | Siemens Ag | Object`s e.g. internal organ, medical image data set position-correct assigning method, involves selecting two sub-areas with respect to objects in each set, and determining local measure for position deviation of both sets in each sub-area |
DE102006010105A1 (en) | 2006-02-28 | 2007-08-30 | Carl Zeiss Meditec Ag | Ophthalmologic device for observation and or documentation of eye, has homogeneous lighting, and lighting equipment consists of light source, homogenization unit and projection device |
JP4344829B2 (en) * | 2006-05-02 | 2009-10-14 | 国立大学法人 筑波大学 | Polarized light receiving image measuring device |
US8644911B1 (en) | 2006-06-30 | 2014-02-04 | Hypermed Imaging, Inc. | OxyVu-1 hyperspectral tissue oxygenation (HTO) measurement system |
US8070682B2 (en) | 2006-07-19 | 2011-12-06 | The University Of Connecticut | Method and apparatus for medical imaging using combined near-infrared optical tomography, fluorescent tomography and ultrasound |
EP2117436A4 (en) * | 2007-03-12 | 2011-03-02 | David Tolkowsky | Devices and methods for performing medical procedures in tree-like luminal structures |
US8982203B2 (en) | 2007-06-06 | 2015-03-17 | Karl Storz Gmbh & Co. Kg | Video system for viewing an object on a body |
WO2008154578A1 (en) * | 2007-06-11 | 2008-12-18 | Board Of Regents, The University Of Texas System | Characterization of a near-infrared laparoscopic hyperspectral imaging system |
JP2009034224A (en) | 2007-07-31 | 2009-02-19 | Olympus Medical Systems Corp | Medical treatment apparatus |
US20100113940A1 (en) * | 2008-01-10 | 2010-05-06 | The Ohio State University Research Foundation | Wound goggles |
US8983580B2 (en) | 2008-01-18 | 2015-03-17 | The Board Of Trustees Of The University Of Illinois | Low-coherence interferometry and optical coherence tomography for image-guided surgical treatment of solid tumors |
WO2010019515A2 (en) * | 2008-08-10 | 2010-02-18 | Board Of Regents, The University Of Texas System | Digital light processing hyperspectral imaging apparatus |
USD612497S1 (en) | 2008-08-29 | 2010-03-23 | Karl Storz Gmbh & Co. Kg | Exoscope |
US8144410B2 (en) | 2008-11-03 | 2012-03-27 | General Electric Company | Multimodal optical imaging |
ES2762203T3 (en) | 2009-03-27 | 2020-05-22 | New View Surgical Inc | Cannula with lighting and integrated camera |
DE102009025662A1 (en) | 2009-06-17 | 2010-12-23 | Karl Storz Gmbh & Co. Kg | Method and apparatus for controlling a multicolor output of an image of a medical object |
WO2011017550A2 (en) | 2009-08-05 | 2011-02-10 | The Johns Hopkins University | Programmable multispectral illumination system for surgery and visualization of light-sensitive tissues |
US9179831B2 (en) * | 2009-11-30 | 2015-11-10 | King Systems Corporation | Visualization instrument |
KR101172745B1 (en) * | 2010-01-29 | 2012-08-14 | 한국전기연구원 | Combined apparatus for detection of multi-spectrum optical imaging coming out of organic body and light therapy |
US9044142B2 (en) | 2010-03-12 | 2015-06-02 | Carl Zeiss Meditec Ag | Surgical optical systems for detecting brain tumors |
JP4937397B2 (en) * | 2010-10-25 | 2012-05-23 | 富士フイルム株式会社 | Medical image diagnosis support apparatus and method, and program |
EP2680743A4 (en) | 2011-03-02 | 2014-08-13 | Diagnostic Photonics Inc | Hand-held fixed-focus optical probe |
DE102011016138A1 (en) | 2011-03-30 | 2012-10-04 | Karl Storz Gmbh & Co. Kg | Device for fluorescence diagnosis |
US8945149B2 (en) | 2011-04-05 | 2015-02-03 | Gabriel Min Kim | Automated surgical illumination system |
EP2514357B1 (en) | 2011-04-18 | 2014-10-29 | Karl Storz GmbH & Co. KG | Exoscope |
US8891087B2 (en) * | 2011-06-01 | 2014-11-18 | Digital Light Innovations | System and method for hyperspectral imaging |
US8878920B2 (en) | 2011-07-12 | 2014-11-04 | Karl Storz Imaging, Inc. | Method and apparatus for protection from high intensity light |
KR101134770B1 (en) | 2011-07-22 | 2012-04-13 | 양완석 | Portable exoscope with ccd camera for human body opening |
US20130044207A1 (en) | 2011-08-16 | 2013-02-21 | Key Technology, Inc. | Imaging apparatus |
US10722318B2 (en) | 2011-08-24 | 2020-07-28 | Mako Surgical Corp. | Surgical tools for selectively illuminating a surgical volume |
US20140378843A1 (en) | 2012-01-20 | 2014-12-25 | The Trustees Of Dartmouth College | Method And Apparatus For Quantitative Hyperspectral Fluorescence And Reflectance Imaging For Surgical Guidance |
-
2014
- 2014-03-14 JP JP2015561867A patent/JP6527086B2/en active Active
- 2014-03-14 EP EP14765228.3A patent/EP2967349B1/en active Active
- 2014-03-14 WO PCT/CA2014/050268 patent/WO2014139020A1/en active Application Filing
- 2014-03-14 US US14/398,649 patent/US10292771B2/en active Active
- 2014-03-14 MY MYPI2015002365A patent/MY177299A/en unknown
- 2014-03-14 CA CA2894133A patent/CA2894133C/en active Active
- 2014-03-14 BR BR112015023545-0A patent/BR112015023545B1/en active IP Right Grant
- 2014-03-14 AU AU2014231342A patent/AU2014231342B2/en active Active
- 2014-03-14 CA CA2942069A patent/CA2942069C/en active Active
- 2014-03-14 SG SG11201507609UA patent/SG11201507609UA/en unknown
- 2014-03-14 CN CN201480015318.5A patent/CN105263398B/en active Active
- 2014-09-15 AU AU2014386696A patent/AU2014386696B2/en active Active
- 2014-09-15 JP JP2016574309A patent/JP2017508579A/en active Pending
- 2014-09-15 US US15/023,120 patent/US20160235482A1/en not_active Abandoned
- 2014-09-15 KR KR1020167028537A patent/KR102364976B1/en active IP Right Grant
- 2014-09-15 CA CA2940297A patent/CA2940297C/en active Active
-
2016
- 2016-06-28 HK HK16107507.3A patent/HK1219405A1/en not_active IP Right Cessation
-
2019
- 2019-05-17 US US16/414,957 patent/US10786314B2/en active Active
- 2019-07-19 US US16/516,496 patent/US20190388160A1/en not_active Abandoned
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20120077743A1 (en) * | 2004-05-25 | 2012-03-29 | Rueger David C | Methods of treating cartilage defects |
US20090018437A1 (en) * | 2007-07-12 | 2009-01-15 | Cooke T Derek | Radiographic imaging method and apparatus |
US20100256504A1 (en) * | 2007-09-25 | 2010-10-07 | Perception Raisonnement Action En Medecine | Methods and apparatus for assisting cartilage diagnostic and therapeutic procedures |
Non-Patent Citations (5)
Title |
---|
Hemanth et al., "In vivo length patterns of the medial collateral ligament during the stance phase of gait", Knee Surg Sports Traumatol Arthrosc., May 2011 * |
Lange, Thomas, "MODELING PRIOR KNOWLEDGE FOR IMAGE REGISTRATION IN LIVER SURGERY", 2011 * |
Perhavc, Jernej, "Image Registration", January 2007 * |
Rashidifard et al., "The Application of Optical Coherence Tomography in Musculoskeletal Disease", Arthritis, Volume 2013, published online 1/15/2013 * |
Rodrigues et al., "MRI Evaluation of Knee Cartilage", 2010, pp 340-346 * |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10535149B2 (en) * | 2014-06-25 | 2020-01-14 | Koninklijke Philips N.V. | Imaging device for registration of different imaging modalities |
US20170148180A1 (en) * | 2014-06-25 | 2017-05-25 | Koninklijke Philips N.V. | Imaging device for registration of different imaging modalities |
US9737370B2 (en) * | 2014-10-14 | 2017-08-22 | Synaptive Medical (Barbados) Inc. | Patient reference tool |
US20160324583A1 (en) * | 2014-10-14 | 2016-11-10 | Leila KHERADPIR | Patient reference tool |
WO2018049215A1 (en) | 2016-09-09 | 2018-03-15 | Intuitive Surgical Operations, Inc. | Simultaneous white light and hyperspectral light imaging systems |
US11911003B2 (en) | 2016-09-09 | 2024-02-27 | Intuitive Surgical Operations, Inc. | Simultaneous white light and hyperspectral light imaging systems |
EP3509529A4 (en) * | 2016-09-09 | 2020-06-03 | Intuitive Surgical Operations Inc. | Simultaneous white light and hyperspectral light imaging systems |
US20180247153A1 (en) * | 2017-02-28 | 2018-08-30 | Verily Life Sciences Llc | System and method for multiclass classification of images using a programmable light source |
US10806334B2 (en) * | 2017-02-28 | 2020-10-20 | Verily Life Sciences Llc | System and method for multiclass classification of images using a programmable light source |
US11229353B2 (en) | 2017-02-28 | 2022-01-25 | Verily Life Sciences Llc | System and method for multiclass classification of images using a programmable light source |
US20220095905A1 (en) * | 2017-02-28 | 2022-03-31 | Verily Life Sciences Llc | System and method for multiclass classification of images using a programmable light source |
US11699102B2 (en) * | 2017-02-28 | 2023-07-11 | Verily Life Sciences Llc | System and method for multiclass classification of images using a programmable light source |
WO2019081807A1 (en) * | 2017-10-27 | 2019-05-02 | Nokia Technologies Oy | Apparatus, systems and methods for detecting light |
KR20200079279A (en) * | 2017-10-27 | 2020-07-02 | 노키아 테크놀로지스 오와이 | Apparatus, system and method for detecting light |
KR102343626B1 (en) | 2017-10-27 | 2021-12-27 | 노키아 테크놀로지스 오와이 | Apparatus, system and method for detecting light |
US11441889B2 (en) * | 2017-10-27 | 2022-09-13 | Nokia Technologies Oy | Apparatus, systems, and methods for detecting light |
EP3477246A1 (en) * | 2017-10-27 | 2019-05-01 | Nokia Technologies Oy | Apparatus, systems and methods for detecting light |
WO2019152261A1 (en) * | 2018-01-30 | 2019-08-08 | Ingle Manish Eknath | Surgical visualization and recording system |
EP4073748A4 (en) * | 2019-12-13 | 2024-01-17 | ChemImage Corporation | Methods for improved operative surgical report generation using machine learning and devices thereof |
Also Published As
Publication number | Publication date |
---|---|
MY177299A (en) | 2020-09-11 |
AU2014386696A1 (en) | 2016-10-13 |
BR112015023545B1 (en) | 2022-05-10 |
EP2967349B1 (en) | 2021-09-22 |
EP2967349A4 (en) | 2017-03-22 |
KR102364976B1 (en) | 2022-02-17 |
US10292771B2 (en) | 2019-05-21 |
US20150109427A1 (en) | 2015-04-23 |
CA2942069A1 (en) | 2014-09-18 |
CA2942069C (en) | 2022-03-01 |
EP2967349A1 (en) | 2016-01-20 |
AU2014231342B2 (en) | 2018-03-29 |
US10786314B2 (en) | 2020-09-29 |
AU2014231342A1 (en) | 2015-11-05 |
CA2894133C (en) | 2016-11-01 |
WO2014139020A1 (en) | 2014-09-18 |
US20190388160A1 (en) | 2019-12-26 |
CA2940297A1 (en) | 2015-09-17 |
BR112015023545A2 (en) | 2018-05-08 |
JP2017508579A (en) | 2017-03-30 |
SG11201507609UA (en) | 2015-10-29 |
KR20170044610A (en) | 2017-04-25 |
HK1219405A1 (en) | 2017-04-07 |
JP2016515848A (en) | 2016-06-02 |
CN105263398A (en) | 2016-01-20 |
JP6527086B2 (en) | 2019-06-05 |
US20200015908A1 (en) | 2020-01-16 |
AU2014386696B2 (en) | 2020-02-27 |
CA2940297C (en) | 2023-12-05 |
CA2894133A1 (en) | 2014-09-18 |
CN105263398B (en) | 2018-05-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20190388160A1 (en) | Methods and systems for intraoperatively confirming location of tissue structures | |
EP3116376B1 (en) | Methods and systems for intraoperatively confirming location of tissue structures | |
US11857317B2 (en) | Method and apparatus for quantitative and depth resolved hyperspectral fluorescence and reflectance imaging for surgical guidance | |
US20140378843A1 (en) | Method And Apparatus For Quantitative Hyperspectral Fluorescence And Reflectance Imaging For Surgical Guidance | |
WO2015023990A1 (en) | Method and apparatus for quantitative and depth resolved hyperspectral fluorescence and reflectance imaging for surgical guidance | |
CN116528742A (en) | Surgical visualization and particle trend analysis system | |
CA2956626C (en) | System and method using a combined modality optical probe | |
US20230239583A1 (en) | Method and system for joint demosaicking and spectral signature estimation | |
Ebner et al. | Intraoperative hyperspectral label-free imaging: from system design to first-in-patient translation | |
Puustinen et al. | Towards clinical hyperspectral imaging (HSI) standards: initial design for a microneurosurgical HSI database | |
Vandebriel et al. | Integrating hyperspectral imaging in an existing intra-operative environment for detection of intrinsic brain tumors | |
Coole et al. | Multimodal optical imaging with real-time projection of cancer risk and biopsy guidance maps for early oral cancer diagnosis and treatment | |
Bartczak et al. | Spectral video in image-guided microsurgical applications: Integrating imaging technology into the clinical environment and ergonomic considerations |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SYNAPTIVE MEDICAL (BARBADOS) INC., BARBADOS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PIRON, CAMERON ANTHONY;RICHMOND, JOSHUA LEE;MCFADYEN, STEPHEN B.E.;AND OTHERS;SIGNING DATES FROM 20140924 TO 20140930;REEL/FRAME:038029/0432 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |
|
AS | Assignment |
Owner name: SYNAPTIVE MEDICAL INC., CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SYNAPTIVE MEDICAL (BARBADOS) INC.;REEL/FRAME:054251/0337 Effective date: 20200902 |
|
AS | Assignment |
Owner name: SYNAPTIVE MEDICAL INC., CANADA Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE 16/935440 APPLICATION PREVIOUSLY RECORDED ON REEL 054251 FRAME 0337. ASSIGNOR(S) HEREBY CONFIRMS THE ASSIGNMENT;ASSIGNOR:SYNAPTIVE MEDICAL (BARBADOS) INC.;REEL/FRAME:055059/0725 Effective date: 20200902 |
|
AS | Assignment |
Owner name: ESPRESSO CAPITAL LTD., CANADA Free format text: SECURITY INTEREST;ASSIGNOR:SYNAPTIVE MEDICAL INC.;REEL/FRAME:054922/0791 Effective date: 20201223 |