WO2005039391A2 - Systems and methods for intraoperative targetting - Google Patents

Systems and methods for intraoperative targetting Download PDF

Info

Publication number
WO2005039391A2
WO2005039391A2 PCT/US2004/035014 US2004035014W WO2005039391A2 WO 2005039391 A2 WO2005039391 A2 WO 2005039391A2 US 2004035014 W US2004035014 W US 2004035014W WO 2005039391 A2 WO2005039391 A2 WO 2005039391A2
Authority
WO
WIPO (PCT)
Prior art keywords
image
patient
ultrasound
target site
target
Prior art date
Application number
PCT/US2004/035014
Other languages
French (fr)
Other versions
WO2005039391A3 (en
Inventor
Ramin Shahidi
Original Assignee
The Board Of Trustees Of The Leland Stanford Junior University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US10/764,650 external-priority patent/US20050085717A1/en
Priority claimed from US10/764,651 external-priority patent/US20050085718A1/en
Application filed by The Board Of Trustees Of The Leland Stanford Junior University filed Critical The Board Of Trustees Of The Leland Stanford Junior University
Priority to EP20040796074 priority Critical patent/EP1680024A2/en
Priority to US10/576,632 priority patent/US20070276234A1/en
Priority to JP2006536816A priority patent/JP2007508913A/en
Publication of WO2005039391A2 publication Critical patent/WO2005039391A2/en
Publication of WO2005039391A3 publication Critical patent/WO2005039391A3/en

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/46Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
    • A61B8/461Displaying means of special interest
    • A61B8/463Displaying means of special interest characterised by displaying multiple images or images and diagnostic data on one display
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/06Devices, other than using radiation, for detecting or locating foreign bodies ; determining position of probes within or on the body of the patient
    • A61B5/061Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body
    • A61B5/064Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body using markers
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0833Detecting organic movements or changes, e.g. tumours, cysts, swellings involving detecting or locating foreign bodies or organic structures
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0833Detecting organic movements or changes, e.g. tumours, cysts, swellings involving detecting or locating foreign bodies or organic structures
    • A61B8/0841Detecting organic movements or changes, e.g. tumours, cysts, swellings involving detecting or locating foreign bodies or organic structures for locating instruments
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/12Diagnosis using ultrasonic, sonic or infrasonic waves in body cavities or body tracts, e.g. by using catheters
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/42Details of probe positioning or probe attachment to the patient
    • A61B8/4245Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/42Details of probe positioning or probe attachment to the patient
    • A61B8/4245Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient
    • A61B8/4254Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient using sensors mounted on the probe
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/44Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
    • A61B8/4416Constructional features of the ultrasonic, sonic or infrasonic diagnostic device related to combined acquisition of different diagnostic modalities, e.g. combination of ultrasound and X-ray acquisitions
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/48Diagnostic techniques
    • A61B8/483Diagnostic techniques involving the acquisition of a 3D volume of data
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/107Visualisation of planned trajectories or target regions
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • A61B2034/2046Tracking techniques
    • A61B2034/2051Electromagnetic tracking systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • A61B2034/2046Tracking techniques
    • A61B2034/2055Optical tracking systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B2090/364Correlation of different images or relation of image positions in respect to the body
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/37Surgical systems with images on a monitor during operation
    • A61B2090/378Surgical systems with images on a monitor during operation using ultrasound
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/39Markers, e.g. radio-opaque or breast lesions markers
    • A61B2090/3995Multi-modality markers
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/25User interfaces for surgical systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/08Detecting organic movements or changes, e.g. tumours, cysts, swellings
    • A61B8/0808Detecting organic movements or changes, e.g. tumours, cysts, swellings for diagnosis of the brain
    • A61B8/0816Detecting organic movements or changes, e.g. tumours, cysts, swellings for diagnosis of the brain using echo-encephalography
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/361Image-producing devices, e.g. surgical cameras

Definitions

  • endoscopes can display only visible surfaces and it is therefore often difficult to visualize tumors, vessels, and other anatomical structures that lie beneath opaque tissue (e.g., targeting of pancreatic adenocarcinomas via gastro-intestinal endoscopy, or targeting of submucosal lesions to sample peri-intestinal structures such as masses in the liver, or targeting of subluminal lesion in the bronchi).
  • image-guided therapy (IGT) systems have been introduced. These systems complement conventional endoscopy and have been used predominantly in neurological, sinus, and spinal surgery, where bony or marker-based registration can provide adequate target accuracy using pre-operative images (typically 1-3 mm).
  • volumetric surgical navigation has been limited by the lack of the computational power required to produce real-time 3D images.
  • the use of various volumetric imaging modalities has progressed to permit the physician to visualize and quantify the extent of disease in 3D in order to plan and execute treatment.
  • Systems are currently able to provide real-time fusion of pre-operative 3D data with intraoperative 2D data images from video cameras, ultrasound probes, surgical microscopes, and endoscopes. These systems have been used predominantly in neurological, sinus, and spinal surgery, where direct access to the pre-operative data plays a major role in the execution of the surgical task. This is despite the fact that, because of movement and deformation of the tissue during the surgery, these IGT procedures tend to lose their spatial registration with respect to the pre-operatively acquired image.
  • the method of some embodiments of the invention assists a user in guiding a medical instrument to a subsurface target site in a patient.
  • This method generates at least one intraoperative ultrasonic images.
  • the method indicates a target site on the ultrasonic image(s).
  • the method determines 3-D coordinates of the target site in a reference coordinate system.
  • the method (1) tracks the position of the instrument in the reference coordinate system, (2) projects onto a display device a view field as seen from the position with respect to the tool in the reference coordinate system, and (3) projects onto the displayed view field indicia of the target site corresponding to the position.
  • the field of view is a view not only from the position of the instrument but also from a known orientation of the instrument in the reference coordinate system.
  • the user can guide the instrument toward the target site by moving the instrument so that the indicia are placed or held in a given state in the displayed field of view.
  • Figs. 1-2 show exemplary flowcharts of the operation of the system of some embodiments of the invention.
  • Figs. 3-4 shows exemplary user interface displays of the system of some embodiments of the invention.
  • Figs. 5-6 shows exemplary operating set-up arrangements in accordance with one aspect of the system.
  • Figure 1 illustrates a process 100 of some embodiments of the invention. This process guides a medical instrument to a desired position in a patient. As shown in this figure, the process 100 initially acquires (at 105) one or more intraoperative images of the target site.
  • the process 100 registers (at 110) the intraoperative images, the patient target site, and the surgical instruments into a common coordinate system.
  • the patient, the imaging source(s) responsible for the intraoperative images and surgical tool must all be placed in the same frame of reference (in registration). This can be done by a variety of methods, three of which are described below.
  • a wall-mounted tracking device can be used to track the patient, imaging source(s), and the surgical tool (e.g., endoscope).
  • the surgical tool e.g., endoscope
  • only the position of the tool can be tracked. Under such an approach, the tool can be placed in registration with the patient and imaging source by touching the tool point to fiducials on the body and to the positions of the imaging source(s).
  • the device could be registered by tool-to-patient contacts. That is, once the images are made, from known coordinates, it is no longer necessary to further track the position of the image source(s).
  • the patient and image sources are placed in registration by fiducials on the patient and in the images, or alternatively, by placing the imaging device at known coordinates with respect to the patient.
  • the patient and tool are placed in registration by detecting the positions of fiducials with respect to the tool, e.g., by using a detector on the 6 STAN.P0009PCT tool for detecting the positions of the patient fiducials.
  • the patient and an endoscope tool can be placed in registration by imaging the fiducials in the endoscope, and matching the imaged positions with the position of the endoscope.
  • the process 100 tracks (at 115) the position of the surgical instrument with respect to the patient target site.
  • a magnetic tracking system is used to track the endoscope for navigation integration in one implementation.
  • the system provides a magnetic transducer into the working channel at the endoscope tip, positioning the field generator so that the optimal sensing volume encompasses the range of sensor positions.
  • a miniaturized magnetic tracking system with metal insensitivity can be used.
  • the tracking system may be calibrated using a calibration jig.
  • a calibration target is modified from a uniform to a non-uniform grid of points by reverse- mapping the perspective transform, so that the calibration target point density is approximately equal throughout the endoscope image.
  • the calibration jig is waterproofed and designed to operate in a submerged environment. Where appropriate, calibration will be performed while the jig is immersed in a liquid with refractive properties similar to the operating environment.
  • an ultrasound calibration system can be used for accurate reconstruction of volumetric ultrasound data.
  • An optical tracking system is used to measure the position and orientation of a tracking device that will be attached to the ultrasound probe.
  • a spatial calibration of intrinsic and extrinsic parameters of the ultrasound probe is performed. These parameters are used to transform the ultrasound image into the co-ordinate frame of the endoscope 's field of view.
  • a magnetic tracking system is used for the ultrasound probe. Using only one tracking system for both the endoscope and
  • the ultrasound probe reduces obstructions in the environment, and avoids a line-of-sight tracking requirement.
  • tracking of the probe is done using an optical tracking system.
  • the calibration of the 3D probe is done in a manner similar to a 2D ultrasound probe calibration using intensity-based registration. Intensity-based registration is fully automatic and does not require segmentation or feature identification. In the typical 2D case, acquired images are subject to scaling in the video generation and capture process. This transformation and the known position of the tracking ultrasonic calibration device (calibration phantom) are used to determine the relationship between the ultrasound imaging volume and the ultrasound probe's tracking device. Successful calibration requires an unchanged geometry. The calibration phantom will be designed to withstand relocation and handling without deformation.
  • a quick-release clamp attached to the phantom will hold the ultrasound probe during the calibration process.
  • a spatial correlation of the endoscopic video with dynamic ultrasound images is then done.
  • the processing internal to each tracking system, endoscope, and ultrasound machine causes a unique time delay between the real-time input and output of each device.
  • the output data streams are not synchronized and are refreshed at different intervals.
  • the time taken by the navigation system to acquire and process these outputs is stream- dependant. Consequently, motion due to breathing and other actions can combine with these independent latencies to cause real-time display of dynamic device positions different to those when the imaging is actually being acquired.
  • a computer is used to perform the spatial correlation.
  • the computer can handle a larger image volume, allowing for increased size of the physical imaged volume or higher image resolution (up to 512 _ 512 _ 512 instead of 256 _ 256 _ 64).
  • the computer also provides faster 3D reconstruction and merging, and a higher-quality
  • the computer time-stamps and buffers the tracking and data streams, and then interpolating tracked device position and orientation to match the image data timestamps.
  • the ultrasound probe is moved across a step surface in the calibration phantom to create a temporal step function in both the tracking system and image data stream.
  • the relative delay is determined by comparing the timestamps of the observed step function in each data stream.
  • the endoscope latency is determined similarly using the same phantom. In some embodiments, this is done whenever the ultrasound system is reconfigured. The endoscope latency will not need to be recalculated unless the endoscope electronics are changed, however.
  • the process shows (at 120) on a display device one or more images of the patient target site.
  • the process receives (at 125) a user's indication of a spatial feature of the patient target site on the images of the patient target site.
  • the process projects (at 130) an indicia on the images relating the position and orientation of the surgical instruments to the spatial feature of the patient target site.
  • the methodology illustrated in Figure 1 dynamically tracks and targets lesions in motion beyond the visible endoscopic view.
  • the subregion surrounding the target in the ultrasound volume will be stored as a reference, together with the tracked orientation of the volume.
  • a subregion of each successively-acquired ultrasound volume, centered at the target position in the preceding volume, will be re-sampled using the orientation of the reference target subregion.
  • Three-dimensional cross-correlation of the re- sampled subregion with the reference subregion will be used to find the new location of the
  • the designated target point or region is then displayed to the surgeon during a surgical operation, to guide the position and orientation of the tool toward the target site.
  • the target area is displayed to the user by (1) displaying a field representing the patient target area, and (2) using the tracked position of the tool with respect to the patient to superimpose on the field one or more indicia whose position in the displayed field is indicative of the relative position of the tool with respect to the marked target position.
  • the tool is equipped with a laser pointer that directs a laser beam onto the patient to indicate the position and orientation of a trajectory for accessing the target region. The user can follow this trajectory by aligning the tool with the laser-beam.
  • the displayed image is the image seen by the endoscope, and the indicia are displayed on this image.
  • the indicia may indicate target position as the center point of the indicia, e.g., arrows, and tool orientation for reaching the target from that position, by the degree of elongation of arrows, such that the indicia are brought to equal sizes when the tool is properly oriented.
  • the indicia may indicate target position as the center point of the indicia, e.g., arrows, and tool orientation for reaching the target from that position, by the degree of elongation of arrows, such that the indicia are brought to equal sizes when the tool is properly oriented.
  • the indicia may
  • STAN.P0009PCT indicate the surface point for entry and the elongation of the arrows, the tool orientation- trajectory for reaching the target from that surface point.
  • Some embodiments enable surgeons to visualize a field of view of the surgical endoscope overlaid with volumetrically-reconstructed medical images of a localized area of the patient's anatomy. Using this volumetric navigation system, the surgeon visualizes the surgical site via the surgical endoscope, while exploring the inner layers of the patient's anatomy through the three-dimensionally reconstructed pre-operative MRI or CT images. Given the endoscope 's position and orientation, and given the characteristics of the camera, a perspective volume-rendered view matching that of the optical image obtained by the endoscope is rendered.
  • This system allows the surgeon to virtually fly through and around the site of the surgery to visualize alternative approaches and qualitatively determine the best one.
  • the volumetrically reconstructed images are generated using intensity based filtering and direct perspective volume rendering, which removes the need for conventional segmentation of high-contrast images.
  • the real-time 3D-rendered radiographic reconstruction images matched with the intra-operative endoscopic images provide a new capability in minimally-invasive endoscopic surgery. Since hitting vascular structures remains the greatest hazard in endoscopic procedures, this new technology represents a marked improvement over conventional image-guidance systems, which generally display 2D reconstructed images .
  • the user makes a marking on the image corresponding to the target region or site. This marking may be a point, line or area. From this, and by tracking the position of the tool in the patient coordinate system, the system functions to provide the user with visual information indicating the position of the target identified from the ultrasonic image.
  • the navigation system that uses the process 100 of Figure 1 operates in three distinct modes.
  • the first is target identification mode.
  • the imaged ultrasound volume will be displayed to allow the surgeon to locate one or more target regions of interest and mark them for targeting.
  • the system will show an interactive volumetric rendering as well as up to three user positionable orthogonal cross-sectional planes for precise 2D location of the target.
  • the endoscope will be used to set the position and orientation of the frame of reference. Based on these parameters and using the optical characteristics of the endoscope, the system will overlay target navigation data on the endoscope video. This will allow the surgeon to target regions of interest beyond the visual range of the endoscope 's field of view.
  • Displayed data will include the directions of, and distances to, the target regions relative to the endoscope tip, as well as a potential range of error in this data.
  • the third mode will be used to perform the actual interventional procedure (such as biopsy or ablation) once the endoscope is in the correct position.
  • the interactive imaged ultrasound volume and cross-sectional planes will be displayed, with the location of the endoscope and the trajectory through its tip projected onto each of the views.
  • the endoscope needle itself will also be visible in the ultrasound displays.
  • the navigation system allows the interventional tool to be positioned in the center of the lesion without being limited to a single, fixed 2D ultrasound plane emanating from the endoscope tip.
  • a magnetic sensor will need to be removed from the working channel in order to perform the biopsy, and the navigation display will use the stored position observed immediately prior to its removal.
  • a sensor is integrated into the needle assembly, which will be in place at calibration.
  • the navigation system provides real-time data on the position and orientation of the endoscope, and the ultrasound system provides the dynamic image data.
  • the tip position data is used to calculate the location of the endoscope tip in the image volume, and the probe orientation data will be used to determine the rendering camera position and orientation. Surgeon feedback will be used to improve and refine the navigation system. Procedure durations and outcomes will be compared to those of the conventional biopsy procedure, performed on the phantom without navigation and image-enhanced endoscopy assistance.
  • some embodiments store the subregion surrounding the target in the ultrasound volume as a reference, together with the tracked orientation of the volume.
  • FIG. 2 illustrates a process 200 of some embodiments of the invention. Like the process 100 of Figure 1, the process 200 guides a medical instrument to a desired position in a patient. As shown Figure 2, the process 200 initially acquires (at 205) one or more 2D or
  • the process tracks (at 210) the position of a surgical instrument with respect to the patient target site.
  • the process registers (at 215) the intraoperative images of the patient site, the patient target site, and the surgical instrument into a common 3D reference coordinate system.
  • the process renders (at 220) the image of the patient target site on a display
  • the process also specifies (at 225) a spatial feature (shape and position) of the patient target site on the image.
  • the process correlates (at 230) the position and orientation of the surgical instrument with respect to the target feature.
  • the process projects (at 235) an indicia (e.g., a three-dimensional shape, points and/or lines) on the intraoperative image relating the position and orientation of the surgical instrument to the target spatial feature.
  • Figures 3 and 4 illustrate exemplary user interfaces for the imaging systems that use the processes illustrated in Figures 1 and 2.
  • Figure 3 shows an exemplary user interface (UI) for ultrasound-enhanced endoscopy.
  • the left panel shows the endoscopic view with a superimposed targeting vector and a distance measurement.
  • the right panels show reformatted cross-sectional planes through the acquired 3D ultrasound volume.
  • Figure 4 shows another UI for ultrasound-enhanced endoscopy.
  • the left panel shows the endoscopic view with virtual tool tracking and visualization and vascularure acquired through Doppler imaging.
  • the lower right panel shows volume-rendered 3D ultrasound.
  • the UIs of Figures 3 and 4 support interactive rendering of the ultrasound data to allow a user to locate and mark the desired region of interest in the ultrasound image volume.
  • the UIs allow the user to locate and mark target regions of interest. Hitting vascular structures is a serious hazard in endoscopic procedures. Visualization of the vasculature behind the surface tissue in the endoscopic view would assist in avoiding the vascular structures (anti-targeting).
  • Figures 5 and 6 respectively illustrate exemplary surgical arrangements according to some embodiments of the invention. These systems, can: • track 500+ mm flexible endoscopes with an accuracy of 1.8 mm in position and 1 ° in orientation • acquire external 3D ultrasound images and process them for navigation in near real-time
  • STAN.P0009PCT allow dynamic target identification on any reformatted 3D ultrasound cross- sectional plane view. • optionally overlay dynamic Doppler ultrasound data, rendered using intensity based opacity filters, on the endoscopic view. • provide an overall coarse target accuracy of 10 mm, with a refined target accuracy of 5mm during breath-holds.
  • a video source 500 e.g., microscopic or camcorder
  • An intra-operative imaging system 502 e.g., an ultrasonic system captures an intra-operative imaging data stream 103. The information is displayed on an ultrasonic display 504.
  • a trackable intra-operative imaging probe 505 is also deployed in one or more trackable surgical tools 506.
  • Other tools include a trackable endoscope 507 or any intraoperative video source.
  • the tracking device 508 has tracking wires 509 that communicate a tracking data stream 510.
  • a navigation system 511 with a navigation interface 512 is provided to allow the user to work with an intra-operative video image 513 (perspective view). In the absence of video source this could be blank.
  • Primary targeting markers 514 pointing to a target outside the field of view
  • secondary targeting markers 515 pointing to a target inside the field of view
  • FIG. 6 shows another exemplary surgical set-up.
  • FIG. 6 shows an exemplary surgical set-up.
  • FIG 6 several infrared vision cameras capture patient images.
  • An ultrasonic probe positions an ultra-sound sensor in the
  • Surgical tools such as an endoscope are then positioned in the patient.
  • the infrared vision cameras report the position of the sensors to a computer, which in turn forwards the collected information to a workstation that generates a 3D image reconstruction.
  • the workstation also registers, manipulates the data and visualizes the patient data on a screen.
  • the workstation also receives data from an ultrasound machine that captures 2D images of the patient. Since the geometry of a flexible endoscope in use changes continually, the field of view at the endoscope tip is not directly dependent on the position of a tracking device attached to some other part of the endoscope.
  • Accurate volume reconstruction from ultrasound images requires precise estimation of six extrinsic parameters (position and orientation) and any required intrinsic parameters such as scale.
  • the calibration procedure should be not only accurate but also simple and quick, since it should be performed whenever the tracking sensor is mounted on the ultrasound probe or any of the relevant ultrasound imaging parameters, such as imaging depth or frequency of operation, are modified.
  • An optical tracking system is used to measure the position and orientation of a tracking device that will be attached to the ultrasound probe.
  • 17 STAN.P0009PCT acquired images are subject to scaling in the video generation and capture process. Since video output is not used, but the volumetric ultrasound data is accessed directly, this will not be an issue.
  • the intrinsic parameters of the 3D probe which will have been calibrated by the manufacturer, will be unmodified.
  • a 200 _ 200 _ 200 mm phantom of tissue-mimicking material is used with an integrated CT-visible tracking device. Distributed along all three dimensions within the phantom will be cylinders and cubes, 20 mm in diameter and containing CT contrast material with modified acoustic impedance.
  • the phantom will be imaged using the ultrasound probe; the transformation between the ultrasound volume and a previously acquired, reference CT volumetric image will be computed using intensity-based rigid registration (which requires the intensities of the two images to be similar in structure, but not in value). This transformation and the known position of the phantom's tracking device will be used to determine the relationship between the ultrasound imaging volume and the ultrasound probe's tracking device.
  • Successful calibration requires an unchanged geometry.
  • the phantom will be designed to withstand relocation and handling without deformation. A quick-release clamp attached to the phantom will hold the ultrasound probe during the calibration process.
  • an interface supports interactive rendering of the ultrasound data.
  • An interactive navigation system requires a way for the user to locate and mark target regions of interest. Respiration and other movements will cause the original location of any target to shift. If targets are not dynamically tracked, navigation information will degrade over time. The visibility of regular biopsy needles under ultrasound is poor and hitting vascular structures is a serious hazard in endoscopic procedures. Visualization of the vasculature behind the surface tissue in the endoscopic view would assist in avoiding it (anti-targeting), but
  • the navigation system operates in three distinct modes.
  • the first is target identification mode.
  • the imaged ultrasound volume will be displayed to allow the surgeon to locate one or more target regions of interest and mark them for targeting.
  • the system will show an interactive volumetric rendering as well as up to three user positionable orthogonal cross-sectional planes for precise 2D location of the target.
  • the endoscope will be used to set the position and orientation of the frame of reference. Based on these parameters and using the optical characteristics of the endoscope, the system will overlay target navigation data on the endoscope video. This will allow the surgeon to target regions of interest beyond the visual range of the endoscope 's field of view.
  • Displayed data will include the directions of, and distances to, the target regions relative to the endoscope tip, as well as a potential range of error in this data.
  • the final mode will be used to perform the actual biopsy once the endoscope is in the correct position.
  • the interactive imaged ultrasound volume and cross-sectional planes will be displayed, with the location of the endoscope and the trajectory through its tip projected onto each of the views.
  • the endoscope needle itself will also be visible in the ultrasound displays. This will help to position the biopsy needle in the center of the lesion without being limited to a single, fixed 2D ultrasound plane emanating from the endoscope tip, as is currently the case.
  • the magnetic sensor will need to be removed from the working channel in order to perform the biopsy, and the navigation display will use the stored position observed immediately before its removal.
  • a sensor will be integrated into the needle assembly, which will be in place at calibration.
  • the dynamic ultrasound data may be rendered in real time using intensity-based opacity filters, making nonvascular structures transparent. This effectively isolates the vascular structure without requiring computationally-demanding deformable geometric models for segmentation, thus being able to follow movement and deformation in real time. If the lag is significant, navigation accuracy will be degraded when the target moves. Where optimal accuracy is required, such as when the actual biopsy is perfonned, a brief motionless breath-hold may be required. Lens distortion compensation is performed for the data display in real time, so that the superimposed navigation display maps accurately to the underlying endoscope video. A new ultrasound volume will replace the next most recent volume in its entirety, much as it does on the display of the ultrasound machine itself, although possibly at a different spatial location. This avoids many problematic areas such as misleading old data, data expiration, unbounded imaging volumes, and locking rendering data. Instead, a simple ping-pong buffer pair may be used; one may be used for navigation and display while the
  • CT or MR scans or both will be performed.
  • high resolution contrast enhanced MR images and MR angiograms would be obtained.
  • the image data would be transferred to the
  • Chroma key techniques will automatically identify the flashing markers, enabling automatic and continuous registration and overlay of the patient's physical anatomy with the 3D image data sets.
  • Chroma key is a video special effect technique that allows unique detection of flashing objects with known frequencies in 3D space (e.g., flashing light emitting diodes attached to the patient's head).
  • Diode markers can also be added to conventional ultrasound probes and surgical tools (e.g., probes, scalpels) for their tracking in the stereotactic space.
  • the markers are automatically recognized and overlaid on the display of the registered 3D image.
  • a triangle formation of three markers on the ultrasound probe allows the tracking of its movement as it scans the surgical site, thus providing the system with volumetric ultrasound images.
  • Intra-operatively acquired 3D ultrasound images are then being fused with the pre-operative CT or MR imagery using anatomical features visible to both modalities, such as vascular structures and the lesion.
  • an extrapolated line is then being fused with the pre-operative CT or MR imagery using anatomical features visible to both modalities, such as vascular structures and the lesion.
  • STAN.P0009PCT extending from the displayed image of a surgical device, indicates the trajectory of the planned approach.
  • Moving the tool automatically leads to a change in the displayed potential trajectory, and the location of the L.E.D. on the tool enables determination of the precise depth and location of the tool, thereby enabling precise determination of the depth and location of the operative site. Therefore, this system not only simplifies the planning of a minimally invasion approach to a direct and interactive task, but it is also more precise than the conventional systems due to its intra-operative image updates and registration using ultrasound imagery.
  • the information provided by the video camera-based object recognition system, the laser targeting system will further aid in localization of the surgical site, thus increasing the registration and image overlay performance and accuracy by localizing the area for which re-registration is needed.
  • This information allows the workstation to automatically display the real time 3D image of the operative field in context, oriented to (and if desired overlaid with) the approach.
  • the 3D reformatting uses volume display techniques and allows instantaneous variation of transparency. With this technique, deep as well as superficial structures, can be seen in context, thereby considerably enhancing intraoperative guidance.
  • the system software has two aspects, one dealing with enhancements to the user interface, and the second focusing on algorithms for image manipulation and registration. These algorithms consist of the means for image segmentation, volumetric visualization, image fusion and image overlay.
  • the system provides: i) A "user friendly" interface, to facilitate usage in the operating room. ii) Interactive image analysis and manipulation routines (e.g. arbitrary cuts, image segmentation, image magnification and transformation) on the workstation system.
  • STAN.P0009PCT increased chance for the complete resection of the abnormality.
  • These systems allow accurate soft-tissue navigation.
  • the systems also provide effective calibration and correlation of intraoperative volumetric imaging data with video endoscopy images. Some of these systems acquire external 3D ultrasound images and process them for navigation in near real-time.
  • These system allow dynamic target identification on any reformatted 3D ultrasound cross-sectional plane.
  • the system can automatically track the movement of the target as tissue moves or deforms during the procedure.
  • These systems can dynamically map the target location onto the endoscopic view in form of a direction vector and display quantifiable data such as distance to target.
  • the systems can provide targeting information on the dynamic 3D ultrasound view.
  • the systems can virtually visualize the position and orientation of tracked surgical tools in the ultrasound view, and optionally also in the endoscopic view. These systems also can overlay dynamic Doppler ultrasound data, rendered using intensity based opacity filters, on the endoscopic view.
  • the invention has been described in terms of specific examples, which are illustrative only and are not to be construed as limiting.
  • the invention may be implemented in digital electronic circuitry or in computer hardware, firmware, software, or in combinations of them.
  • Apparatus of the invention may be implemented in a computer program product tangibly embodied in a machine-readable storage device for execution by a computer processor; and method steps of the invention may be performed by a computer processor executing a program to perform functions of the invention by operating on input data and generating output.
  • Suitable processors include, by way of example, both general and special purpose microprocessors.
  • Storage devices suitable for tangibly embodying computer program instructions include all forms of non-volatile memory including, but not limited to: semiconductor memory devices such as EPROM, EEPROM, and flash devices; magnetic disks (fixed, floppy, and removable); other magnetic media such as tape; optical media such

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Surgery (AREA)
  • Biomedical Technology (AREA)
  • Veterinary Medicine (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Medical Informatics (AREA)
  • Molecular Biology (AREA)
  • Animal Behavior & Ethology (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Pathology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Physics & Mathematics (AREA)
  • Biophysics (AREA)
  • Radiology & Medical Imaging (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Robotics (AREA)
  • Ultra Sonic Daignosis Equipment (AREA)
  • Magnetic Resonance Imaging Apparatus (AREA)
  • Surgical Instruments (AREA)
  • Endoscopes (AREA)

Abstract

Systems and methods are disclosed for assisting a user in guiding a medical instrument to a subsurface target site in a patient by indicating a spatial feature of a patient target site on an intraoperative image (e.g., endoscopie image), determining 3-D coordinates of the patient target site spatial feature in a reference coordinate system using the spatial feature of the target site indicated on the intraoperative image (e.g., ultrasound image, determining a position of the instrument in the reference coordinate system, projecting onto a display device a view field from a predetermined position relative to the instrument in the reference coordinate system, and projecting onto the view field an indicia of the spatial feature of the target site corresponding to the predetermined position.

Description

SYSTEMS AND METHODS FOR INTRAOPERATIVE TARGETING BACKGROUND In recent years, the medical community has been increasingly focused on minimizing the invasiveness of surgical procedures. Advances in imaging technology and instrumentation have enabled procedures using minimally-invasive surgery with very small incisions. Growth in this category is being driven by a reduction in morbidity relative to traditional open procedures, because the smaller incisions minimize damage to healthy tissue, reduce patient pain, and speed patient recovery. The introduction of miniature CCD cameras and their associated micro-electronics has broadened the application of endoscopy from an occasional biopsy to full minimally-invasive surgical ablation and aspiration. Minimally-invasive endoscopic surgery Offers advantages of a reduced likelihood of intraoperative and post-operative complications, less pain, and faster patient recovery. However, the small field of view, the lack of orientation cues, and the presence of blood and obscuring tissues combine to make video endoscopic procedures in general disorienting and challenging to perform. Modern volumetric surgical navigation techniques have promised better exposure and orientation for minimally-invasive procedures, but the effective use of current surgical navigation techniques for soft tissue endoscopy is still hampered by two difficulties: (1) accurately tracking all six degrees of freedom (DOF) on a flexible endoscope within the body, and (2) compensating for tissue deformations and. target movements during an interventional procedure. To illustrate, when using an endoscope, the surgeon's vision is limited to the camera's narrow field of view and the lens is often obstructed by blood or fog, resulting in the surgeon suffering a loss of orientation. Moreover, endoscopes can display only visible surfaces and it is therefore often difficult to visualize tumors, vessels, and other anatomical structures that lie beneath opaque tissue (e.g., targeting of pancreatic adenocarcinomas via gastro-intestinal endoscopy, or targeting of submucosal lesions to sample peri-intestinal structures such as masses in the liver, or targeting of subluminal lesion in the bronchi). Recently, image-guided therapy (IGT) systems have been introduced. These systems complement conventional endoscopy and have been used predominantly in neurological, sinus, and spinal surgery, where bony or marker-based registration can provide adequate target accuracy using pre-operative images (typically 1-3 mm). While IGT enhances the surgeon's ability to direct instruments and target specific anatomical structures, in soft tissue these systems lack sufficient targeting accuracy due to intra-operative tissue movement and deformation. In addition, since an endoscope provides a video representation of a 3D environment, it is difficult to correlate the conventional, purely 2D IGT images with the endoscope video. Correlation of information obtained from intra-operative 3D ultrasonic imaging with video endoscopy can significantly improve the accuracy of localization and targeting in minimally-invasive IGT procedures. Until the mid 1990's, the most common use of image guidance was for stereotactic biopsies, in which a surgical trajectory device and a frame of reference were used. Traditional frame-based methods of stereotaxis defined the intracranial anatomy with reference to a set of fiducial markers, which were attached to a frame that was screwed into the patient's skull. These fiducials were measured on pre-operative tomographic (MRI or CT) images. A trajectory-enforcement device was placed on top of the frame of reference and used to guide the biopsy tool to the target lesion, based on prior calculations obtained from pre- operative data. The use of a mechanical frame allowed for high localization accuracy, but caused patient discomfort, limited surgical flexibility, and did not allow the surgeon to visualize the approach of the biopsy tool to the lesion. There has been a gradual emergence of image guided techniques that eliminate the need for the frame altogether. The first frameless stereotactic system used an articulated
2 STAN.P0009PCT robotic arm to register pre-operative imaging with the patient's anatomy in the operating room. This was followed by the use of acoustic devices for tracking instruments in the operating environment. The acoustic devices eventually were superceded by optical tracking systems, which use a camera and infrared diodes (or reflectors) attached to a moving object to accurately track its position and orientation. These systems use markers placed externally on the patient to register pre-operative imaging with the patient's anatomy in the operating room. Such intra-operative navigation techniques use pre-operative CT or MR images to provide localized information during surgery. In addition, all systems enhance intra-operative localization by providing feedback regarding the location of the surgical instruments with respect to 2D preoperative data. Until recently, volumetric surgical navigation has been limited by the lack of the computational power required to produce real-time 3D images. The use of various volumetric imaging modalities has progressed to permit the physician to visualize and quantify the extent of disease in 3D in order to plan and execute treatment. Systems are currently able to provide real-time fusion of pre-operative 3D data with intraoperative 2D data images from video cameras, ultrasound probes, surgical microscopes, and endoscopes. These systems have been used predominantly in neurological, sinus, and spinal surgery, where direct access to the pre-operative data plays a major role in the execution of the surgical task. This is despite the fact that, because of movement and deformation of the tissue during the surgery, these IGT procedures tend to lose their spatial registration with respect to the pre-operatively acquired image.
STAN.P0009PCT SUMMARY The method of some embodiments of the invention assists a user in guiding a medical instrument to a subsurface target site in a patient. This method generates at least one intraoperative ultrasonic images. The method indicates a target site on the ultrasonic image(s). The method determines 3-D coordinates of the target site in a reference coordinate system. The method (1) tracks the position of the instrument in the reference coordinate system, (2) projects onto a display device a view field as seen from the position with respect to the tool in the reference coordinate system, and (3) projects onto the displayed view field indicia of the target site corresponding to the position. In some embodiments, the field of view is a view not only from the position of the instrument but also from a known orientation of the instrument in the reference coordinate system. By observing the indicia, the user can guide the instrument toward the target site by moving the instrument so that the indicia are placed or held in a given state in the displayed field of view.
STAN.P0009PCT BRIEF DESCRIPTION OF THE DRAWINGS The novel features of the invention are set forth in the appended claims. However, for purpose of explanation, several embodiments of the invention are set forth in the following figures. Figs. 1-2 show exemplary flowcharts of the operation of the system of some embodiments of the invention. Figs. 3-4 shows exemplary user interface displays of the system of some embodiments of the invention. Figs. 5-6 shows exemplary operating set-up arrangements in accordance with one aspect of the system.
STAN.P0009PCT DETAILED DESCRIPTION OF THE INVENTION
In the following description, numerous details are set forth for purpose of explanation.
However, one of ordinary skill in the art will realize that the invention may be practiced without the use of these specific details. In other instances, well-known structures and devices are shown in block diagram form in order not to obscure the description of the invention with unnecessary detail. Figure 1 illustrates a process 100 of some embodiments of the invention. This process guides a medical instrument to a desired position in a patient. As shown in this figure, the process 100 initially acquires (at 105) one or more intraoperative images of the target site.
Next, the process 100 registers (at 110) the intraoperative images, the patient target site, and the surgical instruments into a common coordinate system. The patient, the imaging source(s) responsible for the intraoperative images and surgical tool must all be placed in the same frame of reference (in registration). This can be done by a variety of methods, three of which are described below. First, a wall-mounted tracking device can be used to track the patient, imaging source(s), and the surgical tool (e.g., endoscope). Second, only the position of the tool can be tracked. Under such an approach, the tool can be placed in registration with the patient and imaging source by touching the tool point to fiducials on the body and to the positions of the imaging source(s). Thereafter, if the patient moves, the device could be registered by tool-to-patient contacts. That is, once the images are made, from known coordinates, it is no longer necessary to further track the position of the image source(s). Third the patient and image sources are placed in registration by fiducials on the patient and in the images, or alternatively, by placing the imaging device at known coordinates with respect to the patient. The patient and tool are placed in registration by detecting the positions of fiducials with respect to the tool, e.g., by using a detector on the 6 STAN.P0009PCT tool for detecting the positions of the patient fiducials. Alternatively, the patient and an endoscope tool can be placed in registration by imaging the fiducials in the endoscope, and matching the imaged positions with the position of the endoscope. After the registration operation at 110, the process 100 tracks (at 115) the position of the surgical instrument with respect to the patient target site. In some embodiments, a magnetic tracking system is used to track the endoscope for navigation integration in one implementation. The system provides a magnetic transducer into the working channel at the endoscope tip, positioning the field generator so that the optimal sensing volume encompasses the range of sensor positions. In one implementation that provides for six degrees of freedom (6 DOF), a miniaturized magnetic tracking system with metal insensitivity can be used. The tracking system may be calibrated using a calibration jig. A calibration target is modified from a uniform to a non-uniform grid of points by reverse- mapping the perspective transform, so that the calibration target point density is approximately equal throughout the endoscope image. The calibration jig is waterproofed and designed to operate in a submerged environment. Where appropriate, calibration will be performed while the jig is immersed in a liquid with refractive properties similar to the operating environment. In one embodiment, an ultrasound calibration system can be used for accurate reconstruction of volumetric ultrasound data. An optical tracking system is used to measure the position and orientation of a tracking device that will be attached to the ultrasound probe. A spatial calibration of intrinsic and extrinsic parameters of the ultrasound probe is performed. These parameters are used to transform the ultrasound image into the co-ordinate frame of the endoscope 's field of view. In another embodiment, a magnetic tracking system is used for the ultrasound probe. Using only one tracking system for both the endoscope and
STAN.P0009PCT the ultrasound probe reduces obstructions in the environment, and avoids a line-of-sight tracking requirement. In another embodiment, tracking of the probe is done using an optical tracking system. The calibration of the 3D probe is done in a manner similar to a 2D ultrasound probe calibration using intensity-based registration. Intensity-based registration is fully automatic and does not require segmentation or feature identification. In the typical 2D case, acquired images are subject to scaling in the video generation and capture process. This transformation and the known position of the tracking ultrasonic calibration device (calibration phantom) are used to determine the relationship between the ultrasound imaging volume and the ultrasound probe's tracking device. Successful calibration requires an unchanged geometry. The calibration phantom will be designed to withstand relocation and handling without deformation. A quick-release clamp attached to the phantom will hold the ultrasound probe during the calibration process. A spatial correlation of the endoscopic video with dynamic ultrasound images is then done. The processing internal to each tracking system, endoscope, and ultrasound machine causes a unique time delay between the real-time input and output of each device. The output data streams are not synchronized and are refreshed at different intervals. In addition, the time taken by the navigation system to acquire and process these outputs is stream- dependant. Consequently, motion due to breathing and other actions can combine with these independent latencies to cause real-time display of dynamic device positions different to those when the imaging is actually being acquired. In some embodiments, a computer is used to perform the spatial correlation. The computer can handle a larger image volume, allowing for increased size of the physical imaged volume or higher image resolution (up to 512 _ 512 _ 512 instead of 256 _ 256 _ 64). The computer also provides faster 3D reconstruction and merging, and a higher-quality
8 STAN.P0009PCT perspective volume rendering at a higher frame rate. The computer time-stamps and buffers the tracking and data streams, and then interpolating tracked device position and orientation to match the image data timestamps. In determining the required time offset, the ultrasound probe is moved across a step surface in the calibration phantom to create a temporal step function in both the tracking system and image data stream. The relative delay is determined by comparing the timestamps of the observed step function in each data stream. The endoscope latency is determined similarly using the same phantom. In some embodiments, this is done whenever the ultrasound system is reconfigured. The endoscope latency will not need to be recalculated unless the endoscope electronics are changed, however. The patient is imaged through the ultrasound probe, and the endoscope becomes the frame of reference for the surgeon. The important information is contained in the dynamic relationship of the ultrasound data to the endoscope video, which is known through calibration and tracking of both devices. Turning now to Figure 1, the process shows (at 120) on a display device one or more images of the patient target site. Next, the process receives (at 125) a user's indication of a spatial feature of the patient target site on the images of the patient target site. The process then projects (at 130) an indicia on the images relating the position and orientation of the surgical instruments to the spatial feature of the patient target site. The methodology illustrated in Figure 1 dynamically tracks and targets lesions in motion beyond the visible endoscopic view. When a target is identified, the subregion surrounding the target in the ultrasound volume will be stored as a reference, together with the tracked orientation of the volume. A subregion of each successively-acquired ultrasound volume, centered at the target position in the preceding volume, will be re-sampled using the orientation of the reference target subregion. Three-dimensional cross-correlation of the re- sampled subregion with the reference subregion will be used to find the new location of the
9 STAN.P0009PCT target. This dynamic tracking will follow each target over time; if the system is displaying target navigation data, the data will change in real time to follow the updated location of the target relative to the endoscope. Vascular structures return a strong, well differentiated Doppler signal. The dynamic ultrasound data may be rendered in real time using intensity-based opacity filters, making nonvascular structures transparent. This effectively isolates the vascular structure without requiring computationally-demanding deformable geometric models for segmentation, thus the system can follow movements and deformations in real time. The methodology illustrated in Figure 1 allows a user such a surgeon to mark a selected target point or region on intraoperative ultrasonic images (one or more 3-D ultrasound images). The designated target point or region is then displayed to the surgeon during a surgical operation, to guide the position and orientation of the tool toward the target site. In some embodiments, the target area is displayed to the user by (1) displaying a field representing the patient target area, and (2) using the tracked position of the tool with respect to the patient to superimpose on the field one or more indicia whose position in the displayed field is indicative of the relative position of the tool with respect to the marked target position. Also, in some embodiments, the tool is equipped with a laser pointer that directs a laser beam onto the patient to indicate the position and orientation of a trajectory for accessing the target region. The user can follow this trajectory by aligning the tool with the laser-beam. In the embodiments where the tool is an endoscope, the displayed image is the image seen by the endoscope, and the indicia are displayed on this image. The indicia may indicate target position as the center point of the indicia, e.g., arrows, and tool orientation for reaching the target from that position, by the degree of elongation of arrows, such that the indicia are brought to equal sizes when the tool is properly oriented. Alternatively, the indicia may
10 STAN.P0009PCT indicate the surface point for entry and the elongation of the arrows, the tool orientation- trajectory for reaching the target from that surface point. Some embodiments enable surgeons to visualize a field of view of the surgical endoscope overlaid with volumetrically-reconstructed medical images of a localized area of the patient's anatomy. Using this volumetric navigation system, the surgeon visualizes the surgical site via the surgical endoscope, while exploring the inner layers of the patient's anatomy through the three-dimensionally reconstructed pre-operative MRI or CT images. Given the endoscope 's position and orientation, and given the characteristics of the camera, a perspective volume-rendered view matching that of the optical image obtained by the endoscope is rendered. This system allows the surgeon to virtually fly through and around the site of the surgery to visualize alternative approaches and qualitatively determine the best one. The volumetrically reconstructed images are generated using intensity based filtering and direct perspective volume rendering, which removes the need for conventional segmentation of high-contrast images. The real-time 3D-rendered radiographic reconstruction images matched with the intra-operative endoscopic images provide a new capability in minimally-invasive endoscopic surgery. Since hitting vascular structures remains the greatest hazard in endoscopic procedures, this new technology represents a marked improvement over conventional image-guidance systems, which generally display 2D reconstructed images . In operation, and with respect to embodiments that use ultrasonic images, the user makes a marking on the image corresponding to the target region or site. This marking may be a point, line or area. From this, and by tracking the position of the tool in the patient coordinate system, the system functions to provide the user with visual information indicating the position of the target identified from the ultrasonic image.
11 STAN.P0009PCT The navigation system that uses the process 100 of Figure 1 operates in three distinct modes. The first is target identification mode. The imaged ultrasound volume will be displayed to allow the surgeon to locate one or more target regions of interest and mark them for targeting. The system will show an interactive volumetric rendering as well as up to three user positionable orthogonal cross-sectional planes for precise 2D location of the target. In the second mode, the endoscope will be used to set the position and orientation of the frame of reference. Based on these parameters and using the optical characteristics of the endoscope, the system will overlay target navigation data on the endoscope video. This will allow the surgeon to target regions of interest beyond the visual range of the endoscope 's field of view. Displayed data will include the directions of, and distances to, the target regions relative to the endoscope tip, as well as a potential range of error in this data. The third mode will be used to perform the actual interventional procedure (such as biopsy or ablation) once the endoscope is in the correct position. The interactive imaged ultrasound volume and cross-sectional planes will be displayed, with the location of the endoscope and the trajectory through its tip projected onto each of the views. The endoscope needle itself will also be visible in the ultrasound displays. The navigation system allows the interventional tool to be positioned in the center of the lesion without being limited to a single, fixed 2D ultrasound plane emanating from the endoscope tip. (That 2D view capability can be duplicated by optionally aligning a cross sectional ultrasound plane with the endoscope.) In the first implementation of the endoscope tracking system, a magnetic sensor will need to be removed from the working channel in order to perform the biopsy, and the navigation display will use the stored position observed immediately prior to its removal. In another embodiment, a sensor is integrated into the needle assembly, which will be in place at calibration.
12 STAN.P0009PCT The navigation system provides real-time data on the position and orientation of the endoscope, and the ultrasound system provides the dynamic image data. The tip position data is used to calculate the location of the endoscope tip in the image volume, and the probe orientation data will be used to determine the rendering camera position and orientation. Surgeon feedback will be used to improve and refine the navigation system. Procedure durations and outcomes will be compared to those of the conventional biopsy procedure, performed on the phantom without navigation and image-enhanced endoscopy assistance. When a target is identified, some embodiments store the subregion surrounding the target in the ultrasound volume as a reference, together with the tracked orientation of the volume. These embodiments will then re-sample a subregion of each successively-acquired ultrasound volume, centered at the target position in the preceding volume, by using the orientation of the reference target subregion. Some embodiments will use three-dimensional cross-correlation of the re-sampled subregion with the reference subregion to find the new location of the target. This dynamic tracking will follow each target over time; if the system is displaying target navigation data, the data will change in real time to follow the updated location of the target relative to the endoscope. Figure 2 illustrates a process 200 of some embodiments of the invention. Like the process 100 of Figure 1, the process 200 guides a medical instrument to a desired position in a patient. As shown Figure 2, the process 200 initially acquires (at 205) one or more 2D or
3D intraoperative images of the patient target site from a given orientation. Next, the process tracks (at 210) the position of a surgical instrument with respect to the patient target site. The process then registers (at 215) the intraoperative images of the patient site, the patient target site, and the surgical instrument into a common 3D reference coordinate system. Next, the process renders (at 220) the image of the patient target site on a display
13 STAN.P0009PCT device. The process also specifies (at 225) a spatial feature (shape and position) of the patient target site on the image. The process then correlates (at 230) the position and orientation of the surgical instrument with respect to the target feature. The process projects (at 235) an indicia (e.g., a three-dimensional shape, points and/or lines) on the intraoperative image relating the position and orientation of the surgical instrument to the target spatial feature. Figures 3 and 4 illustrate exemplary user interfaces for the imaging systems that use the processes illustrated in Figures 1 and 2. Figure 3 shows an exemplary user interface (UI) for ultrasound-enhanced endoscopy. The left panel shows the endoscopic view with a superimposed targeting vector and a distance measurement. The right panels show reformatted cross-sectional planes through the acquired 3D ultrasound volume. Figure 4 shows another UI for ultrasound-enhanced endoscopy. The left panel shows the endoscopic view with virtual tool tracking and visualization and vascularure acquired through Doppler imaging. The lower right panel shows volume-rendered 3D ultrasound. The UIs of Figures 3 and 4 support interactive rendering of the ultrasound data to allow a user to locate and mark the desired region of interest in the ultrasound image volume. The UIs allow the user to locate and mark target regions of interest. Hitting vascular structures is a serious hazard in endoscopic procedures. Visualization of the vasculature behind the surface tissue in the endoscopic view would assist in avoiding the vascular structures (anti-targeting). Figures 5 and 6 respectively illustrate exemplary surgical arrangements according to some embodiments of the invention. These systems, can: • track 500+ mm flexible endoscopes with an accuracy of 1.8 mm in position and 1 ° in orientation • acquire external 3D ultrasound images and process them for navigation in near real-time
14 STAN.P0009PCT • allow dynamic target identification on any reformatted 3D ultrasound cross- sectional plane view. • optionally overlay dynamic Doppler ultrasound data, rendered using intensity based opacity filters, on the endoscopic view. • provide an overall coarse target accuracy of 10 mm, with a refined target accuracy of 5mm during breath-holds. In the system of Figure 5, a video source 500 (e.g., microscopic or camcorder)is used to generate a video signal 501. In some embodiments described below, the video source 500 is an endoscopic system. An intra-operative imaging system 502 (e.g., an ultrasonic system) captures an intra-operative imaging data stream 103. The information is displayed on an ultrasonic display 504. A trackable intra-operative imaging probe 505 is also deployed in one or more trackable surgical tools 506. Other tools include a trackable endoscope 507 or any intraoperative video source. The tracking device 508 has tracking wires 509 that communicate a tracking data stream 510. A navigation system 511 with a navigation interface 512 is provided to allow the user to work with an intra-operative video image 513 (perspective view). In the absence of video source this could be blank. Primary targeting markers 514 (pointing to a target outside the field of view) as well as secondary targeting markers 515 (pointing to a target inside the field of view) can be used. An intra-operative image 516 and an image of the lesion target 517 are shown with a virtual representation of surgical tools or video source 518 (e.g., endoscope) as an orthographic view 519 (outside view). Additionally, an image overlay 520 of any arbitrary 3D shape (anatomical representation or tool representation) can also be shown. Figure 6 shows another exemplary surgical set-up. In Figure 6, several infrared vision cameras capture patient images. An ultrasonic probe positions an ultra-sound sensor in the
15 STAN.P0009PCT patient. Surgical tools such as an endoscope are then positioned in the patient. The infrared vision cameras report the position of the sensors to a computer, which in turn forwards the collected information to a workstation that generates a 3D image reconstruction. The workstation also registers, manipulates the data and visualizes the patient data on a screen. The workstation also receives data from an ultrasound machine that captures 2D images of the patient. Since the geometry of a flexible endoscope in use changes continually, the field of view at the endoscope tip is not directly dependent on the position of a tracking device attached to some other part of the endoscope. This precludes direct optical or mechanical tracking; while useful and accurate, these systems require an uninhibited line of sight or an obtrusive mechanical linkage, and thus cannot be used when tracking a flexible device within the body. In order to make use of tracked endoscope video, six extrinsic parameters (position and orientation) and five intrinsic parameters (focal length, optical center co-ordinates, aspect ratio, and lens distortion coefficient) of the imaging system are required to determine the pose of the endoscope tip and its optical characteristics. The values of these parameters for any given configuration are initially unknown. A magnetic transducer is inserted into the working channel at the endoscope tip, positioning the field generator so that the optimal sensing volume encompasses the range of sensor positions. At this time, a 6 DOF miniaturized magnetic tracking system with metal insensitivity is used, although recent developments promise improved systems in the near future. In order to correctly insert acquired ultrasound images into the volume dataset, the world co-ordinates of each pixel in the image must be determined. This requires precise tracking of the ultrasound probe as well as calibration of the ultrasound image. Current
16 STAN.P0009PCT calibration techniques are too cumbersome and time consuming to be performed prior to each use of the 3D ultrasound system. When tracking ultrasound data, the region of interest may be a significant distance from the probe itself. Consequently, any tracking error is magnified when the probe's orientation is projected to locate the region being imaged. One of the advantages of the ultrasound reconstruction engine is that it can be adapted to any existing ultrasound system configuration. In order to exploit this versatility, a simple and reliable tracking-sensor mount capability for a variety of types and sizes of ultrasound probes is used, as it is essential that the tracking sensor and ultrasound probe maintain a fixed position relative to each another after calibration. The surgeon may also wish to use the probe independently of the tracking system and its probe attachment. Accurate volume reconstruction from ultrasound images requires precise estimation of six extrinsic parameters (position and orientation) and any required intrinsic parameters such as scale. The calibration procedure should be not only accurate but also simple and quick, since it should be performed whenever the tracking sensor is mounted on the ultrasound probe or any of the relevant ultrasound imaging parameters, such as imaging depth or frequency of operation, are modified. An optical tracking system is used to measure the position and orientation of a tracking device that will be attached to the ultrasound probe.
In order to make the system practical to use in a clinical environment, spatially calibration of the intrinsic and extrinsic parameters of the ultrasound probe is done. These parameters will then be used to properly transform the ultrasound image into the co-ordinate frame of the endoscope 's field of view. The initial solution will be to use magnetic tracking for the ultrasound probe. An alternative solution is to track the probe using an optical tracking system. Tracking devices and a corresponding universal mounting bracket are deployed. In the typical 2D case,
17 STAN.P0009PCT acquired images are subject to scaling in the video generation and capture process. Since video output is not used, but the volumetric ultrasound data is accessed directly, this will not be an issue. The intrinsic parameters of the 3D probe, which will have been calibrated by the manufacturer, will be unmodified. A 200 _ 200 _ 200 mm phantom of tissue-mimicking material is used with an integrated CT-visible tracking device. Distributed along all three dimensions within the phantom will be cylinders and cubes, 20 mm in diameter and containing CT contrast material with modified acoustic impedance. The phantom will be imaged using the ultrasound probe; the transformation between the ultrasound volume and a previously acquired, reference CT volumetric image will be computed using intensity-based rigid registration (which requires the intensities of the two images to be similar in structure, but not in value). This transformation and the known position of the phantom's tracking device will be used to determine the relationship between the ultrasound imaging volume and the ultrasound probe's tracking device. Successful calibration requires an unchanged geometry. The phantom will be designed to withstand relocation and handling without deformation. A quick-release clamp attached to the phantom will hold the ultrasound probe during the calibration process. In order to locate and mark the desired region of interest in the ultrasound image volume, an interface supports interactive rendering of the ultrasound data. An interactive navigation system requires a way for the user to locate and mark target regions of interest. Respiration and other movements will cause the original location of any target to shift. If targets are not dynamically tracked, navigation information will degrade over time. The visibility of regular biopsy needles under ultrasound is poor and hitting vascular structures is a serious hazard in endoscopic procedures. Visualization of the vasculature behind the surface tissue in the endoscopic view would assist in avoiding it (anti-targeting), but
18 STAN.P0009PCT segmentation is a difficult, computationally intensive task. To address the foregoing, the navigation system operates in three distinct modes. The first is target identification mode. The imaged ultrasound volume will be displayed to allow the surgeon to locate one or more target regions of interest and mark them for targeting. The system will show an interactive volumetric rendering as well as up to three user positionable orthogonal cross-sectional planes for precise 2D location of the target. In the second mode, the endoscope will be used to set the position and orientation of the frame of reference. Based on these parameters and using the optical characteristics of the endoscope, the system will overlay target navigation data on the endoscope video. This will allow the surgeon to target regions of interest beyond the visual range of the endoscope 's field of view. Displayed data will include the directions of, and distances to, the target regions relative to the endoscope tip, as well as a potential range of error in this data. The final mode will be used to perform the actual biopsy once the endoscope is in the correct position. The interactive imaged ultrasound volume and cross-sectional planes will be displayed, with the location of the endoscope and the trajectory through its tip projected onto each of the views. The endoscope needle itself will also be visible in the ultrasound displays. This will help to position the biopsy needle in the center of the lesion without being limited to a single, fixed 2D ultrasound plane emanating from the endoscope tip, as is currently the case. (That 2D view capability will however be duplicated by optionally aligning a cross-sectional ultrasound plane with the endoscope.) In the first implementation of the flexible endoscope tracking system, the magnetic sensor will need to be removed from the working channel in order to perform the biopsy, and the navigation display will use the stored position observed immediately before its removal. Ultimately, though, a sensor will be integrated into the needle assembly, which will be in place at calibration.
19 STAN.P0009PCT When a target is identified, the subregion surrounding the target in the ultrasound volume will be stored as a reference, together with the tracked orientation of the volume. A subregion of each successively-acquired ultrasound volume, centered at the target position in the preceding volume, will be re-sampled using the orientation of the reference target subregion. Three-dimensional cross-correlation of the re-sampled subregion with the reference subregion will be used to find the new location of the target. This dynamic tracking will follow each target over time; if the system is displaying target navigation data, the data will change in real time to follow the updated location of the target relative to the endoscope. Vascular structures return a strong, well differentiated Doppler signal. The dynamic ultrasound data may be rendered in real time using intensity-based opacity filters, making nonvascular structures transparent. This effectively isolates the vascular structure without requiring computationally-demanding deformable geometric models for segmentation, thus being able to follow movement and deformation in real time. If the lag is significant, navigation accuracy will be degraded when the target moves. Where optimal accuracy is required, such as when the actual biopsy is perfonned, a brief motionless breath-hold may be required. Lens distortion compensation is performed for the data display in real time, so that the superimposed navigation display maps accurately to the underlying endoscope video. A new ultrasound volume will replace the next most recent volume in its entirety, much as it does on the display of the ultrasound machine itself, although possibly at a different spatial location. This avoids many problematic areas such as misleading old data, data expiration, unbounded imaging volumes, and locking rendering data. Instead, a simple ping-pong buffer pair may be used; one may be used for navigation and display while the
20 STAN.P0009PCT other is being updated. Another benefit of this approach is that the reduced computational complexity contributes to better interactive performance and a smaller memory footprint. All phantoms will be manufactured to tolerances at least 40 times smaller than the expected system error for the test associated with the phantom. This degree of inaccuracy is small enough to be included in the total system error without any significant impact on specifications. Computerized object recognition and surveillance are used in order to overlay images of the operative field from ultrasound onto a 3D image of the object using the Laser Targeting System and internal anatomical markers. The 3D images will be created in the workstation both from high resolution MR and CT images and intra-operative ultrasound, obtained with volume acquisition techniques. The system enables an interactive and 3D guidance system for surgeons with maximal flexibility and accuracy. With these techniques, surgery will be performed with the same tools and basic procedures as with non-guided operations, yet the precision and minimization of trauma provided by frame-based stereotaxy. An exemplary operation of the system will be discussed next help to delineate its features. Take the case of a deep intra-axial brain lesion for which resection is planned. Prior to obtaining the pre-operative images, 4 markers are placed on the patient. These can be small (-2mm), flashing diodes surrounded by pantopaque filled spheres glued to the skin. Pantopaque is an oil-based, iodine containing X-ray contrast agent that, until recently, was used for myelograms. The iodine makes it visible on CT images while the oil base renders it visible on MRI examinations. Depending on the imaging characteristics of the lesion and the important surrounding anatomy, (i.e., whether there are calcifications, whether it enhances with gadolinium, etc.) CT or MR scans or both will be performed. Typically high resolution contrast enhanced MR images and MR angiograms would be obtained. The image data would be transferred to the
21 STAN.P0009PCT workstation (via the hospital's computer network), volume rendered, and (in the case of multiple imaging modalities) fused. The image data will be segmented to allow detailed visualization in appropriate anatomic context of the lesion selected intra- and extra-axial structures, and the fiducials. Segmentation of the fiducial markers, brain, the vascular system and surface of the scalp would be fully automatic. The segmentation of the lesion, however, would be only partly automatic, as the irregular anatomy surrounding such lesions is currently too unpredictable for automatic segmentation algorithms. In the operating room, the patient would be positioned in the usual fashion. The optical tracking system would be positioned above and to the side of the patient's head. Chroma key techniques will automatically identify the flashing markers, enabling automatic and continuous registration and overlay of the patient's physical anatomy with the 3D image data sets. Chroma key is a video special effect technique that allows unique detection of flashing objects with known frequencies in 3D space (e.g., flashing light emitting diodes attached to the patient's head). Diode markers can also be added to conventional ultrasound probes and surgical tools (e.g., probes, scalpels) for their tracking in the stereotactic space. Using chroma key techniques, the markers are automatically recognized and overlaid on the display of the registered 3D image. A triangle formation of three markers on the ultrasound probe allows the tracking of its movement as it scans the surgical site, thus providing the system with volumetric ultrasound images. Continuous intra-operative registration of the patient's anatomy with the intra-operative ultrasound images is very important due to the movement and deformation of the brain tissue during the surgery. Intra-operatively acquired 3D ultrasound images are then being fused with the pre-operative CT or MR imagery using anatomical features visible to both modalities, such as vascular structures and the lesion. In addition, an extrapolated line
22 STAN.P0009PCT extending from the displayed image of a surgical device, indicates the trajectory of the planned approach. Moving the tool automatically leads to a change in the displayed potential trajectory, and the location of the L.E.D. on the tool enables determination of the precise depth and location of the tool, thereby enabling precise determination of the depth and location of the operative site. Therefore, this system not only simplifies the planning of a minimally invasion approach to a direct and interactive task, but it is also more precise than the conventional systems due to its intra-operative image updates and registration using ultrasound imagery. Furthermore, the information provided by the video camera-based object recognition system, the laser targeting system will further aid in localization of the surgical site, thus increasing the registration and image overlay performance and accuracy by localizing the area for which re-registration is needed. This information allows the workstation to automatically display the real time 3D image of the operative field in context, oriented to (and if desired overlaid with) the approach. The 3D reformatting uses volume display techniques and allows instantaneous variation of transparency. With this technique, deep as well as superficial structures, can be seen in context, thereby considerably enhancing intraoperative guidance. The system software has two aspects, one dealing with enhancements to the user interface, and the second focusing on algorithms for image manipulation and registration. These algorithms consist of the means for image segmentation, volumetric visualization, image fusion and image overlay. In one embodiment, the system provides: i) A "user friendly" interface, to facilitate usage in the operating room. ii) Interactive image analysis and manipulation routines (e.g. arbitrary cuts, image segmentation, image magnification and transformation) on the workstation system.
23 STAN.P0009PCT iii) Seamless Interface between the optical tracking system (provided with a diode encoded pointer) with a computer workstation. iv) Seamless interface of a video camera and the laser targeting system with the optical tracking system and workstation. v) Overlay of video images from (iv) on the 3D data from (iii) using diode markers positioned on the surface of the test objects. Test the accuracy of the pointer's guidance in objects. vi) Inclusion of an ultrasound probe with the optical tracking system and workstation in order to obtain 3D ultrasound images. vii) Fusion of 3D ultrasound images from (vi) to the 3D MR/CT images from (iii) using diode markers positioned on the ultrasound probe. viii) Deform test objects after their scan. Attempt to correct the deforaiation with 3D ultrasound images using linear displacement of the region of interest, then test the accuracy of the pointer's guidance in test objects. Test objects increase in complexity as testing proceeds. The above-described medical systems have numerous advantages. For instance, these systems enhance intra-operative orientation and exposure in endoscopy, which, in turn, increases surgical precision and speeds convalescence and thereby reduces overall costs. The ultrasound-enhanced endoscopy (USEE) improves localization of targets (e.g., peri-lumenal lesions) that lie hidden beyond endoscopic views. On a single endoscopic view, some of these systems dynamically superimpose directional and targeting information calculated from intra-operative ultrasonic images. Magnetic tracking and 3D ultrasound technologies are used in conjunction with dynamic 3D/video calibration and registration algorithms for precise endoscopic targeting. With USEE, clinicians use the same tools and basic procedures as for current endoscopic operations, but with a higher probability of accurate biopsy, and an
24 STAN.P0009PCT increased chance for the complete resection of the abnormality. These systems allow accurate soft-tissue navigation. The systems also provide effective calibration and correlation of intraoperative volumetric imaging data with video endoscopy images. Some of these systems acquire external 3D ultrasound images and process them for navigation in near real-time. These system allow dynamic target identification on any reformatted 3D ultrasound cross-sectional plane. The system can automatically track the movement of the target as tissue moves or deforms during the procedure. These systems can dynamically map the target location onto the endoscopic view in form of a direction vector and display quantifiable data such as distance to target. Optionally, the systems can provide targeting information on the dynamic 3D ultrasound view. The systems can virtually visualize the position and orientation of tracked surgical tools in the ultrasound view, and optionally also in the endoscopic view. These systems also can overlay dynamic Doppler ultrasound data, rendered using intensity based opacity filters, on the endoscopic view. The invention has been described in terms of specific examples, which are illustrative only and are not to be construed as limiting. The invention may be implemented in digital electronic circuitry or in computer hardware, firmware, software, or in combinations of them. Apparatus of the invention may be implemented in a computer program product tangibly embodied in a machine-readable storage device for execution by a computer processor; and method steps of the invention may be performed by a computer processor executing a program to perform functions of the invention by operating on input data and generating output. Suitable processors include, by way of example, both general and special purpose microprocessors. Storage devices suitable for tangibly embodying computer program instructions include all forms of non-volatile memory including, but not limited to: semiconductor memory devices such as EPROM, EEPROM, and flash devices; magnetic disks (fixed, floppy, and removable); other magnetic media such as tape; optical media such
25 STAN.P0009PCT as CD-ROM disks; and magneto-optic devices. Any of the foregoing may be supplemented by, or incorporated in, specially-designed application-specific integrated circuits (ASICs) or suitably programmed field programmable gate arrays (FPGAs). From the a foregoing disclosure and certain variations and modifications already disclosed therein for purposes of illustration, it will be evident to one skilled in the relevant art that the present inventive concept can be embodied in forms different from those described and it will be understood that the invention is intended to extend to such further variations. While the preferred forms of the invention have been shown in the drawings and described herein, the invention should not be construed as limited to the specific forms shown and described since variations of the preferred forms will be apparent to those skilled in the art. Thus the scope of the invention is defined by the following claims and their equivalents.
26 STAN.P0009PCT

Claims

CLAIMSI claim:
1. A method for guiding a medical instrument to a target site within a patient, comprising: capturing at least one intraoperative ultrasonic image from the patient; identifying a spatial feature indication of a patient target site on the intraoperative ultrasonic image, determining coordinates of the patient target site spatial feature in a reference coordinate system, determining a position of the instrument in the reference coordinate system, creating a view field from a predetermined position, and optionally orientation, relative to the instrument in the reference coordinate system, and projecting onto the view field an indicia of the spatial feature of the target site corresponding to the predetermined position, and optionally orientation.
2. The method of claim 1, further comprising using an ultrasonic source to generate the ultrasonic image of the patient, and determining coordinates of a spatial feature indicated on said image from the coordinates of the spatial feature on the image and the position, and optionally orientation, of the ultrasonic source.
3. The method of claim 1, wherein said medical instrument is a source of video and the view field projected onto the display device is the image seen by the video source.
27 STAN.P0009PCT
4. A computer readable medium that stores a computer program that is designed to assist a user in guiding a medical instrument to a target site in a patient, said computer program comprising sets of instructions for: capturing at least one image during an operation of the patient; from a user, receiving an indication of a target site on the captured image; based on the indication, determining coordinates of the patient target site in a reference coordinate system; determining a position of the instrument in the reference coordinate system; projecting onto the display device a field of view from a perspective of the instrument in the reference coordinate system; and projecting onto the field of view the indicia that specifies the position of the target site relative to the position of the instrument.
5. The computer readable medium of claim 4, wherein the computer program further comprises sets of instructions for: using an ultrasonic source to generate the ultrasonic image of the patient, and determining coordinates of a spatial feature indicated on said image from the coordinates of the spatial feature on the image and the position, and optionally orientation, of the ultrasonic source.
6. The computer readable medium of claim 4, wherein said medical instrument is a source of video and the view field projected onto the display device is the image seen by the video source.
28 STAN.P0009PCT
PCT/US2004/035014 2003-10-21 2004-10-21 Systems and methods for intraoperative targetting WO2005039391A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP20040796074 EP1680024A2 (en) 2003-10-21 2004-10-21 Systems and methods for intraoperative targetting
US10/576,632 US20070276234A1 (en) 2003-10-21 2004-10-21 Systems and Methods for Intraoperative Targeting
JP2006536816A JP2007508913A (en) 2003-10-21 2004-10-21 Intraoperative targeting system and method

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US51315703P 2003-10-21 2003-10-21
US60/513,157 2003-10-21
US10/764,650 2004-01-26
US10/764,650 US20050085717A1 (en) 2003-10-21 2004-01-26 Systems and methods for intraoperative targetting
US10/764,651 US20050085718A1 (en) 2003-10-21 2004-01-26 Systems and methods for intraoperative targetting
US10/764,651 2004-01-26

Publications (2)

Publication Number Publication Date
WO2005039391A2 true WO2005039391A2 (en) 2005-05-06
WO2005039391A3 WO2005039391A3 (en) 2005-12-22

Family

ID=34527934

Family Applications (2)

Application Number Title Priority Date Filing Date
PCT/US2004/035024 WO2005043319A2 (en) 2003-10-21 2004-10-21 Systems and methods for intraoperative targeting
PCT/US2004/035014 WO2005039391A2 (en) 2003-10-21 2004-10-21 Systems and methods for intraoperative targetting

Family Applications Before (1)

Application Number Title Priority Date Filing Date
PCT/US2004/035024 WO2005043319A2 (en) 2003-10-21 2004-10-21 Systems and methods for intraoperative targeting

Country Status (4)

Country Link
US (2) US20070276234A1 (en)
EP (2) EP1689290A2 (en)
JP (2) JP2007531553A (en)
WO (2) WO2005043319A2 (en)

Cited By (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1795130A1 (en) * 2005-12-05 2007-06-13 Medison Co., Ltd. Ultrasound system for interventional treatment
EP1779779A3 (en) * 2005-09-16 2007-09-12 Mediguide Ltd. Method and system for delivering a medical device to a selected position within a lumen
WO2008006180A1 (en) * 2006-07-10 2008-01-17 Katholieke Universiteit Leuven Endoscopic vision system
EP1953564A3 (en) * 2007-01-31 2008-08-13 Biosense Webster, Inc. Correlation of ultrasound images and gated position measurements
ITGE20080064A1 (en) * 2008-07-24 2010-01-25 Esaote Spa DEVICE AND METHOD OF GUIDANCE OF SURGICAL UTENSILS BY ECOGRAPHIC IMAGING.
US7697973B2 (en) 1999-05-18 2010-04-13 MediGuide, Ltd. Medical imaging and navigation system
US7778688B2 (en) 1999-05-18 2010-08-17 MediGuide, Ltd. System and method for delivering a stent to a selected position within a lumen
US7840252B2 (en) 1999-05-18 2010-11-23 MediGuide, Ltd. Method and system for determining a three dimensional representation of a tubular organ
US7938777B2 (en) 2006-07-21 2011-05-10 Orthosoft Inc. Non-invasive tracking of bones for surgery
JP2012050888A (en) * 2005-06-06 2012-03-15 Intuitive Surgical Inc Laparoscopic ultrasound robotic surgical system
EP2502558A1 (en) * 2011-03-22 2012-09-26 KUKA Laboratories GmbH Medical workstation
US8303502B2 (en) 2007-03-06 2012-11-06 General Electric Company Method and apparatus for tracking points in an ultrasound image
US9101397B2 (en) 1999-04-07 2015-08-11 Intuitive Surgical Operations, Inc. Real-time generation of three-dimensional ultrasound image using a two-dimensional ultrasound transducer in a robotic system
US9138129B2 (en) 2007-06-13 2015-09-22 Intuitive Surgical Operations, Inc. Method and system for moving a plurality of articulated instruments in tandem back towards an entry guide
WO2015158736A1 (en) * 2014-04-15 2015-10-22 Fiagon Ag Medical Technologies Navigation assistance system for medical instruments
US9333042B2 (en) 2007-06-13 2016-05-10 Intuitive Surgical Operations, Inc. Medical robotic system with coupled control modes
US9345387B2 (en) 2006-06-13 2016-05-24 Intuitive Surgical Operations, Inc. Preventing instrument/tissue collisions
US9469034B2 (en) 2007-06-13 2016-10-18 Intuitive Surgical Operations, Inc. Method and system for switching modes of a robotic system
US9492927B2 (en) 2009-08-15 2016-11-15 Intuitive Surgical Operations, Inc. Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose
US9516996B2 (en) 2008-06-27 2016-12-13 Intuitive Surgical Operations, Inc. Medical robotic system providing computer generated auxiliary views of a camera instrument for controlling the position and orienting of its tip
US9622826B2 (en) 2010-02-12 2017-04-18 Intuitive Surgical Operations, Inc. Medical robotic system providing sensory feedback indicating a difference between a commanded state and a preferred pose of an articulated instrument
US9718190B2 (en) 2006-06-29 2017-08-01 Intuitive Surgical Operations, Inc. Tool position and identification indicator displayed in a boundary area of a computer display screen
US9717563B2 (en) 2008-06-27 2017-08-01 Intuitive Surgical Operations, Inc. Medical robotic system providing an auxilary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide
US9788909B2 (en) 2006-06-29 2017-10-17 Intuitive Surgical Operations, Inc Synthetic representation of a surgical instrument
US9789608B2 (en) 2006-06-29 2017-10-17 Intuitive Surgical Operations, Inc. Synthetic representation of a surgical robot
US9795446B2 (en) 2005-06-06 2017-10-24 Intuitive Surgical Operations, Inc. Systems and methods for interactive user interfaces for robotic minimally invasive surgical systems
US9833167B2 (en) 1999-05-18 2017-12-05 Mediguide Ltd. Method and system for superimposing virtual anatomical landmarks on an image
US9956049B2 (en) 1999-05-18 2018-05-01 Mediguide Ltd. Method and apparatus for invasive device tracking using organ timing signal generated from MPS sensors
US9956044B2 (en) 2009-08-15 2018-05-01 Intuitive Surgical Operations, Inc. Controller assisted reconfiguration of an articulated instrument during movement into and out of an entry guide
US10008017B2 (en) 2006-06-29 2018-06-26 Intuitive Surgical Operations, Inc. Rendering tool information as graphic overlays on displayed images of tools
US10258425B2 (en) 2008-06-27 2019-04-16 Intuitive Surgical Operations, Inc. Medical robotic system providing an auxiliary view of articulatable instruments extending out of a distal end of an entry guide
US10507066B2 (en) 2013-02-15 2019-12-17 Intuitive Surgical Operations, Inc. Providing information of tools by filtering image areas adjacent to or on displayed images of the tools
US10758209B2 (en) 2012-03-09 2020-09-01 The Johns Hopkins University Photoacoustic tracking and registration in interventional ultrasound
US10806346B2 (en) 2015-02-09 2020-10-20 The Johns Hopkins University Photoacoustic tracking and registration in interventional ultrasound
US11259870B2 (en) 2005-06-06 2022-03-01 Intuitive Surgical Operations, Inc. Interactive user interfaces for minimally invasive telesurgical systems
US11357574B2 (en) 2013-10-31 2022-06-14 Intersect ENT International GmbH Surgical instrument and method for detecting the position of a surgical instrument
US11430139B2 (en) 2019-04-03 2022-08-30 Intersect ENT International GmbH Registration method and setup
US11446090B2 (en) 2017-04-07 2022-09-20 Orthosoft Ulc Non-invasive system and method for tracking bones
US11684426B2 (en) 2018-08-31 2023-06-27 Orthosoft Ulc System and method for tracking bones
US11701188B2 (en) 2017-05-10 2023-07-18 Mako Surgical Corp. Robotic spine surgery system and methods

Families Citing this family (210)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2855292B1 (en) * 2003-05-22 2005-12-09 Inst Nat Rech Inf Automat DEVICE AND METHOD FOR REAL TIME REASONING OF PATTERNS ON IMAGES, IN PARTICULAR FOR LOCALIZATION GUIDANCE
EP2316328B1 (en) * 2003-09-15 2012-05-09 Super Dimension Ltd. Wrap-around holding device for use with bronchoscopes
US7379769B2 (en) 2003-09-30 2008-05-27 Sunnybrook Health Sciences Center Hybrid imaging method to monitor medical device delivery and patient support for use in the method
DE102004008164B3 (en) * 2004-02-11 2005-10-13 Karl Storz Gmbh & Co. Kg Method and device for creating at least a section of a virtual 3D model of a body interior
US20060020204A1 (en) * 2004-07-01 2006-01-26 Bracco Imaging, S.P.A. System and method for three-dimensional space management and visualization of ultrasound data ("SonoDEX")
US8795195B2 (en) * 2004-11-29 2014-08-05 Senorx, Inc. Graphical user interface for tissue biopsy system
JP5122743B2 (en) * 2004-12-20 2013-01-16 ゼネラル・エレクトリック・カンパニイ System for aligning 3D images within an interventional system
EP1858418B1 (en) * 2005-02-28 2017-08-16 Robarts Research Institute System for performing a biopsy of a target volume and a computing device for planning the same
US20060235299A1 (en) * 2005-04-13 2006-10-19 Martinelli Michael A Apparatus and method for intravascular imaging
US8257302B2 (en) * 2005-05-10 2012-09-04 Corindus, Inc. User interface for remote control catheterization
US7889905B2 (en) * 2005-05-23 2011-02-15 The Penn State Research Foundation Fast 3D-2D image registration method with application to continuously guided endoscopy
US7681579B2 (en) * 2005-08-02 2010-03-23 Biosense Webster, Inc. Guided procedures for treating atrial fibrillation
WO2007041383A2 (en) * 2005-09-30 2007-04-12 Purdue Research Foundation Endoscopic imaging device
WO2008017051A2 (en) 2006-08-02 2008-02-07 Inneroptic Technology Inc. System and method of providing real-time dynamic imagery of a medical procedure site using multiple modalities
US8248413B2 (en) 2006-09-18 2012-08-21 Stryker Corporation Visual navigation system for endoscopic surgery
US7824328B2 (en) * 2006-09-18 2010-11-02 Stryker Corporation Method and apparatus for tracking a surgical instrument during surgery
US8248414B2 (en) * 2006-09-18 2012-08-21 Stryker Corporation Multi-dimensional navigation of endoscopic video
US7945310B2 (en) * 2006-09-18 2011-05-17 Stryker Corporation Surgical instrument path computation and display for endoluminal surgery
US20080071141A1 (en) * 2006-09-18 2008-03-20 Abhisuek Gattani Method and apparatus for measuring attributes of an anatomical feature during a medical procedure
KR100971417B1 (en) * 2006-10-17 2010-07-21 주식회사 메디슨 Ultrasound system for displaying neddle for medical treatment on compound image of ultrasound image and external medical image
US8340374B2 (en) * 2007-01-11 2012-12-25 Kabushiki Kaisha Toshiba 3-dimensional diagnostic imaging system
US9477686B2 (en) * 2007-01-12 2016-10-25 General Electric Company Systems and methods for annotation and sorting of surgical images
EP2123232A4 (en) * 2007-01-31 2011-02-16 Nat University Corp Hamamatsu University School Of Medicine Device for displaying assistance information for surgical operation, method for displaying assistance information for surgical operation, and program for displaying assistance information for surgical operation
JP4960112B2 (en) * 2007-02-01 2012-06-27 オリンパスメディカルシステムズ株式会社 Endoscopic surgery device
EP2211749B1 (en) * 2007-04-16 2018-10-10 NeuroArm Surgical, Ltd. Methods, devices, and systems useful in registration
JP4934513B2 (en) * 2007-06-08 2012-05-16 株式会社日立メディコ Ultrasonic imaging device
DE102007029888B4 (en) * 2007-06-28 2016-04-07 Siemens Aktiengesellschaft Medical diagnostic imaging and apparatus operating according to this method
US20090074265A1 (en) * 2007-09-17 2009-03-19 Capsovision Inc. Imaging review and navigation workstation system
CA2751629C (en) * 2007-10-19 2016-08-23 Metritrack, Llc Three dimensional mapping display system for diagnostic ultrasound machines and method
US8290569B2 (en) 2007-11-23 2012-10-16 Hologic, Inc. Open architecture tabletop patient support and coil system
WO2009094646A2 (en) 2008-01-24 2009-07-30 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for image guided ablation
JP5154961B2 (en) * 2008-01-29 2013-02-27 テルモ株式会社 Surgery system
CN101951853B (en) * 2008-02-22 2013-01-23 洛马林达大学医学中心 Systems and methods for characterizing spatial distortion in 3D imaging systems
US8340379B2 (en) 2008-03-07 2012-12-25 Inneroptic Technology, Inc. Systems and methods for displaying guidance data based on updated deformable imaging data
JP5288447B2 (en) * 2008-03-28 2013-09-11 学校法人早稲田大学 Surgery support system, approach state detection device and program thereof
WO2009122273A2 (en) 2008-04-03 2009-10-08 Superdimension, Ltd. Magnetic interference detection system and method
WO2009137410A1 (en) 2008-05-06 2009-11-12 Corindus Ltd. Catheter system
US9198597B2 (en) * 2008-05-22 2015-12-01 Christopher Duma Leading-edge cancer treatment
WO2009147671A1 (en) 2008-06-03 2009-12-10 Superdimension Ltd. Feature-based registration method
US8218847B2 (en) 2008-06-06 2012-07-10 Superdimension, Ltd. Hybrid registration method
US20090312629A1 (en) * 2008-06-13 2009-12-17 Inneroptic Technology Inc. Correction of relative tracking errors based on a fiducial
WO2010025338A1 (en) 2008-08-29 2010-03-04 Corindus Ltd. Catheter control system and graphical user interface
JP2010088699A (en) * 2008-10-09 2010-04-22 National Center For Child Health & Development Medical image processing system
US8554307B2 (en) 2010-04-12 2013-10-08 Inneroptic Technology, Inc. Image annotation in image-guided medical procedures
US11464578B2 (en) 2009-02-17 2022-10-11 Inneroptic Technology, Inc. Systems, methods, apparatuses, and computer-readable media for image management in image-guided medical procedures
US8641621B2 (en) 2009-02-17 2014-02-04 Inneroptic Technology, Inc. Systems, methods, apparatuses, and computer-readable media for image management in image-guided medical procedures
US8690776B2 (en) 2009-02-17 2014-04-08 Inneroptic Technology, Inc. Systems, methods, apparatuses, and computer-readable media for image guided surgery
EP2408509B1 (en) 2009-03-18 2023-08-09 Corindus, Inc. Remote catheter system with steerable catheter
EP2429438A1 (en) * 2009-04-24 2012-03-21 Medtronic, Inc. Electromagnetic navigation of medical instruments for cardiothoracic surgery
EP2437676A1 (en) * 2009-06-01 2012-04-11 Koninklijke Philips Electronics N.V. Distance-based position tracking method and system
WO2010148503A1 (en) 2009-06-23 2010-12-29 Sentinelle Medical Inc. Variable angle guide holder for a biopsy guide plug
JP5377153B2 (en) * 2009-08-18 2013-12-25 株式会社東芝 Image processing apparatus, image processing program, and medical diagnostic system
KR101598774B1 (en) * 2009-10-01 2016-03-02 (주)미래컴퍼니 Apparatus and Method for processing surgical image
WO2011040769A2 (en) * 2009-10-01 2011-04-07 주식회사 이턴 Surgical image processing device, image-processing method, laparoscopic manipulation method, surgical robot system and an operation-limiting method therefor
EP3572115B1 (en) 2009-10-12 2024-02-21 Corindus, Inc. Catheter system with percutaneous device movement algorithm
US9962229B2 (en) 2009-10-12 2018-05-08 Corindus, Inc. System and method for navigating a guide wire
US8758263B1 (en) * 2009-10-31 2014-06-24 Voxel Rad, Ltd. Systems and methods for frameless image-guided biopsy and therapeutic intervention
EP2503934B1 (en) * 2009-11-27 2021-10-20 Hologic, Inc. Systems and methods for tracking positions between imaging modalities and transforming a displayed three-dimensional image corresponding to a position and orientation of a probe
US9282947B2 (en) 2009-12-01 2016-03-15 Inneroptic Technology, Inc. Imager focusing based on intraoperative data
US20110160566A1 (en) * 2009-12-24 2011-06-30 Labros Petropoulos Mri and ultrasound guided treatment on a patient
CN102883651B (en) * 2010-01-28 2016-04-27 宾夕法尼亚州研究基金会 Can be applicable to the global registration system and method based on image that bronchoscope guides
JP5551957B2 (en) * 2010-03-31 2014-07-16 富士フイルム株式会社 Projection image generation apparatus, operation method thereof, and projection image generation program
US8696549B2 (en) 2010-08-20 2014-04-15 Veran Medical Technologies, Inc. Apparatus and method for four dimensional soft tissue navigation in endoscopic applications
US9833293B2 (en) 2010-09-17 2017-12-05 Corindus, Inc. Robotic catheter system
US20120083653A1 (en) * 2010-10-04 2012-04-05 Sperling Daniel P Guided procedural treatment device and method
JP5485853B2 (en) * 2010-10-14 2014-05-07 株式会社日立メディコ MEDICAL IMAGE DISPLAY DEVICE AND MEDICAL IMAGE GUIDANCE METHOD
KR101242298B1 (en) 2010-11-01 2013-03-11 삼성메디슨 주식회사 Ultrasound system and method for storing ultrasound images
CN103228210B (en) * 2010-11-24 2016-08-10 美国医软科技公司 The system and method for the interactive three-dimensional operation guiding system for soft organ based on internal anatomy
US9332926B2 (en) 2010-11-25 2016-05-10 Invivo Corporation MRI imaging probe
DE102010062340A1 (en) * 2010-12-02 2012-06-06 Siemens Aktiengesellschaft Method for image support of the navigation of a medical instrument and medical examination device
EP2468207A1 (en) * 2010-12-21 2012-06-27 Renishaw (Ireland) Limited Method and apparatus for analysing images
SG182880A1 (en) * 2011-02-01 2012-08-30 Univ Singapore A method and system for interaction with micro-objects
JP6002695B2 (en) * 2011-03-18 2016-10-05 コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. Tracking brain deformation in neurosurgery
KR101114231B1 (en) 2011-05-16 2012-03-05 주식회사 이턴 Apparatus and Method for processing surgical image
KR101114232B1 (en) 2011-05-17 2012-03-05 주식회사 이턴 Surgical robot system and motion restriction control method thereof
JP5623348B2 (en) * 2011-07-06 2014-11-12 富士フイルム株式会社 Endoscope system, processor device for endoscope system, and method for operating endoscope system
JP6071282B2 (en) * 2011-08-31 2017-02-01 キヤノン株式会社 Information processing apparatus, ultrasonic imaging apparatus, and information processing method
DE102011082444A1 (en) * 2011-09-09 2012-12-20 Siemens Aktiengesellschaft Image-supported navigation method of e.g. endoscope used in medical intervention of human body, involves registering and representing captured image with 3D data set by optical detection system
JP5854399B2 (en) * 2011-11-21 2016-02-09 オリンパス株式会社 Medical system
US8670816B2 (en) 2012-01-30 2014-03-11 Inneroptic Technology, Inc. Multiple medical device guidance
CN103561628B (en) * 2012-03-06 2016-08-17 奥林巴斯株式会社 Endoscopic system
US9561019B2 (en) * 2012-03-07 2017-02-07 Ziteo, Inc. Methods and systems for tracking and guiding sensors and instruments
US9259181B2 (en) 2012-04-26 2016-02-16 Medtronic, Inc. Visualizing tissue activated by electrical stimulation
CA2794226C (en) * 2012-10-31 2020-10-20 Queen's University At Kingston Automated intraoperative ultrasound calibration
US10314559B2 (en) 2013-03-14 2019-06-11 Inneroptic Technology, Inc. Medical device guidance
EP3633543A1 (en) 2013-03-15 2020-04-08 Hologic, Inc. System and method for reviewing and analyzing cytological specimens
US9592095B2 (en) 2013-05-16 2017-03-14 Intuitive Surgical Operations, Inc. Systems and methods for robotic medical system integration with external imaging
WO2015011594A1 (en) * 2013-07-24 2015-01-29 Koninklijke Philips N.V. Non-imaging two dimensional array probe and system for automated screening of carotid stenosis
JP5869541B2 (en) 2013-09-13 2016-02-24 富士フイルム株式会社 ENDOSCOPE SYSTEM, PROCESSOR DEVICE, AND METHOD FOR OPERATING ENDOSCOPE SYSTEM
JP5892985B2 (en) * 2013-09-27 2016-03-23 富士フイルム株式会社 ENDOSCOPE SYSTEM, PROCESSOR DEVICE, AND OPERATION METHOD
JP2016531713A (en) * 2013-09-30 2016-10-13 コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. Image guidance system including user-definable region of interest
US10835203B2 (en) 2013-11-11 2020-11-17 Acessa Health Inc. System for visualization and control of surgical devices utilizing a graphical user interface
JP6270026B2 (en) * 2013-12-05 2018-01-31 国立大学法人名古屋大学 Endoscopic observation support device
WO2015148529A1 (en) * 2014-03-24 2015-10-01 University Of Houston Interactive systems and methods for real-time laparoscopic navigation
US9999772B2 (en) * 2014-04-03 2018-06-19 Pacesetter, Inc. Systems and method for deep brain stimulation therapy
WO2015166413A1 (en) * 2014-04-28 2015-11-05 Tel Hashomer Medical Research, Infrastructure And Services Ltd. High resolution intraoperative mri images
US9770216B2 (en) * 2014-07-02 2017-09-26 Covidien Lp System and method for navigating within the lung
EP4233707A3 (en) * 2014-07-02 2023-10-11 Covidien LP System and program for providing distance and orientation feedback while navigating in 3d
EP3169264A1 (en) * 2014-07-15 2017-05-24 Koninklijke Philips N.V. Image integration and robotic endoscope control in x-ray suite
TWI605795B (en) 2014-08-19 2017-11-21 鈦隼生物科技股份有限公司 Method and system of determining probe position in surgical site
US10583293B2 (en) 2014-09-09 2020-03-10 Medtronic, Inc. Therapy program selection for electrical stimulation therapy based on a volume of tissue activation
US9901406B2 (en) 2014-10-02 2018-02-27 Inneroptic Technology, Inc. Affected region display associated with a medical device
US10617401B2 (en) 2014-11-14 2020-04-14 Ziteo, Inc. Systems for localization of targets inside a body
JP7031950B2 (en) 2014-12-05 2022-03-08 コリンダス、インコーポレイテッド Catheter treatment system
US10188467B2 (en) 2014-12-12 2019-01-29 Inneroptic Technology, Inc. Surgical guidance intersection display
US10154239B2 (en) 2014-12-30 2018-12-11 Onpoint Medical, Inc. Image-guided surgery with surface reconstruction and augmented reality visualization
GB2555012B (en) * 2015-03-17 2020-09-09 Synaptive Medical Barbados Inc Method and device for registering surgical images
JP6766062B2 (en) * 2015-03-17 2020-10-07 インテュイティブ サージカル オペレーションズ, インコーポレイテッド Systems and methods for on-screen identification of instruments in remote-controlled medical systems
US9949700B2 (en) 2015-07-22 2018-04-24 Inneroptic Technology, Inc. Medical device approaches
US10638954B2 (en) 2015-07-23 2020-05-05 Biosense Webster (Israel) Ltd. Surface registration of a CT image with a magnetic tracking system
US20170084036A1 (en) * 2015-09-21 2017-03-23 Siemens Aktiengesellschaft Registration of video camera with medical imaging
EP3367950A4 (en) 2015-10-28 2019-10-02 Endochoice, Inc. Device and method for tracking the position of an endoscope within a patient's body
EP3402408B1 (en) * 2016-01-15 2020-09-02 Koninklijke Philips N.V. Automated probe steering to clinical views using annotations in a fused image guidance system
US9675319B1 (en) 2016-02-17 2017-06-13 Inneroptic Technology, Inc. Loupe display
CN109310476B (en) 2016-03-12 2020-04-03 P·K·朗 Devices and methods for surgery
CN106063726B (en) * 2016-05-24 2019-05-28 中国科学院苏州生物医学工程技术研究所 Navigation system and its air navigation aid are punctured in real time
US10959702B2 (en) 2016-06-20 2021-03-30 Butterfly Network, Inc. Automated image acquisition for assisting a user to operate an ultrasound device
US10278778B2 (en) 2016-10-27 2019-05-07 Inneroptic Technology, Inc. Medical device navigation using a virtual 3D space
US10638952B2 (en) 2016-10-28 2020-05-05 Covidien Lp Methods, systems, and computer-readable media for calibrating an electromagnetic navigation system
US10615500B2 (en) 2016-10-28 2020-04-07 Covidien Lp System and method for designing electromagnetic navigation antenna assemblies
US10517505B2 (en) 2016-10-28 2019-12-31 Covidien Lp Systems, methods, and computer-readable media for optimizing an electromagnetic navigation system
US10792106B2 (en) 2016-10-28 2020-10-06 Covidien Lp System for calibrating an electromagnetic navigation system
US10418705B2 (en) 2016-10-28 2019-09-17 Covidien Lp Electromagnetic navigation antenna assembly and electromagnetic navigation system including the same
US10446931B2 (en) 2016-10-28 2019-10-15 Covidien Lp Electromagnetic navigation antenna assembly and electromagnetic navigation system including the same
US10722311B2 (en) 2016-10-28 2020-07-28 Covidien Lp System and method for identifying a location and/or an orientation of an electromagnetic sensor based on a map
US10751126B2 (en) 2016-10-28 2020-08-25 Covidien Lp System and method for generating a map for electromagnetic navigation
CN109922753B (en) * 2016-12-08 2023-04-14 直观外科手术操作公司 System and method for navigation in image-guided medical procedures
JP6745998B2 (en) * 2016-12-16 2020-08-26 コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. System that provides images to guide surgery
US10918445B2 (en) * 2016-12-19 2021-02-16 Ethicon Llc Surgical system with augmented reality display
WO2018132804A1 (en) 2017-01-16 2018-07-19 Lang Philipp K Optical guidance for surgical, medical, and dental procedures
US10980509B2 (en) * 2017-05-11 2021-04-20 Siemens Medical Solutions Usa, Inc. Deformable registration of preoperative volumes and intraoperative ultrasound images from a tracked transducer
US11259879B2 (en) 2017-08-01 2022-03-01 Inneroptic Technology, Inc. Selective transparency to assist medical device navigation
US10699410B2 (en) * 2017-08-17 2020-06-30 Siemes Healthcare GmbH Automatic change detection in medical images
WO2019051464A1 (en) 2017-09-11 2019-03-14 Lang Philipp K Augmented reality display for vascular and other interventions, compensation for cardiac and respiratory motion
CN109745069A (en) * 2017-11-01 2019-05-14 通用电气公司 ultrasonic imaging method
US11823403B2 (en) 2017-12-27 2023-11-21 Cilag Gmbh International Fluorescence imaging in a light deficient environment
US11484365B2 (en) 2018-01-23 2022-11-01 Inneroptic Technology, Inc. Medical image guidance
WO2019148154A1 (en) 2018-01-29 2019-08-01 Lang Philipp K Augmented reality guidance for orthopedic and other surgical procedures
GB201813450D0 (en) * 2018-08-17 2018-10-03 Hiltermann Sean Augmented reality doll
US20200060643A1 (en) * 2018-08-22 2020-02-27 Bard Access Systems, Inc. Systems and Methods for Infrared-Enhanced Ultrasound Visualization
US10806339B2 (en) 2018-12-12 2020-10-20 Voxel Rad, Ltd. Systems and methods for treating cancer using brachytherapy
US11350847B2 (en) * 2018-12-13 2022-06-07 Biosense Webster (Israel) Ltd. Composite visualization of body part
US11857378B1 (en) 2019-02-14 2024-01-02 Onpoint Medical, Inc. Systems for adjusting and tracking head mounted displays during surgery including with surgical helmets
US11553969B1 (en) 2019-02-14 2023-01-17 Onpoint Medical, Inc. System for computation of object coordinates accounting for movement of a surgical site for spinal and other procedures
JP2020156825A (en) * 2019-03-27 2020-10-01 富士フイルム株式会社 Position information display device, method, and program, and radiography apparatus
WO2020210532A1 (en) 2019-04-09 2020-10-15 Ziteo, Inc. Methods and systems for high performance and versatile molecular imaging
US11265491B2 (en) 2019-06-20 2022-03-01 Cilag Gmbh International Fluorescence imaging with fixed pattern noise cancellation
US11012599B2 (en) 2019-06-20 2021-05-18 Ethicon Llc Hyperspectral imaging in a light deficient environment
US11172811B2 (en) 2019-06-20 2021-11-16 Cilag Gmbh International Image rotation in an endoscopic fluorescence imaging system
US11276148B2 (en) 2019-06-20 2022-03-15 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed fluorescence imaging system
US11533417B2 (en) 2019-06-20 2022-12-20 Cilag Gmbh International Laser scanning and tool tracking imaging in a light deficient environment
US11716533B2 (en) 2019-06-20 2023-08-01 Cilag Gmbh International Image synchronization without input clock and data transmission clock in a pulsed fluorescence imaging system
US11172810B2 (en) 2019-06-20 2021-11-16 Cilag Gmbh International Speckle removal in a pulsed laser mapping imaging system
US11624830B2 (en) 2019-06-20 2023-04-11 Cilag Gmbh International Wide dynamic range using a monochrome image sensor for laser mapping imaging
US11716543B2 (en) 2019-06-20 2023-08-01 Cilag Gmbh International Wide dynamic range using a monochrome image sensor for fluorescence imaging
US11471055B2 (en) 2019-06-20 2022-10-18 Cilag Gmbh International Noise aware edge enhancement in a pulsed fluorescence imaging system
US11233960B2 (en) 2019-06-20 2022-01-25 Cilag Gmbh International Fluorescence imaging with fixed pattern noise cancellation
US11925328B2 (en) 2019-06-20 2024-03-12 Cilag Gmbh International Noise aware edge enhancement in a pulsed hyperspectral imaging system
US11944273B2 (en) 2019-06-20 2024-04-02 Cilag Gmbh International Fluorescence videostroboscopy of vocal cords
US11892403B2 (en) 2019-06-20 2024-02-06 Cilag Gmbh International Image synchronization without input clock and data transmission clock in a pulsed fluorescence imaging system
US11903563B2 (en) 2019-06-20 2024-02-20 Cilag Gmbh International Offset illumination of a scene using multiple emitters in a fluorescence imaging system
US11740448B2 (en) 2019-06-20 2023-08-29 Cilag Gmbh International Driving light emissions according to a jitter specification in a fluorescence imaging system
US11360028B2 (en) 2019-06-20 2022-06-14 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed hyperspectral, fluorescence, and laser mapping imaging system
US11218645B2 (en) 2019-06-20 2022-01-04 Cilag Gmbh International Wide dynamic range using a monochrome image sensor for fluorescence imaging
US11898909B2 (en) 2019-06-20 2024-02-13 Cilag Gmbh International Noise aware edge enhancement in a pulsed fluorescence imaging system
US11937784B2 (en) 2019-06-20 2024-03-26 Cilag Gmbh International Fluorescence imaging in a light deficient environment
US11931009B2 (en) 2019-06-20 2024-03-19 Cilag Gmbh International Offset illumination of a scene using multiple emitters in a hyperspectral imaging system
US11412152B2 (en) 2019-06-20 2022-08-09 Cilag Gmbh International Speckle removal in a pulsed hyperspectral imaging system
US11877065B2 (en) 2019-06-20 2024-01-16 Cilag Gmbh International Image rotation in an endoscopic hyperspectral imaging system
US10841504B1 (en) 2019-06-20 2020-11-17 Ethicon Llc Fluorescence imaging with minimal area monolithic image sensor
US11237270B2 (en) 2019-06-20 2022-02-01 Cilag Gmbh International Hyperspectral, fluorescence, and laser mapping imaging with fixed pattern noise cancellation
US11516387B2 (en) 2019-06-20 2022-11-29 Cilag Gmbh International Image synchronization without input clock and data transmission clock in a pulsed hyperspectral, fluorescence, and laser mapping imaging system
US11700995B2 (en) 2019-06-20 2023-07-18 Cilag Gmbh International Speckle removal in a pulsed fluorescence imaging system
US11633089B2 (en) 2019-06-20 2023-04-25 Cilag Gmbh International Fluorescence imaging with minimal area monolithic image sensor
US11671691B2 (en) 2019-06-20 2023-06-06 Cilag Gmbh International Image rotation in an endoscopic laser mapping imaging system
US11187657B2 (en) 2019-06-20 2021-11-30 Cilag Gmbh International Hyperspectral imaging with fixed pattern noise cancellation
US11221414B2 (en) 2019-06-20 2022-01-11 Cilag Gmbh International Laser mapping imaging with fixed pattern noise cancellation
US11540696B2 (en) 2019-06-20 2023-01-03 Cilag Gmbh International Noise aware edge enhancement in a pulsed fluorescence imaging system
US11793399B2 (en) 2019-06-20 2023-10-24 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed hyperspectral imaging system
US11432706B2 (en) 2019-06-20 2022-09-06 Cilag Gmbh International Hyperspectral imaging with minimal area monolithic image sensor
US12013496B2 (en) 2019-06-20 2024-06-18 Cilag Gmbh International Noise aware edge enhancement in a pulsed laser mapping imaging system
US11550057B2 (en) 2019-06-20 2023-01-10 Cilag Gmbh International Offset illumination of a scene using multiple emitters in a fluorescence imaging system
US11284785B2 (en) 2019-06-20 2022-03-29 Cilag Gmbh International Controlling integral energy of a laser pulse in a hyperspectral, fluorescence, and laser mapping imaging system
US11375886B2 (en) 2019-06-20 2022-07-05 Cilag Gmbh International Optical fiber waveguide in an endoscopic system for laser mapping imaging
US11187658B2 (en) 2019-06-20 2021-11-30 Cilag Gmbh International Fluorescence imaging with fixed pattern noise cancellation
US20200397270A1 (en) 2019-06-20 2020-12-24 Ethicon Llc Optical fiber waveguide in an endoscopic system for fluorescence imaging
US11398011B2 (en) 2019-06-20 2022-07-26 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed laser mapping imaging system
US11674848B2 (en) 2019-06-20 2023-06-13 Cilag Gmbh International Wide dynamic range using a monochrome image sensor for hyperspectral imaging
US11516388B2 (en) 2019-06-20 2022-11-29 Cilag Gmbh International Pulsed illumination in a fluorescence imaging system
US11288772B2 (en) 2019-06-20 2022-03-29 Cilag Gmbh International Super resolution and color motion artifact correction in a pulsed fluorescence imaging system
US11266304B2 (en) 2019-06-20 2022-03-08 Cilag Gmbh International Minimizing image sensor input/output in a pulsed hyperspectral imaging system
US11389066B2 (en) 2019-06-20 2022-07-19 Cilag Gmbh International Noise aware edge enhancement in a pulsed hyperspectral, fluorescence, and laser mapping imaging system
US10979646B2 (en) 2019-06-20 2021-04-13 Ethicon Llc Fluorescence imaging with minimal area monolithic image sensor
US11986160B2 (en) 2019-06-20 2024-05-21 Cllag GmbH International Image synchronization without input clock and data transmission clock in a pulsed hyperspectral imaging system
US11457154B2 (en) 2019-06-20 2022-09-27 Cilag Gmbh International Speckle removal in a pulsed hyperspectral, fluorescence, and laser mapping imaging system
US11294062B2 (en) 2019-06-20 2022-04-05 Cilag Gmbh International Dynamic range using a monochrome image sensor for hyperspectral and fluorescence imaging and topology laser mapping
US11758256B2 (en) 2019-06-20 2023-09-12 Cilag Gmbh International Fluorescence imaging in a light deficient environment
US11122968B2 (en) 2019-06-20 2021-09-21 Cilag Gmbh International Optical fiber waveguide in an endoscopic system for hyperspectral imaging
US10952619B2 (en) 2019-06-20 2021-03-23 Ethicon Llc Hyperspectral and fluorescence imaging and topology laser mapping with minimal area monolithic image sensor
US11589819B2 (en) 2019-06-20 2023-02-28 Cilag Gmbh International Offset illumination of a scene using multiple emitters in a laser mapping imaging system
US11412920B2 (en) 2019-06-20 2022-08-16 Cilag Gmbh International Speckle removal in a pulsed fluorescence imaging system
US11134832B2 (en) 2019-06-20 2021-10-05 Cilag Gmbh International Image rotation in an endoscopic hyperspectral, fluorescence, and laser mapping imaging system
US11622094B2 (en) 2019-06-20 2023-04-04 Cilag Gmbh International Wide dynamic range using a monochrome image sensor for fluorescence imaging
US12089902B2 (en) 2019-07-30 2024-09-17 Coviden Lp Cone beam and 3D fluoroscope lung navigation
US12053247B1 (en) 2020-12-04 2024-08-06 Onpoint Medical, Inc. System for multi-directional tracking of head mounted displays for real-time augmented reality guidance of surgical procedures
WO2022147074A1 (en) * 2020-12-30 2022-07-07 Intuitive Surgical Operations, Inc. Systems and methods for tracking objects crossing body wall for operations associated with a computer-assisted system
US11786206B2 (en) 2021-03-10 2023-10-17 Onpoint Medical, Inc. Augmented reality guidance for imaging systems
US11903656B2 (en) * 2021-09-24 2024-02-20 Biosense Webster (Israel) Ltd. Automatic control and enhancement of 4D ultrasound images
US12004821B2 (en) 2022-02-03 2024-06-11 Medtronic Navigation, Inc. Systems, methods, and devices for generating a hybrid image
EP4338698A1 (en) * 2022-09-13 2024-03-20 Caranx Medical SAS Medical image processing apparatus for generating a dynamic image of a patient

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030011624A1 (en) * 2001-07-13 2003-01-16 Randy Ellis Deformable transformations for interventional guidance
US20040097806A1 (en) * 2002-11-19 2004-05-20 Mark Hunter Navigation system for cardiac therapies

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6405072B1 (en) * 1991-01-28 2002-06-11 Sherwood Services Ag Apparatus and method for determining a location of an anatomical target with reference to a medical apparatus
US6167296A (en) * 1996-06-28 2000-12-26 The Board Of Trustees Of The Leland Stanford Junior University Method for volumetric image navigation

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030011624A1 (en) * 2001-07-13 2003-01-16 Randy Ellis Deformable transformations for interventional guidance
US20040097806A1 (en) * 2002-11-19 2004-05-20 Mark Hunter Navigation system for cardiac therapies

Cited By (92)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9101397B2 (en) 1999-04-07 2015-08-11 Intuitive Surgical Operations, Inc. Real-time generation of three-dimensional ultrasound image using a two-dimensional ultrasound transducer in a robotic system
US10271909B2 (en) 1999-04-07 2019-04-30 Intuitive Surgical Operations, Inc. Display of computer generated image of an out-of-view portion of a medical device adjacent a real-time image of an in-view portion of the medical device
US10433919B2 (en) 1999-04-07 2019-10-08 Intuitive Surgical Operations, Inc. Non-force reflecting method for providing tool force information to a user of a telesurgical system
US9232984B2 (en) 1999-04-07 2016-01-12 Intuitive Surgical Operations, Inc. Real-time generation of three-dimensional ultrasound image using a two-dimensional ultrasound transducer in a robotic system
US9956049B2 (en) 1999-05-18 2018-05-01 Mediguide Ltd. Method and apparatus for invasive device tracking using organ timing signal generated from MPS sensors
US9833167B2 (en) 1999-05-18 2017-12-05 Mediguide Ltd. Method and system for superimposing virtual anatomical landmarks on an image
US10856769B2 (en) 1999-05-18 2020-12-08 St. Jude Medical International Holding S.àr.l. Method and system for superimposing virtual anatomical landmarks on an image
US7697973B2 (en) 1999-05-18 2010-04-13 MediGuide, Ltd. Medical imaging and navigation system
US7778688B2 (en) 1999-05-18 2010-08-17 MediGuide, Ltd. System and method for delivering a stent to a selected position within a lumen
US7840252B2 (en) 1999-05-18 2010-11-23 MediGuide, Ltd. Method and system for determining a three dimensional representation of a tubular organ
US10251712B2 (en) 1999-05-18 2019-04-09 Mediguide Ltd. Method and apparatus for invasive device tracking using organ timing signal generated from MPS sensors
US10363017B2 (en) 2001-09-07 2019-07-30 St. Jude Medical International Holding S.À R.L. System and method for delivering a stent to a selected position within a lumen
US11399909B2 (en) 2005-06-06 2022-08-02 Intuitive Surgical Operations, Inc. Laparoscopic ultrasound robotic surgical system
US10646293B2 (en) 2005-06-06 2020-05-12 Intuitive Surgical Operations, Inc. Laparoscopic ultrasound robotic surgical system
US11717365B2 (en) 2005-06-06 2023-08-08 Intuitive Surgical Operations, Inc. Laparoscopic ultrasound robotic surgical system
US12108998B2 (en) 2005-06-06 2024-10-08 Intuitive Surgical Operations, Inc. Laparoscopic ultrasound robotic surgical system
JP2012050888A (en) * 2005-06-06 2012-03-15 Intuitive Surgical Inc Laparoscopic ultrasound robotic surgical system
US9795446B2 (en) 2005-06-06 2017-10-24 Intuitive Surgical Operations, Inc. Systems and methods for interactive user interfaces for robotic minimally invasive surgical systems
US11259870B2 (en) 2005-06-06 2022-03-01 Intuitive Surgical Operations, Inc. Interactive user interfaces for minimally invasive telesurgical systems
US10603127B2 (en) 2005-06-06 2020-03-31 Intuitive Surgical Operations, Inc. Laparoscopic ultrasound robotic surgical system
EP1779779A3 (en) * 2005-09-16 2007-09-12 Mediguide Ltd. Method and system for delivering a medical device to a selected position within a lumen
EP2080473A2 (en) 2005-09-16 2009-07-22 Mediguide Ltd. Method and system for delivering a medical device to a selected position within a lumen
EP2080473A3 (en) * 2005-09-16 2011-08-17 Mediguide Ltd. Method and system for delivering a medical device to a selected position within a lumen
EP1795130A1 (en) * 2005-12-05 2007-06-13 Medison Co., Ltd. Ultrasound system for interventional treatment
US9345387B2 (en) 2006-06-13 2016-05-24 Intuitive Surgical Operations, Inc. Preventing instrument/tissue collisions
US9801690B2 (en) 2006-06-29 2017-10-31 Intuitive Surgical Operations, Inc. Synthetic representation of a surgical instrument
US10737394B2 (en) 2006-06-29 2020-08-11 Intuitive Surgical Operations, Inc. Synthetic representation of a surgical robot
US10730187B2 (en) 2006-06-29 2020-08-04 Intuitive Surgical Operations, Inc. Tool position and identification indicator displayed in a boundary area of a computer display screen
US10773388B2 (en) 2006-06-29 2020-09-15 Intuitive Surgical Operations, Inc. Tool position and identification indicator displayed in a boundary area of a computer display screen
US9718190B2 (en) 2006-06-29 2017-08-01 Intuitive Surgical Operations, Inc. Tool position and identification indicator displayed in a boundary area of a computer display screen
US11865729B2 (en) 2006-06-29 2024-01-09 Intuitive Surgical Operations, Inc. Tool position and identification indicator displayed in a boundary area of a computer display screen
US9788909B2 (en) 2006-06-29 2017-10-17 Intuitive Surgical Operations, Inc Synthetic representation of a surgical instrument
US9789608B2 (en) 2006-06-29 2017-10-17 Intuitive Surgical Operations, Inc. Synthetic representation of a surgical robot
US10137575B2 (en) 2006-06-29 2018-11-27 Intuitive Surgical Operations, Inc. Synthetic representation of a surgical robot
US10008017B2 (en) 2006-06-29 2018-06-26 Intuitive Surgical Operations, Inc. Rendering tool information as graphic overlays on displayed images of tools
US11638999B2 (en) 2006-06-29 2023-05-02 Intuitive Surgical Operations, Inc. Synthetic representation of a surgical robot
WO2008006180A1 (en) * 2006-07-10 2008-01-17 Katholieke Universiteit Leuven Endoscopic vision system
US8911358B2 (en) 2006-07-10 2014-12-16 Katholieke Universiteit Leuven Endoscopic vision system
US7938777B2 (en) 2006-07-21 2011-05-10 Orthosoft Inc. Non-invasive tracking of bones for surgery
US8152726B2 (en) 2006-07-21 2012-04-10 Orthosoft Inc. Non-invasive tracking of bones for surgery
EP1953564A3 (en) * 2007-01-31 2008-08-13 Biosense Webster, Inc. Correlation of ultrasound images and gated position measurements
US7735349B2 (en) 2007-01-31 2010-06-15 Biosense Websters, Inc. Correlation of ultrasound images and gated position measurements
US8303502B2 (en) 2007-03-06 2012-11-06 General Electric Company Method and apparatus for tracking points in an ultrasound image
US10271912B2 (en) 2007-06-13 2019-04-30 Intuitive Surgical Operations, Inc. Method and system for moving a plurality of articulated instruments in tandem back towards an entry guide
US9629520B2 (en) 2007-06-13 2017-04-25 Intuitive Surgical Operations, Inc. Method and system for moving an articulated instrument back towards an entry guide while automatically reconfiguring the articulated instrument for retraction into the entry guide
US12097002B2 (en) 2007-06-13 2024-09-24 Intuitive Surgical Operations, Inc. Medical robotic system with coupled control modes
US10188472B2 (en) 2007-06-13 2019-01-29 Intuitive Surgical Operations, Inc. Medical robotic system with coupled control modes
US11751955B2 (en) 2007-06-13 2023-09-12 Intuitive Surgical Operations, Inc. Method and system for retracting an instrument into an entry guide
US9138129B2 (en) 2007-06-13 2015-09-22 Intuitive Surgical Operations, Inc. Method and system for moving a plurality of articulated instruments in tandem back towards an entry guide
US9901408B2 (en) 2007-06-13 2018-02-27 Intuitive Surgical Operations, Inc. Preventing instrument/tissue collisions
US11432888B2 (en) 2007-06-13 2022-09-06 Intuitive Surgical Operations, Inc. Method and system for moving a plurality of articulated instruments in tandem back towards an entry guide
US11399908B2 (en) 2007-06-13 2022-08-02 Intuitive Surgical Operations, Inc. Medical robotic system with coupled control modes
US9333042B2 (en) 2007-06-13 2016-05-10 Intuitive Surgical Operations, Inc. Medical robotic system with coupled control modes
US9469034B2 (en) 2007-06-13 2016-10-18 Intuitive Surgical Operations, Inc. Method and system for switching modes of a robotic system
US10695136B2 (en) 2007-06-13 2020-06-30 Intuitive Surgical Operations, Inc. Preventing instrument/tissue collisions
US9717563B2 (en) 2008-06-27 2017-08-01 Intuitive Surgical Operations, Inc. Medical robotic system providing an auxilary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide
US10368952B2 (en) 2008-06-27 2019-08-06 Intuitive Surgical Operations, Inc. Medical robotic system providing an auxiliary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide
US10258425B2 (en) 2008-06-27 2019-04-16 Intuitive Surgical Operations, Inc. Medical robotic system providing an auxiliary view of articulatable instruments extending out of a distal end of an entry guide
US11638622B2 (en) 2008-06-27 2023-05-02 Intuitive Surgical Operations, Inc. Medical robotic system providing an auxiliary view of articulatable instruments extending out of a distal end of an entry guide
US11382702B2 (en) 2008-06-27 2022-07-12 Intuitive Surgical Operations, Inc. Medical robotic system providing an auxiliary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide
US9516996B2 (en) 2008-06-27 2016-12-13 Intuitive Surgical Operations, Inc. Medical robotic system providing computer generated auxiliary views of a camera instrument for controlling the position and orienting of its tip
EP2147636A1 (en) 2008-07-24 2010-01-27 Esaote S.p.A. Device and method for guiding surgical tools by ultrasonic imaging
ITGE20080064A1 (en) * 2008-07-24 2010-01-25 Esaote Spa DEVICE AND METHOD OF GUIDANCE OF SURGICAL UTENSILS BY ECOGRAPHIC IMAGING.
US10492758B2 (en) 2008-07-24 2019-12-03 Esaote, S.P.A. Device and method for guiding surgical tools
US11941734B2 (en) 2009-03-31 2024-03-26 Intuitive Surgical Operations, Inc. Rendering tool information as graphic overlays on displayed images of tools
US10984567B2 (en) 2009-03-31 2021-04-20 Intuitive Surgical Operations, Inc. Rendering tool information as graphic overlays on displayed images of tools
US10282881B2 (en) 2009-03-31 2019-05-07 Intuitive Surgical Operations, Inc. Rendering tool information as graphic overlays on displayed images of tools
US9492927B2 (en) 2009-08-15 2016-11-15 Intuitive Surgical Operations, Inc. Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose
US10772689B2 (en) 2009-08-15 2020-09-15 Intuitive Surgical Operations, Inc. Controller assisted reconfiguration of an articulated instrument during movement into and out of an entry guide
US10271915B2 (en) 2009-08-15 2019-04-30 Intuitive Surgical Operations, Inc. Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose
US10959798B2 (en) 2009-08-15 2021-03-30 Intuitive Surgical Operations, Inc. Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose
US9956044B2 (en) 2009-08-15 2018-05-01 Intuitive Surgical Operations, Inc. Controller assisted reconfiguration of an articulated instrument during movement into and out of an entry guide
US11596490B2 (en) 2009-08-15 2023-03-07 Intuitive Surgical Operations, Inc. Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose
US10828774B2 (en) 2010-02-12 2020-11-10 Intuitive Surgical Operations, Inc. Medical robotic system providing sensory feedback indicating a difference between a commanded state and a preferred pose of an articulated instrument
US10537994B2 (en) 2010-02-12 2020-01-21 Intuitive Surgical Operations, Inc. Medical robotic system providing sensory feedback indicating a difference between a commanded state and a preferred pose of an articulated instrument
US9622826B2 (en) 2010-02-12 2017-04-18 Intuitive Surgical Operations, Inc. Medical robotic system providing sensory feedback indicating a difference between a commanded state and a preferred pose of an articulated instrument
US9872651B2 (en) 2011-03-22 2018-01-23 Kuka Roboter Gmbh Medical workstation
EP2502558A1 (en) * 2011-03-22 2012-09-26 KUKA Laboratories GmbH Medical workstation
US10758209B2 (en) 2012-03-09 2020-09-01 The Johns Hopkins University Photoacoustic tracking and registration in interventional ultrasound
US11389255B2 (en) 2013-02-15 2022-07-19 Intuitive Surgical Operations, Inc. Providing information of tools by filtering image areas adjacent to or on displayed images of the tools
US11806102B2 (en) 2013-02-15 2023-11-07 Intuitive Surgical Operations, Inc. Providing information of tools by filtering image areas adjacent to or on displayed images of the tools
US10507066B2 (en) 2013-02-15 2019-12-17 Intuitive Surgical Operations, Inc. Providing information of tools by filtering image areas adjacent to or on displayed images of the tools
US11357574B2 (en) 2013-10-31 2022-06-14 Intersect ENT International GmbH Surgical instrument and method for detecting the position of a surgical instrument
WO2015158736A1 (en) * 2014-04-15 2015-10-22 Fiagon Ag Medical Technologies Navigation assistance system for medical instruments
US10568713B2 (en) 2014-04-15 2020-02-25 Fiagon Ag Medical Technologies Navigation assistance system for medical instruments
US10806346B2 (en) 2015-02-09 2020-10-20 The Johns Hopkins University Photoacoustic tracking and registration in interventional ultrasound
US11446090B2 (en) 2017-04-07 2022-09-20 Orthosoft Ulc Non-invasive system and method for tracking bones
US11986250B2 (en) 2017-04-07 2024-05-21 Orthosoft Ulc Non-invasive system and method for tracking bones
US11701188B2 (en) 2017-05-10 2023-07-18 Mako Surgical Corp. Robotic spine surgery system and methods
US12035985B2 (en) 2017-05-10 2024-07-16 Mako Surgical Corp. Robotic spine surgery system and methods
US11684426B2 (en) 2018-08-31 2023-06-27 Orthosoft Ulc System and method for tracking bones
US11430139B2 (en) 2019-04-03 2022-08-30 Intersect ENT International GmbH Registration method and setup

Also Published As

Publication number Publication date
US20070225553A1 (en) 2007-09-27
WO2005043319A3 (en) 2005-12-22
JP2007531553A (en) 2007-11-08
JP2007508913A (en) 2007-04-12
WO2005043319A2 (en) 2005-05-12
EP1680024A2 (en) 2006-07-19
US20070276234A1 (en) 2007-11-29
WO2005039391A3 (en) 2005-12-22
EP1689290A2 (en) 2006-08-16

Similar Documents

Publication Publication Date Title
US20070276234A1 (en) Systems and Methods for Intraoperative Targeting
US20050085718A1 (en) Systems and methods for intraoperative targetting
US20050085717A1 (en) Systems and methods for intraoperative targetting
JP7561233B2 (en) Systems and methods for holographic image guidance in non-vascular percutaneous procedures
CN106890025B (en) Minimally invasive surgery navigation system and navigation method
US6850794B2 (en) Endoscopic targeting method and system
US6019724A (en) Method for ultrasound guidance during clinical procedures
EP1103229B1 (en) System and method for use with imaging devices to facilitate planning of interventional procedures
US11026747B2 (en) Endoscopic view of invasive procedures in narrow passages
US10543045B2 (en) System and method for providing a contour video with a 3D surface in a medical navigation system
US9237929B2 (en) System for guiding a medical instrument in a patient body
US8414476B2 (en) Method for using variable direction of view endoscopy in conjunction with image guided surgical systems
US20080234570A1 (en) System For Guiding a Medical Instrument in a Patient Body
WO1996025881A1 (en) Method for ultrasound guidance during clinical procedures
CN101862205A (en) Intraoperative tissue tracking method combined with preoperative image
WO2008036050A2 (en) Methods and systems for providing accuracy evaluation of image guided surgery
Shahidi et al. Volumetric image guidance via a stereotactic endoscope
Lange et al. Development of navigation systems for image-guided laparoscopic tumor resections in liver surgery
Akatsuka et al. Navigation system for neurosurgery with PC platform
Giraldez et al. Multimodal augmented reality system for surgical microscopy
Nakajima et al. Enhanced video image guidance for biopsy using the safety map

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DPEN Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed from 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2006536816

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 2004796074

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 2004796074

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 10576632

Country of ref document: US

WWP Wipo information: published in national office

Ref document number: 10576632

Country of ref document: US