EP0501993B1 - Sondenkorrelierte anzeige von anatomischen bilddaten - Google Patents
Sondenkorrelierte anzeige von anatomischen bilddaten Download PDFInfo
- Publication number
- EP0501993B1 EP0501993B1 EP90916676A EP90916676A EP0501993B1 EP 0501993 B1 EP0501993 B1 EP 0501993B1 EP 90916676 A EP90916676 A EP 90916676A EP 90916676 A EP90916676 A EP 90916676A EP 0501993 B1 EP0501993 B1 EP 0501993B1
- Authority
- EP
- European Patent Office
- Prior art keywords
- probe
- data
- base
- spatial position
- anatomical
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Detecting, measuring or recording devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/11—Measuring movement of the entire body or parts thereof, e.g. head or hand tremor, mobility of a limb
- A61B5/1113—Local tracking of patients, e.g. in a hospital or private home
- A61B5/1114—Tracking parts of the body
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B2090/364—Correlation of different images or relation of image positions in respect to the body
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B2090/364—Correlation of different images or relation of image positions in respect to the body
- A61B2090/365—Correlation of different images or relation of image positions in respect to the body augmented reality, i.e. correlating a live optical image with another image
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/683—Means for maintaining contact with the body
- A61B5/6835—Supports or holders, e.g., articulated arms
Definitions
- the invention relates to a method and a system for visualizing internal regions of an anatomical body according to the preambles of claims 1 and 20. More specifically, the invention relates to a method and a system for determining the position of a probe relative to various anatomical features and displaying the internal anatomical structures corresponding to the position of the probe.
- a process and device for the optical representation of surgical operations whereat a coordinate measuring device is provided for detecting the position of the surgical instrument.
- Tomograms provided with at least three measurement points on the patient are stored in a computer and the position of three measurement points and that of the surgical instrument are determined. These positions are superimposed on corresponding tomographs on the screen.
- the purpose is the documenting of the performance and outcome of a surgical operation, in hidden parts of the body, the instantaneous position of the surgical position or its course are represented during the operation and recorded for subsequent examination.
- tomograms provided with at least three measurement points on the patient are stored in a computer and can be displayed on a screen.
- the described process and device are used in the representation and documentation of surgical operations.
- a stereotactic method and apparatus for treating a region of a patient's body is known defining points in the region using a three-dimensional coordinate system with reference to a ring attached to the patient for establishing a reference point for the three-dimensional coordinate system at the center of the ring.
- the ring and the reference point are used for stereotactically controlling instruments used to treat the region.
- the ring is provided with pins extending parallel to the axis of the ring, and equidistant from the center, for precise location of the center and of a base-line scan for correlation between the location of the region to be treated and the control system for treatment of the region.
- a series of no-ninvasive tomography scans are made through the region and at least part of the pins for determining the coordinate of at least one point of the region selected for the treatment with respect to the center of the ring. All parameters of this described system for stereotactic control of the instruments used for treatment are determined with respect to the ring center as a reference point.
- a process and apparatus particularly suited for guiding neurosurgical operations is described.
- the described process for guiding neurosurgical operations allows a considerable reducton of surgical traumatism during the removal of cerebral lesions by virtue of the possibility of converting into three-dimensional images both the contours of the anatomical structures and the representation of the device for the stereotactic detection of the affected region together with a stereotactic probe which defines a surgical path.
- the apparatus requires a stereotactic detection device being a stereotactic helmet fitted on a patient.
- the images are used to plan the course of a medical procedure, be it diagnostic, therapeutic, or surgical, and for orientation during the procedure.
- the slice images are typically generated by Computerized Tomography (CT) or by Magnetic Resonance Imaging (MRI). Images may also be captured using Angiography, Single-Photon Emission Computed Tomography, and Positron Emission Tomography methods.
- the images typically presented to a user consist of a series of static images on film. These images are very detailed and can resolve anatomical structures less than one millimetre in size. However, their format differs greatly from the actual anatomical features seen during the surgical procedure.
- the images are presented in two-dimensional form rather than in the three-dimensional form of the anatomical features.
- the perspective of the slice image rarely corersponds to the surgeon's viewing angle during the procedure. Consequently, during a procedure, the slice images provide a primitive visualization aid to the patient's anatomy.
- surgeons can make an incision which is larger than the minimum required for the planned procedure. While providing an enlarged window to the patient's anatomy, these larger incisions may result in longer hospital stays and increased risk for the patient. On the other hand, if only a small incision is made, the field of view available to the surgeon is greatly limited. As a result, the surgeon may become disoriented forcing him to correct and recommence the procedure, or to continue at a high risk to the patient.
- Imaging equipment can be used to provide on-the-spot visualization of a patient, it is impractical to use the equipment in the operating room during the procedure. First, the costs of purchasing, operating and maintaining the imaging equipment are prohibitive. Secondly, surgeons have limited access to a patient who is placed in a scanning device. Furthermore, Magnetic Resonance Imaging and Computerized Tomography have side effects which may harm the patient and inhibit the procedures. Magnetic Resonance Imaging produces a very high fixed magnetic field which precludes the use of many instruments. Computerized Tomography, on the other hand, utilizes X-ray radiation which is known to damage human tissue and cause cancer. It is, therefore, not desirable to expose a patient to a computerized tomography scan for a prolonged period.
- a known approach to localizing anatomy during surgery is currently being used for brain lesions.
- the method is known as stereotactic surgery. It involves rigidly attaching the reference frame to the patient's head during the scanning. Using the marks left in the scanned images by the frame, the location of the lesion is computed. During the surgical procedure, a reference frame is again attached to the same location on the patient's head. The frame is used to direct drilling and cutting operations, which are done either manually or automatically.
- Stereotactic surgery has a number of drawbacks. Firstly, it is only suitable for localized brain lesions which have a direct approach path. Secondly, stereotactic surgery requires the use of a cumbersome and uncomfortable reference frame. Furthermore, since the decision to undertake stereotactic surgery is usually done after a first scanning procedure, the patient must undergo a second scan with the frame attached. This results in a prolonged and expensive procedure. Moreover, if the scan utilizes computerized tomography imaging, then the patient is exposed to another dose of radiation.
- a first embodiment of the invention provides a method for visualizing internal regions of an anatomical body in relation to a probe, employing a data-base body of previously acquired images of the anatomical body, the method comprising the steps of:
- the invention provides a system for visualizing internal regions of an anatomical body by utilizing a data-base body of previously acquired images of the anatomical body, the system comprising:
- a probe-correlated system (1) has a probe (10), a computer (11), a data storage unit (12), and a display (13). These components are, individually, well known and common.
- the system (1) is employed to view the anatomical structure of a patient (9) adjacent to the position of the probe (10).
- the computer (11) has ready access to the unit (12) which contains a data-base body (17) representing the anatomical structure of the patient (9).
- the data-base body (17) includes previously acquired digital images (15) of the patient (9). These images (15) can be acquired through various medical-imaging techniques, such as Computerized Tomography, Single-Photon Emission Computed Tomography, Positron Emission Tomography, Magnetic Resonance Imaging, Ultrasound, or Angiography.
- the data-base body (17) can contain pre-processed digital images (16).
- the digital images (15) together with their relative spatial relationship can be pre-processed to represent the various organ surfaces of the patient (9).
- the known system places the pre-processed images (16) in the data-base body (17).
- the probe (10), or any other object which may function as a probe is used by an operator, not shown, to point to a particular location on the anatomical body of the patient (9). The operator can move the probe (10) around or within the anatomical body of the patient (9).
- Spatial coordinates representing the spatial position and possibly the spatial orientation, of the probe relative to a fixed reference point, shown generally at the arrow (20), are conveyed to the computer (11).
- the reference point (20) may either be on the patient (9) as shown, or on some stable platform nearby, not shown.
- the apparatuses described in association with such method will be collectively referred to as spatial determinators.
- an electro-magnetic emitter (20a) is positioned at the reference point (20) and a sensor (20b) is located on the probe (10).
- a sensor (20b) is located on the probe (10).
- the position and orientation of the probe (10) relative to the reference point (20) can be determined.
- a probe (10) using this known locating method is commercially available.
- the computer (11) Given the spatial relationship between the reference point (20) and the patient (9), the computer (11) can determine the position of the probe (10) relative to the patient (9).
- the probe (10) to a multi-joint light-weight arm (25) with a first section (26) and a second section (27) connected together at joint (22).
- the first section (26) of the multi-joint arm (25) is connected to a base (28) at joint (21).
- the base (28) is attached to the patient (9) using adhesive elastic tape (23).
- the probe (10) is attached to the second section (27) at joint (24).
- the joints (21), (22), (24), in combination, provide for a range of motion equal to or greater than that required for a given procedure.
- Angular sensors are located at the joints (21), (22), (24).
- the angular sensors are connected by wire (28a) to one another and to an electronic unit (29).
- the sensors detect any change in the position or orientation of the multi-joint arm (25), and convey this information to the electronic unit (29).
- the unit (29) uses geometric calculations to determine the spatial position and spatial orientation of the probe (10) relative to the base (28) which is used as the reference point.
- the spatial position and spatial orientation of the probe (10) are sent to the computer (11) of fig. 1 through an electronic communication link (27).
- a suitable communication link (27) would be an RS-232 serial communication interface. Since the base (28) is fixed to the body of the patient (9), the computer can use the spatial information to determine the position of the probe (10) relative to the patient (9).
- a dual-arm arrangement shown generally at (31), may be employed.
- the arrangement (31) is particularly effective where the multi-joint arm (30) of fig. 2 cannot be fixed to the patient (9).
- a stand (35) is used to anchor two multi-joint arms (36, 37) similar to the multi-joint arm (30) of fig. 2.
- a probe (10) is attached to the other end of arm (37).
- Arm (36) is attached at its other end to a reference point (40) on the patient (9).
- Sensors are mounted at joints (41, 42, 43) of arm (37), and at joints (44, 45,46) of arm (36).
- the sensors are connected to an electronic unit (39).
- the electronic unit (39) decodes the position and orientation of the probe (10).
- the spatial position and orientation of the probe (10) relative to the patient (9) is obtained.
- the spatial position and orientation of the probe (10) is transmitted to the computer (11) of fig. 1 via the communication link (47).
- the reference arm (36) shown in fig. 3 can be omitted if the patient (9) is fixed to an operating table (48).
- the patient can be fixed to the table (48) using straps (49).
- the reference point (40) can be fixed arbitrarily in space.
- the relative position of the reference point (40) to the joint (41) may be determined once and the relative position of the probe (10) to the reference point (40) determined therefrom. However, if the patient (9) is moved during the procedure, a new reference point (40) or a new spatial relationship must be established.
- mapping is a procedure for determining the current spatial position of the probe (10) and the corresponding adjacent data-base body (17) location.
- This correspondence may be initially determined through a procedure which maps the patient (9) to the data-base body (17). This procedure is known as "registration" since its purpose is to register the correspondence between the anatomical body of the patient (9) and the data-base body (17) with the computer (11).
- a number of registration methods are known. For example, one method involves the marking of reference points on the patent (9). However, this can be inconvenient and there is a risk that the marked positions on the patient (9) may be erased between the time the scan images (15) were generated and the time the surgical procedure is performed.
- Another method involves placing small markers, usually made of cad or ceramic material, on readily identifiable features of the patent, such as the ears or the corners of the eyes.
- the preferred registration method involves using the probe (10) to register with the computer (11) the spatial position of easily identifiable features of the patient, such as the space between the teeth, the nose or the corners of the eyes.
- the previously acquired scan images (15) or the pre-processed images (16) are displayed on the display (13) in such a manner as to allow the user of the system (1) to identify specific points of the chosen features of the patient (9).
- a three dimensional surface format shown in figure 6, is the simplest such format for an unskilled viewer to comprehend.
- Such a three-dimensional surface format can be derived from the pre-processed images (16) in a known manner, and suitable points such as the corners of the eyes (70), space between the teeth (72) are shown in figure 6.
- the method is as follows.
- the probe (10) is placed next to the feature point on the patient (9).
- the spatial position of the probe (10) is then determined.
- a movable marker, e.g. a cursor, on the display (13) is then adjusted so it coincides with a selected feature, e.g. corner of the eyes (70) as seen.
- a proper and unique transformation function can be calculated which maps the spatial position of the probe (10) to the corresponding data-base body location and orientation.
- the accuracy of this transformation function is improved by the use of a larger number of points and a statistical error minimizing techniques, such as the least mean square error method.
- the operator can move the probe (10) in and around the patient (9), and at the same time view the hidden anatomical features of the patient (9) as they appear in the data-base body (17).
- the anatomical features of the patient (9) in the data-base body (17) are presented on the display unit (13) in relationship to the spatial position and possibly orientation of the probe (10).
- the probe (10) may be represented on the display (13) as a point rather than a full probe (10).
- the region adjacent the point probe (10) is then displayed.
- the orientation of the regions displayed is known from the computer (11) and not determined by the orientation of the probe (10).
- a possible presentation format for the data-base images (15) of the patient (9) is shown in fig. 4.
- Two-dimensional representations or slice images are generated by the computer (11) from the data-base images (15).
- the position of the probe (10) relative to the anatomical body (9) is marked on a slice image (50) by the computer (11).
- the slice image (50) together with the probe (52) are displayed on the unit (13).
- the screen of the display unit (13) is divided into 4 separate windows. Three of the windows contain slice images corresponding to three cardinal anatomical planes: sagittal (50); axial (54); and coronal (56). The three slice images (50, 54, 56) intersect at the location of the probe (52). Thus, the operator can observe the anatomical feature of the patient (9) relative to the position of the probe (10) in the six main directions: anterior, posterior, superior, inferior, right and left.
- the fourth window depicted on the display unit (13) can show a slice (57) through the anatomical features in mid-sagittal orientation along the axis of the probe (10). The position and orientation of the probe (10) can be marked on the slice (57), thereby allowing the operator to direct viewing of what lies ahead of the probe (10).
- a three-dimensional model (58) of the patient (9) is generated by the computer (11) from the images (15, 16).
- the computer (11) also generates a three-dimensional model (60) of the probe (10).
- the relative locations of the models (60), (58) correspond to the spatial position and orientation of the probe (10) relative to the patient (9).
- the three-dimensional model (58) of the patient (9) generated from the stored images (15, 16) is presented together with the model (60) on the display unit (13).
- the computer (11) can generate displays directly from the images (15) using a ray-cast method.
- the computer (11) creates the display using the results of simulated X-rays passing through the images (15).
- the simulated X-rays will be affected differently by different elements in the images (15) according to their relative absorption of the X-rays.
- the results may be displayed along with the probe (10) in a manner similar to those described for slices or 3d-images. This produces a simulated X-ray display.
- a display is created using the results of simulated light rays passing through the images (15).
- the elements in the images (15) which do not pass the simulated light rays correspond to surface features and may be used to generate a display similar to the three-dimensional model (58).
- the computer (11) can be used to further process the slice image (50) and three-dimensional images (58) generated from the data-base body (17). For example, a wedge-shaped portion (62) has been cut from the three-dimensional image (58). The cut-out portion (62) exposes various structures adjacent to the probe (10), which would not otherwise be observable. In addition, the cut-out portion (62) gives the operator an unobstructed view of the position of the probe (10) even if it is within the patient (9).
- the slice images (50) and three-dimensional images (58), (60) can also be processed by the computer (11) using other known image processing techniques. For example, the model (60) of the probe (10) can be made translucent, or the slice image (50) can be combined with other slice views.
Claims (26)
- Verfahren zum Sichtbarmachen innerer Bereiche eines anatomischen Körpers (9) in Bezug auf eine Sonde unter Verwendung eines Datenbankkörpers (17) zuvor aufgenommener Bilder des anatomischen Körpers (9), welches folgende Schritte aufweist:a) Einrichten einer räumlichen Position für die Sonde (10) relativ zum anatomischen Körper (9);b) Bestimmen eines Datenbankorts relativ zum Datenbankkörper (17) entsprechend der räumlichen Position der Sonde (10) relativ zum anatomischen Körper (9);c) Abbilden und Registrieren der räumlichen Position der Sonde (10) relativ zum anatomischen Körper (9) auf dem entsprechenden Datenbankort der Sonde (10) relativ zum Datenbankkörper (17); undd) Anzeigen eines Bereichs des Datenbankkörpers (17) neben dem Datenbankort der Sonde (10), wobei der Bereich von einer Vielzahl benachbarter Bilder des Datenbankkörpers (17) abgeleitet wird,gekennzeichnet durch
den Schritt des Erfassens der Bewegung von sowohl der Sonde (10) als auch des anatomischen Körpers (9) zum Erhalten der räumlichen Position für die Sonde (10) relativ zum anatomischen Körper (9), derart daß die Sonde (10) und der anatomische Körper (9) unabhängig verschoben werden, und derart, daß die Registrierung zwischen dem Datenbankkörper (17) und dem anatomischen Körper (9) aufrechterhalten wird. - Verfahren nach Anspruch 1,
dadurch gekennzeichnet,
daß das Anzeigen eines Bereichs des Datenbankkörpers (17) neben dem Datenbankort der Sonde (10) folgende Schritte aufweist:a) Erzeugen eines Schnittbildes aus den zuvor aufgenommenen Bildern und Schneiden einer Vielzahl der benachbarten Bilder, darstellend eines Bereichs des Datenbankkörpers (17) neben einem Datenbankort der Sonde (10); undb) Anzeigen des Schnittbilds. - Verfahren nach Anspruch 1,
dadurch gekennzeichnet,
daß das Anzeigen eines Bereichs des Datenbankkörpers (17) neben dem Datenbankort der Sonde (10) folgende Schritte aufweist:a) Erzeugen eines dreidimensionalen Körpermodells aus dem zuvor aufgenommenen Bildern, darstellend einen Bereich des Datenbankkörpers neben dem Datenbankort der Sonde; undb) Anzeigen des dreidimensionalen Körpermodells. - Verfahren nach Anspruch 1,
dadurch gekennzeichnet,
daß das Anzeigen eines Bereichs des Datenbankkörpers neben dem Datenbankort der Sonde folgende Schritte aufweist:a) Erzeugen eines dreidimensionalen Körpermodells aus zuvor aufgenommenen Bildern, welche zur Abbildung anatomischer Merkmale vorverarbeitet sind und welche einen Bereich des Datenbankkörpers neben dem Datenbankort der Sonde darstellen; undb) Anzeigen des dreidimensionalen Körpermodells. - Verfahren nach Anspruch 4,
dadurch gekennzeichnet,
daß in Schritt (a) ein Bereich des dreidimensionalen Körpermodells entfernt wird, um einen Ort entsprechend dem Ort der Sonde (10) freizulegen. - Verfahren nach Anspruch 1,
dadurch gekennzeichnet,
daß das Anzeigen eines Bereichs des Datenbankkörpers (17) neben dem Datenbankort der Sonde (10) folgende Schritte aufweist:a) Erzeugen eines Anzeigeformats unter Benutzung eines Strahlenabbildungsverfahrens auf zuvor aufgenommenen Bildern zum Darstellen eines Bereichs des Datenbankkörpers neben dem Datenbankort der Sonde; undb) Anzeigen des Anzeigeformats. - Verfahren nach Anspruch 6,
dadurch gekennzeichnet,
daß das Strahlenabbildungsverfahren aus der Gruppe bestehend aus Röntgenstrahlung und Lichtstrahlung ausgewählt wird. - Verfahren nach Anspruch 6 oder 7,
dadurch gekennzeichnet,
daß das Anzeigeformat ein dreidimensionales Format ist. - Verfahren nach Anspruch 1, 2 oder 3,
dadurch gekennzeichnet,
daß die räumliche Orientierung der Sonde (10) zusammen mit ihrer räumlichen Position erhalten wird. - Verfahren nach Anspruch 1, 2 oder 3,
dadurch gekennzeichnet,
daß eine Darstellung der Sonde zusammen mit dem Bereich des Datenbankkörpers (17) neben dem Datenbankort der Sonde (10) angezeigt wird, und daß die relativen Orte der Dar stellung der Sonde (10) und des Datenbankkörpers (17) der räumlichen Position der Sonde (10) relativ zum anatomischen Körper (9) entsprechen. - Verfahren nach Anspruch 5,
dadurch gekennzeichnet,
daß eine Darstellung der Sonde (10) mit dem dreidimensionalen Körpermodell angezeigt wird, wobei der relative Ort der Darstellung der Sonde (10) zum dreidimensionalen Körpermodell der räumlichen Position der Sonde (10) relativ zum anatomischen Körper (9) entspricht. - Verfahren nach Anspruch 11,
dadurch gekennzeichnet,
daß die Darstellung der Sonde (10) genau der tatsächlichen Sonde entspricht, wobei die Darstellung der Sonde zusätzlich so orientiert ist, daß sie der Orientierung der Sonde bezüglich des anatomischen Körpers (9) entspricht, und wobei die Perspektive der Darstellung der Sonde (10) und des dreidimensionalen Körpermodells einander entsprechen. - Verfahren nach Anspruch 1,
gekennzeichnet durch den Schritt der Registrierung vor dem Erhalten der räumlichen Position, wobei die Registrierung folgende Schritte aufweist:a) Positionieren der Sonde (10) neben einem bestimmten Merkmal des anatomischen Körpers (9);b) Bestimmen einer räumlichen Position der Sonde (10);c) Anzeigen eines Bereichs des Datenbankkörpers (17) mit einem Datenbankkörpermerkmal entsprechend dem bestimmten Merkmal;d) Identifizieren des bestimmten Merkmals auf dem angezeigten Bereich; unde) Registrieren der räumlichen Position der Sonde und des Ortes auf dem Datenbankkörper entsprechend der Position des bestimmten Merkmals;wodurch ein Datenbankort bestimmt wird, der einer räumlichen Position der Sonde (10) entspricht. - Verfahren nach Anspruch 1,
gekennzeichnet durch den Schritt der Registrierung vor dem Erhalten der räumlichen Position, wobei die Registrierung folgende Schritte aufweist:a) Markieren von Orten in dem Datenbankkörper (17), welche bestimmten Merkmalen des anatomischen Körpers (9) entsprechen;b) Positionieren der Sonde (10) nahe einem bestimmten Merkmal des anatomischen Körpers (9);c) Bestimmen der räumlichen Position der Sonde (10);d) Registrieren der räumlichen Position der Sonde (10) und ihres entsprechenden Datenbankkörperorts;wodurch ein Datenbankkörperort bestimmt wird, der mit einer räumlichen Position der Sonde (10) übereinstimmt. - Verfahren nach Anspruch 1,
gekennzeichnet durch den Schritt der Registrierung vor dem Erhalten der räumlichen Position, wobei die Registrierung folgende Schritte aufweist:a) Markieren einer Position auf dem anatomischen Körper (9) eines bestimmten gescannten Bildes mit einem entsprechenden Ort im Datenbankkörper (17);b) Positionieren der Sonde (10) nahe der markierten Position auf dem anatomischen Körper (9);c) Bestimmen der räumlichen Position der Sonde (10);d) Registrieren der räumlichen Position der Sonde (10) und ihres entsprechenden Datenbankkörperorts;wodurch ein Datenbankkörperort bestimmt wird, der mit einer räumlichen Position der Sonde (10) übereinstimmt. - Verfahren nach Anspruch 13,
dadurch gekennzeichnet,
daß die Anzeige von Schritt (c) eine dreidimensionale Anzeige ist. - Verfahren nach Anspruch 13 oder 16.
dadurch gekennzeichnet,
daß mehr als drei Datenbankorte identifiziert werden und daß Fehler zwischen den entsprechenden Datenbankorten und den räumlichen Positionen minimalisiert werden, um die Genauigkeit des Registrierungsschritts zu verbessern. - Verfahren nach Anspruch 17,
dadurch gekennzeichnet,
daß die Fehler unter Benutzung einer Analyse der kleinsten Fehlerquadrate minimalisiert werden. - Verfahren nach Anspruch 13,
dadurch gekennzeichnet,
daß die Sonde (10) derart mit dem anatomischen Körper (9) verbunden ist, daß falls der anatomische Körper (9) nach der Registrierung verschoben wird, die Registrierung zwischen dem Datenbankkörper (17) und dem anatomischen Körper (9) aufrechterhalten wird. - System zum Sichtbarmachen innerer Bereiche eines anatomischen Körpers unter Benutzung eines Datenbankkörpers (17) zuvor aufgenommener Bilder des anatomischen Körpers (9), welches aufweist:eine Sonde (10);eine Datenbankspeichereinheit (12) mit den zuvor aufgenommenen Bildern des anatomischen Körpers (9);einen räumlichen Determinator (20a, 20b, 36, 37, 39) zum Bestimmen der räumlichen Position der Sonde (10) relativ zum anatomischen Körper (9);einen Computer (11), welcher die zuvor aufgenommenen Bilder benutzt, um eine Darstellung eines Bereichs des anatomischen Körpers (9) neben der räumlichen Position der Sonde (10) zu erzeugen;eine Einrichtung (11) zum Abbilden und Registrieren der räumlichen Position der Sonde relativ zum anatomischen Körper (9) auf dem entsprechenden Datenbankort der Sonde (10) relativ zum Datenbankkörper (9);eine Anzeigeeinheit (13) zum Anzeigen der Darstellung des anatomischen Körpers,dadurch gekennzeichnet,
daß
die Sonde (10) und der anatomische Körper (9) jeweils eine Einrichtung (20a, 20b, 26, 27) aufweisen, welche es ermöglicht, daß der Determinator die räumliche Position der Sonde (10) relativ zum anatomischen Körper (9) derart bestimmt, daß die Sonde (10) und der anatomische Körper (9) unabhängig verschiebbar sind, und derart, daß die Registrierung zwischen dem Datenbankkörper (17) und dem anatomischen Körper (9) aufrechterhalten bleibt. - System nach Anspruch 20,
dadurch gekennzeichnet,
daß die erzeugten Darstellungen in einem dreidimensionalen Oberflächenformat angezeigt werden. - System nach Anspruch 20,
dadurch gekennzeichnet,
daß der Computer (11) derart gestaltet ist, daß er für den Ort in der Datenbankspeichereinheit entsprechend der räumlichen Position der Sonde (10) so initialisiert wird, daß er die Sonde neben einem bestimmten Merkmalspunkt des anatomischen Körpers (9) positioniert hat, daß eine räumliche Position der Sonde (10) bestimmt wird, daß ein Bereich des Datenbankkörpers (17) mit einem Datenbankkörpermerkmal entsprechend dem bestimmten Merkmal auf dem angezeigten Bereich angezeigt wird und daß die räumliche Position der Sonde (10) und der Ort auf dem Datenbankkörper (17) entsprechend der Position des bestimmten Merkmals registriert werden. - System nach Anspruch 22,
dadurch gekennzeichnet,
daß die erzeugten Bilder in einem dreidimensionalen Format während der Registrierung verschoben werden, und daß die bestimmten Merkmale auf dem dreidimensionalen Format identifiziert werden. - System nach Anspruch 20,
dadurch gekennzeichnet,
daß der räumliche Determinator (20a, 20b, 36, 37, 39) aufweist:a) einen elektromagnetischen Emitter auf einem Referenzpunkt zum Übertragen eines Signales;b) einen Sensor auf der Sonde (10) zum Empfangen des Signals; undc) eine Einrichtung zum Vergleichen des übertragenen Signals mit dem empfangenden Signal zur Bestimmung der Position der Sonde (10). - System nach Anspruch 20,
dadurch gekennzeichnet,
daß der räumliche Determinator (20a, 20b, 36, 37, 39) aufweist:a) erste, zweite, dritte und vierte Abschnitte;b) einen Ständer;c) ein erstes Gelenk zwischen dem ersten Abschnitt und der Sonde;d) ein zweites Gelenk zwischen dem ersten und zweiten Abschnitt;e) ein drittes Gelenk zwischen dem zweiten Abschnitt und dem Ständer;f) ein viertes Gelenk zwischen dem Ständer und dem dritten Abschnitt;g) ein fünftes Gelenk zwischen dem dritten Abschnitt und dem vierten Abschnitt;h) ein sechstes Gelenk zwischen dem vierten Abschnitt und einem Referenzpunkt, dessen räumliche Position relativ zum anatomischen Körper bekannt ist;e) Sensoren, welche an den Gelenken angeordnet sind; undj) eine Einrichtung, die mit den Sensoren verbunden ist, zum Bestimmen der Position der Sonde relativ zum anatomischen Körper. - System nach Anspruch 20,
dadurch gekennzeichnet,
daß der räumliche Determinator (20, 20b, 36, 36, 39) die räumliche Orientierung der Sonde (10) sowie ihre räumliche Position bestimmt.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA3497 | 1989-11-21 | ||
CA002003497A CA2003497C (en) | 1989-11-21 | 1989-11-21 | Probe-correlated viewing of anatomical image data |
PCT/CA1990/000404 WO1991007726A1 (en) | 1989-11-21 | 1990-11-21 | Probe-correlated viewing of anatomical image data |
Publications (2)
Publication Number | Publication Date |
---|---|
EP0501993A1 EP0501993A1 (de) | 1992-09-09 |
EP0501993B1 true EP0501993B1 (de) | 1997-06-11 |
Family
ID=4143598
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP90916676A Expired - Lifetime EP0501993B1 (de) | 1989-11-21 | 1990-11-21 | Sondenkorrelierte anzeige von anatomischen bilddaten |
Country Status (6)
Country | Link |
---|---|
EP (1) | EP0501993B1 (de) |
JP (2) | JP3367663B2 (de) |
AU (1) | AU6726990A (de) |
CA (2) | CA2003497C (de) |
DE (1) | DE69030926T2 (de) |
WO (1) | WO1991007726A1 (de) |
Cited By (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6021343A (en) * | 1997-11-20 | 2000-02-01 | Surgical Navigation Technologies | Image guided awl/tap/screwdriver |
US6167145A (en) | 1996-03-29 | 2000-12-26 | Surgical Navigation Technologies, Inc. | Bone navigation system |
US6226418B1 (en) | 1997-11-07 | 2001-05-01 | Washington University | Rapid convolution based large deformation image matching via landmark and volume imagery |
US6236875B1 (en) | 1994-10-07 | 2001-05-22 | Surgical Navigation Technologies | Surgical navigation systems including reference and localization frames |
US6347240B1 (en) | 1990-10-19 | 2002-02-12 | St. Louis University | System and method for use in displaying images of a body part |
US6370224B1 (en) | 1998-06-29 | 2002-04-09 | Sofamor Danek Group, Inc. | System and methods for the reduction and elimination of image artifacts in the calibration of x-ray imagers |
US6381485B1 (en) | 1999-10-28 | 2002-04-30 | Surgical Navigation Technologies, Inc. | Registration of human anatomy integrated for electromagnetic localization |
US6402762B2 (en) | 1999-10-28 | 2002-06-11 | Surgical Navigation Technologies, Inc. | System for translation of electromagnetic and optical localization systems |
US6408107B1 (en) | 1996-07-10 | 2002-06-18 | Michael I. Miller | Rapid convolution based large deformation image matching via landmark and volume imagery |
US6491702B2 (en) | 1992-04-21 | 2002-12-10 | Sofamor Danek Holdings, Inc. | Apparatus and method for photogrammetric surgical localization |
US6535756B1 (en) | 2000-04-07 | 2003-03-18 | Surgical Navigation Technologies, Inc. | Trajectory storage apparatus and method for surgical navigation system |
US6633686B1 (en) | 1998-11-05 | 2003-10-14 | Washington University | Method and apparatus for image registration using large deformation diffeomorphisms on a sphere |
US6725080B2 (en) | 2000-03-01 | 2004-04-20 | Surgical Navigation Technologies, Inc. | Multiple cannula image guided tool for image guided procedures |
USRE39133E1 (en) * | 1997-09-24 | 2006-06-13 | Surgical Navigation Technologies, Inc. | Percutaneous registration apparatus and method for use in computer-assisted surgical navigation |
US7853307B2 (en) | 2003-08-11 | 2010-12-14 | Veran Medical Technologies, Inc. | Methods, apparatuses, and systems useful in conducting image guided interventions |
US7920909B2 (en) | 2005-09-13 | 2011-04-05 | Veran Medical Technologies, Inc. | Apparatus and method for automatic image guided accuracy verification |
US8046053B2 (en) | 1994-10-07 | 2011-10-25 | Foley Kevin T | System and method for modifying images of a body part |
US8150495B2 (en) | 2003-08-11 | 2012-04-03 | Veran Medical Technologies, Inc. | Bodily sealants and methods and apparatus for image-guided delivery of same |
US8473026B2 (en) | 1994-09-15 | 2013-06-25 | Ge Medical Systems Global Technology Company | System for monitoring a position of a medical instrument with respect to a patient's body |
WO2014036389A1 (en) * | 2012-08-30 | 2014-03-06 | The Regents Of The University Of Michigan | Analytic morphomics: high speed medical image automated analysis method |
US8696549B2 (en) | 2010-08-20 | 2014-04-15 | Veran Medical Technologies, Inc. | Apparatus and method for four dimensional soft tissue navigation in endoscopic applications |
US8781186B2 (en) | 2010-05-04 | 2014-07-15 | Pathfinder Therapeutics, Inc. | System and method for abdominal surface matching using pseudo-features |
US8838199B2 (en) | 2002-04-04 | 2014-09-16 | Medtronic Navigation, Inc. | Method and apparatus for virtual digital subtraction angiography |
US9138165B2 (en) | 2012-02-22 | 2015-09-22 | Veran Medical Technologies, Inc. | Systems, methods and devices for forming respiratory-gated point cloud for four dimensional soft tissue navigation |
US9402691B2 (en) | 2014-09-16 | 2016-08-02 | X-Nav Technologies, LLC | System for determining and tracking movement during a medical procedure |
US9844324B2 (en) | 2013-03-14 | 2017-12-19 | X-Nav Technologies, LLC | Image guided navigation system |
US9943374B2 (en) | 2014-09-16 | 2018-04-17 | X-Nav Technologies, LLC | Image guidance system for detecting and tracking an image pose |
Families Citing this family (95)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5251127A (en) * | 1988-02-01 | 1993-10-05 | Faro Medical Technologies Inc. | Computer-aided surgery apparatus |
FR2652928B1 (fr) | 1989-10-05 | 1994-07-29 | Diadix Sa | Systeme interactif d'intervention locale a l'interieur d'une zone d'une structure non homogene. |
DE69133603D1 (de) * | 1990-10-19 | 2008-10-02 | Univ St Louis | System zur Lokalisierung einer chirurgischen Sonde relativ zum Kopf |
EP0699050B1 (de) * | 1993-04-26 | 2004-03-03 | St. Louis University | Anzeige der lage einer sonde |
DE9403971U1 (de) * | 1994-03-09 | 1994-05-26 | Vierte Art Gmbh Computer Anima | Gerät zur Messung von Bewegungen im Gesicht eines Menschen |
US5617857A (en) * | 1995-06-06 | 1997-04-08 | Image Guided Technologies, Inc. | Imaging system having interactive medical instruments and methods |
US5592939A (en) | 1995-06-14 | 1997-01-14 | Martinelli; Michael A. | Method and system for navigating a catheter probe |
US6256529B1 (en) * | 1995-07-26 | 2001-07-03 | Burdette Medical Systems, Inc. | Virtual reality 3D visualization for surgical procedures |
US5806518A (en) * | 1995-09-11 | 1998-09-15 | Integrated Surgical Systems | Method and system for positioning surgical robot |
US5824085A (en) | 1996-09-30 | 1998-10-20 | Integrated Surgical Systems, Inc. | System and method for cavity generation for surgical planning and initial placement of a bone prosthesis |
US5776136A (en) | 1996-09-30 | 1998-07-07 | Integrated Surgical Systems, Inc. | Method and system for finish cutting bone cavities |
US6016439A (en) * | 1996-10-15 | 2000-01-18 | Biosense, Inc. | Method and apparatus for synthetic viewpoint imaging |
EP1016030A1 (de) | 1997-02-13 | 2000-07-05 | Integrated Surgical Systems, Inc. | Verfahren und system zur registrierung von der position eines chirurgischen systems mit einem preoperativen knochenbild |
WO1999000052A1 (en) * | 1997-06-27 | 1999-01-07 | The Board Of Trustees Of The Leland Stanford Junior University | Method and apparatus for volumetric image navigation |
US6348058B1 (en) | 1997-12-12 | 2002-02-19 | Surgical Navigation Technologies, Inc. | Image guided spinal surgery guide, system, and method for use thereof |
US6482182B1 (en) | 1998-09-03 | 2002-11-19 | Surgical Navigation Technologies, Inc. | Anchoring system for a brain lead |
US6033415A (en) * | 1998-09-14 | 2000-03-07 | Integrated Surgical Systems | System and method for performing image directed robotic orthopaedic procedures without a fiducial reference system |
DE19841951C2 (de) * | 1998-09-14 | 2002-08-29 | Storz Medical Ag Kreuzlingen | Verfahren zur Visualisierung der Ausrichtung von therapeutischen Schallwellen auf einen zu behandelnden bzw. zu bearbeitenden Bereich |
DE19842239A1 (de) | 1998-09-15 | 2000-03-16 | Siemens Ag | Medizintechnische Einrichtung |
US6322567B1 (en) | 1998-12-14 | 2001-11-27 | Integrated Surgical Systems, Inc. | Bone motion tracking system |
US6430434B1 (en) | 1998-12-14 | 2002-08-06 | Integrated Surgical Systems, Inc. | Method for determining the location and orientation of a bone for computer-assisted orthopedic procedures using intraoperatively attached markers |
JP2002533143A (ja) | 1998-12-23 | 2002-10-08 | ヤカブ、ピーター・ディー | 電磁式位置及び向き追跡装置を備える磁気共鳴スキャナ |
US6491699B1 (en) | 1999-04-20 | 2002-12-10 | Surgical Navigation Technologies, Inc. | Instrument guidance method and system for image guided surgery |
DE10040498A1 (de) | 1999-09-07 | 2001-03-15 | Zeiss Carl Fa | Vorrichtung zur bildgestützten Bearbeitung eines Arbeitsobjekts |
US6385476B1 (en) | 1999-09-21 | 2002-05-07 | Biosense, Inc. | Method and apparatus for intracardially surveying a condition of a chamber of a heart |
US6368285B1 (en) | 1999-09-21 | 2002-04-09 | Biosense, Inc. | Method and apparatus for mapping a chamber of a heart |
US6546271B1 (en) | 1999-10-01 | 2003-04-08 | Bioscience, Inc. | Vascular reconstruction |
US11331150B2 (en) | 1999-10-28 | 2022-05-17 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation |
US8644907B2 (en) | 1999-10-28 | 2014-02-04 | Medtronic Navigaton, Inc. | Method and apparatus for surgical navigation |
US6837892B2 (en) * | 2000-07-24 | 2005-01-04 | Mazor Surgical Technologies Ltd. | Miniature bone-mounted surgical robot |
DE10037491A1 (de) | 2000-08-01 | 2002-02-14 | Stryker Leibinger Gmbh & Co Kg | Verfahren zum dreidimensionalen Visualisieren von Strukturen im Körperinneren |
US6650927B1 (en) | 2000-08-18 | 2003-11-18 | Biosense, Inc. | Rendering of diagnostic imaging data on a three-dimensional map |
US6633773B1 (en) | 2000-09-29 | 2003-10-14 | Biosene, Inc. | Area of interest reconstruction for surface of an organ using location data |
US6636757B1 (en) | 2001-06-04 | 2003-10-21 | Surgical Navigation Technologies, Inc. | Method and apparatus for electromagnetic navigation of a surgical probe near a metal object |
US7286866B2 (en) | 2001-11-05 | 2007-10-23 | Ge Medical Systems Global Technology Company, Llc | Method, system and computer product for cardiac interventional procedure planning |
US6947786B2 (en) | 2002-02-28 | 2005-09-20 | Surgical Navigation Technologies, Inc. | Method and apparatus for perspective inversion |
US7499743B2 (en) | 2002-03-15 | 2009-03-03 | General Electric Company | Method and system for registration of 3D images within an interventional system |
US7346381B2 (en) | 2002-11-01 | 2008-03-18 | Ge Medical Systems Global Technology Company Llc | Method and apparatus for medical intervention procedure planning |
US7998062B2 (en) | 2004-03-29 | 2011-08-16 | Superdimension, Ltd. | Endoscope structures and techniques for navigating to a target in branched structure |
US7778686B2 (en) | 2002-06-04 | 2010-08-17 | General Electric Company | Method and apparatus for medical intervention procedure planning and location and navigation of an intervention tool |
US6978167B2 (en) | 2002-07-01 | 2005-12-20 | Claron Technology Inc. | Video pose tracking system and method |
US7166114B2 (en) | 2002-09-18 | 2007-01-23 | Stryker Leibinger Gmbh & Co Kg | Method and system for calibrating a surgical tool and adapter thereof |
US7660623B2 (en) | 2003-01-30 | 2010-02-09 | Medtronic Navigation, Inc. | Six degree of freedom alignment display for medical procedures |
US7747047B2 (en) | 2003-05-07 | 2010-06-29 | Ge Medical Systems Global Technology Company, Llc | Cardiac CT system and method for planning left atrial appendage isolation |
US7343196B2 (en) | 2003-05-09 | 2008-03-11 | Ge Medical Systems Global Technology Company Llc | Cardiac CT system and method for planning and treatment of biventricular pacing using epicardial lead |
US7565190B2 (en) | 2003-05-09 | 2009-07-21 | Ge Medical Systems Global Technology Company, Llc | Cardiac CT system and method for planning atrial fibrillation intervention |
US7813785B2 (en) | 2003-07-01 | 2010-10-12 | General Electric Company | Cardiac imaging system and method for planning minimally invasive direct coronary artery bypass surgery |
US7344543B2 (en) | 2003-07-01 | 2008-03-18 | Medtronic, Inc. | Method and apparatus for epicardial left atrial appendage isolation in patients with atrial fibrillation |
ES2432616T3 (es) | 2003-09-15 | 2013-12-04 | Covidien Lp | Sistema de accesorios para su uso con broncoscopios |
EP2316328B1 (de) | 2003-09-15 | 2012-05-09 | Super Dimension Ltd. | Umhüllungsvorrichtung zur Fixierung von Bronchoskopen |
US7308299B2 (en) | 2003-10-22 | 2007-12-11 | General Electric Company | Method, apparatus and product for acquiring cardiac images |
US7308297B2 (en) | 2003-11-05 | 2007-12-11 | Ge Medical Systems Global Technology Company, Llc | Cardiac imaging system and method for quantification of desynchrony of ventricles for biventricular pacing |
US7873400B2 (en) | 2003-12-10 | 2011-01-18 | Stryker Leibinger Gmbh & Co. Kg. | Adapter for surgical navigation trackers |
US7771436B2 (en) | 2003-12-10 | 2010-08-10 | Stryker Leibinger Gmbh & Co. Kg. | Surgical navigation tracker, system and method |
US7454248B2 (en) | 2004-01-30 | 2008-11-18 | Ge Medical Systems Global Technology, Llc | Method, apparatus and product for acquiring cardiac images |
US8764725B2 (en) | 2004-02-09 | 2014-07-01 | Covidien Lp | Directional anchoring mechanism, method and applications thereof |
JP4615893B2 (ja) * | 2004-05-17 | 2011-01-19 | ジーイー・メディカル・システムズ・グローバル・テクノロジー・カンパニー・エルエルシー | 超音波撮像装置 |
US8515527B2 (en) | 2004-10-13 | 2013-08-20 | General Electric Company | Method and apparatus for registering 3D models of anatomical regions of a heart and a tracking system with projection images of an interventional fluoroscopic system |
US7327872B2 (en) | 2004-10-13 | 2008-02-05 | General Electric Company | Method and system for registering 3D models of anatomical regions with projection images of the same |
WO2007033206A2 (en) | 2005-09-13 | 2007-03-22 | Veran Medical Technologies, Inc. | Apparatus and method for image guided accuracy verification |
US9168102B2 (en) | 2006-01-18 | 2015-10-27 | Medtronic Navigation, Inc. | Method and apparatus for providing a container to a sterile environment |
JP2007244494A (ja) * | 2006-03-14 | 2007-09-27 | J Morita Tokyo Mfg Corp | 歯科診断用oct装置 |
US8660635B2 (en) | 2006-09-29 | 2014-02-25 | Medtronic, Inc. | Method and apparatus for optimizing a computer assisted surgical procedure |
JP5137033B2 (ja) * | 2007-01-31 | 2013-02-06 | 国立大学法人浜松医科大学 | 手術支援情報表示装置、手術支援情報表示方法及び手術支援情報表示プログラム |
US8989842B2 (en) | 2007-05-16 | 2015-03-24 | General Electric Company | System and method to register a tracking system with intracardiac echocardiography (ICE) imaging system |
US8905920B2 (en) | 2007-09-27 | 2014-12-09 | Covidien Lp | Bronchoscope adapter and method |
WO2009122273A2 (en) | 2008-04-03 | 2009-10-08 | Superdimension, Ltd. | Magnetic interference detection system and method |
US8473032B2 (en) | 2008-06-03 | 2013-06-25 | Superdimension, Ltd. | Feature-based registration method |
US8218847B2 (en) | 2008-06-06 | 2012-07-10 | Superdimension, Ltd. | Hybrid registration method |
US8932207B2 (en) | 2008-07-10 | 2015-01-13 | Covidien Lp | Integrated multi-functional endoscopic tool |
JP5569711B2 (ja) | 2009-03-01 | 2014-08-13 | 国立大学法人浜松医科大学 | 手術支援システム |
US8611984B2 (en) | 2009-04-08 | 2013-12-17 | Covidien Lp | Locatable catheter |
WO2011159834A1 (en) | 2010-06-15 | 2011-12-22 | Superdimension, Ltd. | Locatable expandable working channel and method |
US8435033B2 (en) | 2010-07-19 | 2013-05-07 | Rainbow Medical Ltd. | Dental navigation techniques |
JP5657467B2 (ja) * | 2011-05-13 | 2015-01-21 | オリンパスメディカルシステムズ株式会社 | 医療用画像表示システム |
KR20140104502A (ko) * | 2011-12-21 | 2014-08-28 | 메드로보틱스 코포레이션 | 링크를 구비한 고도의 관절 프로브용 안정화 장치, 그 제조 방법 및 그 사용 방법 |
US9545288B2 (en) | 2013-03-14 | 2017-01-17 | Think Surgical, Inc. | Systems and devices for a counter balanced surgical robot |
WO2014152694A1 (en) | 2013-03-14 | 2014-09-25 | Curexo Technology Corporation | Systems and methods for monitoring a surgical procedure with critical regions |
US20150305650A1 (en) | 2014-04-23 | 2015-10-29 | Mark Hunter | Apparatuses and methods for endobronchial navigation to and confirmation of the location of a target tissue and percutaneous interception of the target tissue |
US20150305612A1 (en) | 2014-04-23 | 2015-10-29 | Mark Hunter | Apparatuses and methods for registering a real-time image feed from an imaging device to a steerable catheter |
US10952593B2 (en) | 2014-06-10 | 2021-03-23 | Covidien Lp | Bronchoscope adapter |
EP3229722B1 (de) | 2014-12-08 | 2023-03-22 | Think Surgical, Inc. | Implantatbasierte planung, digitalisierung und registrierung zur totalen gelenkarthroplastie |
US10932866B1 (en) | 2014-12-08 | 2021-03-02 | Think Surgical, Inc. | Implant based planning, digitizing, and registration for total joint arthroplasty |
US10426555B2 (en) | 2015-06-03 | 2019-10-01 | Covidien Lp | Medical instrument with sensor for use in a system and method for electromagnetic navigation |
US9962134B2 (en) | 2015-10-28 | 2018-05-08 | Medtronic Navigation, Inc. | Apparatus and method for maintaining image quality while minimizing X-ray dosage of a patient |
US10478254B2 (en) | 2016-05-16 | 2019-11-19 | Covidien Lp | System and method to access lung tissue |
US10751126B2 (en) | 2016-10-28 | 2020-08-25 | Covidien Lp | System and method for generating a map for electromagnetic navigation |
US10638952B2 (en) | 2016-10-28 | 2020-05-05 | Covidien Lp | Methods, systems, and computer-readable media for calibrating an electromagnetic navigation system |
US10615500B2 (en) | 2016-10-28 | 2020-04-07 | Covidien Lp | System and method for designing electromagnetic navigation antenna assemblies |
US10418705B2 (en) | 2016-10-28 | 2019-09-17 | Covidien Lp | Electromagnetic navigation antenna assembly and electromagnetic navigation system including the same |
US10722311B2 (en) | 2016-10-28 | 2020-07-28 | Covidien Lp | System and method for identifying a location and/or an orientation of an electromagnetic sensor based on a map |
US10446931B2 (en) | 2016-10-28 | 2019-10-15 | Covidien Lp | Electromagnetic navigation antenna assembly and electromagnetic navigation system including the same |
US10517505B2 (en) | 2016-10-28 | 2019-12-31 | Covidien Lp | Systems, methods, and computer-readable media for optimizing an electromagnetic navigation system |
US10792106B2 (en) | 2016-10-28 | 2020-10-06 | Covidien Lp | System for calibrating an electromagnetic navigation system |
US11219489B2 (en) | 2017-10-31 | 2022-01-11 | Covidien Lp | Devices and systems for providing sensors in parallel with medical tools |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4638798A (en) * | 1980-09-10 | 1987-01-27 | Shelden C Hunter | Stereotactic method and apparatus for locating and treating or removing lesions |
IT1227365B (it) * | 1988-11-18 | 1991-04-08 | Istituto Neurologico Carlo Bes | Procedimento ed apparecchiatura particolarmente per la guida di opera zioni neurochirurgiche |
-
1989
- 1989-11-21 CA CA002003497A patent/CA2003497C/en not_active Expired - Fee Related
- 1989-11-21 CA CA002260688A patent/CA2260688A1/en not_active Abandoned
-
1990
- 1990-11-21 EP EP90916676A patent/EP0501993B1/de not_active Expired - Lifetime
- 1990-11-21 DE DE69030926T patent/DE69030926T2/de not_active Expired - Fee Related
- 1990-11-21 WO PCT/CA1990/000404 patent/WO1991007726A1/en active IP Right Grant
- 1990-11-21 AU AU67269/90A patent/AU6726990A/en not_active Abandoned
- 1990-11-21 JP JP51527590A patent/JP3367663B2/ja not_active Expired - Fee Related
-
2002
- 2002-08-08 JP JP2002231280A patent/JP2003159247A/ja active Pending
Cited By (45)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6347240B1 (en) | 1990-10-19 | 2002-02-12 | St. Louis University | System and method for use in displaying images of a body part |
US6490467B1 (en) | 1990-10-19 | 2002-12-03 | Surgical Navigation Technologies, Inc. | Surgical navigation systems including reference and localization frames |
US6434415B1 (en) | 1990-10-19 | 2002-08-13 | St. Louis University | System for use in displaying images of a body part |
US6491702B2 (en) | 1992-04-21 | 2002-12-10 | Sofamor Danek Holdings, Inc. | Apparatus and method for photogrammetric surgical localization |
US8473026B2 (en) | 1994-09-15 | 2013-06-25 | Ge Medical Systems Global Technology Company | System for monitoring a position of a medical instrument with respect to a patient's body |
US6236875B1 (en) | 1994-10-07 | 2001-05-22 | Surgical Navigation Technologies | Surgical navigation systems including reference and localization frames |
US8046053B2 (en) | 1994-10-07 | 2011-10-25 | Foley Kevin T | System and method for modifying images of a body part |
US6167145A (en) | 1996-03-29 | 2000-12-26 | Surgical Navigation Technologies, Inc. | Bone navigation system |
US6408107B1 (en) | 1996-07-10 | 2002-06-18 | Michael I. Miller | Rapid convolution based large deformation image matching via landmark and volume imagery |
USRE45509E1 (en) | 1997-09-24 | 2015-05-05 | Medtronic Navigation, Inc. | Percutaneous registration apparatus and method for use in computer-assisted surgical navigation |
USRE39133E1 (en) * | 1997-09-24 | 2006-06-13 | Surgical Navigation Technologies, Inc. | Percutaneous registration apparatus and method for use in computer-assisted surgical navigation |
USRE44305E1 (en) | 1997-09-24 | 2013-06-18 | Medtronic Navigation, Inc. | Percutaneous registration apparatus and method for use in computer-assisted surgical navigation |
USRE42226E1 (en) * | 1997-09-24 | 2011-03-15 | Medtronic Navigation, Inc. | Percutaneous registration apparatus and method for use in computer-assisted surgical navigation |
USRE42194E1 (en) | 1997-09-24 | 2011-03-01 | Medtronic Navigation, Inc. | Percutaneous registration apparatus and method for use in computer-assisted surgical navigation |
US6226418B1 (en) | 1997-11-07 | 2001-05-01 | Washington University | Rapid convolution based large deformation image matching via landmark and volume imagery |
USRE45484E1 (en) | 1997-11-20 | 2015-04-21 | Medtronic Navigation, Inc. | Image guided awl/tap/screwdriver |
USRE43328E1 (en) | 1997-11-20 | 2012-04-24 | Medtronic Navigation, Inc | Image guided awl/tap/screwdriver |
USRE46422E1 (en) | 1997-11-20 | 2017-06-06 | Medtronic Navigation, Inc. | Image guided awl/tap/screwdriver |
US6021343A (en) * | 1997-11-20 | 2000-02-01 | Surgical Navigation Technologies | Image guided awl/tap/screwdriver |
USRE46409E1 (en) | 1997-11-20 | 2017-05-23 | Medtronic Navigation, Inc. | Image guided awl/tap/screwdriver |
US6370224B1 (en) | 1998-06-29 | 2002-04-09 | Sofamor Danek Group, Inc. | System and methods for the reduction and elimination of image artifacts in the calibration of x-ray imagers |
US6633686B1 (en) | 1998-11-05 | 2003-10-14 | Washington University | Method and apparatus for image registration using large deformation diffeomorphisms on a sphere |
US7657300B2 (en) | 1999-10-28 | 2010-02-02 | Medtronic Navigation, Inc. | Registration of human anatomy integrated for electromagnetic localization |
US6381485B1 (en) | 1999-10-28 | 2002-04-30 | Surgical Navigation Technologies, Inc. | Registration of human anatomy integrated for electromagnetic localization |
US6402762B2 (en) | 1999-10-28 | 2002-06-11 | Surgical Navigation Technologies, Inc. | System for translation of electromagnetic and optical localization systems |
US7881770B2 (en) | 2000-03-01 | 2011-02-01 | Medtronic Navigation, Inc. | Multiple cannula image guided tool for image guided procedures |
US6725080B2 (en) | 2000-03-01 | 2004-04-20 | Surgical Navigation Technologies, Inc. | Multiple cannula image guided tool for image guided procedures |
US6535756B1 (en) | 2000-04-07 | 2003-03-18 | Surgical Navigation Technologies, Inc. | Trajectory storage apparatus and method for surgical navigation system |
US7853305B2 (en) | 2000-04-07 | 2010-12-14 | Medtronic Navigation, Inc. | Trajectory storage apparatus and method for surgical navigation systems |
US8838199B2 (en) | 2002-04-04 | 2014-09-16 | Medtronic Navigation, Inc. | Method and apparatus for virtual digital subtraction angiography |
US8150495B2 (en) | 2003-08-11 | 2012-04-03 | Veran Medical Technologies, Inc. | Bodily sealants and methods and apparatus for image-guided delivery of same |
US8483801B2 (en) | 2003-08-11 | 2013-07-09 | Veran Medical Technologies, Inc. | Methods, apparatuses, and systems useful in conducting image guided interventions |
US10470725B2 (en) | 2003-08-11 | 2019-11-12 | Veran Medical Technologies, Inc. | Method, apparatuses, and systems useful in conducting image guided interventions |
US7853307B2 (en) | 2003-08-11 | 2010-12-14 | Veran Medical Technologies, Inc. | Methods, apparatuses, and systems useful in conducting image guided interventions |
US9218664B2 (en) | 2005-09-13 | 2015-12-22 | Veran Medical Technologies, Inc. | Apparatus and method for image guided accuracy verification |
US9218663B2 (en) | 2005-09-13 | 2015-12-22 | Veran Medical Technologies, Inc. | Apparatus and method for automatic image guided accuracy verification |
US7920909B2 (en) | 2005-09-13 | 2011-04-05 | Veran Medical Technologies, Inc. | Apparatus and method for automatic image guided accuracy verification |
US10617332B2 (en) | 2005-09-13 | 2020-04-14 | Veran Medical Technologies, Inc. | Apparatus and method for image guided accuracy verification |
US8781186B2 (en) | 2010-05-04 | 2014-07-15 | Pathfinder Therapeutics, Inc. | System and method for abdominal surface matching using pseudo-features |
US8696549B2 (en) | 2010-08-20 | 2014-04-15 | Veran Medical Technologies, Inc. | Apparatus and method for four dimensional soft tissue navigation in endoscopic applications |
US9138165B2 (en) | 2012-02-22 | 2015-09-22 | Veran Medical Technologies, Inc. | Systems, methods and devices for forming respiratory-gated point cloud for four dimensional soft tissue navigation |
WO2014036389A1 (en) * | 2012-08-30 | 2014-03-06 | The Regents Of The University Of Michigan | Analytic morphomics: high speed medical image automated analysis method |
US9844324B2 (en) | 2013-03-14 | 2017-12-19 | X-Nav Technologies, LLC | Image guided navigation system |
US9402691B2 (en) | 2014-09-16 | 2016-08-02 | X-Nav Technologies, LLC | System for determining and tracking movement during a medical procedure |
US9943374B2 (en) | 2014-09-16 | 2018-04-17 | X-Nav Technologies, LLC | Image guidance system for detecting and tracking an image pose |
Also Published As
Publication number | Publication date |
---|---|
JP2003159247A (ja) | 2003-06-03 |
AU6726990A (en) | 1991-06-13 |
DE69030926D1 (de) | 1997-07-17 |
WO1991007726A1 (en) | 1991-05-30 |
JPH05504694A (ja) | 1993-07-22 |
CA2003497C (en) | 1999-04-06 |
CA2260688A1 (en) | 1991-05-21 |
CA2003497A1 (en) | 1991-05-21 |
DE69030926T2 (de) | 1997-09-18 |
JP3367663B2 (ja) | 2003-01-14 |
EP0501993A1 (de) | 1992-09-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP0501993B1 (de) | Sondenkorrelierte anzeige von anatomischen bilddaten | |
US6359959B1 (en) | System for determining target positions in the body observed in CT image data | |
US6259943B1 (en) | Frameless to frame-based registration system | |
JP2950340B2 (ja) | 三次元データ組の登録システムおよび登録方法 | |
US5787886A (en) | Magnetic field digitizer for stereotatic surgery | |
EP0997109B1 (de) | Anzeige der Lage einer chirurgischen Sonde | |
US5483961A (en) | Magnetic field digitizer for stereotactic surgery | |
US6064904A (en) | Frameless stereotactic CT scanner with virtual needle display for planning image guided interventional procedures | |
US6491702B2 (en) | Apparatus and method for photogrammetric surgical localization | |
US20040015176A1 (en) | Stereotactic localizer system with dental impression | |
JPH03168139A (ja) | 対話的イメージガイドの手術システム | |
JPH09507131A (ja) | 電算機支援顕微鏡外科手術のための設備と前記設備の使用法 | |
CA2377190A1 (en) | Ultrasonic guidance of target structures for medical procedures | |
JPH09173352A (ja) | 医用ナビゲーションシステム | |
Adams et al. | An optical navigator for brain surgery | |
US20230130653A1 (en) | Apparatus and method for positioning a patient's body and tracking the patient's position during surgery | |
KR20230059157A (ko) | 라이브 및 스캔 이미지 등록 장치 및 방법 | |
Van Geems | The development of a simple stereotactic device for neurosurgical applications |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 19920622 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): DE FR GB IT |
|
17Q | First examination report despatched |
Effective date: 19931110 |
|
GRAG | Despatch of communication of intention to grant |
Free format text: ORIGINAL CODE: EPIDOS AGRA |
|
GRAH | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOS IGRA |
|
GRAH | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOS IGRA |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
ITF | It: translation for a ep patent filed |
Owner name: GUZZI E RAVIZZA S.R.L. |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): DE FR GB IT |
|
REF | Corresponds to: |
Ref document number: 69030926 Country of ref document: DE Date of ref document: 19970717 |
|
ET | Fr: translation filed | ||
PLBQ | Unpublished change to opponent data |
Free format text: ORIGINAL CODE: EPIDOS OPPO |
|
PLBI | Opposition filed |
Free format text: ORIGINAL CODE: 0009260 |
|
PLBF | Reply of patent proprietor to notice(s) of opposition |
Free format text: ORIGINAL CODE: EPIDOS OBSO |
|
26 | Opposition filed |
Opponent name: MEDIVISION AG Effective date: 19980311 Opponent name: SOFAMOR DANEK GROUP, INC. Effective date: 19980311 |
|
PLBF | Reply of patent proprietor to notice(s) of opposition |
Free format text: ORIGINAL CODE: EPIDOS OBSO |
|
PLBF | Reply of patent proprietor to notice(s) of opposition |
Free format text: ORIGINAL CODE: EPIDOS OBSO |
|
PLBO | Opposition rejected |
Free format text: ORIGINAL CODE: EPIDOS REJO |
|
APAC | Appeal dossier modified |
Free format text: ORIGINAL CODE: EPIDOS NOAPO |
|
APAE | Appeal reference modified |
Free format text: ORIGINAL CODE: EPIDOS REFNO |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: IF02 |
|
APAC | Appeal dossier modified |
Free format text: ORIGINAL CODE: EPIDOS NOAPO |
|
RAP2 | Party data changed (patent owner data changed or rights of a patent transferred) |
Owner name: BRAINLAB AG |
|
APAC | Appeal dossier modified |
Free format text: ORIGINAL CODE: EPIDOS NOAPO |
|
PLBN | Opposition rejected |
Free format text: ORIGINAL CODE: 0009273 |
|
PLBP | Opposition withdrawn |
Free format text: ORIGINAL CODE: 0009264 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: OPPOSITION REJECTED |
|
27O | Opposition rejected |
Effective date: 20030301 |
|
APAH | Appeal reference modified |
Free format text: ORIGINAL CODE: EPIDOSCREFNO |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20061124 Year of fee payment: 17 Ref country code: FR Payment date: 20061124 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20061127 Year of fee payment: 17 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: IT Payment date: 20061130 Year of fee payment: 17 |
|
GBPC | Gb: european patent ceased through non-payment of renewal fee |
Effective date: 20071121 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20080603 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: ST Effective date: 20080930 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20071121 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: FR Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20071130 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IT Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20071121 |