CA3031832A1 - Apparatus and method for tracking a volume in a three-dimensional space - Google Patents
Apparatus and method for tracking a volume in a three-dimensional spaceInfo
- Publication number
- CA3031832A1 CA3031832A1 CA3031832A CA3031832A CA3031832A1 CA 3031832 A1 CA3031832 A1 CA 3031832A1 CA 3031832 A CA3031832 A CA 3031832A CA 3031832 A CA3031832 A CA 3031832A CA 3031832 A1 CA3031832 A1 CA 3031832A1
- Authority
- CA
- Canada
- Prior art keywords
- interest
- volume
- dof
- orientation
- working space
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/101—Computer-aided simulation of surgical operations
- A61B2034/105—Modelling of the patient, e.g. for ligaments or bones
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2051—Electromagnetic tracking systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2055—Optical tracking systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/378—Surgical systems with images on a monitor during operation using ultrasound
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Engineering & Computer Science (AREA)
- Surgery (AREA)
- Biomedical Technology (AREA)
- Robotics (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Human Computer Interaction (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
Abstract
Description
62/623,082, filed on 29 January 2018, the contents of which are incorporated herein by reference in their entirety.
Field This invention relates to apparatus, techniques, and methods that provide tracking of a volume of interest in a three-dimensional space in real time.
Background As accuracy of tracking of objects in three-dimensional (3D) space improves, techniques for implementing the technology in different fields are emerging. A technique [1, 2] based on electromagnetic (EM) tracking that has shown good performance in a surgical application for tracking a tissue volume of interest, such as a tumour, relies on two EM
sensors, both reporting position and orientation in six degrees of freedom (6 DOF). According to this technique one 6 DOF sensor is located in or on the tissue volume of interest. In applications such as surgery, the 6 DOF sensor may be attached to a wire or needle that is inserted in the tissue volume of interest.
The second 6 DOF sensor is a reference sensor located within the working space, typically on the patient in the vicinity of the surgical procedure.
Potential drawbacks of this technique include the wire or needle spinning in place within the tissue volume of interest, causing the 6 DOF sensor to spin with it, thus reporting inaccurate orientation of the tissue volume of interest. In addition, because a 6 DOF
sensor is substantially large, it can be an obstacle in applications such as surgery, because it can obstruct the surgeon's hand motions in the working space, or the sensor may not fit inside the tissue volume of interest, or the sensor may obstruct the surgeon's view. Another drawback is that 6 DOF
sensors are more expensive than, e.g., 5 DOF sensors, and there is greater computational complexity when processing the 6 DOF sensor position and orientation.
Summary One aspect of the invention relates to an apparatus for tracking a volume of interest, comprising: a tracking device that tracks at least two sensors in three dimensions, wherein the at least two sensors include: a reference sensor that provides a 6 DOF coordinate system in a working space and a marker sensor adapted to be attached to the volume of interest and to provide at least position of the volume of interest in at least 3 DOF; a tracking device that receives sensor data from the reference sensor and the marker sensor; a processor that receives data from the tracking device and outputs a change in position and/or orientation of the volume of interest within the working space in real time, according to the at least 3 DOF of the marker sensor within the 6 DOF coordinate system.
In one embodiment the marker sensor provides position and orientation of the volume of interest in 5 DOF. In another embodiment the marker sensor provides position and orientation of the volume of interest in 3 DOF.
According to embodiments, the change in position and/or orientation of the volume of interest within the working space in real time may be determined relative to an initial position and orientation of the volume of interest within the working space.
In one embodiment, the at least two sensors are wireless. In one embodiment, the at least two sensors are electromagnetic (EM) or optical.
In one embodiment, the apparatus is part of a surgical or intervention navigation apparatus.
In one embodiment, the apparatus is part of a gaining apparatus, a robotic apparatus, a motion capture apparatus, or a training apparatus.
In one embodiment, the volume of interest is biological tissue.
In one embodiment, the apparatus is part of a surgical navigation apparatus, further comprising: a tool sensor adapted to be attached to a surgical tool and to provide position and orientation of the surgical tool; an imaging probe sensor adapted to be attached to an imaging probe of an imaging device and to provide position and orientation of the imaging probe;
wherein the processor receives images from the imaging device, generates a three-dimensional delineation of the volume of interest from data points in the images, and three-dimensionally
In one embodiment, the marker sensor is associated with a wire-localization needle.
In one embodiment, the volume of interest comprises a tumour.
In one embodiment, the working space comprises breast tissue.
In one embodiment, the surgical tool comprises a cutting tool, a cautery tool, a catheter, a needle, or a radiation therapy device.
Another aspect of the invention relates to programmed media for use with a processor, comprising: code stored on non-transitory storage media compatible with the processor, the code containing instructions to direct the processor to: communicate with a tracking device that tracks at least two sensors in three dimensions, wherein the at least two sensors include a reference sensor that provides a 6 DOF coordinate system in a working space, and a marker sensor adapted to be attached to the volume of interest and to provide at least position of the volume of interest in at least 3 DOF; wherein the processor outputs a change in position and/or orientation of the .. volume of interest within the working space in real time, according to the at least 3 DOF of the marker sensor relative to the 6 DOF coordinate system.
In one embodiment, the marker sensor provides position and orientation of the volume of interest in 5 DOF. In one embodiment, the marker sensor provides position and orientation of the volume of interest in 3 DOF.
In one embodiment, the change in position and/or orientation of the volume of interest within the working space in real time is determined relative to an initial position and orientation of the volume of interest within the working space.
Another aspect of the invention relates to a method for tracking a volume of interest, comprising: tracking at least two sensors in three dimensions, wherein the at least two sensors include: a reference sensor that provides a 6 DOF coordinate system in a working space, and a marker sensor adapted to be attached to the volume of interest and to provide at least position of the volume of interest in at least 3 DOF; outputting a change in position and/or orientation of the volume of interest within the working space in real time, according to the at least 3 DOF of the marker within the 6 DOF coordinate system.
In one embodiment of the method, the change in position and/or orientation of the volume of interest within the working space in real time is determined relative to an initial position and orientation of the volume of interest within the working space.
In one embodiment of the method, the at least two sensors are electromagnetic (EM) or optical.
In one embodiment, the method is applied to surgical navigation or intervention.
3.0 In one embodiment, the method is applied to gaming, robotics, motion capture, or training.
In one embodiment of the method, the volume of interest is biological tissue.
In one embodiment, the method is applied to surgical navigation and further comprises:
sensing position and orientation of a surgical tool within the 6 DOF
coordinate system; sensing position and orientation of an imaging probe of an imaging device within the 6 DOF coordinate system; using a processor to receive images from the imaging device, generate a three-dimensional delineation of the volume of interest from data points in the images, and three-dimensionally register the volume of interest with the surgical tool; and display the three-dimensionally registered volume of interest and the surgical tool in real time on an output device.
In one embodiment of the method, the marker sensor is associated with a wire-localization needle.
In one embodiment of the method, the volume of interest comprises a tumour.
In one embodiment of the method, the working space comprises breast tissue.
In one embodiment of the method, the surgical tool comprises a cutting tool, a cautery tool, a catheter, a needle, or a radiation therapy device.
Figs. 1A-1D are diagrams showing needle movements, including needle shift to and from the initial state, needle tilt, needle spin, needle in the initial state, respectively, according to embodiments described herein.
Fig. lE is a diagram showing a needle coordinate system centered at the needle tip with z-axis parallel to the needle shaft, according to one embodiment.
Fig. 2 is a diagram showing coordinate systems of a surgical tracking and navigation apparatus according to an embodiment of the invention, wherein arrows denote known transformations between coordinate systems.
Figs. 3A and 313 are diagrams showing tissue volume of interest and simulated volume resulting from marker spin, respectively, based on breast cancer patient data.
Figs. 4A and 4B are diagrams showing tissue volume of interest and simulated volume resulting from marker spin, respectively, based on breast cancer patient data.
Detailed Description of Embodiments Embodiments described herein use real-time tracking to three-dimensionally delineate and track a volume of interest in a working space. Whereas embodiments are described primarily with respect to electromagnetic (EM) tracking, it will be appreciated that they may also be implemented using other sensing/tracking modalities, such as those based on optical tracking.
Embodiments are described herein primarily with respect to surgical applications. However, it will be appreciated that the embodiments may be implemented for other applications, such as, but not limited to, gaming, motion capture, training, and robotics.
As used herein, the term "volume of interest" refers to an item, feature, object, etc., that is being tracked.
Prior approaches (e.g., [1, 2]) as applied to surgery provide accurate EM 3D
tracking of a tissue volume of interest. In breast cancer surgery, for example, this results in a significant reduction in the occurrence of positive margins during tumour resection.
However, the prior tracking technique requires a 6 DOF sensor to track the volume of interest (i.e., the tumour). The large and expensive 6 DOF sensor may restrict hand motions in the surgical working space, and may capture and report inaccurate orientation information. Improving accuracy, reducing the size constraints, and reducing cost of the EM sensing system will ensure that the technique becomes adopted and widely used in patients, In particular, insofar as it may be desired that certain sensor configurations are disposable, such as when the sensors are configured with localization needles that are placed in tissue, the cost of using 6 DOF
sensors may be prohibitive.
The apparatus and methods described herein provide more accurate, smaller, and less expensive tracking systems than prior apparatus or methods, by using fewer than 6 DOF to track the volume of interest within a 6 DOF coordinate system. For example, one embodiment uses only 5 DOF tracking of the volume of interest (a position and a three-dimensional vector) together with an external 6 DOF coordinate system, and provides superior performance relative to prior tracking methods based solely on full 6 DOF tracking information.
Another embodiment uses only 3 DOF tracking of the volume of interest (i.e., position) together with an external 6 DOF coordinate system. Thus, it is demonstrated that full 6 DOF tracking (comprising translation and full orientation information) is not required to capture all movements of a volume of interest.
Although the embodiments may be used to track position and orientation of any volume of interest, they are well-suited to tracking a volume that is mobile and/or deformable since the sensors allow estimation of the pose of the moving/deforming volume of interest. In one embodiment the apparatus includes a marker with a 5 DOF sensor, the marker adapted to be disposed in or on the volume of interest. In another embodiment the apparatus includes a marker with a 3 DOF sensor, the marker adapted to be disposed in or on the volume of interest. In various embodiments, the marker may comprise suitable hardware, such as pins, clips, needles, probes, and the like, so that it may be placed and anchored in or on the volume of interest. For
Real-time tracking as described herein may be adapted for a variety of surgical procedures, minimally invasive interventions, and surgical navigation systems.
Embodiments are particularly suitable for procedures in soft or deformable tissue, such as breast. For demonstrative purposes, embodiments relating to breast tumour resection will be described. It will be readily understood by those of ordinary skill in the art that the invention is not limited thereto, as embodiments may be applied to any organ, tissue, or structure, and corresponding procedure. Embodiments may also be applied to other procedures and interventions, such as, for example, radiation therapy interventions such as brachytherapy. In various embodiments and applications, one or more surgical tool is tracked in addition to the volume of interest. The surgical tool may be any tool or device as required by a procedure, e.g., a cutting tool, a cautery tool, a radiotherapy tool (e.g., a catheter, a linear accelerator, a needle, a device to deliver external beam radiation therapy, etc.). Accordingly, as used herein, the term "tool" or "surgical tool" is intended to refer to any tracked instrument, tool, or device that may be used or adapted for use with the apparatus and methods described herein for surgical and radiation therapy interventions.
A surgical navigation system including position and orientation tracking of a tissue volume of interest according to an embodiment described herein may employ an imaging modality based on ultrasound, computed tomography, magnetic resonance imaging, or projection imaging (e.g., X-ray), or a combination thereof. According to such embodiments, surgical navigation is provided by fusion of images obtained from an imaging modality in the same frame of reference as the tracked tissue volume of interest.
Results of preliminary trials in a surgical application (Example 3, below) using 5 DOF
tumour tracking as described herein demonstrate a significant improvement in localization of the tumour during breast conserving surgery in patients. Thus, it is expected that embodiments may
Embodiments will now be described in greater detail by way of the following non-limiting examples.
Example 1, Data Transformations for a 5 DOF Marker Sensor Data transformations used in the embodiments are described using biological tissue as an example of the volume of interest, with reference to Figs. 1A-1D, which are diagrams showing needle movements, including needle shift to and from the initial state (Fig.
1A), needle tilt (Fig.
1B), needle spin (Fig. 1C), needle in an initial state (Figs. 1D), respectively. In the figures, a needle 10 with sensor 12 is shown in an initial state, dispose in a tissue volume. The arrows show the respective movement (shift, tilt, spin). In Figs. lA and 1B, movement of the needle 10 from an initial state is shown by the needle 10a. In Fig. 1C, rotation of the needle 10 from an initial state is shown by new position of the sensor 12a, as the needle 10 is rotated about the axis defined by dashed line 14.
Motions of the tissue volume of interest are restricted to two types: shift (Fig. 1A) and tilt (Fig. 1B). Tilt describes a specific type of rotation relative to an initial orientation and position (Fig. 1D) of the tissue volume of interest. Shift describes translation relative to the initial position of the tissue volume of interest. These motions are captured using 5 DOF tracking information (comprising position and a single orientation vector).
The initial position and orientation of the tissue volume of interest is recorded in 6 DOF
after the sensor has been embedded or attached to the tissue volume of interest, but before the tissue volume of interest is operated on or manipulated. The initial position and orientation then stores the position and orientation of the tissue volume of interest before any external forces are applied (by e.g., cauterizing instruments, catheters, and so on). The 5 DOF
tracking information comprises only position and an orientation vector.
Let the orientation vector represent the direction of a shaft (e.g., needle shaft, sensor shaft) which acts as one basis vector in three dimensions. Two arbitrary orthogonal base vectors
sensor, or another instance of this invention).
Tilt and Shift are then computed continually from real time 5 DOF tracking information.
In the following description, transformation matrices are indicated in bold beginning with a capital letter, three-dimensional vectors are indicated in bold beginning with a lower case letter, and scalar values are indicated in italics. Let Initial and Changed be two matrix transformations describing the position and orientation of the tissue volume of interest relative to a common coordinate system before and after moving, respectively. Let positionlnitial represent the position of the tissue volume of interest before moving. Let positionChanged represent the position of the tissue volume of interest after moving. Let shaftInitial represent the orientation vector described by the 5 DOF tracking information before moving. Let shaftChanged represent the orientation vector described by the 5 DOF tracking information after moving. positionInitial, positionChanged, shaftInitial, shaftChanged are all given by the 5 DOF
tracking information.
Initial is already a fully defined transformation matrix recorded as described above. The corresponding 6 DOF transformation matrix Changed is not yet fully defined, but it can be computed:
Changed = Shift * Tilt * Initial Shift is modeled by a translation that is computed:shiftTranslation =
positionChanged ¨ positionInitial Tilt is modeled as an axis-angle rotation, with axis passing through the origin of the Initial coordinate system. The axis direction and angle are computed:
tiltAxis = Ishaftlnitial x shaftChangedl shaftsDotProduct = shaftInitial = shaftChanged shaftsAngleDif ference = asin(norm( shaftInitial x shaftChanged)) tiltAngle ¨180 ¨ shaftsAngleDif ference given shaftsDotProduct < 0 and shaftAngleDifference < 0 ------ 180 ¨ shaftsAngleDifference
Assuming there is no rotation, the volume of interest is tracked using an initial position and orientation, and shift.
This follows Example 1 regarding transformation matrices, three-dimensional vectors, and scalar values. Let Initial and Changed be two transformation matrices describing the position and orientation of the tissue volume of interest relative to a reference 6 DOF coordinate system before and after moving, respectively. Since the marker sensor only provides a position, there is no orientation information provided. The three orthogonal basis vectors that would normally describe orientation can be chosen arbitrarily.
Let markerPositionInitial be the position of the volume of interest before any movement happens, as measured by the marker sensor. Let markerPositionChanged be the position of the volume of interest after movement happens, as measured by the marker sensor.
Let Shift be described using a translation vector as in Example 1.
shiftTranslation = markerPositionChanged ¨ markerPositionInitial Then:
Changed = Shift * Initial Example 3, Surgical Implementation Background Breast cancer is the most commonly diagnosed cancer in women worldwide. Early-stage breast cancer patients are preferably treated with breast conserving surgery (BCS). BCS is
patients will need re-excision.
Tumour Tracking Needle localization is a standard technique used in BCS to mark the location of the tumour, i.e., the volume of interest. A needle or a hooked wire is implanted into the tumour under image guidance before surgery. A position and orientation sensor is fixed to the needle, and the tumour geometry is defined relative to the sensor. Throughout BCS, the tumour is moved together with the localization needle. The position and orientation sensor attached to the needle is used to capture the tumour movement. However, an important implication of this relationship is that needle movement is limited by the possible movements of the tumour.
Needle movement information was used to derive a system and method based on only 5 DOF tumour tracking. Three different types of needle movements are identified, relative to an initial state when the needle and tumour are not under external force from an instrument, an imaging probe, the surgeon's hands, the procedure, etc. The first type of needle movement is "shift", which is needle tip translation from the initial state (Fig. 1A).
Since the needle tip is embedded in the tumour using a hook, shift provides crucial and reliable information for localizing the tumour. The second type of needle movement is "tilt", which is rotation of the needle that results in a change of needle shaft direction, relative to the initial state (Fig. 1B). It is rotation with axis perpendicular to the needle shaft and passing through the needle tip. The breast is a highly mobile organ, and during BCS it is reasonable to expect that the surgeon will need to lift or manipulate some of the tissue to cut deeper regions. Thus, tilt captures orientation of the tumour. The third type of needle movement is "spin", which is rotation of the needle around the needle shaft itself, relative to the initial state (Fig. 1C).
Together these three needle movements describe a transformation. However, it is recognized herein that since the tumour is embedded within the tissue, it cannot rotate around the
tracking information was used to provide a needle shaft direction vector and a needle translation vector (as described in Example 1). The shaft direction may be used to compute the tilt from the initial state. The translation is immediately taken as the needle shift. Therefore, 5 DOF tracking information provides all necessary information to fully describe the movement of the tumour from an established initial state. This represents a significant reduction in size and cost relative to prior approaches based on 6 DOF sensors, Coordinate Systems Matrix transformations were used to represent all translations and rotations between coordinate systems. The coordinate systems are shown in Fig. 2. The Needle coordinate system measures the position and orientation of the needle, and is centered at the needle tip with z-axis parallel to the needle shaft, as shown in Fig. 1E. The tip and shaft can either be calibrated, or found on tracked ultrasound. The x- and y-axes do not have any distinct meaning, and can be chosen arbitrarily but still orthogonal to each other and to the z-axis. The Reference coordinate system measures the position and anatomical orientation of the patient, and is defined using a sensor secured to the patient. The ultrasound probe and surgical tool (e.g., electro-cauterizer) have coordinate systems and are tracked with their own dedicated sensors. An initial state is defined for the needle and used to compute the movements described above. The initial state is defined using the Initial coordinate system. A transformation matrix relative to the Reference coordinate system is recorded when there is no external force on the needle or on the tumour by, e.g., an ultrasound probe or surgical device. The Ultrasound and Cautery coordinate systems measure the positions of their respective tools.
tracking was compared to a prior 6 DOF method. The phantom was made of soft plastic with a simulated tumour inside (cylindrical, 2 cm in length and 1 cm in diameter). The setup is shown diagrammatically in Fig. 2. The setup included the open-source software and resources described previously [1], a TrakStarTm system (Ascension Technology Corporation, Northern Digital Inc.), and followed the setup protocol for navigated BCS. Tracking data were collected using the open-source PLUS toolkit (www.plusteolkitorg) [3], and relayed to custom software on the 3D Slicer platform [4] using the OpenIGTLink protocol.
Referring to Fig, 2, an ultrasound machine (e.g., SonixGPSTM Tablet, Ultrasonix, Vancouver, CA) was used with an EM tracker 24 having multiple sensor ports. An ultrasound probe 26 with EM sensor and tracker broadcast data through the PLUS toolkit to the SlicerIGT
navigation software running on the navigation computer, providing real-time registration and visualization of the tumour and tool 20 with EM sensor 22 (i.e., tool tip ) position with respect to the reference sensor 28. The EM sensor 22 was attached to the localization needle 20 using a flexible clip, and was inserted into the tumour. The tracked cautery tool is shown at 25. The arrows show transformations between coordinate systems.
In order to make direct comparisons of the tracking methods, a single 6 DOF
sensor was used to collect needle tracking data. The needle tip and shaft were calibrated on tracked ultrasound to create the Needle coordinate system. The 6 DOF tumour tracking method made use of the full transformation from the Needle coordinate system to the Reference, whereas the 5 DOF tumour tracking method made use of only the translation and needle shaft direction, as described above.
Tracked ultrasound was used to segment the simulated tumour and create a tumour shape model. The ground truth position and orientation of the tumour were accurately tracked by a sensor embedded in the tumour separate from the needle. The Initial coordinate system was established when there were no external forces on the phantom. Both 6 DOF- and tracked tumour shape models were copies of the above tumour shape model.
Snapshots of transformations were recorded during various manipulations of the phantom tissue, such as making an incision and then simulating surgical forces on the phantom tissue and
Results It was observed that the 5 DOF method as described herein preserved shift and tilt of the needle, but not spin. This is desirable behaviour, since the tumour itself does not spin.
A total of 21 snapshots of the transformations were recorded, A summary of the results is provided in Table 1, Both metrics indicate that the 5 DOF tumour tracking method as described herein was superior to the 6 DOF method (p = 0.002 for Dice coefficient and p <
0.001 for Hausdorff distance, using one-tailed Wilcoxon signed rank tests).
eased on observations and the reported metrics, it is concluded that the 5 DOF tumour tracking method performs better than the 6 DOF tumour tracking method. The improved performance of the 5 DOF tumour tracking method as described herein over the 6 DOF method is likely explained by the presence of needle spin. The average spin in the phantom experiment was 27.8 degrees.
Advantageously, the benefits of 5 DOF tumour tracking as described herein may be achieved at lower cost and with less complexity than prior tracking/navigation systems based on 6 DOF
tumour tracking.
Table 1. Average metrics (+ stdev) reported for 6 DOF and 5 DOF tumour tracking methods.
Tracking Method Dice Coefficient Hausdorff distance (mm) 6 DOF 0.56 0.18 8.8 4.8 5 DOF 0.71 + 0.10 5.0 1.9 To determine if the experiment was representative of actual surgery, clinical data were analyzed with the goal of quantifying needle spin. Ultrasound and tracking data were previously recorded in 17 BCS patient cases. The needle tip and shaft were segmented on the recorded
Needle spin varied considerably from patient to patient. In seven patients, spin exceeded 30 degrees on average. Some of the 95th percentiles exceeded 90 degrees. It is unlikely that the tumour would have rotated by such large magnitudes inside the breast, even if the tumour was already partially detached. Thus, it is expected that 5 DOF tumour tracking as described herein will result in improved localization of the tumour during BCS in patients.
Example 4. Volume Inaccuracy with 6 DOF Tracking This example demonstrates a deficiency of 6 DOF tracking in breast cancer surgery. This example pertains to the volume outside the tissue volume of interest that could be erroneously tracked using 6 DOF tracking due to erroneous spin (as described above).
A simulation was performed using clinical data of tissue volumes of interest and marker placements for 17 breast cancer patient cases, Spin was calculated around the marker to simulate erroneous rotational data. A union was computed from resulting spun volumes to measure the total amount of extra volume that could be erroneously tracked. Figs. 3A-3B
and 4A-4B show two examples, in which Figs. 3A and 4A show tissue volumes of interest 32, 42 with shafts of inserted markers 30, 40 for two patients. In Figs. 3B and 4B, the corresponding unions of spun volumes 34, 44 are shown. It is appreciated that the unions of spun volumes, representing the volumes that could be tracked if spin is not corrected, are larger than the volumes of interest. For the 17 patients, the average tissue volume of interest was 2,080 cubic millimeters, and the average union volume was 7,010 cubic millimeters. This represents more than a threefold
Thus, accuracy of the procedure may be significantly improved if spin is corrected as described herein.
All cited publications are incorporated herein by reference in their entirety.
Equivalents While the invention has been described with respect to illustrative embodiments thereof, it will be understood that various changes may be made to the embodiments without departing from the scope of the invention. Accordingly, the described embodiments are to be considered merely exemplary and the invention is not to be limited thereby.
[2] Gauvin, G., Ungi, T., Lasso, A., Yeo, C.T., Fichtinger, G., Jabs, D., Walker, R., Merchant, S., Rudan, J., Engel C.J.: Breast-Conserving Surgery using NaviKnife Technology:
Pilot Study on Non-Palpable Tumours. Canadian Surgery Forum, Toronto, Ontario, Canada, September 8-10, 2016. (Can. Surg. J. 2016, 59(4 Suppl. 1): S145-6).
[3] Lasso, A., Hefner, T., Rankin, A., Pinter, Cõ Ungi, T. and Fichtinger, G.: PLUS: open-lo source toolkit for ultrasound-guided intervention systems. IEEE Trans, Bio-Med. Eng. 61(10), 2527-2537 (2014).
[4] Fedorov, A., Beichel, R., Kalpathy-Cramer, J., Finet, J., Fillion-Robin, J.C., Pujol, S., Bauer, C., Jennings, D., Fennessy, F., Sonka, M. and Buatti, J.: 3D Slicer as an image computing platform for the Quantitative Imaging Network, Magn. Reson. Imaging. 30(9), (2012).
Claims (21)
a tracking device that tracks at least two sensors in three dimensions, wherein the at least two sensors include:
a reference sensor that provides a 6 DOF coordinate system in a working space;
a marker sensor adapted to be attached to the volume of interest and to provide at least position of the volume of interest in at least 3 DOF;
a tracking device that receives sensor data from the reference sensor and the marker sensor;
a processor that receives data from the tracking device and outputs a change in position and/or orientation of the volume of interest within the working space in real time, according to the at least 3 DOF of the marker sensor within the 6 DOF coordinate system.
a tool sensor adapted to be attached to a surgical tool and to provide position and orientation of the surgical tool;
an imaging probe sensor adapted to be attached to an imaging probe of an imaging device and to provide position and orientation of the imaging probe;
wherein the processor receives images from the imaging device, generates a three-dimensional delineation of the volume of interest from data points in the images, and three-dimensionally registers the volume of interest with the surgical tool; and an output device that displays the three-dimensionally registered volume of interest and the surgical tool in real time.
code stored on non-transitory storage media compatible with the processor, the code containing instructions to direct the processor to:
communicate with a tracking device that tracks at least two sensors in three dimensions, wherein the at least two sensors include:
a reference sensor that provides a 6 DOF coordinate system in a working space;
a marker sensor adapted to be attached to the volume of interest and to provide at least position of the volume of interest in at least 3 DOF;
wherein the processor outputs a change in position and/or orientation of the volume of interest within the working space in real time, according to the at least 3 DOF of the marker sensor relative to the 6 DOF coordinate system.
tracking at least two sensors in three dimensions, wherein the at least two sensors include:
a reference sensor that provides a 6 DOF coordinate system in a working space;
a marker sensor adapted to be attached to the volume of interest and to provide at least position of the volume of interest in at least 3 DOF;
outputting a change in position and/or orientation of the volume of interest within the working space in real time, according to the at least 3 DOF of the marker within the 6 DOF
coordinate system.
sensing position and orientation of a surgical tool within the 6 DOF
coordinate system;
sensing position and orientation of an imaging probe of an imaging device within the 6 DOF coordinate system;
using a processor to receive images from the imaging device, generate a three-dimensional delineation of the volume of interest from data points in the images, and three-dimensionally register the volume of interest with the surgical tool; and display the three-dimensionally registered volume of interest and the surgical tool in real time on an output device.
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201862623082P | 2018-01-29 | 2018-01-29 | |
| US62/623,082 | 2018-01-29 |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CA3031832A1 true CA3031832A1 (en) | 2019-07-29 |
| CA3031832C CA3031832C (en) | 2025-12-09 |
Family
ID=67391741
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CA3031832A Active CA3031832C (en) | 2018-01-29 | 2019-01-29 | Apparatus and method for tracking a volume in a three-dimensional space |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US11007015B2 (en) |
| CA (1) | CA3031832C (en) |
Families Citing this family (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11638569B2 (en) | 2018-06-08 | 2023-05-02 | Rutgers, The State University Of New Jersey | Computer vision systems and methods for real-time needle detection, enhancement and localization in ultrasound |
| WO2020036968A1 (en) * | 2018-08-13 | 2020-02-20 | Rutgers, The State University Of New Jersey | Computer vision systems and methods for real-time localization of needles in ultrasound images |
| CN117256008A (en) | 2021-04-30 | 2023-12-19 | 农历Ng有限公司 | Real-time localization using image data |
Family Cites Families (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5394875A (en) * | 1993-10-21 | 1995-03-07 | Lewis; Judith T. | Automatic ultrasonic localization of targets implanted in a portion of the anatomy |
| US6973202B2 (en) * | 1998-10-23 | 2005-12-06 | Varian Medical Systems Technologies, Inc. | Single-camera tracking of an object |
| JP4976756B2 (en) * | 2006-06-23 | 2012-07-18 | キヤノン株式会社 | Information processing method and apparatus |
| JP4926817B2 (en) * | 2006-08-11 | 2012-05-09 | キヤノン株式会社 | Index arrangement information measuring apparatus and method |
| CN106714681A (en) * | 2014-07-23 | 2017-05-24 | 凯内蒂科尔股份有限公司 | Systems, devices and methods for tracking and compensating for patient motion during medical imaging scans |
| US10426556B2 (en) * | 2016-06-01 | 2019-10-01 | Vanderbilt University | Biomechanical model assisted image guided surgery system and method |
-
2019
- 2019-01-29 US US16/260,538 patent/US11007015B2/en active Active
- 2019-01-29 CA CA3031832A patent/CA3031832C/en active Active
Also Published As
| Publication number | Publication date |
|---|---|
| US11007015B2 (en) | 2021-05-18 |
| CA3031832C (en) | 2025-12-09 |
| US20190231441A1 (en) | 2019-08-01 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Kim et al. | Ultrasound probe and needle-guide calibration for robotic ultrasound scanning and needle targeting | |
| CA2968917C (en) | Sensor based tracking tool for medical components | |
| US20090221908A1 (en) | System and Method for Alignment of Instrumentation in Image-Guided Intervention | |
| US20120071757A1 (en) | Ultrasound Registration | |
| US10799146B2 (en) | Interactive systems and methods for real-time laparoscopic navigation | |
| US10154882B2 (en) | Global laparoscopy positioning systems and methods | |
| Najmaei et al. | Image‐guided techniques in renal and hepatic interventions | |
| CA2963865C (en) | Phantom to determine positional and angular navigation system error | |
| US20160278746A1 (en) | Real-time tracking for mri-guided breast biopsy | |
| Mohareri et al. | Automatic localization of the da Vinci surgical instrument tips in 3-D transrectal ultrasound | |
| CA3031832C (en) | Apparatus and method for tracking a volume in a three-dimensional space | |
| Lamata et al. | Augmented reality for minimally invasive surgery: overview and some recent advances | |
| US20190133694A1 (en) | Surgical tracking system for tracking and visualizing the relative positioning of two or more surgical components | |
| Adebar et al. | Registration of 3D ultrasound through an air–tissue boundary | |
| Beller et al. | Upgrade of an optical navigation system with a permanent electromagnetic position control: a first step towards “navigated control” for liver surgery | |
| Schneider et al. | Intra-operative “Pick-Up” ultrasound for robot assisted surgery with vessel extraction and registration: a feasibility study | |
| CN109833092A (en) | Internal navigation system and method | |
| Shahin et al. | Ultrasound-based tumor movement compensation during navigated laparoscopic liver interventions | |
| Lange et al. | Development of navigation systems for image-guided laparoscopic tumor resections in liver surgery | |
| JP7495216B2 (en) | Endoscopic surgery support device, endoscopic surgery support method, and program | |
| Marisetty et al. | System design of an automated drilling device for neurosurgical applications | |
| Troccaz et al. | Medical image computing and computer-aided medical interventions applied to soft tissues: Work in progress in urology | |
| Ong et al. | Intraprocedural registration for image-guided kidney surgery | |
| Linte et al. | Targeting accuracy under model-to-subject misalignments in model-guided cardiac surgery | |
| Herrell et al. | Image guidance in robotic-assisted renal surgery |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| EEER | Examination request |
Effective date: 20240123 |
|
| EEER | Examination request |
Effective date: 20240123 |
|
| EEER | Examination request |
Effective date: 20240123 |
|
| EEER | Examination request |
Effective date: 20240123 |
|
| EEER | Examination request |
Effective date: 20240123 |
|
| MFA | Maintenance fee for application paid |
Free format text: FEE DESCRIPTION TEXT: MF (APPLICATION, 6TH ANNIV.) - SMALL Year of fee payment: 6 |
|
| U00 | Fee paid |
Free format text: ST27 STATUS EVENT CODE: A-2-2-U10-U00-U101 (AS PROVIDED BY THE NATIONAL OFFICE); EVENT TEXT: MAINTENANCE REQUEST RECEIVED Effective date: 20241112 |
|
| U11 | Full renewal or maintenance fee paid |
Free format text: ST27 STATUS EVENT CODE: A-2-2-U10-U11-U102 (AS PROVIDED BY THE NATIONAL OFFICE); EVENT TEXT: MAINTENANCE FEE PAYMENT PAID IN FULL Effective date: 20241112 |
|
| D22 | Grant of ip right intended |
Free format text: ST27 STATUS EVENT CODE: A-2-4-D10-D22-D143 (AS PROVIDED BY THE NATIONAL OFFICE); EVENT TEXT: PRE-GRANT Effective date: 20250502 |