US20220054206A1 - Surgical robot platform - Google Patents
Surgical robot platform Download PDFInfo
- Publication number
- US20220054206A1 US20220054206A1 US17/520,196 US202117520196A US2022054206A1 US 20220054206 A1 US20220054206 A1 US 20220054206A1 US 202117520196 A US202117520196 A US 202117520196A US 2022054206 A1 US2022054206 A1 US 2022054206A1
- Authority
- US
- United States
- Prior art keywords
- robot
- surgical instrument
- axis
- surgical
- effectuator
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 claims description 191
- 230000007246 mechanism Effects 0.000 claims description 32
- 238000001356 surgical procedure Methods 0.000 claims description 27
- 238000003780 insertion Methods 0.000 claims description 18
- 230000037431 insertion Effects 0.000 claims description 18
- 238000004364 calculation method Methods 0.000 claims description 14
- 230000033001 locomotion Effects 0.000 abstract description 108
- 239000003550 marker Substances 0.000 description 83
- 239000013598 vector Substances 0.000 description 69
- 210000003484 anatomy Anatomy 0.000 description 60
- 230000015654 memory Effects 0.000 description 42
- 230000008569 process Effects 0.000 description 42
- 239000003795 chemical substances by application Substances 0.000 description 31
- 230000008685 targeting Effects 0.000 description 31
- 230000008859 change Effects 0.000 description 27
- 230000003287 optical effect Effects 0.000 description 25
- 230000000717 retained effect Effects 0.000 description 25
- 230000004044 response Effects 0.000 description 22
- 239000000523 sample Substances 0.000 description 22
- 238000002591 computed tomography Methods 0.000 description 21
- 210000001519 tissue Anatomy 0.000 description 21
- 210000000988 bone and bone Anatomy 0.000 description 17
- 230000009466 transformation Effects 0.000 description 17
- 238000012545 processing Methods 0.000 description 16
- 238000010586 diagram Methods 0.000 description 14
- 230000000694 effects Effects 0.000 description 14
- 238000012360 testing method Methods 0.000 description 13
- 230000007935 neutral effect Effects 0.000 description 12
- 238000000844 transformation Methods 0.000 description 12
- 230000000007 visual effect Effects 0.000 description 12
- 238000004891 communication Methods 0.000 description 10
- 238000001574 biopsy Methods 0.000 description 9
- 230000008901 benefit Effects 0.000 description 7
- 238000013500 data storage Methods 0.000 description 7
- 238000001514 detection method Methods 0.000 description 7
- 239000011324 bead Substances 0.000 description 6
- 238000003384 imaging method Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 5
- 230000035515 penetration Effects 0.000 description 5
- 230000001360 synchronised effect Effects 0.000 description 5
- 238000005452 bending Methods 0.000 description 4
- 238000004422 calculation algorithm Methods 0.000 description 4
- 230000010405 clearance mechanism Effects 0.000 description 4
- 238000005553 drilling Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 238000005286 illumination Methods 0.000 description 4
- 238000002595 magnetic resonance imaging Methods 0.000 description 4
- 238000004519 manufacturing process Methods 0.000 description 4
- 239000000463 material Substances 0.000 description 4
- 239000011159 matrix material Substances 0.000 description 4
- 238000012986 modification Methods 0.000 description 4
- 230000004048 modification Effects 0.000 description 4
- 230000036961 partial effect Effects 0.000 description 4
- 210000004761 scalp Anatomy 0.000 description 4
- 238000013519 translation Methods 0.000 description 4
- 230000014616 translation Effects 0.000 description 4
- 230000004913 activation Effects 0.000 description 3
- 230000006835 compression Effects 0.000 description 3
- 238000007906 compression Methods 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 230000006378 damage Effects 0.000 description 3
- 238000013503 de-identification Methods 0.000 description 3
- 230000000670 limiting effect Effects 0.000 description 3
- 238000005259 measurement Methods 0.000 description 3
- 230000006855 networking Effects 0.000 description 3
- 230000002093 peripheral effect Effects 0.000 description 3
- 210000003625 skull Anatomy 0.000 description 3
- 210000004872 soft tissue Anatomy 0.000 description 3
- VYZAMTAEIAYCRO-UHFFFAOYSA-N Chromium Chemical compound [Cr] VYZAMTAEIAYCRO-UHFFFAOYSA-N 0.000 description 2
- 208000027418 Wounds and injury Diseases 0.000 description 2
- 230000003213 activating effect Effects 0.000 description 2
- 239000002313 adhesive film Substances 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 2
- 230000000903 blocking effect Effects 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 238000002059 diagnostic imaging Methods 0.000 description 2
- 244000144992 flock Species 0.000 description 2
- ACGUYXCXAPNIKK-UHFFFAOYSA-N hexachlorophene Chemical compound OC1=C(Cl)C=C(Cl)C(Cl)=C1CC1=C(O)C(Cl)=CC(Cl)=C1Cl ACGUYXCXAPNIKK-UHFFFAOYSA-N 0.000 description 2
- 239000007943 implant Substances 0.000 description 2
- 210000004705 lumbosacral region Anatomy 0.000 description 2
- 238000012544 monitoring process Methods 0.000 description 2
- 230000029058 respiratory gaseous exchange Effects 0.000 description 2
- 230000000087 stabilizing effect Effects 0.000 description 2
- 239000004593 Epoxy Substances 0.000 description 1
- 244000208060 Lawsonia inermis Species 0.000 description 1
- 241000270295 Serpentes Species 0.000 description 1
- 206010052428 Wound Diseases 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 239000000853 adhesive Substances 0.000 description 1
- 230000001070 adhesive effect Effects 0.000 description 1
- 238000004873 anchoring Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 239000000872 buffer Substances 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 229910052804 chromium Inorganic materials 0.000 description 1
- 239000011651 chromium Substances 0.000 description 1
- 150000001875 compounds Chemical class 0.000 description 1
- 238000013170 computed tomography imaging Methods 0.000 description 1
- 238000012790 confirmation Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000001054 cortical effect Effects 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000002405 diagnostic procedure Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 238000013213 extrapolation Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000003711 image thresholding Methods 0.000 description 1
- 238000002347 injection Methods 0.000 description 1
- 239000007924 injection Substances 0.000 description 1
- 208000014674 injury Diseases 0.000 description 1
- 238000012977 invasive surgical procedure Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 230000007257 malfunction Effects 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 230000005055 memory storage Effects 0.000 description 1
- 229910052751 metal Inorganic materials 0.000 description 1
- 239000002184 metal Substances 0.000 description 1
- 230000003278 mimic effect Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000000414 obstructive effect Effects 0.000 description 1
- 239000003973 paint Substances 0.000 description 1
- 230000001737 promoting effect Effects 0.000 description 1
- 239000002096 quantum dot Substances 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000015541 sensory perception of touch Effects 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 239000000243 solution Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000000638 stimulation Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000005211 surface analysis Methods 0.000 description 1
- 230000001225 therapeutic effect Effects 0.000 description 1
- 238000002560 therapeutic procedure Methods 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/30—Surgical robots
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B10/00—Other methods or instruments for diagnosis, e.g. instruments for taking a cell sample, for biopsy, for vaccination diagnosis; Sex determination; Ovulation-period determination; Throat striking implements
- A61B10/02—Instruments for taking cell samples or for biopsy
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B10/00—Other methods or instruments for diagnosis, e.g. instruments for taking a cell sample, for biopsy, for vaccination diagnosis; Sex determination; Ovulation-period determination; Throat striking implements
- A61B10/02—Instruments for taking cell samples or for biopsy
- A61B10/0233—Pointed or sharp biopsy instruments
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B10/00—Other methods or instruments for diagnosis, e.g. instruments for taking a cell sample, for biopsy, for vaccination diagnosis; Sex determination; Ovulation-period determination; Throat striking implements
- A61B10/02—Instruments for taking cell samples or for biopsy
- A61B10/0233—Pointed or sharp biopsy instruments
- A61B10/0266—Pointed or sharp biopsy instruments means for severing sample
- A61B10/0275—Pointed or sharp biopsy instruments means for severing sample with sample notch, e.g. on the side of inner stylet
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/02—Surgical instruments, devices or methods, e.g. tourniquets for holding wounds open; Tractors
- A61B17/025—Joint distractors
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/16—Bone cutting, breaking or removal means other than saws, e.g. Osteoclasts; Drills or chisels for bones; Trepans
- A61B17/1613—Component parts
- A61B17/1615—Drill bits, i.e. rotating tools extending from a handpiece to contact the worked material
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/16—Bone cutting, breaking or removal means other than saws, e.g. Osteoclasts; Drills or chisels for bones; Trepans
- A61B17/1662—Bone cutting, breaking or removal means other than saws, e.g. Osteoclasts; Drills or chisels for bones; Trepans for particular parts of the body
- A61B17/1671—Bone cutting, breaking or removal means other than saws, e.g. Osteoclasts; Drills or chisels for bones; Trepans for particular parts of the body for the spine
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/16—Bone cutting, breaking or removal means other than saws, e.g. Osteoclasts; Drills or chisels for bones; Trepans
- A61B17/17—Guides or aligning means for drills, mills, pins or wires
- A61B17/1703—Guides or aligning means for drills, mills, pins or wires using imaging means, e.g. by X-rays
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/16—Bone cutting, breaking or removal means other than saws, e.g. Osteoclasts; Drills or chisels for bones; Trepans
- A61B17/17—Guides or aligning means for drills, mills, pins or wires
- A61B17/1739—Guides or aligning means for drills, mills, pins or wires specially adapted for particular parts of the body
- A61B17/1757—Guides or aligning means for drills, mills, pins or wires specially adapted for particular parts of the body for the spine
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/56—Surgical instruments or methods for treatment of bones or joints; Devices specially adapted therefor
- A61B17/58—Surgical instruments or methods for treatment of bones or joints; Devices specially adapted therefor for osteosynthesis, e.g. bone plates, screws, setting implements or the like
- A61B17/68—Internal fixation devices, including fasteners and spinal fixators, even if a part thereof projects from the skin
- A61B17/70—Spinal positioners or stabilisers ; Bone stabilisers comprising fluid filler in an implant
- A61B17/7074—Tools specially adapted for spinal fixation operations other than for bone removal or filler handling
- A61B17/7076—Tools specially adapted for spinal fixation operations other than for bone removal or filler handling for driving, positioning or assembling spinal clamps or bone anchors specially adapted for spinal fixation
- A61B17/7082—Tools specially adapted for spinal fixation operations other than for bone removal or filler handling for driving, positioning or assembling spinal clamps or bone anchors specially adapted for spinal fixation for driving, i.e. rotating, screws or screw parts specially adapted for spinal fixation, e.g. for driving polyaxial or tulip-headed screws
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/56—Surgical instruments or methods for treatment of bones or joints; Devices specially adapted therefor
- A61B17/58—Surgical instruments or methods for treatment of bones or joints; Devices specially adapted therefor for osteosynthesis, e.g. bone plates, screws, setting implements or the like
- A61B17/88—Osteosynthesis instruments; Methods or means for implanting or extracting internal or external fixation devices
- A61B17/8866—Osteosynthesis instruments; Methods or means for implanting or extracting internal or external fixation devices for gripping or pushing bones, e.g. approximators
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/25—User interfaces for surgical systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/30—Surgical robots
- A61B34/32—Surgical robots operating autonomously
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
- A61B34/74—Manipulators with manual electric input means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
- A61B34/76—Manipulators having means for providing feel, e.g. force or tactile feedback
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B46/00—Surgical drapes
- A61B46/20—Surgical drapes specially adapted for patients
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/06—Devices, other than using radiation, for detecting or locating foreign bodies ; determining position of probes within or on the body of the patient
- A61B5/061—Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/06—Devices, other than using radiation, for detecting or locating foreign bodies ; determining position of probes within or on the body of the patient
- A61B5/061—Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body
- A61B5/062—Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body using magnetic field
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/06—Devices, other than using radiation, for detecting or locating foreign bodies ; determining position of probes within or on the body of the patient
- A61B5/065—Determining position of the probe employing exclusively positioning means located on or in the probe, e.g. using position sensors arranged on the probe
- A61B5/066—Superposing sensor position on an image of the patient, e.g. obtained by ultrasound or x-ray imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B50/00—Containers, covers, furniture or holders specially adapted for surgical or diagnostic appliances or instruments, e.g. sterile covers
- A61B50/10—Furniture specially adapted for surgical or diagnostic appliances or instruments
- A61B50/13—Trolleys, e.g. carts
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/10—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
- A61B90/14—Fixators for body parts, e.g. skull clamps; Constructional details of fixators, e.g. pins
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/90—Identification means for patients or instruments, e.g. tags
- A61B90/94—Identification means for patients or instruments, e.g. tags coded with symbols, e.g. text
- A61B90/96—Identification means for patients or instruments, e.g. tags coded with symbols, e.g. text using barcodes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/90—Identification means for patients or instruments, e.g. tags
- A61B90/98—Identification means for patients or instruments, e.g. tags using electromagnetic means, e.g. transponders
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61M—DEVICES FOR INTRODUCING MEDIA INTO, OR ONTO, THE BODY; DEVICES FOR TRANSDUCING BODY MEDIA OR FOR TAKING MEDIA FROM THE BODY; DEVICES FOR PRODUCING OR ENDING SLEEP OR STUPOR
- A61M5/00—Devices for bringing media into the body in a subcutaneous, intra-vascular or intramuscular way; Accessories therefor, e.g. filling or cleaning devices, arm-rests
- A61M5/14—Infusion devices, e.g. infusing by gravity; Blood infusion; Accessories therefor
- A61M5/168—Means for controlling media flow to the body or for metering media to the body, e.g. drip meters, counters ; Monitoring media flow to the body
- A61M5/172—Means for controlling media flow to the body or for metering media to the body, e.g. drip meters, counters ; Monitoring media flow to the body electrical or electronic
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61N—ELECTROTHERAPY; MAGNETOTHERAPY; RADIATION THERAPY; ULTRASOUND THERAPY
- A61N1/00—Electrotherapy; Circuits therefor
- A61N1/02—Details
- A61N1/04—Electrodes
- A61N1/05—Electrodes for implantation or insertion into the body, e.g. heart electrode
- A61N1/0526—Head electrodes
- A61N1/0529—Electrodes for brain stimulation
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/10—Programme-controlled manipulators characterised by positioning means for manipulator elements
- B25J9/106—Programme-controlled manipulators characterised by positioning means for manipulator elements with articulated links
- B25J9/1065—Programme-controlled manipulators characterised by positioning means for manipulator elements with articulated links with parallelograms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/16—Bone cutting, breaking or removal means other than saws, e.g. Osteoclasts; Drills or chisels for bones; Trepans
- A61B17/17—Guides or aligning means for drills, mills, pins or wires
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B10/00—Other methods or instruments for diagnosis, e.g. instruments for taking a cell sample, for biopsy, for vaccination diagnosis; Sex determination; Ovulation-period determination; Throat striking implements
- A61B10/02—Instruments for taking cell samples or for biopsy
- A61B2010/0208—Biopsy devices with actuators, e.g. with triggered spring mechanisms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00115—Electrical control of surgical instruments with audible or visual output
- A61B2017/00119—Electrical control of surgical instruments with audible or visual output alarm; indicating an abnormal situation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00203—Electrical control of surgical instruments with speech control or speech recognition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00207—Electrical control of surgical instruments with hand gesture control or hand gesture recognition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00831—Material properties
- A61B2017/00876—Material properties magnetic
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B17/02—Surgical instruments, devices or methods, e.g. tourniquets for holding wounds open; Tractors
- A61B17/025—Joint distractors
- A61B2017/0256—Joint distractors for the spine
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/107—Visualisation of planned trajectories or target regions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2051—Electromagnetic tracking systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2055—Optical tracking systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2046—Tracking techniques
- A61B2034/2059—Mechanical position encoders
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/20—Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
- A61B2034/2072—Reference field transducer attached to an instrument or patient
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/30—Surgical robots
- A61B2034/301—Surgical robots for introducing or steering flexible instruments inserted into the body, e.g. catheters or endoscopes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
- A61B34/74—Manipulators with manual electric input means
- A61B2034/741—Glove like input devices, e.g. "data gloves"
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
- A61B34/74—Manipulators with manual electric input means
- A61B2034/742—Joysticks
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
- A61B34/74—Manipulators with manual electric input means
- A61B2034/743—Keyboards
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
- A61B34/74—Manipulators with manual electric input means
- A61B2034/744—Mouse
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/03—Automatic limiting or abutting means, e.g. for safety
- A61B2090/033—Abutting means, stops, e.g. abutting on tissue or skin
- A61B2090/034—Abutting means, stops, e.g. abutting on tissue or skin abutting on parts of the device itself
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/06—Measuring instruments not otherwise provided for
- A61B2090/064—Measuring instruments not otherwise provided for for measuring force, pressure or mechanical tension
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/08—Accessories or related features not otherwise provided for
- A61B2090/0807—Indication means
- A61B2090/0811—Indication means for the position of a particular part of an instrument with respect to the rest of the instrument, e.g. position of the anvil of a stapling instrument
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B2090/364—Correlation of different images or relation of image positions in respect to the body
- A61B2090/365—Correlation of different images or relation of image positions in respect to the body augmented reality, i.e. correlating a live optical image with another image
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/374—NMR or MRI
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/376—Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy
- A61B2090/3762—Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy using computed tomography systems [CT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/376—Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy
- A61B2090/3762—Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy using computed tomography systems [CT]
- A61B2090/3764—Surgical systems with images on a monitor during operation using X-rays, e.g. fluoroscopy using computed tomography systems [CT] with a rotating C-arm having a cone beam emitting source
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/36—Image-producing devices or illumination devices not otherwise provided for
- A61B90/37—Surgical systems with images on a monitor during operation
- A61B2090/378—Surgical systems with images on a monitor during operation using ultrasound
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3937—Visible markers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3937—Visible markers
- A61B2090/3941—Photoluminescent markers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3937—Visible markers
- A61B2090/3945—Active visible markers, e.g. light emitting diodes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3937—Visible markers
- A61B2090/395—Visible markers with marking agent for marking skin or other tissue
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3966—Radiopaque markers visible in an X-ray image
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/397—Markers, e.g. radio-opaque or breast lesions markers electromagnetic other than visible, e.g. microwave
- A61B2090/3975—Markers, e.g. radio-opaque or breast lesions markers electromagnetic other than visible, e.g. microwave active
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/397—Markers, e.g. radio-opaque or breast lesions markers electromagnetic other than visible, e.g. microwave
- A61B2090/3975—Markers, e.g. radio-opaque or breast lesions markers electromagnetic other than visible, e.g. microwave active
- A61B2090/3979—Markers, e.g. radio-opaque or breast lesions markers electromagnetic other than visible, e.g. microwave active infrared
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/39—Markers, e.g. radio-opaque or breast lesions markers
- A61B2090/3983—Reference marker arrangements for use with image guided surgery
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/06—Devices, other than using radiation, for detecting or locating foreign bodies ; determining position of probes within or on the body of the patient
- A61B5/061—Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body
- A61B5/064—Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body using markers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B90/00—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
- A61B90/10—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis
- A61B90/11—Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges for stereotaxic surgery, e.g. frame-based stereotaxis with guides for needles or instruments, e.g. arcuate slides or ball joints
Definitions
- a surgical robot (and optionally an imaging system) that utilizes a Cartesian positioning system that allows movement of a surgical instrument to be individually controlled in an x-axis, y-axis and z-axis.
- the surgical robot can include a base, a robot arm coupled to and configured for articulation relative to the base, as well as an end-effectuator coupled to a distal end of the robot arm.
- the effectuator element can include the surgical instrument or can be configured for operative coupling to the surgical instrument.
- Some embodiments of the invention allow the roll, pitch and yaw rotation of the end-effectuator and/or surgical instrument to be controlled without creating movement along the x-axis, y-axis, or z-axis.
- FIG. 1 is a partial perspective view of a room in which a medical procedure is taking place by using a surgical robot.
- FIG. 2 is a perspective view of a surgical robot according to an embodiment of the invention.
- FIGS. 3A-3B are perspective views of the surgical robot illustrated in FIG. 2 , which show the movement of the base of the surgical robot in the z-axis direction in accordance with an embodiment of the invention.
- FIG. 4 is a partial perspective view of the surgical robot of FIG. 2 which shows how the robot arm can be moved in the x-axis direction.
- FIGS. 5A-5B are partial perspective views of the surgical robot of FIG. 2 , which show how the robot arm can be moved in the y-axis direction.
- FIG. 6 is a perspective view of a portion of the robot arm of FIG. 2 showing how an effectuator element can be twisted about a y-axis.
- FIG. 7 is a perspective view of a portion of a robot arm of FIG. 2 showing how an effectuator element can be pivoted about a pivot axis that is perpendicular to the y-axis.
- FIGS. 8A-8B are partial perspective views of the surgical robot of FIG. 2 , which show the movement of a surgical instrument 35 along the z-axis from an effectuator element.
- FIG. 9 is a system diagram which shows local positioning sensors, a controlling PC, and a Radiofrequency (RF) transmitter in accordance with an embodiment of the invention.
- RF Radiofrequency
- FIG. 10 is a system diagram of the controlling PC, user input, and motors for controlling the robot in accordance with an embodiment of the invention.
- FIG. 11 is a flow chart diagram for general operation of a surgical robot in accordance with one embodiment of the invention.
- FIG. 12 is a flow chart diagram for a closed screw/needle insertion performed using a surgical robot in accordance with one embodiment of the invention.
- FIG. 13 is a flow chart diagram of a safe zone surgery performed using a surgical robot as described herein in accordance with one embodiment of the invention.
- FIG. 14 is a flow chart diagram of a flexible catheter insertion procedure performed using a surgical robot as described herein in accordance with one embodiment of the invention.
- FIG. 15A shows a screenshot of a monitor display showing a set up of the anatomy in X, Y and Z views in accordance with one embodiment of the invention.
- FIG. 15B shows a screenshot of a monitor display showing what the user views during an invasive procedure in accordance with one embodiment of the invention.
- FIG. 16 depicts a surgical robot having a plurality of optical markers mounted for tracking movement in an x-direction in accordance with one embodiment of the invention.
- FIGS. 17A-17B depict surgical instruments having a stop mechanism in accordance with one embodiment of the invention.
- FIGS. 17C-17E illustrate tools for manually adjusting a drill stop with reference to drill bit markings in accordance with one embodiment of the invention.
- FIGS. 17F-17J illustrate tools for locking and holding a drill bit in a set position in accordance with one embodiment of the invention.
- FIGS. 18A-18B depicts an end-effectuator having a clearance mechanism in accordance with one embodiment of the invention.
- FIG. 19A-19B depicts an end-effectuator having an attachment element for applying distraction and/or compression forces in accordance with one embodiment of the invention.
- FIGS. 20A-20E show the use of calibration frames with the guidance system in accordance with one embodiment of the invention.
- FIG. 21A depicts flexible roll configurations of a targeting fixture in accordance with one embodiment of the invention.
- FIG. 21B shows possible positions of markers along a line in space in accordance with one embodiment of the invention.
- FIG. 21C depicts flexible roll configurations of a targeting fixture in accordance with one embodiment of the invention.
- FIG. 21D shows a fixture that can be employed to provide desired stiffness to the unrolled fixture such that it maintains its position after unrolling occurs in accordance with one embodiment of the invention.
- FIGS. 22A-22D depict a targeting fixture and method configured for application to the skull of a patient in accordance with one embodiment of the invention.
- FIG. 23 depicts a dynamic tracking device mounted to the spinous process of the lumbar spine of a patient in accordance with one embodiment of the invention.
- FIGS. 24-27 illustrate methods for the robot in accordance with one embodiment of the invention.
- FIGS. 28A-28B illustrate methods for calculating position and/or orientation of an end-effectuator for the robot in accordance with one embodiment of the invention.
- FIGS. 29-33 illustrate methods for the robot in accordance with other embodiments of the invention.
- FIG. 34 illustrates a computing platform that enables implementation of various embodiments of the invention.
- FIGS. 35A-35B display a surgical robot in accordance with one embodiment of the invention.
- the surgical robot system 1 can comprise a surgical robot 15 and one or more positioning sensors 12 .
- the surgical robot 15 can comprise a display means 29 (including for example a display 150 shown in FIG. 10 ), and a housing 27 .
- a display 150 can be attached to the surgical robot 15
- a display means 29 can be detached from surgical robot 15 , either within surgical room 10 or in a remote location.
- the housing 27 can comprise a robot arm 23 , and an end-effectuator 30 coupled to the robot arm 23 controlled by at least one motor 160 .
- the surgical robot system 1 can include a motor assembly 155 comprising at least one motor (represented as 160 in FIG. 10 ).
- the end-effectuator 30 can comprise a surgical instrument 35 .
- the end-effectuator 30 can be coupled to the surgical instrument 35 .
- the term “end-effectuator” is used interchangeably with the terms “end-effectuator,” “effectuator element,” and “effectuator element.”
- the end-effectuator 30 can comprise any known structure for effecting the movement of the surgical instrument 35 in a desired manner.
- a three-dimensional (“3D”) image scan can be taken of a desired surgical area of the patient 18 and sent to a computer platform in communication with surgical robot 15 as described herein (see for example the platform 3400 including the computing device 3401 shown in FIG. 34 ).
- a physician can then program a desired point of insertion and trajectory for surgical instrument 35 to reach a desired anatomical target within or upon the body of patient 18 .
- the desired point of insertion and trajectory can be planned on the 3D image scan, which in some embodiments, can be displayed on display means 29 .
- a physician can plan the trajectory and desired insertion point (if any) on a computed tomography scan (hereinafter referred to as “CT scan”) of a patient 18 .
- CT scan computed tomography scan
- the CT scan can be an isocentric C-arm type scan, an O-arm type scan, or intraoperative CT scan as is known in the art.
- any known 3D image scan can be used in accordance with the embodiments of the invention described herein.
- the surgical robot system 1 can comprise a local positioning system (“LPS”) subassembly to track the position of surgical instrument 35 .
- the LPS subassembly can comprise at least one radio-frequency (RF) transmitter 120 that is coupled were affixed to the end-effectuator 30 or the surgical instrument 35 at a desired location.
- the at least one RF transmitter 120 can comprise a plurality of transmitters 120 , such as, for example, at least three RF transmitters 120 .
- the LPS subassembly can comprise at least one RF receiver 110 configured to receive one or more RF signals produced by the at least one RF transmitter 120 .
- the at least one RF receiver 110 can comprise a plurality of RF receivers 110 , such as, for example, at least three RF receivers 110 .
- the RF receivers 110 can be positioned at known locations within the room 10 where the medical procedure is to take place.
- the RF receivers 110 can be positioned at known locations within the room 10 such that the RF receivers 110 are not coplanar within a plane that is parallel to the floor of the room 10 .
- the time of flight of an RF signal from each RF transmitter 120 of the at least one RF transmitter 120 to each RF receiver 110 of the at least one RF receiver 110 can be measured to calculate the position of each RF transmitter 120 . Because the velocity of the RF signal is known, the time of flight measurements result in at least three distance measurements for each RF transmitter 120 (one to each RF receiver 110 ).
- the surgical robot system 1 can comprise a control device (for example a computer 100 having a processor and a memory coupled to the processor).
- the processor of the control device 100 can be configured to perform time of flight calculations as described herein.
- the processor of the control device 100 can be configured to provide a geometrical description of the location of the at least one RF transmitter 120 with respect to an operative end of the surgical instrument 35 or end-effectuator 30 that is utilized to perform or assist in performing an invasive procedure.
- the position of the RF transmitter 120 , as well as the dimensional profile of the surgical instrument 35 or the effectuator element 30 can be displayed on a monitor (for example on a display means 29 such as the display 150 shown in FIG. 10 ).
- the end-effectuator 30 can be a tubular element (for example a guide tube 50 ) that is positioned at a desired location with respect to, for example, a patient's 18 spine to facilitate the performance of a spinal surgery.
- the guide tube 50 can be aligned with the z axis 70 defined by a corresponding robot motor 160 or, for example, can be disposed at a selected angle relative to the z-axis 70 .
- the processor of the control device i.e. the computer 100
- the memory of the control device (computer 100 for example) can store software for performing the calculations and/or analyses required to perform many of the surgical method steps set forth herein.
- Another embodiment of the disclosed surgical robot system 1 involves the utilization of a robot 15 that is capable of moving the end-effectuator 30 along x-, y-, and z-axes (see 66 , 68 , 70 in FIG. 35B ).
- the x-axis 66 can be orthogonal to the y-axis 68 and z-axis 70
- the y-axis 68 can be orthogonal to the x-axis 66 and z-axis 70
- the z-axis 70 can be orthogonal to the x-axis 66 and the y-axis 68 .
- the robot 15 can be configured to effect movement of the end-effectuator 30 along one axis independently of the other axes. For example, in some embodiments, the robot 15 can cause the end-effectuator 30 to move a given distance along the x-axis 66 without causing any significant movement of the end-effectuator 30 along the y-axis 68 or z-axis 70 .
- the end-effectuator 30 can be configured for selective rotation about one or more of the x-axis 66 , y-axis 68 , and z-axis 70 (such that one or more of the Cardanic Euler Angles (e.g., roll, pitch, and/or yaw) associated with the end-effectuator 30 can be selectively controlled).
- the end-effectuator 30 and/or surgical instrument 35 can be aligned with a selected orientation axis (labeled “Z Tube” in FIG. 35B ) that can be selectively varied and monitored by an agent (for example computer 100 and platform 3400 ) that can operate the surgical robot system 1 .
- selective control of the axial rotation and orientation of the end-effectuator 30 can permit performance of medical procedures with significantly improved accuracy compared to conventional robots that utilize, for example, a six degree of freedom robot arm 23 comprising only rotational axes.
- the robot arm 23 that can be positioned above the body of the patient 18 , with the end-effectuator 30 selectively angled relative to the z-axis toward the body of the patient 18 .
- the robotic surgical system 1 can comprise systems for stabilizing the robotic arm 23 , the end-effectuator 30 , and/or the surgical instrument 35 at their respective positions in the event of power failure.
- the robotic arm 23 , end-effectuator 30 , and/or surgical instrument 35 can comprise a conventional worm-drive mechanism (not shown) coupled to the robotic arm 23 , configured to effect movement of the robotic arm along the z-axis 70 .
- the system for stabilizing the robotic arm 23 , end-effectuator 30 , and/or surgical instrument 35 can comprise a counterbalance coupled to the robotic arm 23 .
- the means for maintaining the robotic arm 23 , end-effectuator 30 , and/or surgical instrument 35 can comprise a conventional brake mechanism (not shown) that is coupled to at least a portion of the robotic arm 23 , such as, for example, the end-effectuator 30 , and that is configured for activation in response to a loss of power or “power off” condition of the surgical robot 15 .
- the surgical robot system 1 can comprise a plurality of positioning sensors 12 configured to receive RF signals from the at least one conventional RF transmitter (not shown) located within room 10 .
- the at least one RF transmitter 120 can be disposed on various points on the surgical robot 15 and/or on patient 18 .
- the at least one RF transmitter 120 can be attached to one or more of the housing 27 , robot arm 23 , end-effectuator 30 , and surgical instrument 35 .
- Some embodiments include positioning sensors 12 that in some embodiments comprise RF receivers 110 .
- RF receivers 110 are in communication with a computer platform as described herein (see for example 3400 comprising a computing device 3401 FIG.
- each transmitter 120 of the at least one RF transmitter 120 can transmit RF energy on a different frequency so that the identity of each transmitter 120 in the room 10 can be determined.
- the location of the at least one RF transmitter 120 , and, consequently, the objects to which the transmitters 120 are attached, are calculated by the computer (e.g., computing device 3401 in FIG. 34 ) using time-of-flight processes.
- the computer (not shown in FIG. 1 ) is also in communication with surgical robot 15 .
- a conventional processor (not shown) of the computer 100 of the computing device 3401 can be configured to effect movement of the surgical robot 15 according to a preplanned trajectory selected prior to the procedure.
- the computer 100 of the computing device 3401 can use robotic guidance software 3406 and robotic guidance data storage 3407 (shown in FIG. 34 ) to effect movement of the surgical robot 15 .
- the position of surgical instrument 35 can be dynamically updated so that surgical robot 15 is aware of the location of surgical instrument 35 at all times during the procedure. Consequently, in some embodiments, the surgical robot 15 can move the surgical instrument 35 to the desired position quickly, with minimal damage to patient 18 , and without any further assistance from a physician (unless the physician so desires). In some further embodiments, the surgical robot 15 can be configured to correct the path of surgical instrument 35 if the surgical instrument 35 strays from the selected, preplanned trajectory.
- the surgical robot 15 can be configured to permit stoppage, modification, and/or manual control of the movement of the end-effectuator 30 and/or surgical instrument 35 .
- an agent e.g., a physician or other user
- tolerance controls can be preprogrammed into the surgical robot 15 and/or processor of the computer platform 3400 (such that the movement of the end-effectuator 30 and/or surgical instrument 35 is adjusted in response to specified conditions being met).
- the surgical robot 15 can be configured to stop movement of end-effectuator 30 and/or surgical instrument 35 .
- the surgical robot 15 detects a resistance, such as a force resistance or a torque resistance above a tolerance level, then the surgical robot 15 can be configured to stop movement of end-effectuator 30 and/or surgical instrument 35 .
- the computer 100 for use in the system can be located within surgical robot 15 , or, alternatively, in another location within surgical room 10 or in a remote location. In some embodiments, the computer 100 can be positioned in operative communication with positioning sensors 12 and surgical robot 15 .
- the surgical robot 15 can also be used with existing conventional guidance systems.
- alternative conventional guidance systems beyond those specifically disclosed herein are within the scope and spirit of the invention.
- a conventional optical tracking system 3417 for tracking the location of the surgical device or a commercially available infrared optical tracking system 3417 , such as Optotrak® (Optotrak® is a registered trademark of Northern Digital Inc. Northern Digital, Waterloo, Ontario, Canada), can be used to track the patient 18 movement and the robot's base 25 location and/or intermediate axis location, and used with the surgical robot system 1 .
- Optotrak® is a registered trademark of Northern Digital Inc. Northern Digital, Waterloo, Ontario, Canada
- the surgical robot system 1 can comprise conventional optical markers attached to selected locations on the end-effectuator 30 and/or the surgical instrument 35 that are configured to emit or reflect light.
- the light emitted from and/or reflected by the markers can be read by cameras and/or optical sensors and the location of the object can be calculated through triangulation methods (such as stereo-photogrammetry).
- the surgical robot 15 can comprise a base 25 connected to wheels 31 .
- the size and mobility of these embodiments can enable the surgical robot to be readily moved from patient to patient and room to room as desired.
- the surgical robot 15 can further comprise a case 40 that is slidably attached to base 25 such that the case 40 can slide up and down along the z-axis 70 substantially perpendicular to the surface on which base 25 sits.
- the surgical robot 15 can include a display means 29 , and a housing 27 which contains robot arm 23 .
- the end-effectuator 30 can comprise a surgical instrument 35 , whereas in other embodiments, the end-effectuator 30 can be coupled to the surgical instrument 35 . In some embodiments, it is arm 23 can be connected to the end-effectuator 30 , with surgical instrument 35 being removably attached to the end-effectuator 30 .
- the effectuator element 30 can include an outer surface 30 d , and can comprise a distal end 30 a defining a beveled leading edge 30 b and a non-beveled leading edge 30 c .
- the surgical instrument 35 can be any known conventional instrument, device, hardware component, and/or attachment that is used during performance of a an invasive or non-invasive medical procedure (including surgical, therapeutic, and diagnostic procedures).
- the surgical instrument 35 can be embodied in or can comprise a needle 7405 , 7410 , a conventional probe, a conventional screw, a conventional drill, a conventional tap, a conventional catheter, a conventional scalpel forceps, or the like.
- the surgical instrument 35 can be a biological delivery device, such as, for example and without limitation, a conventional syringe, which can distribute biologically acting compounds throughout the body of a patient 18 .
- the surgical instrument 35 can comprise a guide tube 50 (also referred to herein as a “Z-tube 50 ”) that defines a central bore configured for receipt of one or more additional surgical instruments 35 .
- the surgical robot 15 is moveable in a plurality of axes (for instance x-axis 66 , y-axis 68 , and z-axis 70 ) in order to improve the ability to accurately and precisely reach a target location.
- Some embodiments include a robot 15 that moves on a Cartesian positioning system; that is, movements in different axes can occur relatively independently of one another instead of at the end of a series of joints.
- case 40 can be configured to be raised and lowered relative to the base 25 along the z-axis.
- Some embodiments include a housing 27 that can be attached to case 40 and be configured to move in the z-direction (defined by z-frame 72 ) with case 40 when case 40 is raised and lowered. Consequently, in some embodiments, arm 23 , the end-effectuator 30 , and surgical instrument 35 can be configured to move with case 40 as case 40 is raised and lowered relative to base 25 .
- housing 27 can be slidably attached to case 40 so that it can extend and retract along the x-axis 66 relative to case 40 and substantially perpendicularly to the direction case 40 moves relative to base 25 . Consequently, in some embodiments, the robot arm 23 , the end-effectuator 30 , and surgical instrument 35 can be configured to move with housing 27 as housing 27 is extended and retracted relative to case 40 .
- robot arm 23 can be extendable along the y-axis 68 relative to case 40 , base 25 , and housing 27 . Consequently, in some embodiments, the end-effectuator 30 and surgical instrument 35 can be configured to move with arm 23 as arm 23 is extended and retracted relative to housing 27 .
- arm 23 can be attached to a low profile rail system (not shown) which is encased by housing 27 .
- FIG. 6 shows an embodiment of an end-effectuator 30 that is configured to rotate about the y-axis 68 , performing a rotation having a specific roll 62 .
- FIG. 7 shows an embodiment of an end-effectuator 30 that is configured to rotate about the x-axis 66 , performing a rotation having a specific pitch 60 .
- FIG. 8 shows an embodiment of an end-effectuator 30 that is configured to raise and lower surgical instrument 35 along a substantially vertical axis, which can be a secondary movable axis 64 , referred to as “Z-tube axis 64 ”.
- the orientation of the guide tube 50 can be initially aligned with z-axis 70 , but such orientation can change in response to changes in roll 62 and/or pitch 60 .
- FIG. 9 shows a system diagram of the 3D positioning sensors 110 , computer 100 , and RF transmitters 120 in accordance with some embodiments of the invention is provided.
- computer 100 is in communication with positioning sensors 110 .
- RF transmitters 120 are attached to various points on the surgical robot 15 .
- the RF transmitters 120 can also be attached to various points on or around an anatomical target of a patient 18 .
- computer 100 can be configured to send a signal to the RF transmitters 120 , prompting the RF transmitters 120 to transmit RF signals that are read by the positioning sensors 110 .
- the computer 100 can be coupled to the RF transmitters 120 using any conventional communication means, whether wired or wireless.
- the positioning sensors 110 can be in communication with computer 100 , which can be configured to calculate the location of the positions of all the RF transmitters 120 based on time-of-flight information received from the positioning sensors 110 .
- computer 100 can be configured to dynamically update the calculated location of the surgical instrument 35 and/or end-effectuator 30 being used in the procedure, which can be displayed to the agent.
- Some embodiments can include a system diagram of surgical robot system 1 having a computer 100 , a display means 29 comprising a display 150 , user input 170 , and motors 160 , provided as illustrated in FIG. 10 .
- motors 160 can be installed in the surgical robot 15 and control the movement of the end-effectuator 30 and/or surgical instrument 35 as described above.
- computer 100 can be configured to dynamically update the location of the surgical instrument 35 being used in the procedure, and can be configured to send appropriate signals to the motors 160 such that the surgical robot 15 has a corresponding response to the information received by computer 100 .
- the computer 100 in response to information received by computer 100 , the computer 100 can be configured to prompt the motors 160 to move the surgical instrument 35 along a preplanned trajectory.
- user input 170 can be used to plan the trajectory for a desired navigation.
- a user can use the user input 170 to input the desired changes, and the computer 100 can be configured to transmit corresponding signals to the motors 160 in response to the user input 170 .
- the motors 160 can be or can comprise conventional pulse motors.
- the pulse motors can be in a conventional direct drive configuration or a belt drive and pulley combination attached to the surgical instrument 35 .
- the motors 160 can be conventional pulse motors that are attached to a conventional belt drive rack-and-pinion system or equivalent conventional power transmission component.
- the use of conventional linear pulse motors within the surgical robot 15 can permit establishment of a non-rigid position for the end-effectuator 30 and/or surgical instrument 35 .
- the end-effectuator 30 and/or surgical instrument 35 will not be fixed in a completely rigid position, but rather the end-effectuator 30 and/or the surgical instrument 35 can be configured such that an agent (e.g., a surgeon or other user) can overcome the x-axis 66 and y-axis 68 , and force the end-effectuator 30 and/or surgical instrument 35 from its current position.
- an agent e.g., a surgeon or other user
- the amount of force necessary to overcome such axes can be adjusted and configured automatically or by an agent.
- the surgical robot 15 can comprise circuitry configured to monitor one or more of: (a) the position of the robot arm 23 , the end-effectuator 30 , and/or the surgical instrument 35 along the x-axis 66 , y-axis 68 , and z-axis 70 ; (b) the rotational position (e.g., roll 62 and pitch 60 ) of the robot arm 23 , the end-effectuator 30 , and/or the surgical instrument 35 relative to the x-( 66 ), y-( 68 ), and z-( 70 ) axes; and (c) the position of the end-effectuator 30 , and/or the surgical instrument 35 along the travel of the re-orientable axis that is parallel at all times to the end-effectuator 30 and surgical instrument 35 (the Z-tube axis 64 ).
- the rotational position e.g., roll 62 and pitch 60
- circuitry for monitoring the positions of the x-axis 66 , y-axis 68 , z-axis 70 , Z-tube axis 64 , roll 62 , and/or pitch 60 can comprise relative or absolute conventional encoder units (also referred to as encoders) embedded within or functionally coupled to conventional actuators and/or bearings of at least one of the motors 160 .
- the circuitry of the surgical robot 15 can be configured to provide auditory, visual, and/or tactile feedback to the surgeon or other user when the desired amount of positional tolerance (e.g., rotational tolerance, translational tolerance, a combination thereof, or the like) for the trajectory has been exceeded.
- the positional tolerance can be configurable and defined, for example, in units of degrees and/or millimeters.
- the robot 15 moves into a selected position, ready for the surgeon to deliver a selected surgical instrument 35 , such as, for example and without limitation, a conventional screw, a biopsy needle 8110 , and the like.
- a selected surgical instrument 35 such as, for example and without limitation, a conventional screw, a biopsy needle 8110 , and the like.
- the system 1 can be configured to provide an audible warning and/or a visual warning.
- the system 1 can produce audible beeps and/or display a warning message on the display means 29 , such as “Warning: Off Trajectory,” while also displaying the axes for which an acceptable tolerance has been exceeded.
- a light illumination may be directed to the end-effectuator 30 , the guide tube 50 , the operation area (i.e. the surgical field 17 ) of the patient 18 , or a combination of these regions.
- some embodiments include at least one visual indication 900 capable of illuminating a surgical field 17 of a patient 18 .
- Some embodiments include at least one visual indication 900 capable of indicating a target lock by projecting an illumination on a surgical field 17 .
- the system 1 can provide feedback to the user regarding whether the robot 15 is locked on target.
- the system 1 can provide an alert to the user regarding whether at least one marker 720 is blocked, or whether the system 1 is actively seeking one or more markers 720 .
- the visual indication 900 can be projected by one or more conventional light emitting diodes mounted on or near the robot end-effectuator 30 .
- the visual indication can comprise lights projected on the surgical field 17 including a color indicative of the current situation.
- a green projected light could represent a locked-on-target situation, whereas in some embodiments, a red illumination could indicate a trajectory error, or obscured markers 720 .
- a yellow illumination could indicate the system 1 is actively seeking one or more markers 720 .
- the robot 15 can be configured to provide mechanical resistance (“push back” or haptic feedback) to the movement of the end-effectuator 30 and/or surgical instrument 35 in this manner, thereby promoting movement of the end-effectuator 30 and/or surgical instrument 35 back to the correct, selected orientation.
- the robot 15 can be configured to substantially immediately return the end-effectuator 30 and/or surgical instrument 35 back to the desired trajectory, at which time the audible and visual warnings and alerts can be configured to cease.
- the visual warning could include a visual indication 900 that may include a green light if no tolerances have been exceeded, or a red light if tolerances are about to, or have been exceeded.
- a conventional worm-drive system would be absolutely rigid, and a robot 15 having such a worm-drive system would be unable to be passively moved (without breaking the robot 15 ) no matter how hard the surgeon pushed.
- a completely rigid articulation system can be inherently unsafe to a patient 18 .
- a robot 15 were moving toward the patient 18 and inadvertently collided with tissues, then these tissues could be damaged.
- conventional sensors can be placed on the surface of such a robot 15 to compensate for these risks, such sensors can add considerable complexity to the overall system 1 and would be difficult to operate in a fail-safe mode.
- the end-effectuator 30 and/or surgical instrument 35 inadvertently collides with tissues of the patient 18 , a collision would occur with a more tolerable force that would be unlikely to damage such tissues.
- auditory and/or visual feedback as described above can be provided to indicate an increase in the current required to overcome the obstacle.
- the end-effectuator 30 of the robot 15 can be configured to displace itself (move away) from the inadvertently contacted tissue if a threshold required motor 160 current is encountered. In some embodiments, this threshold could be configured (by a control component, for example) for each axis such that the moderate forces associated with engagement between the tissue and the end-effectuator 30 can be recognized and/or avoided.
- the amount of rigidity associated with the positioning and orientation of the end-effectuator 30 and/or the surgical instrument 35 can be selectively varied.
- the robot 15 can be configured to shift between a high-rigidity mode and a low-rigidity mode.
- the robot 15 can be programmed so that it automatically shifts to the low-rigidity mode as the end-effectuator 30 and surgical instrument 35 are shifted from one trajectory to another, from a starting position as they approach a target trajectory and/or target position.
- the robot 15 can be configured to shift to the high-rigidity mode. In some embodiments, this mechanism may improve safety because the robot 15 would be unlikely to cause injury if it inadvertently collided with the patient 18 while in the low-rigidity mode.
- Some embodiments include a robot 15 that can be configured to effect movement of the end-effectuator 30 and/or surgical instrument 35 in a selected sequence of distinct movements.
- the x-axis 66 , y-axis 68 , roll 62 , and 60 pitch 60 orientations are all changed simultaneously, and the speed of movement of the end-effectuator 30 can be increased. Consequently, because of the range of positions through which the end-effectuator 30 travels, the likelihood of a collision with the tissue of the patient 18 can also be increased.
- the robot 15 can be configured to effect movement of the end-effectuator 30 and/or surgical instrument 35 such that the position of the end-effectuator 30 and/or surgical instrument 35 within the x-axis 66 and the y-axis 68 are adjusted before the roll 62 and pitch 60 of the end-effectuator 30 and/or surgical instrument 35 are adjusted.
- the robot 15 can be configured to effect movement of the end-effectuator 30 and/or surgical instrument 35 so that the roll 62 and pitch 60 are shifted to 0°. The position of the end-effectuator 30 and/or surgical instrument 35 within the x-axis 66 and the y-axis 68 are adjusted, and then the roll 62 and pitch 60 of the end-effectuator 30 and/or surgical instrument 35 are adjusted.
- Some embodiments include a robot 15 that can be optionally configured to ensure that the end-effectuator 30 and/or surgical instrument 35 are moved vertically along the z-axis 70 (away from the patient 18 ) by a selected amount before a change in the position and/or trajectory of the end-effectuator 30 and/or surgical instrument 35 is effected.
- the robot 15 when an agent (for example, a surgeon or other user, or equipment) changes the trajectory of the end-effectuator 30 and/or surgical instrument 35 from a first trajectory to a second trajectory, the robot 15 can be configured to vertically displace the end-effectuator 30 and/or surgical instrument 35 from the body of the patient 18 along the z-axis 70 by the selected amount (while adjusting x-axis 66 and y-axis 68 configurations to remain on the first trajectory vector, for example), and then effecting the change in position and/or orientation of the end-effectuator 30 and/or surgical instrument 35 . This ensures that the end-effectuator 30 and/or surgical instrument 35 do not move laterally while embedded within the tissue of the patient 18 .
- an agent for example, a surgeon or other user, or equipment
- the robot 15 can be configured to produce a warning message that seeks confirmation from the agent (for example, a surgeon or other user, or equipment) that it is safe to proceed with a change in the trajectory of the end-effectuator 30 and/or surgical instrument 35 without first displacing the end-effectuator 30 and/or surgical instrument 35 along the z-axis.
- the agent for example, a surgeon or other user, or equipment
- At least one conventional force sensor can be coupled to the end-effectuator 30 and/or surgical instrument 35 such that the at least one force sensor receives forces applied along the orientation axis (Z-tube axis 64 ) to the surgical instrument 35 .
- the at least one force sensor can be configured to produce a digital signal.
- the digital signal can be indicative of the force that is applied in the direction of the Z-tube axis 64 to the surgical instrument 35 by the body of the patient 18 as the surgical instrument 35 advances into the tissue of the patient 18 .
- the at least one force sensor can be a small conventional uniaxial load cell based on a conventional strain gauge mechanism.
- the uniaxial load cell can be coupled to, for example, analog-to-digital filtering to supply a continuous digital data stream to the system 1 .
- the at least one force sensor can be configured to substantially continuously produce signals indicative of the force that is currently being applied to the surgical instrument 35 .
- the surgical instrument 35 can be advanced into the tissue of the patient 18 by lowering the z-axis 70 while the position of the end-effectuator 30 and/or surgical instrument 35 along the x-axis 66 and y-axes 68 is adjusted such that alignment with the selected trajectory vector is substantially maintained.
- the roll 62 and pitch 60 orientations can remain constant or self-adjust during movement of the x-( 66 ), y-( 68 ), and z-( 70 ) axes such that the surgical instrument 35 remains oriented along the selected trajectory vector.
- the position of the end-effectuator 30 along the z-axis 70 can be locked at a selected mid-range position (spaced a selected distance from the patient 18 ) as the surgical instrument 35 advances into the tissue of the patient 18 .
- the stiffness of the end-effectuator 30 and/or the surgical instrument 35 can be set at a selected level as further described herein.
- the stiffness of the Z-tube axis 64 position of the end-effectuator 30 and/or the surgical instrument 35 can be coupled to a conventional mechanical lock (not shown) configured to impart desired longitudinal stiffness characteristics to the end-effectuator 30 and/or surgical instrument 35 .
- the counterforce applied by the tissue of the patient 18 during penetration of the surgical instrument 35 can oppose the direction of advancement of the surgical instrument 35 such that the surgical instrument 35 cannot advance along the selected trajectory vector.
- the Z-tube axis 64 can be forced up and there can be no net advancement of the surgical instrument 35 .
- the at least one force sensor can permit an agent (for example, a surgeon or other user, or equipment) to determine, (based on sudden increase in the level of applied force monitored by the force sensor at the end-effectuator 30 and/or the surgical instrument 35 ), when the surgical instrument 35 has encountered a bone or other specific structure within the body of the patient 18 .
- an agent for example, a surgeon or other user, or equipment
- the orientation angle of the end-effectuator 30 and/or surgical instrument 35 and the x-axis 66 and y-axis 68 can be configured to align the Z-tube axis 64 with the desired trajectory vector at a fully retracted Z-tube position, while a z-axis 70 position is set in which the distal tip of the surgical instrument 35 is poised to enter tissue.
- the end-effectuator 30 can be positioned in a manner that the end-effectuator 30 can move, for example, exactly or substantially exactly down the trajectory vector if it were advanced only along guide tube 50 .
- advancing the Z-tube axis 64 can cause the guide tube 50 to enter into tissue, and an agent (a surgeon or other user, equipment, etc.) can monitor change in force from the load sensor. Advancement can continue until a sudden increase in applied force is detected at the time the surgical instrument 35 contacts bone.
- an agent a surgeon or other user, equipment, etc.
- the robot 15 can be configured to deactivate the one or more motors 160 that advance the Z-tube axis 64 such that the end-effectuator 30 and/or the surgical instrument 35 can move freely in the Z-tube axis 64 direction while the position of the end-effectuator 30 and/or the surgical instrument 35 continues to be monitored.
- the surgeon can then push the end-effectuator 30 down along the Z-tube axis 64 , (which coincides with the desired trajectory vector) by hand.
- the position of the surgical instrument 35 can be corrected by adjustment along the x-( 66 ) and/or y-( 68 ) axes and/or in the roll 62 and/or pitch 60 directions.
- the agent for example, a surgeon or other user, or equipment
- the surgical instrument 35 can manually force the surgical instrument 35 to advance until a tactile sense of the surgical instrument 35 contacts bone, or another known region of the body).
- the robotic surgical system 1 can comprise a plurality of conventional tracking markers 720 configured to track the movement of the robot arm 23 , the end-effectuator 30 , and/or the surgical instrument 35 in three dimensions. It should be appreciated that three dimensional positional information from tracking markers 720 can be used in conjunction with the one dimensional linear positional information from absolute or relative conventional linear encoders on each axis of the robot 15 to maintain a high degree of accuracy.
- the plurality of tracking markers 720 can be mounted (or otherwise secured) thereon an outer surface of the robot 15 , such as, for example and without limitation, on the base 25 of the robot 15 , or the robot arm 23 .
- the plurality of tracking markers 720 can be configured to track the movement of the robot 15 arm, the end-effectuator 30 , and/or the surgical instrument 35 .
- the computer 100 can utilize the tracking information to calculate the orientation and coordinates of the distal tip 30 a of the surgical instrument 35 based on encoder counts along the x-axis 66 , y-axis 68 , z-axis 70 , the Z-tube axis 64 , and the roll 62 and pitch 60 axes.
- the plurality of tracking markers 720 can be positioned on the base 25 of the robot 15 spaced from the surgical field 17 to reduce the likelihood of being obscured by the surgeon, surgical tools, or other parts of the robot 15 .
- At least one tracking marker 720 of the plurality of tracking markers 720 can be mounted or otherwise secured to the end-effectuator 30 .
- the positioning of one or more tracking markers 720 on the end-effectuator 30 can maximize the accuracy of the positional measurements by serving to check or verify the end-effectuator 30 position (calculated from the positional information from the markers on the base 25 of the robot 15 and the encoder counts of the x-( 66 ), y-( 68 ), roll 62 , pitch 60 , and Z-tube axes 64 ).
- At least one optical marker of the plurality of optical tracking markers 720 can be positioned on the robot 15 between the base 25 of the robot 15 and the end-effectuator 30 instead of, or in addition to, the markers 720 on the base 25 of the robot 15 , (see FIG. 16 ).
- the at least one tracking marker 720 can be mounted to a portion of the robot 15 that effects movement of the end-effectuator 30 and/or surgical instrument 35 along the x-axis to enable the tracking marker 720 to move along the x-axis 66 as the end-effectuator 30 and surgical instrument 35 move along the x-axis 66 .
- the placement of the tracking markers 720 in this way can reduce the likelihood of a surgeon blocking the tracking marker 720 from the cameras or detection device, or the tracking marker 720 becoming an obstruction to surgery.
- the tracking marker 720 because of the high accuracy in calculating the orientation and position of the end-effectuator 30 based on the tracking marker 720 outputs and/or encoder counts from each axis, it can be possible to very accurately determine the position of the end-effectuator 30 .
- the placement of markers 720 on any intermediate axis of the robot 15 can permit the exact position of the end-effectuator 30 to be calculated based on location of such markers 720 and counts of encoders on axes ( 66 , 62 , 60 , 64 ) between the markers 720 and the end-effectuator 30 .
- the order of axes from the base 25 to the end-effectuator 30 is z-( 70 ) then x-( 66 ) then y-( 68 ) then roll 62 then pitch 60 then Z-tube 64 .
- tracking markers 720 are placed on the housing 27 of the robot 15 that moves with the roll 62 axis, the locations of such tracking markers 720 and the encoder counts of the pitch 60 and Z-tube axes 64 can be sufficient to calculate the end-effectuator 30 position.
- the surgical instrument 35 when the surgical instrument 35 is advanced into the tissue of the patient 18 with the assistance of a guide tube 50 , the surgical instrument 35 can comprise a stop mechanism 52 that is configured to prevent the surgical instrument 35 from advancing when it reaches a predetermined amount of protrusion (see for example, FIGS. 17A-B ).
- a stop mechanism 52 that is configured to prevent the surgical instrument 35 from advancing when it reaches a predetermined amount of protrusion (see for example, FIGS. 17A-B ).
- the robot 15 can substantially continuously monitor the protrusion distance, and in some embodiments, the distance can be displayed on a display (such as display means 29 ).
- protrusion distance can be substantially continuously monitored using a spring-loaded plunger 54 including a spring-loaded mechanism 55 a and sensor pad 55 b that has a coupled wiper 56 (see for example FIG. 17B ).
- the stop mechanism 52 on the surgical instrument 35 can be configured to contact the spring-loaded mechanism 55 well before it encounters the end of the guide tube 50 .
- any conventional linear encoding mechanism can be used to monitor the plunger's depth of depression and transmit that information to the computer 100 as further described herein.
- FIGS. 17C-17E depict tools for manually adjusting a drill stop 46 with reference to drill bit markings 44 in accordance with one embodiment of the invention.
- the drill bit 42 can include release mechanisms 48 on each end of the drill stop 46 .
- the release 48 on one end of the drill stop 46 is pulled, it is possible to move the drill stop 46 up the shaft of the drill bit 42 .
- the release 48 on the other end of the drill stop 46 is pulled, it is possible to move the drill stop 46 down the shaft (see the direction of movement in FIGS. 17D and 17E ).
- neither release mechanism 48 is pulled, the drill stop 46 will not move in either direction, even if bumped.
- Some embodiments include the ability to lock and hold the drill bit 42 in a set position relative to the tube 50 in which it is housed.
- the drill bit 42 can be locked by locking the drill stop 46 relative to the tube 50 using a locking mechanism.
- FIGS. 17F-J illustrates tools for locking and holding a drill bit 42 in a set position in accordance with one embodiment of the invention.
- the locking mechanism 49 shown in FIG. 17 H can comprise two clam shells 49 (shown in FIG. 17F ).
- a drill bit 42 can be locked into position by assembling the clam shells around the drill stop 46 (shown in FIG. 17G ).
- This feature allows the user to lock the drill bit 42 in a position such that the tip slightly protrudes past the end of the tube 50 (see FIGS. 171 and 17J ). In this position, the user can force the tube 50 to penetrate through soft tissues to force the tube 50 to contact bone (for example during a percutaneous spine screw insertion).
- the end-effectuator 30 can be configured not block the tracking optical markers 720 or interfere with the surgeon.
- the end-effectuator 30 can comprise a clearance mechanism 33 including an actuator 33 a that permits this configuration, as depicted in FIGS. 18A and 18B .
- the guide tube 50 can be secured within a housing of the end-effectuator 30 with two shafts 32 .
- the shafts 32 move relative to one other, due to a parallelogram effect of the clearance mechanism 33 , the position of the guide tube 50 can mimic the position of the end-effectuator 30 (see FIG. 18B ).
- the end-effectuator 30 can comprise an attachment element 37 that is configured to apply such forces (see for example FIGS. 19A-B ).
- the end-effectuator 30 attachment element 37 can be configured for coupling to the end-effectuator 30 at substantially the same location as the clearance mechanism 33 .
- the end-effectuator 30 with attachment element 37 snaps into the same place as the end-effectuator 30 without the attachment element 37 .
- the relative movement of the two shafts 32 caused by angulation 30 e will not cause movement in the pitch 60 direction and will instead cause distraction (illustrated as moving from an attachment element 37 distance 37 a in FIG. 19A to distance 37 b in FIG. 19B ).
- distraction illustrated as moving from an attachment element 37 distance 37 a in FIG. 19A to distance 37 b in FIG. 19B .
- shaft 32 movement as shown in FIGS. 19A-B would cause distraction
- rotation of the actuator 33 a in the opposite direction to that represented by 30 e would cause compression (i.e. the distance 37 b in FIG. 19B would move towards the distance 37 a in FIG. 19A ).
- the methods disclosed in the various embodiments described throughout the subject specification can be stored on an article of manufacture, or computer-readable medium, to facilitate transporting and transferring such methods to a computing device (e.g., a desktop computer, a mobile computer, a mobile telephone, a blade computer, a programmable logic controller, and the like) for execution, and thus implementation, by a processor of the computing device or for storage in a memory thereof.
- a computing device e.g., a desktop computer, a mobile computer, a mobile telephone, a blade computer, a programmable logic controller, and the like
- the surgical robot 15 can adjust its position automatically continuously or substantially continuously in order to move the end-effectuator 30 to an intended (i.e. planned) position.
- the surgical robot 15 can adjust its position automatically continuously or substantially continuously based on the current position of the end-effectuator 30 and surgical target as provided by a current snapshot of tracking markers, LPS, or other tracking data. It should further be appreciated that certain position adjustment strategies can be inefficient.
- an inefficient strategy for the robot 15 to find a target location can be an iterative algorithm to estimate the necessary direction of movement, move toward the target location, and then assess a mismatch between a current location and the target location (the mismatch referred to as an error), and estimate a new direction, repeating the cycle of estimate-movement-assessment until the target location is reached within a satisfactory error.
- the position adjustment strategies in accordance with some embodiments of the invention are substantively more efficient than iterative strategies.
- a surgical robot 15 can make movements and adjust its location by calibrating the relative directions of motions in each axis (permitting computation via execution of software or firmware with the computer 100 ) at each frame of tracking data, of a unique set of necessary motor encoder counts that can cause each of the individual axes to move to the correct location.
- the Cartesian design of the disclosed robot 15 can permit such a calibration to be made by establishing a coordinate system for the robot 15 and determining key axes of rotation.
- methods for calibrating the relative directions of the robot's 15 axes can utilize a sequence of carefully planned movements, each in a single axis.
- temporary tracking markers 720 are attached to the end-effectuator 30 to capture the motion of the end-effectuator 30 . It should be appreciated that the disclosed methods do not require the axes of the robot 15 to be exactly or substantially perpendicular, nor do they require the vector along which a particular axis moves (such as the x-axis 66 ) to coincide with the vector about which rotation occurs (such as pitch 60 , which occurs primarily about the x-axis 66 ).
- the disclosed methods include motion along a specific robot 15 axis that occurs in a straight line.
- the disclosed methods for calibrating the relative directions of movement of the robot's 15 axes can utilize one or more frames of tracking data captured at the ends of individual moves made in x-( 66 ), y-( 68 ), roll ( 62 ), pitch ( 60 ), and Z-tube axes 64 from markers 720 temporarily attached to the end-effectuator's 30 guide tube 50 .
- all other axes can be configured at the zero position (for example, the position where the encoder for the axis reads 0 counts).
- one or more frames of tracking data with all robot 15 axes at 0 counts may be necessary, and one or more frames of data with the temporary markers 720 rotated to a different position about the longitudinal axis of the guide tube 50 may be necessary.
- the marker 720 positions from these moves can be used to establish a Cartesian coordinate system for the robot 15 in which the origin (0,0,0) is through the center of the end-effectuator 30 and is at the location along the end-effectuator 30 closest to where pitch 60 occurs.
- this coordinate system can be rotated to an alignment in which y-axis 68 movement of the robot 15 can occur exactly or substantially along the coordinate system's y-axis 68 , while x-axis 66 movement of the robot 15 occurs substantially perpendicular to the y-axis 68 , but by construction of the coordinate system, without resulting in any change in the z-axis 70 coordinate.
- the steps for establishing the robot's 15 coordinate system based at least on the foregoing individual moves can comprise the following: First, from the initial and final positions of the manual rotation of tracking markers 720 about the long axis of the end-effectuator 30 , a finite helical axis of motion is calculated, which can be represented by a vector that is centered in and aligned with the end-effectuator 30 . It should be appreciated that methods for calculating a finite helical axis of motion from two positions of three or more markers are described in the literature, for example, by Spoor and Veldpaus (Spoor, C. W. and F. E.
- the vector that is centered in and aligned with the end-effectuator 30 can be defined, or constructed, by interconnecting two points that are attached to two separate rigid bodies that can be temporarily affixed to the entry and exit of the guide tube 50 on the Z-tube axis 64 .
- each of the two rigid bodies can include at least one tracking marker 720 (e.g., one tracking marker 720 , two tracking markers 720 , three tracking markers 720 , more than three tracking markers 720 , etc.), and a calibration can be performed that provides information indicative of the locations on the rigid bodies that are adjacent to the entry and exit of the guide tube 50 relative to the tracking markers.
- a tracking marker 720 e.g., one tracking marker 720 , two tracking markers 720 , three tracking markers 720 , more than three tracking markers 720 , etc.
- a second helical axis can be calculated from the pitch 60 movements, providing a vector substantially parallel to the x-axis of the robot 15 but also close to perpendicular with the first helical axis calculated.
- the closest point on the first helical axis to the second helical axis (or vector aligned with the end-effectuator 30 ) is calculated using simple geometry and used to define the origin of the robot's coordinate system (0,0,0).
- a third helical axis is calculated from the two positions of the roll 62 axis. In certain scenarios, it cannot be assumed that the vector about which roll occurs (third helical axis) and the vector along which the y-axis 68 moves are exactly or substantially parallel.
- Vectors for x-axis 66 and y-axis 68 motion can be determined from neutral and extended positions of x-axis 66 and y-axis 68 and stored separately.
- the coordinate system can be realigned to enable y-axis movement of the robot 15 to occur exactly or substantially in the y-axis 68 direction of the coordinate system, and x-axis 66 movement of the robot 15 without any change in the z-coordinate ( 70 ).
- a series of rotations about a coordinate axis is performed and applied to every point of interest in the current coordinate system. Each point is then considered to be represented in the new coordinate system.
- the vector can be pre-multiplied by a 3 ⁇ 3 rotation matrix.
- the 3 ⁇ 3 rotation matrix for a rotation of Rx degrees about the x-axis is:
- the 3 ⁇ 3 rotation matrix for a rotation of R y degrees about the y-axis is:
- the 3 ⁇ 3 rotation matrix for a rotation of R z degrees about the z-axis is:
- a series of three rotations can be performed.
- a sequence of three rotations can comprise (1) a rotation about x using an R x value appropriate to rotate the y-movement vector until its z coordinate equal 0, followed by (2) a rotation about z using an R z value appropriate to rotate the y-movement vector until its x coordinate equal 0, followed by (3) a rotation about y using an R y value appropriate to rotate the x-movement vector until its z coordinate equals 0.
- the arctangent function can be utilized.
- the new coordinate system is aligned such that the y-movement axis of the surgical robot 15 is exactly or substantially exactly aligned with the coordinate system's y-axis 68
- the roll 62 rotation movement of the robot 15 should not be assumed to occur exactly or substantially exactly about a vector aligned with the coordinate system's y-axis 68
- the pitch 60 movement of the surgical robot 15 should not be assumed to occur exactly or substantially exactly about a vector aligned with the coordinate system's x-axis.
- rotational movement there can be linear and orientational “offsets” from the helical axis of motion to the nearest coordinate axis.
- offsets can be calculated and retained (e.g., stored in a computing device's memory) so that for any rotation occurring during operation, the offsets can be applied, rotation can be performed, and then negative offsets can be applied so that positional change occurring with rotation motion accounts for the true center of rotation.
- the desired trajectory can be first calculated in the medical image coordinate system, then transformed to the robot 15 coordinate system based at least on known relative locations of active markers.
- active markers for example, in some embodiments, conventional light-emitting markers and/or conventional reflective markers associated with an optical tracking system 3417 can be used (see for example active markers 720 in FIG. 20A ).
- conventional electromagnetic sensors associated with an electromagnetic tracking system can be used.
- radio-opaque markers for example markers 730 shown in FIG. 20A
- radio-opaque markers 730 (spheres formed, at least in part from metal or other dense material), can be used to provide a marker 730 that can at least partially absorb x-rays to produce a highly contrasted image of the sphere in a CT scan image.
- the number of counts necessary to achieve the desired pitch 60 can be determined, keeping track of how this pitch 60 can affect roll 62 .
- the change in pitch angle 60 can be multiplied by the previously calibrated motor counts per degree for pitch.
- the change in roll 62 caused by this change in pitch 60 can be calculated from the orientation of the helical axis and the rotation angle (pitch) about the helical axis. Then, the necessary roll 62 to get to the desired roll 62 to reach the planned trajectory alignment can be calculated, with the benefit that applying roll 62 does not, by definition of the coordinate system, result in any further change in pitch.
- the coordinate system is then shifted back to the previously described robot 15 coordinate system by the inverse of the three rotations applied above. Then the necessary counts to reach the desired x-axis 66 position can be calculated, also keeping track of how this x-axis 66 position change will affect y-axis 68 position. Then the necessary y-axis 68 counts to reach the desired y-axis position can be readily calculated with the benefit that changing the y-axis 68 coordinate can have no effect on any other axis since the y-axis motion vector is by definition aligned with the robot's y-axis 68 .
- the orientation of the Z-tube 50 movement vector is adjusted when adjusting roll 62 and pitch 60 and the counts necessary to move it to the desired position along the trajectory vector is calculated from the offset.
- these counts can be sent as computer-accessible instructions (e.g., computer-readable and/or computer-executable instructions) to respective controllers for each axis in order to move the axes to the computed positions.
- FIG. 24 is a flowchart of a method 2400 for positioning and advancing through soft tissue in accordance with one or more aspects according to one embodiment of the invention.
- a medical image is accessed (e.g., received, retrieved, or otherwise acquired).
- the medical image can be a 3D anatomical image scan including, but not limited to a CT scan, a magnetic resonance imaging scan (hereinafter referred to as an “MRI scan”), an X-ray image, or other anatomical scan. It should be appreciated that any 3D anatomical scan may be utilized with the surgical robot 15 and is within the scope of the present invention.
- a targeting fixture 690 is calibrated to the medical image.
- the calibration can be semi-automated or automated.
- data indicative of an intended trajectory associated with the medical image is received.
- a robot 15 is substantially maintained on the intended trajectory.
- a control platform for example, platform 3400 shown in FIG. 34 ) can adjust movement of the robot 15 in order to substantially maintain the intended trajectory.
- FIGS. 25-26 are flowcharts of methods for calibrating a targeting fixture 690 to a medical image in accordance with one or more embodiments of the invention.
- the method 2500 can embody a semi-automated calibration method and can be implemented (e.g., executed) as part of block 2420 in certain scenarios.
- data indicative of a medical image having a representation of a plurality of radio-opaque markers is received.
- such plurality can contain four radio-opaque markers 730 .
- a geometrical center for each radio-opaque marker 730 is determined in a coordinate system associated with the medical image.
- image thresholding can be utilized to define one or more edges of each radio-opaque marker 730 and a geometrical center thereof. Thresholding refers to an image processing technique in which pixel intensity within a 2D region can be monitored. For example, the x, y positions (for instance expressed in mm) of pixels of an intensity that reach a predetermined value can be retrieved. Stated similarly, the threshold refers to the transition pixel intensity from light to dark.
- the radio-opaque marker 730 can appear light and the adjacent space (such as tissue or air) can appear dark. In some embodiments, displaying pixels that satisfy a thresholding criterion at an intensity encountered at the edge of a radio-opaque marker can yield a largely circular trace outlining the marker on the medical image. Since in some embodiments, markers 730 can be spherical, a method for finding the center of the marker 730 in a 2D view can include firstly restricting the 2D view to a sampling region with the high-intensity image of the sphere toward the center of the region and pixels of lower intensity toward the outer edges of the region.
- the method can include finding the mean x threshold position (e.g., the maximum x coordinate of pixels satisfying the threshold criterion plus minimum x coordinate of pixels satisfying the threshold criterion divided by two), and finding the mean y threshold position using a similar method.
- the center of the sphere can be found by determining 2D centers of slices through the same marker 730 in two orthogonal views.
- the method can include finding mean x and mean y from an xy slice, then finding mean x and mean z from an xz slice to get a mean x, y, and z axis coordinate representing the center of the marker 730 .
- new xy and xz slices can be evaluated again and the maximum and minimum x, y, and z threshold values can be again determined to evaluate the dimensions of the thresholded object in each view.
- a non-spherical object of high intensity such as a small process of cortical bone extending away from the side of the spine, may fail to satisfy (1) a condition where there is high intensity near the middle of the region, but low intensity all around, since the process may extend out of the region in one or more directions; or (2) a condition where the dimensions in x, y, and z of the centered object do not match each other (e.g., non-spherical case).
- each centered sphere is determined for each radio-opaque marker 730 for the fixture being calibrated. In some embodiments, when at least one such sphere is not determined, or identified, the threshold setting is adjusted and flow is directed to block 2510 . In some embodiments, at block 2540 , each centered sphere is mapped to each radio-opaque marker 730 of the plurality of radio-opaque markers 730 .
- block 2540 can represent a mapping action which, in some embodiments, can comprise implementing a sorting process to establish a specific centered sphere is associated with a specific one of the plurality of radio-opaque markers 730 .
- a plurality of radio-opaque markers 730 contains four radio-opaque markers 730 (represented, for example, as OP1, OP2, OP3, and OP4).
- the sorting process can map each one of four centered markers 730 to one of OP1, OP2, OP3, or OP4.
- the sorting process can distinguish a specific marker 730 by measuring inter-marker distances from mean positions of the four unidentified markers 730 , and comparing such distances to extant inter-marker distances (for example, those that are pre-measured and retained in memory, such as mass storage device 3404 ) for each marker 730 on a marker fixture.
- the opaque markers 730 on the fixture 690 can be placed asymmetrically, each marker 730 can be identified from a unique set of inter-marker distances corresponding to such marker 730 .
- a single physical marker 730 (one of OP1, OP2, OP3, or OP4) can have a matching inter-marker distance sum within a specified tolerance (such as ⁇ 1 mm) of D.
- coordinates of each centered sphere can be retained (for example in memory of a computer platform 3400 ). As described herein, in some embodiments, such coordinates can be utilized in a process for tracking movement of a robot 15 .
- Some embodiments include method 2600 (shown as a flowchart in FIG. 26 ) that can embody an automated calibration method and can be implemented (e.g., executed) as part of block 2420 in certain scenarios.
- an x-y grid of test area squares is created at block 2605 .
- each test area square can be larger than the diameter of a sphere (a radio-opaque marker 730 ) associated with a targeting fixture 690 comprised of material that, when imaged, appears as opaque.
- each test area square can be at least partially overlapping with at least one adjacent test area square.
- a nearly half the surface of a test area square can overlap with the surface of an adjacent test area square.
- borders of a medical image within a first test area square are determined. It should be appreciated that in some embodiments, a sphere can be rendered as a circular area, but a section of bone represented in the medical image can be asymmetrical.
- a thresholding process in accordance with one or more aspects described herein can be implemented to exclude one or more invalid markers 730 by assessing if the x, y, and z axes boundaries of the object are of substantially equivalent dimensions, consistent with the shape being spherical.
- a maximum (max) border coordinate is less than the maximum coordinate of the test area, and a minimum (min) border coordinate is greater than the minimum coordinate of the test area, and vertical span of features rendered in the image are equal or substantially equal to horizontal span of such features.
- flow is directed to block 2645 , at which the first test area is moved to next grid location and next Z plane.
- flow is directed to block 2625 , at which X coordinate and Y coordinate are centered at the center of the current test area.
- Z coordinate is probed by creating a second test area square spanning upwards and downwards in XZ plane and/or YZ plane to determine one or more borders of an object.
- borders are of different spacing, flow is directed to block 2645 .
- a sphere having a center at X coordinate, Y coordinate, and Z coordinate is identified at block 2640 and flow is directed to block 2645 .
- the flow is directed to block 2615 , in which the first area is the updated instance of a prior first area, with the flow reiterating one or more of blocks 2620 through 2645 .
- flow is directed to block 2655 at which invalid marker(s) 730 can be excluded.
- a paring process can be implemented to exclude one or more invalid markers 730 .
- the known spacings between each of the N radio-opaque markers 730 (with N a natural number) on the targeting fixture 690 and each other radio-opaque marker 730 on the targeting fixture 690 can be compared to the markers 730 that have been found on the medical image.
- any sphere found on the medical image that does not have spacings relative to N ⁇ 1 other markers 730 that are within an acceptable tolerance of known spacings retained, for example, on a list can be considered to be invalid.
- a targeting fixture 690 has four radio-opaque markers 730 , there are six known spacings, with each marker 730 having a quantifiable spacing relative to three other markers 730 : the inter-marker spacings for markers 1-2, 1-3, 1-4, 2-3, 2-4, and 3-4.
- the inter-marker spacings for markers 1-2, 1-3, 1-4, 2-3, 2-4, and 3-4.
- their inter-marker spacings can be calculated.
- each of the five potential markers 730 individually, if any one of such five markers 730 does not have three of its four inter-marker spacings within a very small distance of the spacings on the list of six previously quantified known spacings, it is considered invalid.
- each centered radio-opaque marker 730 can be mapped to each radio-opaque marker 730 of a plurality of radio-opaque markers 730 .
- a sorting process in accordance with one or more aspects described herein can be implemented to map such markers 730 to radio opaque markers 730 .
- coordinates of each centered sphere can be retained (e.g., in memory of a computer platform 3400 ). As described herein, in some embodiments, such coordinates can be utilized in a process for tracking movement of a robot 15 .
- the established (e.g., calibrated) spatial relationship between active markers 720 and radio-opaque markers 730 can be utilized to transform the coordinate system from the coordinate system of the medical image to the coordinate system of the tracking system 3417 , or vice versa.
- Some embodiments include a process for transforming coordinates from the medical image's coordinate system to the tracking system's coordinate system can include a fixture 690 comprising four radio-opaque markers OP1, OP2, OP3, and OP4 (for example radio-opaque markers 730 ) in a rigidly fixed position relative to four active markers AM1, AM2, AM3, AM4 (for example, active markers 720 ).
- this positional relationship can be retained in a computer memory (e.g., system memory 3412 ) for later access on real-time or substantially on real-time in a set of four arbitrary reference Cartesian coordinate systems that can be readily reachable through transformations at any later frame of data.
- each reference coordinate system can utilize an unambiguous positioning of three of the active markers 720 .
- the information retained for later access can be the coordinates of AM1-AM3 and OP1-OP4 in such reference coordinate system.
- another such reference coordinate system can transform OP1-OP4 by utilizing AM2, AM3, and AM4.
- another such reference coordinate system can transform OP1-OP4 by utilizing AM1, AM3, and AM4.
- another such reference coordinate system can transform OP1-OP4 by utilizing AM1, AM2, and AM4.
- the coordinates of the active markers AM1-AM4 can be provided by the tracking system 3417 .
- transformations suitable to reach the conditions of the reference coordinate system can be applied.
- such transformations can position AM1, AM2, and AM3 on the x-y plane in a position in proximity to the position that was earlier stored in computer memory for this reference coordinate system.
- a least squares algorithm can be utilized to apply an offset and rotation to the triad of markers 720 .
- the least squares algorithm can be implemented as described by Sneath (Sneath P. H. A., Trend-surface analysis of transformation grids, J. Zoology 151, 65-122 (1967)).
- transformations suitable to reach the reference coordinate system can be retained in memory (e.g., system memory 3412 and/or mass storage device 3404 ).
- the retained coordinates of OP1-OP4 in such reference coordinate system can be retrieved and the inverse of the retained transformations to reach the reference coordinate system can be applied to such coordinates.
- the new coordinates of OP1-OP4 (the coordinates resulting from application of the inverse of the transformations) are in the coordinate system of the tracking system 3417 .
- the coordinates of OP1-OP4 can be retrieved.
- the four sets of OP1-OP4 coordinates in the tracking system's coordinate system that can be calculated from different triads of active markers 720 are contemplated to have coordinates that are approximately equivalent.
- the data set can be analyzed to determine which of the active markers 720 provides non-suitable (or poor) data by assessing how accurately each triad of active markers 720 at the current frame overlays onto the retained positions of active markers 720 .
- a mean value obtained from the four sets can be utilized for each radio-opaque marker 730 .
- the same transformations can be applied to the data.
- the tip and tail of a trajectory vector can be transformed to the four reference coordinate systems and then retrieved with triads of active markers 720 at any frame of data and transformed to the tracking system's coordinate system.
- FIG. 27 is a flowchart of a method 2700 for automatically maintaining a surgical robot 15 substantially on a trajectory in accordance some embodiments of the invention.
- data indicative of position of one or more of Z-frame 72 or Z-tube 50 are received.
- data indicative of each robot 15 joint in the surgical robot 15 such as encoder counts from each axis motor 160 , are accessed.
- a current robot 15 position on a planned trajectory is accessed (the position being represented in a camera coordinate system or the coordinate system of other tracking device).
- the planned trajectory can be generated by an operator.
- the operator e.g., a surgeon
- the operator can scroll and rotate through the image slices until the desired anatomy can be viewed on three windows representing three orthogonal planes (typically sagittal, coronal, and axial slices).
- the operator can then draw a line at the desired slope and location on one window; the line simultaneously is calculated and appears on the other two windows, constrained by the views of the screens and orientation on the window on which it was drawn.
- a line (e.g., referred to as line t) that is fixed on the image both in angle and position represents the desired trajectory; the surgeon has to rotate and scroll the images to align this trajectory to the desired location and orientation on the anatomy.
- a planned trajectory can be retained in a memory of a computing device (for example, computing device 3401 ) that controls the surgical robot 15 or is coupled thereto for use during a specific procedure.
- each planned trajectory can be associated with a descriptor that can be retained in memory with the planned trajectory.
- the descriptor can be the level and side of the spine where screw insertion is planned.
- the line t that is (fixed on the image both in angle and position representing the desired trajectory) is dictated by the current position of the robot's end effectuator 30 , or by an extrapolation of the end effectuator guide tube 50 if an instrument 35 were to extend from it along the same vector.
- the position of this extrapolated line is updated on the medical image, based on markers 720 attached to the robot, conventional encoders showing current position of an axis, or a combination of these registers.
- that vector's position in the medical image coordinate system is stored into the computer memory (for example in memory of a computer platform 3400 ) so that later, when recalled, the robot 15 will move automatically in the horizontal plane to intersect with this vector.
- the robot's axes can be put in a passive state.
- the markers 720 in the passive state, continue to collect data on the robot arm 23 position and encoders on each axis 64 , 66 , 68 , 70 continue to provide information regarding the position of the axis; therefore the position of an extrapolated line can be updated on the medical image as the passive robot 15 is dragged into any orientation and position in the horizontal plane.
- the position when a desired trajectory is reached, the position can be stored into the computer memory.
- Some embodiments include conventional software control or a conventional switch activation capable of placing the robot 15 into an active state to immediately rigidly hold the position or trajectory, and to begin compensating for movement of the patient 18 .
- the computing device that implements the method 2700 or that is coupled to the surgical robot 15 can render one or more planned trajectories. Such information can permit confirming that the trajectories planned are within the range of the robot's 15 reach by calculating the necessary motor 160 encoder counts to reach each desired trajectory, and assessing if the counts are within the range of possible counts of each axis.
- information including whether each trajectory is in range, and how close each trajectory is to being out of range can be provided to an agent (such as a surgeon or other user, or equipment).
- a display means 29 (such as a display device 3411 ) can render (i.e. display) the limits of axis counts or linear or angular positions of one or more axes and the position on each axis where each targeted trajectory is currently located.
- the display device 3411 (for example, a display 150 ) can render a view of the horizontal work field as a rectangle with the robot's x-axis 66 movement and y-axis 68 movement ranges defining the horizontal and vertical dimensions of the rectangle, respectively.
- marks (for example, circles) on the rectangle can represent the position of each planned trajectory at the current physical location of the robot 15 relative to the patient 18 .
- a 3D Cartesian volume can represent the x-axis 66 movement, y-axis 68 movement and z-axis 70 movement ranges of the robot 15 .
- line segments or cylinders rendered in the volume can represent the position of each planned trajectory at the current location of the robot 15 relative to the patient 18 . Repositioning of the robot 15 or a patient 18 is performed at this time to a location that is within range of the desired trajectories.
- the surgeon can adjust the Z Frame 72 position, which can affect the x-axis 66 range and the y-axis 68 range of trajectories that the robot 15 is capable of reaching (for example, converging trajectories require less x-axis 66 or y-axis reach the lower the robot 15 is in the z-axis 70 ).
- a screen shows whether tracking markers on the patient 18 and robot 15 are in view of the detection device of the tracking system (for example, optical tracking system 3417 shown in FIG. 34 ). Repositioning of the cameras, if necessary, is also performed at this time for good visibility or optimal detection of tracking sensors.
- the detection device of the tracking system for example, optical tracking system 3417 shown in FIG. 34 .
- orientation of an end-effectuator 30 in a robot 15 coordinate system is calculated.
- position of the end-effectuator 30 in the robot 15 coordinate system is calculated.
- a line t defining the planned trajectory in the robot 15 coordinate system is determined.
- robot 15 position is locked on the planned trajectory at a current Z level.
- information indicative of quality of the trajectory lock can be supplied.
- actual coordinate(s) of the surgical robot 15 can be rendered in conjunction with respective coordinate(s) of the planned trajectory.
- aural indicia can be provided based on such quality.
- a high-frequency and/or high-amplitude noise can embody aural indicia suitable to represent a low-quality lock.
- a brief melody may be repeatedly played, such as the sound associated with successful recognition of a USB memory device by a computer, to indicate successful lock on the planned trajectory.
- a buzz or other warning noise may be played if the robot 15 is unable to reach its target due to the axis being mechanically overpowered, or if the tracking markers 720 are undetectable by cameras 8200 or other marker position sensors.
- the flow is directed to block 2710 .
- the method 2700 can be implemented (i.e., executed) as part of block 2440 in certain scenarios. It should be appreciated that in some embodiments, the method 2700 also can be implemented for any robot 15 having at least one feature that enable movement of the robot 15 .
- FIG. 28A is a flowchart of a method 2800 a for calculating position and/or orientation of an end-effectuator 30 in a robot 15 according to one at least one embodiment of the invention.
- the position and/or orientation can be calculated based at least on monitored position of a tracking array 690 mounted on the robot's x-axis 66 and monitored counts of encoders on the y-axis 68 , roll 62 , pitch 60 , and Z-tube axis 64 actuators.
- position and/or orientation are calculated in a robot 15 coordinate system.
- the method 2800 a can embody one or more of blocks 2740 or 2750 .
- the current position (i.e., 3D position) of x-axis 66 mounted robot 15 tracking markers is accessed.
- the current position of the tracking array 690 mounted to the robot 15 is transformed to neutral position. This is a position that was previously stored and represents the position of the tracking array 690 when the robot 15 was at zero counts on each axis between the tracker 690 and the robot 15 base (x-axis 66 and z-axis 70 in this configuration).
- the set of transformations (T1) to transform from the current position to the neutral position can be retained in computer memory (for example, the system memory 3412 and/or mass storage device 3404 ).
- a tip and tail of a line segment representing the vector in line with the end-effectuator 30 can be computed based at least on the process described herein. In some embodiments, this process can establish a robot 15 coordinate system and calibrate the relative orientations of the axes of movement of the robot 15 where tracking markers 720 can be attached temporarily to the end-effectuator 30 . In some embodiments, the vector's position in space can be determined by finding the finite helical axis of motion of markers 720 manually rotated to two positions around the guide tube 50 .
- the vector's position in space can be determined by connecting a point located at the entry of the guide tube 50 (identified by a temporarily mounted rigid body 690 with tracking markers 720 ) to a point located at the exit of the guide tube 50 (identified by a second temporarily mounted rigid body 690 with tracking markers 720 ).
- the tip of the line segment can be obtained as the point along the vector that is closest to the vector representing the helical axis of motion during pitch.
- the tail of the line segment can be set an arbitrary distance (for example about 100 mm) up the vector aligned with the guide tube 50 and/or first helical axis.
- the Cartesian coordinates of such tip and tail positions can be transformed to a coordinate system described herein in which the y-axis 68 movement can coincide with the y-axis 68 of the coordinate system, and the x-axis 66 can be aligned such that x-axis 66 movement can cause the greatest change in direction in the x-axis 66 , moderate change in the y-axis 68 , and no change in the z-axis 70 .
- these coordinates can be retained in a computer memory (for example system memory 3412 ) for later retrieval.
- tip and tail coordinates for neutral are accessed (i.e., retrieved).
- tip and tail are translated along Z-tube 50 neutral unit vector by monitored Z-tube 50 counts.
- an instantaneous axis of rotation (“IAR”) is accessed.
- the IAR is the same as the helical axis of motion ignoring the element of translation along the helical axis for pitch 60 for neutral.
- the vectors for this IAR were previously stored in computer memory at the time the coordinate system of the robot 15 was calibrated.
- tip coordinate, tail coordinate, and IAR vector direction and location coordinates are transformed (for example, iteratively transformed) to a new coordinate system in which IAR is aligned with X axis.
- data indicative of such transformations can be stored.
- tip coordinate and tail coordinate are rotated about X axis by pitch 60 angle.
- tip coordinate and tail coordinate are transformed back by inverse of T2 to the previous coordinate system.
- previously stored vectors that represent the IAR for roll 62 are accessed.
- tip coordinate, tail coordinate, IAR coordinate are transformed (for example, iteratively transformed) to a new coordinate system in which IAR is aligned with y-axis 68 .
- data indicative of such transformation(s) (T3) can be retained in memory.
- tip coordinate and tail coordinate are rotated about y-axis 68 by roll 62 angle.
- tip coordinate and tail coordinate are transformed back by inverse of T3 to the previous coordinate system.
- tip coordinate and tail coordinate are translated along a y-axis 68 unit vector (e.g., a vector aligned in this coordinate system with the y-axis 68 ) by monitored counts.
- tip coordinate and tail coordinate are transformed back by inverse of T1 to the current coordinate system monitored by the tracking system 3417 .
- FIG. 28B is a flowchart of a method 2800 b for calculating position and/or orientation of an end-effectuator 30 in a robot 15 in accordance with one embodiment of the invention.
- the position and/or the orientation can be calculated based at least on monitored position of a tracking array 690 mounted on the robot's 15 roll 62 axis and monitored counts of encoders on the pitch 60 and Z-tube 50 actuators.
- position and/or orientation can be calculated in a robot 15 coordinate system.
- the method 2800 B can embody one or more of blocks 2740 or 2750 .
- current position of an array of one or more robot 15 tracking markers 720 is accessed.
- the current position is a 3D position and the array of robot 15 tracking markers 720 can be mounted to the roll 62 axis of the robot 15 .
- the current position of the array of robot 15 tracking markers 720 mounted to the robot 15 is transformed to neutral position.
- the neutral position can be a position that was previously stored and can represent the position of the robot 15 tracking array 690 when the robot 15 had zero counts on each axis between the tracker 690 and the robot base 25 (e.g., z-axis 70 , x-axis 68 , y-axis 66 , and roll 62 axis in this configuration).
- data indicative of a set of transformations (T1) to go from the current position to the neutral position can be stored in a computer memory (for example, mass storage device 3404 or system memory 3412 ).
- a tip and tail of a line segment representing the vector in line with the end-effectuator 30 with temporarily attached tracking markers 720 is located.
- the vector's position in space can be determined by finding the finite helical axis of motion of markers manually rotated to two positions around the guide tube 50 .
- the vector's position in space can be determined by connecting a point located at the entry of the guide tube 50 (identified by a temporarily mounted rigid body 690 with tracking markers 720 ) to a point located at the exit of the guide tube 50 (identified by a second temporarily mounted rigid body 690 with tracking markers 720 ).
- the tip of the line segment can be found as the point along the vector that is closest to the vector representing the helical axis of motion during pitch.
- the tail of the line segment can be set an arbitrary distance (for example, nearly 100 mm) up the vector aligned with the guide tube/first helical axis.
- the Cartesian coordinates of these tip and tail positions can be transformed to a coordinate system described herein in which the y-axis 68 movement substantially coincides with the y-axis 68 of the coordinate system, and the x-axis 66 movement is aligned in a manner that, in some embodiments, x-axis 66 movement causes the greatest change in direction in the x-axis 66 , slight change in y-axis 68 , and no change in the z-axis 70 . It should be appreciated that such coordinates can be retained in memory (for example system memory 3412 ) for later retrieval.
- tip and tail coordinates for the neutral position are accessed (i.e., retrieved or otherwise obtained).
- tip and tail are translated along Z-tube 50 neutral unit vector by monitored Z-tube 50 counts.
- IAR is accessed.
- the vectors for this IAR may be available in a computer memory, for example, such vectors may be retained in the computer memory at the time the coordinate system of the robot 15 is calibrated in accordance with one or more embodiments described herein.
- tip coordinate, tail coordinate, and IAR vector direction and location coordinates are transformed to a new coordinate system in which IAR is aligned with x-axis 66 .
- data indicative of the applied transformations (T2) can be retained in a computer memory.
- tip coordinate and tail coordinate are rotated about x-axis 66 by pitch 60 angle.
- tip coordinate and tail coordinate are transformed back by applying the inverse of T2 to the previous coordinate system.
- tip coordinate and tail coordinate are transformed back by applying the inverse of T1 to the current coordinate system monitored by the tracking system 3417 .
- FIG. 29 is a flowchart of a method 2900 for determining a line indicative of a trajectory in a robot 15 coordinate system in accordance with one embodiment of the invention.
- the trajectory can be a planned trajectory associated with a surgical procedure.
- respective opaque marker 730 positions are accessed from a rigid body source (such as a fixture 690 ).
- an opaque marker 730 position is transformed from a representation in an image coordinate system to a representation in a current active marker 720 coordinate system.
- a planned trajectory is transformed from a representation in the image coordinate system to a representation in the current active maker 720 coordinate system.
- a set of current marker 720 positions on a robot 15 is transformed to a representation in a robot 15 coordinate system.
- the planned trajectory is transformed from a representation in the current active marker 720 coordinate system to the robot 15 coordinate system.
- FIG. 30 is a flowchart of a method 3000 for adjusting a robot 15 position to lock on a trajectory in accordance in accordance with one embodiment of the invention.
- the trajectory can be locked at a current Z plane, or level above the surgical field 17 .
- an instruction to move a roll 62 axis is transmitted at block 3010 and flow is directed to block 3015 .
- flow is directed to block 3015 where it is determined if the pitch 60 of the end-effectuator 30 has matched the pitch of the trajectory.
- an instruction to move a pitch 60 axis is transmitted at block 3020 and flow is directed to block 3025 .
- flow is directed to block 3025 where it is determined if x-axis 66 coordinates of points on the vector of the end-effectuator 30 intercept the x-axis 66 coordinates of the desired trajectory vector.
- an instruction to move the x-axis 66 can be transmitted and flow is directed to 3035 .
- flow is directed to block 3035 where it is determined if y-axis 68 coordinates of points on the vector of the end-effectuator 30 intercept the y-axis 68 coordinates of the desired trajectory vector.
- an instruction to move the y-axis 68 can be transmitted at block 3040 and flow is directed to block 3045 .
- flow is directed to block 3045 in which it is determined if a Z-tube 50 is being adjusted.
- an end-user can configure information (i.e., data or metadata) indicative of the Z-tube 50 being adjusted to control it to a desired position, for example.
- flow is terminated.
- flow is directed to block 3050 where it is determined if the Z-tube 50 is positioned at a predetermined distance from anatomy.
- flow terminates and the Z-tube 50 is located at a desired position with respect to a target location in the anatomy (bone, biopsy site, etc.).
- an instruction to move the Z-tube axis 64 is transmitted at block 3055 and the flow is directed to block 3050 .
- the subject method 3000 in some embodiments, but not all embodiments, may require that movement in each of the indicated axes (x-axis 66 , y-axis 68 , Z-tube axis 64 , roll 62 , and pitch 60 ) occurs without affecting the other axes earlier in the method flow.
- the y-axis 68 movement at block 3040 should not cause change in the position of x-axis 66 coordinate, which was already checked at block 3025 .
- the method 3000 can be implemented iteratively in order to reach a desired final position in instances where the axes do not move completely independently.
- the method 3000 can account for all axis positions nearly simultaneously, and can determine the exact amount of movement necessary in each axis, and thus it can be more efficient.
- FIG. 31 is a flowchart of a method 3100 for positioning an end-effectuator 30 in space in accordance with one embodiment of the invention.
- the positioning can comprise positioning the end-effectuator 30 in a first plane (for example, the x-y plane or horizontal plane) and moving the end-effectuator 30 along a direction substantially normal to the first plane.
- FIGS. 32-33 are flowcharts of methods for driving an end-effectuator 30 to a procedure location in accordance with one embodiment of the invention.
- the procedure location can be a position at the surface of a bone into which a conventional screw of other piece of hardware is to be inserted.
- the end-effectuator 30 can be fitted with a guide tube 50 or conventional dilator.
- the method 3200 can be implemented (i.e., executed).
- the surgeon may want to move the end-effectuator tip 30 , fitted with a guide tube 50 or a conventional dilator, all the way down to the bone. It should be appreciated that in some embodiments, since the first lateral movement occurs above the level where the patient 18 is lying, the methods depicted in FIGS.
- the method can also utilize the robot's Cartesian architecture, and the ease with which a coordinated movement down the infinite trajectory vector can be made. That is, in some embodiments, to move down this vector, the roll 62 and pitch 60 axes need no adjustment, while the x-axis 66 , y-axis 68 , and Z-frame 72 axes are moved at a fixed rate. In certain embodiments, for an articular robot 15 to make such a move, the multiple angular axes would have to be synchronized nonlinearly, with all axes simultaneously moved at varying rates.
- FIG. 34 illustrates a block diagram of a computer platform 3400 having a computing device 3401 that enables various features of the invention, and performance of the various methods disclosed herein in accordance with some embodiments of the invention.
- the computing device 3401 can control operation of a surgical robot 15 and an optical tracking system 3417 in accordance with aspects described herein.
- control can comprise calibration of relative systems of coordinates, generation of planned trajectories, monitoring of position of various units of the surgical robots 15 and/or units functionally coupled thereto, and implementation of safety protocols, and the like.
- computing device 3401 can embody a programmable controller that can control operation of a surgical robot 15 as described herein.
- the operating environment 3400 is only an example of an operating environment and is not intended to suggest any limitation as to the scope of use or functionality of operating environment architecture.
- the operating environment 3400 should not be interpreted as having any dependency or requirement relating to any one functional element or combination of functional elements (e.g., units, components, adapters, or the like).
- the various embodiments of the invention can be operational with numerous other general purpose or special purpose computing system environments or configurations.
- Examples of well-known computing systems, environments, and/or configurations that can be suitable for use with the systems and methods of the invention comprise personal computers, server computers, laptop devices or handheld devices, and multiprocessor systems. Additional examples comprise mobile devices, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that comprise any of the above systems or devices, and the like.
- the processing effected in the disclosed systems and methods can be performed by software components.
- the disclosed systems and methods can be described in the general context of computer-executable instructions, such as program modules, being executed by one or more computers, such as computing device 3401 , or other computing devices.
- program modules comprise computer code, routines, programs, objects, components, data structures, etc., that perform particular tasks or implement particular abstract data types.
- the disclosed methods also can be practiced in grid-based and distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network.
- program modules can be located in both local and remote computer storage media including memory storage devices.
- the components of the computing device 3401 can comprise, but are not limited to, one or more processors 3403 , or processing units 3403 , a system memory 3412 , and a system bus 3413 that couples various system components including the processor 3403 to the system memory 3412 .
- the system in the case of multiple processing units 3403 , the system can utilize parallel computing.
- a processor 3403 or a processing unit 3403 refers to any computing processing unit or processing device comprising, but not limited to, single-core processors; single-processors with software multithread execution capability; multi-core processors; multi-core processors with software multithread execution capability; multi-core processors with hardware multithread technology; parallel platforms; and parallel platforms with distributed shared memory.
- a processor 3403 or processing unit 3403 can refer to an integrated circuit, an application specific integrated circuit (ASIC), a digital signal processor (DSP), a field programmable gate array (FPGA), a programmable logic controller (PLC), a complex programmable logic device (CPLD), a discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein.
- ASIC application specific integrated circuit
- DSP digital signal processor
- FPGA field programmable gate array
- PLC programmable logic controller
- CPLD complex programmable logic device
- processors or processing units referred to herein can exploit nano-scale architectures such as, molecular and quantum-dot based transistors, switches and gates, in order to optimize space usage or enhance performance of the computing devices that can implement the various aspects of the subject invention.
- processor 3403 or processing unit 3403 also can be implemented as a combination of computing processing units.
- the system bus 3413 represents one or more of several possible types of bus structures, including a memory bus or memory controller, a peripheral bus, an accelerated graphics port, and a processor or local bus using any of a variety of bus architectures.
- bus architectures can comprise an Industry Standard Architecture (ISA) bus, a Micro Channel Architecture (MCA) bus, an Enhanced ISA (EISA) bus, a Video Electronics Standards Association (VESA) local bus, an Accelerated Graphics Port (AGP) bus, and a Peripheral Component Interconnects (PCI), a PCI-Express bus, a Personal Computer Memory Card Industry Association (PCMCIA), Universal Serial Bus (USB) and the like.
- ISA Industry Standard Architecture
- MCA Micro Channel Architecture
- EISA Enhanced ISA
- VESA Video Electronics Standards Association
- AGP Accelerated Graphics Port
- PCI Peripheral Component Interconnects
- PCI-Express PCI-Express
- PCMCIA Personal Computer Memory Card Industry Association
- USB Universal Serial Bus
- the bus 3413 and all buses specified in this specification and annexed drawings also can be implemented over a wired or wireless network connection and each of the subsystems, including the processor 3403 , a mass storage device 3404 , an operating system 3405 , robotic guidance software 3406 , robotic guidance data storage 3407 , a network adapter 3408 , system memory 3412 , an input/output interface 3410 , a display adapter 3409 , a display device 3411 , and a human machine interface 3402 , can be contained within one or more remote computing devices 3414 a,b at physically separate locations, functionally coupled (e.g., communicatively coupled) through buses of this form, in effect implementing a fully distributed system.
- functionally coupled e.g., communicatively coupled
- robotic guidance software 3406 can configure the computing device 3401 , or a processor thereof, to perform the automated control of position of the local robot 3416 (for example, surgical robot 15 ) in accordance with aspects of the invention. Such control can be enabled, at least in part, by a tracking system 3417 .
- robotic guidance software 3406 can configure such computer 100 to perform the functionality described in the subject invention.
- robotic guidance software 3406 can be retained in a memory as a group of computer-accessible instructions (for instance, computer-readable instructions, computer-executable instructions, or computer-readable computer-executable instructions).
- the group of computer-accessible instructions can encode the methods of the invention (such as the methods illustrated in FIGS. 24-33 in accordance with some embodiments of the invention).
- the group of computer-accessible instructions can encode various formalisms (e.g., image segmentation) for computer vision tracking.
- robotic guidance software 3406 can include a compiled instance of such computer-accessible instructions, a linked instance of such computer-accessible instructions, a compiled and linked instance of such computer-executable instructions, or an otherwise executable instance of the group of computer-accessible instructions.
- robotic guidance data storage 3407 can comprise various types of data that can permit implementation (e.g., compilation, linking, execution, and combinations thereof) of the robotic guidance software 3406 .
- robotic guidance data storage 3407 can comprise data associated with intraoperative imaging, automated adjustment of position of the local robot 3416 and/or remote robot 3422 , or the like.
- the data retained in the robotic guidance data storage 3407 can be formatted according to any image data in industry standard format.
- a remote tracking system 3424 can enable, at least in part, control of the remote robot 3422 .
- the information can comprise tracking information, trajectory information, surgical procedure information, safety protocols, and so forth.
- the computing device 3401 typically comprises a variety of computer readable media.
- the readable media can be any available media that is accessible by the computer 3401 and comprises, for example and not meant to be limiting, both volatile and non-volatile media, removable and non-removable media.
- the system memory 3412 comprises computer readable media in the form of volatile memory, such as random access memory (RAM), and/or non-volatile memory, such as read only memory (ROM).
- the system memory 3412 typically contains data (such as a group of tokens employed for code buffers) and/or program modules such as operating system 3405 and robotic guidance software 3406 that are immediately accessible to, and/or are presently operated-on by the processing unit 3403 .
- operating system 3405 can comprise operating systems such as Windows operating system, Unix, Linux, Symbian, Android, Apple iOS operating system, Chromium, and substantially any operating system for wireless computing devices or tethered computing devices.
- Apple® is a trademark of Apple Computer, Inc., registered in the United States and other countries.
- iOS® is a registered trademark of Cisco and used under license by Apple Inc.
- Microsoft® and Windows® are either registered trademarks or trademarks of Microsoft Corporation in the United States and/or other countries.
- Android® and Chrome® operating system are a registered trademarks of Google Inc.
- Symbian® is a registered trademark of Symbian Ltd.
- Linux® is a registered trademark of Linus Torvalds.
- UNIX® is a registered trademark of The Open Group.
- computing device 3401 can comprise other removable/non-removable, volatile/non-volatile computer storage media. As illustrated, in some embodiments, computing device 3401 comprises a mass storage device 3404 which can provide non-volatile storage of computer code (e.g., computer-executable instructions), computer-readable instructions, data structures, program modules, and other data for the computing device 3401 .
- computer code e.g., computer-executable instructions
- computer-readable instructions e.g., data structures, program modules, and other data for the computing device 3401 .
- a mass storage device 3404 can be a hard disk, a removable magnetic disk, a removable optical disk, magnetic cassettes or other magnetic storage devices, flash memory cards, CD-ROM, digital versatile disks (DVD) or other optical storage, random access memories (RAM), read only memories (ROM), electrically erasable programmable read-only memory (EEPROM), and the like.
- any number of program modules can be stored on the mass storage device 3404 , including by way of example, an operating system 3405 , and tracking software 3406 .
- each of the operating system 3405 and tracking software 3406 (or some combination thereof) can comprise elements of the programming and the tracking software 3406 .
- data and code (for example, computer-executable instructions, patient-specific trajectories, and patient 18 anatomical data) can be retained as part of tracking software 3406 and stored on the mass storage device 3404 .
- tracking software 3406 , and related data and code can be stored in any of one or more databases known in the art. Examples of such databases comprise, DB2®, Microsoft® Access, Microsoft® SQL Server, Oracle®, mySQL, PostgreSQL, and the like. Further examples include membase databases and flat file databases. The databases can be centralized or distributed across multiple systems.
- DB2® is a registered trademark of IBM in the United States.
- Microsoft®, Microsoft® Access®, and Microsoft® SQL ServerTM are either registered trademarks or trademarks of Microsoft Corporation in the United States and/or other countries.
- Oracle® is a registered trademark of Oracle Corporation and/or its affiliates.
- MySQL® is a registered trademark of MySQL AB in the United States, the European Union and other countries.
- PostgreSQL® and the PostgreSQL® logo are trademarks or registered trademarks of The PostgreSQL Global Development Group, in the U.S. and other countries.
- an agent for example, a surgeon or other user, or equipment
- an input device can comprise, but are not limited to, a camera (or other detection device for non-optical tracking markers), a keyboard, a pointing device (for example, a mouse), a microphone, a joystick, a scanner (for example, a barcode scanner), a reader device such as a radiofrequency identification (RFID) readers or magnetic stripe readers, gesture-based input devices such as tactile input devices (for example, touch screens, gloves and other body coverings or wearable devices), speech recognition devices, or natural interfaces, and the like.
- RFID radiofrequency identification
- these and other input devices can be connected to the processing unit 3403 via a human machine interface 3402 that is coupled to the system bus 3413 .
- they can be connected by other interface and bus structures, such as a parallel port, game port, an IEEE 1394 port (also known as a firewire port), a serial port, or a universal serial bus (USB).
- a display device 3411 can also be functionally coupled to the system bus 3413 via an interface, such as a display adapter 3409 .
- the computer 3401 can have more than one display adapter 3409 and the computer 3401 can have more than one display device 3411 .
- a display device 3411 can be a monitor, a liquid crystal display, or a projector.
- some embodiments can include other output peripheral devices that can comprise components such as speakers (not shown) and a printer (not shown) capable of being connected to the computer 3401 via input/output Interface 3410 .
- the input/output interface 3410 can be a pointing device, either tethered to, or wirelessly coupled to the computing device 3410 .
- any step and/or result of the methods can be output in any form to an output device.
- the output can be any form of visual representation, including, but not limited to, textual, graphical, animation, audio, tactile, and the like.
- one or more cameras can be contained or functionally coupled to the tracking system 3417 , which is functionally coupled to the system bus 3413 via an input/output interface of the one or more input/output interfaces 3410 . Such functional coupling can permit the one or more camera(s) to be coupled to other functional elements of the computing device 3401 .
- the input/output interface, at least a portion of the system bus 3413 , and the system memory 3412 can embody a frame grabber unit that can permit receiving imaging data acquired by at least one of the one or more cameras.
- the frame grabber can be an analog frame grabber, a digital frame grabber, or a combination thereof.
- the processor 3403 can provide analog-to-digital conversion functionality and decoder functionality to enable the frame grabber to operate with medical imaging data.
- the input/output interface can include circuitry to collect the analog signal received from at least one camera of the one or more cameras.
- tracking software 3406 in response to execution by processor 3403 , can operate the frame grabber to receive imaging data in accordance with various aspects described herein.
- Some embodiments include a computing device 3401 that can operate in a networked environment (for example, an industrial environment) using logical connections to one or more remote computing devices 3414 a,b , a remote robot 3422 , and a tracking system 3424 .
- a remote computing device can be a personal computer, portable computer, a mobile telephone, a server, a router, a network computer, a peer device or other common network node, and so on.
- an agent for example, a surgeon or other user, or equipment
- logical connections between the computer 3401 and a remote computing device 3414 a,b can be made via a local area network (LAN) and a general wide area network (WAN).
- the network connections can be implemented through a network adapter 3408 .
- the network adapter 3408 can be implemented in both wired and wireless environments.
- Some embodiments include networking environments that can be conventional and commonplace in offices, enterprise-wide computer networks, intranets.
- the networking environments generally can be embodied in wire-line networks or wireless networks (for example, cellular networks, such as third generation (“3G”) and fourth generation (“4G”) cellular networks, facility-based networks (for example, femtocell, picocell, wifi networks).
- a group of one or more networks 3415 can provide such networking environments.
- the one or more network(s) can comprise a LAN deployed in an industrial environment comprising the system 1 described herein.
- application programs and other executable program components such as the operating system 3405 are illustrated herein as discrete blocks, although it is recognized that such programs and components reside at various times in different storage components of the computing device 3401 , and are executed by the data processor(s) of the computer 100 .
- Some embodiments include an implementation of tracking software 3406 that can be stored on or transmitted across some form of computer readable media. Any of the disclosed methods can be performed by computer readable instructions embodied on computer readable media.
- Computer readable media can be any available media that can be accessed by a computer.
- Computer-readable media can comprise “computer storage media,” or “computer-readable storage media,” and “communications media.”
- “Computer storage media” comprise volatile and non-volatile, removable and non-removable media implemented in any methods or technology for storage of information such as computer readable instructions, data structures, program modules, or other data.
- computer storage media comprises, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by a computer.
- some embodiments include the computing device 3401 that can control operation of local robots 3416 and/or remote robots 3422 .
- the computing device 3401 can execute robotic guidance software 3407 to control such robots 3416 , 3422 , 15 .
- the robotic guidance software 3407 in response to execution, can utilize trajectories (such as, tip and tail coordinates) that can be planned and/or configured remotely or locally.
- the robotic guidance software 3407 in response to execution, can implement one or more of the methods described herein in a local robot's computer or a remote robot's computer to cause movement of the remote robot 15 or the local robot 15 according to one or more trajectories.
- the computing device 3401 can enable pre-operative planning of the surgical procedure. In some embodiments, the computing device 3401 can permit spatial positioning and orientation of a surgical tool (for example, instrument 35 ) during intraoperative procedures. In some further embodiments, the computing device 3401 can enable open procedures. In some other embodiments, the computing device 3401 can enable percutaneous procedures. In certain embodiments, the computing device 3401 and the robotic guidance software 3407 can embody a 3D tracking system 3417 to simultaneously monitor the positions of the device and the anatomy of the patient 18 . In some embodiments, the 3D tracking system 3417 can be configured to cast the patient's anatomy and the end-effectuator 30 in a common coordinate system.
- the computing device 3401 can access (i.e., load) image data from a conventional static storage device.
- the computing device 3401 can permit a 3D volumetric representation of patient 18 anatomy to be loaded into memory (for example, system memory 3412 ) and displayed (for example, via display device 3411 ).
- the computing device 3401 in response to execution of the robotic guidance software 3407 can enable navigation through the 3D volume representation of a patient's anatomy.
- the computing device 3401 can operate with a conventional power source required to offer the device for sale in the specified country.
- a conventional power cable that supplies power can be a sufficient length to access conventional hospital power outlets.
- the computing device 3401 in the event of a power loss, can hold the current end-effectuator 30 in a position unless an agent (for example, a surgeon or other user, or equipment) manually moves the end-effectuator 30 .
- the computing device 3401 can monitor system physical condition data. In some embodiments, the computing device 3401 can report to an operator (for example, a surgeon) each of the physical condition data and indicate an out-of-range value. In some embodiments, the computing device 3401 can enable entry and storage of manufacturing calibration values for end-effectuator 30 positioning using, for example, the input/output interface 3410 . In some embodiments, the computing device 3401 can enable access to manufacturing calibration values by an agent (for example, a surgeon or other user, or equipment) authenticated to an appropriate access level. In some embodiments, the data can be retained in robotic guidance data storage 3407 , or can be accessed via network(s) 3415 when the data is retained in a remote computing device 3414 a.
- the computing device 3401 can render (using for example display device 3411 ) a technical screen with a subset of the end-effectuator 30 positioning calibration and system health data.
- the information is only accessible to an agent (for example, a surgeon or other user, or equipment) authenticated to an appropriate level.
- the computing device 3401 can enable field calibration of end-effectuator 30 positioning only by an agent (for example, a surgeon or other user, or equipment) authenticated to an appropriate access level.
- the computing device 3401 can convey the status of local robot 3416 , remote robot 3422 , and/or other device being locked in position using a visual or aural alert.
- the computing device 3401 can include an emergency stop control that upon activation, disables power to the device's motors 160 but not to the processor 3403 .
- the emergency stop control can be accessible by the operator of computing device 3401 .
- the computing device 3401 can monitor the emergency stop status and indicate to the operator that the emergency stop has been activated.
- the computing device 3401 can be operated in a mode that permits manual positioning of the end-effectuator 30 .
- the computing device 3401 can boot directly to an application representing the robotic guidance software 3406 .
- computing device 3401 can perform a system check prior to each use. In scenarios in which the system check fails, the computing device 3401 can notify an operator.
- the computing device 3401 can generate an indicator for reporting system status. Some embodiments include the computing device 3401 that can minimize or can mitigate delays in processing, and in the event of a delay in processing, notify an agent (for example, a surgeon or other user, or equipment). For example, in some embodiments, a delay may occur while a system scan is being performed to assess system status, and consequently the computing device 3401 can schedule (for example, generate a process queue) system scans to occur at low usage times. In some embodiments, a system clock of the computing device 3401 can be read before and after key processes to assess the length of time required to complete computation of a process. In some embodiments, the actual time to complete the process can be compared to the expected time.
- a discrepancy if a discrepancy is found to be beyond an acceptable tolerance, the agent can be notified, and/or concurrently running non-essential computational tasks can be terminated.
- a conventional system clock (not shown) can be part of processor 3403 .
- the computing device 3401 can generate a display that follows a standardized workflow. In some embodiments, the computing device 3401 can render or ensure that text is rendered in a font of sufficient size and contrast to be readable from an appropriate distance. In some embodiments, the computing device 3401 can enable an operator to locate the intended position of a surgical implant or tool.
- the computing device 3401 can determine the relative position of the end-effectuator 30 to the anatomy of the patient 18 . For example, to at least such end, the computing device 3401 can collect data the optical tracking system 3417 , and can analyze the data to generate data indicative of such relative position. In some embodiments, the computing device 3401 can indicate the end-effectuator 30 position and orientation. In some embodiments, the computing device 3401 can enable continuous control of end-effectuator 30 position relative to the anatomy of a patient 18 .
- the computing device 3401 can enable an agent (for example, a surgeon or other user, or equipment) to mark the intended position of a surgical implant or tool (for example, instrument 35 ).
- the computing device 3401 can allow the position and orientation of a conventional hand-held probe (or an instrument 35 ) to be displayed overlaid on images of the patient's anatomy.
- the computing device 3401 can enable an agent (for example, a surgeon or other user, or equipment) to position conventional surgical screws. In some embodiments, the computing device 3401 can enable selection of the length and diameter of surgical screws by the agent. In yet another aspect, the computing device can ensure that the relative position, size and scale of screws are maintained on the display 3411 when in graphical representation. In some embodiments, the computing device 3401 can verify screw path plans against an operation envelope and reject screw path plans outside this envelope. In still another aspect, the computing device 3401 can enable hiding of a graphical screw representation.
- an agent for example, a surgeon or other user, or equipment
- the computing device 3401 can enable a function that allows the current view to be stored. In some embodiments, the computing device 3401 can enable a view reset function that sets the current view back to a previously stored view. In some embodiments, the computing device 3401 can enable an authentication based tiered access system. In some embodiments, the computing device 3401 can log and store system activity. In some embodiments, the computing device 3401 can enable access to the system activity log to an agent authorized to an appropriate level. In some embodiments, the computing device 3401 can enable entry and storage of patient 18 data.
- the computing device 3401 can enable the appropriate disposition of patient 18 data and/or procedure data. For example, in a scenario in which such data are being collected for research, the computing device 3401 can implement de-identification of the data in order to meet patient 18 privacy requirements. In some embodiments, the de-identification can be implemented in response to execution of computer-executable instruction(s) retained in memory 3412 or any other memory accessible to the computing device 3401 . In some embodiments, the de-identification can be performed automatically before the patient 18 data and/or procedure data are sent to a repository or any other data storage (including mass storage device 3404 , for example).
- indicia e.g., a dialog box
- an agent e.g., machine or human
- FIG. 11 shows a flow chart diagram 1100 for general operation of the robot 15 according to some embodiments is shown.
- the local positioning system (herein referred to as “LPS”) establishes a spatial coordinate measuring system for the room 10 where the invasive procedure is to occur; in other words, the LPS is calibrated.
- a conventional mechanical fixture that includes a plurality of attached calibrating transmitters 120 is placed within the room 10 where positioning sensors 12 are located.
- at least three calibrating transmitters 120 are required, but any number of calibrating transmitters 120 above three is within the scope of the invention.
- at least three positioning sensors 12 are required, but any number of positioning sensors 12 above three is also within the scope of the invention, and the accuracy of the system is increased with the addition of more positioning sensors.
- the distance between each of the calibrating transmitters 120 relative to each other is measured prior to calibration step 210 .
- Each calibrating transmitter 120 transmits RF signals on a different frequency so that the positioning sensors 12 can determine which transmitter 120 emitted a particular RF signal.
- the signal of each of these transmitters 120 is received by positioning sensors 12 .
- the positioning sensors 12 are able to calculate the spatial distance of each of the positioning sensors 12 relative to each other.
- the system 1 is now calibrated. As a result, in some embodiments, the positioning sensors 12 can now determine the spatial position of any new RF transmitter 120 introduced into the room 10 relative to the positioning sensors 12 .
- the positions of the RF transmitters 120 tracking the anatomical target are read by positioning sensors 110 . These transmitters 120 identify the initial position of the anatomical target and any changes in position during the procedure. In some embodiments, if any RF transmitters 120 must transmit through a medium that changes the RF signal characteristics, then the system will compensate for these changes when determining the transmitter's 120 position.
- the positions of the transmitters 120 on the anatomy are calibrated relative to the LPS coordinate system.
- the LPS provides a reference system, and the location of the anatomical target is calculated relative to the LPS coordinates.
- the positions of transmitters 120 affixed to the anatomical target are recorded at the same time as positions of temporary transmitters 120 placed on precisely known anatomical landmarks also identified on the anatomical image. This calculation is performed by a computer 100 .
- the positions of the RF transmitters 120 that track the anatomical target are read. Since the locations of the transmitters 120 on the anatomical target have already been calibrated, the system can easily determine if there has been any change in position of the anatomical target.
- Some embodiments include a step 260 , where the positions of the transmitters 120 on the surgical instrument 35 are read.
- the transmitters 120 may be located on the surgical instrument 35 itself, and/or there may be transmitters 120 attached to various points of the surgical robot 15 .
- the surgical robot 15 can also include a plurality of attached conventional position encoders that help determine the position of the surgical instrument 35 .
- the position encoders can be devices used to generate an electronic signal that indicates a position or movement relative to a reference position.
- a position signal can be generated using conventional magnetic sensors, conventional capacitive sensors, and conventional optical sensors.
- position data read from the position encoders may be used to determine the position of the surgical instrument 35 used in the procedure.
- the data may be redundant of position data calculated from RF transmitters 120 located on the surgical instrument 35 . Therefore, in some embodiments, position data from the position encoders may be used to double-check the position being read from the LPS.
- the coordinates of the positions of the transmitters 120 on the surgical instrument 35 , and/or the positions read from the position encoders is calibrated relative to the anatomical coordinate system.
- the position data of the surgical instrument 35 is synchronized into the same coordinate system as the patient's anatomy. In some embodiments, this calculation is performed automatically by the computer 100 since the positions of the transmitters 120 on the anatomical target and the positions of the transmitters 120 on the surgical instrument 35 are in the same coordinate system, and the positions of the transmitters 120 on the anatomical target are already calibrated relative to the anatomy.
- the computer 100 superimposes a representation of the location calculated in step 270 of the surgical device on the 3D anatomical image of the patient 18 taken in step 220 .
- the superimposed image can be displayed to an agent.
- the computer 100 sends the appropriate signals to the motors 160 to drive the surgical robot 15 .
- the robot 15 is driven so that the surgical instrument 35 follows the preprogrammed trajectory if there is no further input from the agent.
- the computer 100 drives the robot 15 in response to the agent input.
- the computer 100 determines whether the anatomy needs to be recalibrated.
- the agent may choose to recalibrate the anatomy, in which case the computer 100 responds to agent input.
- the computer 100 may be programmed to recalibrate the anatomy in response to certain events. For instance, in some embodiments, the computer 100 may be programmed to recalibrate the anatomy if the RF transmitters 120 on the anatomical target indicate that the location of the anatomical target has shifted relative to the RF transmitters 120 (i.e. this spatial relationship should be fixed).
- an indicator that the anatomical target location has shifted relative to the transmitters 120 is if the computer 100 calculates that the surgical instrument 35 appears to be inside bone when no drilling or penetration is actually occurring.
- the process beginning at step 230 is repeated. In some embodiments, if the anatomy does not need to be recalibrated, then the process beginning at step 250 is repeated. In some embodiments, at any time during the procedure, certain fault conditions may cause the computer 100 to interrupt the program and respond accordingly. For instance, in some embodiments, if the signal from the RF transmitters 120 cannot be read, then the computer 100 may be programmed to stop the movement of the robot 15 , or remove the surgical instrument 35 from the patient 18 . Another example of a fault condition is if the robot 15 encounters a resistance above a preprogrammed tolerance level.
- FIG. 12 shows a flow chart diagram 1200 for a closed screw/needle insertion procedure according to an embodiment of the invention is shown.
- the robot 15 holds a guide tube 50 adjacent to the patient 18 in the correct angular orientation at the point where a conventional pedicle screw is to be inserted through the tissue and into the bone of the patient 18 .
- the distance between each of the calibrating transmitters 120 relative to each other is measured prior to calibration step 300 .
- each calibrating transmitter 120 transmits RF signals on a different frequency so the positioning sensors 12 can determine which transmitter 120 emitted a particular RF signal.
- the signal of each of these transmitters 120 is received by positioning sensors 12 . Since the distance between each of the calibrating transmitters 120 is known, and the sensors 12 can identify the signals from each of the calibrating transmitters 120 based on the known frequency, using time of flight calculation, in some embodiments, the positioning sensors 12 are able to calculate the spatial distance of each of the positioning sensors 12 relative to each other.
- the system 1 is now calibrated. As a result, in some embodiments, the positioning sensors 12 can now determine the spatial position of any new RF transmitter 120 introduced into the room 10 relative to the positioning sensors 12 .
- a 3D anatomical image scan such as a CT scan, is taken of the anatomical target. Any 3D anatomical image scan may be used with the surgical robot 15 and is within the scope of the present invention.
- the operator selects a desired trajectory and insertion point of the surgical instrument 35 on the anatomical image captured at step 310 .
- the desired trajectory and insertion point is programmed into the computer 100 so that the robot 15 can drive a guide tube 50 automatically to follow the trajectory.
- the positions of the RF transmitters 120 tracking the anatomical target are read by positioning sensors 110 . In some embodiments, these transmitters 120 identify the initial position of the anatomical target and any changes in position during the procedure.
- the system will compensate for these changes when determining the transmitter's 120 position.
- the positions of the transmitters 120 on the anatomy are calibrated relative to the LPS coordinate system.
- the LPS provides a reference system, and the location of the anatomical target is calculated relative to the LPS coordinates.
- the positions of transmitters 120 affixed to the anatomical target are recorded at the same time as positions of temporary transmitters 120 on precisely known anatomical landmarks also identified on the anatomical image. This calculation is performed by a computer.
- the positions of the RF transmitters 120 that track the anatomical target are read. Since the locations of the transmitters 120 on the anatomical target have already been calibrated, in some embodiments, the system can easily determine if there has been any change in position of the anatomical target.
- the positions of the transmitters 120 on the surgical instrument 35 are read. In some embodiments, the transmitters 120 may be located on the surgical instrument 35 , and/or attached to various points of the surgical robot 15 .
- the coordinates of the positions of the transmitters 120 on the surgical instrument 35 , and/or the positions read from the position encoders, are calibrated relative to the anatomical coordinate system.
- the position data of the surgical instrument 35 is synchronized into the same coordinate system as the anatomy. This calculation is performed automatically by the computer 100 since the positions of the transmitters 120 on the anatomical target and the positions of the transmitters 120 on the surgical instrument 35 are in the same coordinate system and the positions of the transmitters 120 on the anatomical target are already calibrated relative to the anatomy.
- the computer 100 superimposes a representation of the location calculated in step 370 of the surgical device on the 3D anatomical image of the patient 18 taken in step 310 .
- the superimposed image can be displayed to the user.
- the computer 100 determines whether the guide tube 50 is in the correct orientation and position to follow the trajectory planned at step 320 . If it is not, then step 393 is reached. If it is in the correct orientation and position to follow the trajectory, then step 395 is reached.
- the computer 100 determines what adjustments it needs to make in order to make the guide tube 50 follow the preplanned trajectory.
- the computer 100 sends the appropriate signals to drive the motors 160 in order to correct the movement of the guide tube.
- the computer 100 determines whether the procedure has been completed. If the procedure has not been completed, then the process beginning at step 350 is repeated.
- certain fault conditions may cause the computer 100 to interrupt the program and respond accordingly. For instance, if the signal from the RF transmitters 120 cannot be read, then the computer 100 may be programmed to stop the movement of the robot 15 or lift the guide tube 50 away from the patient 18 . Another example of a fault condition is if the robot 15 encounters a resistance above a preprogrammed tolerance level. Another example of a fault condition is if the RF transmitters 120 on the anatomical target shift so that actual and calculated positions of the anatomy no longer match. One indicator that the anatomical target location has shifted relative to the transmitters 120 is if the computer 100 calculates that the surgical instrument 35 appears to be inside bone when no drilling or penetration is actually occurring.
- the proper response to each condition may be programmed into the system, or a specific response may be user-initiated.
- the computer 100 may determine that in response to an anatomy shift, the anatomy would have to be recalibrated, and the process beginning at step 330 should be repeated.
- a fault condition may require the flowchart to repeat from step 300 .
- the user may decide that recalibration from step 330 is desired, and initiate that step himself.
- a flow chart diagram 1300 for a safe zone surgical procedure performed using the system described herein is shown in accordance with some embodiments of the invention.
- a safe zone surgical procedure there is a defined safe zone around the surgical area within which the surgical device must stay.
- the physician manually controls the surgical device that is attached to the end-effectuator 30 of the surgical robot 15 . If the physician moves the surgical device outside of the safe zone, then the surgical robot 15 stiffens the arm 23 so that the physician cannot move the instrument 35 in any direction that would move the surgical instrument 35 outside the safe zone.
- the distance between each of the calibrating transmitters 120 relative to each other is measured prior to calibration step 400 .
- Each calibrating transmitter 120 transmits RF signals on a different frequency so the positioning sensors 12 can determine which transmitter 120 emitted a particular RF signal.
- the signal of each of these transmitters 120 is received by positioning sensors 12 . Since the distance between each of the calibrating transmitters 120 is known, and the sensors 12 can identify the signals from each of the calibrating transmitters 120 based on the known frequency, the positioning sensors 12 are able to calculate, using time of flight calculation, the spatial distance of each of the positioning sensors 12 relative to each other.
- the system 1 is now calibrated. As a result, the positioning sensors 12 can now determine the spatial position of any new RF transmitter 120 introduced into the room 10 relative to the positioning sensors 12 .
- a 3D anatomical image scan such as a CT scan, is taken of the anatomical target. Any 3D anatomical image scan may be used with the surgical robot 15 and is within the scope of the present invention.
- the operator inputs a desired safe zone on the anatomical image taken in step 410 .
- the operator uses an input to the computer 100 to draw a safe zone on a CT scan taken of the patient 18 in step 410 .
- the positions of the RF transmitters 120 tracking the anatomical target are read by positioning sensors. These transmitters 120 identify the initial position of the anatomical target and any changes in position during the procedure. In some embodiments, if any RF transmitters 120 must transmit through a medium that changes the RF signal characteristics, then the system will compensate for these changes when determining the transmitter's 120 position.
- the positions of the transmitters 120 on the anatomy are calibrated relative to the LPS coordinate system.
- the LPS provides a reference system, and the location of the anatomical target is calculated relative to the LPS coordinates.
- the positions of transmitters 120 affixed to the anatomical target are recorded at the same time as positions of temporary transmitters 120 on precisely known landmarks on the anatomy that can also be identified on the anatomical image. This calculation is performed by a computer 100 .
- the positions of the RF transmitters 120 that track the anatomical target are read. Since the locations of the transmitters 120 on the anatomical target have already been calibrated, the system can easily determine if there has been any change in position of the anatomical target.
- the positions of the transmitters 120 on the surgical instrument 35 are read.
- the transmitters 120 may be located on the surgical instrument 35 itself, and/or there may be transmitters 120 attached to various points of the surgical robot 15 .
- the coordinates of the positions of the transmitters 120 on the surgical instrument 35 , and/or the positions read from the position encoders are calibrated relative to the anatomical coordinate system. In other words, the position data of the surgical instrument 35 is synchronized into the same coordinate system as the anatomy.
- This calculation is performed automatically by the computer 100 since the positions of the transmitters 120 on the anatomical target and the positions of the transmitters 120 on the surgical instrument 35 are in the same coordinate system and the positions of the transmitters 120 on the anatomical target are already calibrated relative to the anatomy.
- the computer 100 superimposes a representation of the location calculated in step 470 of the surgical device on the 3D anatomical image of the patient 18 taken in step 410 . In some embodiments, the superimposed image can be displayed to the user.
- the computer 100 determines whether the surgical device attached to the end-effectuator 30 of the surgical robot 15 is within a specified range of the safe zone boundary (for example, within 1 millimeter of reaching the safe zone boundary). In some embodiments, if the end-effectuator 30 is almost to the boundary, then step 493 is reached. In some embodiments, if it is well within the safe zone boundary, then step 495 is reached.
- the computer 100 stiffens the arm of the surgical robot 15 in any direction that would allow the user to move the surgical device closer to the safe zone boundary.
- the computer 100 determines whether the anatomy needs to be recalibrated. In some embodiments, the user may choose to recalibrate the anatomy, in which case the computer 100 responds to user input. Alternatively, in some embodiments, the computer 100 may be programmed to recalibrate the anatomy in response to certain events. For instance, in some embodiments, the computer 100 may be programmed to recalibrate the anatomy if the RF transmitters 120 on the anatomical target indicate that the location of the anatomical target has shifted relative to the RF transmitters 120 (i.e. this spatial relationship should be fixed.) In some embodiments, an indicator that the anatomical target location has shifted relative to the transmitters 120 is if the computer 100 calculates that the surgical instrument 35 appears to be inside bone when no drilling or penetration is actually occurring.
- the process beginning at step 430 is repeated. In some embodiments, if the anatomy does not need to be recalibrated, then the process beginning at step 450 is repeated. In some embodiments, at any time during the procedure, certain fault conditions may cause the computer 100 to interrupt the program and respond accordingly. For instance, in some embodiments, if the signal from the RF transmitters 120 cannot be read, then the computer 100 may be programmed to stop the movement of the robot 15 or remove the surgical instrument 35 from the patient 18 . Another example of a fault condition is if the robot 15 encounters a resistance above a preprogrammed tolerance level.
- FIG. 14 a flow chart diagram 1400 for a conventional flexible catheter or wire insertion procedure according to an embodiment of the invention is shown.
- Catheters are used in a variety of medical procedures to deliver medicaments to a specific site in a patient's body. Often, delivery to a specific location is needed so a targeted diseased area can then be treated. Sometimes instead of inserting the catheter directly, a flexible wire is first inserted, over which the flexible catheter can be slid.
- the distance between each of the calibrating transmitters 120 relative to each other is measured prior to calibration step 500 .
- each calibrating transmitter 120 transmits RF signals on a different frequency so the positioning sensors 12 , 110 can determine which transmitter 120 emitted a particular RF signal.
- the signal from each of these transmitters 120 is received by positioning sensors 12 , 110 . Since the distance between each of the calibrating transmitters 120 is known, and the sensors can identify the signals from each of the calibrating transmitters 120 based on the known frequency, in some embodiments, using time of flight calculation, the positioning sensors 12 , 110 are able to calculate the spatial distance of each of the positioning sensors 12 , 110 relative to each other. The system is now calibrated. As a result, in some embodiments, the positioning sensors 12 , 110 can now determine the spatial position of any new RF transmitter 120 introduced into the room 10 relative to the positioning sensors 12 , 110 .
- reference needles that contain the RF transmitters 120 are inserted into the body.
- the purpose of these needles is to track movement of key regions of soft tissue that will deform during the procedure or with movement of the patient 18 .
- a 3D anatomical image scan (such as a CT scan) is taken of the anatomical target. Any 3D anatomical image scan may be used with the surgical robot 15 and is within the scope of the present invention.
- the anatomical image capture area includes the tips of the reference needles so that their transmitters' 120 positions can be determined relative to the anatomy.
- the RF signals from the catheter tip and reference needles are read.
- the position of the catheter tip is calculated. Because the position of the catheter tip relative to the reference needles and the positions of the reference needles relative to the anatomy are known, the computer 100 can calculate the position of the catheter tip relative to the anatomy. In some embodiments, at step 550 , the superimposed catheter tip and the shaft representation is displayed on the anatomical image taken in step 520 . In some embodiments, at step 560 , the computer 100 determines whether the catheter tip is advancing toward the anatomical target. If it is not moving to the anatomical target, then step 563 is reached. If it is correctly moving, then step 570 is reached.
- the robot 15 arm is adjusted to guide the catheter tip in the desired direction. If the anatomy needs to be calibrated, then in some embodiments, the process beginning at step 520 is repeated. If the anatomy does not need to be recalibrated, then the process beginning at step 540 is repeated. In some embodiments, at step 570 , the computer 100 determines whether the procedure has been completed. If the procedure has not been completed, then the process beginning at step 540 is repeated.
- certain fault conditions may cause the computer 100 to interrupt the program and respond accordingly. For instance, in some embodiments, if the signal from the RF transmitter's 120 cannot be read, then the computer 100 may be programmed to stop the movement of the robot 15 or remove the flexible catheter from the patient 18 . Another example of a fault condition is if the robot 15 encounters a resistance above a preprogrammed tolerance level. A further example of a fault condition is if the RF transmitter's 120 on the anatomical target indicate the location of the anatomical target shift so that actual and calculated positions of the anatomy no longer match. In some embodiments, one indicator that the anatomical target location has shifted relative to the transmitter's 120 is if the computer 100 calculates that the surgical instrument 35 appears to be inside bone when no drilling or penetration is actually occurring.
- the proper response to each condition may be programmed into the system, or a specific response may be user-initiated.
- the computer 100 may determine that in response to an anatomy shift, the anatomy would have to be recalibrated, and the process beginning at step 520 should be repeated.
- a fault condition may require the flowchart to repeat from step 500 .
- the user may decide that recalibration from step 520 is desired, and initiate that step himself.
- the software provides the method to select the target area of surgery, plan the surgical path, check the planned trajectory of the surgical path, synchronize the medical images to the positioning system and precisely control the positioning system during surgery.
- the surgical positioning system and navigation software includes an optical guidance system or RF Local Positioning System (RF-LPS), which are in communication with the positioning system.
- RF-LPS RF Local Positioning System
- FIG. 15A shows a screen shot 600 of the selection step for a user using a software program as described herein in accordance with some embodiments of the invention.
- Screen shot 600 includes windows 615 , 625 , and 635 , which show a 3D anatomical image of surgical target 630 on different planes.
- the user selects the appropriate 3D image corresponding to anatomical location of where the procedure will occur.
- the user uses a graphic control to change the perspective of the image in order to more easily view the image from different angles.
- the user can view the surgical target 630 within separate coordinated views for each of the x-axis, y-axis and z-axis coordinates for each anatomical location in the database in each window 615 , 625 and 635 , respectively.
- an input control is used with the software in order to plan the trajectory of the surgical instrument 35 .
- the input control is in the shape of a biopsy needle 8110 for which the user can plan a trajectory.
- FIG. 15B shows a screen shot 650 during the medical procedure in accordance with some embodiments of the invention.
- the user can still view the anatomical target 630 in different x-axis, y-axis and z-axis coordinate views on windows 615 , 625 , and 635 .
- the user can see the planned trajectory line 670 in multiple windows 615 and 625 .
- the actual trajectory and location of the surgical instrument 35 is superimposed on the image (shown as line segment 660 ).
- the actual trajectory and location of the surgical instrument 35 is dynamically updated and displayed, and is shown as a line segment 660 .
- the actual trajectory and location of the surgical instrument 35 could be shown as a trapezoid or a solid central line surrounded by a blurred or semi-transparent fringe to represent the region of uncertainty.
- the tracking system 3417 and robot 15 encoders calculate that the surgical instrument 35 should be located at the solid line or center of the trapezoid.
- the size of this edge could be estimated knowing the stiffness and tolerance of the surgical instrument 35 within the guide tube 50 , and by using experimental data collected for the same instrument 35 under previous controlled conditions. In some embodiments, displaying this region of uncertainty helps prevent the user from expecting the system to deliver a tool to a target trajectory with a physically impossible level of precision.
- the surgical robot 15 can be used with alternate guidance systems other than an LPS.
- the surgical robot system 1 can comprise a targeting fixture 690 for use with a guidance system.
- one targeting fixture 690 comprises a calibration frame 700 , as shown in FIGS. 20A-20E .
- a calibration frame 700 can be used in connection with many invasive procedures; for example, it can be used in thoracolumbar pedicle screw insertion in order to help achieve a more accurate trajectory position.
- the use of the calibration frame 700 can simplify the calibration procedure.
- the calibration frame 700 can be temporarily affixed to the skin of a patient 18 surrounding a selected site for a medical procedure, and then the medical procedure can be performed through a window defined by the calibration frame.
- the calibration frame 700 can comprise a combination of radio-opaque markers 730 and infrared, or “active,” markers 720 .
- the radio-opaque markers 730 can be located within the CT scan region 710
- the active markers 720 can be located outside of the CT scan region 710 .
- a surgical field 17 i.e., the area where the invasive procedure will occur
- the actual distances of the radio-opaque 730 and active markers 720 relative to each other can be measured from a high-precision laser scan of the calibration frame.
- the actual relative distances can be measured by actively measuring the positions of active markers 720 while nearly simultaneously or simultaneously pointing with a pointing device, such as a conventional digitizing probe, to one or more locations on the surface of the radio-opaque markers 730 .
- digitizing probes can comprise active markers 720 embedded in a rigid body 690 and a tip extending from the rigid body.
- the location of the probe tip relative to the rigid body of the probe can be established. In some embodiments, it can then be possible to calculate the location of the probe's tip from the probe's active markers 720 . In some embodiments, for a probe with a concave tip that is calibrated as previously described, the point in space returned during operation of the probe can represent a point distal to the tip of the probe at the center of the tip's concavity.
- the probe when a probe (configured with a concave tip and calibrated to marker 730 of the same or nearly the same diameter as the targeting fixture's radio-opaque marker 730 ) is touched to the radio-opaque marker 730 , the probe can register the center of the sphere.
- active markers 720 can also be placed on the robot in order to monitor a position of the robot 15 and calibration frame 700 simultaneously or nearly simultaneously.
- the calibration frame 700 is mounted on the patient's skin before surgery/biopsy, and will stay mounted during the entire procedure. Surgery/biopsy takes place through the center of the frame 700 .
- the CT scan contains both the medical images of the patient's bony anatomy, and spherical representations of the radio-opaque markers 730 .
- software is used to determine the locations of the centers of the markers 730 relative to the trajectories defined by the surgeon on the medical images.
- the pixel spacing of the CT scan can be conveyed within encoded headers in DICOM images, or can be otherwise available to a tracking software (for example, the robotic guidance software 3406 ), it can, in some embodiments, be possible to register locations of the centers of the markers 730 in Cartesian coordinates (in millimeters, for example, or other length units). In some embodiments, it can be possible to register the Cartesian coordinates of the tip and tail of each trajectory in the same length units.
- the system knows the positions of the trajectories relative to the radio-opaque markers 730 , the positions of the radio-opaque markers 730 relative to the active markers 720 , and the positions of the active markers 720 on the calibration frame 700 relative to the active markers on the robot 15 (not shown), the system has all information necessary to position the robot's end-effectuator 30 relative to the defined trajectories.
- the calibration frame 700 can comprise at least three radio-opaque markers 730 embedded in the periphery of the calibration frame 700 .
- the at least three radio-opaque markers 730 can be positioned asymmetrically about the periphery of the calibration frame 700 such that the software, as described herein, can sort the at least three radio-opaque markers 730 based only on the geometric coordinates of each marker 730 .
- the calibration frame 700 can comprise at least one bank of active markers 720 .
- each bank of the at least one bank can comprise at least three active markers 720 .
- the at least one bank of active markers 720 can comprise four banks of active markers 720 .
- the calibration frame 700 can comprise a plurality of leveling posts 77 coupled to respective corner regions of the calibration frame 700 .
- the corner regions of the calibration frame 700 can include leveling posts 77 that can comprise radiolucent materials.
- the plurality of leveling posts 77 can be configured to promote uniform, rigid contact between the calibration frame 700 and the skin of the patient 18 .
- a surgical-grade adhesive film such as, for example and without limitation, IobanTM from 3MTM, can be used to temporarily adhere the calibration frame 700 to the skin of the patient 18 .
- 3MTM and IobanTM are registered trademarks of 3M Company.
- the calibration frame 700 can comprise a plurality of upright posts 75 that are angled away from the frame 700 (see FIG. 20B ).
- the plurality of active markers 720 can be mounted on the plurality of upright posts 75 .
- radio-opaque markers 730 there are four radio-opaque markers 730 (non-metallic BBs from an air gun) embedded in the periphery of the frame, labeled OP1, OP2, OP3, OP4.
- the radio-opaque markers 730 are placed in an asymmetrical configuration (notice how OP1 and OP2 are separated from each other by more distance than OP3 and OP4, and OP1 and OP4 are aligned with each other across the gap, however OP3 is positioned more toward the center than OP2). The reason for this arrangement is so that a computer algorithm can automatically sort the markers to determine which is which if only given the raw coordinates of the four markers and not their identification.
- the angulation of the upright posts can permit the active markers 720 to face toward the cameras or detection devices of the tracking system (for example, the tracking system 3417 ).
- the upright posts can be angled away from the calibration frame by about 10°.
- a conventional digitizing probe such as a 6-marker probe, embedded with active markers 720 in a known relationship to the probe's tip (see for example FIG. 20C ) can be used to point to each of the radio-opaque markers 730 .
- the probe can point to locations on two opposite surfaces of the spherical radio-opaque markers 730 while recording the position of the probe tip and the active markers 720 on the frame 700 simultaneously. Then, the average position of the two surface coordinates can be taken, corresponding to the center of the sphere.
- An image of the robot 15 used with this targeting fixture 690 is shown in FIG. 20D .
- the robot 15 can work through the window formed by the frame 700 .
- the working portal is kept on the interior of the frame 700 and the markers 720 on the exterior of the frame 700 can improve accuracy over a system where fiducials are mounted away from the area where surgery is being performed.
- simulation, and/or modeling it is believed that a reason for improved accuracy is that optimal accuracy of tracking markers 720 can be achieved if tracking markers 720 are placed around the perimeter of the frame 700 being tracked.
- FIG. 20E illustrating a calibration frame 700 .
- This fixture 690 is simplified to make it less obstructive to the surgeon.
- the calibration frame 700 can comprise four active markers 720 having a lower profile than the active markers 720 described above and depicted in FIGS. 20A-20D .
- the calibration frame 700 can comprise a plurality of upright posts 75 that are angled away from the calibration frame by about 10°.
- the active markers 720 are mounted on the posts 75 that are angled back by 10°, and this angulation keeps the markers 720 facing toward the cameras despite the patient being horizontal.
- the front markers 720 can have less chance of obscuring the rear markers 720 .
- posts 75 that are farthest away from the camera or farthest from a detection device of the tracking system 3417 can be taller and spaced farther laterally than the posts 75 closest to the camera.
- the calibration frame 700 can comprise markers 730 that are both radio-opaque for detection by a medical imaging scanner, and visible by the cameras or otherwise detectable by the real-time tracking system 3417 .
- the relationship between radio-opaque 730 and active markers ( 730 , 720 ) does not need to be measured or established because they are one in the same. Therefore, in some embodiments, as soon as the position is determined from the CT scan (or other imaging scan), the spatial relationship between the robot 15 and anatomy of the patient 18 can be defined.
- the targeting fixture 690 can comprise a flexible roll configuration.
- the targeting fixture 690 can comprise three or more radio-opaque markers 730 that define a rigid outer frame and nine or more active markers 720 embedded in a flexible roll of material (for example, the flexible roll 705 in FIG. 21A ).
- radio-opaque markers 730 are visible on CT scans and/or other medical diagnostic images, such as MRI, or reconstructions from O-arm or Iso-C scans, and their centroids can be determined from the 3D image.
- Active markers 720 include tracked markers 720 that have 3D coordinates that are detectable in real-time using cameras or other means.
- Some embodiments can utilize active marker systems based on reflective optical systems such as Motion Analysis Inc., or Peak Performance.
- Other suitable technologies include infrared-emitting marker systems such as Optotrak, electromagnetic systems such as Medtronic's Axiem®, or Flock of Birds®, or a local positioning system (“LPS”) described by Smith et al. in U.S. Patent Publication No. 2007/0238985.
- Flock Of Birds® is a registered trademark of Ascension Technology Corporation.
- Axiem is a trademark of Medtronic, Inc., and its affiliated companies.
- Medtronic® is a registered trademark used for Surgical and Medical Apparatus, Appliances and Instruments.
- the flexible roll 705 can comprise self-adhering film, such as, for example and without limitation, 3MTMIobanTM adhesive film (iodine-impregnated transparent surgical drape) similar to routinely used operating room product model 6651 EZ (3M, St. Paul, Minn.). IobanTM is a trademark of 3M company.
- the radio-opaque and active markers 730 , 720
- each radio-opaque marker 730 coupled to three or more active markers 720 .
- the markers can simultaneously serve as radio-opaque and active markers (for example, an active marker 720 whose position can be detected from cameras or other sensors), and the position determined from the 3D medical image can substantially exactly correspond to the center of the marker 720 .
- active markers for example, an active marker 720 whose position can be detected from cameras or other sensors
- the position determined from the 3D medical image can substantially exactly correspond to the center of the marker 720 .
- as few as three such markers 720 could be embedded in the flexible roll 705 and still permit determination of the spatial relationship between the robot 15 and the anatomy of the patient 18 .
- radio-opaque markers 730 and active markers 720 are not one in the same, in some embodiments the at least three active markers 720 must be rigidly connected to each radio-opaque marker 730 because three separate non-collinear points are needed to unambiguously define the relative positions of points on a rigid body. That is, if only one or 2 active markers 720 are viewed, there is more than one possible calculated position where a
- other considerations can be used to permit the use of two active markers 720 per radio-opaque marker 730 .
- the location of the radio-opaque marker 730 can be determined as the mean location of the two active markers 720 .
- the two active markers 720 and the radio-opaque marker 730 are intentionally positioned collinearly but with the radio-opaque marker 730 closer to one active marker 720 than the other (see for example FIG.
- the radio-opaque marker 730 must be at one of two possible positions along the line in space formed by the two active markers 720 (see FIG. 21B ).
- the flexible roll 705 is configured so that each pair of active markers 720 is oriented (when in its final position) with one marker 720 more toward the center of the flexible roll 705 , then it can be determined from the orientations of all markers or certain combinations of markers from different regions which of the two possible positions within each region is the correct position for the radio-opaque marker 730 (see FIG. 21C showing flexible roll 705 showed rolled on a torso and shown unrolled on a torso with markers 720 , 730 in place).
- the radio-opaque markers 730 can be positioned toward the inside of the frame 705 ), with marker groups nearer to the top of the figure having the radio-opaque marker 730 positioned below the active markers 720 and marker groups near the bottom of the figure having the radio-opaque marker positioned above the active markers 720 .
- the flexible roll 705 can be positioned across the patient's back or other area, and adhered to the skin of the patient 18 as it is unrolled. In some embodiments, knowing the spatial relationship between each triad of active markers 720 and the rigidly coupled radio-opaque marker 730 , it is possible to establish the relationship between the robot 15 (position established by its own active markers 720 ) and the anatomy (visualized together with radio-opaque markers 730 on MM, CT, or other 3D scan). In some embodiments, the flexible roll 705 can be completely disposable.
- the flexible roll 705 can comprise reusable marker groups integrated with a disposable roll with medical grade adhesive on each side to adhere to the patient 18 and the marker groups 720 , 730 .
- the flexible roll 705 can comprise a drape incorporated into the flexible roll 705 for covering the patient 18 , with the drape configured to fold outwardly from the roll 705 .
- the roll 705 can have a desired stiffness such that the roll 705 does not substantially change its position relative to the bony anatomy of the patient 18 .
- a conventional radiolucent wire can be embedded in the perimeter of the frame 700 .
- a chain of plastic beads such as the commercially available tripods shown in FIG. 21D , or a commercially available “snake light” type fixture, can be employed to provide desired stiffness to the unrolled fixture such that it maintains its position after unrolling occurs.
- the beads of the chain of plastic beads as shown can be affixed to each other with a high friction so that they hold their position once shifted.
- chains of beads can be incorporated into, and define, a perimeter of the frame 700 .
- this type of frame could be loaded with conventional chemicals that mix at the time of application.
- components of a conventional two-part epoxy could be held in separate fragile baggies within the frame that pop open when the user first starts to manipulate the beads.
- the user would attach the frame to the patient 18 , and mold it to the contours of the patient's body. After a short period of time, the frame 700 would solidify to form a very rigid frame, locking the beads in their current orientation.
- the targeting fixture 690 can be an adherable fixture, configured for temporary attachment to the skin of a patient 18 .
- the targeting fixture 690 can be temporarily adhered to the patient 18 during imaging, removed, and then subsequently reattached during a follow-up medical procedure, such as a surgery.
- the targeting fixture 690 can be applied to the skull of a patient 18 for use in placement of electrodes for deep brain stimulation.
- this method can use a single fixture 690 , or two related fixtures. In this instance, the two related fixtures can share the same surface shape.
- one fixture 690 can be temporarily attached at the time of medical image scanning, and can include radio-opaque markers 730 (but not active markers 720 ), and the second fixture 690 can be attached at the time of surgery, and can include active markers 720 (but not radio-opaque markers 730 ).
- the first fixture (for scanning) can comprise a frame 690 with three or more embedded radio-opaque markers 730 , and two or more openings 740 for application of markings (the markings shown as 750 in FIG. 22B ).
- the device 690 can be adhered to the scalp of a patient 18 , and the openings 740 can be used to paint marks on the scalp with, for example, henna or dye (shown as “+” marks 750 in FIG. 22B ).
- the patient 18 can receive a 3D scan (for example an MRI or CT scan) in which the radio-opaque markers 730 are captured. As illustrated by FIG.
- the fixture 690 can then be removed, leaving the dye marks 750 on the scalp.
- the patient 18 can return, and the surgeon or technician can attach the 2nd fixture (for surgery) containing active markers 720 for intraoperative tracking (see FIG. 22C-22D ).
- the fixture 690 shown in FIG. 22C can be mounted to the scalp, spatially positioned and oriented in the same position as the previously adhered first fixture (shown in FIG. 22A ) by ensuring that the previously placed dye marks 750 line up with holes in the second fixture 690 (see the alignment arrows depicted in FIG. 22C ).
- the fixture 690 can have a transparent frame for good visualization.
- the above-described method assumes that the locations of the marks 750 do not change over the period of time between the scan and the return of the patient 18 for surgery. Since the relative positions between the radio-opaque markers 730 from the temporary (first) fixture 690 (which appear in the scan) and the active markers 720 on the second applied fixture 690 are known through a calibration and/or by careful manufacturing of the fixtures 690 , the coordinate system of the anatomy and the coordinate system of the active markers 720 can be synchronized so that the robot 15 can target any planned trajectory on the 3D image as described further herein.
- this method can enable image guidance with only one pre-op scan and without requiring the patient 18 to go home after a pre-op scan. This circumvents the need for a patient 18 to take care of wounds from targeting screws that are invasively drilled into the skull of the patient 18 .
- the targeting fixture 690 can comprise a conventional clamping mechanism for securely attaching the targeting fixture 690 to the patient 18 .
- the targeting fixture 690 can be configured to clamp to the spinous process 6301 of a patient 18 after the surgeon has surgically exposed the spinous process.
- FIG. 23 shows a dynamic tracking device 2300 mounted to the spinous process 2310 in the lumbar spine of a patient 18 in accordance with some embodiments of the invention. This targeting fixture is used with Medtronic's StealthStation.
- the relationship between the markers 720 , 730 and the bony anatomy of the patient 18 can be established using a registration process wherein known landmarks are touched with a digitizing probe at the same time that the markers on the tracker are visible.
- the probe itself can have a shaft protruding from a group of markers 720 , 730 , thereby permitting the tracking system 3417 to calculate the coordinates of the probe tip relative to the markers 720 , 730 .
- the clamping mechanism of the targeting fixture 690 can be configured for clamping to the spinous process 2310 , or can be configured for anchoring to bone of the patient 18 such that the fixture 690 is substantially stationary and not easily moved.
- the targeting fixture 690 can comprise at least three active markers 720 and distinct radio-opaque markers 730 that are detected on the CT or other 3D image, preferably near the clamp (to be close to bone).
- the active markers 720 themselves must be configured to be visualized accurately on CT or other 3D image.
- the portion of the fixture 690 containing a radio-opaque marker 730 can be made to be detachable to enable removal from the fixture after the 3D image is obtained.
- a combination of radio-opaque 730 and active markers 720 can allow tracking with the robot 15 in the same way that is possible with the frame-type targeting fixtures 690 described above.
- one aspect of the software and/or firmware disclosed herein is a unique process for locating the center of the above-described markers 730 that takes advantage of the fact that a CT scan can comprise slices, typically spaced 1.5 mm or more apart in the z direction, and sampled with about 0.3 mm resolution in the x-axis and y-axis directions.
- a CT scan can comprise slices, typically spaced 1.5 mm or more apart in the z direction, and sampled with about 0.3 mm resolution in the x-axis and y-axis directions.
- the diameter of the radio-opaque markers 730 is several times larger than this slice spacing, different z slices of the sphere will appear as circles of different diameters on each successive x-y planar slice.
- a z slice substantially exactly through the center of the sphere can yield a circle with a radius R that is substantially the same as that of the sphere.
- a z slice through a point at the top or bottom of the sphere can yield a circle with a radius R approximating zero.
- a calibration frame 700 is attached to the patient's 18 skin, substantially within the region at which surgery/biopsy is to take place.
- the patient 18 receives a CT scan either supine or prone, whichever positioning orients the calibration frame 700 upward.
- the surgeon subsequently manipulates three planar views of the patient's 18 CT images with rotations and translations.
- the surgeon then draws trajectories on the images that define the desired position, and strike angle of the end-effectuator 30 .
- automatic calibration can be performed in order to obtain the centers of radio-opaque makers 730 of the calibration frame 700 , and to utilize the stored relationship between the active markers 720 and radio-opaque markers 730 . This procedure permits the robot 15 to move in the coordinate system of the anatomy and/or drawn trajectories.
- the robot 15 then will move to the desired position. In some embodiments, if forceful resistance beyond a pre-set tolerance is exceeded, the robot 15 will halt. In some further embodiments, the robot 15 can hold the guide tube 50 at the desired position and strike angle to allow the surgeon to insert a conventional screw or needle (for example, needle 7405 , 7410 or biopsy needle 8110 ). In some embodiments, if tissues move in response to applied force or due to breathing, the movement will be tracked by optical markers 720 , and the robot's position will automatically be adjusted.
- a conventional screw or needle for example, needle 7405 , 7410 or biopsy needle 8110
- a targeting fixture 690 comprising a small tree of optical markers, for example, can be attached to a bony prominence in the area of interest.
- conventional calibration procedures for image guidance can be utilized to establish the anatomy relative to the optical tracking system 3417 and medical images.
- the targeting fixture 690 can contain rigidly mounted, substantially permanent or detachable radio-opaque markers 730 that can be imaged with a CT scan.
- the calibration procedures consistent with those stated for the calibration frame 700 can be utilized to establish the anatomy relative to the robot 15 and the medical image.
- the surgeon manipulates three planar views of the patient's CT images with rotations and translations. In some embodiments, the surgeon then draws trajectories on the images that define the desired position and strike angle of the end-effectuator 30 . In some embodiments, the robot 15 moves to the desired position. In some embodiments, if forceful resistance beyond a pre-set tolerance is exceeded, the robot 15 will halt. In some embodiments, the robot 15 holds the guide tube 50 at the desired position and strike angle to allow the surgeon to insert a conventional screw. In some embodiments, if tissues move in response to applied force or due to breathing, the movement will be tracked by optical markers 720 , and the robot's position will automatically be adjusted.
Abstract
A medical robot system, including a robot coupled to an effectuator element with the robot configured for controlled movement and positioning. The system may include a transmitter configured to emit one or more signals, and the transmitter is coupled to an instrument coupled to the effectuator element. The system may further include a motor assembly coupled to the robot and a plurality of receivers configured to receive the one or more signals emitted by the transmitter. A control unit is coupled to the motor assembly and the plurality of receivers, and the control unit is configured to supply one or more instruction signals to the motor assembly. The instruction signals can be configured to cause the motor assembly to selectively move the effectuator element
Description
- This application is a continuation of U.S. patent application Ser. No. 15/609,305 filed on May 31, 2017 (published U.S. Pat. Pub. No. 2017-0281145), which is a continuation of U.S. patent application Ser. No. 13/924,505 filed on Jun. 21, 2013 (now U.S. Pat. No. 9,782,229), which is incorporated herein by reference in its entirety for all purposes. Application Ser. No. 13/924,505 claims priority to U.S. Provisional Pat. App. No. 61/662,702 filed Jun. 21, 2012 (expired) and U.S. Provisional Pat. App. No. 61/800,527 filed Mar. 15, 2013 (expired), which are incorporated herein by reference in their entireties for all purposes.
- Various medical procedures require the precise localization of a three-dimensional position of a surgical instrument within the body in order to effect optimized treatment. Limited robotic assistance for surgical procedures is currently available. One of the characteristics of many of the current robots used in surgical applications which make them error prone is that they use an articular arm based on a series of rotational joints. The use of an articular system may create difficulties in arriving at an accurately targeted location because the level of any error is increased over each joint in the articular system.
- Some embodiments of the invention provide a surgical robot (and optionally an imaging system) that utilizes a Cartesian positioning system that allows movement of a surgical instrument to be individually controlled in an x-axis, y-axis and z-axis. In some embodiments, the surgical robot can include a base, a robot arm coupled to and configured for articulation relative to the base, as well as an end-effectuator coupled to a distal end of the robot arm. The effectuator element can include the surgical instrument or can be configured for operative coupling to the surgical instrument. Some embodiments of the invention allow the roll, pitch and yaw rotation of the end-effectuator and/or surgical instrument to be controlled without creating movement along the x-axis, y-axis, or z-axis.
-
FIG. 1 is a partial perspective view of a room in which a medical procedure is taking place by using a surgical robot. -
FIG. 2 is a perspective view of a surgical robot according to an embodiment of the invention. -
FIGS. 3A-3B are perspective views of the surgical robot illustrated inFIG. 2 , which show the movement of the base of the surgical robot in the z-axis direction in accordance with an embodiment of the invention. -
FIG. 4 is a partial perspective view of the surgical robot ofFIG. 2 which shows how the robot arm can be moved in the x-axis direction. -
FIGS. 5A-5B are partial perspective views of the surgical robot ofFIG. 2 , which show how the robot arm can be moved in the y-axis direction. -
FIG. 6 is a perspective view of a portion of the robot arm ofFIG. 2 showing how an effectuator element can be twisted about a y-axis. -
FIG. 7 is a perspective view of a portion of a robot arm ofFIG. 2 showing how an effectuator element can be pivoted about a pivot axis that is perpendicular to the y-axis. -
FIGS. 8A-8B are partial perspective views of the surgical robot ofFIG. 2 , which show the movement of asurgical instrument 35 along the z-axis from an effectuator element. -
FIG. 9 is a system diagram which shows local positioning sensors, a controlling PC, and a Radiofrequency (RF) transmitter in accordance with an embodiment of the invention. -
FIG. 10 is a system diagram of the controlling PC, user input, and motors for controlling the robot in accordance with an embodiment of the invention. -
FIG. 11 is a flow chart diagram for general operation of a surgical robot in accordance with one embodiment of the invention. -
FIG. 12 is a flow chart diagram for a closed screw/needle insertion performed using a surgical robot in accordance with one embodiment of the invention. -
FIG. 13 is a flow chart diagram of a safe zone surgery performed using a surgical robot as described herein in accordance with one embodiment of the invention. -
FIG. 14 is a flow chart diagram of a flexible catheter insertion procedure performed using a surgical robot as described herein in accordance with one embodiment of the invention. -
FIG. 15A shows a screenshot of a monitor display showing a set up of the anatomy in X, Y and Z views in accordance with one embodiment of the invention. -
FIG. 15B shows a screenshot of a monitor display showing what the user views during an invasive procedure in accordance with one embodiment of the invention. -
FIG. 16 depicts a surgical robot having a plurality of optical markers mounted for tracking movement in an x-direction in accordance with one embodiment of the invention. -
FIGS. 17A-17B depict surgical instruments having a stop mechanism in accordance with one embodiment of the invention. -
FIGS. 17C-17E illustrate tools for manually adjusting a drill stop with reference to drill bit markings in accordance with one embodiment of the invention. -
FIGS. 17F-17J illustrate tools for locking and holding a drill bit in a set position in accordance with one embodiment of the invention. -
FIGS. 18A-18B depicts an end-effectuator having a clearance mechanism in accordance with one embodiment of the invention. -
FIG. 19A-19B depicts an end-effectuator having an attachment element for applying distraction and/or compression forces in accordance with one embodiment of the invention. -
FIGS. 20A-20E show the use of calibration frames with the guidance system in accordance with one embodiment of the invention. -
FIG. 21A depicts flexible roll configurations of a targeting fixture in accordance with one embodiment of the invention. -
FIG. 21B shows possible positions of markers along a line in space in accordance with one embodiment of the invention. -
FIG. 21C depicts flexible roll configurations of a targeting fixture in accordance with one embodiment of the invention. -
FIG. 21D shows a fixture that can be employed to provide desired stiffness to the unrolled fixture such that it maintains its position after unrolling occurs in accordance with one embodiment of the invention. -
FIGS. 22A-22D depict a targeting fixture and method configured for application to the skull of a patient in accordance with one embodiment of the invention. -
FIG. 23 depicts a dynamic tracking device mounted to the spinous process of the lumbar spine of a patient in accordance with one embodiment of the invention. -
FIGS. 24-27 illustrate methods for the robot in accordance with one embodiment of the invention. -
FIGS. 28A-28B illustrate methods for calculating position and/or orientation of an end-effectuator for the robot in accordance with one embodiment of the invention. -
FIGS. 29-33 illustrate methods for the robot in accordance with other embodiments of the invention. -
FIG. 34 illustrates a computing platform that enables implementation of various embodiments of the invention. -
FIGS. 35A-35B display a surgical robot in accordance with one embodiment of the invention. - Referring now to
FIGS. 1 and 35A , some embodiments include asurgical robot system 1 is disclosed in aroom 10 where a medical procedure is occurring. In some embodiments, thesurgical robot system 1 can comprise asurgical robot 15 and one ormore positioning sensors 12. In this aspect, thesurgical robot 15 can comprise a display means 29 (including for example adisplay 150 shown inFIG. 10 ), and ahousing 27. In some embodiments adisplay 150 can be attached to thesurgical robot 15, whereas in other embodiments, a display means 29 can be detached fromsurgical robot 15, either withinsurgical room 10 or in a remote location. In some embodiments, thehousing 27 can comprise arobot arm 23, and an end-effectuator 30 coupled to therobot arm 23 controlled by at least onemotor 160. For example, in some embodiments, thesurgical robot system 1 can include amotor assembly 155 comprising at least one motor (represented as 160 inFIG. 10 ). In some embodiments, the end-effectuator 30 can comprise asurgical instrument 35. In other embodiments, the end-effectuator 30 can be coupled to thesurgical instrument 35. As used herein, the term “end-effectuator” is used interchangeably with the terms “end-effectuator,” “effectuator element,” and “effectuator element.” In some embodiments, the end-effectuator 30 can comprise any known structure for effecting the movement of thesurgical instrument 35 in a desired manner. - In some embodiments, prior to performance of an invasive procedure, a three-dimensional (“3D”) image scan can be taken of a desired surgical area of the
patient 18 and sent to a computer platform in communication withsurgical robot 15 as described herein (see for example theplatform 3400 including thecomputing device 3401 shown inFIG. 34 ). In some embodiments, a physician can then program a desired point of insertion and trajectory forsurgical instrument 35 to reach a desired anatomical target within or upon the body ofpatient 18. In some embodiments, the desired point of insertion and trajectory can be planned on the 3D image scan, which in some embodiments, can be displayed on display means 29. In some embodiments, a physician can plan the trajectory and desired insertion point (if any) on a computed tomography scan (hereinafter referred to as “CT scan”) of apatient 18. In some embodiments, the CT scan can be an isocentric C-arm type scan, an O-arm type scan, or intraoperative CT scan as is known in the art. However, in some embodiments, any known 3D image scan can be used in accordance with the embodiments of the invention described herein. - In some embodiments, the
surgical robot system 1 can comprise a local positioning system (“LPS”) subassembly to track the position ofsurgical instrument 35. The LPS subassembly can comprise at least one radio-frequency (RF)transmitter 120 that is coupled were affixed to the end-effectuator 30 or thesurgical instrument 35 at a desired location. In some embodiments, the at least oneRF transmitter 120 can comprise a plurality oftransmitters 120, such as, for example, at least threeRF transmitters 120. In another embodiment, the LPS subassembly can comprise at least oneRF receiver 110 configured to receive one or more RF signals produced by the at least oneRF transmitter 120. In some embodiments, the at least oneRF receiver 110 can comprise a plurality ofRF receivers 110, such as, for example, at least threeRF receivers 110. In these embodiments, theRF receivers 110 can be positioned at known locations within theroom 10 where the medical procedure is to take place. In some embodiments, theRF receivers 110 can be positioned at known locations within theroom 10 such that theRF receivers 110 are not coplanar within a plane that is parallel to the floor of theroom 10. - In some embodiments, during use, the time of flight of an RF signal from each
RF transmitter 120 of the at least oneRF transmitter 120 to eachRF receiver 110 of the at least one RF receiver 110 (e.g., one RF receiver, two RF receivers, three RF receivers, etc.) can be measured to calculate the position of eachRF transmitter 120. Because the velocity of the RF signal is known, the time of flight measurements result in at least three distance measurements for each RF transmitter 120 (one to each RF receiver 110). - In some embodiments, the
surgical robot system 1 can comprise a control device (for example acomputer 100 having a processor and a memory coupled to the processor). In some embodiments, the processor of thecontrol device 100 can be configured to perform time of flight calculations as described herein. Further, in some embodiments, can be configured to provide a geometrical description of the location of the at least oneRF transmitter 120 with respect to an operative end of thesurgical instrument 35 or end-effectuator 30 that is utilized to perform or assist in performing an invasive procedure. In some further embodiments, the position of theRF transmitter 120, as well as the dimensional profile of thesurgical instrument 35 or theeffectuator element 30 can be displayed on a monitor (for example on a display means 29 such as thedisplay 150 shown inFIG. 10 ). In one embodiment, the end-effectuator 30 can be a tubular element (for example a guide tube 50) that is positioned at a desired location with respect to, for example, a patient's 18 spine to facilitate the performance of a spinal surgery. In some embodiments, theguide tube 50 can be aligned with the z axis 70 defined by a correspondingrobot motor 160 or, for example, can be disposed at a selected angle relative to the z-axis 70. In either case, the processor of the control device (i.e. the computer 100) can be configured to account for the orientation of the tubular element and the position of theRF transmitter 120. As further described herein, in some embodiments, the memory of the control device (computer 100 for example) can store software for performing the calculations and/or analyses required to perform many of the surgical method steps set forth herein. - Another embodiment of the disclosed
surgical robot system 1 involves the utilization of arobot 15 that is capable of moving the end-effectuator 30 along x-, y-, and z-axes (see 66, 68, 70 inFIG. 35B ). In this embodiment, the x-axis 66 can be orthogonal to the y-axis 68 and z-axis 70, the y-axis 68 can be orthogonal to the x-axis 66 and z-axis 70, and the z-axis 70 can be orthogonal to the x-axis 66 and the y-axis 68. In some embodiments, therobot 15 can be configured to effect movement of the end-effectuator 30 along one axis independently of the other axes. For example, in some embodiments, therobot 15 can cause the end-effectuator 30 to move a given distance along the x-axis 66 without causing any significant movement of the end-effectuator 30 along the y-axis 68 or z-axis 70. - In some further embodiments, the end-
effectuator 30 can be configured for selective rotation about one or more of the x-axis 66, y-axis 68, and z-axis 70 (such that one or more of the Cardanic Euler Angles (e.g., roll, pitch, and/or yaw) associated with the end-effectuator 30 can be selectively controlled). In some embodiments, during operation, the end-effectuator 30 and/orsurgical instrument 35 can be aligned with a selected orientation axis (labeled “Z Tube” inFIG. 35B ) that can be selectively varied and monitored by an agent (forexample computer 100 and platform 3400) that can operate thesurgical robot system 1. In some embodiments, selective control of the axial rotation and orientation of the end-effectuator 30 can permit performance of medical procedures with significantly improved accuracy compared to conventional robots that utilize, for example, a six degree offreedom robot arm 23 comprising only rotational axes. - In some embodiments, as shown in
FIG. 1 , therobot arm 23 that can be positioned above the body of thepatient 18, with the end-effectuator 30 selectively angled relative to the z-axis toward the body of thepatient 18. In this aspect, in some embodiments, the roboticsurgical system 1 can comprise systems for stabilizing therobotic arm 23, the end-effectuator 30, and/or thesurgical instrument 35 at their respective positions in the event of power failure. In some embodiments, therobotic arm 23, end-effectuator 30, and/orsurgical instrument 35 can comprise a conventional worm-drive mechanism (not shown) coupled to therobotic arm 23, configured to effect movement of the robotic arm along the z-axis 70. In some embodiments, the system for stabilizing therobotic arm 23, end-effectuator 30, and/orsurgical instrument 35 can comprise a counterbalance coupled to therobotic arm 23. In another embodiment, the means for maintaining therobotic arm 23, end-effectuator 30, and/orsurgical instrument 35 can comprise a conventional brake mechanism (not shown) that is coupled to at least a portion of therobotic arm 23, such as, for example, the end-effectuator 30, and that is configured for activation in response to a loss of power or “power off” condition of thesurgical robot 15. - Referring to
FIG. 1 , in some embodiments, thesurgical robot system 1 can comprise a plurality ofpositioning sensors 12 configured to receive RF signals from the at least one conventional RF transmitter (not shown) located withinroom 10. In some embodiments, the at least oneRF transmitter 120 can be disposed on various points on thesurgical robot 15 and/or onpatient 18. For example, in some embodiments, the at least oneRF transmitter 120 can be attached to one or more of thehousing 27,robot arm 23, end-effectuator 30, andsurgical instrument 35. Some embodiments includepositioning sensors 12 that in some embodiments compriseRF receivers 110. In some embodiments,RF receivers 110 are in communication with a computer platform as described herein (see for example 3400 comprising acomputing device 3401FIG. 34 ) that receives the signal from theRF transmitters 120. In some embodiments, eachtransmitter 120 of the at least oneRF transmitter 120 can transmit RF energy on a different frequency so that the identity of eachtransmitter 120 in theroom 10 can be determined. In some embodiments, the location of the at least oneRF transmitter 120, and, consequently, the objects to which thetransmitters 120 are attached, are calculated by the computer (e.g.,computing device 3401 inFIG. 34 ) using time-of-flight processes. - In some embodiments, the computer (not shown in
FIG. 1 ) is also in communication withsurgical robot 15. In some embodiments, a conventional processor (not shown) of thecomputer 100 of thecomputing device 3401 can be configured to effect movement of thesurgical robot 15 according to a preplanned trajectory selected prior to the procedure. For example, in some embodiments, thecomputer 100 of thecomputing device 3401 can userobotic guidance software 3406 and robotic guidance data storage 3407 (shown inFIG. 34 ) to effect movement of thesurgical robot 15. - In some embodiments, the position of
surgical instrument 35 can be dynamically updated so thatsurgical robot 15 is aware of the location ofsurgical instrument 35 at all times during the procedure. Consequently, in some embodiments, thesurgical robot 15 can move thesurgical instrument 35 to the desired position quickly, with minimal damage topatient 18, and without any further assistance from a physician (unless the physician so desires). In some further embodiments, thesurgical robot 15 can be configured to correct the path ofsurgical instrument 35 if thesurgical instrument 35 strays from the selected, preplanned trajectory. - In some embodiments, the
surgical robot 15 can be configured to permit stoppage, modification, and/or manual control of the movement of the end-effectuator 30 and/orsurgical instrument 35. Thus, in use, in some embodiments, an agent (e.g., a physician or other user) that can operate thesystem 1 has the option to stop, modify, or manually control the autonomous movement of end-effectuator 30 and/orsurgical instrument 35. Further, in some embodiments, tolerance controls can be preprogrammed into thesurgical robot 15 and/or processor of the computer platform 3400 (such that the movement of the end-effectuator 30 and/orsurgical instrument 35 is adjusted in response to specified conditions being met). For example, in some embodiments, if thesurgical robot 15 cannot detect the position ofsurgical instrument 35 because of a malfunction in the at least oneRF transmitter 120, then thesurgical robot 15 can be configured to stop movement of end-effectuator 30 and/orsurgical instrument 35. In some embodiments, ifsurgical robot 15 detects a resistance, such as a force resistance or a torque resistance above a tolerance level, then thesurgical robot 15 can be configured to stop movement of end-effectuator 30 and/orsurgical instrument 35. - In some embodiments, the
computer 100 for use in the system (for example represented by computing device 3401), as further described herein, can be located withinsurgical robot 15, or, alternatively, in another location withinsurgical room 10 or in a remote location. In some embodiments, thecomputer 100 can be positioned in operative communication withpositioning sensors 12 andsurgical robot 15. - In some further embodiments, the
surgical robot 15 can also be used with existing conventional guidance systems. Thus, alternative conventional guidance systems beyond those specifically disclosed herein are within the scope and spirit of the invention. For instance, a conventionaloptical tracking system 3417 for tracking the location of the surgical device, or a commercially available infraredoptical tracking system 3417, such as Optotrak® (Optotrak® is a registered trademark of Northern Digital Inc. Northern Digital, Waterloo, Ontario, Canada), can be used to track the patient 18 movement and the robot's base 25 location and/or intermediate axis location, and used with thesurgical robot system 1. In some embodiments in which thesurgical robot system 1 comprises a conventional infraredoptical tracking system 3417, thesurgical robot system 1 can comprise conventional optical markers attached to selected locations on the end-effectuator 30 and/or thesurgical instrument 35 that are configured to emit or reflect light. In some embodiments, the light emitted from and/or reflected by the markers can be read by cameras and/or optical sensors and the location of the object can be calculated through triangulation methods (such as stereo-photogrammetry). - Referring now to
FIG. 2 , it is seen that, in some embodiments, thesurgical robot 15 can comprise a base 25 connected to wheels 31. The size and mobility of these embodiments can enable the surgical robot to be readily moved from patient to patient and room to room as desired. As shown, in some embodiments, thesurgical robot 15 can further comprise a case 40 that is slidably attached tobase 25 such that the case 40 can slide up and down along the z-axis 70 substantially perpendicular to the surface on whichbase 25 sits. In some embodiments, thesurgical robot 15 can include a display means 29, and ahousing 27 which containsrobot arm 23. - As described earlier, the end-
effectuator 30 can comprise asurgical instrument 35, whereas in other embodiments, the end-effectuator 30 can be coupled to thesurgical instrument 35. In some embodiments, it isarm 23 can be connected to the end-effectuator 30, withsurgical instrument 35 being removably attached to the end-effectuator 30. - Referring now to
FIGS. 2, 3A-3B, 4, 5A-5B, 6, 7, and 8A-8B , in some embodiments, theeffectuator element 30 can include anouter surface 30 d, and can comprise adistal end 30 a defining a beveled leadingedge 30 b and a non-beveledleading edge 30 c. In some embodiments, thesurgical instrument 35 can be any known conventional instrument, device, hardware component, and/or attachment that is used during performance of a an invasive or non-invasive medical procedure (including surgical, therapeutic, and diagnostic procedures). For example and without limitation, in some embodiments, thesurgical instrument 35 can be embodied in or can comprise a needle 7405, 7410, a conventional probe, a conventional screw, a conventional drill, a conventional tap, a conventional catheter, a conventional scalpel forceps, or the like. In addition or in the alternative, in some embodiments, thesurgical instrument 35 can be a biological delivery device, such as, for example and without limitation, a conventional syringe, which can distribute biologically acting compounds throughout the body of apatient 18. In some embodiments, thesurgical instrument 35 can comprise a guide tube 50 (also referred to herein as a “Z-tube 50”) that defines a central bore configured for receipt of one or more additionalsurgical instruments 35. - In some embodiments, the
surgical robot 15 is moveable in a plurality of axes (for instance x-axis 66, y-axis 68, and z-axis 70) in order to improve the ability to accurately and precisely reach a target location. Some embodiments include arobot 15 that moves on a Cartesian positioning system; that is, movements in different axes can occur relatively independently of one another instead of at the end of a series of joints. - Referring now to
FIGS. 3A and 3B , the movement of case 40 relative to base 25 ofsurgical robot 15 is represented as a change of height of thesystem 1 and the position of the case 40 with respect to thebase 25. As illustrated, in some embodiments, case 40 can be configured to be raised and lowered relative to thebase 25 along the z-axis. Some embodiments include ahousing 27 that can be attached to case 40 and be configured to move in the z-direction (defined by z-frame 72) with case 40 when case 40 is raised and lowered. Consequently, in some embodiments,arm 23, the end-effectuator 30, andsurgical instrument 35 can be configured to move with case 40 as case 40 is raised and lowered relative tobase 25. - In a further embodiment, referring now to
FIG. 4 ,housing 27 can be slidably attached to case 40 so that it can extend and retract along the x-axis 66 relative to case 40 and substantially perpendicularly to the direction case 40 moves relative tobase 25. Consequently, in some embodiments, therobot arm 23, the end-effectuator 30, andsurgical instrument 35 can be configured to move withhousing 27 ashousing 27 is extended and retracted relative to case 40. - Referring now to
FIGS. 5A and 5B , the extension ofarm 23 along the y-axis 68 is shown. In some embodiments,robot arm 23 can be extendable along the y-axis 68 relative to case 40,base 25, andhousing 27. Consequently, in some embodiments, the end-effectuator 30 andsurgical instrument 35 can be configured to move witharm 23 asarm 23 is extended and retracted relative tohousing 27. In some embodiments,arm 23 can be attached to a low profile rail system (not shown) which is encased byhousing 27. - Referring now to
FIGS. 6, 7 andFIGS. 8A-B , the movement of the end-effectuator 30 is shown.FIG. 6 shows an embodiment of an end-effectuator 30 that is configured to rotate about the y-axis 68, performing a rotation having aspecific roll 62.FIG. 7 shows an embodiment of an end-effectuator 30 that is configured to rotate about the x-axis 66, performing a rotation having aspecific pitch 60.FIG. 8 shows an embodiment of an end-effectuator 30 that is configured to raise and lowersurgical instrument 35 along a substantially vertical axis, which can be a secondary movable axis 64, referred to as “Z-tube axis 64”. In some embodiments, the orientation of theguide tube 50 can be initially aligned with z-axis 70, but such orientation can change in response to changes inroll 62 and/orpitch 60. -
FIG. 9 shows a system diagram of the3D positioning sensors 110,computer 100, andRF transmitters 120 in accordance with some embodiments of the invention is provided. As shown,computer 100 is in communication withpositioning sensors 110. In some embodiments, during operation,RF transmitters 120 are attached to various points on thesurgical robot 15. In some embodiments, theRF transmitters 120 can also be attached to various points on or around an anatomical target of apatient 18. In some embodiments,computer 100 can be configured to send a signal to theRF transmitters 120, prompting theRF transmitters 120 to transmit RF signals that are read by thepositioning sensors 110. In some embodiments, thecomputer 100 can be coupled to theRF transmitters 120 using any conventional communication means, whether wired or wireless. In some embodiments, thepositioning sensors 110 can be in communication withcomputer 100, which can be configured to calculate the location of the positions of all theRF transmitters 120 based on time-of-flight information received from thepositioning sensors 110. In some embodiments,computer 100 can be configured to dynamically update the calculated location of thesurgical instrument 35 and/or end-effectuator 30 being used in the procedure, which can be displayed to the agent. - Some embodiments can include a system diagram of
surgical robot system 1 having acomputer 100, a display means 29 comprising adisplay 150,user input 170, andmotors 160, provided as illustrated inFIG. 10 . In some embodiments,motors 160 can be installed in thesurgical robot 15 and control the movement of the end-effectuator 30 and/orsurgical instrument 35 as described above. In some embodiments,computer 100 can be configured to dynamically update the location of thesurgical instrument 35 being used in the procedure, and can be configured to send appropriate signals to themotors 160 such that thesurgical robot 15 has a corresponding response to the information received bycomputer 100. For example, in some embodiments, in response to information received bycomputer 100, thecomputer 100 can be configured to prompt themotors 160 to move thesurgical instrument 35 along a preplanned trajectory. - In some embodiments, prior to performance of a medical procedure, such as, for example, an invasive surgical procedure,
user input 170 can be used to plan the trajectory for a desired navigation. After the medical procedure has commenced, if changes in the trajectory and/or movement of the end-effectuator 30 and/orsurgical instrument 35 are desired, a user can use theuser input 170 to input the desired changes, and thecomputer 100 can be configured to transmit corresponding signals to themotors 160 in response to theuser input 170. - In some embodiments, the
motors 160 can be or can comprise conventional pulse motors. In this aspect, in some embodiments, the pulse motors can be in a conventional direct drive configuration or a belt drive and pulley combination attached to thesurgical instrument 35. Alternatively, in other embodiments, themotors 160 can be conventional pulse motors that are attached to a conventional belt drive rack-and-pinion system or equivalent conventional power transmission component. - In some embodiments, the use of conventional linear pulse motors within the
surgical robot 15 can permit establishment of a non-rigid position for the end-effectuator 30 and/orsurgical instrument 35. Thus, in some embodiments, the end-effectuator 30 and/orsurgical instrument 35 will not be fixed in a completely rigid position, but rather the end-effectuator 30 and/or thesurgical instrument 35 can be configured such that an agent (e.g., a surgeon or other user) can overcome the x-axis 66 and y-axis 68, and force the end-effectuator 30 and/orsurgical instrument 35 from its current position. For example, in some embodiments, the amount of force necessary to overcome such axes can be adjusted and configured automatically or by an agent. In some embodiments, thesurgical robot 15 can comprise circuitry configured to monitor one or more of: (a) the position of therobot arm 23, the end-effectuator 30, and/or thesurgical instrument 35 along the x-axis 66, y-axis 68, and z-axis 70; (b) the rotational position (e.g., roll 62 and pitch 60) of therobot arm 23, the end-effectuator 30, and/or thesurgical instrument 35 relative to the x-(66), y-(68), and z-(70) axes; and (c) the position of the end-effectuator 30, and/or thesurgical instrument 35 along the travel of the re-orientable axis that is parallel at all times to the end-effectuator 30 and surgical instrument 35 (the Z-tube axis 64). - In one embodiment, circuitry for monitoring the positions of the x-axis 66, y-axis 68, z-axis 70, Z-tube axis 64,
roll 62, and/or pitch 60 can comprise relative or absolute conventional encoder units (also referred to as encoders) embedded within or functionally coupled to conventional actuators and/or bearings of at least one of themotors 160. Optionally, in some embodiments, the circuitry of thesurgical robot 15 can be configured to provide auditory, visual, and/or tactile feedback to the surgeon or other user when the desired amount of positional tolerance (e.g., rotational tolerance, translational tolerance, a combination thereof, or the like) for the trajectory has been exceeded. In some embodiments, the positional tolerance can be configurable and defined, for example, in units of degrees and/or millimeters. - In some embodiments, the
robot 15 moves into a selected position, ready for the surgeon to deliver a selectedsurgical instrument 35, such as, for example and without limitation, a conventional screw, a biopsy needle 8110, and the like. In some embodiments, as the surgeon works, if the surgeon inadvertently forces the end-effectuator 30 and/orsurgical instrument 35 off of the desired trajectory, then thesystem 1 can be configured to provide an audible warning and/or a visual warning. For example, in some embodiments, thesystem 1 can produce audible beeps and/or display a warning message on the display means 29, such as “Warning: Off Trajectory,” while also displaying the axes for which an acceptable tolerance has been exceeded. - In some embodiments, in addition to, or in place of the audible warning, a light illumination may be directed to the end-
effectuator 30, theguide tube 50, the operation area (i.e. the surgical field 17) of thepatient 18, or a combination of these regions. For example, some embodiments include at least one visual indication 900 capable of illuminating asurgical field 17 of apatient 18. Some embodiments include at least one visual indication 900 capable of indicating a target lock by projecting an illumination on asurgical field 17. In some embodiments, thesystem 1 can provide feedback to the user regarding whether therobot 15 is locked on target. In some other embodiments, thesystem 1 can provide an alert to the user regarding whether at least onemarker 720 is blocked, or whether thesystem 1 is actively seeking one ormore markers 720. - In some embodiments, the visual indication 900 can be projected by one or more conventional light emitting diodes mounted on or near the robot end-
effectuator 30. In some embodiments, the visual indication can comprise lights projected on thesurgical field 17 including a color indicative of the current situation. In some embodiments, a green projected light could represent a locked-on-target situation, whereas in some embodiments, a red illumination could indicate a trajectory error, or obscuredmarkers 720. In some other embodiments, a yellow illumination could indicate thesystem 1 is actively seeking one ormore markers 720. - In some embodiments, if the surgeon attempts to exceed the acceptable tolerances, the
robot 15 can be configured to provide mechanical resistance (“push back” or haptic feedback) to the movement of the end-effectuator 30 and/orsurgical instrument 35 in this manner, thereby promoting movement of the end-effectuator 30 and/orsurgical instrument 35 back to the correct, selected orientation. In some embodiments, when the surgeon then begins to correct the improper position, therobot 15 can be configured to substantially immediately return the end-effectuator 30 and/orsurgical instrument 35 back to the desired trajectory, at which time the audible and visual warnings and alerts can be configured to cease. For example, in some embodiments, the visual warning could include a visual indication 900 that may include a green light if no tolerances have been exceeded, or a red light if tolerances are about to, or have been exceeded. - As one will appreciate, a conventional worm-drive system would be absolutely rigid, and a
robot 15 having such a worm-drive system would be unable to be passively moved (without breaking the robot 15) no matter how hard the surgeon pushed. Furthermore, a completely rigid articulation system can be inherently unsafe to apatient 18. For example, if such arobot 15 were moving toward thepatient 18 and inadvertently collided with tissues, then these tissues could be damaged. Although conventional sensors can be placed on the surface of such arobot 15 to compensate for these risks, such sensors can add considerable complexity to theoverall system 1 and would be difficult to operate in a fail-safe mode. In contrast, during use of therobot 15 described herein, if the end-effectuator 30 and/orsurgical instrument 35 inadvertently collides with tissues of thepatient 18, a collision would occur with a more tolerable force that would be unlikely to damage such tissues. Additionally, in some embodiments, auditory and/or visual feedback as described above can be provided to indicate an increase in the current required to overcome the obstacle. Furthermore, in some embodiments, the end-effectuator 30 of therobot 15 can be configured to displace itself (move away) from the inadvertently contacted tissue if a threshold requiredmotor 160 current is encountered. In some embodiments, this threshold could be configured (by a control component, for example) for each axis such that the moderate forces associated with engagement between the tissue and the end-effectuator 30 can be recognized and/or avoided. - In some embodiments, the amount of rigidity associated with the positioning and orientation of the end-
effectuator 30 and/or thesurgical instrument 35 can be selectively varied. For example, in some embodiments, therobot 15 can be configured to shift between a high-rigidity mode and a low-rigidity mode. In some embodiments, therobot 15 can be programmed so that it automatically shifts to the low-rigidity mode as the end-effectuator 30 andsurgical instrument 35 are shifted from one trajectory to another, from a starting position as they approach a target trajectory and/or target position. Moreover, in some embodiment, once the end-effectuator 30 and/orsurgical instrument 35 is within a selected distance of the target trajectory and/or target position, such as, for example, within about 1° and about 1 mm of the target, therobot 15 can be configured to shift to the high-rigidity mode. In some embodiments, this mechanism may improve safety because therobot 15 would be unlikely to cause injury if it inadvertently collided with the patient 18 while in the low-rigidity mode. - Some embodiments include a
robot 15 that can be configured to effect movement of the end-effectuator 30 and/orsurgical instrument 35 in a selected sequence of distinct movements. In some embodiments, during movement of the end-effectuator 30 and/orsurgical instrument 35 from one trajectory to another trajectory, the x-axis 66, y-axis 68,roll pitch 60 orientations are all changed simultaneously, and the speed of movement of the end-effectuator 30 can be increased. Consequently, because of the range of positions through which the end-effectuator 30 travels, the likelihood of a collision with the tissue of the patient 18 can also be increased. Hence, in some embodiments, therobot 15 can be configured to effect movement of the end-effectuator 30 and/orsurgical instrument 35 such that the position of the end-effectuator 30 and/orsurgical instrument 35 within the x-axis 66 and the y-axis 68 are adjusted before theroll 62 and pitch 60 of the end-effectuator 30 and/orsurgical instrument 35 are adjusted. In some alternative embodiments, therobot 15 can be configured to effect movement of the end-effectuator 30 and/orsurgical instrument 35 so that theroll 62 andpitch 60 are shifted to 0°. The position of the end-effectuator 30 and/orsurgical instrument 35 within the x-axis 66 and the y-axis 68 are adjusted, and then theroll 62 and pitch 60 of the end-effectuator 30 and/orsurgical instrument 35 are adjusted. - Some embodiments include a
robot 15 that can be optionally configured to ensure that the end-effectuator 30 and/orsurgical instrument 35 are moved vertically along the z-axis 70 (away from the patient 18) by a selected amount before a change in the position and/or trajectory of the end-effectuator 30 and/orsurgical instrument 35 is effected. For example, in some embodiments, when an agent (for example, a surgeon or other user, or equipment) changes the trajectory of the end-effectuator 30 and/orsurgical instrument 35 from a first trajectory to a second trajectory, therobot 15 can be configured to vertically displace the end-effectuator 30 and/orsurgical instrument 35 from the body of thepatient 18 along the z-axis 70 by the selected amount (while adjusting x-axis 66 and y-axis 68 configurations to remain on the first trajectory vector, for example), and then effecting the change in position and/or orientation of the end-effectuator 30 and/orsurgical instrument 35. This ensures that the end-effectuator 30 and/orsurgical instrument 35 do not move laterally while embedded within the tissue of thepatient 18. Optionally, in some embodiments, therobot 15 can be configured to produce a warning message that seeks confirmation from the agent (for example, a surgeon or other user, or equipment) that it is safe to proceed with a change in the trajectory of the end-effectuator 30 and/orsurgical instrument 35 without first displacing the end-effectuator 30 and/orsurgical instrument 35 along the z-axis. - In some embodiments, at least one conventional force sensor (not shown) can be coupled to the end-
effectuator 30 and/orsurgical instrument 35 such that the at least one force sensor receives forces applied along the orientation axis (Z-tube axis 64) to thesurgical instrument 35. In some embodiments, the at least one force sensor can be configured to produce a digital signal. In some embodiments for example, the digital signal can be indicative of the force that is applied in the direction of the Z-tube axis 64 to thesurgical instrument 35 by the body of the patient 18 as thesurgical instrument 35 advances into the tissue of thepatient 18. In some embodiments, the at least one force sensor can be a small conventional uniaxial load cell based on a conventional strain gauge mechanism. In some embodiments, the uniaxial load cell can be coupled to, for example, analog-to-digital filtering to supply a continuous digital data stream to thesystem 1. Optionally, in some embodiments, the at least one force sensor can be configured to substantially continuously produce signals indicative of the force that is currently being applied to thesurgical instrument 35. In some embodiments, thesurgical instrument 35 can be advanced into the tissue of the patient 18 by lowering the z-axis 70 while the position of the end-effectuator 30 and/orsurgical instrument 35 along the x-axis 66 and y-axes 68 is adjusted such that alignment with the selected trajectory vector is substantially maintained. Furthermore, in some embodiments, theroll 62 and pitch 60 orientations can remain constant or self-adjust during movement of the x-(66), y-(68), and z-(70) axes such that thesurgical instrument 35 remains oriented along the selected trajectory vector. In some embodiments, the position of the end-effectuator 30 along the z-axis 70 can be locked at a selected mid-range position (spaced a selected distance from the patient 18) as thesurgical instrument 35 advances into the tissue of thepatient 18. In some embodiments, the stiffness of the end-effectuator 30 and/or thesurgical instrument 35 can be set at a selected level as further described herein. For example, in some embodiments, the stiffness of the Z-tube axis 64 position of the end-effectuator 30 and/or thesurgical instrument 35 can be coupled to a conventional mechanical lock (not shown) configured to impart desired longitudinal stiffness characteristics to the end-effectuator 30 and/orsurgical instrument 35. In some embodiments, if the end-effectuator 30 and/orsurgical instrument 35 lack sufficient longitudinal stiffness, then the counterforce applied by the tissue of the patient 18 during penetration of thesurgical instrument 35 can oppose the direction of advancement of thesurgical instrument 35 such that thesurgical instrument 35 cannot advance along the selected trajectory vector. In other words, as the z-axis 70 advances downwards, the Z-tube axis 64 can be forced up and there can be no net advancement of thesurgical instrument 35. In some embodiments, the at least one force sensor can permit an agent (for example, a surgeon or other user, or equipment) to determine, (based on sudden increase in the level of applied force monitored by the force sensor at the end-effectuator 30 and/or the surgical instrument 35), when thesurgical instrument 35 has encountered a bone or other specific structure within the body of thepatient 18. - In some alternative embodiments, the orientation angle of the end-
effectuator 30 and/orsurgical instrument 35 and the x-axis 66 and y-axis 68 can be configured to align the Z-tube axis 64 with the desired trajectory vector at a fully retracted Z-tube position, while a z-axis 70 position is set in which the distal tip of thesurgical instrument 35 is poised to enter tissue. In this configuration, in some embodiments, the end-effectuator 30 can be positioned in a manner that the end-effectuator 30 can move, for example, exactly or substantially exactly down the trajectory vector if it were advanced only alongguide tube 50. In such scenario, in some embodiments, advancing the Z-tube axis 64 can cause theguide tube 50 to enter into tissue, and an agent (a surgeon or other user, equipment, etc.) can monitor change in force from the load sensor. Advancement can continue until a sudden increase in applied force is detected at the time thesurgical instrument 35 contacts bone. - In some embodiments, the
robot 15 can be configured to deactivate the one ormore motors 160 that advance the Z-tube axis 64 such that the end-effectuator 30 and/or thesurgical instrument 35 can move freely in the Z-tube axis 64 direction while the position of the end-effectuator 30 and/or thesurgical instrument 35 continues to be monitored. In some embodiments, the surgeon can then push the end-effectuator 30 down along the Z-tube axis 64, (which coincides with the desired trajectory vector) by hand. In some embodiments, if the end-effectuator 30 position has been forced out of alignment with the trajectory vector, the position of thesurgical instrument 35 can be corrected by adjustment along the x-(66) and/or y-(68) axes and/or in theroll 62 and/or pitch 60 directions. In some embodiments, whenmotor 160 associated with the Z-tube 50 movement of thesurgical instrument 35 is deactivated, the agent (for example, a surgeon or other user, or equipment) can manually force thesurgical instrument 35 to advance until a tactile sense of thesurgical instrument 35 contacts bone, or another known region of the body). - In some further embodiments, the robotic
surgical system 1 can comprise a plurality ofconventional tracking markers 720 configured to track the movement of therobot arm 23, the end-effectuator 30, and/or thesurgical instrument 35 in three dimensions. It should be appreciated that three dimensional positional information from trackingmarkers 720 can be used in conjunction with the one dimensional linear positional information from absolute or relative conventional linear encoders on each axis of therobot 15 to maintain a high degree of accuracy. In some embodiments, the plurality of trackingmarkers 720 can be mounted (or otherwise secured) thereon an outer surface of therobot 15, such as, for example and without limitation, on thebase 25 of therobot 15, or therobot arm 23. In some embodiments, the plurality of trackingmarkers 720 can be configured to track the movement of therobot 15 arm, the end-effectuator 30, and/or thesurgical instrument 35. In some embodiments, thecomputer 100 can utilize the tracking information to calculate the orientation and coordinates of thedistal tip 30 a of thesurgical instrument 35 based on encoder counts along the x-axis 66, y-axis 68, z-axis 70, the Z-tube axis 64, and theroll 62 and pitch 60 axes. Further, in some embodiments, the plurality of trackingmarkers 720 can be positioned on thebase 25 of therobot 15 spaced from thesurgical field 17 to reduce the likelihood of being obscured by the surgeon, surgical tools, or other parts of therobot 15. In some embodiments, at least onetracking marker 720 of the plurality of trackingmarkers 720 can be mounted or otherwise secured to the end-effectuator 30. In some embodiments, the positioning of one ormore tracking markers 720 on the end-effectuator 30 can maximize the accuracy of the positional measurements by serving to check or verify the end-effectuator 30 position (calculated from the positional information from the markers on thebase 25 of therobot 15 and the encoder counts of the x-(66), y-(68),roll 62,pitch 60, and Z-tube axes 64). - In some further embodiments, at least one optical marker of the plurality of
optical tracking markers 720 can be positioned on therobot 15 between the base 25 of therobot 15 and the end-effectuator 30 instead of, or in addition to, themarkers 720 on thebase 25 of therobot 15, (seeFIG. 16 ). In some embodiments, the at least onetracking marker 720 can be mounted to a portion of therobot 15 that effects movement of the end-effectuator 30 and/orsurgical instrument 35 along the x-axis to enable thetracking marker 720 to move along the x-axis 66 as the end-effectuator 30 andsurgical instrument 35 move along the x-axis 66. The placement of the trackingmarkers 720 in this way can reduce the likelihood of a surgeon blocking thetracking marker 720 from the cameras or detection device, or thetracking marker 720 becoming an obstruction to surgery. In certain embodiments, because of the high accuracy in calculating the orientation and position of the end-effectuator 30 based on thetracking marker 720 outputs and/or encoder counts from each axis, it can be possible to very accurately determine the position of the end-effectuator 30. For example, in some embodiments, without requiring knowledge of the counts of axis encoders for the z-axis 70, which is between the x-axis 66 and thebase 25, knowing only the position of themarkers 720 on the x-axis 66 and the counts of encoders on the y-(68),roll 62,pitch 60, and Z-tube axes 64 can enable computation of the position of the end-effectuator 30. In some embodiments, the placement ofmarkers 720 on any intermediate axis of therobot 15 can permit the exact position of the end-effectuator 30 to be calculated based on location ofsuch markers 720 and counts of encoders on axes (66, 62, 60, 64) between themarkers 720 and the end-effectuator 30. In some embodiments, from the configuration of the robot 15 (see for example,FIG. 2 ), the order of axes from the base 25 to the end-effectuator 30 is z-(70) then x-(66) then y-(68) then roll 62 then pitch 60 then Z-tube 64. Therefore, for example, within embodiments in whichtracking markers 720 are placed on thehousing 27 of therobot 15 that moves with theroll 62 axis, the locations ofsuch tracking markers 720 and the encoder counts of thepitch 60 and Z-tube axes 64 can be sufficient to calculate the end-effectuator 30 position. - In some embodiments, when the
surgical instrument 35 is advanced into the tissue of the patient 18 with the assistance of aguide tube 50, thesurgical instrument 35 can comprise a stop mechanism 52 that is configured to prevent thesurgical instrument 35 from advancing when it reaches a predetermined amount of protrusion (see for example,FIGS. 17A-B ). In some embodiments, by knowing the lengths of theguide tube 50 and thesurgical instrument 35, the distance between the respective ends of thesurgical instrument 35, and the location where the stop mechanism 52 is attached, it is possible to determine the maximum distance past the end of theguide tube 50 that thesurgical instrument 35 can protrude. - In some embodiments, it can be desirable to monitor not just the maximum protrusion distance of the
surgical instrument 35, but also the actual protrusion distance at any instant during the insertion process. Therefore, in some embodiments, therobot 15 can substantially continuously monitor the protrusion distance, and in some embodiments, the distance can be displayed on a display (such as display means 29). In some embodiments, protrusion distance can be substantially continuously monitored using a spring-loadedplunger 54 including a spring-loaded mechanism 55 a and sensor pad 55 b that has a coupled wiper 56 (see for exampleFIG. 17B ). In some embodiments, the stop mechanism 52 on thesurgical instrument 35 can be configured to contact the spring-loaded mechanism 55 well before it encounters the end of theguide tube 50. In some embodiments, when thewiper 56 moves across the position sensor pad 55 b, its linear position is sampled, thereby permitting calculation of the distance by which thesurgical instrument 35 protrudes past the end of theguide tube 50 substantially in real-time. In some embodiments, any conventional linear encoding mechanism can be used to monitor the plunger's depth of depression and transmit that information to thecomputer 100 as further described herein. - Some embodiments include instruments that enable the stop on a
drill bit 42 to be manually adjusted with reference to markings 44 on thedrill bit 42. For example,FIGS. 17C-17E depict tools for manually adjusting a drill stop 46 with reference to drill bit markings 44 in accordance with one embodiment of the invention. As shown, in some embodiments, thedrill bit 42 can include release mechanisms 48 on each end of the drill stop 46. In some embodiments, if the release 48 on one end of the drill stop 46 is pulled, it is possible to move the drill stop 46 up the shaft of thedrill bit 42. In some embodiments, if the release 48 on the other end of the drill stop 46 is pulled, it is possible to move the drill stop 46 down the shaft (see the direction of movement inFIGS. 17D and 17E ). In some embodiments, if neither release mechanism 48 is pulled, the drill stop 46 will not move in either direction, even if bumped. - Some embodiments include the ability to lock and hold the
drill bit 42 in a set position relative to thetube 50 in which it is housed. For example, in some embodiments, thedrill bit 42 can be locked by locking the drill stop 46 relative to thetube 50 using a locking mechanism.FIGS. 17F-J illustrates tools for locking and holding adrill bit 42 in a set position in accordance with one embodiment of the invention. In some embodiments, thelocking mechanism 49 shown in FIG. 17H can comprise two clam shells 49 (shown inFIG. 17F ). In some embodiments, adrill bit 42 can be locked into position by assembling the clam shells around the drill stop 46 (shown inFIG. 17G ). This feature allows the user to lock thedrill bit 42 in a position such that the tip slightly protrudes past the end of the tube 50 (seeFIGS. 171 and 17J ). In this position, the user can force thetube 50 to penetrate through soft tissues to force thetube 50 to contact bone (for example during a percutaneous spine screw insertion). - In some further embodiments, the end-
effectuator 30 can be configured not block the trackingoptical markers 720 or interfere with the surgeon. For example, in some embodiments, the end-effectuator 30 can comprise a clearance mechanism 33 including an actuator 33 a that permits this configuration, as depicted inFIGS. 18A and 18B . As shown, theguide tube 50 can be secured within a housing of the end-effectuator 30 with twoshafts 32. In some embodiments, theshafts 32 move relative to one other, due to a parallelogram effect of the clearance mechanism 33, the position of theguide tube 50 can mimic the position of the end-effectuator 30 (seeFIG. 18B ). - In applications such as cervical or lumbar fusion surgery, it can be beneficial to apply distraction or compression across one or more levels of the spine (anteriorly or posteriorly) before locking hardware in place. In some embodiments, the end-
effectuator 30 can comprise an attachment element 37 that is configured to apply such forces (see for exampleFIGS. 19A-B ). In some embodiments, the end-effectuator 30 attachment element 37 can be configured for coupling to the end-effectuator 30 at substantially the same location as the clearance mechanism 33. In some embodiments, the end-effectuator 30 with attachment element 37 snaps into the same place as the end-effectuator 30 without the attachment element 37. In some embodiments, during use of the end-effectuator 30 attachment element 37, the relative movement of the twoshafts 32 caused by angulation 30 e will not cause movement in thepitch 60 direction and will instead cause distraction (illustrated as moving from an attachment element 37 distance 37 a inFIG. 19A to distance 37 b inFIG. 19B ). Further, althoughshaft 32 movement as shown inFIGS. 19A-B would cause distraction, rotation of the actuator 33 a in the opposite direction to that represented by 30 e would cause compression (i.e. the distance 37 b inFIG. 19B would move towards the distance 37 a inFIG. 19A ). - In view of the embodiments described hereinbefore, some embodiments that can be implemented in accordance with the disclosed subject matter can be better appreciated with reference to the flowcharts in
FIGS. 24-33 . For purposes of simplicity of explanation, the method disclosed by the embodiments described herein is presented and described as a series of steps; however, it is to be understood and appreciated that the claimed subject matter is not limited by the order of acts, as some acts may occur in different orders and/or concurrently with other acts from that shown and described herein. For example, the various methods or processes of some embodiments of the invention can alternatively be represented as a series of interrelated states or events, such as in a state diagram. Furthermore, not all illustrated acts may be required to implement a method in accordance with some embodiments of the invention. Further yet, two or more of the disclosed methods or processes can be implemented in combination with each other, to accomplish one or more features or advantages herein described. - It should be further appreciated that the methods disclosed in the various embodiments described throughout the subject specification can be stored on an article of manufacture, or computer-readable medium, to facilitate transporting and transferring such methods to a computing device (e.g., a desktop computer, a mobile computer, a mobile telephone, a blade computer, a programmable logic controller, and the like) for execution, and thus implementation, by a processor of the computing device or for storage in a memory thereof.
- In some embodiments, the
surgical robot 15 can adjust its position automatically continuously or substantially continuously in order to move the end-effectuator 30 to an intended (i.e. planned) position. For example, in some embodiments, thesurgical robot 15 can adjust its position automatically continuously or substantially continuously based on the current position of the end-effectuator 30 and surgical target as provided by a current snapshot of tracking markers, LPS, or other tracking data. It should further be appreciated that certain position adjustment strategies can be inefficient. For example, an inefficient strategy for therobot 15 to find a target location can be an iterative algorithm to estimate the necessary direction of movement, move toward the target location, and then assess a mismatch between a current location and the target location (the mismatch referred to as an error), and estimate a new direction, repeating the cycle of estimate-movement-assessment until the target location is reached within a satisfactory error. Conversely, the position adjustment strategies in accordance with some embodiments of the invention are substantively more efficient than iterative strategies. For example, in some embodiments, asurgical robot 15 can make movements and adjust its location by calibrating the relative directions of motions in each axis (permitting computation via execution of software or firmware with the computer 100) at each frame of tracking data, of a unique set of necessary motor encoder counts that can cause each of the individual axes to move to the correct location. In some embodiments, the Cartesian design of the disclosedrobot 15 can permit such a calibration to be made by establishing a coordinate system for therobot 15 and determining key axes of rotation. - As described in greater detail below, in some embodiments, methods for calibrating the relative directions of the robot's 15 axes can utilize a sequence of carefully planned movements, each in a single axis. In some embodiments, during these moves,
temporary tracking markers 720 are attached to the end-effectuator 30 to capture the motion of the end-effectuator 30. It should be appreciated that the disclosed methods do not require the axes of therobot 15 to be exactly or substantially perpendicular, nor do they require the vector along which a particular axis moves (such as the x-axis 66) to coincide with the vector about which rotation occurs (such aspitch 60, which occurs primarily about the x-axis 66). In certain embodiments, the disclosed methods include motion along aspecific robot 15 axis that occurs in a straight line. In some embodiments, the disclosed methods for calibrating the relative directions of movement of the robot's 15 axes can utilize one or more frames of tracking data captured at the ends of individual moves made in x-(66), y-(68), roll (62), pitch (60), and Z-tube axes 64 frommarkers 720 temporarily attached to the end-effectuator's 30guide tube 50. In some embodiments, when moving individual axes, all other axes can be configured at the zero position (for example, the position where the encoder for the axis reads 0 counts). Additionally or alternatively, one or more frames of tracking data with allrobot 15 axes at 0 counts (neutral position) may be necessary, and one or more frames of data with thetemporary markers 720 rotated to a different position about the longitudinal axis of theguide tube 50 may be necessary. In some embodiments, themarker 720 positions from these moves can be used to establish a Cartesian coordinate system for therobot 15 in which the origin (0,0,0) is through the center of the end-effectuator 30 and is at the location along the end-effectuator 30 closest to wherepitch 60 occurs. Additionally or alternatively, in some embodiments, this coordinate system can be rotated to an alignment in which y-axis 68 movement of therobot 15 can occur exactly or substantially along the coordinate system's y-axis 68, while x-axis 66 movement of therobot 15 occurs substantially perpendicular to the y-axis 68, but by construction of the coordinate system, without resulting in any change in the z-axis 70 coordinate. In certain embodiments, the steps for establishing the robot's 15 coordinate system based at least on the foregoing individual moves can comprise the following: First, from the initial and final positions of the manual rotation of trackingmarkers 720 about the long axis of the end-effectuator 30, a finite helical axis of motion is calculated, which can be represented by a vector that is centered in and aligned with the end-effectuator 30. It should be appreciated that methods for calculating a finite helical axis of motion from two positions of three or more markers are described in the literature, for example, by Spoor and Veldpaus (Spoor, C. W. and F. E. Veldpaus, “Rigid body motion calculated from spatial co-ordinates of markers,” J Biomech 13(4): 391-393 (1980)). In some embodiments, rather than calculating the helical axis, the vector that is centered in and aligned with the end-effectuator 30 can be defined, or constructed, by interconnecting two points that are attached to two separate rigid bodies that can be temporarily affixed to the entry and exit of theguide tube 50 on the Z-tube axis 64. In this instance, each of the two rigid bodies can include at least one tracking marker 720 (e.g., onetracking marker 720, two trackingmarkers 720, three trackingmarkers 720, more than three trackingmarkers 720, etc.), and a calibration can be performed that provides information indicative of the locations on the rigid bodies that are adjacent to the entry and exit of theguide tube 50 relative to the tracking markers. - A second helical axis can be calculated from the
pitch 60 movements, providing a vector substantially parallel to the x-axis of therobot 15 but also close to perpendicular with the first helical axis calculated. In some embodiments, the closest point on the first helical axis to the second helical axis (or vector aligned with the end-effectuator 30) is calculated using simple geometry and used to define the origin of the robot's coordinate system (0,0,0). A third helical axis is calculated from the two positions of theroll 62 axis. In certain scenarios, it cannot be assumed that the vector about which roll occurs (third helical axis) and the vector along which the y-axis 68 moves are exactly or substantially parallel. Moreover, it cannot be assumed that the vector about which pitch 60 occurs and the vector along which x-axis 66 motion occurs are exactly or substantially parallel. Vectors for x-axis 66 and y-axis 68 motion can be determined from neutral and extended positions of x-axis 66 and y-axis 68 and stored separately. As described herein, in some embodiments, the coordinate system can be realigned to enable y-axis movement of therobot 15 to occur exactly or substantially in the y-axis 68 direction of the coordinate system, and x-axis 66 movement of therobot 15 without any change in the z-coordinate (70). In general, to perform such a transformation of coordinate systems, a series of rotations about a coordinate axis is performed and applied to every point of interest in the current coordinate system. Each point is then considered to be represented in the new coordinate system. In some embodiments, to apply a rotation of a point represented by a 3×1 vector about a particular axis, the vector can be pre-multiplied by a 3×3 rotation matrix. The 3×3 rotation matrix for a rotation of Rx degrees about the x-axis is: -
- The 3×3 rotation matrix for a rotation of Ry degrees about the y-axis is:
-
- The 3×3 rotation matrix for a rotation of Rz degrees about the z-axis is:
-
- In some embodiments, to transform coordinate systems, a series of three rotations can be performed. For example, such rotations can be applied to all vectors and points of interest in the current coordinate system, including the x-movement vector, y-movement vector and each of the helical axe, to align the y movement vector with the new coordinate system's y-axis, and to align the x movement vector as closely as possible to the new coordinate system's x-axis at z=0. It should be appreciated that more than one possible sequence of three rotations can be performed to achieve substantially the same goal. For example, in some embodiments, a sequence of three rotations can comprise (1) a rotation about x using an Rx value appropriate to rotate the y-movement vector until its z coordinate equal 0, followed by (2) a rotation about z using an Rz value appropriate to rotate the y-movement vector until its x coordinate equal 0, followed by (3) a rotation about y using an Ry value appropriate to rotate the x-movement vector until its z coordinate equals 0. In some embodiments, to find the rotation angle appropriate to achieve a given rotation, the arctangent function can be utilized. For example, in some embodiments, the angle needed to rotate a point or vector (x1,y1,z1) about the z axis to y1=0 is −arctan(y1/x1).
- It should be appreciated that after transformation of the coordinate system, in some embodiments, although the new coordinate system is aligned such that the y-movement axis of the
surgical robot 15 is exactly or substantially exactly aligned with the coordinate system's y-axis 68, theroll 62 rotation movement of therobot 15 should not be assumed to occur exactly or substantially exactly about a vector aligned with the coordinate system's y-axis 68. Similarly, in some embodiments, thepitch 60 movement of thesurgical robot 15 should not be assumed to occur exactly or substantially exactly about a vector aligned with the coordinate system's x-axis. In some embodiments, inroll 62 and pitch 60 rotational movement there can be linear and orientational “offsets” from the helical axis of motion to the nearest coordinate axis. In some embodiments, from the helical axes determined above using tracked markers, such offsets can be calculated and retained (e.g., stored in a computing device's memory) so that for any rotation occurring during operation, the offsets can be applied, rotation can be performed, and then negative offsets can be applied so that positional change occurring with rotation motion accounts for the true center of rotation. - In some embodiments, during tracking, the desired trajectory can be first calculated in the medical image coordinate system, then transformed to the
robot 15 coordinate system based at least on known relative locations of active markers. For example, in some embodiments, conventional light-emitting markers and/or conventional reflective markers associated with anoptical tracking system 3417 can be used (see for exampleactive markers 720 inFIG. 20A ). In other embodiments, conventional electromagnetic sensors associated with an electromagnetic tracking system can be used. In some other embodiments, radio-opaque markers (forexample markers 730 shown inFIG. 20A ) can be used with a CT imaging system. In some embodiments, radio-opaque markers 730 (spheres formed, at least in part from metal or other dense material), can be used to provide amarker 730 that can at least partially absorb x-rays to produce a highly contrasted image of the sphere in a CT scan image. - In some embodiments, the necessary counts for the end-
effectuator 30 to reach the desired position in the robot's 15 coordinate system can be calculated based on the following example process. First the necessary counts to reach the desired angular orientation can be calculated. In some embodiments, a series of three rotations can be applied to shift the coordinate system temporarily to a new coordinate system in which the y-axis 68 coincides or substantially coincides with the helical axis of motion forroll 62, and the x-axis 66 is largely aligned with the helical axis of motion forpitch 60 and by definition, and the helical axis of motion forpitch 60 has constant z=0. Then, the number of counts necessary to achieve the desiredpitch 60 can be determined, keeping track of how thispitch 60 can affectroll 62. In one implementation, to find the necessary counts to achieve the desired pitch, the change inpitch angle 60 can be multiplied by the previously calibrated motor counts per degree for pitch. The change inroll 62 caused by this change inpitch 60 can be calculated from the orientation of the helical axis and the rotation angle (pitch) about the helical axis. Then, thenecessary roll 62 to get to the desiredroll 62 to reach the planned trajectory alignment can be calculated, with the benefit that applyingroll 62 does not, by definition of the coordinate system, result in any further change in pitch. The coordinate system is then shifted back to the previously describedrobot 15 coordinate system by the inverse of the three rotations applied above. Then the necessary counts to reach the desired x-axis 66 position can be calculated, also keeping track of how this x-axis 66 position change will affect y-axis 68 position. Then the necessary y-axis 68 counts to reach the desired y-axis position can be readily calculated with the benefit that changing the y-axis 68 coordinate can have no effect on any other axis since the y-axis motion vector is by definition aligned with the robot's y-axis 68. In a scenario in which the Z-tube 50 position is being actively controlled, the orientation of the Z-tube 50 movement vector is adjusted when adjustingroll 62 andpitch 60 and the counts necessary to move it to the desired position along the trajectory vector is calculated from the offset. In some embodiments, after the necessary counts to achieve the desired positions in all axes are calculated as described, these counts can be sent as computer-accessible instructions (e.g., computer-readable and/or computer-executable instructions) to respective controllers for each axis in order to move the axes to the computed positions. -
FIG. 24 is a flowchart of a method 2400 for positioning and advancing through soft tissue in accordance with one or more aspects according to one embodiment of the invention. As shown, in some embodiments, at block 2410, a medical image is accessed (e.g., received, retrieved, or otherwise acquired). As described herein, the medical image can be a 3D anatomical image scan including, but not limited to a CT scan, a magnetic resonance imaging scan (hereinafter referred to as an “MRI scan”), an X-ray image, or other anatomical scan. It should be appreciated that any 3D anatomical scan may be utilized with thesurgical robot 15 and is within the scope of the present invention. In some embodiments, at block 2420, a targetingfixture 690 is calibrated to the medical image. In some embodiments, the calibration can be semi-automated or automated. In some embodiments, at block 2430, data indicative of an intended trajectory associated with the medical image is received. In some embodiments, at block 2440, arobot 15 is substantially maintained on the intended trajectory. In some embodiments, a control platform (for example,platform 3400 shown inFIG. 34 ) can adjust movement of therobot 15 in order to substantially maintain the intended trajectory. -
FIGS. 25-26 are flowcharts of methods for calibrating a targetingfixture 690 to a medical image in accordance with one or more embodiments of the invention. As shown inFIG. 25 , in some embodiments, themethod 2500 can embody a semi-automated calibration method and can be implemented (e.g., executed) as part of block 2420 in certain scenarios. In some embodiments, atblock 2510, data indicative of a medical image having a representation of a plurality of radio-opaque markers (for example radio-opaque markers 730) is received. In one embodiment, as described herein, such plurality can contain four radio-opaque markers 730. In some embodiments, atblock 2520, a geometrical center for each radio-opaque marker 730 is determined in a coordinate system associated with the medical image. In some embodiments, image thresholding can be utilized to define one or more edges of each radio-opaque marker 730 and a geometrical center thereof. Thresholding refers to an image processing technique in which pixel intensity within a 2D region can be monitored. For example, the x, y positions (for instance expressed in mm) of pixels of an intensity that reach a predetermined value can be retrieved. Stated similarly, the threshold refers to the transition pixel intensity from light to dark. In some embodiments, on 2D slices of the medical image, the radio-opaque marker 730 can appear light and the adjacent space (such as tissue or air) can appear dark. In some embodiments, displaying pixels that satisfy a thresholding criterion at an intensity encountered at the edge of a radio-opaque marker can yield a largely circular trace outlining the marker on the medical image. Since in some embodiments,markers 730 can be spherical, a method for finding the center of themarker 730 in a 2D view can include firstly restricting the 2D view to a sampling region with the high-intensity image of the sphere toward the center of the region and pixels of lower intensity toward the outer edges of the region. Secondly, the method can include finding the mean x threshold position (e.g., the maximum x coordinate of pixels satisfying the threshold criterion plus minimum x coordinate of pixels satisfying the threshold criterion divided by two), and finding the mean y threshold position using a similar method. In some embodiments, the center of the sphere can be found by determining 2D centers of slices through thesame marker 730 in two orthogonal views. For example, in some embodiments, the method can include finding mean x and mean y from an xy slice, then finding mean x and mean z from an xz slice to get a mean x, y, and z axis coordinate representing the center of themarker 730. Further, upon or after the mean x, mean y, and mean z are found, new xy and xz slices can be evaluated again and the maximum and minimum x, y, and z threshold values can be again determined to evaluate the dimensions of the thresholded object in each view. It can be appreciated from this method that in some embodiments, a non-spherical object of high intensity, such as a small process of cortical bone extending away from the side of the spine, may fail to satisfy (1) a condition where there is high intensity near the middle of the region, but low intensity all around, since the process may extend out of the region in one or more directions; or (2) a condition where the dimensions in x, y, and z of the centered object do not match each other (e.g., non-spherical case). - As shown in
FIG. 25 , in some embodiments, atblock 2530, it is ascertained if one centered sphere is determined for each radio-opaque marker 730 for the fixture being calibrated. In some embodiments, when at least one such sphere is not determined, or identified, the threshold setting is adjusted and flow is directed to block 2510. In some embodiments, atblock 2540, each centered sphere is mapped to each radio-opaque marker 730 of the plurality of radio-opaque markers 730. As shown, in some embodiments, block 2540 can represent a mapping action which, in some embodiments, can comprise implementing a sorting process to establish a specific centered sphere is associated with a specific one of the plurality of radio-opaque markers 730. In some embodiments, a plurality of radio-opaque markers 730 contains four radio-opaque markers 730 (represented, for example, as OP1, OP2, OP3, and OP4). In some embodiments, the sorting process can map each one of four centeredmarkers 730 to one of OP1, OP2, OP3, or OP4. In some embodiments, the sorting process can distinguish aspecific marker 730 by measuring inter-marker distances from mean positions of the fourunidentified markers 730, and comparing such distances to extant inter-marker distances (for example, those that are pre-measured and retained in memory, such as mass storage device 3404) for eachmarker 730 on a marker fixture. In some embodiments, theopaque markers 730 on thefixture 690 can be placed asymmetrically, eachmarker 730 can be identified from a unique set of inter-marker distances corresponding tosuch marker 730. For example, in some embodiments where the sum of inter-marker distances of oneunknown marker 730 relative to theother threes markers 730 measured from the medical image is D, a single physical marker 730 (one of OP1, OP2, OP3, or OP4) can have a matching inter-marker distance sum within a specified tolerance (such as ±1 mm) of D. In some embodiments, at block 2550, coordinates of each centered sphere can be retained (for example in memory of a computer platform 3400). As described herein, in some embodiments, such coordinates can be utilized in a process for tracking movement of arobot 15. - Some embodiments include method 2600 (shown as a flowchart in
FIG. 26 ) that can embody an automated calibration method and can be implemented (e.g., executed) as part of block 2420 in certain scenarios. In some embodiments, atblock 2605, for a Z position, an x-y grid of test area squares is created. In some embodiments, each test area square can be larger than the diameter of a sphere (a radio-opaque marker 730) associated with a targetingfixture 690 comprised of material that, when imaged, appears as opaque. In some embodiments, each test area square can be at least partially overlapping with at least one adjacent test area square. In one embodiment of the invention, a nearly half the surface of a test area square can overlap with the surface of an adjacent test area square. In some embodiments, atblock 2610, calibration is initiated at place Z=0,x-y grid row 0,x-y grid column 0. In some embodiments, atblock 2615, borders of a medical image within a first test area square are determined. It should be appreciated that in some embodiments, a sphere can be rendered as a circular area, but a section of bone represented in the medical image can be asymmetrical. In some embodiments, a thresholding process in accordance with one or more aspects described herein can be implemented to exclude one or moreinvalid markers 730 by assessing if the x, y, and z axes boundaries of the object are of substantially equivalent dimensions, consistent with the shape being spherical. - In some embodiments, at
block 2620, it is determined if a maximum (max) border coordinate is less than the maximum coordinate of the test area, and a minimum (min) border coordinate is greater than the minimum coordinate of the test area, and vertical span of features rendered in the image are equal or substantially equal to horizontal span of such features. As shown inFIG. 26 , in some embodiments, in the negative case, flow is directed to block 2645, at which the first test area is moved to next grid location and next Z plane. Conversely, in case the three foregoing conditions are fulfilled, flow is directed to block 2625, at which X coordinate and Y coordinate are centered at the center of the current test area. In some embodiments, atblock 2630, Z coordinate is probed by creating a second test area square spanning upwards and downwards in XZ plane and/or YZ plane to determine one or more borders of an object. In some embodiments, atblock 2635, it is determined if borders of the object observed in XZ plane are of substantially equivalent relative spacing (vertically and horizontally) to borders in the x-y plane, consistent with the shape of the object being spherical. In some embodiments, when such borders are of different spacing, flow is directed to block 2645. Conversely, when spacing of such borders is substantially equivalent between views, a sphere having a center at X coordinate, Y coordinate, and Z coordinate is identified atblock 2640 and flow is directed to block 2645. - In some embodiments, at
block 2650, it is determined if last row and column in x-y grid are reached and last Z plane is reached as a result of updating the first test area atblock 2645. In some embodiments, in the negative case, flow is directed to block 2615, in which the first area is the updated instance of a prior first area, with the flow reiterating one or more ofblocks 2620 through 2645. Conversely, in the affirmative case, flow is directed to block 2655 at which invalid marker(s) 730 can be excluded. In some embodiments, a paring process can be implemented to exclude one or moreinvalid markers 730. For this paring process, in some embodiments, the known spacings between each of the N radio-opaque markers 730 (with N a natural number) on the targetingfixture 690 and each other radio-opaque marker 730 on the targetingfixture 690 can be compared to themarkers 730 that have been found on the medical image. In a scenario in which more than N number ofmarkers 730 can be found on the medical image, any sphere found on the medical image that does not have spacings relative to N−1other markers 730 that are within an acceptable tolerance of known spacings retained, for example, on a list can be considered to be invalid. For example, if a targetingfixture 690 has four radio-opaque markers 730, there are six known spacings, with eachmarker 730 having a quantifiable spacing relative to three other markers 730: the inter-marker spacings for markers 1-2, 1-3, 1-4, 2-3, 2-4, and 3-4. On the 3D medical image of the targetingfixture 690, in some embodiments, if fivepotential markers 730 are found on the medical image, their inter-marker spacings can be calculated. In this scenario, there are 10 inter-marker spacings: 1-2, 1-3, 1-4, 1-5, 2-3, 2-4, 2-5, 3-4, 3-5, and 4-5, with each sphere having a quantifiable spacing relative to fourother markers 730. Considering each of the fivepotential markers 730 individually, if any one of such fivemarkers 730 does not have three of its four inter-marker spacings within a very small distance of the spacings on the list of six previously quantified known spacings, it is considered invalid. - In some embodiments, at
block 2660, each centered radio-opaque marker 730, identified atblock 2640, can be mapped to each radio-opaque marker 730 of a plurality of radio-opaque markers 730. In some embodiments, a sorting process in accordance with one or more aspects described herein can be implemented to mapsuch markers 730 to radioopaque markers 730. In some embodiments, atblock 2665, coordinates of each centered sphere can be retained (e.g., in memory of a computer platform 3400). As described herein, in some embodiments, such coordinates can be utilized in a process for tracking movement of arobot 15. In some embodiments, during tracking, the established (e.g., calibrated) spatial relationship betweenactive markers 720 and radio-opaque markers 730 can be utilized to transform the coordinate system from the coordinate system of the medical image to the coordinate system of thetracking system 3417, or vice versa. Some embodiments include a process for transforming coordinates from the medical image's coordinate system to the tracking system's coordinate system can include afixture 690 comprising four radio-opaque markers OP1, OP2, OP3, and OP4 (for example radio-opaque markers 730) in a rigidly fixed position relative to four active markers AM1, AM2, AM3, AM4 (for example, active markers 720). In some embodiments, at the time the calibration of thefixture 690 occurred, this positional relationship can be retained in a computer memory (e.g., system memory 3412) for later access on real-time or substantially on real-time in a set of four arbitrary reference Cartesian coordinate systems that can be readily reachable through transformations at any later frame of data. In some embodiments, each reference coordinate system can utilize an unambiguous positioning of three of theactive markers 720. Some embodiments can include a reference coordinate system for AM1, AM2, and AM3 can be coordinate system in which AM1 can be positioned at the origin (e.g., the three-dimensional vector (0,0,0)); AM2 can be positioned on the x-axis (e.g., x-coordinate AM2x>0, y-coordinate AM2y=0, and z-coordinate AM2z=0); and AM3 can be positioned on the x-y plane (e.g., x-coordinate AM3x unrestricted, y-coordinate AM3y>0, and z-coordinated AM3z=0). Some embodiments include a method to generate a transformation to such coordinate system can comprise (1) translation of AM1, AM2, AM3, OP1, OP2, OP3, and OP4 in a manner that AM1 vector position is (0,0,0); (2) rotation about the x-axis by an angle suitable to position AM2 at z=0 (e.g., rotation applied to AM2, AM3 and OP1-OP4); (3) rotation about the z-axis by an angle suitable to position AM2 at y=0 and x>0 (e.g., rotation applied to AM2, AM3 and OP1-OP4); (4) rotation about the x-axis by an angle suitable to position AM3 at z=0 and y>0 (e.g., rotation applied to AM3 and OP1-OP4). It should be appreciated that, in some embodiments, it is unnecessary to retain these transformations in computer memory, for example; rather, the information retained for later access can be the coordinates of AM1-AM3 and OP1-OP4 in such reference coordinate system. In some embodiments, another such reference coordinate system can transform OP1-OP4 by utilizing AM2, AM3, and AM4. In some embodiments, another such reference coordinate system can transform OP1-OP4 by utilizing AM1, AM3, and AM4. In some further embodiments, another such reference coordinate system can transform OP1-OP4 by utilizing AM1, AM2, and AM4. - In some embodiments, at the time of tracking, during any given frame of data, the coordinates of the active markers AM1-AM4 can be provided by the
tracking system 3417. In some embodiments, by utilizing markers AM1, AM2, and AM3, transformations suitable to reach the conditions of the reference coordinate system can be applied. In some embodiments, such transformations can position AM1, AM2, and AM3 on the x-y plane in a position in proximity to the position that was earlier stored in computer memory for this reference coordinate system. In some embodiments, for example, to achieve a best fit of the triad ofactive markers 720 on their stored location, a least squares algorithm can be utilized to apply an offset and rotation to the triad ofmarkers 720. In one implementation, the least squares algorithm can be implemented as described by Sneath (Sneath P. H. A., Trend-surface analysis of transformation grids, J. Zoology 151, 65-122 (1967)). In some embodiments, transformations suitable to reach the reference coordinate system, including the least squares adjustment, can be retained in memory (e.g.,system memory 3412 and/or mass storage device 3404). In some embodiments, the retained coordinates of OP1-OP4 in such reference coordinate system can be retrieved and the inverse of the retained transformations to reach the reference coordinate system can be applied to such coordinates. It should be appreciated that the new coordinates of OP1-OP4 (the coordinates resulting from application of the inverse of the transformations) are in the coordinate system of thetracking system 3417. Similarly, in some embodiments, by utilizing the remaining three triads ofactive markers 720, the coordinates of OP1-OP4 can be retrieved. - In some embodiments, the four sets of OP1-OP4 coordinates in the tracking system's coordinate system that can be calculated from different triads of
active markers 720 are contemplated to have coordinates that are approximately equivalent. In some embodiments, when coordinates are not equivalent, the data set can be analyzed to determine which of theactive markers 720 provides non-suitable (or poor) data by assessing how accurately each triad ofactive markers 720 at the current frame overlays onto the retained positions ofactive markers 720. In some other embodiments, when the coordinates are nearly equivalent, a mean value obtained from the four sets can be utilized for each radio-opaque marker 730. In some embodiments, to transform coordinates of other data (such as trajectories from the medical image coordinate system) to the tracking system's coordinate system, the same transformations can be applied to the data. For example, in some embodiments, the tip and tail of a trajectory vector can be transformed to the four reference coordinate systems and then retrieved with triads ofactive markers 720 at any frame of data and transformed to the tracking system's coordinate system. -
FIG. 27 is a flowchart of amethod 2700 for automatically maintaining asurgical robot 15 substantially on a trajectory in accordance some embodiments of the invention. In some embodiments, at block 2710, data indicative of position of one or more of Z-frame 72 or Z-tube 50 are received. In some embodiments, atblock 2720, data indicative of eachrobot 15 joint in thesurgical robot 15, such as encoder counts from eachaxis motor 160, are accessed. In some embodiments, atblock 2730, acurrent robot 15 position on a planned trajectory is accessed (the position being represented in a camera coordinate system or the coordinate system of other tracking device). In one embodiment, the planned trajectory can be generated by an operator. For example, the operator (e.g., a surgeon) can scroll and rotate through the image slices until the desired anatomy can be viewed on three windows representing three orthogonal planes (typically sagittal, coronal, and axial slices). The operator can then draw a line at the desired slope and location on one window; the line simultaneously is calculated and appears on the other two windows, constrained by the views of the screens and orientation on the window on which it was drawn. - In another embodiment, a line (e.g., referred to as line t) that is fixed on the image both in angle and position represents the desired trajectory; the surgeon has to rotate and scroll the images to align this trajectory to the desired location and orientation on the anatomy. At least one advantage of such embodiment is that it can provide a more complete, holistic picture of the anatomy in relationship to the desired trajectory that may not require the operator to erase and start over or nudge the line after it is drawn, and this process was therefore adopted. In some embodiments, a planned trajectory can be retained in a memory of a computing device (for example, computing device 3401) that controls the
surgical robot 15 or is coupled thereto for use during a specific procedure. In some embodiments, each planned trajectory can be associated with a descriptor that can be retained in memory with the planned trajectory. As an example, the descriptor can be the level and side of the spine where screw insertion is planned. - In another embodiment, the line t that is (fixed on the image both in angle and position representing the desired trajectory) is dictated by the current position of the robot's
end effectuator 30, or by an extrapolation of the endeffectuator guide tube 50 if aninstrument 35 were to extend from it along the same vector. In some embodiments, as therobot 15 is driven manually out over the patient 18 by activatingmotors 160 controlling individual or combined axes 64, 66, 68, 70, the position of this extrapolated line (robot's end effectuator 30) is updated on the medical image, based onmarkers 720 attached to the robot, conventional encoders showing current position of an axis, or a combination of these registers. In some embodiments, when the desired trajectory is reached, that vector's position in the medical image coordinate system is stored into the computer memory (for example in memory of a computer platform 3400) so that later, when recalled, therobot 15 will move automatically in the horizontal plane to intersect with this vector. In some embodiments, instead of manually driving therobot 15 by activatingmotors 160, the robot's axes can be put in a passive state. In some embodiments, in the passive state, themarkers 720 continue to collect data on therobot arm 23 position and encoders on each axis 64, 66, 68, 70 continue to provide information regarding the position of the axis; therefore the position of an extrapolated line can be updated on the medical image as thepassive robot 15 is dragged into any orientation and position in the horizontal plane. In some embodiments, when a desired trajectory is reached, the position can be stored into the computer memory. Some embodiments include conventional software control or a conventional switch activation capable of placing therobot 15 into an active state to immediately rigidly hold the position or trajectory, and to begin compensating for movement of thepatient 18. - In some further embodiments, the computing device that implements the
method 2700 or that is coupled to thesurgical robot 15 can render one or more planned trajectories. Such information can permit confirming that the trajectories planned are within the range of the robot's 15 reach by calculating thenecessary motor 160 encoder counts to reach each desired trajectory, and assessing if the counts are within the range of possible counts of each axis. - In some embodiments, information including whether each trajectory is in range, and how close each trajectory is to being out of range can be provided to an agent (such as a surgeon or other user, or equipment). For example, in some embodiments, a display means 29 (such as a display device 3411) can render (i.e. display) the limits of axis counts or linear or angular positions of one or more axes and the position on each axis where each targeted trajectory is currently located.
- In another embodiment, the display device 3411 (for example, a display 150) can render a view of the horizontal work field as a rectangle with the robot's x-axis 66 movement and y-axis 68 movement ranges defining the horizontal and vertical dimensions of the rectangle, respectively. In some embodiments, marks (for example, circles) on the rectangle can represent the position of each planned trajectory at the current physical location of the
robot 15 relative to thepatient 18. In another embodiment, a 3D Cartesian volume can represent the x-axis 66 movement, y-axis 68 movement and z-axis 70 movement ranges of therobot 15. In some embodiments, line segments or cylinders rendered in the volume can represent the position of each planned trajectory at the current location of therobot 15 relative to thepatient 18. Repositioning of therobot 15 or apatient 18 is performed at this time to a location that is within range of the desired trajectories. In other embodiments, the surgeon can adjust the Z Frame 72 position, which can affect the x-axis 66 range and the y-axis 68 range of trajectories that therobot 15 is capable of reaching (for example, converging trajectories require less x-axis 66 or y-axis reach the lower therobot 15 is in the z-axis 70). During this time, simultaneously, a screen shows whether tracking markers on thepatient 18 androbot 15 are in view of the detection device of the tracking system (for example,optical tracking system 3417 shown inFIG. 34 ). Repositioning of the cameras, if necessary, is also performed at this time for good visibility or optimal detection of tracking sensors. - In some embodiments, at
block 2740, orientation of an end-effectuator 30 in arobot 15 coordinate system is calculated. In some embodiments, atblock 2750, position of the end-effectuator 30 in therobot 15 coordinate system is calculated. In some embodiments, atblock 2760, a line t defining the planned trajectory in therobot 15 coordinate system is determined. In some embodiments, at block 2770,robot 15 position is locked on the planned trajectory at a current Z level. In some embodiments, atblock 2780, information indicative of quality of the trajectory lock can be supplied. In some embodiments, actual coordinate(s) of thesurgical robot 15 can be rendered in conjunction with respective coordinate(s) of the planned trajectory. In some embodiments, aural indicia can be provided based on such quality. For instance, in some embodiments, a high-frequency and/or high-amplitude noise can embody aural indicia suitable to represent a low-quality lock. In some alternative embodiments, a brief melody may be repeatedly played, such as the sound associated with successful recognition of a USB memory device by a computer, to indicate successful lock on the planned trajectory. In other embodiments, a buzz or other warning noise may be played if therobot 15 is unable to reach its target due to the axis being mechanically overpowered, or if the trackingmarkers 720 are undetectable by cameras 8200 or other marker position sensors. - In some embodiments, at block 2790, it is determined if a surgical procedure is finished and, in the affirmative case, the flow terminates. In other embodiments, the flow is directed to block 2710. In some embodiments, the
method 2700 can be implemented (i.e., executed) as part of block 2440 in certain scenarios. It should be appreciated that in some embodiments, themethod 2700 also can be implemented for anyrobot 15 having at least one feature that enable movement of therobot 15. -
FIG. 28A is a flowchart of a method 2800 a for calculating position and/or orientation of an end-effectuator 30 in arobot 15 according to one at least one embodiment of the invention. In some embodiments, the position and/or orientation can be calculated based at least on monitored position of atracking array 690 mounted on the robot's x-axis 66 and monitored counts of encoders on the y-axis 68,roll 62,pitch 60, and Z-tube axis 64 actuators. In some embodiments, position and/or orientation are calculated in arobot 15 coordinate system. In some embodiments, the method 2800 a can embody one or more ofblocks block 2805 a, the current position (i.e., 3D position) of x-axis 66 mountedrobot 15 tracking markers is accessed. In some embodiments, at block 2810 a, the current position of thetracking array 690 mounted to therobot 15 is transformed to neutral position. This is a position that was previously stored and represents the position of thetracking array 690 when therobot 15 was at zero counts on each axis between thetracker 690 and therobot 15 base (x-axis 66 and z-axis 70 in this configuration). In some embodiments, the set of transformations (T1) to transform from the current position to the neutral position can be retained in computer memory (for example, thesystem memory 3412 and/or mass storage device 3404). In some embodiments, a tip and tail of a line segment representing the vector in line with the end-effectuator 30 can be computed based at least on the process described herein. In some embodiments, this process can establish arobot 15 coordinate system and calibrate the relative orientations of the axes of movement of therobot 15 where trackingmarkers 720 can be attached temporarily to the end-effectuator 30. In some embodiments, the vector's position in space can be determined by finding the finite helical axis of motion ofmarkers 720 manually rotated to two positions around theguide tube 50. In some embodiments, the vector's position in space can be determined by connecting a point located at the entry of the guide tube 50 (identified by a temporarily mountedrigid body 690 with tracking markers 720) to a point located at the exit of the guide tube 50 (identified by a second temporarily mountedrigid body 690 with tracking markers 720). - In some embodiments, the tip of the line segment can be obtained as the point along the vector that is closest to the vector representing the helical axis of motion during pitch. In some embodiments, the tail of the line segment can be set an arbitrary distance (for example about 100 mm) up the vector aligned with the
guide tube 50 and/or first helical axis. In some embodiments, the Cartesian coordinates of such tip and tail positions can be transformed to a coordinate system described herein in which the y-axis 68 movement can coincide with the y-axis 68 of the coordinate system, and the x-axis 66 can be aligned such that x-axis 66 movement can cause the greatest change in direction in the x-axis 66, moderate change in the y-axis 68, and no change in the z-axis 70. In some embodiments, these coordinates can be retained in a computer memory (for example system memory 3412) for later retrieval. In some embodiments, at block 2815 a, tip and tail coordinates for neutral are accessed (i.e., retrieved). In some embodiments, at block 2820 a, tip and tail are translated along Z-tube 50 neutral unit vector by monitored Z-tube 50 counts. In some embodiments, at block 2825 a, an instantaneous axis of rotation (“IAR”) is accessed. The IAR is the same as the helical axis of motion ignoring the element of translation along the helical axis forpitch 60 for neutral. As described earlier, in some embodiments, the vectors for this IAR were previously stored in computer memory at the time the coordinate system of therobot 15 was calibrated. In some embodiments, at block 2830 a, tip coordinate, tail coordinate, and IAR vector direction and location coordinates are transformed (for example, iteratively transformed) to a new coordinate system in which IAR is aligned with X axis. In some embodiments, data indicative of such transformations (T2) can be stored. In some embodiments, atblock 2835 a, tip coordinate and tail coordinate are rotated about X axis bypitch 60 angle. In some embodiments, at block 2840 a, tip coordinate and tail coordinate are transformed back by inverse of T2 to the previous coordinate system. In some embodiments, at block 2845 a, previously stored vectors that represent the IAR forroll 62 are accessed. In some embodiments, at block 2850 a, tip coordinate, tail coordinate, IAR coordinate are transformed (for example, iteratively transformed) to a new coordinate system in which IAR is aligned with y-axis 68. In some embodiments, data indicative of such transformation(s) (T3) can be retained in memory. In some embodiments, at block 2855 a, tip coordinate and tail coordinate are rotated about y-axis 68 byroll 62 angle. In some embodiments, at block 2860 a, tip coordinate and tail coordinate are transformed back by inverse of T3 to the previous coordinate system. In some embodiments, at block 2865 a, tip coordinate and tail coordinate are translated along a y-axis 68 unit vector (e.g., a vector aligned in this coordinate system with the y-axis 68) by monitored counts. In some embodiments, at block 2870 a, tip coordinate and tail coordinate are transformed back by inverse of T1 to the current coordinate system monitored by thetracking system 3417. -
FIG. 28B is a flowchart of a method 2800 b for calculating position and/or orientation of an end-effectuator 30 in arobot 15 in accordance with one embodiment of the invention. In some embodiments, the position and/or the orientation can be calculated based at least on monitored position of atracking array 690 mounted on the robot's 15roll 62 axis and monitored counts of encoders on thepitch 60 and Z-tube 50 actuators. In some embodiments, position and/or orientation can be calculated in arobot 15 coordinate system. In accordance with some embodiments of the invention, the method 2800B can embody one or more ofblocks block 2805 b, current position of an array of one ormore robot 15tracking markers 720 is accessed. In some embodiments, the current position is a 3D position and the array ofrobot 15tracking markers 720 can be mounted to theroll 62 axis of therobot 15. In some embodiments, at block 2810 b, the current position of the array ofrobot 15tracking markers 720 mounted to therobot 15 is transformed to neutral position. In some embodiments, the neutral position can be a position that was previously stored and can represent the position of therobot 15tracking array 690 when therobot 15 had zero counts on each axis between thetracker 690 and the robot base 25 (e.g., z-axis 70, x-axis 68, y-axis 66, and roll 62 axis in this configuration). In some embodiments, data indicative of a set of transformations (T1) to go from the current position to the neutral position can be stored in a computer memory (for example,mass storage device 3404 or system memory 3412). - In some embodiments, in order to establish a
robot 15 coordinate system and calibrate the relative orientations of the axes of movement of therobot 15, a tip and tail of a line segment representing the vector in line with the end-effectuator 30 with temporarily attached trackingmarkers 720 is located. In some embodiments, the vector's position in space can be determined by finding the finite helical axis of motion of markers manually rotated to two positions around theguide tube 50. In other embodiments, the vector's position in space can be determined by connecting a point located at the entry of the guide tube 50 (identified by a temporarily mountedrigid body 690 with tracking markers 720) to a point located at the exit of the guide tube 50 (identified by a second temporarily mountedrigid body 690 with tracking markers 720). - In some embodiments, the tip of the line segment can be found as the point along the vector that is closest to the vector representing the helical axis of motion during pitch. In some embodiments, the tail of the line segment can be set an arbitrary distance (for example, nearly 100 mm) up the vector aligned with the guide tube/first helical axis. In some embodiments, the Cartesian coordinates of these tip and tail positions can be transformed to a coordinate system described herein in which the y-axis 68 movement substantially coincides with the y-axis 68 of the coordinate system, and the x-axis 66 movement is aligned in a manner that, in some embodiments, x-axis 66 movement causes the greatest change in direction in the x-axis 66, slight change in y-axis 68, and no change in the z-axis 70. It should be appreciated that such coordinates can be retained in memory (for example system memory 3412) for later retrieval. In some embodiments, at block 2815 b, tip and tail coordinates for the neutral position are accessed (i.e., retrieved or otherwise obtained). In some embodiments, at block 2820 b, tip and tail are translated along Z-
tube 50 neutral unit vector by monitored Z-tube 50 counts. In some embodiments, atblock 2825 b, IAR is accessed. In one implementation, the vectors for this IAR may be available in a computer memory, for example, such vectors may be retained in the computer memory at the time the coordinate system of therobot 15 is calibrated in accordance with one or more embodiments described herein. In some embodiments, atblock 2830 b, tip coordinate, tail coordinate, and IAR vector direction and location coordinates are transformed to a new coordinate system in which IAR is aligned with x-axis 66. In some embodiments, data indicative of the applied transformations (T2) can be retained in a computer memory. In some embodiments, atblock 2835 b, tip coordinate and tail coordinate are rotated about x-axis 66 bypitch 60 angle. In some embodiments, at block 2840 b, tip coordinate and tail coordinate are transformed back by applying the inverse of T2 to the previous coordinate system. In some embodiments, atblock 2870 b, tip coordinate and tail coordinate are transformed back by applying the inverse of T1 to the current coordinate system monitored by thetracking system 3417. -
FIG. 29 is a flowchart of amethod 2900 for determining a line indicative of a trajectory in arobot 15 coordinate system in accordance with one embodiment of the invention. In some embodiments, the trajectory can be a planned trajectory associated with a surgical procedure. In some embodiments, atblock 2910, for a set of currentactive marker 720 positions on a targetingfixture 690, respectiveopaque marker 730 positions are accessed from a rigid body source (such as a fixture 690). In some embodiments, atblock 2920, anopaque marker 730 position is transformed from a representation in an image coordinate system to a representation in a currentactive marker 720 coordinate system. In some embodiments, atblock 2930, a planned trajectory is transformed from a representation in the image coordinate system to a representation in the currentactive maker 720 coordinate system. In some embodiments, at block 2940, a set ofcurrent marker 720 positions on arobot 15 is transformed to a representation in arobot 15 coordinate system. In some embodiments, atblock 2950, the planned trajectory is transformed from a representation in the currentactive marker 720 coordinate system to therobot 15 coordinate system. -
FIG. 30 is a flowchart of amethod 3000 for adjusting arobot 15 position to lock on a trajectory in accordance in accordance with one embodiment of the invention. As illustrated, in some embodiments, the trajectory can be locked at a current Z plane, or level above thesurgical field 17. In some embodiments, atblock 3005, it is determined ifroll 62 of an end-effectuator 30 matches roll of the trajectory (represented by a line t (or t)). In the negative case, in some embodiments, an instruction to move aroll 62 axis is transmitted atblock 3010 and flow is directed to block 3015. Conversely, in the affirmative case, in some embodiments, flow is directed to block 3015 where it is determined if thepitch 60 of the end-effectuator 30 has matched the pitch of the trajectory. In the negative case, an instruction to move a pitch 60 axis is transmitted atblock 3020 and flow is directed to block 3025. In the affirmative case, in some embodiments, flow is directed to block 3025 where it is determined if x-axis 66 coordinates of points on the vector of the end-effectuator 30 intercept the x-axis 66 coordinates of the desired trajectory vector. In the negative case, in some embodiments, an instruction to move the x-axis 66 can be transmitted and flow is directed to 3035. In the affirmative case, in some embodiments, flow is directed to block 3035 where it is determined if y-axis 68 coordinates of points on the vector of the end-effectuator 30 intercept the y-axis 68 coordinates of the desired trajectory vector. In the negative case, in some embodiments, an instruction to move the y-axis 68 can be transmitted atblock 3040 and flow is directed to block 3045. In the affirmative case, in some embodiments, flow is directed to block 3045 in which it is determined if a Z-tube 50 is being adjusted. In some embodiments, an end-user can configure information (i.e., data or metadata) indicative of the Z-tube 50 being adjusted to control it to a desired position, for example. In the negative case, in some embodiments, flow is terminated. In the affirmative case, in some embodiments, flow is directed to block 3050 where it is determined if the Z-tube 50 is positioned at a predetermined distance from anatomy. In the affirmative case, in some embodiments, flow terminates and the Z-tube 50 is located at a desired position with respect to a target location in the anatomy (bone, biopsy site, etc.). In the negative case, in some embodiments, an instruction to move the Z-tube axis 64 is transmitted atblock 3055 and the flow is directed to block 3050. It should be noted that thesubject method 3000 in some embodiments, but not all embodiments, may require that movement in each of the indicated axes (x-axis 66, y-axis 68, Z-tube axis 64,roll 62, and pitch 60) occurs without affecting the other axes earlier in the method flow. For example, in some embodiments, the y-axis 68 movement atblock 3040 should not cause change in the position of x-axis 66 coordinate, which was already checked atblock 3025. In some embodiments, themethod 3000 can be implemented iteratively in order to reach a desired final position in instances where the axes do not move completely independently. In certain embodiments, themethod 3000 can account for all axis positions nearly simultaneously, and can determine the exact amount of movement necessary in each axis, and thus it can be more efficient. -
FIG. 31 is a flowchart of amethod 3100 for positioning an end-effectuator 30 in space in accordance with one embodiment of the invention. In some embodiments, the positioning can comprise positioning the end-effectuator 30 in a first plane (for example, the x-y plane or horizontal plane) and moving the end-effectuator 30 along a direction substantially normal to the first plane.FIGS. 32-33 are flowcharts of methods for driving an end-effectuator 30 to a procedure location in accordance with one embodiment of the invention. As an example, in some embodiments, the procedure location can be a position at the surface of a bone into which a conventional screw of other piece of hardware is to be inserted. In some embodiments, the end-effectuator 30 can be fitted with aguide tube 50 or conventional dilator. In some embodiments, in scenarios in which a Z-tube 50 of asurgical robot 15 comprising the end-effectuator 30 is to be locked and a Z-frame 72 is to be advanced, themethod 3200 can be implemented (i.e., executed). In applications where conventional screws are to be driven into bone, the surgeon may want to move the end-effectuator tip 30, fitted with aguide tube 50 or a conventional dilator, all the way down to the bone. It should be appreciated that in some embodiments, since the first lateral movement occurs above the level where thepatient 18 is lying, the methods depicted inFIGS. 31-33 can mitigate the likelihood that therobot 15 randomly collides with apatient 18. In some embodiments, the method can also utilize the robot's Cartesian architecture, and the ease with which a coordinated movement down the infinite trajectory vector can be made. That is, in some embodiments, to move down this vector, theroll 62 and pitch 60 axes need no adjustment, while the x-axis 66, y-axis 68, and Z-frame 72 axes are moved at a fixed rate. In certain embodiments, for anarticular robot 15 to make such a move, the multiple angular axes would have to be synchronized nonlinearly, with all axes simultaneously moved at varying rates. -
FIG. 34 illustrates a block diagram of acomputer platform 3400 having acomputing device 3401 that enables various features of the invention, and performance of the various methods disclosed herein in accordance with some embodiments of the invention. In some embodiments, thecomputing device 3401 can control operation of asurgical robot 15 and anoptical tracking system 3417 in accordance with aspects described herein. In some embodiments, control can comprise calibration of relative systems of coordinates, generation of planned trajectories, monitoring of position of various units of thesurgical robots 15 and/or units functionally coupled thereto, and implementation of safety protocols, and the like. For example, in some embodiments,computing device 3401 can embody a programmable controller that can control operation of asurgical robot 15 as described herein. It should be appreciated that in accordance with some embodiments of the invention, theoperating environment 3400 is only an example of an operating environment and is not intended to suggest any limitation as to the scope of use or functionality of operating environment architecture. In some embodiments of the invention, theoperating environment 3400 should not be interpreted as having any dependency or requirement relating to any one functional element or combination of functional elements (e.g., units, components, adapters, or the like). - The various embodiments of the invention can be operational with numerous other general purpose or special purpose computing system environments or configurations. Examples of well-known computing systems, environments, and/or configurations that can be suitable for use with the systems and methods of the invention comprise personal computers, server computers, laptop devices or handheld devices, and multiprocessor systems. Additional examples comprise mobile devices, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that comprise any of the above systems or devices, and the like.
- In some embodiments, the processing effected in the disclosed systems and methods can be performed by software components. In some embodiments, the disclosed systems and methods can be described in the general context of computer-executable instructions, such as program modules, being executed by one or more computers, such as
computing device 3401, or other computing devices. Generally, program modules comprise computer code, routines, programs, objects, components, data structures, etc., that perform particular tasks or implement particular abstract data types. The disclosed methods also can be practiced in grid-based and distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules can be located in both local and remote computer storage media including memory storage devices. - Further, one skilled in the art will appreciate that the systems and methods disclosed herein can be implemented via a general-purpose computing device in the form of the
computing device 3401. In some embodiments, the components of thecomputing device 3401 can comprise, but are not limited to, one ormore processors 3403, orprocessing units 3403, asystem memory 3412, and a system bus 3413 that couples various system components including theprocessor 3403 to thesystem memory 3412. In some embodiments, in the case ofmultiple processing units 3403, the system can utilize parallel computing. - In general, a
processor 3403 or aprocessing unit 3403 refers to any computing processing unit or processing device comprising, but not limited to, single-core processors; single-processors with software multithread execution capability; multi-core processors; multi-core processors with software multithread execution capability; multi-core processors with hardware multithread technology; parallel platforms; and parallel platforms with distributed shared memory. Additionally or alternatively, aprocessor 3403 orprocessing unit 3403 can refer to an integrated circuit, an application specific integrated circuit (ASIC), a digital signal processor (DSP), a field programmable gate array (FPGA), a programmable logic controller (PLC), a complex programmable logic device (CPLD), a discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. Processors or processing units referred to herein can exploit nano-scale architectures such as, molecular and quantum-dot based transistors, switches and gates, in order to optimize space usage or enhance performance of the computing devices that can implement the various aspects of the subject invention. In some embodiments,processor 3403 orprocessing unit 3403 also can be implemented as a combination of computing processing units. - The system bus 3413 represents one or more of several possible types of bus structures, including a memory bus or memory controller, a peripheral bus, an accelerated graphics port, and a processor or local bus using any of a variety of bus architectures. By way of example, such architectures can comprise an Industry Standard Architecture (ISA) bus, a Micro Channel Architecture (MCA) bus, an Enhanced ISA (EISA) bus, a Video Electronics Standards Association (VESA) local bus, an Accelerated Graphics Port (AGP) bus, and a Peripheral Component Interconnects (PCI), a PCI-Express bus, a Personal Computer Memory Card Industry Association (PCMCIA), Universal Serial Bus (USB) and the like. The bus 3413, and all buses specified in this specification and annexed drawings also can be implemented over a wired or wireless network connection and each of the subsystems, including the
processor 3403, amass storage device 3404, anoperating system 3405,robotic guidance software 3406, roboticguidance data storage 3407, anetwork adapter 3408,system memory 3412, an input/output interface 3410, adisplay adapter 3409, adisplay device 3411, and ahuman machine interface 3402, can be contained within one or moreremote computing devices 3414 a,b at physically separate locations, functionally coupled (e.g., communicatively coupled) through buses of this form, in effect implementing a fully distributed system. - In some embodiments,
robotic guidance software 3406 can configure thecomputing device 3401, or a processor thereof, to perform the automated control of position of the local robot 3416 (for example, surgical robot 15) in accordance with aspects of the invention. Such control can be enabled, at least in part, by atracking system 3417. In some embodiments, when thecomputing device 3401 embodies thecomputer 100 functionally coupled tosurgical robot 15,robotic guidance software 3406 can configuresuch computer 100 to perform the functionality described in the subject invention. In some embodiments,robotic guidance software 3406 can be retained in a memory as a group of computer-accessible instructions (for instance, computer-readable instructions, computer-executable instructions, or computer-readable computer-executable instructions). In some embodiments, the group of computer-accessible instructions can encode the methods of the invention (such as the methods illustrated inFIGS. 24-33 in accordance with some embodiments of the invention). In some embodiments, the group of computer-accessible instructions can encode various formalisms (e.g., image segmentation) for computer vision tracking. Some embodiments includerobotic guidance software 3406 that can include a compiled instance of such computer-accessible instructions, a linked instance of such computer-accessible instructions, a compiled and linked instance of such computer-executable instructions, or an otherwise executable instance of the group of computer-accessible instructions. - Some embodiments include robotic
guidance data storage 3407 that can comprise various types of data that can permit implementation (e.g., compilation, linking, execution, and combinations thereof) of therobotic guidance software 3406. In some embodiments, roboticguidance data storage 3407 can comprise data associated with intraoperative imaging, automated adjustment of position of thelocal robot 3416 and/orremote robot 3422, or the like. In some embodiments, the data retained in the roboticguidance data storage 3407 can be formatted according to any image data in industry standard format. As illustrated, in some embodiments, aremote tracking system 3424 can enable, at least in part, control of theremote robot 3422. In some embodiments, the information can comprise tracking information, trajectory information, surgical procedure information, safety protocols, and so forth. - In some embodiments of the invention, the
computing device 3401 typically comprises a variety of computer readable media. The readable media can be any available media that is accessible by thecomputer 3401 and comprises, for example and not meant to be limiting, both volatile and non-volatile media, removable and non-removable media. In some embodiments, thesystem memory 3412 comprises computer readable media in the form of volatile memory, such as random access memory (RAM), and/or non-volatile memory, such as read only memory (ROM). In some embodiments, thesystem memory 3412 typically contains data (such as a group of tokens employed for code buffers) and/or program modules such asoperating system 3405 androbotic guidance software 3406 that are immediately accessible to, and/or are presently operated-on by theprocessing unit 3403. In some embodiments,operating system 3405 can comprise operating systems such as Windows operating system, Unix, Linux, Symbian, Android, Apple iOS operating system, Chromium, and substantially any operating system for wireless computing devices or tethered computing devices. Apple® is a trademark of Apple Computer, Inc., registered in the United States and other countries. iOS® is a registered trademark of Cisco and used under license by Apple Inc. Microsoft® and Windows® are either registered trademarks or trademarks of Microsoft Corporation in the United States and/or other countries. Android® and Chrome® operating system are a registered trademarks of Google Inc. Symbian® is a registered trademark of Symbian Ltd. Linux® is a registered trademark of Linus Torvalds. UNIX® is a registered trademark of The Open Group. - In some embodiments,
computing device 3401 can comprise other removable/non-removable, volatile/non-volatile computer storage media. As illustrated, in some embodiments,computing device 3401 comprises amass storage device 3404 which can provide non-volatile storage of computer code (e.g., computer-executable instructions), computer-readable instructions, data structures, program modules, and other data for thecomputing device 3401. For instance, in some embodiments, amass storage device 3404 can be a hard disk, a removable magnetic disk, a removable optical disk, magnetic cassettes or other magnetic storage devices, flash memory cards, CD-ROM, digital versatile disks (DVD) or other optical storage, random access memories (RAM), read only memories (ROM), electrically erasable programmable read-only memory (EEPROM), and the like. - In some embodiments, optionally, any number of program modules can be stored on the
mass storage device 3404, including by way of example, anoperating system 3405, andtracking software 3406. In some embodiments, each of theoperating system 3405 and tracking software 3406 (or some combination thereof) can comprise elements of the programming and thetracking software 3406. In some embodiments, data and code (for example, computer-executable instructions, patient-specific trajectories, andpatient 18 anatomical data) can be retained as part oftracking software 3406 and stored on themass storage device 3404. In some embodiments,tracking software 3406, and related data and code, can be stored in any of one or more databases known in the art. Examples of such databases comprise, DB2®, Microsoft® Access, Microsoft® SQL Server, Oracle®, mySQL, PostgreSQL, and the like. Further examples include membase databases and flat file databases. The databases can be centralized or distributed across multiple systems. - DB2® is a registered trademark of IBM in the United States. Microsoft®, Microsoft® Access®, and Microsoft® SQL Server™ are either registered trademarks or trademarks of Microsoft Corporation in the United States and/or other countries. Oracle® is a registered trademark of Oracle Corporation and/or its affiliates. MySQL® is a registered trademark of MySQL AB in the United States, the European Union and other countries. PostgreSQL® and the PostgreSQL® logo are trademarks or registered trademarks of The PostgreSQL Global Development Group, in the U.S. and other countries.
- In some embodiments, an agent (for example, a surgeon or other user, or equipment) can enter commands and information into the
computing device 3401 via an input device (not shown). Examples of such input devices can comprise, but are not limited to, a camera (or other detection device for non-optical tracking markers), a keyboard, a pointing device (for example, a mouse), a microphone, a joystick, a scanner (for example, a barcode scanner), a reader device such as a radiofrequency identification (RFID) readers or magnetic stripe readers, gesture-based input devices such as tactile input devices (for example, touch screens, gloves and other body coverings or wearable devices), speech recognition devices, or natural interfaces, and the like. In some embodiments, these and other input devices can be connected to theprocessing unit 3403 via ahuman machine interface 3402 that is coupled to the system bus 3413. In some other embodiments, they can be connected by other interface and bus structures, such as a parallel port, game port, an IEEE 1394 port (also known as a firewire port), a serial port, or a universal serial bus (USB). - In some further embodiments, a
display device 3411 can also be functionally coupled to the system bus 3413 via an interface, such as adisplay adapter 3409. In some embodiments, thecomputer 3401 can have more than onedisplay adapter 3409 and thecomputer 3401 can have more than onedisplay device 3411. For example, in some embodiments, adisplay device 3411 can be a monitor, a liquid crystal display, or a projector. Further, in addition to thedisplay device 3411, some embodiments can include other output peripheral devices that can comprise components such as speakers (not shown) and a printer (not shown) capable of being connected to thecomputer 3401 via input/output Interface 3410. In some embodiments, the input/output interface 3410 can be a pointing device, either tethered to, or wirelessly coupled to thecomputing device 3410. In some embodiments, any step and/or result of the methods can be output in any form to an output device. In some embodiments, the output can be any form of visual representation, including, but not limited to, textual, graphical, animation, audio, tactile, and the like. - In certain embodiments, one or more cameras can be contained or functionally coupled to the
tracking system 3417, which is functionally coupled to the system bus 3413 via an input/output interface of the one or more input/output interfaces 3410. Such functional coupling can permit the one or more camera(s) to be coupled to other functional elements of thecomputing device 3401. In one embodiment, the input/output interface, at least a portion of the system bus 3413, and thesystem memory 3412 can embody a frame grabber unit that can permit receiving imaging data acquired by at least one of the one or more cameras. In some embodiments, the frame grabber can be an analog frame grabber, a digital frame grabber, or a combination thereof. In some embodiments, where the frame grabber is an analog frame grabber, theprocessor 3403 can provide analog-to-digital conversion functionality and decoder functionality to enable the frame grabber to operate with medical imaging data. Further, in some embodiments, the input/output interface can include circuitry to collect the analog signal received from at least one camera of the one or more cameras. In some embodiments, in response to execution byprocessor 3403,tracking software 3406 can operate the frame grabber to receive imaging data in accordance with various aspects described herein. - Some embodiments include a
computing device 3401 that can operate in a networked environment (for example, an industrial environment) using logical connections to one or moreremote computing devices 3414 a,b, aremote robot 3422, and atracking system 3424. By way of example, in some embodiments, a remote computing device can be a personal computer, portable computer, a mobile telephone, a server, a router, a network computer, a peer device or other common network node, and so on. In particular, in some embodiments, an agent (for example, a surgeon or other user, or equipment) can point to other tracked structures, including anatomy of apatient 18, using a remote computing device 3414 such as a hand-held probe that is capable of being tracked and sterilized. In some embodiments, logical connections between thecomputer 3401 and aremote computing device 3414 a,b can be made via a local area network (LAN) and a general wide area network (WAN). In some embodiments, the network connections can be implemented through anetwork adapter 3408. In some embodiments, thenetwork adapter 3408 can be implemented in both wired and wireless environments. Some embodiments include networking environments that can be conventional and commonplace in offices, enterprise-wide computer networks, intranets. In some embodiments, the networking environments generally can be embodied in wire-line networks or wireless networks (for example, cellular networks, such as third generation (“3G”) and fourth generation (“4G”) cellular networks, facility-based networks (for example, femtocell, picocell, wifi networks). In some embodiments, a group of one ormore networks 3415 can provide such networking environments. In some embodiments of the invention, the one or more network(s) can comprise a LAN deployed in an industrial environment comprising thesystem 1 described herein. - As an illustration, in some embodiments, application programs and other executable program components such as the
operating system 3405 are illustrated herein as discrete blocks, although it is recognized that such programs and components reside at various times in different storage components of thecomputing device 3401, and are executed by the data processor(s) of thecomputer 100. Some embodiments include an implementation oftracking software 3406 that can be stored on or transmitted across some form of computer readable media. Any of the disclosed methods can be performed by computer readable instructions embodied on computer readable media. Computer readable media can be any available media that can be accessed by a computer. By way of example and not meant to be limiting, computer-readable media can comprise “computer storage media,” or “computer-readable storage media,” and “communications media.” “Computer storage media” comprise volatile and non-volatile, removable and non-removable media implemented in any methods or technology for storage of information such as computer readable instructions, data structures, program modules, or other data. In some embodiments of the invention, computer storage media comprises, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by a computer. - As described herein, some embodiments include the
computing device 3401 that can control operation oflocal robots 3416 and/orremote robots 3422. Within embodiments in which thelocal robot 3416 or theremote robot 3422 aresurgical robots 15, thecomputing device 3401 can executerobotic guidance software 3407 to controlsuch robots robotic guidance software 3407, in response to execution, can utilize trajectories (such as, tip and tail coordinates) that can be planned and/or configured remotely or locally. In an additional or alternative aspect, in response to execution, therobotic guidance software 3407 can implement one or more of the methods described herein in a local robot's computer or a remote robot's computer to cause movement of theremote robot 15 or thelocal robot 15 according to one or more trajectories. - In some embodiments, the
computing device 3401 can enable pre-operative planning of the surgical procedure. In some embodiments, thecomputing device 3401 can permit spatial positioning and orientation of a surgical tool (for example, instrument 35) during intraoperative procedures. In some further embodiments, thecomputing device 3401 can enable open procedures. In some other embodiments, thecomputing device 3401 can enable percutaneous procedures. In certain embodiments, thecomputing device 3401 and therobotic guidance software 3407 can embody a3D tracking system 3417 to simultaneously monitor the positions of the device and the anatomy of thepatient 18. In some embodiments, the3D tracking system 3417 can be configured to cast the patient's anatomy and the end-effectuator 30 in a common coordinate system. - In some embodiments, the
computing device 3401 can access (i.e., load) image data from a conventional static storage device. In some embodiments, thecomputing device 3401 can permit a 3D volumetric representation ofpatient 18 anatomy to be loaded into memory (for example, system memory 3412) and displayed (for example, via display device 3411). In some embodiments, thecomputing device 3401, in response to execution of therobotic guidance software 3407 can enable navigation through the 3D volume representation of a patient's anatomy. - In some embodiments, the
computing device 3401 can operate with a conventional power source required to offer the device for sale in the specified country. A conventional power cable that supplies power can be a sufficient length to access conventional hospital power outlets. In some embodiments, in the event of a power loss, thecomputing device 3401 can hold the current end-effectuator 30 in a position unless an agent (for example, a surgeon or other user, or equipment) manually moves the end-effectuator 30. - In some embodiments, the
computing device 3401 can monitor system physical condition data. In some embodiments, thecomputing device 3401 can report to an operator (for example, a surgeon) each of the physical condition data and indicate an out-of-range value. In some embodiments, thecomputing device 3401 can enable entry and storage of manufacturing calibration values for end-effectuator 30 positioning using, for example, the input/output interface 3410. In some embodiments, thecomputing device 3401 can enable access to manufacturing calibration values by an agent (for example, a surgeon or other user, or equipment) authenticated to an appropriate access level. In some embodiments, the data can be retained in roboticguidance data storage 3407, or can be accessed via network(s) 3415 when the data is retained in aremote computing device 3414 a. - In some embodiments, the
computing device 3401 can render (using for example display device 3411) a technical screen with a subset of the end-effectuator 30 positioning calibration and system health data. The information is only accessible to an agent (for example, a surgeon or other user, or equipment) authenticated to an appropriate level. - In some embodiments, the
computing device 3401 can enable field calibration of end-effectuator 30 positioning only by an agent (for example, a surgeon or other user, or equipment) authenticated to an appropriate access level. In some embodiments, thecomputing device 3401 can convey the status oflocal robot 3416,remote robot 3422, and/or other device being locked in position using a visual or aural alert. - In some further embodiments, the
computing device 3401 can include an emergency stop control that upon activation, disables power to the device'smotors 160 but not to theprocessor 3403. In some embodiments, the emergency stop control can be accessible by the operator ofcomputing device 3401. In some embodiments, thecomputing device 3401 can monitor the emergency stop status and indicate to the operator that the emergency stop has been activated. - In some other embodiments, the
computing device 3401 can be operated in a mode that permits manual positioning of the end-effectuator 30. In some embodiments, thecomputing device 3401 can boot directly to an application representing therobotic guidance software 3406. In some embodiments,computing device 3401 can perform a system check prior to each use. In scenarios in which the system check fails, thecomputing device 3401 can notify an operator. - In some embodiments, the
computing device 3401 can generate an indicator for reporting system status. Some embodiments include thecomputing device 3401 that can minimize or can mitigate delays in processing, and in the event of a delay in processing, notify an agent (for example, a surgeon or other user, or equipment). For example, in some embodiments, a delay may occur while a system scan is being performed to assess system status, and consequently thecomputing device 3401 can schedule (for example, generate a process queue) system scans to occur at low usage times. In some embodiments, a system clock of thecomputing device 3401 can be read before and after key processes to assess the length of time required to complete computation of a process. In some embodiments, the actual time to complete the process can be compared to the expected time. In some embodiments, if a discrepancy is found to be beyond an acceptable tolerance, the agent can be notified, and/or concurrently running non-essential computational tasks can be terminated. In one embodiment, a conventional system clock (not shown) can be part ofprocessor 3403. - In some embodiments, the
computing device 3401 can generate a display that follows a standardized workflow. In some embodiments, thecomputing device 3401 can render or ensure that text is rendered in a font of sufficient size and contrast to be readable from an appropriate distance. In some embodiments, thecomputing device 3401 can enable an operator to locate the intended position of a surgical implant or tool. - In some further embodiments, the
computing device 3401 can determine the relative position of the end-effectuator 30 to the anatomy of thepatient 18. For example, to at least such end, thecomputing device 3401 can collect data theoptical tracking system 3417, and can analyze the data to generate data indicative of such relative position. In some embodiments, thecomputing device 3401 can indicate the end-effectuator 30 position and orientation. In some embodiments, thecomputing device 3401 can enable continuous control of end-effectuator 30 position relative to the anatomy of apatient 18. - In some embodiments, the
computing device 3401 can enable an agent (for example, a surgeon or other user, or equipment) to mark the intended position of a surgical implant or tool (for example, instrument 35). In some embodiments, thecomputing device 3401 can allow the position and orientation of a conventional hand-held probe (or an instrument 35) to be displayed overlaid on images of the patient's anatomy. - In some embodiments, the
computing device 3401 can enable an agent (for example, a surgeon or other user, or equipment) to position conventional surgical screws. In some embodiments, thecomputing device 3401 can enable selection of the length and diameter of surgical screws by the agent. In yet another aspect, the computing device can ensure that the relative position, size and scale of screws are maintained on thedisplay 3411 when in graphical representation. In some embodiments, thecomputing device 3401 can verify screw path plans against an operation envelope and reject screw path plans outside this envelope. In still another aspect, thecomputing device 3401 can enable hiding of a graphical screw representation. - In some embodiments, the
computing device 3401 can enable a function that allows the current view to be stored. In some embodiments, thecomputing device 3401 can enable a view reset function that sets the current view back to a previously stored view. In some embodiments, thecomputing device 3401 can enable an authentication based tiered access system. In some embodiments, thecomputing device 3401 can log and store system activity. In some embodiments, thecomputing device 3401 can enable access to the system activity log to an agent authorized to an appropriate level. In some embodiments, thecomputing device 3401 can enable entry and storage ofpatient 18 data. - In some embodiments, the
computing device 3401 can enable the appropriate disposition ofpatient 18 data and/or procedure data. For example, in a scenario in which such data are being collected for research, thecomputing device 3401 can implement de-identification of the data in order to meetpatient 18 privacy requirements. In some embodiments, the de-identification can be implemented in response to execution of computer-executable instruction(s) retained inmemory 3412 or any other memory accessible to thecomputing device 3401. In some embodiments, the de-identification can be performed automatically before the patient 18 data and/or procedure data are sent to a repository or any other data storage (includingmass storage device 3404, for example). In some embodiments, indicia (e.g., a dialog box) can be rendered (for example, at display device 3411) to prompt an agent (e.g., machine or human) to permanently deletepatient 18 data and/or procedure data at the end of a procedure. -
FIG. 11 shows a flow chart diagram 1100 for general operation of therobot 15 according to some embodiments is shown. In some embodiments, atstep 210, the local positioning system (herein referred to as “LPS”) establishes a spatial coordinate measuring system for theroom 10 where the invasive procedure is to occur; in other words, the LPS is calibrated. In some embodiments, in order to calibrate the LPS, a conventional mechanical fixture that includes a plurality of attached calibratingtransmitters 120 is placed within theroom 10 wherepositioning sensors 12 are located. In some embodiments of the invention, at least three calibratingtransmitters 120 are required, but any number of calibratingtransmitters 120 above three is within the scope of the invention. Also, in some embodiments, at least threepositioning sensors 12 are required, but any number ofpositioning sensors 12 above three is also within the scope of the invention, and the accuracy of the system is increased with the addition of more positioning sensors. - In some embodiments, the distance between each of the calibrating
transmitters 120 relative to each other is measured prior tocalibration step 210. Each calibratingtransmitter 120 transmits RF signals on a different frequency so that thepositioning sensors 12 can determine whichtransmitter 120 emitted a particular RF signal. In some embodiments, the signal of each of thesetransmitters 120 is received by positioningsensors 12. In some embodiments, since the distance between each of the calibratingtransmitters 120 is known, and thesensors 12 can identify the signals from each of the calibratingtransmitters 120 based on the known frequency, using time of flight calculation, thepositioning sensors 12 are able to calculate the spatial distance of each of thepositioning sensors 12 relative to each other. Thesystem 1 is now calibrated. As a result, in some embodiments, thepositioning sensors 12 can now determine the spatial position of anynew RF transmitter 120 introduced into theroom 10 relative to thepositioning sensors 12. - In some embodiments, a step 220 a in which a 3D anatomical image scan, such as a CT scan, is taken of the anatomical target. Any 3D anatomical image scan may be used with the
surgical robot 15 and is within the scope of the present invention. In some embodiments, atstep 230, the positions of theRF transmitters 120 tracking the anatomical target are read by positioningsensors 110. Thesetransmitters 120 identify the initial position of the anatomical target and any changes in position during the procedure. In some embodiments, if anyRF transmitters 120 must transmit through a medium that changes the RF signal characteristics, then the system will compensate for these changes when determining the transmitter's 120 position. - In some embodiments, at
step 240, the positions of thetransmitters 120 on the anatomy are calibrated relative to the LPS coordinate system. In other words, the LPS provides a reference system, and the location of the anatomical target is calculated relative to the LPS coordinates. In some embodiments, to calibrate the anatomy relative to the LPS, the positions oftransmitters 120 affixed to the anatomical target are recorded at the same time as positions oftemporary transmitters 120 placed on precisely known anatomical landmarks also identified on the anatomical image. This calculation is performed by acomputer 100. - In some embodiments, at
step 250, the positions of theRF transmitters 120 that track the anatomical target are read. Since the locations of thetransmitters 120 on the anatomical target have already been calibrated, the system can easily determine if there has been any change in position of the anatomical target. Some embodiments include astep 260, where the positions of thetransmitters 120 on thesurgical instrument 35 are read. Thetransmitters 120 may be located on thesurgical instrument 35 itself, and/or there may betransmitters 120 attached to various points of thesurgical robot 15. - In some embodiments of the invention, the
surgical robot 15 can also include a plurality of attached conventional position encoders that help determine the position of thesurgical instrument 35. In some embodiments, the position encoders can be devices used to generate an electronic signal that indicates a position or movement relative to a reference position. In some other embodiments, a position signal can be generated using conventional magnetic sensors, conventional capacitive sensors, and conventional optical sensors. - In some embodiments, position data read from the position encoders may be used to determine the position of the
surgical instrument 35 used in the procedure. In some embodiments, the data may be redundant of position data calculated fromRF transmitters 120 located on thesurgical instrument 35. Therefore, in some embodiments, position data from the position encoders may be used to double-check the position being read from the LPS. - In some embodiments, at
step 270, the coordinates of the positions of thetransmitters 120 on thesurgical instrument 35, and/or the positions read from the position encoders, is calibrated relative to the anatomical coordinate system. In other words, in some embodiments, the position data of thesurgical instrument 35 is synchronized into the same coordinate system as the patient's anatomy. In some embodiments, this calculation is performed automatically by thecomputer 100 since the positions of thetransmitters 120 on the anatomical target and the positions of thetransmitters 120 on thesurgical instrument 35 are in the same coordinate system, and the positions of thetransmitters 120 on the anatomical target are already calibrated relative to the anatomy. - In some embodiments, at
step 280, thecomputer 100 superimposes a representation of the location calculated instep 270 of the surgical device on the 3D anatomical image of the patient 18 taken instep 220. In some embodiments, the superimposed image can be displayed to an agent. In some embodiments, atstep 290, thecomputer 100 sends the appropriate signals to themotors 160 to drive thesurgical robot 15. In some embodiments, if the agent preprogrammed a trajectory, then therobot 15 is driven so that thesurgical instrument 35 follows the preprogrammed trajectory if there is no further input from the agent. In some embodiments, if there is agent input, then thecomputer 100 drives therobot 15 in response to the agent input. - In some embodiments, at
step 295, thecomputer 100 determines whether the anatomy needs to be recalibrated. In some embodiments, the agent may choose to recalibrate the anatomy, in which case thecomputer 100 responds to agent input. Alternatively, in some embodiments, thecomputer 100 may be programmed to recalibrate the anatomy in response to certain events. For instance, in some embodiments, thecomputer 100 may be programmed to recalibrate the anatomy if theRF transmitters 120 on the anatomical target indicate that the location of the anatomical target has shifted relative to the RF transmitters 120 (i.e. this spatial relationship should be fixed). In some embodiments, an indicator that the anatomical target location has shifted relative to thetransmitters 120 is if thecomputer 100 calculates that thesurgical instrument 35 appears to be inside bone when no drilling or penetration is actually occurring. - In some embodiments, if the anatomy needs to be calibrated, then the process beginning at
step 230 is repeated. In some embodiments, if the anatomy does not need to be recalibrated, then the process beginning atstep 250 is repeated. In some embodiments, at any time during the procedure, certain fault conditions may cause thecomputer 100 to interrupt the program and respond accordingly. For instance, in some embodiments, if the signal from theRF transmitters 120 cannot be read, then thecomputer 100 may be programmed to stop the movement of therobot 15, or remove thesurgical instrument 35 from thepatient 18. Another example of a fault condition is if therobot 15 encounters a resistance above a preprogrammed tolerance level. -
FIG. 12 shows a flow chart diagram 1200 for a closed screw/needle insertion procedure according to an embodiment of the invention is shown. In a closed pedicle screw insertion procedure, in some embodiments, therobot 15 holds aguide tube 50 adjacent to the patient 18 in the correct angular orientation at the point where a conventional pedicle screw is to be inserted through the tissue and into the bone of thepatient 18. - In some embodiments, the distance between each of the calibrating
transmitters 120 relative to each other is measured prior tocalibration step 300. In some embodiments, each calibratingtransmitter 120 transmits RF signals on a different frequency so thepositioning sensors 12 can determine whichtransmitter 120 emitted a particular RF signal. In some embodiments, the signal of each of thesetransmitters 120 is received by positioningsensors 12. Since the distance between each of the calibratingtransmitters 120 is known, and thesensors 12 can identify the signals from each of the calibratingtransmitters 120 based on the known frequency, using time of flight calculation, in some embodiments, thepositioning sensors 12 are able to calculate the spatial distance of each of thepositioning sensors 12 relative to each other. Thesystem 1 is now calibrated. As a result, in some embodiments, thepositioning sensors 12 can now determine the spatial position of anynew RF transmitter 120 introduced into theroom 10 relative to thepositioning sensors 12. - In some embodiments, at
step 310, a 3D anatomical image scan, such as a CT scan, is taken of the anatomical target. Any 3D anatomical image scan may be used with thesurgical robot 15 and is within the scope of the present invention. In some embodiments, atstep 320, the operator selects a desired trajectory and insertion point of thesurgical instrument 35 on the anatomical image captured atstep 310. In some embodiments, the desired trajectory and insertion point is programmed into thecomputer 100 so that therobot 15 can drive aguide tube 50 automatically to follow the trajectory. In some embodiments, atstep 330, the positions of theRF transmitters 120 tracking the anatomical target are read by positioningsensors 110. In some embodiments, thesetransmitters 120 identify the initial position of the anatomical target and any changes in position during the procedure. - In some embodiments, if any
RF transmitters 120 must transmit through a medium that changes the RF signal characteristics, the system will compensate for these changes when determining the transmitter's 120 position. In some embodiments, atstep 340, the positions of thetransmitters 120 on the anatomy are calibrated relative to the LPS coordinate system. In other words, the LPS provides a reference system, and the location of the anatomical target is calculated relative to the LPS coordinates. In some embodiments, to calibrate the anatomy relative to the LPS, the positions oftransmitters 120 affixed to the anatomical target are recorded at the same time as positions oftemporary transmitters 120 on precisely known anatomical landmarks also identified on the anatomical image. This calculation is performed by a computer. - In some embodiments, at
step 350, the positions of theRF transmitters 120 that track the anatomical target are read. Since the locations of thetransmitters 120 on the anatomical target have already been calibrated, in some embodiments, the system can easily determine if there has been any change in position of the anatomical target. In some embodiments, atstep 360, the positions of thetransmitters 120 on thesurgical instrument 35 are read. In some embodiments, thetransmitters 120 may be located on thesurgical instrument 35, and/or attached to various points of thesurgical robot 15. - In some embodiments, at
step 370, the coordinates of the positions of thetransmitters 120 on thesurgical instrument 35, and/or the positions read from the position encoders, are calibrated relative to the anatomical coordinate system. In other words, the position data of thesurgical instrument 35 is synchronized into the same coordinate system as the anatomy. This calculation is performed automatically by thecomputer 100 since the positions of thetransmitters 120 on the anatomical target and the positions of thetransmitters 120 on thesurgical instrument 35 are in the same coordinate system and the positions of thetransmitters 120 on the anatomical target are already calibrated relative to the anatomy. - In some embodiments, at
step 380, thecomputer 100 superimposes a representation of the location calculated instep 370 of the surgical device on the 3D anatomical image of the patient 18 taken instep 310. The superimposed image can be displayed to the user. In some embodiments, atstep 390, thecomputer 100 determines whether theguide tube 50 is in the correct orientation and position to follow the trajectory planned atstep 320. If it is not, then step 393 is reached. If it is in the correct orientation and position to follow the trajectory, then step 395 is reached. - In some embodiments, at
step 393, thecomputer 100 determines what adjustments it needs to make in order to make theguide tube 50 follow the preplanned trajectory. Thecomputer 100 sends the appropriate signals to drive themotors 160 in order to correct the movement of the guide tube. In some embodiments, atstep 395, thecomputer 100 determines whether the procedure has been completed. If the procedure has not been completed, then the process beginning atstep 350 is repeated. - In some embodiments, at any time during the procedure, certain fault conditions may cause the
computer 100 to interrupt the program and respond accordingly. For instance, if the signal from theRF transmitters 120 cannot be read, then thecomputer 100 may be programmed to stop the movement of therobot 15 or lift theguide tube 50 away from thepatient 18. Another example of a fault condition is if therobot 15 encounters a resistance above a preprogrammed tolerance level. Another example of a fault condition is if theRF transmitters 120 on the anatomical target shift so that actual and calculated positions of the anatomy no longer match. One indicator that the anatomical target location has shifted relative to thetransmitters 120 is if thecomputer 100 calculates that thesurgical instrument 35 appears to be inside bone when no drilling or penetration is actually occurring. - In some embodiments, the proper response to each condition may be programmed into the system, or a specific response may be user-initiated. For example, the
computer 100 may determine that in response to an anatomy shift, the anatomy would have to be recalibrated, and the process beginning atstep 330 should be repeated. Alternatively, a fault condition may require the flowchart to repeat fromstep 300. Another alternative is the user may decide that recalibration fromstep 330 is desired, and initiate that step himself. - Referring now to
FIG. 13 , a flow chart diagram 1300 for a safe zone surgical procedure performed using the system described herein is shown in accordance with some embodiments of the invention. In a safe zone surgical procedure, there is a defined safe zone around the surgical area within which the surgical device must stay. The physician manually controls the surgical device that is attached to the end-effectuator 30 of thesurgical robot 15. If the physician moves the surgical device outside of the safe zone, then thesurgical robot 15 stiffens thearm 23 so that the physician cannot move theinstrument 35 in any direction that would move thesurgical instrument 35 outside the safe zone. - In some embodiments, the distance between each of the calibrating
transmitters 120 relative to each other is measured prior tocalibration step 400. Each calibratingtransmitter 120 transmits RF signals on a different frequency so thepositioning sensors 12 can determine whichtransmitter 120 emitted a particular RF signal. The signal of each of thesetransmitters 120 is received by positioningsensors 12. Since the distance between each of the calibratingtransmitters 120 is known, and thesensors 12 can identify the signals from each of the calibratingtransmitters 120 based on the known frequency, thepositioning sensors 12 are able to calculate, using time of flight calculation, the spatial distance of each of thepositioning sensors 12 relative to each other. Thesystem 1 is now calibrated. As a result, thepositioning sensors 12 can now determine the spatial position of anynew RF transmitter 120 introduced into theroom 10 relative to thepositioning sensors 12. - In some embodiments, at
step 410, a 3D anatomical image scan, such as a CT scan, is taken of the anatomical target. Any 3D anatomical image scan may be used with thesurgical robot 15 and is within the scope of the present invention. In some embodiments, atstep 420, the operator inputs a desired safe zone on the anatomical image taken instep 410. In an embodiment of the invention, the operator uses an input to thecomputer 100 to draw a safe zone on a CT scan taken of the patient 18 instep 410. In some embodiments, atstep 430, the positions of theRF transmitters 120 tracking the anatomical target are read by positioning sensors. Thesetransmitters 120 identify the initial position of the anatomical target and any changes in position during the procedure. In some embodiments, if anyRF transmitters 120 must transmit through a medium that changes the RF signal characteristics, then the system will compensate for these changes when determining the transmitter's 120 position. - In some embodiments, at
step 440, the positions of thetransmitters 120 on the anatomy are calibrated relative to the LPS coordinate system. In other words, the LPS provides a reference system, and the location of the anatomical target is calculated relative to the LPS coordinates. To calibrate the anatomy relative to the LPS, the positions oftransmitters 120 affixed to the anatomical target are recorded at the same time as positions oftemporary transmitters 120 on precisely known landmarks on the anatomy that can also be identified on the anatomical image. This calculation is performed by acomputer 100. In some embodiments, atstep 450, the positions of theRF transmitters 120 that track the anatomical target are read. Since the locations of thetransmitters 120 on the anatomical target have already been calibrated, the system can easily determine if there has been any change in position of the anatomical target. - In some embodiments, at
step 460, the positions of thetransmitters 120 on thesurgical instrument 35 are read. Thetransmitters 120 may be located on thesurgical instrument 35 itself, and/or there may betransmitters 120 attached to various points of thesurgical robot 15. In some embodiments, atstep 470, the coordinates of the positions of thetransmitters 120 on thesurgical instrument 35, and/or the positions read from the position encoders, are calibrated relative to the anatomical coordinate system. In other words, the position data of thesurgical instrument 35 is synchronized into the same coordinate system as the anatomy. This calculation is performed automatically by thecomputer 100 since the positions of thetransmitters 120 on the anatomical target and the positions of thetransmitters 120 on thesurgical instrument 35 are in the same coordinate system and the positions of thetransmitters 120 on the anatomical target are already calibrated relative to the anatomy. - In some embodiments, at
step 480, thecomputer 100 superimposes a representation of the location calculated instep 470 of the surgical device on the 3D anatomical image of the patient 18 taken instep 410. In some embodiments, the superimposed image can be displayed to the user. In some embodiments, atstep 490, thecomputer 100 determines whether the surgical device attached to the end-effectuator 30 of thesurgical robot 15 is within a specified range of the safe zone boundary (for example, within 1 millimeter of reaching the safe zone boundary). In some embodiments, if the end-effectuator 30 is almost to the boundary, then step 493 is reached. In some embodiments, if it is well within the safe zone boundary, then step 495 is reached. - In some embodiments, at
step 493, thecomputer 100 stiffens the arm of thesurgical robot 15 in any direction that would allow the user to move the surgical device closer to the safe zone boundary. In some embodiments, atstep 495, thecomputer 100 determines whether the anatomy needs to be recalibrated. In some embodiments, the user may choose to recalibrate the anatomy, in which case thecomputer 100 responds to user input. Alternatively, in some embodiments, thecomputer 100 may be programmed to recalibrate the anatomy in response to certain events. For instance, in some embodiments, thecomputer 100 may be programmed to recalibrate the anatomy if theRF transmitters 120 on the anatomical target indicate that the location of the anatomical target has shifted relative to the RF transmitters 120 (i.e. this spatial relationship should be fixed.) In some embodiments, an indicator that the anatomical target location has shifted relative to thetransmitters 120 is if thecomputer 100 calculates that thesurgical instrument 35 appears to be inside bone when no drilling or penetration is actually occurring. - In some embodiments, if the anatomy needs to be calibrated, then the process beginning at
step 430 is repeated. In some embodiments, if the anatomy does not need to be recalibrated, then the process beginning atstep 450 is repeated. In some embodiments, at any time during the procedure, certain fault conditions may cause thecomputer 100 to interrupt the program and respond accordingly. For instance, in some embodiments, if the signal from theRF transmitters 120 cannot be read, then thecomputer 100 may be programmed to stop the movement of therobot 15 or remove thesurgical instrument 35 from thepatient 18. Another example of a fault condition is if therobot 15 encounters a resistance above a preprogrammed tolerance level. - Referring now to
FIG. 14 , a flow chart diagram 1400 for a conventional flexible catheter or wire insertion procedure according to an embodiment of the invention is shown. Catheters are used in a variety of medical procedures to deliver medicaments to a specific site in a patient's body. Often, delivery to a specific location is needed so a targeted diseased area can then be treated. Sometimes instead of inserting the catheter directly, a flexible wire is first inserted, over which the flexible catheter can be slid. - In some embodiments, the distance between each of the calibrating
transmitters 120 relative to each other is measured prior tocalibration step 500. In some embodiments, each calibratingtransmitter 120 transmits RF signals on a different frequency so thepositioning sensors transmitter 120 emitted a particular RF signal. In some embodiments, the signal from each of thesetransmitters 120 is received by positioningsensors transmitters 120 is known, and the sensors can identify the signals from each of the calibratingtransmitters 120 based on the known frequency, in some embodiments, using time of flight calculation, thepositioning sensors positioning sensors positioning sensors new RF transmitter 120 introduced into theroom 10 relative to thepositioning sensors - In some embodiments, at
step 510, reference needles that contain theRF transmitters 120 are inserted into the body. The purpose of these needles is to track movement of key regions of soft tissue that will deform during the procedure or with movement of thepatient 18. - In some embodiments, at
step 520, a 3D anatomical image scan (such as a CT scan) is taken of the anatomical target. Any 3D anatomical image scan may be used with thesurgical robot 15 and is within the scope of the present invention. In some embodiments, the anatomical image capture area includes the tips of the reference needles so that their transmitters' 120 positions can be determined relative to the anatomy. In some embodiments, atstep 530, the RF signals from the catheter tip and reference needles are read. - In some embodiments, at
step 540, the position of the catheter tip is calculated. Because the position of the catheter tip relative to the reference needles and the positions of the reference needles relative to the anatomy are known, thecomputer 100 can calculate the position of the catheter tip relative to the anatomy. In some embodiments, atstep 550, the superimposed catheter tip and the shaft representation is displayed on the anatomical image taken instep 520. In some embodiments, atstep 560, thecomputer 100 determines whether the catheter tip is advancing toward the anatomical target. If it is not moving to the anatomical target, then step 563 is reached. If it is correctly moving, then step 570 is reached. - In some embodiments, at
step 563, therobot 15 arm is adjusted to guide the catheter tip in the desired direction. If the anatomy needs to be calibrated, then in some embodiments, the process beginning atstep 520 is repeated. If the anatomy does not need to be recalibrated, then the process beginning atstep 540 is repeated. In some embodiments, atstep 570, thecomputer 100 determines whether the procedure has been completed. If the procedure has not been completed, then the process beginning atstep 540 is repeated. - In some embodiments, at any time during the procedure, certain fault conditions may cause the
computer 100 to interrupt the program and respond accordingly. For instance, in some embodiments, if the signal from the RF transmitter's 120 cannot be read, then thecomputer 100 may be programmed to stop the movement of therobot 15 or remove the flexible catheter from thepatient 18. Another example of a fault condition is if therobot 15 encounters a resistance above a preprogrammed tolerance level. A further example of a fault condition is if the RF transmitter's 120 on the anatomical target indicate the location of the anatomical target shift so that actual and calculated positions of the anatomy no longer match. In some embodiments, one indicator that the anatomical target location has shifted relative to the transmitter's 120 is if thecomputer 100 calculates that thesurgical instrument 35 appears to be inside bone when no drilling or penetration is actually occurring. - In some embodiments, the proper response to each condition may be programmed into the system, or a specific response may be user-initiated. For example, in some embodiments, the
computer 100 may determine that in response to an anatomy shift, the anatomy would have to be recalibrated, and the process beginning atstep 520 should be repeated. Alternatively, in some embodiments, a fault condition may require the flowchart to repeat fromstep 500. In other embodiments, the user may decide that recalibration fromstep 520 is desired, and initiate that step himself. - Referring now to
FIGS. 15A & 15B , screenshots of software for use with the described system is provided in accordance with some embodiments of the invention. The software provides the method to select the target area of surgery, plan the surgical path, check the planned trajectory of the surgical path, synchronize the medical images to the positioning system and precisely control the positioning system during surgery. The surgical positioning system and navigation software includes an optical guidance system or RF Local Positioning System (RF-LPS), which are in communication with the positioning system. -
FIG. 15A shows a screen shot 600 of the selection step for a user using a software program as described herein in accordance with some embodiments of the invention. Screen shot 600 includeswindows 615, 625, and 635, which show a 3D anatomical image of surgical target 630 on different planes. In this step, the user selects the appropriate 3D image corresponding to anatomical location of where the procedure will occur. In some embodiments, the user uses a graphic control to change the perspective of the image in order to more easily view the image from different angles. In some embodiments, the user can view the surgical target 630 within separate coordinated views for each of the x-axis, y-axis and z-axis coordinates for each anatomical location in the database in eachwindow 615, 625 and 635, respectively. - In some embodiments, after selecting the desired 3D image of the surgical target 630, the user will plan the appropriate trajectory on the selected image. In some embodiments, an input control is used with the software in order to plan the trajectory of the
surgical instrument 35. In one embodiment of the invention, the input control is in the shape of a biopsy needle 8110 for which the user can plan a trajectory. -
FIG. 15B shows a screen shot 650 during the medical procedure in accordance with some embodiments of the invention. In some embodiments, the user can still view the anatomical target 630 in different x-axis, y-axis and z-axis coordinate views onwindows 615, 625, and 635. As shown in screen shot 650, the user can see the planned trajectory line 670 inmultiple windows 615 and 625. The actual trajectory and location of thesurgical instrument 35 is superimposed on the image (shown as line segment 660). In some embodiments, the actual trajectory and location of thesurgical instrument 35 is dynamically updated and displayed, and is shown as a line segment 660. In some other embodiments, the actual trajectory and location of thesurgical instrument 35 could be shown as a trapezoid or a solid central line surrounded by a blurred or semi-transparent fringe to represent the region of uncertainty. In some embodiments (under perfect conditions with no bending of the surgical instrument as it enters tissues) thetracking system 3417 androbot 15 encoders calculate that thesurgical instrument 35 should be located at the solid line or center of the trapezoid. In some embodiments, due to bending of theinstrument 35 that might occur if tissues of different densities are crossed, there might be bending, with the amount of reasonably expected bending displayed as the edges of the trapezoid or fringe. In some embodiments, the size of this edge could be estimated knowing the stiffness and tolerance of thesurgical instrument 35 within theguide tube 50, and by using experimental data collected for thesame instrument 35 under previous controlled conditions. In some embodiments, displaying this region of uncertainty helps prevent the user from expecting the system to deliver a tool to a target trajectory with a physically impossible level of precision. - As described earlier, in some embodiments, the
surgical robot 15 can be used with alternate guidance systems other than an LPS. In some embodiments, thesurgical robot system 1 can comprise a targetingfixture 690 for use with a guidance system. In some embodiments, one targetingfixture 690 comprises acalibration frame 700, as shown inFIGS. 20A-20E . Acalibration frame 700 can be used in connection with many invasive procedures; for example, it can be used in thoracolumbar pedicle screw insertion in order to help achieve a more accurate trajectory position. In some embodiments, the use of thecalibration frame 700 can simplify the calibration procedure. In some embodiments of the invention, thecalibration frame 700 can be temporarily affixed to the skin of apatient 18 surrounding a selected site for a medical procedure, and then the medical procedure can be performed through a window defined by the calibration frame. - As shown in
FIGS. 20A and 20B , in some embodiments of the invention, thecalibration frame 700 can comprise a combination of radio-opaque markers 730 and infrared, or “active,”markers 720. In some embodiments, the radio-opaque markers 730 can be located within theCT scan region 710, and theactive markers 720 can be located outside of theCT scan region 710. In some embodiments, a surgical field 17 (i.e., the area where the invasive procedure will occur) can be located within the perimeter created by radio-opaque markers 730. In some embodiments, the actual distances of the radio-opaque 730 andactive markers 720 relative to each other can be measured from a high-precision laser scan of the calibration frame. Additionally or alternatively, in some embodiments, the actual relative distances can be measured by actively measuring the positions ofactive markers 720 while nearly simultaneously or simultaneously pointing with a pointing device, such as a conventional digitizing probe, to one or more locations on the surface of the radio-opaque markers 730. In certain embodiments, digitizing probes can compriseactive markers 720 embedded in arigid body 690 and a tip extending from the rigid body. - In some embodiments, through factory calibration or other calibration method(s), such as pivoting calibration, the location of the probe tip relative to the rigid body of the probe can be established. In some embodiments, it can then be possible to calculate the location of the probe's tip from the probe's
active markers 720. In some embodiments, for a probe with a concave tip that is calibrated as previously described, the point in space returned during operation of the probe can represent a point distal to the tip of the probe at the center of the tip's concavity. Therefore, in some embodiments, when a probe (configured with a concave tip and calibrated tomarker 730 of the same or nearly the same diameter as the targeting fixture's radio-opaque marker 730) is touched to the radio-opaque marker 730, the probe can register the center of the sphere. In some embodiments,active markers 720 can also be placed on the robot in order to monitor a position of therobot 15 andcalibration frame 700 simultaneously or nearly simultaneously. In some embodiments, thecalibration frame 700 is mounted on the patient's skin before surgery/biopsy, and will stay mounted during the entire procedure. Surgery/biopsy takes place through the center of theframe 700. - In some embodiments, when the region of the plate with the radio-
opaque markers 730 is scanned intra-operatively or prior to surgery (for example, using a CT scanner), the CT scan contains both the medical images of the patient's bony anatomy, and spherical representations of the radio-opaque markers 730. In some embodiments, software is used to determine the locations of the centers of themarkers 730 relative to the trajectories defined by the surgeon on the medical images. Because the pixel spacing of the CT scan can be conveyed within encoded headers in DICOM images, or can be otherwise available to a tracking software (for example, the robotic guidance software 3406), it can, in some embodiments, be possible to register locations of the centers of themarkers 730 in Cartesian coordinates (in millimeters, for example, or other length units). In some embodiments, it can be possible to register the Cartesian coordinates of the tip and tail of each trajectory in the same length units. - In some embodiments, because the system knows the positions of the trajectories relative to the radio-
opaque markers 730, the positions of the radio-opaque markers 730 relative to theactive markers 720, and the positions of theactive markers 720 on thecalibration frame 700 relative to the active markers on the robot 15 (not shown), the system has all information necessary to position the robot's end-effectuator 30 relative to the defined trajectories. - In some other embodiments of the invention, the
calibration frame 700 can comprise at least three radio-opaque markers 730 embedded in the periphery of thecalibration frame 700. In some embodiments, the at least three radio-opaque markers 730 can be positioned asymmetrically about the periphery of thecalibration frame 700 such that the software, as described herein, can sort the at least three radio-opaque markers 730 based only on the geometric coordinates of eachmarker 730. In some embodiments, thecalibration frame 700 can comprise at least one bank ofactive markers 720. In some embodiments, each bank of the at least one bank can comprise at least threeactive markers 720. In some embodiments, the at least one bank ofactive markers 720 can comprise four banks ofactive markers 720. In yet another aspect, thecalibration frame 700 can comprise a plurality of levelingposts 77 coupled to respective corner regions of thecalibration frame 700. In some embodiments, the corner regions of thecalibration frame 700 can include levelingposts 77 that can comprise radiolucent materials. In some embodiments, the plurality of levelingposts 77 can be configured to promote uniform, rigid contact between thecalibration frame 700 and the skin of thepatient 18. In some embodiments, a surgical-grade adhesive film, such as, for example and without limitation, Ioban™ from 3M™, can be used to temporarily adhere thecalibration frame 700 to the skin of thepatient 18. 3M™ and Ioban™ are registered trademarks of 3M Company. In some further embodiments, thecalibration frame 700 can comprise a plurality of upright posts 75 that are angled away from the frame 700 (seeFIG. 20B ). In some embodiments, the plurality ofactive markers 720 can be mounted on the plurality of upright posts 75. - As shown in
FIG. 20B , in some embodiments, there are four radio-opaque markers 730 (non-metallic BBs from an air gun) embedded in the periphery of the frame, labeled OP1, OP2, OP3, OP4. In some embodiments, only threemarkers 730 are needed for determining the orientation of a rigid body in space (the 4th marker is there for added accuracy). In some embodiments, the radio-opaque markers 730 are placed in an asymmetrical configuration (notice how OP1 and OP2 are separated from each other by more distance than OP3 and OP4, and OP1 and OP4 are aligned with each other across the gap, however OP3 is positioned more toward the center than OP2). The reason for this arrangement is so that a computer algorithm can automatically sort the markers to determine which is which if only given the raw coordinates of the four markers and not their identification. - In some embodiments, there are four banks of active markers 720 (three
markers 720 per bank). Only one bank of threemarkers 720 is needed (redundancy is for added accuracy and so that the system will still work if the surgeon, tools, or robot are blocking some of the markers. In some embodiments, despite the horizontal orientation of thepatient 18, the angulation of the upright posts can permit theactive markers 720 to face toward the cameras or detection devices of the tracking system (for example, the tracking system 3417). In some embodiments, the upright posts can be angled away from the calibration frame by about 10°. - In some applications, to establish the spatial relationship between the active 720 and radio-
opaque markers 730, a conventional digitizing probe, such as a 6-marker probe, embedded withactive markers 720 in a known relationship to the probe's tip (see for exampleFIG. 20C ) can be used to point to each of the radio-opaque markers 730. In some embodiments, the probe can point to locations on two opposite surfaces of the spherical radio-opaque markers 730 while recording the position of the probe tip and theactive markers 720 on theframe 700 simultaneously. Then, the average position of the two surface coordinates can be taken, corresponding to the center of the sphere. An image of therobot 15 used with this targetingfixture 690 is shown inFIG. 20D . For placement of conventional surgical screws, a biopsy, injection, or other procedures, in some embodiments, therobot 15 can work through the window formed by theframe 700. During a surgical procedure, in some embodiments, the working portal is kept on the interior of theframe 700 and themarkers 720 on the exterior of theframe 700 can improve accuracy over a system where fiducials are mounted away from the area where surgery is being performed. Without wishing to be bound by theory, simulation, and/or modeling, it is believed that a reason for improved accuracy is that optimal accuracy of trackingmarkers 720 can be achieved if trackingmarkers 720 are placed around the perimeter of theframe 700 being tracked. - Further embodiments of the invention are shown in
FIG. 20E illustrating acalibration frame 700. Thisfixture 690 is simplified to make it less obstructive to the surgeon. In some embodiments, thecalibration frame 700 can comprise fouractive markers 720 having a lower profile than theactive markers 720 described above and depicted inFIGS. 20A-20D . For example, thecalibration frame 700 can comprise a plurality of upright posts 75 that are angled away from the calibration frame by about 10°. In some embodiments, theactive markers 720 are mounted on the posts 75 that are angled back by 10°, and this angulation keeps themarkers 720 facing toward the cameras despite the patient being horizontal. - Moreover, in some embodiments, the
front markers 720 can have less chance of obscuring therear markers 720. For example, posts 75 that are farthest away from the camera or farthest from a detection device of thetracking system 3417 can be taller and spaced farther laterally than the posts 75 closest to the camera. In some further embodiments of the invention, thecalibration frame 700 can comprisemarkers 730 that are both radio-opaque for detection by a medical imaging scanner, and visible by the cameras or otherwise detectable by the real-time tracking system 3417. In some embodiments, the relationship between radio-opaque 730 and active markers (730, 720) does not need to be measured or established because they are one in the same. Therefore, in some embodiments, as soon as the position is determined from the CT scan (or other imaging scan), the spatial relationship between therobot 15 and anatomy of the patient 18 can be defined. - In other embodiments, the targeting
fixture 690 can comprise a flexible roll configuration. In some embodiments, the targetingfixture 690 can comprise three or more radio-opaque markers 730 that define a rigid outer frame and nine or moreactive markers 720 embedded in a flexible roll of material (for example, theflexible roll 705 inFIG. 21A ). As described earlier, radio-opaque markers 730 are visible on CT scans and/or other medical diagnostic images, such as MRI, or reconstructions from O-arm or Iso-C scans, and their centroids can be determined from the 3D image.Active markers 720 include trackedmarkers 720 that have 3D coordinates that are detectable in real-time using cameras or other means. Some embodiments can utilize active marker systems based on reflective optical systems such as Motion Analysis Inc., or Peak Performance. Other suitable technologies include infrared-emitting marker systems such as Optotrak, electromagnetic systems such as Medtronic's Axiem®, or Flock of Birds®, or a local positioning system (“LPS”) described by Smith et al. in U.S. Patent Publication No. 2007/0238985. Flock Of Birds® is a registered trademark of Ascension Technology Corporation. Axiem is a trademark of Medtronic, Inc., and its affiliated companies. Medtronic® is a registered trademark used for Surgical and Medical Apparatus, Appliances and Instruments. - In some embodiments of the invention, at least a portion of the
flexible roll 705 can comprise self-adhering film, such as, for example and without limitation, 3M™Ioban™ adhesive film (iodine-impregnated transparent surgical drape) similar to routinely used operating room product model 6651 EZ (3M, St. Paul, Minn.). Ioban™ is a trademark of 3M company. In some embodiments, within theflexible roll 705, the radio-opaque and active markers (730, 720) can be rigidly coupled to each other, with each radio-opaque marker 730 coupled to three or moreactive markers 720. Alternatively, in some embodiments, the markers can simultaneously serve as radio-opaque and active markers (for example, anactive marker 720 whose position can be detected from cameras or other sensors), and the position determined from the 3D medical image can substantially exactly correspond to the center of themarker 720. In some embodiments, as few as threesuch markers 720 could be embedded in theflexible roll 705 and still permit determination of the spatial relationship between therobot 15 and the anatomy of thepatient 18. If radio-opaque markers 730 andactive markers 720 are not one in the same, in some embodiments the at least threeactive markers 720 must be rigidly connected to each radio-opaque marker 730 because three separate non-collinear points are needed to unambiguously define the relative positions of points on a rigid body. That is, if only one or 2active markers 720 are viewed, there is more than one possible calculated position where a rigidly coupled radio-opaque marker could be. - In some embodiments of the invention, other considerations can be used to permit the use of two
active markers 720 per radio-opaque marker 730. For example, in some embodiments, if twoactive markers 720 and one radio-opaque marker 730 are intentionally positioned collinearly, with the radio-opaque marker 730 exactly at the midpoint between the twoactive markers 720, the location of the radio-opaque marker 730 can be determined as the mean location of the twoactive markers 720. Alternatively, in some embodiments, if the twoactive markers 720 and the radio-opaque marker 730 are intentionally positioned collinearly but with the radio-opaque marker 730 closer to oneactive marker 720 than the other (see for exampleFIG. 21B ), then in some embodiments, the radio-opaque marker 730 must be at one of two possible positions along the line in space formed by the two active markers 720 (seeFIG. 21B ). In this case, in some embodiments, if theflexible roll 705 is configured so that each pair ofactive markers 720 is oriented (when in its final position) with onemarker 720 more toward the center of theflexible roll 705, then it can be determined from the orientations of all markers or certain combinations of markers from different regions which of the two possible positions within each region is the correct position for the radio-opaque marker 730 (seeFIG. 21C showingflexible roll 705 showed rolled on a torso and shown unrolled on a torso withmarkers opaque markers 730 can be positioned toward the inside of the frame 705), with marker groups nearer to the top of the figure having the radio-opaque marker 730 positioned below theactive markers 720 and marker groups near the bottom of the figure having the radio-opaque marker positioned above theactive markers 720. - In some embodiments, the
flexible roll 705 can be positioned across the patient's back or other area, and adhered to the skin of the patient 18 as it is unrolled. In some embodiments, knowing the spatial relationship between each triad ofactive markers 720 and the rigidly coupled radio-opaque marker 730, it is possible to establish the relationship between the robot 15 (position established by its own active markers 720) and the anatomy (visualized together with radio-opaque markers 730 on MM, CT, or other 3D scan). In some embodiments, theflexible roll 705 can be completely disposable. Alternatively, in some other embodiments, theflexible roll 705 can comprise reusable marker groups integrated with a disposable roll with medical grade adhesive on each side to adhere to thepatient 18 and themarker groups flexible roll 705 can comprise a drape incorporated into theflexible roll 705 for covering thepatient 18, with the drape configured to fold outwardly from theroll 705. - In some embodiments, after the
roll 705 has been unrolled, theroll 705 can have a desired stiffness such that theroll 705 does not substantially change its position relative to the bony anatomy of thepatient 18. In some embodiments of the invention, a conventional radiolucent wire can be embedded in the perimeter of theframe 700. In some embodiments, it a chain of plastic beads, such as the commercially available tripods shown inFIG. 21D , or a commercially available “snake light” type fixture, can be employed to provide desired stiffness to the unrolled fixture such that it maintains its position after unrolling occurs. For example, in some embodiments, the beads of the chain of plastic beads as shown can be affixed to each other with a high friction so that they hold their position once shifted. Further, in some embodiments, chains of beads can be incorporated into, and define, a perimeter of theframe 700. In some embodiments, this type of frame could be loaded with conventional chemicals that mix at the time of application. For example, in some embodiments, components of a conventional two-part epoxy could be held in separate fragile baggies within the frame that pop open when the user first starts to manipulate the beads. In some embodiments, the user would attach the frame to thepatient 18, and mold it to the contours of the patient's body. After a short period of time, theframe 700 would solidify to form a very rigid frame, locking the beads in their current orientation. - In some embodiments of the invention, the targeting
fixture 690 can be an adherable fixture, configured for temporary attachment to the skin of apatient 18. For example, in some embodiments, the targetingfixture 690 can be temporarily adhered to the patient 18 during imaging, removed, and then subsequently reattached during a follow-up medical procedure, such as a surgery. In some embodiments, the targetingfixture 690 can be applied to the skull of apatient 18 for use in placement of electrodes for deep brain stimulation. In some embodiments, this method can use asingle fixture 690, or two related fixtures. In this instance, the two related fixtures can share the same surface shape. However, onefixture 690 can be temporarily attached at the time of medical image scanning, and can include radio-opaque markers 730 (but not active markers 720), and thesecond fixture 690 can be attached at the time of surgery, and can include active markers 720 (but not radio-opaque markers 730). - In some embodiments, the first fixture (for scanning) can comprise a
frame 690 with three or more embedded radio-opaque markers 730, and two or more openings 740 for application of markings (the markings shown as 750 inFIG. 22B ). In some embodiments, thedevice 690 can be adhered to the scalp of apatient 18, and the openings 740 can be used to paint marks on the scalp with, for example, henna or dye (shown as “+” marks 750 inFIG. 22B ). With thisfixture 690 in place, in some embodiments, the patient 18 can receive a 3D scan (for example an MRI or CT scan) in which the radio-opaque markers 730 are captured. As illustrated byFIG. 22B , in some embodiments, thefixture 690 can then be removed, leaving the dye marks 750 on the scalp. In some embodiments, on a later date (before the dye marks 750 wear off), the patient 18 can return, and the surgeon or technician can attach the 2nd fixture (for surgery) containingactive markers 720 for intraoperative tracking (seeFIG. 22C-22D ). In some embodiments, thefixture 690 shown inFIG. 22C can be mounted to the scalp, spatially positioned and oriented in the same position as the previously adhered first fixture (shown inFIG. 22A ) by ensuring that the previously placed dye marks 750 line up with holes in the second fixture 690 (see the alignment arrows depicted inFIG. 22C ). Optionally, in some embodiments, thefixture 690 can have a transparent frame for good visualization. The above-described method assumes that the locations of the marks 750 do not change over the period of time between the scan and the return of thepatient 18 for surgery. Since the relative positions between the radio-opaque markers 730 from the temporary (first) fixture 690 (which appear in the scan) and theactive markers 720 on the second appliedfixture 690 are known through a calibration and/or by careful manufacturing of thefixtures 690, the coordinate system of the anatomy and the coordinate system of theactive markers 720 can be synchronized so that therobot 15 can target any planned trajectory on the 3D image as described further herein. Further, in some embodiments, this method can enable image guidance with only one pre-op scan and without requiring the patient 18 to go home after a pre-op scan. This circumvents the need for a patient 18 to take care of wounds from targeting screws that are invasively drilled into the skull of thepatient 18. - In some embodiments of the invention, the targeting
fixture 690 can comprise a conventional clamping mechanism for securely attaching the targetingfixture 690 to thepatient 18. For example, in some embodiments, the targetingfixture 690 can be configured to clamp to the spinous process 6301 of a patient 18 after the surgeon has surgically exposed the spinous process.FIG. 23 shows adynamic tracking device 2300 mounted to thespinous process 2310 in the lumbar spine of a patient 18 in accordance with some embodiments of the invention. This targeting fixture is used with Medtronic's StealthStation. This figure is reprinted from Bartolomei J, Henn J S, Lemole G M Jr., Lynch J, Dickman C A, Sonntag V K H, Application of frameless stereotaxy to spinal surgery, Barrow Quarterly 17(1), 35-43 (2001). StealthStation® is a trademark of Medtronic, Inc., and its affiliated companies. - In some embodiments, during use of a targeting
fixture 690 having a conventional clamping mechanism with image guidance, the relationship between themarkers markers tracking system 3417 to calculate the coordinates of the probe tip relative to themarkers - In some embodiments, the clamping mechanism of the targeting
fixture 690 can be configured for clamping to thespinous process 2310, or can be configured for anchoring to bone of the patient 18 such that thefixture 690 is substantially stationary and not easily moved. In some further embodiments, the targetingfixture 690 can comprise at least threeactive markers 720 and distinct radio-opaque markers 730 that are detected on the CT or other 3D image, preferably near the clamp (to be close to bone). In some alternative embodiments, theactive markers 720 themselves must be configured to be visualized accurately on CT or other 3D image. In certain embodiments, the portion of thefixture 690 containing a radio-opaque marker 730 can be made to be detachable to enable removal from the fixture after the 3D image is obtained. In some further embodiments, a combination of radio-opaque 730 andactive markers 720 can allow tracking with therobot 15 in the same way that is possible with the frame-type targeting fixtures 690 described above. - In some embodiments, one aspect of the software and/or firmware disclosed herein is a unique process for locating the center of the above-described
markers 730 that takes advantage of the fact that a CT scan can comprise slices, typically spaced 1.5 mm or more apart in the z direction, and sampled with about 0.3 mm resolution in the x-axis and y-axis directions. In some embodiments, since the diameter of the radio-opaque markers 730 is several times larger than this slice spacing, different z slices of the sphere will appear as circles of different diameters on each successive x-y planar slice. In some embodiments, since the diameter of the sphere is defined beforehand, the necessary z position of the center of the sphere relative to the slices can be calculated to provide the given set of circles of various diameters. Stated similarly, in some embodiments, a z slice substantially exactly through the center of the sphere can yield a circle with a radius R that is substantially the same as that of the sphere. In some embodiments, a z slice through a point at the top or bottom of the sphere can yield a circle with a radius R approximating zero. In some other embodiments, a z slice through a z-axis coordinate Z1 between the center and top or bottom of the sphere can yield a circle with a radius R1=R cos(arcsin(Z1/R)). - In some embodiments of the invention, the observed radii of circles on z slices of known inter-slice spacing can be analyzed using the equation defined by R1=R cos(arcsin(Z1/R)). This provides a unique mathematical solution permitting the determination of the distance of each slice away from the center of the sphere. In cases in which a sphere has a diameter small enough that only a few slices through the sphere appear on a medical image, this process can provide a more precise the center of a sphere.
- Some embodiments of the use of the
calibration frame 700 are described to further clarify the methods of use. For example, some embodiments include the steps of a conventional closed screw or conventional needle (for example, a biopsy needle 8110) insertion procedure utilizing acalibration frame 700 as follows. In some embodiments, acalibration frame 700 is attached to the patient's 18 skin, substantially within the region at which surgery/biopsy is to take place. In some embodiments, thepatient 18 receives a CT scan either supine or prone, whichever positioning orients thecalibration frame 700 upward. In some embodiments, the surgeon subsequently manipulates three planar views of the patient's 18 CT images with rotations and translations. In some embodiments, the surgeon then draws trajectories on the images that define the desired position, and strike angle of the end-effectuator 30. In some embodiments, automatic calibration can be performed in order to obtain the centers of radio-opaque makers 730 of thecalibration frame 700, and to utilize the stored relationship between theactive markers 720 and radio-opaque markers 730. This procedure permits therobot 15 to move in the coordinate system of the anatomy and/or drawn trajectories. - In some embodiments, the
robot 15 then will move to the desired position. In some embodiments, if forceful resistance beyond a pre-set tolerance is exceeded, therobot 15 will halt. In some further embodiments, therobot 15 can hold theguide tube 50 at the desired position and strike angle to allow the surgeon to insert a conventional screw or needle (for example, needle 7405, 7410 or biopsy needle 8110). In some embodiments, if tissues move in response to applied force or due to breathing, the movement will be tracked byoptical markers 720, and the robot's position will automatically be adjusted. - As a further illustration of a procedure using an alternate guidance system, in some embodiments, the steps of an open screw insertion procedure utilizing an optical guidance system is described. In some embodiments, after surgical exposure, a targeting
fixture 690 comprising a small tree of optical markers, for example, can be attached to a bony prominence in the area of interest. In some embodiments, conventional calibration procedures for image guidance can be utilized to establish the anatomy relative to theoptical tracking system 3417 and medical images. For another example, the targetingfixture 690 can contain rigidly mounted, substantially permanent or detachable radio-opaque markers 730 that can be imaged with a CT scan. In some embodiments, the calibration procedures consistent with those stated for thecalibration frame 700 can be utilized to establish the anatomy relative to therobot 15 and the medical image. - In some embodiments, the surgeon manipulates three planar views of the patient's CT images with rotations and translations. In some embodiments, the surgeon then draws trajectories on the images that define the desired position and strike angle of the end-
effectuator 30. In some embodiments, therobot 15 moves to the desired position. In some embodiments, if forceful resistance beyond a pre-set tolerance is exceeded, therobot 15 will halt. In some embodiments, therobot 15 holds theguide tube 50 at the desired position and strike angle to allow the surgeon to insert a conventional screw. In some embodiments, if tissues move in response to applied force or due to breathing, the movement will be tracked byoptical markers 720, and the robot's position will automatically be adjusted. - Although several embodiments of the invention have been disclosed in the foregoing specification, it is understood that many modifications and other embodiments of the invention will come to mind to which the invention pertains, having the benefit of the teaching presented in the foregoing description and associated drawings. It is thus understood that the invention is not limited to the specific embodiments disclosed hereinabove, and that many modifications and other embodiments are intended to be included within the scope of the appended claims. Moreover, although specific terms are employed herein, as well as in the claims which follow, they are used only in a generic and descriptive sense, and not for the purposes of limiting the described invention, nor the claims which follow.
- It will be appreciated by those skilled in the art that while the invention has been described above in connection with particular embodiments and examples, the invention is not necessarily so limited, and that numerous other embodiments, examples, uses, modifications and departures from the embodiments, examples and uses are intended to be encompassed by the claims attached hereto. The entire disclosure of each patent and publication cited herein is incorporated by reference, as if each such patent or publication were individually incorporated by reference herein. Various features and advantages of the invention are set forth in the following claims.
Claims (20)
1. A method of planning a surgery using a surgical robot system, said method comprising:
planning a desired trajectory for a surgical instrument to reach a desired anatomical target; and
moving the surgical robot system to a position corresponding to the desired trajectory for the surgical instrument,
wherein the surgical robot system includes:
a surgical robot having a controllable robot arm, the robot arm having an end effectuator comprising a guide tube;
a surgical instrument configured to be positioned through the guide tube and configured to be advanced into tissue of a patient; and
a stop mechanism configured to prevent the surgical instrument from advancing through the guide tube when the surgical instrument reaches a predetermined amount of protrusion.
2. The method of claim 1 , wherein the surgical robot is able to determine a maximum protrusion distance past an end of the guide tube that the surgical instrument is able to protrude.
3. The method of claim 2 , wherein the maximum protrusion distance is determined from known lengths of the guide tube and the surgical instrument, a known distance between respective ends of the surgical instrument, and a known location where the stop mechanism is attached.
4. The method of claim 2 , wherein the maximum protrusion distance of the surgical instrument is monitored.
5. The method of claim 1 , wherein an actual protrusion distance of the surgical instrument is monitored during insertion.
6. The method of claim 5 , wherein the actual protrusion distance of the surgical instrument is substantially continuously monitored and is displayed on a display.
7. The method of claim 5 , wherein the actual protrusion distance is monitored with a spring-loaded plunger including a spring-loaded mechanism and sensor pad having a coupled wiper.
8. The method of claim 7 , wherein the stop mechanism is configured to contact the spring-loaded mechanism before it encounters the end of the guide tube.
9. The method of claim 7 , wherein, when the wiper moves across the sensor pad, a linear position of the wiper is sampled, thereby permitting calculation of the actual protrusion distance.
10. The method of claim 1 , wherein the surgical instrument is a drill bit and the drill bit includes the stop mechanism.
11. The method of claim 10 , wherein the stop mechanism on the drill bit is manually adjustable with reference to markings on the drill bit.
12. The method of claim 10 , wherein the drill bit includes release mechanisms on each end of the stop mechanism.
13. The method of claim 10 , wherein the drill bit includes a locking mechanism configured to lock and hold the drill bit in a set position relative to the guide tube.
14. The method of claim 13 , wherein the locking mechanism includes two clam shells, and the drill bit is locked into position by closing the clam shells around the stop mechanism.
15. A method of planning a surgery using a surgical robot system, said method comprising:
planning a desired trajectory for a surgical instrument to reach a desired anatomical target; and
moving the surgical robot system to a position corresponding to the desired trajectory for the surgical instrument,
wherein the surgical robot system includes:
a surgical robot having a controllable robot arm, the robot arm having an end effectuator comprising a guide tube; and
a surgical instrument configured to slide through the guide tube, wherein the surgical instrument or the guide tube includes a stop mechanism to prevent the surgical instrument from advancing through the guide tube at a predetermined location.
16. The method of claim 15 , wherein the surgical robot is able to determine a maximum protrusion distance past an end of the guide tube that the surgical instrument is able to protrude.
17. The method of claim 16 , wherein the maximum protrusion distance is determined from known lengths of the guide tube and the surgical instrument, a known distance between respective ends of the surgical instrument, and a known location where the stop mechanism is attached.
18. The method of claim 16 , wherein the maximum protrusion distance of the surgical instrument is monitored.
19. The method of claim 15 , wherein an actual protrusion distance of the surgical instrument is monitored during insertion.
20. The method of claim 19 , wherein the actual protrusion distance of the surgical instrument is substantially continuously monitored and is displayed on a display.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/520,196 US20230021298A9 (en) | 2012-06-21 | 2021-11-05 | Surgical robot platform |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261662702P | 2012-06-21 | 2012-06-21 | |
US201361800527P | 2013-03-15 | 2013-03-15 | |
US13/924,505 US9782229B2 (en) | 2007-02-16 | 2013-06-21 | Surgical robot platform |
US15/609,305 US11191598B2 (en) | 2012-06-21 | 2017-05-31 | Surgical robot platform |
US17/520,196 US20230021298A9 (en) | 2012-06-21 | 2021-11-05 | Surgical robot platform |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/609,305 Continuation US11191598B2 (en) | 2006-02-16 | 2017-05-31 | Surgical robot platform |
Publications (2)
Publication Number | Publication Date |
---|---|
US20220054206A1 true US20220054206A1 (en) | 2022-02-24 |
US20230021298A9 US20230021298A9 (en) | 2023-01-19 |
Family
ID=49769457
Family Applications (24)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/924,505 Active 2027-08-27 US9782229B2 (en) | 2006-02-16 | 2013-06-21 | Surgical robot platform |
US14/476,101 Active 2036-02-21 US10531927B2 (en) | 2012-06-21 | 2014-09-03 | Methods for performing invasive medical procedures using a surgical robot |
US15/449,260 Active 2036-02-18 US11135022B2 (en) | 2012-06-21 | 2017-03-03 | Surgical robot platform |
US15/448,830 Abandoned US20170239002A1 (en) | 2012-06-21 | 2017-03-03 | Surgical robot platform |
US15/454,379 Active 2036-03-25 US10912617B2 (en) | 2012-06-21 | 2017-03-09 | Surgical robot platform |
US15/455,662 Active 2036-06-20 US11103317B2 (en) | 2012-06-21 | 2017-03-10 | Surgical robot platform |
US15/460,974 Active 2033-11-20 US11284949B2 (en) | 2012-06-21 | 2017-03-16 | Surgical robot platform |
US15/462,280 Active US10485617B2 (en) | 2012-06-21 | 2017-03-17 | Surgical robot platform |
US15/466,937 Active 2033-10-18 US11331153B2 (en) | 2012-06-21 | 2017-03-23 | Surgical robot platform |
US15/467,005 Active 2036-03-15 US11684431B2 (en) | 2012-06-21 | 2017-03-23 | Surgical robot platform |
US15/473,698 Active 2036-03-29 US11026756B2 (en) | 2012-06-21 | 2017-03-30 | Surgical robot platform |
US15/609,322 Pending US20170265949A1 (en) | 2012-06-21 | 2017-05-31 | Surgical robot platform |
US15/609,305 Active 2034-02-16 US11191598B2 (en) | 2006-02-16 | 2017-05-31 | Surgical robot platform |
US15/648,621 Active US10835326B2 (en) | 2012-06-21 | 2017-07-13 | Surgical robot platform |
US15/697,943 Abandoned US20170360517A1 (en) | 2006-02-16 | 2017-09-07 | Surgical robot platform |
US15/704,636 Active US10835328B2 (en) | 2012-06-21 | 2017-09-14 | Surgical robot platform |
US16/267,833 Pending US20190167362A1 (en) | 2006-02-16 | 2019-02-05 | Surgical robot platform |
US16/595,578 Pending US20200155243A1 (en) | 2012-06-21 | 2019-10-08 | Surgical robot platform |
US16/708,545 Active 2034-09-28 US11690687B2 (en) | 2012-06-21 | 2019-12-10 | Methods for performing medical procedures using a surgical robot |
US17/071,307 Pending US20210022814A1 (en) | 2012-06-21 | 2020-10-15 | Surgical robot platform |
US17/520,196 Pending US20230021298A9 (en) | 2012-06-21 | 2021-11-05 | Surgical robot platform |
US17/652,723 Pending US20220409306A1 (en) | 2012-06-21 | 2022-02-28 | Surgical robot platform |
US17/724,092 Pending US20220233262A1 (en) | 2012-06-21 | 2022-04-19 | Surgical robot platform |
US17/727,013 Pending US20220241037A1 (en) | 2012-06-21 | 2022-04-22 | Surgical robot platform |
Family Applications Before (20)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/924,505 Active 2027-08-27 US9782229B2 (en) | 2006-02-16 | 2013-06-21 | Surgical robot platform |
US14/476,101 Active 2036-02-21 US10531927B2 (en) | 2012-06-21 | 2014-09-03 | Methods for performing invasive medical procedures using a surgical robot |
US15/449,260 Active 2036-02-18 US11135022B2 (en) | 2012-06-21 | 2017-03-03 | Surgical robot platform |
US15/448,830 Abandoned US20170239002A1 (en) | 2012-06-21 | 2017-03-03 | Surgical robot platform |
US15/454,379 Active 2036-03-25 US10912617B2 (en) | 2012-06-21 | 2017-03-09 | Surgical robot platform |
US15/455,662 Active 2036-06-20 US11103317B2 (en) | 2012-06-21 | 2017-03-10 | Surgical robot platform |
US15/460,974 Active 2033-11-20 US11284949B2 (en) | 2012-06-21 | 2017-03-16 | Surgical robot platform |
US15/462,280 Active US10485617B2 (en) | 2012-06-21 | 2017-03-17 | Surgical robot platform |
US15/466,937 Active 2033-10-18 US11331153B2 (en) | 2012-06-21 | 2017-03-23 | Surgical robot platform |
US15/467,005 Active 2036-03-15 US11684431B2 (en) | 2012-06-21 | 2017-03-23 | Surgical robot platform |
US15/473,698 Active 2036-03-29 US11026756B2 (en) | 2012-06-21 | 2017-03-30 | Surgical robot platform |
US15/609,322 Pending US20170265949A1 (en) | 2012-06-21 | 2017-05-31 | Surgical robot platform |
US15/609,305 Active 2034-02-16 US11191598B2 (en) | 2006-02-16 | 2017-05-31 | Surgical robot platform |
US15/648,621 Active US10835326B2 (en) | 2012-06-21 | 2017-07-13 | Surgical robot platform |
US15/697,943 Abandoned US20170360517A1 (en) | 2006-02-16 | 2017-09-07 | Surgical robot platform |
US15/704,636 Active US10835328B2 (en) | 2012-06-21 | 2017-09-14 | Surgical robot platform |
US16/267,833 Pending US20190167362A1 (en) | 2006-02-16 | 2019-02-05 | Surgical robot platform |
US16/595,578 Pending US20200155243A1 (en) | 2012-06-21 | 2019-10-08 | Surgical robot platform |
US16/708,545 Active 2034-09-28 US11690687B2 (en) | 2012-06-21 | 2019-12-10 | Methods for performing medical procedures using a surgical robot |
US17/071,307 Pending US20210022814A1 (en) | 2012-06-21 | 2020-10-15 | Surgical robot platform |
Family Applications After (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/652,723 Pending US20220409306A1 (en) | 2012-06-21 | 2022-02-28 | Surgical robot platform |
US17/724,092 Pending US20220233262A1 (en) | 2012-06-21 | 2022-04-19 | Surgical robot platform |
US17/727,013 Pending US20220241037A1 (en) | 2012-06-21 | 2022-04-22 | Surgical robot platform |
Country Status (4)
Country | Link |
---|---|
US (24) | US9782229B2 (en) |
EP (1) | EP2863827B1 (en) |
JP (1) | JP2015528713A (en) |
WO (1) | WO2013192598A1 (en) |
Families Citing this family (466)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8010180B2 (en) * | 2002-03-06 | 2011-08-30 | Mako Surgical Corp. | Haptic guidance system and method |
US9155544B2 (en) | 2002-03-20 | 2015-10-13 | P Tech, Llc | Robotic systems and methods |
US8361067B2 (en) | 2002-09-30 | 2013-01-29 | Relievant Medsystems, Inc. | Methods of therapeutically heating a vertebral body to treat back pain |
US8219178B2 (en) * | 2007-02-16 | 2012-07-10 | Catholic Healthcare West | Method and system for performing invasive medical procedures using a surgical robot |
US10357184B2 (en) | 2012-06-21 | 2019-07-23 | Globus Medical, Inc. | Surgical tool systems and method |
US10653497B2 (en) | 2006-02-16 | 2020-05-19 | Globus Medical, Inc. | Surgical tool systems and methods |
US10893912B2 (en) * | 2006-02-16 | 2021-01-19 | Globus Medical Inc. | Surgical tool systems and methods |
US9204923B2 (en) | 2008-07-16 | 2015-12-08 | Intuitive Surgical Operations, Inc. | Medical instrument electronically energized using drive cables |
US8672837B2 (en) | 2010-06-24 | 2014-03-18 | Hansen Medical, Inc. | Methods and devices for controlling a shapeable medical device |
US9119655B2 (en) * | 2012-08-03 | 2015-09-01 | Stryker Corporation | Surgical manipulator capable of controlling a surgical instrument in multiple modes |
US9921712B2 (en) | 2010-12-29 | 2018-03-20 | Mako Surgical Corp. | System and method for providing substantially stable control of a surgical tool |
US9308050B2 (en) | 2011-04-01 | 2016-04-12 | Ecole Polytechnique Federale De Lausanne (Epfl) | Robotic system and method for spinal and other surgeries |
US20150127154A1 (en) * | 2011-06-02 | 2015-05-07 | Brain Corporation | Reduced degree of freedom robotic controller apparatus and methods |
US9566710B2 (en) | 2011-06-02 | 2017-02-14 | Brain Corporation | Apparatus and methods for operating robotic devices using selective state space training |
CA2841459C (en) | 2011-07-11 | 2020-07-28 | Board Of Regents Of The University Of Nebraska | Robotic surgical devices, systems, and related methods |
EP2750620B1 (en) * | 2011-09-02 | 2017-04-26 | Stryker Corporation | Surgical instrument including a cutting accessory extending from a housing and actuators that establish the position of the cutting accessory relative to the housing |
US20130072982A1 (en) * | 2011-09-19 | 2013-03-21 | Peter Melott Simonson | Spinal assembly load gauge |
US9572682B2 (en) * | 2011-09-29 | 2017-02-21 | Arthromeda, Inc. | System and method for precise prosthesis positioning in hip arthroplasty |
JP5893330B2 (en) * | 2011-10-18 | 2016-03-23 | オリンパス株式会社 | Operation input device and method for initializing operation input device |
US10390877B2 (en) | 2011-12-30 | 2019-08-27 | Relievant Medsystems, Inc. | Systems and methods for treating back pain |
EP4357083A2 (en) | 2012-05-01 | 2024-04-24 | Board of Regents of the University of Nebraska | Single site robotic device and related systems and methods |
US11135026B2 (en) | 2012-05-11 | 2021-10-05 | Peter L. Bono | Robotic surgical system |
EP2861173B1 (en) * | 2012-06-19 | 2016-04-06 | Brainlab AG | Method and apparatus for detecting undesirable rotation of medical markers |
US10799298B2 (en) | 2012-06-21 | 2020-10-13 | Globus Medical Inc. | Robotic fluoroscopic navigation |
US11857149B2 (en) | 2012-06-21 | 2024-01-02 | Globus Medical, Inc. | Surgical robotic systems with target trajectory deviation monitoring and related methods |
US11589771B2 (en) | 2012-06-21 | 2023-02-28 | Globus Medical Inc. | Method for recording probe movement and determining an extent of matter removed |
US10136954B2 (en) | 2012-06-21 | 2018-11-27 | Globus Medical, Inc. | Surgical tool systems and method |
US11116576B2 (en) * | 2012-06-21 | 2021-09-14 | Globus Medical Inc. | Dynamic reference arrays and methods of use |
US10758315B2 (en) | 2012-06-21 | 2020-09-01 | Globus Medical Inc. | Method and system for improving 2D-3D registration convergence |
US11317971B2 (en) | 2012-06-21 | 2022-05-03 | Globus Medical, Inc. | Systems and methods related to robotic guidance in surgery |
US11607149B2 (en) | 2012-06-21 | 2023-03-21 | Globus Medical Inc. | Surgical tool systems and method |
US20170172669A1 (en) * | 2012-06-21 | 2017-06-22 | Globus Medical, Inc. | System and method for a surveillance marker in robotic-assisted surgery |
US10646280B2 (en) | 2012-06-21 | 2020-05-12 | Globus Medical, Inc. | System and method for surgical tool insertion using multiaxis force and moment feedback |
US11786324B2 (en) * | 2012-06-21 | 2023-10-17 | Globus Medical, Inc. | Surgical robotic automation with tracking markers |
US11253327B2 (en) | 2012-06-21 | 2022-02-22 | Globus Medical, Inc. | Systems and methods for automatically changing an end-effector on a surgical robot |
US11896446B2 (en) | 2012-06-21 | 2024-02-13 | Globus Medical, Inc | Surgical robotic automation with tracking markers |
US10350013B2 (en) | 2012-06-21 | 2019-07-16 | Globus Medical, Inc. | Surgical tool systems and methods |
US11857266B2 (en) | 2012-06-21 | 2024-01-02 | Globus Medical, Inc. | System for a surveillance marker in robotic-assisted surgery |
US20170258535A1 (en) * | 2012-06-21 | 2017-09-14 | Globus Medical, Inc. | Surgical robotic automation with tracking markers |
US11399900B2 (en) | 2012-06-21 | 2022-08-02 | Globus Medical, Inc. | Robotic systems providing co-registration using natural fiducials and related methods |
US10842461B2 (en) | 2012-06-21 | 2020-11-24 | Globus Medical, Inc. | Systems and methods of checking registrations for surgical systems |
US10624710B2 (en) | 2012-06-21 | 2020-04-21 | Globus Medical, Inc. | System and method for measuring depth of instrumentation |
US11864745B2 (en) | 2012-06-21 | 2024-01-09 | Globus Medical, Inc. | Surgical robotic system with retractor |
US11298196B2 (en) | 2012-06-21 | 2022-04-12 | Globus Medical Inc. | Surgical robotic automation with tracking markers and controlled tool advancement |
US10874466B2 (en) | 2012-06-21 | 2020-12-29 | Globus Medical, Inc. | System and method for surgical tool insertion using multiaxis force and moment feedback |
US11045267B2 (en) | 2012-06-21 | 2021-06-29 | Globus Medical, Inc. | Surgical robotic automation with tracking markers |
US11864839B2 (en) | 2012-06-21 | 2024-01-09 | Globus Medical Inc. | Methods of adjusting a virtual implant and related surgical navigation systems |
US11395706B2 (en) | 2012-06-21 | 2022-07-26 | Globus Medical Inc. | Surgical robot platform |
JP2015528713A (en) * | 2012-06-21 | 2015-10-01 | グローバス メディカル インコーポレイティッド | Surgical robot platform |
US11793570B2 (en) * | 2012-06-21 | 2023-10-24 | Globus Medical Inc. | Surgical robotic automation with tracking markers |
US9226796B2 (en) | 2012-08-03 | 2016-01-05 | Stryker Corporation | Method for detecting a disturbance as an energy applicator of a surgical instrument traverses a cutting path |
CN107198567B (en) | 2012-08-03 | 2021-02-09 | 史赛克公司 | Systems and methods for robotic surgery |
EP2882331A4 (en) | 2012-08-08 | 2016-03-23 | Univ Nebraska | Robotic surgical devices, systems, and related methods |
US10588691B2 (en) | 2012-09-12 | 2020-03-17 | Relievant Medsystems, Inc. | Radiofrequency ablation of tissue within a vertebral body |
US20160164976A1 (en) * | 2012-09-24 | 2016-06-09 | Suitable Technologies, Inc. | Systems and methods for remote presence |
US9008757B2 (en) | 2012-09-26 | 2015-04-14 | Stryker Corporation | Navigation system including optical and non-optical sensors |
US9724156B2 (en) * | 2012-10-09 | 2017-08-08 | Mohammadreza Meymanat | System for detecting, actively targeting and selectively treating malignant cells in human body by non-invasive procedure |
KR101433242B1 (en) * | 2012-11-16 | 2014-08-25 | 경북대학교 산학협력단 | Reduction surgical robot and method for driving control thereof |
US10013531B2 (en) * | 2012-12-20 | 2018-07-03 | Accenture Global Services Limited | Context based augmented reality |
CA2896947C (en) * | 2013-01-04 | 2021-02-09 | DePuy Synthes Products, Inc. | Method for designing and manufacturing a bone implant |
US9993273B2 (en) | 2013-01-16 | 2018-06-12 | Mako Surgical Corp. | Bone plate and tracking device using a bone plate for attaching to a patient's anatomy |
US9566120B2 (en) | 2013-01-16 | 2017-02-14 | Stryker Corporation | Navigation systems and methods for indicating and reducing line-of-sight errors |
US9057600B2 (en) | 2013-03-13 | 2015-06-16 | Hansen Medical, Inc. | Reducing incremental measurement sensor error |
US9014851B2 (en) | 2013-03-15 | 2015-04-21 | Hansen Medical, Inc. | Systems and methods for tracking robotically controlled medical instruments |
US10274553B2 (en) | 2013-03-15 | 2019-04-30 | Canon U.S.A., Inc. | Needle placement manipulator with attachment for RF-coil |
US9271663B2 (en) | 2013-03-15 | 2016-03-01 | Hansen Medical, Inc. | Flexible instrument localization from both remote and elongation sensors |
US9629595B2 (en) | 2013-03-15 | 2017-04-25 | Hansen Medical, Inc. | Systems and methods for localizing, tracking and/or controlling medical instruments |
US9764468B2 (en) | 2013-03-15 | 2017-09-19 | Brain Corporation | Adaptive predictor apparatus and methods |
US9037396B2 (en) * | 2013-05-23 | 2015-05-19 | Irobot Corporation | Simultaneous localization and mapping for a mobile robot |
US11020016B2 (en) | 2013-05-30 | 2021-06-01 | Auris Health, Inc. | System and method for displaying anatomy and devices on a movable display |
US9242372B2 (en) | 2013-05-31 | 2016-01-26 | Brain Corporation | Adaptive robotic interface apparatus and methods |
US10368878B2 (en) | 2013-06-11 | 2019-08-06 | Orthotaxy | System for positioning a surgical device |
US9179051B1 (en) * | 2013-06-13 | 2015-11-03 | Clara Stoudt | Voice-activated hands-free camera holder systems |
WO2014199413A1 (en) | 2013-06-13 | 2014-12-18 | テルモ株式会社 | Medical manipulator, and control method therefor |
US9314924B1 (en) | 2013-06-14 | 2016-04-19 | Brain Corporation | Predictive robotic controller apparatus and methods |
US9792546B2 (en) | 2013-06-14 | 2017-10-17 | Brain Corporation | Hierarchical robotic controller apparatus and methods |
DE102013213727A1 (en) * | 2013-07-12 | 2015-01-15 | Siemens Aktiengesellschaft | Interventional imaging system |
WO2015009949A2 (en) | 2013-07-17 | 2015-01-22 | Board Of Regents Of The University Of Nebraska | Robotic surgical devices, systems and related methods |
KR101534097B1 (en) * | 2013-08-13 | 2015-07-06 | 삼성전자주식회사 | Apparatus for obtaining medical image and method for adjusting location of table by using the same |
FR3010628B1 (en) | 2013-09-18 | 2015-10-16 | Medicrea International | METHOD FOR REALIZING THE IDEAL CURVATURE OF A ROD OF A VERTEBRAL OSTEOSYNTHESIS EQUIPMENT FOR STRENGTHENING THE VERTEBRAL COLUMN OF A PATIENT |
WO2015040623A1 (en) * | 2013-09-22 | 2015-03-26 | Mazor Robotics Ltd. | Automatic dilator |
US9579789B2 (en) | 2013-09-27 | 2017-02-28 | Brain Corporation | Apparatus and methods for training of robotic control arbitration |
US9283048B2 (en) * | 2013-10-04 | 2016-03-15 | KB Medical SA | Apparatus and systems for precise guidance of surgical tools |
FR3012030B1 (en) | 2013-10-18 | 2015-12-25 | Medicrea International | METHOD FOR REALIZING THE IDEAL CURVATURE OF A ROD OF A VERTEBRAL OSTEOSYNTHESIS EQUIPMENT FOR STRENGTHENING THE VERTEBRAL COLUMN OF A PATIENT |
JP7107635B2 (en) * | 2013-10-24 | 2022-07-27 | グローバス メディカル インコーポレイティッド | Surgical tool system and method |
US9597797B2 (en) | 2013-11-01 | 2017-03-21 | Brain Corporation | Apparatus and methods for haptic training of robots |
US9463571B2 (en) | 2013-11-01 | 2016-10-11 | Brian Corporation | Apparatus and methods for online training of robots |
KR20150053860A (en) * | 2013-11-08 | 2015-05-19 | 삼성전자주식회사 | Method and apparatus for controlling a movement of a medical device |
DE102013112375A1 (en) * | 2013-11-11 | 2015-05-13 | Aesculap Ag | Surgical referencing device, surgical navigation system and method |
US20150157197A1 (en) * | 2013-12-09 | 2015-06-11 | Omer Aslam Ilahi | Endoscopic image overlay |
CA2930172C (en) * | 2013-12-10 | 2018-08-07 | Halliburton Energy Services, Inc. | Continuous live tracking system for placement of cutting elements |
US9066755B1 (en) * | 2013-12-13 | 2015-06-30 | DePuy Synthes Products, Inc. | Navigable device recognition system |
US9358685B2 (en) | 2014-02-03 | 2016-06-07 | Brain Corporation | Apparatus and methods for control of robot actions based on corrective user inputs |
EP3104803B1 (en) | 2014-02-11 | 2021-09-15 | KB Medical SA | Sterile handle for controlling a robotic surgical system from a sterile field |
EP3107478B1 (en) | 2014-02-20 | 2020-12-09 | Intuitive Surgical Operations, Inc. | Limited movement of a surgical mounting platform controlled by manual motion of robotic arms |
US9867667B2 (en) | 2014-02-27 | 2018-01-16 | Canon Usa Inc. | Placement apparatus |
US10561469B2 (en) * | 2014-02-28 | 2020-02-18 | Sony Corporation | Robot arm apparatus and robot arm control method |
CN106470634B (en) | 2014-03-17 | 2019-06-14 | 直观外科手术操作公司 | System and method for being indicated outside the screen of the instrument in remote control operation medical system |
US20170296273A9 (en) * | 2014-03-17 | 2017-10-19 | Roy Anthony Brown | Surgical Targeting Systems and Methods |
KR102387096B1 (en) * | 2014-03-28 | 2022-04-15 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | Quantitative three-dimensional visualization of instruments in a field of view |
DE102014104800A1 (en) | 2014-04-03 | 2015-10-08 | Aesculap Ag | Medical fastening device and referencing device and medical instruments |
US20150282735A1 (en) * | 2014-04-04 | 2015-10-08 | Izi Medical Products,Llc | Reference device for surgical navigation system |
JP6557255B2 (en) * | 2014-04-17 | 2019-08-07 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | Method and system for detecting high speed movement of a surgical device |
CN106659537B (en) | 2014-04-24 | 2019-06-11 | Kb医疗公司 | The surgical instrument holder used in conjunction with robotic surgical system |
US9346167B2 (en) | 2014-04-29 | 2016-05-24 | Brain Corporation | Trainable convolutional network apparatus and methods for operating a robotic vehicle |
WO2015171988A1 (en) | 2014-05-09 | 2015-11-12 | Canon U.S.A., Inc. | Positioning apparatus |
EP3811891A3 (en) | 2014-05-14 | 2021-05-05 | Stryker European Holdings I, LLC | Navigation system and processor arrangement for tracking the position of a work target |
US10836038B2 (en) | 2014-05-21 | 2020-11-17 | Fanuc America Corporation | Learning path control |
KR101633774B1 (en) * | 2014-05-29 | 2016-06-28 | 주식회사 고영테크놀러지 | Assistance robot for vertebra surgery |
WO2015184146A1 (en) * | 2014-05-30 | 2015-12-03 | Sameh Mesallum | Systems for automated biomechanical computerized surgery |
US9717443B2 (en) * | 2014-06-12 | 2017-08-01 | Coloplast A/S | Surgical tool and method for identifying an incision site |
US10828120B2 (en) * | 2014-06-19 | 2020-11-10 | Kb Medical, Sa | Systems and methods for performing minimally invasive surgery |
US9603668B2 (en) | 2014-07-02 | 2017-03-28 | Covidien Lp | Dynamic 3D lung map view for tool navigation inside the lung |
USD823312S1 (en) * | 2014-08-11 | 2018-07-17 | Sony Corporation | Display panel or screen with graphical user interface |
JP2016045874A (en) * | 2014-08-26 | 2016-04-04 | ソニー株式会社 | Information processor, method for information processing, and program |
US9993177B2 (en) | 2014-08-28 | 2018-06-12 | DePuy Synthes Products, Inc. | Systems and methods for intraoperatively measuring anatomical orientation |
EP3188662A4 (en) * | 2014-09-05 | 2018-05-23 | Procept Biorobotics Corporation | Physician controlled tissue resection integrated with treatment mapping of target organ images |
US10285670B2 (en) | 2014-09-12 | 2019-05-14 | Canon U.S.A., Inc. | Needle positioning apparatus |
US10463434B2 (en) * | 2014-09-24 | 2019-11-05 | 7D Surgical Inc. | Tracking marker support structure and surface registration methods employing the same for performing navigated surgical procedures |
US9630318B2 (en) | 2014-10-02 | 2017-04-25 | Brain Corporation | Feature detection apparatus and methods for training of robotic navigation |
WO2016059603A1 (en) * | 2014-10-17 | 2016-04-21 | Koninklijke Philips N.V. | System for real-time organ segmentation and tool navigation during tool insertion in interventional therapy and method of opeperation thereof |
EP3009095A1 (en) * | 2014-10-17 | 2016-04-20 | Imactis | Method for planning the introduction of a needle in a patient's body |
EP3009091A1 (en) * | 2014-10-17 | 2016-04-20 | Imactis | Medical system for use in interventional radiology |
EP3009096A1 (en) * | 2014-10-17 | 2016-04-20 | Imactis | Method and system for displaying the position and orientation of a linear instrument navigated with respect to a 3D medical image |
US10441366B2 (en) * | 2014-10-22 | 2019-10-15 | Think Surgical, Inc. | Actively controlled optical tracker with a robot |
US20160119593A1 (en) * | 2014-10-24 | 2016-04-28 | Nurep, Inc. | Mobile console |
CN111839731A (en) | 2014-10-27 | 2020-10-30 | 直观外科手术操作公司 | System and method for monitoring control points during reactive motion |
CN107072864B (en) | 2014-10-27 | 2019-06-14 | 直观外科手术操作公司 | For being registrated to the system and method for operating table |
CN115444567A (en) | 2014-10-27 | 2022-12-09 | 直观外科手术操作公司 | System and method for instrument interference compensation |
EP3212149A4 (en) | 2014-10-27 | 2018-05-30 | Intuitive Surgical Operations, Inc. | System and method for integrated surgical table |
EP3212151B1 (en) | 2014-10-27 | 2020-07-29 | Intuitive Surgical Operations, Inc. | System for integrated surgical table motion |
CN111166476B (en) | 2014-10-27 | 2023-05-02 | 直观外科手术操作公司 | Medical device with active brake release control |
JP6731920B2 (en) | 2014-12-02 | 2020-07-29 | カーベー メディカル エスアー | Robot-assisted volume removal during surgery |
DE102014226240A1 (en) * | 2014-12-17 | 2016-06-23 | Kuka Roboter Gmbh | System for robot-assisted medical treatment |
CN107110631B (en) | 2015-02-02 | 2020-01-31 | 西安大略大学 | System for tracking objects |
WO2016131648A1 (en) * | 2015-02-17 | 2016-08-25 | Koninklijke Philips N.V. | Device for positioning a marker in a 3d ultrasonic image volume |
US10555782B2 (en) | 2015-02-18 | 2020-02-11 | Globus Medical, Inc. | Systems and methods for performing minimally invasive spinal surgery with a robotic surgical system using a percutaneous technique |
CN111839732A (en) | 2015-02-25 | 2020-10-30 | 马科外科公司 | Navigation system and method for reducing tracking interruptions during surgical procedures |
US9717387B1 (en) | 2015-02-26 | 2017-08-01 | Brain Corporation | Apparatus and methods for programming and training of robotic household appliances |
CN104644205A (en) * | 2015-03-02 | 2015-05-27 | 上海联影医疗科技有限公司 | Method and system for positioning patient during diagnostic imaging |
US11576645B2 (en) | 2015-03-02 | 2023-02-14 | Shanghai United Imaging Healthcare Co., Ltd. | Systems and methods for scanning a patient in an imaging system |
US11576578B2 (en) | 2015-03-02 | 2023-02-14 | Shanghai United Imaging Healthcare Co., Ltd. | Systems and methods for scanning a patient in an imaging system |
CN107405171A (en) * | 2015-03-11 | 2017-11-28 | 柯惠Lp公司 | Operating robot trolley fixator |
EP3282997B1 (en) | 2015-04-15 | 2021-06-16 | Mobius Imaging, LLC | Integrated medical imaging and surgical robotic system |
WO2016181320A1 (en) | 2015-05-12 | 2016-11-17 | Navix International Limited | Fiducial marking for image-electromagnetic field registration |
US10376335B2 (en) * | 2015-05-20 | 2019-08-13 | Siemens Healthcare Gmbh | Method and apparatus to provide updated patient images during robotic surgery |
US9622831B2 (en) * | 2015-05-20 | 2017-04-18 | Siemens Healthcare Gmbh | Method and apparatus to provide updated patient images during robotic surgery |
FR3036279B1 (en) * | 2015-05-21 | 2017-06-23 | Medtech Sa | NEUROSURGICAL ASSISTANCE ROBOT |
KR102371053B1 (en) * | 2015-06-04 | 2022-03-10 | 큐렉소 주식회사 | Surgical robot system |
GB201509887D0 (en) * | 2015-06-08 | 2015-07-22 | Univ Strathclyde | Remote operations system |
DE102015109371A1 (en) * | 2015-06-12 | 2016-12-15 | avateramedical GmBH | Apparatus and method for robotic surgery |
US9867673B2 (en) | 2015-07-14 | 2018-01-16 | Canon U.S.A, Inc. | Medical support device |
US10639065B2 (en) | 2015-07-21 | 2020-05-05 | Canon U.S.A., Inc. | Medical assist device |
US10646298B2 (en) * | 2015-07-31 | 2020-05-12 | Globus Medical, Inc. | Robot arm and methods of use |
EP3331453A4 (en) | 2015-08-03 | 2019-04-24 | Board of Regents of the University of Nebraska | Robotic surgical devices systems and related methods |
US10702226B2 (en) * | 2015-08-06 | 2020-07-07 | Covidien Lp | System and method for local three dimensional volume reconstruction using a standard fluoroscope |
WO2017028916A1 (en) | 2015-08-19 | 2017-02-23 | Brainlab Ag | Reference array holder |
US10687905B2 (en) * | 2015-08-31 | 2020-06-23 | KB Medical SA | Robotic surgical systems and methods |
US10034716B2 (en) * | 2015-09-14 | 2018-07-31 | Globus Medical, Inc. | Surgical robotic systems and methods thereof |
EP3349649B1 (en) | 2015-09-18 | 2022-03-09 | Auris Health, Inc. | Navigation of tubular networks |
EP3147086B1 (en) * | 2015-09-22 | 2020-11-04 | Airbus Defence and Space GmbH | Automation of robot operations in aircraft construction |
DE112016004725B4 (en) * | 2015-10-14 | 2021-09-16 | Kawasaki Jukogyo Kabushiki Kaisha | Method for teaching a robot and robot arm control device |
FI20155784A (en) * | 2015-11-02 | 2017-05-03 | Cryotech Nordic Oü | Automated system for laser-assisted dermatological treatment and control procedure |
EP3200719B1 (en) | 2015-11-02 | 2019-05-22 | Brainlab AG | Determining a configuration of a medical robotic arm |
EP3370657B1 (en) | 2015-11-04 | 2023-12-27 | Medicrea International | Apparatus for spinal reconstructive surgery and measuring spinal length |
GB2546388B (en) * | 2015-11-24 | 2020-10-21 | Cmr Surgical Ltd | Port Control |
US10143526B2 (en) | 2015-11-30 | 2018-12-04 | Auris Health, Inc. | Robot-assisted driving systems and methods |
CN105616004B (en) * | 2015-12-09 | 2018-10-26 | 北京工业大学 | A kind of surgical robot system and operating method of radioactive particle interventional treatment tumour |
CN106913383B (en) * | 2015-12-25 | 2020-04-21 | 先健科技(深圳)有限公司 | Developing structure and implantable medical device with same |
US10806523B2 (en) | 2015-12-28 | 2020-10-20 | Xact Robotics Ltd. | Adjustable registration frame |
WO2017115235A1 (en) * | 2015-12-29 | 2017-07-06 | Koninklijke Philips N.V. | Image-based adaptive path planning for a robot |
US10335241B2 (en) | 2015-12-30 | 2019-07-02 | DePuy Synthes Products, Inc. | Method and apparatus for intraoperative measurements of anatomical orientation |
US9554411B1 (en) | 2015-12-30 | 2017-01-24 | DePuy Synthes Products, Inc. | Systems and methods for wirelessly powering or communicating with sterile-packed devices |
US10888483B2 (en) | 2016-01-12 | 2021-01-12 | Virginia Commonwealth University | Systems, devices, and methods for position monitoring and motion compensation |
WO2017132505A1 (en) | 2016-01-29 | 2017-08-03 | Canon U.S.A., Inc. | Tool placement manipulator |
US11883217B2 (en) | 2016-02-03 | 2024-01-30 | Globus Medical, Inc. | Portable medical imaging system and method |
WO2017139556A1 (en) * | 2016-02-12 | 2017-08-17 | Medos International Sarl | Systems and methods for intraoperatively measuring anatomical orientation |
EP3417428B1 (en) * | 2016-02-17 | 2023-06-07 | De La Rue Authentication Solutions, Inc. | Method for determining authenticity using images that exhibit parallax |
GB2578422B (en) * | 2016-02-25 | 2021-09-15 | Noel Dyer Kelly | System and method for automatic muscle movement detection |
US11064904B2 (en) | 2016-02-29 | 2021-07-20 | Extremity Development Company, Llc | Smart drill, jig, and method of orthopedic surgery |
WO2017151751A1 (en) * | 2016-03-02 | 2017-09-08 | Think Surgical, Inc. | Method for recovering a registration of a bone |
US11353326B2 (en) * | 2016-03-06 | 2022-06-07 | Arizona Board Of Regents On Behalf Of The University Of Arizona | Traverse and trajectory optimization and multi-purpose tracking |
CN105536093B (en) * | 2016-03-10 | 2019-05-03 | 京东方科技集团股份有限公司 | Transfusion system and the infusion method for using transfusion system |
WO2017158397A1 (en) | 2016-03-13 | 2017-09-21 | Synaptive Medical (Barbados) Inc. | System and method for sensing tissue deformation |
EP3988027A1 (en) * | 2016-03-13 | 2022-04-27 | Vuze Medical Ltd. | Apparatus and methods for use with skeletal procedures |
JP6733239B2 (en) * | 2016-03-18 | 2020-07-29 | セイコーエプソン株式会社 | Controller and robot system |
WO2017169898A1 (en) * | 2016-03-30 | 2017-10-05 | パナソニックIpマネジメント株式会社 | Data storage device, robot system, and data storage method |
US20170281976A1 (en) * | 2016-04-01 | 2017-10-05 | Varian Medical Systems International Ag | Systems and methods for detecting and/or measuring motion and position associated with a patient |
CN105997250B (en) * | 2016-04-27 | 2018-11-20 | 中国科学院深圳先进技术研究院 | A kind of spinal lamina grinding attachment |
US20170312033A1 (en) * | 2016-04-27 | 2017-11-02 | Metal Industries Research&Development Centre | Surgery navigation system |
WO2017197170A1 (en) * | 2016-05-12 | 2017-11-16 | The Regents Of The University Of California | Safely controlling an autonomous entity in presence of intelligent agents |
EP4353182A2 (en) | 2016-05-18 | 2024-04-17 | Virtual Incision Corporation | Robotic surgical devices and systems |
US10537395B2 (en) | 2016-05-26 | 2020-01-21 | MAKO Surgical Group | Navigation tracker with kinematic connector assembly |
US10331138B2 (en) * | 2016-07-05 | 2019-06-25 | Baidu Usa Llc | Standard scene-based planning control methods for operating autonomous vehicles |
EP3484376A4 (en) | 2016-07-12 | 2020-07-22 | Mobius Imaging LLC | Multi-stage dilator and cannula system and method |
EP3484583B1 (en) * | 2016-07-13 | 2021-07-07 | Sensus Healthcare, Inc. | Robotic intraoperative radiation therapy |
EP3484362A1 (en) | 2016-07-14 | 2019-05-22 | Navix International Limited | Characteristic track catheter navigation |
US10052163B2 (en) * | 2016-07-19 | 2018-08-21 | Hcl Technologies Limited | Assisting a surgeon to operate a surgical device |
DE102016214319A1 (en) * | 2016-08-03 | 2018-02-08 | Siemens Healthcare Gmbh | biopsy unit |
CA3022448A1 (en) | 2016-08-05 | 2018-02-08 | Brainlab Ag | Automatic image registration of scans for image-guided surgery |
US11058496B2 (en) * | 2016-08-15 | 2021-07-13 | Biosense Webster (Israel) Ltd. | Registering probe and sheath images on a display |
KR101848027B1 (en) * | 2016-08-16 | 2018-04-12 | 주식회사 고영테크놀러지 | Surgical robot system for stereotactic surgery and method for controlling a stereotactic surgery robot |
KR101861176B1 (en) * | 2016-08-16 | 2018-05-28 | 주식회사 고영테크놀러지 | Surgical robot for stereotactic surgery and method for controlling a stereotactic surgery robot |
CN106241686B (en) * | 2016-08-31 | 2019-07-05 | 山东新华医疗器械股份有限公司 | Raw material explosive barrel automatic cap fastening machine |
EP3509527A4 (en) * | 2016-09-09 | 2020-12-30 | Mobius Imaging LLC | Methods and systems for display of patient data in computer-assisted surgery |
US10993771B2 (en) * | 2016-09-12 | 2021-05-04 | Synaptive Medical Inc. | Trackable apparatuses and methods |
US10650621B1 (en) | 2016-09-13 | 2020-05-12 | Iocurrents, Inc. | Interfacing with a vehicular controller area network |
WO2018053282A1 (en) * | 2016-09-16 | 2018-03-22 | GYS Tech, LLC d/b/a Cardan Robotics | System and method for mounting a robotic arm in a surgical robotic system |
CN109952070B (en) * | 2016-10-05 | 2022-02-01 | 纽文思公司 | Surgical navigation system and related methods |
CN206062123U (en) * | 2016-10-10 | 2017-04-05 | 韩力 | A kind of gas heating type smoking product |
US10695087B2 (en) | 2016-10-19 | 2020-06-30 | Canon U.S.A., Inc. | Placement manipulator and attachment for positioning a puncture instrument |
CN111417352A (en) | 2016-10-21 | 2020-07-14 | Gys科技有限责任公司(经营名称为卡丹机器人) | Method and system for setting trajectory and target position for image-guided surgery |
FR3057757B1 (en) * | 2016-10-21 | 2021-04-16 | Medtech | AUTOMATIC REGISTRATION DEVICE AND METHOD FOR 3D INTRA-OPERATIVE IMAGES |
US11751948B2 (en) | 2016-10-25 | 2023-09-12 | Mobius Imaging, Llc | Methods and systems for robot-assisted surgery |
DE102016221222A1 (en) * | 2016-10-27 | 2018-05-03 | Siemens Healthcare Gmbh | A method of operating a collision protection system for a medical surgical device, medical surgical device, computer program and data carrier |
CN114795479A (en) * | 2016-10-28 | 2022-07-29 | 奥尔索夫特Ulc公司 | Robotic cutting workflow |
EP3318213B1 (en) * | 2016-11-04 | 2022-10-12 | Globus Medical, Inc | System for measuring depth of instrumentation |
EP3534817A4 (en) * | 2016-11-04 | 2020-07-29 | Intuitive Surgical Operations Inc. | Reconfigurable display in computer-assisted tele-operated surgery |
JP2018110841A (en) * | 2016-11-10 | 2018-07-19 | グローバス メディカル インコーポレイティッド | Systems and methods of checking positioning for surgical systems |
US10350010B2 (en) * | 2016-11-14 | 2019-07-16 | Intai Technology Corp. | Method and system for verifying panoramic images of implants |
WO2018092059A1 (en) | 2016-11-16 | 2018-05-24 | Navix International Limited | Tissue model dynamic visual rendering |
US11331029B2 (en) | 2016-11-16 | 2022-05-17 | Navix International Limited | Esophagus position detection by electrical mapping |
US11284813B2 (en) | 2016-11-16 | 2022-03-29 | Navix International Limited | Real-time display of tissue deformation by interactions with an intra-body probe |
EP3541313B1 (en) | 2016-11-16 | 2023-05-10 | Navix International Limited | Estimators for ablation effectiveness |
WO2018092063A1 (en) * | 2016-11-16 | 2018-05-24 | Navix International Limited | Real-time display of treatment-related tissue changes using virtual material |
WO2018109556A1 (en) | 2016-12-12 | 2018-06-21 | Medicrea International | Systems and methods for patient-specific spinal implants |
WO2018112025A1 (en) | 2016-12-16 | 2018-06-21 | Mako Surgical Corp. | Techniques for modifying tool operation in a surgical robotic system based on comparing actual and commanded states of the tool relative to a surgical site |
US10918445B2 (en) | 2016-12-19 | 2021-02-16 | Ethicon Llc | Surgical system with augmented reality display |
US10244926B2 (en) | 2016-12-28 | 2019-04-02 | Auris Health, Inc. | Detecting endolumenal buckling of flexible instruments |
US11173003B2 (en) | 2017-01-10 | 2021-11-16 | Intuitive Surgical Operations, Inc. | Systems and methods for using a robotic medical system |
JP7304820B2 (en) * | 2017-02-14 | 2023-07-07 | アトラクシス エス・アー・エール・エル | Fast optical tracking with compression and/or CMOS windowing |
US11158415B2 (en) * | 2017-02-16 | 2021-10-26 | Mako Surgical Corporation | Surgical procedure planning system with multiple feedback loops |
US10010379B1 (en) * | 2017-02-21 | 2018-07-03 | Novarad Corporation | Augmented reality viewing and tagging for medical procedures |
EP3369394B1 (en) * | 2017-03-03 | 2020-06-10 | Globus Medical, Inc. | System for a surveillance marker in robotic-assisted surgery |
US10687863B2 (en) * | 2017-03-10 | 2020-06-23 | Board Of Trustees Of Michigan State University | Apparatus and method for minimally invasive osteosynthesis of sacroiliac luxations/fractures |
JP7018604B2 (en) * | 2017-03-16 | 2022-02-14 | 東芝エネルギーシステムズ株式会社 | Subject positioning device, subject positioning method, subject positioning program and radiation therapy system |
US10682129B2 (en) | 2017-03-23 | 2020-06-16 | Mobius Imaging, Llc | Robotic end effector with adjustable inner diameter |
JP7170979B2 (en) | 2017-03-31 | 2022-11-15 | エンピリアン メディカル システムズ,インコーポレイテッド | X-ray source that forms a three-dimensional beam |
CN108990412B (en) | 2017-03-31 | 2022-03-22 | 奥瑞斯健康公司 | Robot system for cavity network navigation compensating physiological noise |
US10349986B2 (en) | 2017-04-20 | 2019-07-16 | Warsaw Orthopedic, Inc. | Spinal implant system and method |
EP4108201B1 (en) | 2017-04-21 | 2024-03-27 | Medicrea International | A system for developing one or more patient-specific spinal implants |
US10471605B2 (en) * | 2017-04-26 | 2019-11-12 | Ready Robotics | Programmable adapters for detachably connecting robotic peripherals to adaptively retool robots |
EP3621545B1 (en) | 2017-05-10 | 2024-02-21 | MAKO Surgical Corp. | Robotic spine surgery system |
US11033341B2 (en) | 2017-05-10 | 2021-06-15 | Mako Surgical Corp. | Robotic spine surgery system and methods |
CN110799145B (en) * | 2017-05-12 | 2024-03-01 | 网络手术公司 | Surgical robot for orthopedic interventions |
AU2018273693A1 (en) * | 2017-05-25 | 2019-11-28 | Covidien Lp | Systems and methods for detection of objects within a field of view of an image capture device |
TWI670681B (en) * | 2017-06-04 | 2019-09-01 | 鈦隼生物科技股份有限公司 | Method and system of determining one or more points on operation pathway |
EP3412232B1 (en) * | 2017-06-09 | 2024-03-27 | K2M, Inc. | Fixation systems for repairing a pars fracture |
US10022192B1 (en) | 2017-06-23 | 2018-07-17 | Auris Health, Inc. | Automatically-initialized robotic systems for navigation of luminal networks |
CN116725667A (en) | 2017-06-28 | 2023-09-12 | 奥瑞斯健康公司 | System for providing positioning information and method for positioning an instrument within an anatomical structure |
WO2019005696A1 (en) | 2017-06-28 | 2019-01-03 | Auris Health, Inc. | Electromagnetic distortion detection |
JP6959428B2 (en) * | 2017-07-07 | 2021-11-02 | キヤノン ユーエスエイ, インコーポレイテッドCanon U.S.A., Inc | Planning for multiple probe ablation |
WO2019012520A1 (en) | 2017-07-08 | 2019-01-17 | Vuze Medical Ltd. | Apparatus and methods for use with image-guided skeletal procedures |
CN111148547A (en) | 2017-07-18 | 2020-05-12 | 胜赛斯医疗有限责任公司 | Real-time X-ray dosimetry in intraoperative radiation therapy |
US10675094B2 (en) * | 2017-07-21 | 2020-06-09 | Globus Medical Inc. | Robot surgical platform |
US20210361357A1 (en) * | 2017-07-21 | 2021-11-25 | Globus Medical, Inc. | Robot surgical platform |
US10448978B2 (en) * | 2017-07-27 | 2019-10-22 | Warsaw Orthopedic, Inc. | Spinal implant system and method |
JP2019025572A (en) * | 2017-07-28 | 2019-02-21 | セイコーエプソン株式会社 | Control device of robot, the robot, robot system, and method of checking abnormality of the robot |
WO2019032849A2 (en) | 2017-08-11 | 2019-02-14 | GYS Tech, LLC d/b/a Cardan Robotics | Method and apparatus for attaching a reference marker to a patient |
CN107551387B (en) * | 2017-08-17 | 2020-05-15 | 济南优科医疗技术有限公司 | Multifunctional auxiliary treatment robot |
US20190059850A1 (en) * | 2017-08-25 | 2019-02-28 | Neural Analytics, Inc. | Portable headset |
US10605875B2 (en) * | 2017-08-28 | 2020-03-31 | Synaptive Medical (Barbados) Inc. | Contrast system and methods for reflective markers |
US11051894B2 (en) | 2017-09-27 | 2021-07-06 | Virtual Incision Corporation | Robotic surgical devices with tracking camera technology and related systems and methods |
EP3691545A4 (en) | 2017-10-04 | 2022-02-16 | Mobius Imaging, LLC | Systems and methods for performing lateral-access spine surgery |
EP3691558A4 (en) | 2017-10-05 | 2021-07-21 | Mobius Imaging LLC | Methods and systems for performing computer assisted surgery |
US11197723B2 (en) | 2017-10-09 | 2021-12-14 | Canon U.S.A., Inc. | Medical guidance system and method using localized insertion plane |
US10555778B2 (en) | 2017-10-13 | 2020-02-11 | Auris Health, Inc. | Image-based branch detection and mapping for navigation |
US11058493B2 (en) | 2017-10-13 | 2021-07-13 | Auris Health, Inc. | Robotic system configured for navigation path tracing |
WO2019075719A1 (en) * | 2017-10-20 | 2019-04-25 | 深圳华大智造科技有限公司 | Ultrasonic detection device, ultrasonic control device, ultrasonic system and ultrasonic imaging method |
CA3080151A1 (en) | 2017-10-23 | 2019-05-02 | Peter L. BONO | Rotary oscillating/reciprocating surgical tool |
EP3476358B8 (en) * | 2017-10-27 | 2020-10-21 | Siemens Healthcare GmbH | System for tracking a position of a target object |
CN110383336B (en) * | 2017-11-15 | 2022-11-01 | 深圳市瑞立视多媒体科技有限公司 | Rigid body configuration method, device, terminal equipment and computer storage medium |
US10918422B2 (en) | 2017-12-01 | 2021-02-16 | Medicrea International | Method and apparatus for inhibiting proximal junctional failure |
US11452572B2 (en) * | 2017-12-14 | 2022-09-27 | Intuitive Surgical Operations, Inc. | Medical tools having tension bands |
AU2018384820A1 (en) | 2017-12-14 | 2020-05-21 | Auris Health, Inc. | System and method for estimating instrument location |
EP3684283A4 (en) | 2017-12-18 | 2021-07-14 | Auris Health, Inc. | Methods and systems for instrument tracking and navigation within luminal networks |
CN117140580A (en) | 2018-01-05 | 2023-12-01 | 内布拉斯加大学董事会 | Single arm robotic device with compact joint design and related systems and methods |
EP3510927A1 (en) * | 2018-01-10 | 2019-07-17 | Globus Medical, Inc. | Surgical robotic systems with target trajectory deviation monitoring |
CA3088311A1 (en) * | 2018-01-12 | 2019-07-18 | Peter L. BONO | Surgical sensor anchor system |
AU2019207913A1 (en) * | 2018-01-12 | 2020-09-03 | Capstone Surgical Technologies, Llc | Robotic surgical control system |
CN111655187A (en) | 2018-01-26 | 2020-09-11 | 马科外科公司 | End effector, system, and method for impacting a prosthesis guided by a surgical robot |
US20190254753A1 (en) * | 2018-02-19 | 2019-08-22 | Globus Medical, Inc. | Augmented reality navigation systems for use with robotic surgical systems and methods of their use |
US10856942B2 (en) * | 2018-03-02 | 2020-12-08 | Ethicon Llc | System and method for closed-loop surgical tool homing |
KR102060998B1 (en) * | 2018-03-23 | 2019-12-31 | 주식회사 제이엘유 | Laser medical treatment |
MX2020010112A (en) | 2018-03-28 | 2020-11-06 | Auris Health Inc | Systems and methods for registration of location sensors. |
US10827913B2 (en) | 2018-03-28 | 2020-11-10 | Auris Health, Inc. | Systems and methods for displaying estimated location of instrument |
JP6810087B2 (en) * | 2018-03-29 | 2021-01-06 | ファナック株式会社 | Machine learning device, robot control device and robot vision system using machine learning device, and machine learning method |
US11672491B2 (en) | 2018-03-30 | 2023-06-13 | Empyrean Medical Systems, Inc. | Validation of therapeutic radiation treatment |
WO2019198061A1 (en) | 2018-04-13 | 2019-10-17 | Universidade Do Minho | Guidance system, method and devices thereof |
US11039894B2 (en) | 2018-04-20 | 2021-06-22 | Verb Surgical Inc. | Robotic port placement guide and method of use |
WO2019211741A1 (en) * | 2018-05-02 | 2019-11-07 | Augmedics Ltd. | Registration of a fiducial marker for an augmented reality system |
TWI740138B (en) * | 2018-05-04 | 2021-09-21 | 醫百科技股份有限公司 | Composite surgical probe device |
EP3566669A1 (en) | 2018-05-10 | 2019-11-13 | Globus Medical, Inc. | Systems and methods related to robotic guidance in surgery |
CN112074235A (en) * | 2018-05-10 | 2020-12-11 | 美国西门子医疗系统股份有限公司 | Visual indicator system for hospital beds |
US10966736B2 (en) * | 2018-05-21 | 2021-04-06 | Warsaw Orthopedic, Inc. | Spinal implant system and methods of use |
DE102018208203B4 (en) * | 2018-05-24 | 2020-02-13 | Carl Zeiss Industrielle Messtechnik Gmbh | Target body, arrangement with target body and method for determining a position and / or an orientation of a target body |
US11191594B2 (en) * | 2018-05-25 | 2021-12-07 | Mako Surgical Corp. | Versatile tracking arrays for a navigation system and methods of recovering registration using the same |
US10905499B2 (en) | 2018-05-30 | 2021-02-02 | Auris Health, Inc. | Systems and methods for location sensor-based branch prediction |
US10898275B2 (en) | 2018-05-31 | 2021-01-26 | Auris Health, Inc. | Image-based airway analysis and mapping |
JP7214757B2 (en) | 2018-05-31 | 2023-01-30 | オーリス ヘルス インコーポレイテッド | Robotic system and method for navigation of luminal networks that detect physiological noise |
EP3801189A4 (en) | 2018-05-31 | 2022-02-23 | Auris Health, Inc. | Path-based navigation of tubular networks |
CN110575255B (en) | 2018-06-07 | 2022-08-16 | 格罗伯斯医疗有限公司 | Robotic system and related methods for providing co-registration using natural fiducials |
ES1215739Y (en) * | 2018-06-15 | 2018-10-17 | Fundacion Instituto De Investig Marques De Valdecilla | Navigation reference team in robotic surgery assisted by stereotactic navigation of the organs and soft parts of a patient's pelvic area |
DE102019004233B4 (en) | 2018-06-15 | 2022-09-22 | Mako Surgical Corp. | SYSTEMS AND METHODS FOR TRACKING OBJECTS |
US10589423B2 (en) * | 2018-06-18 | 2020-03-17 | Shambhu Nath Roy | Robot vision super visor for hybrid homing, positioning and workspace UFO detection enabling industrial robot use for consumer applications |
US11094221B2 (en) | 2018-06-21 | 2021-08-17 | University Of Utah Research Foundation | Visual guidance system and method for posing a physical object in three dimensional space |
JP7082090B2 (en) | 2018-06-27 | 2022-06-07 | グローバス メディカル インコーポレイティッド | How to tune virtual implants and related surgical navigation systems |
JP6936282B2 (en) | 2018-06-28 | 2021-09-15 | グローバス メディカル インコーポレイティッド | Controlling surgical robots to avoid robot arm collisions |
DE102018116558A1 (en) * | 2018-07-09 | 2020-01-09 | Aesculap Ag | Medical technology instruments and processes |
CN112351722A (en) * | 2018-07-11 | 2021-02-09 | 奥林巴斯株式会社 | Endoscope system, endoscope calibration method, and endoscope control device |
WO2020020433A1 (en) * | 2018-07-23 | 2020-01-30 | Brainlab Ag | Planning of surgical anchor placement location data |
TWI695765B (en) * | 2018-07-31 | 2020-06-11 | 國立臺灣大學 | Robotic arm |
TWI733151B (en) * | 2018-08-01 | 2021-07-11 | 鈦隼生物科技股份有限公司 | Method, system and readable storage media of tracking patient position in operation |
US11298186B2 (en) * | 2018-08-02 | 2022-04-12 | Point Robotics Medtech Inc. | Surgery assistive system and method for obtaining surface information thereof |
US11409276B2 (en) * | 2018-08-17 | 2022-08-09 | Michael Isakov | Method and system for controlling robots within in an interactive arena and generating a virtual overlayed |
US11911111B2 (en) * | 2018-08-24 | 2024-02-27 | University Of Hawaii | Autonomous system and method for planning, tracking, and controlling the operation of steerable surgical devices |
EP3849656A4 (en) * | 2018-09-14 | 2022-04-27 | Neuralink Corp. | Device implantation using a cartridge |
KR20210130698A (en) | 2018-09-14 | 2021-11-01 | 뉴럴링크 코포레이션 | computer vision skills |
US11197728B2 (en) | 2018-09-17 | 2021-12-14 | Auris Health, Inc. | Systems and methods for concomitant medical procedures |
US11160672B2 (en) | 2018-09-24 | 2021-11-02 | Simplify Medical Pty Ltd | Robotic systems and methods for distraction in intervertebral disc prosthesis implantation |
US11648058B2 (en) | 2018-09-24 | 2023-05-16 | Simplify Medical Pty Ltd | Robotic system and method for bone preparation for intervertebral disc prosthesis implantation |
US11819424B2 (en) | 2018-09-24 | 2023-11-21 | Simplify Medical Pty Ltd | Robot assisted intervertebral disc prosthesis selection and implantation system |
WO2020075502A1 (en) | 2018-10-12 | 2020-04-16 | Sony Corporation | An operating room control system, method, and program |
US10940334B2 (en) | 2018-10-19 | 2021-03-09 | Sensus Healthcare, Inc. | Systems and methods for real time beam sculpting intra-operative-radiation-therapy treatment planning |
US20200129241A1 (en) * | 2018-10-30 | 2020-04-30 | Think Surgical, Inc. | Surgical markers for robotic surgery with reduced bone penetration |
EP3876860A1 (en) | 2018-11-06 | 2021-09-15 | Bono, Peter L. | Robotic surgical system and method |
WO2020097481A1 (en) * | 2018-11-08 | 2020-05-14 | Mako Surgical Corp. | Robotic spine surgery system and methods |
US11278360B2 (en) | 2018-11-16 | 2022-03-22 | Globus Medical, Inc. | End-effectors for surgical robotic systems having sealed optical components |
US11287874B2 (en) | 2018-11-17 | 2022-03-29 | Novarad Corporation | Using optical codes with augmented reality displays |
US11364084B2 (en) | 2018-11-21 | 2022-06-21 | Biosense Webster (Israel) Ltd. | Contact force compensation in a robot manipulator |
JP6882249B2 (en) * | 2018-11-29 | 2021-06-02 | ファナック株式会社 | Operation device for robots |
CN109240092B (en) * | 2018-11-30 | 2021-09-10 | 长春工业大学 | Reconfigurable modular flexible mechanical arm trajectory tracking control method based on multiple intelligent agents |
US11744655B2 (en) | 2018-12-04 | 2023-09-05 | Globus Medical, Inc. | Drill guide fixtures, cranial insertion fixtures, and related methods and robotic systems |
US11123142B2 (en) | 2018-12-06 | 2021-09-21 | Biosense Webster (Israel) Ltd. | Quick registration of coordinate systems for robotic surgery |
EP3666212A1 (en) * | 2018-12-14 | 2020-06-17 | Globus Medical, Inc. | Surgical robotic automation with tracking markers |
US11254009B2 (en) | 2018-12-20 | 2022-02-22 | Auris Health, Inc. | Systems and methods for robotic arm alignment and docking |
EP3908171A4 (en) | 2019-01-07 | 2022-09-14 | Virtual Incision Corporation | Robotically assisted surgical system and related devices and methods |
EP3909516A4 (en) * | 2019-01-11 | 2022-03-02 | Toppan Printing Co., Ltd. | Seal unit for use in inspection |
US20200237459A1 (en) * | 2019-01-25 | 2020-07-30 | Biosense Webster (Israel) Ltd. | Flexible multi-coil tracking sensor |
US11426242B2 (en) | 2019-01-30 | 2022-08-30 | Medtronic Navigation, Inc. | System and method for registration between coordinate systems and navigation of selected members |
US11911110B2 (en) * | 2019-01-30 | 2024-02-27 | Medtronic Navigation, Inc. | System and method for registration between coordinate systems and navigation of selected members |
EP3890644A4 (en) | 2019-02-08 | 2022-11-16 | Auris Health, Inc. | Robotically controlled clot manipulation and removal |
US11065065B2 (en) * | 2019-02-22 | 2021-07-20 | Warsaw Orthopedic, Inc. | Spinal implant system and methods of use |
US11162825B2 (en) * | 2019-02-26 | 2021-11-02 | Humanetics Innovative Solutions, Inc. | System and method for calibrating an optical fiber measurement system |
EP3937817A4 (en) * | 2019-03-12 | 2022-11-09 | Intuitive Surgical Operations, Inc. | Layered functionality for a user input mechanism in a computer-assisted surgical system |
KR102269772B1 (en) * | 2019-03-13 | 2021-06-28 | 큐렉소 주식회사 | End effector for surgical robot |
US11490981B2 (en) * | 2019-03-15 | 2022-11-08 | Cilag Gmbh International | Robotic surgical controls having feedback capabilities |
US11666401B2 (en) | 2019-03-15 | 2023-06-06 | Cilag Gmbh International | Input controls for robotic surgery |
US11471229B2 (en) * | 2019-03-15 | 2022-10-18 | Cilag Gmbh International | Robotic surgical systems with selectively lockable end effectors |
US11583350B2 (en) | 2019-03-15 | 2023-02-21 | Cilag Gmbh International | Jaw coordination of robotic surgical controls |
US11690690B2 (en) | 2019-03-15 | 2023-07-04 | Cilag Gmbh International | Segmented control inputs for surgical robotic systems |
US11701190B2 (en) | 2019-03-15 | 2023-07-18 | Cilag Gmbh International | Selectable variable response of shaft motion of surgical robotic systems |
US11284957B2 (en) | 2019-03-15 | 2022-03-29 | Cilag Gmbh International | Robotic surgical controls with force feedback |
US11806084B2 (en) | 2019-03-22 | 2023-11-07 | Globus Medical, Inc. | System for neuronavigation registration and robotic trajectory guidance, and related methods and devices |
US11317978B2 (en) | 2019-03-22 | 2022-05-03 | Globus Medical, Inc. | System for neuronavigation registration and robotic trajectory guidance, robotic surgery, and related methods and devices |
US20200297357A1 (en) * | 2019-03-22 | 2020-09-24 | Globus Medical, Inc. | System for neuronavigation registration and robotic trajectory guidance, robotic surgery, and related methods and devices |
US11382549B2 (en) | 2019-03-22 | 2022-07-12 | Globus Medical, Inc. | System for neuronavigation registration and robotic trajectory guidance, and related methods and devices |
US20200297426A1 (en) | 2019-03-22 | 2020-09-24 | Globus Medical, Inc. | System for neuronavigation registration and robotic trajectory guidance, and related methods and devices |
US11571265B2 (en) | 2019-03-22 | 2023-02-07 | Globus Medical Inc. | System for neuronavigation registration and robotic trajectory guidance, robotic surgery, and related methods and devices |
US11419616B2 (en) | 2019-03-22 | 2022-08-23 | Globus Medical, Inc. | System for neuronavigation registration and robotic trajectory guidance, robotic surgery, and related methods and devices |
US11925417B2 (en) | 2019-04-02 | 2024-03-12 | Medicrea International | Systems, methods, and devices for developing patient-specific spinal implants, treatments, operations, and/or procedures |
US11877801B2 (en) | 2019-04-02 | 2024-01-23 | Medicrea International | Systems, methods, and devices for developing patient-specific spinal implants, treatments, operations, and/or procedures |
FR3094627B1 (en) * | 2019-04-04 | 2022-09-02 | Quantum Surgical | Device for guiding a medical needle |
US11608148B2 (en) * | 2019-04-05 | 2023-03-21 | Fmc Technologies, Inc. | Submersible remote operated vehicle tool change control |
KR20210149805A (en) | 2019-04-08 | 2021-12-09 | 아우리스 헬스, 인코포레이티드 | Systems, Methods, and Workflows for Concurrent Procedures |
US11083501B2 (en) | 2019-04-24 | 2021-08-10 | Warsaw Orthopedic, Inc. | Surgical system and method |
EP3733112A1 (en) | 2019-05-03 | 2020-11-04 | Globus Medical, Inc. | System for robotic trajectory guidance for navigated biopsy needle |
WO2020243631A1 (en) * | 2019-05-30 | 2020-12-03 | Icahn School Of Medicine At Mount Sinai | Robot mounted camera registration and tracking system for orthopedic and neurological surgery |
US11129588B2 (en) * | 2019-06-19 | 2021-09-28 | Paul Adams | Ultrasound probe with an integrated needle assembly and a computer program product, a method and a system for providing a path for inserting a needle of the ultrasound probe |
CN112137722A (en) | 2019-06-26 | 2020-12-29 | 格罗伯斯医疗有限公司 | Surgical robotic systems, methods, and devices |
WO2020263520A1 (en) * | 2019-06-26 | 2020-12-30 | Auris Health, Inc. | Systems and methods for robotic arm alignment and docking |
CN114072258A (en) * | 2019-07-01 | 2022-02-18 | 瓦斯菲·阿希达法特 | Medical artificial intelligent robot arrangement for robot doctor |
US11628023B2 (en) | 2019-07-10 | 2023-04-18 | Globus Medical, Inc. | Robotic navigational system for interbody implants |
US11730560B2 (en) | 2019-07-12 | 2023-08-22 | Stuckenbrock Medizintechnik Gmbh | Fast-action clamping device with locking mechanism, and surgical device |
WO2021016066A1 (en) * | 2019-07-19 | 2021-01-28 | Mako Surgical Corp. | Tool assembly and methods for robotic-assisted surgery |
US10820871B1 (en) | 2019-08-09 | 2020-11-03 | GE Precision Healthcare LLC | Mobile X-ray imaging system including a parallel robotic structure |
WO2021030536A1 (en) * | 2019-08-13 | 2021-02-18 | Duluth Medical Technologies Inc. | Robotic surgical methods and apparatuses |
KR20220058918A (en) | 2019-08-30 | 2022-05-10 | 아우리스 헬스, 인코포레이티드 | Instrument image reliability system and method |
CN114340542B (en) | 2019-08-30 | 2023-07-21 | 奥瑞斯健康公司 | Systems and methods for weight-based registration of position sensors |
KR20220056220A (en) | 2019-09-03 | 2022-05-04 | 아우리스 헬스, 인코포레이티드 | Electromagnetic Distortion Detection and Compensation |
US11612440B2 (en) | 2019-09-05 | 2023-03-28 | Nuvasive, Inc. | Surgical instrument tracking devices and related methods |
US11234780B2 (en) | 2019-09-10 | 2022-02-01 | Auris Health, Inc. | Systems and methods for kinematic optimization with shared robotic degrees-of-freedom |
US11958183B2 (en) | 2019-09-19 | 2024-04-16 | The Research Foundation For The State University Of New York | Negotiation-based human-robot collaboration via augmented reality |
WO2021058087A1 (en) * | 2019-09-24 | 2021-04-01 | Brainlab Ag | Method and system for projecting an incision marker onto a patient |
US20210093407A1 (en) * | 2019-09-26 | 2021-04-01 | Auris Health, Inc. | Systems and methods for collision avoidance using object models |
US10959792B1 (en) | 2019-09-26 | 2021-03-30 | Auris Health, Inc. | Systems and methods for collision detection and avoidance |
US11123870B2 (en) * | 2019-09-27 | 2021-09-21 | HighRes Biosolutions, Inc. | Robotic transport system and method therefor |
EP4041115A1 (en) * | 2019-10-10 | 2022-08-17 | Intersect ENT International GmbH | Registration method and navigation system |
EP3824839A1 (en) | 2019-11-19 | 2021-05-26 | Koninklijke Philips N.V. | Robotic positioning of a device |
JP7239545B2 (en) * | 2019-11-26 | 2023-03-14 | グローバス メディカル インコーポレイティッド | Systems for neuronavigation registration and robotic trajectory guidance, robotic surgery, and related methods and devices |
US11769251B2 (en) | 2019-12-26 | 2023-09-26 | Medicrea International | Systems and methods for medical image analysis |
EP4084721A4 (en) | 2019-12-31 | 2024-01-03 | Auris Health Inc | Anatomical feature identification and targeting |
EP4084722A4 (en) | 2019-12-31 | 2024-01-10 | Auris Health Inc | Alignment interfaces for percutaneous access |
CN114901192A (en) | 2019-12-31 | 2022-08-12 | 奥瑞斯健康公司 | Alignment technique for percutaneous access |
CN111012456B (en) * | 2020-01-06 | 2021-01-19 | 赛诺联合医疗科技(北京)有限公司 | Ablation operation automatic positioning and needle inserting device based on image guide equipment |
CN113081269A (en) | 2020-01-08 | 2021-07-09 | 格罗伯斯医疗有限公司 | Surgical robotic system for performing surgery on anatomical features of a patient |
WO2021146503A1 (en) | 2020-01-16 | 2021-07-22 | Spinal Innovations, Llc | Pressure activated surgical tool for use in spinal decompression procedures and methods of using the same |
US11237627B2 (en) | 2020-01-16 | 2022-02-01 | Novarad Corporation | Alignment of medical images in augmented reality displays |
US11154405B2 (en) | 2020-01-23 | 2021-10-26 | Globus Medical, Inc. | Articulating expandable interbody fusions devices |
US20210251591A1 (en) * | 2020-02-17 | 2021-08-19 | Globus Medical, Inc. | System and method of determining optimal 3-dimensional position and orientation of imaging device for imaging patient bones |
US11207150B2 (en) * | 2020-02-19 | 2021-12-28 | Globus Medical, Inc. | Displaying a virtual model of a planned instrument attachment to ensure correct selection of physical instrument attachment |
US20210275257A1 (en) * | 2020-03-06 | 2021-09-09 | Covidien Lp | Uterine manipulator including position sensor |
EP3881791A1 (en) | 2020-03-18 | 2021-09-22 | Globus Medical, Inc. | System for neuronavigation registration and robotic trajectory guidance, and related methods and devices |
JP7006714B2 (en) * | 2020-03-23 | 2022-01-24 | カシオ計算機株式会社 | Positioning system, position measuring device, position measuring method and program |
US20210298830A1 (en) * | 2020-03-25 | 2021-09-30 | Covidien Lp | Robotic surgical system and methods of use thereof |
US20210322112A1 (en) * | 2020-04-21 | 2021-10-21 | Mazor Robotics Ltd. | System and method for aligning an imaging device |
CN111546378B (en) * | 2020-04-22 | 2022-05-17 | 哈尔滨工业大学 | Rapid collision detection method for space manipulator |
US20230184606A1 (en) * | 2020-04-28 | 2023-06-15 | Xerxes Maximilian Libsch | System for substrate detection and mapping using force sensing and magnetic coupling |
US11253216B2 (en) | 2020-04-28 | 2022-02-22 | Globus Medical Inc. | Fixtures for fluoroscopic imaging systems and related navigation systems and methods |
CN111544120A (en) * | 2020-05-14 | 2020-08-18 | 上海交通大学医学院附属第九人民医院 | Precision measuring device for surgical robot navigation system |
US20210378756A1 (en) * | 2020-06-09 | 2021-12-09 | Globus Medical, Inc. | Surgical object tracking in visible light via fiducial seeding and synthetic image registration |
CN113799115B (en) * | 2020-06-11 | 2023-03-24 | 台达电子工业股份有限公司 | Coordinate correction method of robot arm |
EP4171427A1 (en) | 2020-06-29 | 2023-05-03 | Auris Health, Inc. | Systems and methods for detecting contact between a link and an external object |
EP4171428A1 (en) | 2020-06-30 | 2023-05-03 | Auris Health, Inc. | Robotic medical system with collision proximity indicators |
US11357586B2 (en) | 2020-06-30 | 2022-06-14 | Auris Health, Inc. | Systems and methods for saturated robotic movement |
DE102020208325A1 (en) * | 2020-07-02 | 2022-01-05 | Siemens Healthcare Gmbh | Method and system for creating a navigation plan for a catheter with a robot |
US11357640B2 (en) | 2020-07-08 | 2022-06-14 | Globus Medical Inc. | Expandable interbody fusions devices |
US11877807B2 (en) | 2020-07-10 | 2024-01-23 | Globus Medical, Inc | Instruments for navigated orthopedic surgeries |
US20220016782A1 (en) * | 2020-07-20 | 2022-01-20 | Humatics Corporation | Radio-frequency systems and methods for co-localization of multiple devices and/or people |
US20220047297A1 (en) * | 2020-08-12 | 2022-02-17 | Mazor Robotics Ltd. | Cannulation devices, systems, and methods |
US11737831B2 (en) * | 2020-09-02 | 2023-08-29 | Globus Medical Inc. | Surgical object tracking template generation for computer assisted navigation during surgical procedure |
CN111905273B (en) * | 2020-09-21 | 2022-06-10 | 哈尔滨市全科医疗技术发展有限责任公司 | Rotary movable far infrared physiotherapy instrument |
US20220125405A1 (en) * | 2020-10-22 | 2022-04-28 | Yanwei Wang | System and Methods for Ultrasound Imaging with Modularized Frontend and Personal Computer System |
US11911112B2 (en) | 2020-10-27 | 2024-02-27 | Globus Medical, Inc. | Robotic navigational system |
WO2022094090A1 (en) * | 2020-10-30 | 2022-05-05 | Mako Surgical Corp. | Robotic surgical system with recovery alignment |
US11941814B2 (en) * | 2020-11-04 | 2024-03-26 | Globus Medical Inc. | Auto segmentation using 2-D images taken during 3-D imaging spin |
US11813746B2 (en) | 2020-12-30 | 2023-11-14 | Cilag Gmbh International | Dual driving pinion crosscheck |
CN116669652A (en) * | 2020-12-30 | 2023-08-29 | 奥瑞斯健康公司 | Method and device for transmitting emergency signals via differential signal lines in a robotic system |
CN112869778B (en) * | 2021-01-11 | 2022-04-15 | 北京航空航天大学 | Variable rigidity throat swab holder |
US20220249172A1 (en) * | 2021-02-05 | 2022-08-11 | Biosense Webster (Israel) Ltd. | System and method for alignment of balloon ablation catheter in pulmonary vein |
WO2022170201A1 (en) * | 2021-02-08 | 2022-08-11 | Medtronic Navigation, Inc | Systems and methods for intraoperative re-registration |
DE102021201727A1 (en) | 2021-02-24 | 2022-08-25 | Siemens Healthcare Gmbh | Device for moving a medical object and method for providing a correction specification |
DE102021201729A1 (en) | 2021-02-24 | 2022-08-25 | Siemens Healthcare Gmbh | Device for moving a medical object and method for providing a control specification |
WO2022197550A1 (en) * | 2021-03-15 | 2022-09-22 | Relievant Medsystems, Inc. | Robotic spine systems and robotic-assisted methods for tissue modulation |
DE102021202753A1 (en) | 2021-03-22 | 2022-09-22 | Siemens Healthcare Gmbh | Providing a corrected data set |
US11844583B2 (en) | 2021-03-31 | 2023-12-19 | Moon Surgical Sas | Co-manipulation surgical system having an instrument centering mode for automatic scope movements |
US11819302B2 (en) | 2021-03-31 | 2023-11-21 | Moon Surgical Sas | Co-manipulation surgical system having user guided stage control |
CA3212211A1 (en) | 2021-03-31 | 2022-10-06 | David Paul Noonan | Co-manipulation surgical system for use with surgical instruments for performing laparoscopic surgery |
US11812938B2 (en) | 2021-03-31 | 2023-11-14 | Moon Surgical Sas | Co-manipulation surgical system having a coupling mechanism removeably attachable to surgical instruments |
US11832909B2 (en) | 2021-03-31 | 2023-12-05 | Moon Surgical Sas | Co-manipulation surgical system having actuatable setup joints |
DE102021204728A1 (en) | 2021-05-10 | 2022-11-10 | Siemens Healthcare Gmbh | Method and system for providing a corrected data set |
EP4349295A1 (en) * | 2021-06-02 | 2024-04-10 | Shanghai Microport Medbot (Group) Co., Ltd. | Remote center of motion follow-up adjustment system for support apparatus, intraoperative remote center of motion adjustment method, readable storage medium and surgical robot system |
CN113303848B (en) * | 2021-06-02 | 2022-10-18 | 南充市中心医院 | Neurosurgery posterior skull distraction drainage device |
EP4104786A1 (en) | 2021-06-15 | 2022-12-21 | Stryker European Operations Limited | Technique for determining poses of tracked vertebrae |
US11896445B2 (en) | 2021-07-07 | 2024-02-13 | Augmedics Ltd. | Iliac pin and adapter |
US11925426B2 (en) | 2021-07-16 | 2024-03-12 | DePuy Synthes Products, Inc. | Surgical robot with anti-skive feature |
US11534309B1 (en) | 2021-07-20 | 2022-12-27 | Globus Medical Inc. | Interlaminar lumbar interbody fusion implants, intradiscal implants, instruments, and methods |
US20230023449A1 (en) | 2021-07-20 | 2023-01-26 | Globus Medical, Inc. | Interlaminar lumbar interbody fusion system and associated robotic systems |
CN113397617B (en) * | 2021-08-04 | 2022-11-01 | 平昌县人民医院 | Clinical operation tractive equipment of using of department of general surgery |
US20230045275A1 (en) * | 2021-08-05 | 2023-02-09 | GE Precision Healthcare LLC | Methods and system for guided device insertion during medical imaging |
US20230083538A1 (en) * | 2021-09-15 | 2023-03-16 | Globus Medical, Inc. | Robot surgical platform for cranial surgery |
US20230135286A1 (en) * | 2021-11-01 | 2023-05-04 | Medtronic Navigation, Inc. | Systems, devices, and methods for tracking one or more surgical landmarks |
US20230145909A1 (en) * | 2021-11-05 | 2023-05-11 | Avent, Inc. | Configurable System and Method for Indicating Deviation from a Medical Device Placement Pathway |
CN114129263B (en) * | 2021-11-29 | 2023-07-25 | 武汉联影智融医疗科技有限公司 | Surgical robot path planning method, system, equipment and storage medium |
US11911115B2 (en) * | 2021-12-20 | 2024-02-27 | Globus Medical Inc. | Flat panel registration fixture and method of using same |
CN114191108A (en) * | 2021-12-20 | 2022-03-18 | 王超 | General anesthesia operation vehicle |
CN114102612B (en) * | 2022-01-24 | 2022-05-03 | 河北工业大学 | Robot tail end path contour error control method |
EP4238523A1 (en) * | 2022-03-01 | 2023-09-06 | Stryker European Operations Limited | Technique for determining a marker arrangement that defines marker positions of a tracker |
CN114587560B (en) * | 2022-03-16 | 2023-08-25 | 燕山大学 | Remote control injection device for bone cement and injection method thereof |
GB2621584A (en) * | 2022-08-15 | 2024-02-21 | Cmr Surgical Ltd | Alarm system for a surgical robot |
WO2024039796A1 (en) * | 2022-08-19 | 2024-02-22 | Method Ai, Inc. | Surgical procedure segmentation |
WO2024041730A1 (en) * | 2022-08-23 | 2024-02-29 | Abb Schweiz Ag | Method of determining relationship between robot coordinate system and external surface |
US11839442B1 (en) | 2023-01-09 | 2023-12-12 | Moon Surgical Sas | Co-manipulation surgical system for use with surgical instruments for performing laparoscopic surgery while estimating hold force |
CN116158851B (en) * | 2023-03-01 | 2024-03-01 | 哈尔滨工业大学 | Scanning target positioning system and method of medical remote ultrasonic automatic scanning robot |
CN116839482B (en) * | 2023-08-23 | 2023-12-12 | 成都利尼科医学技术发展有限公司 | Mechanical isocenter measuring device and method for medical accelerator |
CN116983058A (en) * | 2023-09-27 | 2023-11-03 | 北京红立方医疗设备有限公司 | Auxiliary medical instrument for implantation of biological materials in minimally invasive surgery |
Family Cites Families (700)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE2614083B2 (en) | 1976-04-01 | 1979-02-08 | Siemens Ag, 1000 Berlin Und 8000 Muenchen | X-ray film device for the production of transverse slice images |
US4979949A (en) * | 1988-04-26 | 1990-12-25 | The Board Of Regents Of The University Of Washington | Robot-aided system for surgery |
US5354314A (en) | 1988-12-23 | 1994-10-11 | Medical Instrumentation And Diagnostics Corporation | Three-dimensional beam localization apparatus and microscope for stereotactic diagnoses or surgery mounted on robotic type arm |
US5086401A (en) * | 1990-05-11 | 1992-02-04 | International Business Machines Corporation | Image-directed robotic system for precise robotic surgery including redundant consistency checking |
US5246010A (en) | 1990-12-11 | 1993-09-21 | Biotrine Corporation | Method and apparatus for exhalation analysis |
US5417210A (en) | 1992-05-27 | 1995-05-23 | International Business Machines Corporation | System and method for augmentation of endoscopic surgery |
US5184601A (en) * | 1991-08-05 | 1993-02-09 | Putman John M | Endoscope stabilizer |
CN1071291A (en) | 1991-09-30 | 1993-04-21 | 莫托罗拉公司 | The portable communications receiver that has compact virtual image display |
US6963792B1 (en) | 1992-01-21 | 2005-11-08 | Sri International | Surgical method |
DE69331789T2 (en) | 1992-01-21 | 2003-03-13 | Stanford Res Inst Int | Endoscopic surgical instrument |
US5631973A (en) | 1994-05-05 | 1997-05-20 | Sri International | Method for telemanipulation with telepresence |
US5657429A (en) | 1992-08-10 | 1997-08-12 | Computer Motion, Inc. | Automated endoscope system optimal positioning |
US5397323A (en) | 1992-10-30 | 1995-03-14 | International Business Machines Corporation | Remote center-of-motion robot for surgery |
US5732703A (en) * | 1992-11-30 | 1998-03-31 | The Cleveland Clinic Foundation | Stereotaxy wand and tool guide |
US6406472B1 (en) * | 1993-05-14 | 2002-06-18 | Sri International, Inc. | Remote center positioner |
JP2665052B2 (en) | 1993-05-14 | 1997-10-22 | エスアールアイ インターナショナル | Remote center positioning device |
US5413454A (en) * | 1993-07-09 | 1995-05-09 | Movsesian; Peter | Mobile robotic arm |
US5531227A (en) * | 1994-01-28 | 1996-07-02 | Schneider Medical Technologies, Inc. | Imaging device and method |
US6096048A (en) * | 1994-04-20 | 2000-08-01 | Howard, Iii; Matthew A. | Noninvasive, reattachable skull fiducial marker system |
JP3378401B2 (en) | 1994-08-30 | 2003-02-17 | 株式会社日立メディコ | X-ray equipment |
US5492527A (en) | 1994-09-09 | 1996-02-20 | Linvatec Corporation | Arthroscopic shaver with rotatable collet and slide aspiration control valve |
US5803089A (en) * | 1994-09-15 | 1998-09-08 | Visualization Technology, Inc. | Position tracking and imaging system for use in medical applications |
US6646541B1 (en) | 1996-06-24 | 2003-11-11 | Computer Motion, Inc. | General purpose distributed operating room control system |
US6978166B2 (en) | 1994-10-07 | 2005-12-20 | Saint Louis University | System for use in displaying images of a body part |
DE69534862T2 (en) | 1994-10-07 | 2006-08-17 | St. Louis University | Surgical navigation arrangement including reference and location systems |
US6246898B1 (en) * | 1995-03-28 | 2001-06-12 | Sonometrics Corporation | Method for carrying out a medical procedure using a three-dimensional tracking and imaging system |
US5882206A (en) | 1995-03-29 | 1999-03-16 | Gillio; Robert G. | Virtual surgery system |
US5887121A (en) | 1995-04-21 | 1999-03-23 | International Business Machines Corporation | Method of constrained Cartesian control of robotic mechanisms with active and passive joints |
US6122541A (en) | 1995-05-04 | 2000-09-19 | Radionics, Inc. | Head band for frameless stereotactic registration |
US5576830A (en) | 1995-05-23 | 1996-11-19 | Eastman Kodak Company | Position sensor with half tone optical gradient wedge |
US5649956A (en) | 1995-06-07 | 1997-07-22 | Sri International | System and method for releasably holding a surgical instrument |
CA2809745C (en) | 1995-06-07 | 2014-09-09 | Sri International | Method for releasably holding a surgical instrument |
US5638819A (en) * | 1995-08-29 | 1997-06-17 | Manwaring; Kim H. | Method and apparatus for guiding an instrument to a target |
US5825982A (en) | 1995-09-15 | 1998-10-20 | Wright; James | Head cursor control interface for an automated endoscope system for optimal positioning |
US6351659B1 (en) * | 1995-09-28 | 2002-02-26 | Brainlab Med. Computersysteme Gmbh | Neuro-navigation system |
US5772594A (en) | 1995-10-17 | 1998-06-30 | Barrick; Earl F. | Fluoroscopic image guided orthopaedic surgery system with intraoperative registration |
US5682886A (en) * | 1995-12-26 | 1997-11-04 | Musculographics Inc | Computer-assisted surgical system |
US5855583A (en) | 1996-02-20 | 1999-01-05 | Computer Motion, Inc. | Method and apparatus for performing minimally invasive cardiac procedures |
SG64340A1 (en) | 1996-02-27 | 1999-04-27 | Inst Of Systems Science Nation | Curved surgical instruments and methods of mapping a curved path for stereotactic surgery |
US5636255A (en) * | 1996-03-05 | 1997-06-03 | Queen's University At Kingston | Method and apparatus for CT image registration |
US6167145A (en) | 1996-03-29 | 2000-12-26 | Surgical Navigation Technologies, Inc. | Bone navigation system |
US5799055A (en) * | 1996-05-15 | 1998-08-25 | Northwestern University | Apparatus and method for planning a stereotactic surgical procedure using coordinated fluoroscopy |
US5792135A (en) | 1996-05-20 | 1998-08-11 | Intuitive Surgical, Inc. | Articulated surgical instrument for performing minimally invasive surgery with enhanced dexterity and sensitivity |
US6167296A (en) | 1996-06-28 | 2000-12-26 | The Board Of Trustees Of The Leland Stanford Junior University | Method for volumetric image navigation |
US7302288B1 (en) | 1996-11-25 | 2007-11-27 | Z-Kat, Inc. | Tool position indicator |
US8529582B2 (en) | 1996-12-12 | 2013-09-10 | Intuitive Surgical Operations, Inc. | Instrument interface of a robotic surgical system |
US7727244B2 (en) | 1997-11-21 | 2010-06-01 | Intuitive Surgical Operation, Inc. | Sterile surgical drape |
US6205411B1 (en) | 1997-02-21 | 2001-03-20 | Carnegie Mellon University | Computer-assisted surgery planner and intra-operative guidance system |
US6012216A (en) | 1997-04-30 | 2000-01-11 | Ethicon, Inc. | Stand alone swage apparatus |
US5820559A (en) | 1997-03-20 | 1998-10-13 | Ng; Wan Sing | Computerized boundary estimation in medical images |
US5911449A (en) | 1997-04-30 | 1999-06-15 | Ethicon, Inc. | Semi-automated needle feed method and apparatus |
US6276471B1 (en) | 1997-06-06 | 2001-08-21 | EXPRESSO DEUTSCHLAND TRANSPOTGERäTE GMBH | Delivery cart |
US6231565B1 (en) | 1997-06-18 | 2001-05-15 | United States Surgical Corporation | Robotic arm DLUs for performing surgical tasks |
EP2362283B1 (en) | 1997-09-19 | 2015-11-25 | Massachusetts Institute Of Technology | Robotic apparatus |
US6226548B1 (en) | 1997-09-24 | 2001-05-01 | Surgical Navigation Technologies, Inc. | Percutaneous registration apparatus and method for use in computer-assisted surgical navigation |
US5951475A (en) | 1997-09-25 | 1999-09-14 | International Business Machines Corporation | Methods and apparatus for registering CT-scan data to multiple fluoroscopic images |
US5999837A (en) * | 1997-09-26 | 1999-12-07 | Picker International, Inc. | Localizing and orienting probe for view devices |
US5987960A (en) | 1997-09-26 | 1999-11-23 | Picker International, Inc. | Tool calibrator |
US6157853A (en) | 1997-11-12 | 2000-12-05 | Stereotaxis, Inc. | Method and apparatus using shaped field of repositionable magnet to guide implant |
US6212419B1 (en) | 1997-11-12 | 2001-04-03 | Walter M. Blume | Method and apparatus using shaped field of repositionable magnet to guide implant |
US6031888A (en) | 1997-11-26 | 2000-02-29 | Picker International, Inc. | Fluoro-assist feature for a diagnostic imaging device |
US6348058B1 (en) * | 1997-12-12 | 2002-02-19 | Surgical Navigation Technologies, Inc. | Image guided spinal surgery guide, system, and method for use thereof |
US6165170A (en) | 1998-01-29 | 2000-12-26 | International Business Machines Corporation | Laser dermablator and dermablation |
US7169141B2 (en) | 1998-02-24 | 2007-01-30 | Hansen Medical, Inc. | Surgical instrument |
FR2779339B1 (en) | 1998-06-09 | 2000-10-13 | Integrated Surgical Systems Sa | MATCHING METHOD AND APPARATUS FOR ROBOTIC SURGERY, AND MATCHING DEVICE COMPRISING APPLICATION |
US6477400B1 (en) | 1998-08-20 | 2002-11-05 | Sofamor Danek Holdings, Inc. | Fluoroscopic image guided orthopaedic surgery system with intraoperative registration |
DE19839825C1 (en) | 1998-09-01 | 1999-10-07 | Siemens Ag | Diagnostic X=ray device |
US6033415A (en) | 1998-09-14 | 2000-03-07 | Integrated Surgical Systems | System and method for performing image directed robotic orthopaedic procedures without a fiducial reference system |
DE19842798C1 (en) | 1998-09-18 | 2000-05-04 | Howmedica Leibinger Gmbh & Co | Calibration device |
US6340363B1 (en) | 1998-10-09 | 2002-01-22 | Surgical Navigation Technologies, Inc. | Image guided vertebral distractor and method for tracking the position of vertebrae |
US6659939B2 (en) | 1998-11-20 | 2003-12-09 | Intuitive Surgical, Inc. | Cooperative minimally invasive telesurgical system |
US8527094B2 (en) | 1998-11-20 | 2013-09-03 | Intuitive Surgical Operations, Inc. | Multi-user medical robotic system for collaboration or training in minimally invasive surgical procedures |
US6459926B1 (en) * | 1998-11-20 | 2002-10-01 | Intuitive Surgical, Inc. | Repositioning and reorientation of master/slave relationship in minimally invasive telesurgery |
US8600551B2 (en) * | 1998-11-20 | 2013-12-03 | Intuitive Surgical Operations, Inc. | Medical robotic system with operatively couplable simulator unit for surgeon training |
US6951535B2 (en) | 2002-01-16 | 2005-10-04 | Intuitive Surgical, Inc. | Tele-medicine system that transmits an entire state of a subsystem |
US6325808B1 (en) | 1998-12-08 | 2001-12-04 | Advanced Realtime Control Systems, Inc. | Robotic system, docking station, and surgical tool for collaborative control in minimally invasive surgery |
US7125403B2 (en) | 1998-12-08 | 2006-10-24 | Intuitive Surgical | In vivo accessories for minimally invasive robotic surgery |
US6322567B1 (en) | 1998-12-14 | 2001-11-27 | Integrated Surgical Systems, Inc. | Bone motion tracking system |
US6451027B1 (en) | 1998-12-16 | 2002-09-17 | Intuitive Surgical, Inc. | Devices and methods for moving an image capture device in telesurgical systems |
US7016457B1 (en) | 1998-12-31 | 2006-03-21 | General Electric Company | Multimode imaging system for generating high quality images |
JP3270834B2 (en) | 1999-01-27 | 2002-04-02 | ファイザー・プロダクツ・インク | Heteroaromatic bicyclic derivatives useful as anticancer agents |
DE19905974A1 (en) | 1999-02-12 | 2000-09-07 | Siemens Ag | Computer tomography scanning method using multi-line detector |
US6560354B1 (en) | 1999-02-16 | 2003-05-06 | University Of Rochester | Apparatus and method for registration of images to physical space using a weighted combination of points and surfaces |
US6692090B1 (en) | 1999-03-08 | 2004-02-17 | Continental Teves, Ag & Co. Ohg | Method of maintaining the effect of the engine brake |
US6778850B1 (en) | 1999-03-16 | 2004-08-17 | Accuray, Inc. | Frameless radiosurgery treatment system and method |
US6144875A (en) | 1999-03-16 | 2000-11-07 | Accuray Incorporated | Apparatus and method for compensating for respiratory and patient motion during treatment |
US6501981B1 (en) | 1999-03-16 | 2002-12-31 | Accuray, Inc. | Apparatus and method for compensating for respiratory and patient motions during treatment |
US6080161A (en) | 1999-03-19 | 2000-06-27 | Eaves, Iii; Felmont F. | Fastener and method for bone fixation |
US6470207B1 (en) | 1999-03-23 | 2002-10-22 | Surgical Navigation Technologies, Inc. | Navigational guidance via computer-assisted fluoroscopic imaging |
JP2000271110A (en) | 1999-03-26 | 2000-10-03 | Hitachi Medical Corp | Medical x-ray system |
US6594552B1 (en) | 1999-04-07 | 2003-07-15 | Intuitive Surgical, Inc. | Grip strength with tactile feedback for robotic surgery |
US6565554B1 (en) | 1999-04-07 | 2003-05-20 | Intuitive Surgical, Inc. | Friction compensation in a minimally invasive surgical apparatus |
US6424885B1 (en) | 1999-04-07 | 2002-07-23 | Intuitive Surgical, Inc. | Camera referenced control in a minimally invasive surgical apparatus |
US6301495B1 (en) | 1999-04-27 | 2001-10-09 | International Business Machines Corporation | System and method for intra-operative, image-based, interactive verification of a pre-operative surgical plan |
DE19927953A1 (en) | 1999-06-18 | 2001-01-11 | Siemens Ag | X=ray diagnostic apparatus |
US7203491B2 (en) | 2001-04-18 | 2007-04-10 | Space Data Corporation | Unmanned lighter-than-air safe termination and recovery methods |
US6314311B1 (en) | 1999-07-28 | 2001-11-06 | Picker International, Inc. | Movable mirror laser registration system |
DE19936364A1 (en) * | 1999-08-03 | 2001-02-15 | Siemens Ag | Identification and localisation of marks in a 3D medical scanning process |
US6788018B1 (en) | 1999-08-03 | 2004-09-07 | Intuitive Surgical, Inc. | Ceiling and floor mounted surgical robot set-up arms |
US8271130B2 (en) | 2009-03-09 | 2012-09-18 | Intuitive Surgical Operations, Inc. | Master controller having redundant degrees of freedom and added forces to create internal motion |
US8004229B2 (en) | 2005-05-19 | 2011-08-23 | Intuitive Surgical Operations, Inc. | Software center and highly configurable robotic systems for surgery and other uses |
US7594912B2 (en) | 2004-09-30 | 2009-09-29 | Intuitive Surgical, Inc. | Offset remote center manipulator for robotic surgery |
US6312435B1 (en) | 1999-10-08 | 2001-11-06 | Intuitive Surgical, Inc. | Surgical instrument with extended reach for use in minimally invasive surgery |
US8644907B2 (en) | 1999-10-28 | 2014-02-04 | Medtronic Navigaton, Inc. | Method and apparatus for surgical navigation |
US6499488B1 (en) | 1999-10-28 | 2002-12-31 | Winchester Development Associates | Surgical sensor |
US8239001B2 (en) | 2003-10-17 | 2012-08-07 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation |
US6235038B1 (en) | 1999-10-28 | 2001-05-22 | Medtronic Surgical Navigation Technologies | System for translation of electromagnetic and optical localization systems |
US6379302B1 (en) | 1999-10-28 | 2002-04-30 | Surgical Navigation Technologies Inc. | Navigation information overlay onto ultrasound imagery |
US7366562B2 (en) | 2003-10-17 | 2008-04-29 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation |
US6245028B1 (en) * | 1999-11-24 | 2001-06-12 | Marconi Medical Systems, Inc. | Needle biopsy system |
US20010036302A1 (en) | 1999-12-10 | 2001-11-01 | Miller Michael I. | Method and apparatus for cross modality image registration |
US7635390B1 (en) | 2000-01-14 | 2009-12-22 | Marctec, Llc | Joint replacement component having a modular articulating surface |
US6377011B1 (en) | 2000-01-26 | 2002-04-23 | Massachusetts Institute Of Technology | Force feedback user interface for minimally invasive surgical simulator and teleoperator and other similar apparatus |
WO2001056007A1 (en) | 2000-01-28 | 2001-08-02 | Intersense, Inc. | Self-referenced tracking |
US6725080B2 (en) * | 2000-03-01 | 2004-04-20 | Surgical Navigation Technologies, Inc. | Multiple cannula image guided tool for image guided procedures |
AU2001248161A1 (en) | 2000-03-15 | 2001-09-24 | Orthosoft Inc. | Automatic calibration system for computer-aided surgical instruments |
US6535756B1 (en) | 2000-04-07 | 2003-03-18 | Surgical Navigation Technologies, Inc. | Trajectory storage apparatus and method for surgical navigation system |
US6949189B2 (en) | 2000-04-20 | 2005-09-27 | Cuno Incorporated | Keyed filter assembly |
US6856827B2 (en) | 2000-04-28 | 2005-02-15 | Ge Medical Systems Global Technology Company, Llc | Fluoroscopic tracking and visualization system |
US6484049B1 (en) | 2000-04-28 | 2002-11-19 | Ge Medical Systems Global Technology Company, Llc | Fluoroscopic tracking and visualization system |
US6856826B2 (en) | 2000-04-28 | 2005-02-15 | Ge Medical Systems Global Technology Company, Llc | Fluoroscopic tracking and visualization system |
US6614453B1 (en) | 2000-05-05 | 2003-09-02 | Koninklijke Philips Electronics, N.V. | Method and apparatus for medical image display for surgical tool planning and navigation in clinical environments |
ATE221751T1 (en) * | 2000-05-09 | 2002-08-15 | Brainlab Ag | METHOD FOR REGISTERING A PATIENT DATA SET FROM AN IMAGING PROCESS IN NAVIGATION-ASSISTED SURGICAL PROCEDURES USING X-RAY IMAGE ASSIGNMENT |
US7085400B1 (en) * | 2000-06-14 | 2006-08-01 | Surgical Navigation Technologies, Inc. | System and method for image based sensor calibration |
US6645196B1 (en) | 2000-06-16 | 2003-11-11 | Intuitive Surgical, Inc. | Guided tool change |
US6782287B2 (en) | 2000-06-27 | 2004-08-24 | The Board Of Trustees Of The Leland Stanford Junior University | Method and apparatus for tracking a medical instrument based on image registration |
US6837892B2 (en) | 2000-07-24 | 2005-01-04 | Mazor Surgical Technologies Ltd. | Miniature bone-mounted surgical robot |
US6902560B1 (en) | 2000-07-27 | 2005-06-07 | Intuitive Surgical, Inc. | Roll-pitch-roll surgical tool |
DE10037491A1 (en) | 2000-08-01 | 2002-02-14 | Stryker Leibinger Gmbh & Co Kg | Process for three-dimensional visualization of structures inside the body |
US6823207B1 (en) | 2000-08-26 | 2004-11-23 | Ge Medical Systems Global Technology Company, Llc | Integrated fluoroscopic surgical navigation and imaging workstation with command protocol |
JP4022145B2 (en) | 2000-09-25 | 2007-12-12 | ゼット − キャット、インコーポレイテッド | Fluoroscopic superposition structure with optical and / or magnetic markers |
AU2002215822A1 (en) | 2000-10-23 | 2002-05-06 | Deutsches Krebsforschungszentrum Stiftung Des Offentlichen Rechts | Method, device and navigation aid for navigation during medical interventions |
US6582390B1 (en) * | 2000-11-08 | 2003-06-24 | Endovascular Technologies, Inc. | Dual lumen peel-away sheath introducer |
US6718194B2 (en) | 2000-11-17 | 2004-04-06 | Ge Medical Systems Global Technology Company, Llc | Computer assisted intramedullary rod surgery system with enhanced features |
US6666579B2 (en) | 2000-12-28 | 2003-12-23 | Ge Medical Systems Global Technology Company, Llc | Method and apparatus for obtaining and displaying computed tomography images using a fluoroscopy imaging system |
US6840938B1 (en) | 2000-12-29 | 2005-01-11 | Intuitive Surgical, Inc. | Bipolar cauterizing instrument |
WO2002060653A2 (en) * | 2001-01-29 | 2002-08-08 | The Acrobot Company Limited | Active-constraint robots |
US7043961B2 (en) | 2001-01-30 | 2006-05-16 | Z-Kat, Inc. | Tool calibrator and tracker system |
WO2002071369A1 (en) | 2001-03-01 | 2002-09-12 | Xitact S.A. | Device for determining the position of a device used for simulating surgical operations |
US7220262B1 (en) | 2001-03-16 | 2007-05-22 | Sdgi Holdings, Inc. | Spinal fixation system and related methods |
WO2002076302A2 (en) * | 2001-03-26 | 2002-10-03 | Lb Medical Gmbh | Method and device system for removing material or for working material |
FR2822674B1 (en) | 2001-04-03 | 2003-06-27 | Scient X | STABILIZED INTERSOMATIC MELTING SYSTEM FOR VERTEBERS |
US8046057B2 (en) | 2001-04-11 | 2011-10-25 | Clarke Dana S | Tissue structure identification in advance of instrument |
US8398634B2 (en) | 2002-04-18 | 2013-03-19 | Intuitive Surgical Operations, Inc. | Wristed robotic surgical tool for pluggable end-effectors |
US6783524B2 (en) | 2001-04-19 | 2004-08-31 | Intuitive Surgical, Inc. | Robotic surgical tool with ultrasound cauterizing and cutting instrument |
US7824401B2 (en) | 2004-10-08 | 2010-11-02 | Intuitive Surgical Operations, Inc. | Robotic tool with wristed monopolar electrosurgical end effectors |
US6994708B2 (en) | 2001-04-19 | 2006-02-07 | Intuitive Surgical | Robotic tool with monopolar electro-surgical scissors |
US6636757B1 (en) | 2001-06-04 | 2003-10-21 | Surgical Navigation Technologies, Inc. | Method and apparatus for electromagnetic navigation of a surgical probe near a metal object |
US7607440B2 (en) | 2001-06-07 | 2009-10-27 | Intuitive Surgical, Inc. | Methods and apparatus for surgical planning |
JP2005500096A (en) | 2001-06-13 | 2005-01-06 | ヴォリューム・インタラクションズ・プライヴェート・リミテッド | Guide system |
US6584339B2 (en) | 2001-06-27 | 2003-06-24 | Vanderbilt University | Method and apparatus for collecting and processing physical space data for use while performing image-guided surgery |
US7063705B2 (en) | 2001-06-29 | 2006-06-20 | Sdgi Holdings, Inc. | Fluoroscopic locator and registration device |
WO2003001987A2 (en) | 2001-06-29 | 2003-01-09 | Intuitive Surgical, Inc. | Platform link wrist mechanism |
US20040243147A1 (en) | 2001-07-03 | 2004-12-02 | Lipow Kenneth I. | Surgical robot and robotic controller |
ITMI20011759A1 (en) | 2001-08-09 | 2003-02-09 | Nuovo Pignone Spa | SCRAPER DEVICE FOR PISTON ROD OF ALTERNATIVE COMPRESSORS |
US7708741B1 (en) * | 2001-08-28 | 2010-05-04 | Marctec, Llc | Method of preparing bones for knee replacement surgery |
US6728599B2 (en) * | 2001-09-07 | 2004-04-27 | Computer Motion, Inc. | Modularity system for computer assisted surgery |
US6587750B2 (en) * | 2001-09-25 | 2003-07-01 | Intuitive Surgical, Inc. | Removable infinite roll master grip handle and touch sensor for robotic surgery |
US6619840B2 (en) | 2001-10-15 | 2003-09-16 | Koninklijke Philips Electronics N.V. | Interventional volume scanner |
US6839612B2 (en) | 2001-12-07 | 2005-01-04 | Institute Surgical, Inc. | Microwrist system for surgical procedures |
US7634306B2 (en) * | 2002-02-13 | 2009-12-15 | Kinamed, Inc. | Non-image, computer assisted navigation system for joint replacement surgery with modular implant system |
US6711431B2 (en) * | 2002-02-13 | 2004-03-23 | Kinamed, Inc. | Non-imaging, computer assisted navigation system for hip replacement surgery |
US6947786B2 (en) | 2002-02-28 | 2005-09-20 | Surgical Navigation Technologies, Inc. | Method and apparatus for perspective inversion |
US8996169B2 (en) | 2011-12-29 | 2015-03-31 | Mako Surgical Corp. | Neural monitor-based dynamic haptics |
US8010180B2 (en) * | 2002-03-06 | 2011-08-30 | Mako Surgical Corp. | Haptic guidance system and method |
US7206626B2 (en) * | 2002-03-06 | 2007-04-17 | Z-Kat, Inc. | System and method for haptic sculpting of physical objects |
US7831292B2 (en) * | 2002-03-06 | 2010-11-09 | Mako Surgical Corp. | Guidance system and method for surgical procedures with improved feedback |
US7108421B2 (en) | 2002-03-19 | 2006-09-19 | Breakaway Imaging, Llc | Systems and methods for imaging large field-of-view objects |
WO2003086714A2 (en) | 2002-04-05 | 2003-10-23 | The Trustees Of Columbia University In The City Of New York | Robotic scrub nurse |
AU2003234576A1 (en) * | 2002-05-17 | 2003-12-02 | Case Western Reserve University | Chemical shift markers for improved wireless fiducial marker tracking |
US7099428B2 (en) | 2002-06-25 | 2006-08-29 | The Regents Of The University Of Michigan | High spatial resolution X-ray computed tomography (CT) system |
US7248914B2 (en) | 2002-06-28 | 2007-07-24 | Stereotaxis, Inc. | Method of navigating medical devices in the presence of radiopaque material |
US7630752B2 (en) | 2002-08-06 | 2009-12-08 | Stereotaxis, Inc. | Remote control of medical devices using a virtual device interface |
US7231063B2 (en) | 2002-08-09 | 2007-06-12 | Intersense, Inc. | Fiducial detection system |
AU2003264048A1 (en) | 2002-08-09 | 2004-02-25 | Intersense, Inc. | Motion tracking system and method |
WO2004014244A2 (en) * | 2002-08-13 | 2004-02-19 | Microbotics Corporation | Microsurgical robot system |
US6892090B2 (en) | 2002-08-19 | 2005-05-10 | Surgical Navigation Technologies, Inc. | Method and apparatus for virtual endoscopy |
US7331967B2 (en) | 2002-09-09 | 2008-02-19 | Hansen Medical, Inc. | Surgical instrument coupling mechanism |
ES2204322B1 (en) | 2002-10-01 | 2005-07-16 | Consejo Sup. De Invest. Cientificas | FUNCTIONAL BROWSER. |
JP3821435B2 (en) | 2002-10-18 | 2006-09-13 | 松下電器産業株式会社 | Ultrasonic probe |
US9308002B2 (en) * | 2002-11-07 | 2016-04-12 | Crescent H Trust | Precise hip component positioning for hip replacement surgery |
US7599730B2 (en) * | 2002-11-19 | 2009-10-06 | Medtronic Navigation, Inc. | Navigation system for cardiac therapies |
US7318827B2 (en) | 2002-12-02 | 2008-01-15 | Aesculap Ag & Co. Kg | Osteotomy procedure |
US7319897B2 (en) | 2002-12-02 | 2008-01-15 | Aesculap Ag & Co. Kg | Localization device display method and apparatus |
US8814793B2 (en) | 2002-12-03 | 2014-08-26 | Neorad As | Respiration monitor |
US7386365B2 (en) | 2004-05-04 | 2008-06-10 | Intuitive Surgical, Inc. | Tool grip calibration for robotic surgery |
US7945021B2 (en) | 2002-12-18 | 2011-05-17 | Varian Medical Systems, Inc. | Multi-mode cone beam CT radiotherapy simulator and treatment machine with a flat panel imager |
US7505809B2 (en) | 2003-01-13 | 2009-03-17 | Mediguide Ltd. | Method and system for registering a first image with a second image relative to the body of a patient |
US7660623B2 (en) | 2003-01-30 | 2010-02-09 | Medtronic Navigation, Inc. | Six degree of freedom alignment display for medical procedures |
US7542791B2 (en) | 2003-01-30 | 2009-06-02 | Medtronic Navigation, Inc. | Method and apparatus for preplanning a surgical procedure |
WO2004069040A2 (en) | 2003-02-04 | 2004-08-19 | Z-Kat, Inc. | Method and apparatus for computer assistance with intramedullary nail procedure |
US6988009B2 (en) | 2003-02-04 | 2006-01-17 | Zimmer Technology, Inc. | Implant registration device for surgical navigation system |
US7083615B2 (en) | 2003-02-24 | 2006-08-01 | Intuitive Surgical Inc | Surgical tool having electrocautery energy supply conductor with inhibited current leakage |
US20110015521A1 (en) * | 2003-03-27 | 2011-01-20 | Boulder Innovation Group, Inc. | Means of Tracking Movement of Bodies During Medical Treatment |
JP4163991B2 (en) | 2003-04-30 | 2008-10-08 | 株式会社モリタ製作所 | X-ray CT imaging apparatus and imaging method |
US20060281991A1 (en) * | 2003-05-09 | 2006-12-14 | Fitzpatrick J M | Fiducial marker holder system for surgery |
US9060770B2 (en) | 2003-05-20 | 2015-06-23 | Ethicon Endo-Surgery, Inc. | Robotically-driven surgical instrument with E-beam driver |
US7194120B2 (en) | 2003-05-29 | 2007-03-20 | Board Of Regents, The University Of Texas System | Methods and systems for image-guided placement of implants |
US7171257B2 (en) | 2003-06-11 | 2007-01-30 | Accuray Incorporated | Apparatus and method for radiosurgery |
US9002518B2 (en) | 2003-06-30 | 2015-04-07 | Intuitive Surgical Operations, Inc. | Maximum torque driving of robotic surgical tools in robotic surgical systems |
US7960935B2 (en) | 2003-07-08 | 2011-06-14 | The Board Of Regents Of The University Of Nebraska | Robotic devices with agent delivery components and related methods |
US7042184B2 (en) | 2003-07-08 | 2006-05-09 | Board Of Regents Of The University Of Nebraska | Microrobot for surgical applications |
EP1646316B1 (en) | 2003-07-15 | 2009-12-16 | Koninklijke Philips Electronics N.V. | Computed tomography scanner with large gantry bore |
US7313430B2 (en) | 2003-08-28 | 2007-12-25 | Medtronic Navigation, Inc. | Method and apparatus for performing stereotactic surgery |
US7835778B2 (en) | 2003-10-16 | 2010-11-16 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation of a multiple piece construct for implantation |
US20050171558A1 (en) | 2003-10-17 | 2005-08-04 | Abovitz Rony A. | Neurosurgery targeting and delivery system for brain structures |
US7840253B2 (en) | 2003-10-17 | 2010-11-23 | Medtronic Navigation, Inc. | Method and apparatus for surgical navigation |
US20050096502A1 (en) | 2003-10-29 | 2005-05-05 | Khalili Theodore M. | Robotic surgical device |
US8206304B1 (en) * | 2003-12-16 | 2012-06-26 | Vascular Technology Incorporated | Doppler transceiver and probe for use in minimally invasive procedures |
US9393039B2 (en) | 2003-12-17 | 2016-07-19 | Brainlab Ag | Universal instrument or instrument set for computer guided surgery |
US7466303B2 (en) | 2004-02-10 | 2008-12-16 | Sunnybrook Health Sciences Center | Device and process for manipulating real and virtual objects in three-dimensional space |
WO2005086062A2 (en) | 2004-03-05 | 2005-09-15 | Depuy International Limited | Registration methods and apparatus |
US20060100610A1 (en) | 2004-03-05 | 2006-05-11 | Wallace Daniel T | Methods using a robotic catheter system |
EP1722705A2 (en) | 2004-03-10 | 2006-11-22 | Depuy International Limited | Orthopaedic operating systems, methods, implants and instruments |
US7657298B2 (en) | 2004-03-11 | 2010-02-02 | Stryker Leibinger Gmbh & Co. Kg | System, device, and method for determining a position of an object |
CA2556082A1 (en) * | 2004-03-12 | 2005-09-29 | Bracco Imaging S.P.A. | Accuracy evaluation of video-based augmented reality enhanced surgical navigation systems |
US7163542B2 (en) * | 2004-03-30 | 2007-01-16 | Synthes (U.S.A.) | Adjustable depth drill bit |
CN1938422A (en) * | 2004-04-08 | 2007-03-28 | 金克克国际有限公司 | Mutant alpha-amylases |
US8475495B2 (en) | 2004-04-08 | 2013-07-02 | Globus Medical | Polyaxial screw |
US8860753B2 (en) | 2004-04-13 | 2014-10-14 | University Of Georgia Research Foundation, Inc. | Virtual surgical system and methods |
KR100617974B1 (en) | 2004-04-22 | 2006-08-31 | 한국과학기술원 | Command-following laparoscopic system |
US20050245820A1 (en) * | 2004-04-28 | 2005-11-03 | Sarin Vineet K | Method and apparatus for verifying and correcting tracking of an anatomical structure during surgery |
US7567834B2 (en) | 2004-05-03 | 2009-07-28 | Medtronic Navigation, Inc. | Method and apparatus for implantation between two vertebral bodies |
US7379790B2 (en) | 2004-05-04 | 2008-05-27 | Intuitive Surgical, Inc. | Tool memory-based software upgrades for robotic surgery |
US8528565B2 (en) | 2004-05-28 | 2013-09-10 | St. Jude Medical, Atrial Fibrillation Division, Inc. | Robotic surgical system and method for automated therapy delivery |
US7974674B2 (en) | 2004-05-28 | 2011-07-05 | St. Jude Medical, Atrial Fibrillation Division, Inc. | Robotic surgical system and method for surface modeling |
FR2871363B1 (en) | 2004-06-15 | 2006-09-01 | Medtech Sa | ROBOTIZED GUIDING DEVICE FOR SURGICAL TOOL |
FR2872152B1 (en) | 2004-06-24 | 2006-08-11 | Inst Francais Du Petrole | HIERARCHISED POROSITY MATERIAL COMPRISING SILICON |
US7327865B2 (en) | 2004-06-30 | 2008-02-05 | Accuray, Inc. | Fiducial-less tracking with non-rigid image registration |
GB2415838B (en) * | 2004-07-01 | 2008-05-21 | Abb Offshore Systems Ltd | Cable connection |
ITMI20041448A1 (en) | 2004-07-20 | 2004-10-20 | Milano Politecnico | APPARATUS FOR THE MERGER AND NAVIGATION OF ECOGRAPHIC AND VOLUMETRIC IMAGES OF A PATIENT USING A COMBINATION OF ACTIVE AND PASSIVE OPTICAL MARKERS FOR THE LOCALIZATION OF ECHOGRAPHIC PROBES AND SURGICAL INSTRUMENTS COMPARED TO THE PATIENT |
US7440793B2 (en) | 2004-07-22 | 2008-10-21 | Sunita Chauhan | Apparatus and method for removing abnormal tissue |
CA2513202C (en) | 2004-07-23 | 2015-03-31 | Mehran Anvari | Multi-purpose robotic operating system and method |
GB2422759B (en) | 2004-08-05 | 2008-07-16 | Elekta Ab | Rotatable X-ray scan apparatus with cone beam offset |
US8182491B2 (en) | 2004-08-06 | 2012-05-22 | Depuy Spine, Inc. | Rigidly guided implant placement |
US8016835B2 (en) * | 2004-08-06 | 2011-09-13 | Depuy Spine, Inc. | Rigidly guided implant placement with control assist |
US7702379B2 (en) | 2004-08-25 | 2010-04-20 | General Electric Company | System and method for hybrid tracking in surgical navigation |
US7555331B2 (en) | 2004-08-26 | 2009-06-30 | Stereotaxis, Inc. | Method for surgical navigation utilizing scale-invariant registration between a navigation system and a localization system |
DE102004042489B4 (en) | 2004-08-31 | 2012-03-29 | Siemens Ag | Medical examination or treatment facility with associated method |
CA2581009C (en) | 2004-09-15 | 2011-10-04 | Synthes (U.S.A.) | Calibrating device |
EP1799107A1 (en) | 2004-10-06 | 2007-06-27 | Philips Intellectual Property & Standards GmbH | Computed tomography method |
US7831294B2 (en) | 2004-10-07 | 2010-11-09 | Stereotaxis, Inc. | System and method of surgical imagining with anatomical overlay for navigation of surgical devices |
US7318327B2 (en) | 2004-10-26 | 2008-01-15 | Respironics In-X, Inc. | Liquefying and storing a gas |
US7983733B2 (en) | 2004-10-26 | 2011-07-19 | Stereotaxis, Inc. | Surgical navigation using a three-dimensional user interface |
US20060111704A1 (en) * | 2004-11-22 | 2006-05-25 | Rox Medical, Inc. | Devices, systems, and methods for energy assisted arterio-venous fistula creation |
US7062006B1 (en) | 2005-01-19 | 2006-06-13 | The Board Of Trustees Of The Leland Stanford Junior University | Computed tomography with increased field of view |
US7837674B2 (en) | 2005-01-24 | 2010-11-23 | Intuitive Surgical Operations, Inc. | Compact counter balance for robotic surgical systems |
US7763015B2 (en) | 2005-01-24 | 2010-07-27 | Intuitive Surgical Operations, Inc. | Modular manipulator support for robotic surgery |
US20060184396A1 (en) | 2005-01-28 | 2006-08-17 | Dennis Charles L | System and method for surgical navigation |
US7231014B2 (en) | 2005-02-14 | 2007-06-12 | Varian Medical Systems Technologies, Inc. | Multiple mode flat panel X-ray imaging system |
CN101160104B (en) | 2005-02-22 | 2012-07-04 | 马科外科公司 | Haptic guidance system and method |
EP1861010B1 (en) | 2005-03-07 | 2019-12-25 | Hector O. Pacheco | Cannula for improved access to vertebral bodies for kyphoplasty, vertebroplasty, vertebral body biopsy or screw placement |
US8465771B2 (en) | 2005-03-30 | 2013-06-18 | The University Of Western Ontario | Anisotropic hydrogels |
US8496647B2 (en) | 2007-12-18 | 2013-07-30 | Intuitive Surgical Operations, Inc. | Ribbed force sensor |
US8375808B2 (en) | 2005-12-30 | 2013-02-19 | Intuitive Surgical Operations, Inc. | Force sensing for surgical instruments |
US9421019B2 (en) * | 2005-04-07 | 2016-08-23 | Omnilife Science, Inc. | Robotic guide assembly for use in computer-aided surgery |
US7720523B2 (en) | 2005-04-20 | 2010-05-18 | General Electric Company | System and method for managing power deactivation within a medical imaging system |
US8208988B2 (en) | 2005-05-13 | 2012-06-26 | General Electric Company | System and method for controlling a medical imaging device |
US8147503B2 (en) * | 2007-09-30 | 2012-04-03 | Intuitive Surgical Operations Inc. | Methods of locating and tracking robotic instruments in robotic surgical systems |
US8073528B2 (en) * | 2007-09-30 | 2011-12-06 | Intuitive Surgical Operations, Inc. | Tool tracking systems, methods and computer products for image guided surgery |
US8398541B2 (en) | 2006-06-06 | 2013-03-19 | Intuitive Surgical Operations, Inc. | Interactive user interfaces for robotic minimally invasive surgical systems |
KR101258912B1 (en) | 2005-06-06 | 2013-04-30 | 인튜어티브 서지컬 인코포레이티드 | Laparoscopic ultrasound robotic surgical system |
JP2007000406A (en) | 2005-06-24 | 2007-01-11 | Ge Medical Systems Global Technology Co Llc | X-ray ct method and x-ray ct apparatus |
US7840256B2 (en) | 2005-06-27 | 2010-11-23 | Biomet Manufacturing Corporation | Image guided tracking array and method |
US20070005002A1 (en) | 2005-06-30 | 2007-01-04 | Intuitive Surgical Inc. | Robotic surgical instruments for irrigation, aspiration, and blowing |
EP1906858B1 (en) * | 2005-07-01 | 2016-11-16 | Hansen Medical, Inc. | Robotic catheter system |
US20070038059A1 (en) | 2005-07-07 | 2007-02-15 | Garrett Sheffer | Implant and instrument morphing |
JP2007029232A (en) * | 2005-07-25 | 2007-02-08 | Hitachi Medical Corp | System for supporting endoscopic operation |
US20080302950A1 (en) | 2005-08-11 | 2008-12-11 | The Brigham And Women's Hospital, Inc. | System and Method for Performing Single Photon Emission Computed Tomography (Spect) with a Focal-Length Cone-Beam Collimation |
ATE555737T1 (en) * | 2005-08-11 | 2012-05-15 | Navotek Medical Ltd | LOCALIZATION OF A RADIOACTIVE SOURCE |
US7787699B2 (en) | 2005-08-17 | 2010-08-31 | General Electric Company | Real-time integration and recording of surgical image data |
US8800838B2 (en) | 2005-08-31 | 2014-08-12 | Ethicon Endo-Surgery, Inc. | Robotically-controlled cable-based surgical end effectors |
US7643862B2 (en) | 2005-09-15 | 2010-01-05 | Biomet Manufacturing Corporation | Virtual mouse for use in surgical navigation |
US20070073133A1 (en) | 2005-09-15 | 2007-03-29 | Schoenefeld Ryan J | Virtual mouse for use in surgical navigation |
US7835784B2 (en) | 2005-09-21 | 2010-11-16 | Medtronic Navigation, Inc. | Method and apparatus for positioning a reference frame |
US8079950B2 (en) | 2005-09-29 | 2011-12-20 | Intuitive Surgical Operations, Inc. | Autofocus and/or autoscaling in telesurgery |
US8224024B2 (en) | 2005-10-04 | 2012-07-17 | InterSense, LLC | Tracking objects with markers |
EP1932245A4 (en) | 2005-10-04 | 2010-07-07 | Eugene J Alexander | Method for synchronizing the operation of multiple devices for generating three dimensional surface models of moving objects |
US20070233185A1 (en) * | 2005-10-20 | 2007-10-04 | Thomas Anderson | Systems and methods for sealing a vascular opening |
US8182444B2 (en) * | 2005-11-04 | 2012-05-22 | Medrad, Inc. | Delivery of agents such as cells to tissue |
WO2007059233A2 (en) * | 2005-11-15 | 2007-05-24 | Johns Hopkins University | An active cannula for bio-sensing and surgical intervention |
US20090216113A1 (en) | 2005-11-17 | 2009-08-27 | Eric Meier | Apparatus and Methods for Using an Electromagnetic Transponder in Orthopedic Procedures |
US7711406B2 (en) | 2005-11-23 | 2010-05-04 | General Electric Company | System and method for detection of electromagnetic radiation by amorphous silicon x-ray detector for metal detection in x-ray imaging |
DE602005007509D1 (en) | 2005-11-24 | 2008-07-24 | Brainlab Ag | Medical referencing system with gamma camera |
US8190238B2 (en) * | 2005-12-09 | 2012-05-29 | Hansen Medical, Inc. | Robotic catheter system and methods |
US8498691B2 (en) * | 2005-12-09 | 2013-07-30 | Hansen Medical, Inc. | Robotic catheter system and methods |
US8672922B2 (en) | 2005-12-20 | 2014-03-18 | Intuitive Surgical Operations, Inc. | Wireless communication in a robotic surgical system |
US7762825B2 (en) | 2005-12-20 | 2010-07-27 | Intuitive Surgical Operations, Inc. | Electro-mechanical interfaces to mount robotic surgical arms |
US7689320B2 (en) | 2005-12-20 | 2010-03-30 | Intuitive Surgical Operations, Inc. | Robotic surgical system with joint motion controller adapted to reduce instrument tip vibrations |
US7955322B2 (en) | 2005-12-20 | 2011-06-07 | Intuitive Surgical Operations, Inc. | Wireless communication in a robotic surgical system |
US8182470B2 (en) | 2005-12-20 | 2012-05-22 | Intuitive Surgical Operations, Inc. | Telescoping insertion axis of a robotic surgical system |
US7819859B2 (en) | 2005-12-20 | 2010-10-26 | Intuitive Surgical Operations, Inc. | Control system for reducing internally generated frictional and inertial resistance to manual positioning of a surgical manipulator |
US8054752B2 (en) | 2005-12-22 | 2011-11-08 | Intuitive Surgical Operations, Inc. | Synchronous data communication |
ES2292327B1 (en) | 2005-12-26 | 2009-04-01 | Consejo Superior Investigaciones Cientificas | MINI CAMERA GAMMA AUTONOMA AND WITH LOCATION SYSTEM, FOR INTRACHIRURGICAL USE. |
US7907166B2 (en) | 2005-12-30 | 2011-03-15 | Intuitive Surgical Operations, Inc. | Stereo telestration for robotic surgery |
US7930065B2 (en) | 2005-12-30 | 2011-04-19 | Intuitive Surgical Operations, Inc. | Robotic surgery system including position sensors using fiber bragg gratings |
EP1965717B1 (en) | 2005-12-30 | 2012-05-16 | Intuitive Surgical Operations, Inc. | Surgical instrument with modular force sensor |
US7533892B2 (en) | 2006-01-05 | 2009-05-19 | Intuitive Surgical, Inc. | Steering system for heavy mobile medical equipment |
KR100731052B1 (en) | 2006-01-23 | 2007-06-22 | 한양대학교 산학협력단 | Bi-planar fluoroscopy guided robot system for a minimally invasive surgical |
US8142420B2 (en) | 2006-01-25 | 2012-03-27 | Intuitive Surgical Operations Inc. | Robotic arm with five-bar spherical linkage |
US8162926B2 (en) | 2006-01-25 | 2012-04-24 | Intuitive Surgical Operations Inc. | Robotic arm with five-bar spherical linkage |
US7845537B2 (en) | 2006-01-31 | 2010-12-07 | Ethicon Endo-Surgery, Inc. | Surgical instrument having recording capabilities |
US20110290856A1 (en) | 2006-01-31 | 2011-12-01 | Ethicon Endo-Surgery, Inc. | Robotically-controlled surgical instrument with force-feedback capabilities |
EP1815950A1 (en) | 2006-02-03 | 2007-08-08 | The European Atomic Energy Community (EURATOM), represented by the European Commission | Robotic surgical system for performing minimally invasive medical procedures |
US8219178B2 (en) * | 2007-02-16 | 2012-07-10 | Catholic Healthcare West | Method and system for performing invasive medical procedures using a surgical robot |
US8219177B2 (en) | 2006-02-16 | 2012-07-10 | Catholic Healthcare West | Method and system for performing invasive medical procedures using a surgical robot |
US10893912B2 (en) * | 2006-02-16 | 2021-01-19 | Globus Medical Inc. | Surgical tool systems and methods |
CA2642481C (en) | 2006-02-16 | 2016-04-05 | David W. Smith | System utilizing radio frequency signals for tracking and improving navigation of slender instruments during insertion into the body |
US10357184B2 (en) * | 2012-06-21 | 2019-07-23 | Globus Medical, Inc. | Surgical tool systems and method |
US8526688B2 (en) | 2006-03-09 | 2013-09-03 | General Electric Company | Methods and systems for registration of surgical navigation data and image data |
US20090048514A1 (en) * | 2006-03-09 | 2009-02-19 | Slender Medical Ltd. | Device for ultrasound monitored tissue treatment |
JP5248474B2 (en) | 2006-03-30 | 2013-07-31 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | Targeting method, targeting device, computer-readable medium, and program element |
US20070233238A1 (en) | 2006-03-31 | 2007-10-04 | Medtronic Vascular, Inc. | Devices for Imaging and Navigation During Minimally Invasive Non-Bypass Cardiac Procedures |
WO2007120744A2 (en) | 2006-04-14 | 2007-10-25 | William Beaumont Hospital | Scanning slot cone-beam computed tomography and scanning focus spot cone-beam computed tomography |
US8021310B2 (en) | 2006-04-21 | 2011-09-20 | Nellcor Puritan Bennett Llc | Work of breathing display for a ventilation system |
US8112292B2 (en) | 2006-04-21 | 2012-02-07 | Medtronic Navigation, Inc. | Method and apparatus for optimizing a therapy |
US7940999B2 (en) | 2006-04-24 | 2011-05-10 | Siemens Medical Solutions Usa, Inc. | System and method for learning-based 2D/3D rigid registration for image-guided surgery using Jensen-Shannon divergence |
EP1854425A1 (en) * | 2006-05-11 | 2007-11-14 | BrainLAB AG | Position determination for medical devices with redundant position measurement and weighting to prioritise measurements |
DE112007001214T5 (en) | 2006-05-16 | 2009-04-02 | Surgiceye Gmbh | Method and apparatus for 3D acquisition, 3D visualization and computer-guided operation with nuclear probes |
EP1857070A1 (en) * | 2006-05-18 | 2007-11-21 | BrainLAB AG | Contactless medical registration with distance measurement |
US9492237B2 (en) * | 2006-05-19 | 2016-11-15 | Mako Surgical Corp. | Method and apparatus for controlling a haptic device |
US7683565B2 (en) * | 2006-05-19 | 2010-03-23 | Mako Surgical Corp. | Method and apparatus for controlling a haptic device |
US20080004523A1 (en) | 2006-06-29 | 2008-01-03 | General Electric Company | Surgical tool guide |
US9718190B2 (en) * | 2006-06-29 | 2017-08-01 | Intuitive Surgical Operations, Inc. | Tool position and identification indicator displayed in a boundary area of a computer display screen |
DE102006032127B4 (en) | 2006-07-05 | 2008-04-30 | Aesculap Ag & Co. Kg | Calibration method and calibration device for a surgical referencing unit |
US20080013809A1 (en) | 2006-07-14 | 2008-01-17 | Bracco Imaging, Spa | Methods and apparatuses for registration in image guided surgery |
US9084622B2 (en) * | 2006-08-02 | 2015-07-21 | Omnitek Partners Llc | Automated laser-treatment system with real-time integrated 3D vision system for laser debridement and the like |
EP1886640B1 (en) | 2006-08-08 | 2009-11-18 | BrainLAB AG | Planning method and system for adjusting a free-shaped bone implant |
WO2008021671A2 (en) | 2006-08-17 | 2008-02-21 | Koninklijke Philips Electronics N. V. | Computed tomography image acquisition |
US20100158331A1 (en) * | 2006-08-22 | 2010-06-24 | Jacobs James P | System and method for determining absolute position using a multiple wavelength signal |
DE102006041033B4 (en) | 2006-09-01 | 2017-01-19 | Siemens Healthcare Gmbh | Method for reconstructing a three-dimensional image volume |
US8231610B2 (en) | 2006-09-06 | 2012-07-31 | National Cancer Center | Robotic surgical system for laparoscopic surgery |
US8532741B2 (en) | 2006-09-08 | 2013-09-10 | Medtronic, Inc. | Method and apparatus to optimize electrode placement for neurological stimulation |
US8150497B2 (en) | 2006-09-08 | 2012-04-03 | Medtronic, Inc. | System for navigating a planned procedure within a body |
US20080082109A1 (en) | 2006-09-08 | 2008-04-03 | Hansen Medical, Inc. | Robotic surgical system with forward-oriented field of view guide instrument navigation |
US8150498B2 (en) | 2006-09-08 | 2012-04-03 | Medtronic, Inc. | System for identification of anatomical landmarks |
US8248413B2 (en) | 2006-09-18 | 2012-08-21 | Stryker Corporation | Visual navigation system for endoscopic surgery |
CN103961130B (en) | 2006-09-25 | 2017-08-15 | 马佐尔机器人有限公司 | So that C-arm system adapts to the method to provide three-dimensional imaging information |
US8660635B2 (en) | 2006-09-29 | 2014-02-25 | Medtronic, Inc. | Method and apparatus for optimizing a computer assisted surgical procedure |
US8052688B2 (en) | 2006-10-06 | 2011-11-08 | Wolf Ii Erich | Electromagnetic apparatus and method for nerve localization during spinal surgery |
US20080144906A1 (en) | 2006-10-09 | 2008-06-19 | General Electric Company | System and method for video capture for fluoroscopy and navigation |
US8401620B2 (en) * | 2006-10-16 | 2013-03-19 | Perfint Healthcare Private Limited | Needle positioning apparatus and method |
US20080109012A1 (en) | 2006-11-03 | 2008-05-08 | General Electric Company | System, method and apparatus for tableside remote connections of medical instruments and systems using wireless communications |
US8551114B2 (en) | 2006-11-06 | 2013-10-08 | Human Robotics S.A. De C.V. | Robotic surgical device |
US20080108912A1 (en) | 2006-11-07 | 2008-05-08 | General Electric Company | System and method for measurement of clinical parameters of the knee for use during knee replacement surgery |
US20080108991A1 (en) | 2006-11-08 | 2008-05-08 | General Electric Company | Method and apparatus for performing pedicle screw fusion surgery |
US8682413B2 (en) | 2006-11-15 | 2014-03-25 | General Electric Company | Systems and methods for automated tracker-driven image selection |
US7935130B2 (en) | 2006-11-16 | 2011-05-03 | Intuitive Surgical Operations, Inc. | Two-piece end-effectors for robotic surgical tools |
CA2670261A1 (en) | 2006-11-16 | 2008-05-29 | Vanderbilt University | Apparatus and methods of compensating for organ deformation, registration of internal structures to images, and applications of same |
US8727618B2 (en) | 2006-11-22 | 2014-05-20 | Siemens Aktiengesellschaft | Robotic device and method for trauma patient diagnosis and therapy |
US7835557B2 (en) | 2006-11-28 | 2010-11-16 | Medtronic Navigation, Inc. | System and method for detecting status of imaging device |
US8320991B2 (en) | 2006-12-01 | 2012-11-27 | Medtronic Navigation Inc. | Portable electromagnetic navigation system |
US7683331B2 (en) | 2006-12-08 | 2010-03-23 | Rush University Medical Center | Single photon emission computed tomography (SPECT) system for cardiac imaging |
US7683332B2 (en) | 2006-12-08 | 2010-03-23 | Rush University Medical Center | Integrated single photon emission computed tomography (SPECT)/transmission computed tomography (TCT) system for cardiac imaging |
US8556807B2 (en) | 2006-12-21 | 2013-10-15 | Intuitive Surgical Operations, Inc. | Hermetically sealed distal sensor endoscope |
US20080177203A1 (en) | 2006-12-22 | 2008-07-24 | General Electric Company | Surgical navigation planning system and method for placement of percutaneous instrumentation and implants |
US8357165B2 (en) * | 2006-12-22 | 2013-01-22 | Depuy Products, Inc. | Reference array mounting bracket for use with a computer assisted orthopaedic surgery system |
DE102006061178A1 (en) | 2006-12-22 | 2008-06-26 | Siemens Ag | Medical system for carrying out and monitoring a minimal invasive intrusion, especially for treating electro-physiological diseases, has X-ray equipment and a control/evaluation unit |
US20080161680A1 (en) | 2006-12-29 | 2008-07-03 | General Electric Company | System and method for surgical navigation of motion preservation prosthesis |
US9220573B2 (en) | 2007-01-02 | 2015-12-29 | Medtronic Navigation, Inc. | System and method for tracking positions of uniform marker geometries |
US7780349B2 (en) * | 2007-01-03 | 2010-08-24 | James G. Schwade | Apparatus and method for robotic radiosurgery beam geometry quality assurance |
US8684253B2 (en) | 2007-01-10 | 2014-04-01 | Ethicon Endo-Surgery, Inc. | Surgical instrument with wireless communication between a control unit of a robotic system and remote sensor |
US8374673B2 (en) | 2007-01-25 | 2013-02-12 | Warsaw Orthopedic, Inc. | Integrated surgical navigational and neuromonitoring system having automated surgical assistance and control |
EP2124799B1 (en) | 2007-02-01 | 2012-10-31 | Interactive Neuroscience Center, Llc | Surgical navigation |
WO2008097853A2 (en) | 2007-02-02 | 2008-08-14 | Hansen Medical, Inc. | Mounting support assembly for suspending a medical instrument driver above an operating table |
WO2008101228A2 (en) * | 2007-02-15 | 2008-08-21 | Hansen Medical, Inc. | Robotic medical instrument system |
US8233963B2 (en) | 2007-02-19 | 2012-07-31 | Medtronic Navigation, Inc. | Automatic identification of tracked surgical devices using an electromagnetic localization system |
US8600478B2 (en) | 2007-02-19 | 2013-12-03 | Medtronic Navigation, Inc. | Automatic identification of instruments used with a surgical navigation system |
DE102007009017B3 (en) | 2007-02-23 | 2008-09-25 | Siemens Ag | Arrangement for supporting a percutaneous procedure |
US8784425B2 (en) * | 2007-02-28 | 2014-07-22 | Smith & Nephew, Inc. | Systems and methods for identifying landmarks on orthopedic implants |
US10039613B2 (en) | 2007-03-01 | 2018-08-07 | Surgical Navigation Technologies, Inc. | Method for localizing an imaging device with a surgical navigation system |
US8098914B2 (en) | 2007-03-05 | 2012-01-17 | Siemens Aktiengesellschaft | Registration of CT volumes with fluoroscopic images |
US20080228068A1 (en) | 2007-03-13 | 2008-09-18 | Viswanathan Raju R | Automated Surgical Navigation with Electro-Anatomical and Pre-Operative Image Data |
US8821511B2 (en) | 2007-03-15 | 2014-09-02 | General Electric Company | Instrument guide for use with a surgical navigation system |
US20080235052A1 (en) | 2007-03-19 | 2008-09-25 | General Electric Company | System and method for sharing medical information between image-guided surgery systems |
US8150494B2 (en) | 2007-03-29 | 2012-04-03 | Medtronic Navigation, Inc. | Apparatus for registering a physical space to image space |
US7879045B2 (en) | 2007-04-10 | 2011-02-01 | Medtronic, Inc. | System for guiding instruments having different sizes |
EP2142132B1 (en) | 2007-04-16 | 2012-09-26 | NeuroArm Surgical, Ltd. | System for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis |
WO2009044287A2 (en) * | 2007-04-16 | 2009-04-09 | The Governors Of The University Of Calgary | Methods, devices, and systems for automated movements involving medical robots |
US8311611B2 (en) | 2007-04-24 | 2012-11-13 | Medtronic, Inc. | Method for performing multiple registrations in a navigated procedure |
US8108025B2 (en) | 2007-04-24 | 2012-01-31 | Medtronic, Inc. | Flexible array for use in navigated surgery |
US20090012509A1 (en) | 2007-04-24 | 2009-01-08 | Medtronic, Inc. | Navigated Soft Tissue Penetrating Laser System |
US8301226B2 (en) | 2007-04-24 | 2012-10-30 | Medtronic, Inc. | Method and apparatus for performing a navigated procedure |
US8010177B2 (en) | 2007-04-24 | 2011-08-30 | Medtronic, Inc. | Intraoperative image registration |
US8062364B1 (en) | 2007-04-27 | 2011-11-22 | Knee Creations, Llc | Osteoarthritis treatment and device |
DE102007022122B4 (en) | 2007-05-11 | 2019-07-11 | Deutsches Zentrum für Luft- und Raumfahrt e.V. | Gripping device for a surgery robot arrangement |
US8057397B2 (en) | 2007-05-16 | 2011-11-15 | General Electric Company | Navigation and imaging system sychronized with respiratory and/or cardiac activity |
US20080287771A1 (en) | 2007-05-17 | 2008-11-20 | General Electric Company | Surgical navigation system with electrostatic shield |
US8934961B2 (en) | 2007-05-18 | 2015-01-13 | Biomet Manufacturing, Llc | Trackable diagnostic scope apparatus and methods of use |
US20080300477A1 (en) | 2007-05-30 | 2008-12-04 | General Electric Company | System and method for correction of automated image registration |
US20080300478A1 (en) | 2007-05-30 | 2008-12-04 | General Electric Company | System and method for displaying real-time state of imaged anatomy during a surgical procedure |
US8945148B2 (en) | 2007-06-13 | 2015-02-03 | Intuitive Surgical Operations, Inc. | Surgical system instrument manipulator |
US20080314181A1 (en) * | 2007-06-19 | 2008-12-25 | Bruce Schena | Robotic Manipulator with Remote Center of Motion and Compact Drive |
US9468412B2 (en) | 2007-06-22 | 2016-10-18 | General Electric Company | System and method for accuracy verification for image based surgical navigation |
EP3673855B1 (en) | 2007-07-12 | 2021-09-08 | Board of Regents of the University of Nebraska | Systems of actuation in robotic devices |
US7834484B2 (en) | 2007-07-16 | 2010-11-16 | Tyco Healthcare Group Lp | Connection cable and method for activating a voltage-controlled generator |
JP2009045428A (en) | 2007-07-25 | 2009-03-05 | Terumo Corp | Operating mechanism, medical manipulator and surgical robot system |
US8100950B2 (en) | 2007-07-27 | 2012-01-24 | The Cleveland Clinic Foundation | Oblique lumbar interbody fusion |
US8035685B2 (en) | 2007-07-30 | 2011-10-11 | General Electric Company | Systems and methods for communicating video data between a mobile imaging system and a fixed monitor system |
US8328818B1 (en) | 2007-08-31 | 2012-12-11 | Globus Medical, Inc. | Devices and methods for treating bone |
JP5452826B2 (en) | 2007-09-19 | 2014-03-26 | ロバーツ、ウォルター、エー. | Intraoperative radiotherapy applicator device with direct visualization robot |
US20090080737A1 (en) | 2007-09-25 | 2009-03-26 | General Electric Company | System and Method for Use of Fluoroscope and Computed Tomography Registration for Sinuplasty Navigation |
US8224484B2 (en) | 2007-09-30 | 2012-07-17 | Intuitive Surgical Operations, Inc. | Methods of user interface with alternate tool mode for robotic surgical tools |
US8452443B2 (en) * | 2007-10-01 | 2013-05-28 | Abb Research Ltd | Method for controlling a plurality of axes in an industrial robot system and an industrial robot system |
US9522046B2 (en) | 2010-08-23 | 2016-12-20 | Gip | Robotic surgery system |
US9265589B2 (en) * | 2007-11-06 | 2016-02-23 | Medtronic Navigation, Inc. | System and method for navigated drill guide |
US8462911B2 (en) | 2007-11-06 | 2013-06-11 | Koninklijke Philips Electronics N.V. | Nuclear medicine SPECT-CT machine with integrated asymmetric flat panel cone-beam CT and SPECT system |
DE102007055203A1 (en) | 2007-11-19 | 2009-05-20 | Kuka Roboter Gmbh | A robotic device, medical workstation and method for registering an object |
US8561473B2 (en) | 2007-12-18 | 2013-10-22 | Intuitive Surgical Operations, Inc. | Force sensor temperature compensation |
US8400094B2 (en) | 2007-12-21 | 2013-03-19 | Intuitive Surgical Operations, Inc. | Robotic surgical system with patient support |
US20100274120A1 (en) | 2007-12-21 | 2010-10-28 | Koninklijke Philips Electronics N.V. | Synchronous interventional scanner |
US8864798B2 (en) | 2008-01-18 | 2014-10-21 | Globus Medical, Inc. | Transverse connector |
KR20100120183A (en) | 2008-01-30 | 2010-11-12 | 더 트러스티이스 오브 콜롬비아 유니버시티 인 더 시티 오브 뉴욕 | Systems, devices, and methods for robot-assisted micro-surgical stenting |
US20090198121A1 (en) | 2008-02-01 | 2009-08-06 | Martin Hoheisel | Method and apparatus for coordinating contrast agent injection and image acquisition in c-arm computed tomography |
US8573465B2 (en) | 2008-02-14 | 2013-11-05 | Ethicon Endo-Surgery, Inc. | Robotically-controlled surgical end effector system with rotary actuated closure systems |
US8696458B2 (en) | 2008-02-15 | 2014-04-15 | Thales Visionix, Inc. | Motion tracking system and method using camera and non-camera sensors |
US7925653B2 (en) | 2008-02-27 | 2011-04-12 | General Electric Company | Method and system for accessing a group of objects in an electronic document |
US9220514B2 (en) * | 2008-02-28 | 2015-12-29 | Smith & Nephew, Inc. | System and method for identifying a landmark |
US20090223019A1 (en) | 2008-03-10 | 2009-09-10 | Bigfish Consulting Llc | Apparatus for wrapping around one or more target objects for gripping, as a hanging structure and/or to provide an extension |
US20090228019A1 (en) | 2008-03-10 | 2009-09-10 | Yosef Gross | Robotic surgical system |
US8282653B2 (en) | 2008-03-24 | 2012-10-09 | Board Of Regents Of The University Of Nebraska | System and methods for controlling surgical tool elements |
US8808164B2 (en) | 2008-03-28 | 2014-08-19 | Intuitive Surgical Operations, Inc. | Controlling a robotic surgical tool with a display monitor |
WO2009120122A1 (en) | 2008-03-28 | 2009-10-01 | Telefonaktiebolaget L M Ericsson (Publ) | Identification of a manipulated or defect base station during handover |
US8333755B2 (en) | 2008-03-31 | 2012-12-18 | Intuitive Surgical Operations, Inc. | Coupler to transfer controller motion from a robotic manipulator to an attached instrument |
US7843158B2 (en) | 2008-03-31 | 2010-11-30 | Intuitive Surgical Operations, Inc. | Medical robotic system adapted to inhibit motions resulting in excessive end effector forces |
US7886743B2 (en) | 2008-03-31 | 2011-02-15 | Intuitive Surgical Operations, Inc. | Sterile drape interface for robotic surgical instrument |
US9002076B2 (en) | 2008-04-15 | 2015-04-07 | Medtronic, Inc. | Method and apparatus for optimal trajectory planning |
US9345875B2 (en) | 2008-04-17 | 2016-05-24 | Medtronic, Inc. | Method and apparatus for cannula fixation for an array insertion tube set |
US8228368B2 (en) | 2008-04-26 | 2012-07-24 | Intuitive Surgical Operations, Inc. | Augmented stereoscopic visualization for a surgical robot using a captured fluorescence image and captured stereoscopic visible images |
US8364243B2 (en) | 2008-04-30 | 2013-01-29 | Nanosys, Inc. | Non-fouling surfaces for reflective spheres |
US20090281452A1 (en) * | 2008-05-02 | 2009-11-12 | Marcus Pfister | System and method for a medical procedure using computed tomography |
US9579161B2 (en) | 2008-05-06 | 2017-02-28 | Medtronic Navigation, Inc. | Method and apparatus for tracking a patient |
WO2009144730A1 (en) * | 2008-05-28 | 2009-12-03 | Technion Research & Development Foundation Ltd. | Ultrasound guided robot for flexible needle steering |
US20110022229A1 (en) | 2008-06-09 | 2011-01-27 | Bae Sang Jang | Master interface and driving method of surgical robot |
TW201004607A (en) | 2008-07-25 | 2010-02-01 | Been-Der Yang | Image guided navigation system and method thereof |
US8054184B2 (en) | 2008-07-31 | 2011-11-08 | Intuitive Surgical Operations, Inc. | Identification of surgical instrument attached to surgical robot |
US8771170B2 (en) | 2008-08-01 | 2014-07-08 | Microaccess, Inc. | Methods and apparatus for transesophageal microaccess surgery |
JP2010035984A (en) | 2008-08-08 | 2010-02-18 | Canon Inc | X-ray imaging apparatus |
US9248000B2 (en) | 2008-08-15 | 2016-02-02 | Stryker European Holdings I, Llc | System for and method of visualizing an interior of body |
US8500728B2 (en) | 2008-08-18 | 2013-08-06 | Encision, Inc. | Enhanced control systems including flexible shielding and support systems for electrosurgical applications |
DE102008041813B4 (en) | 2008-09-04 | 2013-06-20 | Carl Zeiss Microscopy Gmbh | Method for the depth analysis of an organic sample |
US7900524B2 (en) | 2008-09-09 | 2011-03-08 | Intersense, Inc. | Monitoring tools |
US8165658B2 (en) | 2008-09-26 | 2012-04-24 | Medtronic, Inc. | Method and apparatus for positioning a guide relative to a base |
US8073335B2 (en) | 2008-09-30 | 2011-12-06 | Intuitive Surgical Operations, Inc. | Operator input device for a robotic surgical system |
CN102177430B (en) | 2008-10-10 | 2014-04-02 | 皇家飞利浦电子股份有限公司 | Method and apparatus to improve CT image acquisition using a displaced geometry |
KR100944412B1 (en) | 2008-10-13 | 2010-02-25 | (주)미래컴퍼니 | Surgical slave robot |
WO2010044852A2 (en) | 2008-10-14 | 2010-04-22 | University Of Florida Research Foundation, Inc. | Imaging platform to provide integrated navigation capabilities for surgical guidance |
EP2355711B1 (en) | 2008-10-20 | 2019-09-11 | The Johns Hopkins University | Environment property estimation and graphical display |
EP2179703B1 (en) | 2008-10-21 | 2012-03-28 | BrainLAB AG | Integration of surgical instrument and display device for supporting image-based surgery |
KR101075363B1 (en) | 2008-10-31 | 2011-10-19 | 정창욱 | Surgical Robot System Having Tool for Minimally Invasive Surgery |
US20110150765A1 (en) | 2008-10-31 | 2011-06-23 | Searete Llc, A Limited Liability Corporation Of The State Of Delaware | Frozen compositions and methods for piercing a substrate |
US9033958B2 (en) | 2008-11-11 | 2015-05-19 | Perception Raisonnement Action En Medecine | Surgical robotic system |
TWI435705B (en) | 2008-11-20 | 2014-05-01 | Been Der Yang | Surgical position device and image guided navigation system using the same |
TWI498827B (en) | 2008-11-21 | 2015-09-01 | Verayo Inc | Non-networked rfid-puf authentication |
WO2010061810A1 (en) | 2008-11-27 | 2010-06-03 | 株式会社 日立メディコ | Radiation image pickup device |
US8483800B2 (en) | 2008-11-29 | 2013-07-09 | General Electric Company | Surgical navigation enabled imaging table environment |
WO2010064234A2 (en) | 2008-12-01 | 2010-06-10 | Mazor Surgical Technologies Ltd. | Robot guided oblique spinal stabilization |
ES2341079B1 (en) | 2008-12-11 | 2011-07-13 | Fundacio Clinic Per A La Recerca Biomedica | EQUIPMENT FOR IMPROVED VISION BY INFRARED VASCULAR STRUCTURES, APPLICABLE TO ASSIST PHYTOSCOPIC, LAPAROSCOPIC AND ENDOSCOPIC INTERVENTIONS AND SIGNAL TREATMENT PROCESS TO IMPROVE SUCH VISION. |
US8021393B2 (en) | 2008-12-12 | 2011-09-20 | Globus Medical, Inc. | Lateral spinous process spacer with deployable wings |
US8594841B2 (en) | 2008-12-31 | 2013-11-26 | Intuitive Surgical Operations, Inc. | Visual force feedback in a minimally invasive surgical procedure |
US8830224B2 (en) | 2008-12-31 | 2014-09-09 | Intuitive Surgical Operations, Inc. | Efficient 3-D telestration for local robotic proctoring |
US8374723B2 (en) | 2008-12-31 | 2013-02-12 | Intuitive Surgical Operations, Inc. | Obtaining force information in a minimally invasive surgical procedure |
US8184880B2 (en) | 2008-12-31 | 2012-05-22 | Intuitive Surgical Operations, Inc. | Robust sparse image matching for robotic surgery |
EP2586374B1 (en) | 2009-01-21 | 2015-03-18 | Koninklijke Philips N.V. | Method and apparatus for large field of view imaging and detection and compensation of motion artifacts |
WO2010085759A1 (en) * | 2009-01-23 | 2010-07-29 | Intersect Ent, Inc . | Devices and methods for dilating tissues |
US20100198052A1 (en) * | 2009-01-28 | 2010-08-05 | Kimble Jenkins | Mri-compatible articulating arms and related systems and methods |
US8611985B2 (en) | 2009-01-29 | 2013-12-17 | Imactis | Method and device for navigation of a surgical tool |
KR101038417B1 (en) | 2009-02-11 | 2011-06-01 | 주식회사 이턴 | Surgical robot system and control method thereof |
US8662801B2 (en) * | 2009-02-24 | 2014-03-04 | Black & Decker Inc. | Depth gauge for drill bit |
US8418073B2 (en) | 2009-03-09 | 2013-04-09 | Intuitive Surgical Operations, Inc. | User interfaces for electrosurgical tools in robotic surgical systems |
US8423182B2 (en) * | 2009-03-09 | 2013-04-16 | Intuitive Surgical Operations, Inc. | Adaptable integrated energy control system for electrosurgical tools in robotic surgical systems |
US8120301B2 (en) | 2009-03-09 | 2012-02-21 | Intuitive Surgical Operations, Inc. | Ergonomic surgeon control console in robotic surgical systems |
US8918207B2 (en) | 2009-03-09 | 2014-12-23 | Intuitive Surgical Operations, Inc. | Operator input device for a robotic surgical system |
US9737235B2 (en) | 2009-03-09 | 2017-08-22 | Medtronic Navigation, Inc. | System and method for image-guided navigation |
US20120053597A1 (en) | 2009-03-10 | 2012-03-01 | Mcmaster University | Mobile robotic surgical system |
US8335552B2 (en) | 2009-03-20 | 2012-12-18 | Medtronic, Inc. | Method and apparatus for instrument placement |
CN105342705A (en) | 2009-03-24 | 2016-02-24 | 伊顿株式会社 | Surgical robot system using augmented reality, and method for controlling same |
US20100249571A1 (en) | 2009-03-31 | 2010-09-30 | General Electric Company | Surgical navigation system with wireless magnetoresistance tracking sensors |
US8882803B2 (en) | 2009-04-01 | 2014-11-11 | Globus Medical, Inc. | Orthopedic clamp and extension rod |
WO2010122563A1 (en) | 2009-04-23 | 2010-10-28 | M.S.T. Medical Surgery Technologies Ltd. | Two-part endoscope surgical device |
US20100280363A1 (en) | 2009-04-24 | 2010-11-04 | Medtronic, Inc. | Electromagnetic Navigation of Medical Instruments for Cardiothoracic Surgery |
EP2424455A4 (en) * | 2009-04-27 | 2017-08-02 | Smith & Nephew, Inc. | System and method for identifying a landmark |
US8945147B2 (en) * | 2009-04-27 | 2015-02-03 | Smith & Nephew, Inc. | System and method for identifying a landmark |
US9031637B2 (en) * | 2009-04-27 | 2015-05-12 | Smith & Nephew, Inc. | Targeting an orthopaedic implant landmark |
EP2432372B1 (en) | 2009-05-18 | 2018-12-26 | Teleflex Medical Incorporated | Devices for performing minimally invasive surgery |
ES2388029B1 (en) | 2009-05-22 | 2013-08-13 | Universitat Politècnica De Catalunya | ROBOTIC SYSTEM FOR LAPAROSCOPIC SURGERY. |
CN101897593B (en) | 2009-05-26 | 2014-08-13 | 清华大学 | Computer chromatography imaging device and method |
US8121249B2 (en) | 2009-06-04 | 2012-02-21 | Virginia Tech Intellectual Properties, Inc. | Multi-parameter X-ray computed tomography |
WO2011013164A1 (en) | 2009-07-27 | 2011-02-03 | 株式会社島津製作所 | Radiographic apparatus |
CN102598059B (en) | 2009-08-06 | 2015-07-01 | 皇家飞利浦电子股份有限公司 | Method and apparatus for generating computed tomography images with offset detector geometries |
US10828786B2 (en) | 2009-08-17 | 2020-11-10 | Mazor Robotics Ltd. | Device for improving the accuracy of manual operations |
US9844414B2 (en) | 2009-08-31 | 2017-12-19 | Gregory S. Fischer | System and method for robotic surgical intervention in a magnetic resonance imager |
EP2298223A1 (en) | 2009-09-21 | 2011-03-23 | Stryker Leibinger GmbH & Co. KG | Technique for registering image data of an object |
US8465476B2 (en) | 2009-09-23 | 2013-06-18 | Intuitive Surgical Operations, Inc. | Cannula mounting fixture |
US9044269B2 (en) | 2009-09-30 | 2015-06-02 | Brainlab Ag | Two-part medical tracking marker |
NL1037348C2 (en) | 2009-10-02 | 2011-04-05 | Univ Eindhoven Tech | Surgical robot, instrument manipulator, combination of an operating table and a surgical robot, and master-slave operating system. |
US8556979B2 (en) | 2009-10-15 | 2013-10-15 | Globus Medical, Inc. | Expandable fusion device and method of installation thereof |
US8062375B2 (en) | 2009-10-15 | 2011-11-22 | Globus Medical, Inc. | Expandable fusion device and method of installation thereof |
US8679183B2 (en) | 2010-06-25 | 2014-03-25 | Globus Medical | Expandable fusion device and method of installation thereof |
US8685098B2 (en) | 2010-06-25 | 2014-04-01 | Globus Medical, Inc. | Expandable fusion device and method of installation thereof |
US8709086B2 (en) | 2009-10-15 | 2014-04-29 | Globus Medical, Inc. | Expandable fusion device and method of installation thereof |
EP2493411A4 (en) | 2009-10-28 | 2015-04-29 | Imris Inc | Automatic registration of images for image guided surgery |
US8411927B2 (en) * | 2009-10-29 | 2013-04-02 | Siemens Aktiengesellschaft | Marker detection in X-ray images |
USD631966S1 (en) | 2009-11-10 | 2011-02-01 | Globus Medical, Inc. | Basilar invagination implant |
US8521331B2 (en) | 2009-11-13 | 2013-08-27 | Intuitive Surgical Operations, Inc. | Patient-side surgeon interface for a minimally invasive, teleoperated surgical instrument |
US9082036B2 (en) * | 2009-11-25 | 2015-07-14 | Dental Imaging Technologies Corporation | Method for accurate sub-pixel localization of markers on X-ray images |
US9082177B2 (en) * | 2009-11-25 | 2015-07-14 | Dental Imaging Technologies Corporation | Method for tracking X-ray markers in serial CT projection images |
US8363919B2 (en) * | 2009-11-25 | 2013-01-29 | Imaging Sciences International Llc | Marker identification and processing in x-ray images |
US20110137152A1 (en) | 2009-12-03 | 2011-06-09 | General Electric Company | System and method for cooling components of a surgical navigation system |
US8277509B2 (en) | 2009-12-07 | 2012-10-02 | Globus Medical, Inc. | Transforaminal prosthetic spinal disc apparatus |
WO2011070519A1 (en) | 2009-12-10 | 2011-06-16 | Koninklijke Philips Electronics N.V. | Scanning system for differential phase contrast imaging |
US8694075B2 (en) | 2009-12-21 | 2014-04-08 | General Electric Company | Intra-operative registration for navigated surgical procedures |
US8353963B2 (en) | 2010-01-12 | 2013-01-15 | Globus Medical | Expandable spacer and method for use thereof |
US9381045B2 (en) | 2010-01-13 | 2016-07-05 | Jcbd, Llc | Sacroiliac joint implant and sacroiliac joint instrument for fusing a sacroiliac joint |
CN102711650B (en) | 2010-01-13 | 2015-04-01 | 皇家飞利浦电子股份有限公司 | Image integration based registration and navigation for endoscopic surgery |
US9030444B2 (en) | 2010-01-14 | 2015-05-12 | Brainlab Ag | Controlling and/or operating a medical device by means of a light pointer |
US9039769B2 (en) | 2010-03-17 | 2015-05-26 | Globus Medical, Inc. | Intervertebral nucleus and annulus implants and method of use thereof |
US20140330288A1 (en) | 2010-03-25 | 2014-11-06 | Precision Automation And Robotics India Ltd. | Articulating Arm for a Robotic Surgical Instrument System |
US20110238080A1 (en) | 2010-03-25 | 2011-09-29 | Date Ranjit | Robotic Surgical Instrument System |
IT1401669B1 (en) | 2010-04-07 | 2013-08-02 | Sofar Spa | ROBOTIC SURGERY SYSTEM WITH PERFECT CONTROL. |
US8870880B2 (en) | 2010-04-12 | 2014-10-28 | Globus Medical, Inc. | Angling inserter tool for expandable vertebral implant |
EP2558007A4 (en) * | 2010-04-13 | 2017-11-29 | Sentreheart, Inc. | Methods and devices for accessing and delivering devices to a heart |
US8717430B2 (en) | 2010-04-26 | 2014-05-06 | Medtronic Navigation, Inc. | System and method for radio-frequency imaging, registration, and localization |
IT1399603B1 (en) | 2010-04-26 | 2013-04-26 | Scuola Superiore Di Studi Universitari E Di Perfez | ROBOTIC SYSTEM FOR MINIMUM INVASIVE SURGERY INTERVENTIONS |
CA2797302C (en) | 2010-04-28 | 2019-01-15 | Ryerson University | System and methods for intraoperative guidance feedback |
EP2566392A4 (en) | 2010-05-04 | 2015-07-15 | Pathfinder Therapeutics Inc | System and method for abdominal surface matching using pseudo-features |
US8738115B2 (en) | 2010-05-11 | 2014-05-27 | Siemens Aktiengesellschaft | Method and apparatus for selective internal radiation therapy planning and implementation |
DE102010020284A1 (en) | 2010-05-12 | 2011-11-17 | Siemens Aktiengesellschaft | Determination of 3D positions and orientations of surgical objects from 2D X-ray images |
US8603077B2 (en) | 2010-05-14 | 2013-12-10 | Intuitive Surgical Operations, Inc. | Force transmission for robotic surgical instrument |
US8883210B1 (en) | 2010-05-14 | 2014-11-11 | Musculoskeletal Transplant Foundation | Tissue-derived tissuegenic implants, and methods of fabricating and using same |
KR101181569B1 (en) | 2010-05-25 | 2012-09-10 | 정창욱 | Surgical robot system capable of implementing both of single port surgery mode and multi-port surgery mode and method for controlling same |
US20110295370A1 (en) | 2010-06-01 | 2011-12-01 | Sean Suh | Spinal Implants and Methods of Use Thereof |
DE102010026674B4 (en) | 2010-07-09 | 2012-09-27 | Siemens Aktiengesellschaft | Imaging device and radiotherapy device |
US8675939B2 (en) | 2010-07-13 | 2014-03-18 | Stryker Leibinger Gmbh & Co. Kg | Registration of anatomical data sets |
WO2012007036A1 (en) | 2010-07-14 | 2012-01-19 | Brainlab Ag | Method and system for determining an imaging direction and calibration of an imaging apparatus |
DE102010031943A1 (en) * | 2010-07-22 | 2012-01-26 | Siemens Aktiengesellschaft | Method for marking a predetermined guide path of a medical instrument and orientation device |
US20120035507A1 (en) | 2010-07-22 | 2012-02-09 | Ivan George | Device and method for measuring anatomic geometries |
US8740882B2 (en) | 2010-07-30 | 2014-06-03 | Lg Electronics Inc. | Medical robotic system and method of controlling the same |
KR101911843B1 (en) * | 2010-08-02 | 2018-10-25 | 더 존스 홉킨스 유니버시티 | Tool exchange interface and control algorithm for cooperative surgical robots |
WO2012024686A2 (en) | 2010-08-20 | 2012-02-23 | Veran Medical Technologies, Inc. | Apparatus and method for four dimensional soft tissue navigation |
JP2012045278A (en) | 2010-08-30 | 2012-03-08 | Fujifilm Corp | X-ray imaging apparatus and x-ray imaging method |
WO2012030304A1 (en) | 2010-09-01 | 2012-03-08 | Agency For Science, Technology And Research | A robotic device for use in image-guided robot assisted surgical training |
KR20120030174A (en) | 2010-09-17 | 2012-03-28 | 삼성전자주식회사 | Surgery robot system and surgery apparatus and method for providing tactile feedback |
EP2431003B1 (en) | 2010-09-21 | 2018-03-21 | Medizinische Universität Innsbruck | Registration device, system, kit and method for a patient registration |
US8679125B2 (en) | 2010-09-22 | 2014-03-25 | Biomet Manufacturing, Llc | Robotic guided femoral head reshaping |
US8657809B2 (en) | 2010-09-29 | 2014-02-25 | Stryker Leibinger Gmbh & Co., Kg | Surgical navigation system |
US8627996B2 (en) | 2010-10-06 | 2014-01-14 | Sonics & Materials Inc. | System and method for terminating aluminum conductors |
US8718346B2 (en) | 2011-10-05 | 2014-05-06 | Saferay Spine Llc | Imaging system and method for use in surgical and interventional medical procedures |
US8526700B2 (en) | 2010-10-06 | 2013-09-03 | Robert E. Isaacs | Imaging system and method for surgical and interventional medical procedures |
US9913693B2 (en) | 2010-10-29 | 2018-03-13 | Medtronic, Inc. | Error correction techniques in surgical navigation |
US20130072784A1 (en) * | 2010-11-10 | 2013-03-21 | Gnanasekar Velusamy | Systems and methods for planning image-guided interventional procedures |
US8876866B2 (en) | 2010-12-13 | 2014-11-04 | Globus Medical, Inc. | Spinous process fusion devices and methods thereof |
WO2012082615A2 (en) | 2010-12-13 | 2012-06-21 | Ortho Kinematics, Inc. | Methods, systems and devices for clinical data reporting and surgical navigation |
WO2012088471A1 (en) * | 2010-12-22 | 2012-06-28 | Veebot, Llc | Systems and methods for autonomous intravenous needle insertion |
JP2014502541A (en) | 2010-12-22 | 2014-02-03 | ビューレイ・インコーポレイテッド | System and method for image guidance during medical procedures |
WO2012095755A1 (en) | 2011-01-13 | 2012-07-19 | Koninklijke Philips Electronics N.V. | Intraoperative camera calibration for endoscopic surgery |
KR101181613B1 (en) | 2011-02-21 | 2012-09-10 | 윤상진 | Surgical robot system for performing surgery based on displacement information determined by user designation and control method therefor |
US20120226145A1 (en) | 2011-03-03 | 2012-09-06 | National University Of Singapore | Transcutaneous robot-assisted ablation-device insertion navigation system |
US9026247B2 (en) | 2011-03-30 | 2015-05-05 | University of Washington through its Center for Communication | Motion and video capture for tracking and evaluating robotic surgery and associated systems and methods |
US9308050B2 (en) | 2011-04-01 | 2016-04-12 | Ecole Polytechnique Federale De Lausanne (Epfl) | Robotic system and method for spinal and other surgeries |
US8951195B2 (en) * | 2011-04-05 | 2015-02-10 | Houston Medical Robotics, Inc. | Motorized systems and methods for accessing the lumen of a vessel |
US20120256092A1 (en) | 2011-04-06 | 2012-10-11 | General Electric Company | Ct system for use in multi-modality imaging system |
WO2012139031A1 (en) | 2011-04-06 | 2012-10-11 | The Trustees Of Columbia University In The City Of New York | System, method and computer-accessible medium for providing a panoramic cone beam computed tomography (cbct) |
US10426554B2 (en) | 2011-04-29 | 2019-10-01 | The Johns Hopkins University | System and method for tracking and navigation |
US9072535B2 (en) | 2011-05-27 | 2015-07-07 | Ethicon Endo-Surgery, Inc. | Surgical stapling instruments with rotatable staple deployment arrangements |
JPWO2012169642A1 (en) | 2011-06-06 | 2015-02-23 | 希 松本 | Registration template manufacturing method |
US8498744B2 (en) | 2011-06-30 | 2013-07-30 | Mako Surgical Corporation | Surgical robotic systems with manual and haptic and/or active control modes |
JP6008960B2 (en) * | 2011-07-06 | 2016-10-19 | シー・アール・バード・インコーポレーテッドC R Bard Incorporated | Needle length determination and calibration for insertion guidance systems |
CA2841459C (en) | 2011-07-11 | 2020-07-28 | Board Of Regents Of The University Of Nebraska | Robotic surgical devices, systems, and related methods |
US8818105B2 (en) | 2011-07-14 | 2014-08-26 | Accuray Incorporated | Image registration for image-guided surgery |
KR20130015146A (en) | 2011-08-02 | 2013-02-13 | 삼성전자주식회사 | Method and apparatus for processing medical image, robotic surgery system using image guidance |
US10866783B2 (en) | 2011-08-21 | 2020-12-15 | Transenterix Europe S.A.R.L. | Vocally activated surgical control system |
EP2750620B1 (en) * | 2011-09-02 | 2017-04-26 | Stryker Corporation | Surgical instrument including a cutting accessory extending from a housing and actuators that establish the position of the cutting accessory relative to the housing |
US9427330B2 (en) | 2011-09-06 | 2016-08-30 | Globus Medical, Inc. | Spinal plate |
US8864833B2 (en) | 2011-09-30 | 2014-10-21 | Globus Medical, Inc. | Expandable fusion device and method of installation thereof |
US9060794B2 (en) | 2011-10-18 | 2015-06-23 | Mako Surgical Corp. | System and method for robotic surgery |
US8894688B2 (en) | 2011-10-27 | 2014-11-25 | Globus Medical Inc. | Adjustable rod devices and methods of using the same |
DE102011054910B4 (en) | 2011-10-28 | 2013-10-10 | Ovesco Endoscopy Ag | Magnetic end effector and means for guiding and positioning same |
CA3041707C (en) | 2011-11-15 | 2021-04-06 | Manickam UMASUTHAN | Method of real-time tracking of moving/flexible surfaces |
FR2983059B1 (en) | 2011-11-30 | 2014-11-28 | Medtech | ROBOTIC-ASSISTED METHOD OF POSITIONING A SURGICAL INSTRUMENT IN RELATION TO THE BODY OF A PATIENT AND DEVICE FOR CARRYING OUT SAID METHOD |
US9713499B2 (en) | 2011-12-05 | 2017-07-25 | Mazor Robotics Ltd. | Active bed mount for surgical robot |
KR101901580B1 (en) | 2011-12-23 | 2018-09-28 | 삼성전자주식회사 | Surgical robot and control method thereof |
FR2985167A1 (en) | 2011-12-30 | 2013-07-05 | Medtech | ROBOTISE MEDICAL METHOD FOR MONITORING PATIENT BREATHING AND CORRECTION OF ROBOTIC TRAJECTORY. |
US9259278B2 (en) | 2011-12-30 | 2016-02-16 | Mako Surgical Corp. | System for image-based robotic surgery |
US9265583B2 (en) | 2011-12-30 | 2016-02-23 | Mako Surgical Corp. | Method for image-based robotic surgery |
KR20130080909A (en) | 2012-01-06 | 2013-07-16 | 삼성전자주식회사 | Surgical robot and method for controlling the same |
US9138297B2 (en) | 2012-02-02 | 2015-09-22 | Intuitive Surgical Operations, Inc. | Systems and methods for controlling a robotic surgical system |
WO2013116140A1 (en) * | 2012-02-03 | 2013-08-08 | Intuitive Surgical Operations, Inc. | Steerable flexible needle with embedded shape sensing |
US10249036B2 (en) | 2012-02-22 | 2019-04-02 | Veran Medical Technologies, Inc. | Surgical catheter having side exiting medical instrument and related systems and methods for four dimensional soft tissue navigation |
US11207132B2 (en) | 2012-03-12 | 2021-12-28 | Nuvasive, Inc. | Systems and methods for performing spinal surgery |
US8855822B2 (en) | 2012-03-23 | 2014-10-07 | Innovative Surgical Solutions, Llc | Robotic surgical system with mechanomyography feedback |
DE102012006352B4 (en) * | 2012-03-28 | 2014-07-17 | Mbda Deutschland Gmbh | Device for testing and / or operating an active unit |
KR101946000B1 (en) | 2012-03-28 | 2019-02-08 | 삼성전자주식회사 | Robot system and Control Method thereof for surgery |
US8888821B2 (en) | 2012-04-05 | 2014-11-18 | Warsaw Orthopedic, Inc. | Spinal implant measuring system and method |
US10354377B2 (en) * | 2012-04-11 | 2019-07-16 | The Trustees Of Columbia University In The City Of New York | Techniques for segmentation of lymph nodes, lung lesions and other solid or part-solid objects |
WO2013158655A1 (en) | 2012-04-16 | 2013-10-24 | Neurologica Corp. | Imaging system with rigidly mounted fiducial markers |
JP2015521056A (en) | 2012-04-16 | 2015-07-27 | ニューロロジカ・コーポレーション | Wireless imaging system |
US10383765B2 (en) | 2012-04-24 | 2019-08-20 | Auris Health, Inc. | Apparatus and method for a global coordinate system for use in robotic surgery |
US20140142591A1 (en) | 2012-04-24 | 2014-05-22 | Auris Surgical Robotics, Inc. | Method, apparatus and a system for robotic assisted surgery |
WO2013166098A1 (en) | 2012-05-01 | 2013-11-07 | The Johns Hopkins University | Improved method and apparatus for robotically assisted cochlear implant surgery |
US20140234804A1 (en) | 2012-05-02 | 2014-08-21 | Eped Inc. | Assisted Guidance and Navigation Method in Intraoral Surgery |
US9125556B2 (en) | 2012-05-14 | 2015-09-08 | Mazor Robotics Ltd. | Robotic guided endoscope |
JP2015516278A (en) | 2012-05-18 | 2015-06-11 | ケアストリーム ヘルス インク | Volumetric imaging system for cone-beam computed tomography |
KR20130132109A (en) | 2012-05-25 | 2013-12-04 | 삼성전자주식회사 | Supporting device and surgical robot system adopting the same |
CN104334112B (en) | 2012-06-01 | 2017-10-27 | 直观外科手术操作公司 | Multiport surgical robot system framework |
CN107595392B (en) | 2012-06-01 | 2020-11-27 | 直观外科手术操作公司 | Avoidance of manipulator arm collisions with patient using null space |
US20150032164A1 (en) * | 2012-06-21 | 2015-01-29 | Globus Medical, Inc. | Methods for Performing Invasive Medical Procedures Using a Surgical Robot |
US11589771B2 (en) * | 2012-06-21 | 2023-02-28 | Globus Medical Inc. | Method for recording probe movement and determining an extent of matter removed |
US11395706B2 (en) * | 2012-06-21 | 2022-07-26 | Globus Medical Inc. | Surgical robot platform |
US10350013B2 (en) * | 2012-06-21 | 2019-07-16 | Globus Medical, Inc. | Surgical tool systems and methods |
JP2015528713A (en) * | 2012-06-21 | 2015-10-01 | グローバス メディカル インコーポレイティッド | Surgical robot platform |
US10231791B2 (en) * | 2012-06-21 | 2019-03-19 | Globus Medical, Inc. | Infrared signal based position recognition system for use with a robot-assisted surgery |
US11786324B2 (en) * | 2012-06-21 | 2023-10-17 | Globus Medical, Inc. | Surgical robotic automation with tracking markers |
US20170258535A1 (en) * | 2012-06-21 | 2017-09-14 | Globus Medical, Inc. | Surgical robotic automation with tracking markers |
US10646280B2 (en) * | 2012-06-21 | 2020-05-12 | Globus Medical, Inc. | System and method for surgical tool insertion using multiaxis force and moment feedback |
US11116576B2 (en) * | 2012-06-21 | 2021-09-14 | Globus Medical Inc. | Dynamic reference arrays and methods of use |
US10624710B2 (en) * | 2012-06-21 | 2020-04-21 | Globus Medical, Inc. | System and method for measuring depth of instrumentation |
US20170172669A1 (en) * | 2012-06-21 | 2017-06-22 | Globus Medical, Inc. | System and method for a surveillance marker in robotic-assisted surgery |
US11896446B2 (en) * | 2012-06-21 | 2024-02-13 | Globus Medical, Inc | Surgical robotic automation with tracking markers |
EP2864086B1 (en) | 2012-06-22 | 2017-04-19 | Board of Regents of the University of Nebraska | Local control robotic surgical devices |
US20130345757A1 (en) | 2012-06-22 | 2013-12-26 | Shawn D. Stad | Image Guided Intra-Operative Contouring Aid |
US20140001231A1 (en) * | 2012-06-28 | 2014-01-02 | Ethicon Endo-Surgery, Inc. | Firing system lockout arrangements for surgical instruments |
US9072536B2 (en) * | 2012-06-28 | 2015-07-07 | Ethicon Endo-Surgery, Inc. | Differential locking arrangements for rotary powered surgical instruments |
US8747238B2 (en) * | 2012-06-28 | 2014-06-10 | Ethicon Endo-Surgery, Inc. | Rotary drive shaft assemblies for surgical instruments with articulatable end effectors |
US9408606B2 (en) * | 2012-06-28 | 2016-08-09 | Ethicon Endo-Surgery, Llc | Robotically powered surgical device with manually-actuatable reversing system |
US8880223B2 (en) | 2012-07-16 | 2014-11-04 | Florida Institute for Human & Maching Cognition | Anthro-centric multisensory interface for sensory augmentation of telesurgery |
US20140031664A1 (en) | 2012-07-30 | 2014-01-30 | Mako Surgical Corp. | Radiographic imaging device |
KR101997566B1 (en) | 2012-08-07 | 2019-07-08 | 삼성전자주식회사 | Surgical robot system and control method thereof |
US9770305B2 (en) | 2012-08-08 | 2017-09-26 | Board Of Regents Of The University Of Nebraska | Robotic surgical devices, systems, and related methods |
EP2882331A4 (en) | 2012-08-08 | 2016-03-23 | Univ Nebraska | Robotic surgical devices, systems, and related methods |
US10110785B2 (en) | 2012-08-10 | 2018-10-23 | Karl Storz Imaging, Inc. | Deployable imaging system equipped with solid state imager |
CN104717935B (en) | 2012-08-15 | 2018-04-06 | 直观外科手术操作公司 | Cancel the system of joint motions using kernel |
RU2015110290A (en) | 2012-08-24 | 2016-10-20 | Юниверсити Оф Хьюстон | ROBOTIC DEVICE AND SYSTEMS FOR SURGERY USING ROBOT AND UNDER VISUALIZED CONTROL |
US20140080086A1 (en) | 2012-09-20 | 2014-03-20 | Roger Chen | Image Navigation Integrated Dental Implant System |
US8892259B2 (en) | 2012-09-26 | 2014-11-18 | Innovative Surgical Solutions, LLC. | Robotic surgical system with mechanomyography feedback |
US9757160B2 (en) | 2012-09-28 | 2017-09-12 | Globus Medical, Inc. | Device and method for treatment of spinal deformity |
US20140107473A1 (en) * | 2012-10-17 | 2014-04-17 | Cincinnati Children's Hospital Medical Center | Laser Guidance System for Interventions |
US9226579B2 (en) | 2012-11-02 | 2016-01-05 | Paragon Furniture LP | Modular desk systems and methods |
KR102038632B1 (en) | 2012-11-06 | 2019-10-30 | 삼성전자주식회사 | surgical instrument, supporting device, and surgical robot system adopting the same |
CN104780862A (en) | 2012-11-14 | 2015-07-15 | 直观外科手术操作公司 | Smart drapes for collision avoidance |
KR20140064329A (en) | 2012-11-20 | 2014-05-28 | 삼성전기주식회사 | Printed circuit board and manufacturing method thereof |
KR102079945B1 (en) | 2012-11-22 | 2020-02-21 | 삼성전자주식회사 | Surgical robot and method for controlling the surgical robot |
US9008752B2 (en) | 2012-12-14 | 2015-04-14 | Medtronic, Inc. | Method to determine distribution of a material by an infused magnetic resonance image contrast agent |
US9393361B2 (en) | 2012-12-14 | 2016-07-19 | Medtronic, Inc. | Method to determine a material distribution |
DE102012025101A1 (en) | 2012-12-20 | 2014-06-26 | avateramedical GmBH | Active positioning device of a surgical instrument and a surgical robotic system comprising it |
DE102013004459A1 (en) | 2012-12-20 | 2014-06-26 | avateramedical GmBH | Holding and positioning device of a surgical instrument and / or an endoscope for minimally invasive surgery and a robotic surgical system |
US9001962B2 (en) | 2012-12-20 | 2015-04-07 | Triple Ring Technologies, Inc. | Method and apparatus for multiple X-ray imaging applications |
US9002437B2 (en) | 2012-12-27 | 2015-04-07 | General Electric Company | Method and system for position orientation correction in navigation |
US10028788B2 (en) | 2012-12-31 | 2018-07-24 | Mako Surgical Corp. | System for image-based robotic surgery |
KR20140090374A (en) | 2013-01-08 | 2014-07-17 | 삼성전자주식회사 | Single port surgical robot and control method thereof |
CN103969269B (en) | 2013-01-31 | 2018-09-18 | Ge医疗系统环球技术有限公司 | Method and apparatus for geometric calibration CT scanner |
US20140221819A1 (en) | 2013-02-01 | 2014-08-07 | David SARMENT | Apparatus, system and method for surgical navigation |
EP2951743B1 (en) | 2013-02-04 | 2020-04-29 | Children's National Medical Center | Hybrid control surgical robotic system |
KR20140102465A (en) | 2013-02-14 | 2014-08-22 | 삼성전자주식회사 | Surgical robot and method for controlling the same |
KR102117270B1 (en) | 2013-03-06 | 2020-06-01 | 삼성전자주식회사 | Surgical robot system and method for controlling the same |
KR20140110685A (en) | 2013-03-08 | 2014-09-17 | 삼성전자주식회사 | Method for controlling of single port surgical robot |
KR20140110620A (en) | 2013-03-08 | 2014-09-17 | 삼성전자주식회사 | surgical robot system and operating method thereof |
KR102119534B1 (en) | 2013-03-13 | 2020-06-05 | 삼성전자주식회사 | Surgical robot and method for controlling the same |
KR20140112207A (en) | 2013-03-13 | 2014-09-23 | 삼성전자주식회사 | Augmented reality imaging display system and surgical robot system comprising the same |
US9314308B2 (en) | 2013-03-13 | 2016-04-19 | Ethicon Endo-Surgery, Llc | Robotic ultrasonic surgical device with articulating end effector |
CA2905948C (en) | 2013-03-14 | 2022-01-11 | Board Of Regents Of The University Of Nebraska | Methods, systems, and devices relating to robotic surgical devices, end effectors, and controllers |
US9629595B2 (en) | 2013-03-15 | 2017-04-25 | Hansen Medical, Inc. | Systems and methods for localizing, tracking and/or controlling medical instruments |
EP2996545B1 (en) | 2013-03-15 | 2021-10-20 | Board of Regents of the University of Nebraska | Robotic surgical systems |
EP2969409A4 (en) * | 2013-03-15 | 2016-11-16 | Intuitive Surgical Operations | Inter-operative switching of tools in a robotic surgical system |
US20140276880A1 (en) * | 2013-03-15 | 2014-09-18 | Blackstone Medical, Inc. | Locking assembly for a surgical drill bit guide |
KR102117273B1 (en) | 2013-03-21 | 2020-06-01 | 삼성전자주식회사 | Surgical robot system and method for controlling the same |
KR20140121581A (en) | 2013-04-08 | 2014-10-16 | 삼성전자주식회사 | Surgical robot system |
KR20140123122A (en) | 2013-04-10 | 2014-10-22 | 삼성전자주식회사 | Surgical Robot and controlling method of thereof |
US9414859B2 (en) | 2013-04-19 | 2016-08-16 | Warsaw Orthopedic, Inc. | Surgical rod measuring system and method |
US8964934B2 (en) | 2013-04-25 | 2015-02-24 | Moshe Ein-Gal | Cone beam CT scanning |
KR20140129702A (en) | 2013-04-30 | 2014-11-07 | 삼성전자주식회사 | Surgical robot system and method for controlling the same |
US20140364720A1 (en) | 2013-06-10 | 2014-12-11 | General Electric Company | Systems and methods for interactive magnetic resonance imaging |
KR20150000171A (en) | 2013-06-24 | 2015-01-02 | 삼성전자주식회사 | Movable medical apparatus and method for controlling of movement of the same |
DE102013012397B4 (en) | 2013-07-26 | 2018-05-24 | Rg Mechatronics Gmbh | Surgical robot system |
US10786283B2 (en) | 2013-08-01 | 2020-09-29 | Musc Foundation For Research Development | Skeletal bone fixation mechanism |
US20150085970A1 (en) | 2013-09-23 | 2015-03-26 | General Electric Company | Systems and methods for hybrid scanning |
CA2926621C (en) | 2013-10-07 | 2021-05-25 | Technion Research & Development Foundation Ltd. | Needle steering by shaft manipulation |
US9848922B2 (en) | 2013-10-09 | 2017-12-26 | Nuvasive, Inc. | Systems and methods for performing spine surgery |
CN110123448A (en) | 2013-10-09 | 2019-08-16 | 纽文思公司 | The method for being designed in art during vertebra program of performing the operation and evaluating spine malformation correction |
ITBO20130599A1 (en) | 2013-10-31 | 2015-05-01 | Cefla Coop | METHOD AND APPARATUS TO INCREASE THE FIELD OF VIEW IN A COMPUTERIZED TOMOGRAPHIC ACQUISITION WITH CONE-BEAM TECHNIQUE |
US20150146847A1 (en) | 2013-11-26 | 2015-05-28 | General Electric Company | Systems and methods for providing an x-ray imaging system with nearly continuous zooming capability |
EP3119314B1 (en) | 2014-03-17 | 2020-05-06 | Intuitive Surgical Operations, Inc. | System and method for breakaway clutching in an articulated arm |
WO2015195843A2 (en) | 2014-06-17 | 2015-12-23 | Nuvasive, Inc. | Systems and methods for planning, performing, and assessing spinal correction during surgery |
CN106687062B (en) | 2014-09-17 | 2019-11-22 | 直观外科手术操作公司 | For the system and method mobile using augmented Jacobian matrix control manipulator joint |
US10631907B2 (en) | 2014-12-04 | 2020-04-28 | Mazor Robotics Ltd. | Shaper for vertebral fixation rods |
US20160166329A1 (en) | 2014-12-15 | 2016-06-16 | General Electric Company | Tomographic imaging for interventional tool guidance |
EP3282997B1 (en) | 2015-04-15 | 2021-06-16 | Mobius Imaging, LLC | Integrated medical imaging and surgical robotic system |
US10180404B2 (en) | 2015-04-30 | 2019-01-15 | Shimadzu Corporation | X-ray analysis device |
US10058394B2 (en) * | 2015-07-31 | 2018-08-28 | Globus Medical, Inc. | Robot arm and methods of use |
US10080615B2 (en) * | 2015-08-12 | 2018-09-25 | Globus Medical, Inc. | Devices and methods for temporary mounting of parts to bone |
US10034716B2 (en) * | 2015-09-14 | 2018-07-31 | Globus Medical, Inc. | Surgical robotic systems and methods thereof |
US20170143284A1 (en) | 2015-11-25 | 2017-05-25 | Carestream Health, Inc. | Method to detect a retained surgical object |
US10070939B2 (en) | 2015-12-04 | 2018-09-11 | Zaki G. Ibrahim | Methods for performing minimally invasive transforaminal lumbar interbody fusion using guidance |
KR20180097633A (en) | 2016-01-20 | 2018-08-31 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | System and method for rapid shutdown and recovery of motion departure from medical device repositionable arms |
WO2017127838A1 (en) | 2016-01-22 | 2017-07-27 | Nuvasive, Inc. | Systems and methods for facilitating spine surgery |
US11058378B2 (en) | 2016-02-03 | 2021-07-13 | Globus Medical, Inc. | Portable medical imaging system |
US10448910B2 (en) | 2016-02-03 | 2019-10-22 | Globus Medical, Inc. | Portable medical imaging system |
US10842453B2 (en) | 2016-02-03 | 2020-11-24 | Globus Medical, Inc. | Portable medical imaging system |
US9962133B2 (en) | 2016-03-09 | 2018-05-08 | Medtronic Navigation, Inc. | Transformable imaging system |
US9931025B1 (en) | 2016-09-30 | 2018-04-03 | Auris Surgical Robotics, Inc. | Automated calibration of endoscopes with pull wires |
-
2013
- 2013-06-21 JP JP2015518631A patent/JP2015528713A/en active Pending
- 2013-06-21 EP EP13807863.9A patent/EP2863827B1/en active Active
- 2013-06-21 US US13/924,505 patent/US9782229B2/en active Active
- 2013-06-21 WO PCT/US2013/047194 patent/WO2013192598A1/en unknown
-
2014
- 2014-09-03 US US14/476,101 patent/US10531927B2/en active Active
-
2017
- 2017-03-03 US US15/449,260 patent/US11135022B2/en active Active
- 2017-03-03 US US15/448,830 patent/US20170239002A1/en not_active Abandoned
- 2017-03-09 US US15/454,379 patent/US10912617B2/en active Active
- 2017-03-10 US US15/455,662 patent/US11103317B2/en active Active
- 2017-03-16 US US15/460,974 patent/US11284949B2/en active Active
- 2017-03-17 US US15/462,280 patent/US10485617B2/en active Active
- 2017-03-23 US US15/466,937 patent/US11331153B2/en active Active
- 2017-03-23 US US15/467,005 patent/US11684431B2/en active Active
- 2017-03-30 US US15/473,698 patent/US11026756B2/en active Active
- 2017-05-31 US US15/609,322 patent/US20170265949A1/en active Pending
- 2017-05-31 US US15/609,305 patent/US11191598B2/en active Active
- 2017-07-13 US US15/648,621 patent/US10835326B2/en active Active
- 2017-09-07 US US15/697,943 patent/US20170360517A1/en not_active Abandoned
- 2017-09-14 US US15/704,636 patent/US10835328B2/en active Active
-
2019
- 2019-02-05 US US16/267,833 patent/US20190167362A1/en active Pending
- 2019-10-08 US US16/595,578 patent/US20200155243A1/en active Pending
- 2019-12-10 US US16/708,545 patent/US11690687B2/en active Active
-
2020
- 2020-10-15 US US17/071,307 patent/US20210022814A1/en active Pending
-
2021
- 2021-11-05 US US17/520,196 patent/US20230021298A9/en active Pending
-
2022
- 2022-02-28 US US17/652,723 patent/US20220409306A1/en active Pending
- 2022-04-19 US US17/724,092 patent/US20220233262A1/en active Pending
- 2022-04-22 US US17/727,013 patent/US20220241037A1/en active Pending
Also Published As
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20220054206A1 (en) | Surgical robot platform | |
US11896363B2 (en) | Surgical robot platform | |
CN111417354B (en) | Method and system for robotic assisted surgery | |
US20220346894A1 (en) | Surgical robot platform | |
ES2963154T3 (en) | Medical robot with automatic positioning means | |
EP3824839A1 (en) | Robotic positioning of a device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: GLOBUS MEDICAL, INC., PENNSYLVANIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CRAWFORD, NEIL R.;THEODORE, NICHOLAS;FOSTER, MITCHELL;REEL/FRAME:058034/0387 Effective date: 20131223 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |