US20210401503A1 - Anatomical model generation - Google Patents

Anatomical model generation Download PDF

Info

Publication number
US20210401503A1
US20210401503A1 US17/195,482 US202117195482A US2021401503A1 US 20210401503 A1 US20210401503 A1 US 20210401503A1 US 202117195482 A US202117195482 A US 202117195482A US 2021401503 A1 US2021401503 A1 US 2021401503A1
Authority
US
United States
Prior art keywords
anchor portions
dimensional
surface representation
dimensional surface
anatomic structure
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US17/195,482
Inventor
Doron Harlev
Geoffrey Peter Wright
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Affera Inc
Original Assignee
Affera Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Affera Inc filed Critical Affera Inc
Priority to US17/195,482 priority Critical patent/US20210401503A1/en
Publication of US20210401503A1 publication Critical patent/US20210401503A1/en
Priority to US17/658,252 priority patent/US20230013302A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/742Details of notification to user or communication with user or patient ; user input means using visual displays
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B18/00Surgical instruments, devices or methods for transferring non-mechanical forms of energy to or from the body
    • A61B18/04Surgical instruments, devices or methods for transferring non-mechanical forms of energy to or from the body by heating
    • A61B18/12Surgical instruments, devices or methods for transferring non-mechanical forms of energy to or from the body by heating by passing a current through the tissue to be heated, e.g. high-frequency current
    • A61B18/14Probes or electrodes therefor
    • A61B18/1492Probes or electrodes therefor having a flexible, catheter-like structure, e.g. for heart ablation
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/25User interfaces for surgical systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/50ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for simulation or modelling of medical disorders
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B18/00Surgical instruments, devices or methods for transferring non-mechanical forms of energy to or from the body
    • A61B2018/00636Sensing and controlling the application of energy
    • A61B2018/00642Sensing and controlling the application of energy with feedback, i.e. closed loop control
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B18/00Surgical instruments, devices or methods for transferring non-mechanical forms of energy to or from the body
    • A61B2018/00636Sensing and controlling the application of energy
    • A61B2018/00773Sensed parameters
    • A61B2018/00839Bioelectrical parameters, e.g. ECG, EEG
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B18/00Surgical instruments, devices or methods for transferring non-mechanical forms of energy to or from the body
    • A61B2018/00636Sensing and controlling the application of energy
    • A61B2018/00773Sensed parameters
    • A61B2018/00875Resistance or impedance
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/101Computer-aided simulation of surgical operations
    • A61B2034/102Modelling of surgical devices, implants or prosthesis
    • A61B2034/104Modelling the effect of the tool, e.g. the effect of an implanted prosthesis or for predicting the effect of ablation or burring
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/101Computer-aided simulation of surgical operations
    • A61B2034/105Modelling of the patient, e.g. for ligaments or bones
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • A61B2034/2046Tracking techniques
    • A61B2034/2051Electromagnetic tracking systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • A61B2034/2046Tracking techniques
    • A61B2034/2063Acoustic tracking systems, e.g. using ultrasound
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/06Measuring instruments not otherwise provided for
    • A61B2090/064Measuring instruments not otherwise provided for for measuring force, pressure or mechanical tension
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/06Devices, other than using radiation, for detecting or locating foreign bodies ; determining position of probes within or on the body of the patient
    • A61B5/061Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body
    • A61B5/062Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body using magnetic field
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/06Devices, other than using radiation, for detecting or locating foreign bodies ; determining position of probes within or on the body of the patient
    • A61B5/061Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body
    • A61B5/063Determining position of a probe within the body employing means separate from the probe, e.g. sensing internal probe position employing impedance electrodes on the surface of the body using impedance measurements
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • A61B5/25Bioelectric electrodes therefor
    • A61B5/279Bioelectric electrodes therefor specially adapted for particular uses
    • A61B5/28Bioelectric electrodes therefor specially adapted for particular uses for electrocardiography [ECG]
    • A61B5/283Invasive
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/68Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
    • A61B5/6801Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
    • A61B5/6843Monitoring or controlling sensor contact pressure
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/12Diagnosis using ultrasonic, sonic or infrasonic waves in body cavities or body tracts, e.g. by using catheters
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/24Indexing scheme for image data processing or generation, in general involving graphical user interfaces [GUIs]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/41Medical
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2219/00Indexing scheme for manipulating 3D models or images for computer graphics
    • G06T2219/004Annotating, labelling

Definitions

  • Three-dimensional models can be used to assist in the placement or use of a device when such placement or use is not easily observable or practical.
  • three-dimensional models are used to assist in the placement and use of medical devices for diagnosis or treatment of patients.
  • An example of such a medical procedure carried out with the assistance of a three-dimensional model is the use of a catheter to deliver radio frequency (“RF”) ablation to form lesions that interrupt abnormal conduction in cardiac tissue, thus terminating certain arrhythmias in the heart.
  • RF radio frequency
  • the present disclosure is directed to devices, systems, and methods of generating an accurate three-dimensional model of an anatomic structure of a patient to facilitate, for example, moving a medical device through the anatomic structure during a medical procedure in which the three-dimensional model is used to visualize the medical device in the anatomic structure.
  • the systems and methods of the present disclosure can be used to generate a three-dimensional model based on input (e.g., from a physician) of anchor portions corresponding to the position of anatomic features of the anatomic structure.
  • the systems and methods of the present disclosure can be used to generate a three-dimensional surface representation of the anatomic structure, with the three-dimensional surface representation constrained relative to one or more anchor portions identified on a three-dimensional data structure.
  • the constraint imposed by the one or more anchor portions can change the shape of the three-dimensional surface representation and, thus, can have the appearance of changing the position of tissue in a visual representation of the three-dimensional surface representation
  • the constraint imposed by the anchor portions is sometimes referred to herein as “pinch.”
  • the devices, systems, and methods of the present disclosure can provide a physician with a greater amount of control over a three-dimensional model of an anatomic structure and, additionally or alternatively, can represent salient features of the anatomic structure in greater detail than is typically achievable in a three-dimensional model built based solely on an incomplete or uncertain data set of known locations of a medical device in an anatomic structure.
  • a method includes receiving a plurality of location signals, each received location signal indicative of a respective location of a medical device in an anatomic structure of a patient, forming a three-dimensional data structure representing locations, within the anatomic structure, visited by the medical device at the locations corresponding to the plurality of location signals, receiving one or more anchor portions representing locations relative to the anatomic structure, and generating a three-dimensional surface representation of the anatomic structure of the patient, the three-dimensional surface representation of the anatomic structure of the patient constrained relative to the one or more anchor portions and to contain at least a portion of the three-dimensional data structure.
  • the method can further include displaying, on a graphical user interface, at least one of a two-dimensional projection of the three-dimensional data structure, the one or more anchor portions, and a two-dimensional projection of the three-dimensional surface representation.
  • receiving the one or more anchor portions representing locations relative to the anatomic structure can include receiving, from one or more sensors disposed on the medical device, a signal indicative of contact between the medical device and tissue of the anatomic structure.
  • the signal indicative of contact can be, for example, indicative of a blood-tissue boundary of the anatomic structure of the patient.
  • the signal indicative of contact can include one or more of: a change in impedance detected by one or more electrodes of the medical device, a force detected by a force sensor of the medical device, an ultrasound signal of an ultrasound sensor of the medical device, a deformation of at least a portion of the medical device, and an amplitude derived from an electrogram detected by one or more electrodes of the medical device.
  • receiving the one or more anchor portions can include receiving an input command from a user.
  • receiving the one or more anchor portions can include identifying a subset of the three-dimensional data structure.
  • receiving the one or more anchor portions can include receiving a respective confidence level associated each of the one or more anchor portions, and constraining the three-dimensional surface representation relative to the one or more anchor portions is based on the respective confidence level associated with each of the one or more anchor portions.
  • the method can further include representing, on a graphical user interface, the one or more anchor portions as annotations on the three-dimensional surface representation of the anatomic structure.
  • the method can further include representing, on a graphical user interface, the one or more anchor portions as annotations on the three-dimensional data structure.
  • the method can further include determining whether the one or more anchor portions have been modified and, based on whether the one or more anchor portions have been modified, repeating the generating step. Determining whether the one or more anchor portions have been modified can include, for example, determining whether one or more of previously identified anchor portions have been removed.
  • the three-dimensional surface representation of the anatomic structure can be a continuous mesh.
  • a method includes forming a three-dimensional data structure based on received locations of a tip section of a cardiac catheter in a heart cavity of a patient, receiving one or more anchor portions representing locations relative to the heart cavity, and generating a three-dimensional surface representation of the heart cavity of the patient, the surface representation of the heart cavity of the patient constrained relative to the anchor portions and to contain at least a portion of the three-dimensional data structure.
  • the method can further include displaying, on a graphical user interface, at least one of a two-dimensional projection of the three-dimensional data structure, the one or more anchor portions, and a two-dimensional projection of the generated three-dimensional surface representation.
  • receiving the one or more anchor portions on the three-dimensional data structure can include receiving one or more location signals indicative of one or more respective locations of the cardiac catheter in the heart cavity.
  • receiving the one or more anchor portions can include receiving, from a sensor disposed on the cardiac catheter, a signal indicative of a blood-tissue boundary of the heart cavity of the patient.
  • the signal corresponding to the blood-tissue boundary can include one or more of: a change in impedance detected by one or more electrodes of the cardiac catheter, a force detected by a force sensor of the cardiac catheter, an ultrasound signal of an ultrasound sensor of the cardiac catheter, and a deformation of at least a portion of the cardiac catheter, and an amplitude derived from an electrogram detected by one or more electrodes of the cardiac catheter.
  • receiving the one or more anchor portions on the three-dimensional data structure can include receiving an input command from a user interface.
  • the method can further include representing, on a graphical user interface, the one or more anchor portions as annotations on the three-dimensional surface representation of the heart cavity.
  • a non-transitory, computer-readable storage medium has stored thereon computer executable instructions for causing one or more processors to: receive a plurality of location signals, each received location signal indicative of a respective location of a medical device in an anatomic structure of a patient; form a three-dimensional data structure representing volumes, within the anatomic structure, occupied by the medical device at the locations corresponding to the plurality of location signals; receive one or more anchor portions representing locations relative to the anatomic structure; and generate a three-dimensional surface representation of the anatomic structure of the patient, the three-dimensional surface representation of the anatomic structure of the patient constrained relative to the one or more anchor portions and containing at least a portion of the three-dimensional data structure.
  • Implementations can include one or more of the following advantages.
  • a three-dimensional surface representation of an anatomic structure can be based on one or more anchor portions.
  • the three-dimensional surface representation of the anatomic structure can be constrained to pass near the one or more anchor portions and/or to pass near a fixed position relative to the one or more anchor portions.
  • the three-dimensional surface representation can accurately represent an anatomic structure such as, for example, an anatomic structure with local concavities along a generally convex shape.
  • Such accurate representation can be particularly advantageous for proper manipulation of a catheter for the accurate placement of lesions in the heart, such as placement of lesions in the carina between pulmonary veins.
  • constraining the three-dimensional representation relative to one or more anchor portions can facilitate generating an accurate representation of the anatomic structure based on relatively few data points (e.g., an incomplete or uncertain data set) while still providing a useful representation of salient anatomic features of the anatomic structure.
  • constraining the three-dimensional surface representation relative to one or more anchor portions can facilitate building a useful depiction of the anatomic structure in less time than would ordinarily be required to achieve the same level of detail in a model built based on catheter position alone.
  • constraining the three-dimensional surface representation relative to one or more anchor portions can facilitate shaping the three-dimensional surface representation independently of other parameters used to form the three-dimensional surface representation (e.g., tightness of a surface mesh).
  • one or more anchor portions can be based on feedback from one or more sensors on a medical device within an anatomic structure.
  • the one or more anchor portions can be based on input from the physician.
  • the one or more anchor portions can be based on a useful combination of physician input and feedback from one or more sensors of a medical device positioned within the anatomic structure. Such a combination can be useful for providing insights into the shape of the anatomic structure, while providing the physician with the ability to verify and, if necessary, override feedback from the one or more sensors.
  • FIG. 1 is a schematic representation of a system during a medical procedure.
  • FIG. 2 is a perspective view of an exemplary medical device of the system of FIG. 1 .
  • FIG. 3 is a schematic representation of a tip section of the medical device of FIG. 2 shown in an anatomic structure.
  • FIG. 4 is a schematic depiction of a projection of a three-dimensional data structure and a three-dimensional surface representation of the anatomic structure projected to a graphical user interface of the system of FIG. 1 .
  • FIG. 5A is a schematic representation of a surface of the anatomic structure having superimposed thereon a point cloud, corresponding to known locations of a medical device in an anatomic structure, and a volumetrically smoothed three-dimensional surface representation with a high degree of volumetric smoothing.
  • FIG. 5B is a schematic representation of a surface of the anatomic structure having superimposed thereon the point cloud of FIG. 5A and a three-dimensional surface representation with a low degree of volumetric smoothing.
  • FIG. 5C is a schematic representation of a surface of the anatomic structure having superimposed thereon the point cloud of FIG. 5A and a three-dimensional surface representation constrained relative to one or more anchor portions.
  • FIG. 6A is a schematic depiction of a three-dimensional data structure of FIG. 4 and an unconstrained three-dimensional surface representation displayed on the graphical user interface of the system of FIG. 1 .
  • FIG. 6B is a schematic depiction of the three-dimensional data structure and the unconstrained three-dimensional surface representation of FIG. 6A displayed on the graphical user interface of the system of FIG. 1 with an anchor portion identified on the three-dimensional data structure.
  • FIG. 6C is a schematic depiction of the three-dimensional data structure of FIG. 6A and a three-dimensional surface representation constrained relative to the anchor portion of FIG. 6B , the three-dimensional data structure and the three-dimensional surface representation displayed on the graphical user interface of the system of FIG. 1 , with the three-dimensional surface representation surface smoothed.
  • FIG. 6D is a schematic depiction of the three-dimensional surface representation of FIG. 6C displayed on the graphical user interface of the system of FIG. 1 .
  • FIG. 7 is a flowchart of an exemplary method of representing a surface of an anatomic structure.
  • FIG. 8 is a flowchart of an exemplary method of representing a surface of a heart cavity of a patient.
  • the present disclosure is generally directed to devices, systems, and methods of generating a three-dimensional surface representation of an anatomic structure of a patient. More specifically, the three-dimensional surface representation can accurately represent local anatomic features of the anatomic structure, while being based on an incomplete or uncertain data set, by constraining (e.g., pinching) the three-dimensional surface representation relative to the one or more anchor portions. For at least this reason, the three-dimensional surface representations generated according to the devices, systems and methods of the present disclosure can be generated efficiently and, in use, can be useful for facilitating visualization of a position of a medical device (e.g., a catheter) during a medical procedure (e.g., diagnosis and/or treatment) being performed on the anatomic structure.
  • a medical device e.g., a catheter
  • a medical procedure e.g., diagnosis and/or treatment
  • the systems and methods of the present disclosure can be used for any of various different medical procedures, such as procedures performed on a hollow anatomic structure of a patient, and, more specifically, in a hollow anatomic structure, in which direct visual access to the medical procedure is impractical and/or is improved by the use of a model of the anatomic structure.
  • the systems and methods of the present disclosure can be used to facilitate visualization of a catheter inserted into a heart cavity as part of a medical treatment associated with diagnosis, treatment, or both of a cardiac condition (e.g., cardiac arrhythmia).
  • a cardiac condition e.g., cardiac arrhythmia
  • the systems and methods of the present disclosure can be used in one or more medical procedures associated with interventional pulmonology, brain surgery, or sinus surgery (e.g., sinuplasty).
  • the term “physician” shall be understood to include any type of medical personnel who may be performing or assisting a medical procedure and, thus, is inclusive of a doctor, a nurse, a medical technician, other similar personnel, and any combination thereof.
  • the term “medical procedure” shall be understood to include any manner and form of diagnosis, treatment, or both, inclusive of any preparation activities associated with such diagnosis, treatment, or both.
  • the term “medical procedure” shall be understood to be inclusive of any manner and form of movement or positioning of a medical device in an anatomic chamber.
  • the term “patient” should be considered to include any mammal, including a human, upon which a medical procedure is being performed.
  • FIG. 1 is a schematic representation of a system 100 during a medical procedure performed in an anatomic structure of a patient 102 .
  • the system 100 can include a medical device 104 connected, via an extension cable 106 , to an interface unit 108 .
  • the interface unit 108 can include a processing unit 109 (e.g., one or more processors), a graphical user interface 110 , and a storage medium 111 .
  • the graphical user interface 110 and the storage medium 111 can be in electrical communication (e.g., wired communication, wireless communication, or both) with the processing unit 109 .
  • the medical device 104 can be moved within the anatomic structure (e.g., as part of a medical procedure) such that the processing unit 109 can receive a plurality of location signals of the medical device 104 in the anatomic structure.
  • the processing unit 109 can construct a three-dimensional surface representation of the anatomic structure based on a three-dimensional data structure representing locations, within the anatomic structure, visited by the medical device 104 .
  • a corresponding three-dimensional data structure can be an incomplete or uncertain data set.
  • it can be useful to volumetrically smooth the three-dimensional surface representation generated based on the three-dimensional data structure.
  • the processing unit 109 can receive one or more inputs corresponding to one or more anchor portions for advantageously constraining a three-dimensional surface representation of the anatomic structure.
  • the three-dimensional surface representation can be constrained to include details of the anatomic structure that would not otherwise be represented in a three-dimensional surface representation based on catheter location alone.
  • the three-dimensional surface representation of the anatomic structure formed using system 100 according to any one or more of the methods described herein can be shown on the graphical user interface 110 , and the three-dimensional surface representation can be used to facilitate performance of a medical procedure by a physician.
  • the three-dimensional surface representation of the anatomic structure and the position of the medical device 110 can be shown on the graphical user interface 110 and used as a visual guidance tool (e.g., as an analog) for movement of the medical device 104 in the anatomic structure. It should be appreciated, therefore, that the details provided in the three-dimensional surface representation generated as described herein can facilitate fine movement of the medical device 104 relative to the anatomic structure.
  • the three-dimensional surface representation generated according to any one or more of the methods described herein can more accurately represent anatomic features or landmarks that are useful for positioning the medical device 104 relative to targeted tissue.
  • the three-dimensional surface representation generated according to any one or more of the methods described herein is less likely to be unintentionally distorted in areas in which there are significant spatial gaps in position data of the medical device 104 .
  • the medical device 104 can be any of various different medical devices known in the art for use with respect to an anatomic structure and includes, therefore, any manner and form of medical devices useful for diagnosis, treatment, and combinations thereof.
  • the medical device 104 is described herein as a catheter insertable into an anatomic structure.
  • the medical device 104 can include a handle 120 , a shaft 122 , and a tip section 124 .
  • the shaft 122 can include a proximal portion 126 secured to the handle 120 , and a distal portion 128 coupled to the tip section 124 .
  • the tip section 124 generally includes any portion of the medical device 104 that directly or indirectly engages tissue for the purpose of treatment, diagnosis, or both and, therefore, can include any one or more of all manner and type of contact and/or non-contact interaction with tissue known in the art.
  • the tip section 124 can include one or more of contact and non-contact interaction with tissue in the form of energy interaction (e.g., electrical energy, ultrasound energy, light energy, cooling and any combinations thereof), chemical interaction with tissue, or both.
  • the tip section 124 can deliver energy (e.g., electrical energy) to tissue in the anatomic structure as part of any number of medical procedures.
  • energy e.g., RF energy
  • such ablations created by the tip section 124 along a surface of the anatomic structure can, for example, treat cardiac arrhythmia in patients with this condition.
  • the effectiveness of the ablations created using the tip section 124 in such a cardiac ablation procedure can be dependent upon the location of the ablations.
  • the medical device 104 can include a sensor 125 disposed, for example, along the tip section 124 and in electrical communication with the interface unit 108 (e.g., in communication with the processing unit 109 ).
  • the sensor 125 can be any of various different types of sensors suitable for sensing contact with tissue of an anatomic structure and, therefore, can be useful for providing feedback to the interface unit 108 regarding the location of a blood-tissue boundary.
  • any one or more forms of feedback provided by the sensor 125 can form the basis for generating anchor portions for constraining a three-dimensional surface representation of the anatomic structure. Examples of these forms of feedback provided by the sensor 125 and useful as the basis for generating anchor portions are described below. While the sensor 125 is described herein as a single sensor for the sake of clarity of explanation, the sensor 125 can include an array of any one or more of the sensors described herein, including, for example, any combination of the sensors described herein.
  • the sensor 125 can include a sensing electrode such that changes to an electrical signal measured between the sensor 125 and another electrode (e.g., another sensor carried on the tip section 124 ) can be detected as an indication of the presence of viable tissue in contact with the sensor 125 .
  • viable tissue is tissue that conducts an electrical signal and, thus, includes tissue that has not yet been ablated (e.g., is not scar tissue) as well as tissue that is not otherwise diseased such that conduction of the electrical signal is impaired.
  • the detection of viable tissue in contact with the sensor 125 can include observation of the electrical signal by the physician.
  • the detection of viable tissue in contact with the sensor can be based on a comparison of the electrical signal relative to a predetermined threshold (e.g., for a bipolar electrogram, a threshold of above about 0.1 mV). More generally, any of the various different devices, systems, and methods described herein can be advantageously used in combination with detecting changes to an electrical signal measured between the sensor 125 and another electrode to detect or confirm contact with tissue.
  • a predetermined threshold e.g., for a bipolar electrogram, a threshold of above about 0.1 mV.
  • the sensor 125 can include a force sensor to detect a magnitude and, optionally or additionally, a direction of force exerted on the sensor 125 through contact with a surface of the anatomic structure.
  • a force sensor can include any of various different force sensors responsive to contact between the sensor 125 and tissue of the anatomic structure. Such responsiveness can be independent, or at least substantially independent, of whether or not the contacted tissue is viable.
  • the sensor 125 can be a force sensor including optical fibers, transmitting or sensing coils, and the like, for sensing force. Contact between the sensor 125 and tissue of the anatomic structure can result in an increase in measured force. For example, a contact force greater than 5 g (e.g.
  • the measured force can, in addition or in the alternative, be related to the degree of contact between the sensor 125 and the tissue of the anatomic structure. Additionally, or alternatively, contact between the sensor 125 and tissue of the anatomic structure can result in a measured force in a direction normal to the tissue at a point of contact.
  • the senor 125 can include an ultrasound sensor such that the sensor 125 can detect features of an anatomic structure based on any of various different ultrasound techniques that are known in the art.
  • the sensor 125 can include an ultrasound transducer such that ultrasound reflections can be measured with respect to distance along an axis of the sensor 125 .
  • contact or proximity between the sensor 125 and tissue in the anatomic structure can result in ultrasound reflections at distances corresponding a distance between the sensor 125 and the tissue.
  • the senor 125 can include a deformation sensor to detect deformation (e.g., magnitude, direction, or both) of the tip section 124 as a result of contact between the tip section 124 and a surface of the anatomic structure.
  • the measured deformation can be a substantially monotonic function of the degree of contact between the sensor 125 and the tissue of the anatomic structure.
  • contact between the sensor 125 and tissue of the anatomic structure can result in deformation primarily in a direction normal to the tissue at the point of contact.
  • the senor 125 can include an impedance sensor to detect a change in an electrical impedance as a result of contact between the tip section 124 and tissue of the anatomic structure.
  • contact between the sensor 125 and tissue in the anatomic structure can be detected as an increase in a measured impedance.
  • an increase in a measured impedance larger than the expected variation in the impedance when the sensor 125 is not in contact with tissue e.g. an increase greater than 100 ohms
  • the measured impedance can be a substantially monotonic function of the degree of contact between the sensor 125 and the tissue.
  • contact between the tip section 124 and tissue of the anatomic structure can be based on one or more imaging modalities.
  • the use of one such imaging modality can include observation of one or both of the tip section 124 and the shaft 122 by the physician using fluoroscopy.
  • An additional, or alternative, modality can include observation of one or both of the tip section 124 and the shaft by the physician using intracardiac ultrasound in implementations in which the anatomic structure is a heart cavity.
  • the physician can tag the location of contact with tissue, and the tag can form a basis of the one or more anchor portions near which the three-dimensional surface representation is constrained to pass.
  • information determined automatically from an image can provide an indication of contact between the tip section 124 and tissue of the anatomic structure.
  • anchor portions can be additionally, or alternatively, based on other types of feedback.
  • anchor portions can be placed (e.g., through tags applied by the physician) in locations in which a physician detects a resistance to movement (e.g., rotation, articulation, advancement, or a combination thereof), with resistance being indicative of contact between the medical device 104 and the tissue.
  • the medical device 104 can further, or instead, include a magnetic position sensor 130 along the distal portion 128 of the shaft 122 .
  • the magnetic position sensor 130 can be any of various magnetic position sensors well known in the art and can be positioned at any point along the distal portion 128 .
  • the magnetic position sensor 130 can, for example, include one or more coils that detect signals emanating from magnetic field generators. One or more coils for determining position with five or six degrees of freedom can be used.
  • the magnetic field detected by the magnetic position sensor 130 can be used to determine the position of the distal portion 128 of the catheter shaft 122 according to one or more methods commonly known in the art such as, for example, methods based on using a sensor, such as the magnetic position sensor 130 , to sense magnetic fields indicative of the position of the magnetic position sensor 130 and using a look-up table to determine location of the magnetic position sensor 130 . Accordingly, because the tip section 124 is coupled to the distal portion 128 of the shaft 122 in a known, fixed relationship to the magnetic position sensor 130 , the magnetic position sensor 130 also provides the location of the tip section 124 . While the location of the tip section 124 is described as being determined based on magnetic position sensing, other position sensing methods can additionally or alternatively be used. For example, the location of the tip section 124 can be additionally, or alternatively, based on impedance, ultrasound, and/or imaging (e.g., real time MRI or fluoroscopy).
  • impedance, ultrasound, and/or imaging e.g.,
  • the tip section 124 of the medical device 104 can be moved in an anatomic structure 132 (e.g., prior to application of an ablation treatment or other type of treatment). If the tip section 124 of the medical device 104 is movable in blood in the anatomic structure 132 and obstructed only by a surface 133 of the anatomic cavity 132 , the known positions of the tip section 124 of the medical device 104 can be taken together to provide an indication of the size and shape of a volume defined by the surface 133 of the anatomic structure 132 and can form a basis for a three-dimensional data structure 134 corresponding to a volume defined by the anatomic structure 132 .
  • the three-dimensional data structure 134 can include any of various different data structures known in the art. As an example, the three-dimensional data structure 134 can include an occupancy grid. As another example, the three-dimensional data structure 134 can include an indicator function. Additionally, or alternatively, the three-dimensional data structure 134 can include a segmentation mask. Further, or instead, the three-dimensional data structure 134 can include space carving.
  • the three-dimensional data structure 134 may be an incomplete or uncertain data set. As described in greater detail below, forming the three-dimensional data structure 134 based on such an incomplete or uncertain data set can have implications for accurately representing anatomic features of the surface 133 of the anatomic structure 132 .
  • the three-dimensional data structure 134 can include a probabilistic model as a function of location, and an incomplete or uncertain data set of locations can result in regions of high uncertainty in the probabilistic model.
  • the three-dimensional data structure 134 can include a model with two or more discrete states (e.g., a “blood” state, an “unknown” state, a “tissue” state, a “boundary” state, and combinations thereof) as a function of location, and an incomplete or uncertain data set of locations can result in uncertain states (e.g. “unknown” states).
  • discrete states e.g., a “blood” state, an “unknown” state, a “tissue” state, a “boundary” state, and combinations thereof
  • uncertain states e.g. “unknown” states
  • the three-dimensional data structure 134 can be based on known positions of the tip section 124 of the medical device 104 in the anatomic structure 132 , it should be appreciated that other methods of determining the three-dimensional data structure 134 are additionally or alternatively possible.
  • the three-dimensional data structure 134 can be based on images of the surface 133 of the anatomic structure 132 acquired prior to or during the medical procedure, particularly in use cases in which the images of the surface 133 of the anatomic structure 132 may not be complete. Such images can correspond to any of various, different imaging modalities including, for example, x-ray.
  • the three-dimensional data structure 134 can include, for example, a three-dimensional grid of voxels 135 .
  • the three-dimensional grid of voxels 135 can be any one or more of various different types three-dimensional grids well known in the art.
  • the three-dimensional grid of voxels 135 can include one or more of an occupancy grid and an occupancy field.
  • the three-dimensional grid of voxels 135 can include a volumetric grid representation.
  • Each voxel 135 can be a discrete element of volume. Together, the voxels 135 can form the three-dimensional data structure 134 which, in general, is a three-dimensional notational space. As described in greater detail below, a three-dimensional surface representation 136 can be formed in relation to the three-dimensional data structure 134 such that the three-dimensional surface representation 136 can represent the surface 133 of the anatomic structure 132 . In general, the three-dimensional surface representation 136 can be any one or more of the various different types well-known in the art and, thus, by way of non-exclusive example can include any one or more of the following: a “level set”; a “separating surface”; and an “implicit surface”.
  • the three-dimensional data structure 134 and/or the three-dimensional surface representation 136 can be necessarily based on certain inferences between data points. These inferences, while serving as a useful expedient for generation of the three-dimensional surface representation 136 , can result in discrepancies between the three-dimensional surface representation 136 and the surface 133 of the anatomic structure 132 upon which the three-dimensional surface representation 136 is based. Such discrepancies can, for example, result in one or more anatomic features of the surface 133 of the anatomic structure 132 being obscured, or at least distorted, in the resulting three-dimensional surface representation 136 .
  • volumetric smoothing the three-dimensional surface representation 136 can be based on surface tension methods.
  • volumetric smoothing the three-dimensional surface representation 136 can be based on hole filling methods.
  • volumetric smoothing the three-dimensional surface representation 136 can be based on interpolation.
  • volumetric smoothing of the three-dimensional surface representation 136 can be based on ball-pivoting.
  • discrepancies between the three-dimensional surface representation 136 and the surface 133 of the anatomic structure 132 can arise as a result of a trade-off between the degree of volumetric smoothing used to form the three-dimensional surface representation 136 and incompleteness or uncertainty of the data set forming the three-dimensional data structure 134 . That is, it can be desirable to specify a low degree of volumetric smoothing to achieve resolution of certain anatomic features in the three-dimensional surface representation 136 . Such resolution in one area, however, can have the unintended consequence of creating distortions in areas in which the data set is incomplete or uncertain, with such incompleteness or uncertainty being common in implementations in which the data set is based on locations in the anatomic structure 132 visited by the tip section 124 .
  • the three-dimensional surface representation 136 can appear to include numerous distortions (e.g., undulations or, more specifically, invaginations) that are not representative of the surface 133 of the anatomic structure 132 when a low degree of volumetric smoothing is applied to an incomplete or uncertain data set.
  • numerous distortions e.g., undulations or, more specifically, invaginations
  • increasing the amount of volumetric smoothing of the three-dimensional surface representation 136 to remove such distortions resulting from incompleteness or uncertainty of the data set can have the unintended consequence of obscuring or distorting certain anatomic features.
  • FIGS. 5A-5C are, collectively, a schematic depiction of a generalized example of distortions that can occur as a result of volumetric smoothing a three-dimensional surface representation based solely on surface tension, hole filling, interpolation, ball pivoting, or other similar implicit or explicit surface reconstruction methods applied to an incomplete or uncertain data set.
  • a point cloud 139 is shown superimposed on the surface 133 of the anatomic structure 132 ( FIG. 3 ).
  • the point cloud 139 can correspond to, for example, known locations of the tip section 124 of the medical device 104 in the anatomic structure 132 ( FIG. 3 ).
  • the point cloud 139 should be understood to include any of various different point clouds well known in the art of surface reconstruction. In general, it is desirable to generate a volumetrically smoothed surface mesh from the point cloud 139 to create a three-dimensional surface representation that accurately represents the surface 133 of the anatomic structure 132 ( FIG. 3 ). However, as described in greater detail below, distortions can occur when the point cloud 139 is an incomplete or uncertain data set of an anatomic structure and it is, nevertheless, desirable to represent accurately a geometric feature of the surface 133 .
  • FIG. 5A is a schematic representation of a three-dimensional surface representation 136 a of the point cloud 139 , with the three-dimensional surface representation 136 a volumetrically smoothed to a high degree.
  • a high degree of volumetric smoothing of the three-dimensional surface representation 136 a can result in significant deviations between the three-dimensional surface representation 136 a and the surface 133 (i.e., the intended shape).
  • the three-dimensional surface representation 136 a can be based on an interpolation between data points of the point cloud 139 that are not close to one another, and details of the shape between the data points may be lost.
  • FIG. 5B is a schematic representation of a three-dimensional surface representation 136 b of the point cloud 139 , with the three-dimensional surface representation 136 b smoothed to a low degree of volumetric smoothing. Accordingly, the three-dimensional surface representation 136 b is based on interpolation between data points of the point cloud 139 that are close to one another, as compared to the interpolation associated with FIG. 5A . As shown in FIG. 5B , the interpolation between data points in the point cloud 139 that are relatively close to one another can result in a shape that shows features that are not actually present on the surface 133 .
  • inaccuracies such as those shown in FIG. 5A and FIG. 5B
  • inaccuracies can be particularly prevalent in areas in which the surface 133 is represented by relatively few data points of the point cloud 139 .
  • the time associated with obtaining such a complete data set may be impermissibly long.
  • a useful solution to this trade-off between accuracy and challenges associated with data acquisition is to constrain volumetric smoothing in areas in which the shape of the surface 133 is known or known with some level of confidence.
  • one or more anchor portions 142 can be added to the data set to identify one or more known positions on the surface 133 .
  • the one or more anchor portions 142 can constrain the volumetric smoothing of the three-dimensional surface representation 136 c . Because the three-dimensional surface representation 136 c is constrained relative to the one or more anchor portions 142 , distortions or other types of inaccuracies associated with a high degree of volumetric smoothing ( FIG. 5A ) and a low degree of volumetric smoothing ( FIG. 5B ) are less prevalent in the three-dimensional surface representation 136 c ( FIG. 5C ).
  • the three-dimensional surface representation 136 c is a more accurate representation of the surface 133 than would be obtained without some form of constraint relative to the one or more anchor portions 142 , which represent corresponding positions known, optionally with some degree of confidence, to lie on the surface 133 .
  • the inaccuracies in the three-dimensional surface representation 136 that can result from volumetric smoothing can be particularly prevalent in locally concave areas of the surface 133 along a generally convex portion of the surface 133 of the anatomic structure 132 .
  • the concavity of the surface 133 of the anatomic structure 132 is expressed with respect to the shape of a blood-tissue boundary formed by the surface 133 of the anatomic structure 132 ( FIG. 3 ) around the blood. For example, as shown in FIG.
  • the anatomic structure 132 can include carina 137 a , 137 b , 137 c between anatomic elements 131 a , 131 b , 131 c , 131 d , 131 e of the anatomic structure 132 .
  • the surface 133 should be understood to be locally concave along each carina 137 a , 137 b , 137 c . In such instances, volumetric smoothing the three-dimensional surface representation 136 can result in an inaccurate depiction of the surface 133 of the anatomic structure 132 in the region of the carina 137 a , 137 b , 137 c .
  • Such an inaccurate depiction in the three-dimensional surface representation 136 in the vicinity of one or more of the carina 137 a , 137 b , 137 c can be problematic, for example, in instances in which it may be desirable to apply lesions, and thus position accurately the tip section 124 , in the vicinity of one or more of the carina 137 a , 137 b , 137 c .
  • the anatomic structure 132 is the left atrium
  • LSPV left superior pulmonary vein
  • LIPV left inferior pulmonary vein
  • anatomic structure 132 is depicted as the left atrium in FIG. 3 by way of example and not limitation, and, thus, it should be appreciated that the number and orientation of anatomic elements and corresponding carina (or other similar locally concavities) present in the anatomic structure 132 can depend on the type of anatomic structure.
  • the present disclosure is generally directed to constraining the three-dimensional surface representation 136 according to one or more of the methods described herein.
  • constraining the three-dimensional surface representation 136 as described herein can facilitate accurate representation of anatomic features of the surface 133 of the anatomic structure 132 while allowing the three-dimensional surface representation 136 to be generated efficiently (e.g., based on an incomplete or uncertain data set of locations in the anatomic structure 132 ).
  • the three-dimensional data structure 134 and the three-dimensional surface representation 136 can be stored on the storage medium 111 , along with instructions executable by the processing unit 109 to display the three-dimensional data structure 134 , the three-dimensional surface representation 136 , or both on the graphical user interface 110 , as described in greater detail below.
  • the instructions stored on the storage medium 111 and executable by the processing unit 109 to display one or both of the three-dimensional data structure 134 and the three-dimensional surface representation can be, for example, an application built using Visualization Toolkit, an open-source 3D computer graphics toolkit, available at www.vtk.org.
  • FIG. 4 is a schematic representation of the three-dimensional data structure 134 and the three-dimensional surface representation 136 projected onto a viewing window 138 of an image plane 140 . While the three-dimensional data structure 134 and the three-dimensional surface representation 136 are described as both being projected onto the viewing window 138 , it should be understood that the three-dimensional data structure 134 and the three-dimensional surface representation 136 can be individually projected to the viewing window 138 . For example, it may be desirable to project both the three-dimensional data structure 134 and the three-dimensional surface representation 136 onto the viewing window 138 while the three-dimensional surface representation 136 is being built.
  • the graphical user interface 110 can be two-dimensional such that the image plane 140 corresponds to a plane of the two-dimensional display of the graphical user interface 110 , and the viewing window 138 corresponds to a field of view of the two-dimensional display of the graphical user interface 110 . Accordingly, the image formed by projecting one or both of the three-dimensional data structure 134 and the three-dimensional surface representation 136 onto the viewing window 138 can be displayed on the graphical user interface 110 .
  • a physician can, in certain instances, interact with the projection of the three-dimensional data structure 134 on the graphical user interface 110 to identify one or more locations of anatomic features on the three-dimensional data structure 134 such that the three-dimensional surface representation 136 , generated based on the three-dimensional data structure 134 , reflects these identified anatomic features.
  • FIGS. 6A-6D collectively represent a sequence of images formed on the viewing window 138 and displayed on the graphical user interface 110 .
  • the sequence illustrates, in general, the generation of the three-dimensional surface representation 136 .
  • the three-dimensional data structure 134 and an unconstrained three-dimensional surface representation 136 ′ can be displayed on the graphical user interface 110 .
  • the tricuspid valve, the inferior vena cava, and the superior vena cava have each been cut in the unconstrained three-dimensional surface representation 136 ′ to facilitate visualization. These cuts do not constrain the surface extracted from the three-dimensional data structure 134 .
  • the unconstrained three-dimensional surface representation 136 ′ can represent a surface extracted from the three-dimensional data structure 134 without specific input regarding the position of anatomic features in the three-dimensional data structure 134 .
  • anatomic features of the surface 133 of the anatomic structure 132 may be obscured, distorted, or otherwise misrepresented in the unconstrained three-dimensional surface representation 136 ′.
  • the unconstrained three-dimensional surface representation 136 ′ may not accurately represent anatomic features of the surface 133 of the anatomic structure 132 ( FIG. 3 ) if the three-dimensional data structure 134 is based on an incomplete or uncertain data set.
  • attempts to form a surface mesh based on the three-dimensional data structure 134 can result in an obscured and/or distorted representation of the surface 133 of the anatomic structure 132 ( FIG. 3 ), particularly along locally concave portions of the surface 133 of the anatomic structure 132 ( FIG. 3 ).
  • volumetric smoothing required to create the three-dimensional surface representation 136 ′ based on an incomplete or uncertain data set can obscure or distort areas that are concave because the difference between missing data and an actual concave region is often not discernible by a smoothing algorithm, such as a surface tension algorithm, a hole filling algorithm, an interpolation algorithm, a ball-pivoting algorithm, or other similar algorithms.
  • a smoothing algorithm such as a surface tension algorithm, a hole filling algorithm, an interpolation algorithm, a ball-pivoting algorithm, or other similar algorithms.
  • volumetric smoothing required to form the three-dimensional surface representation 136 ′ based on an incomplete or uncertain data set can have the unintended consequence of covering over concave regions.
  • a low degree of volumetric smoothing can create distortions along other portions of the three-dimensional surface representation 136 ′, such as where the data set is incomplete or uncertain. Accordingly, while the degree of volumetric smoothing can be adjusted to capture aspects of local detail, such adjustments can have a negative impact on the accuracy of the overall shape of the three-dimensional surface representation 136 ′.
  • an advantageous alternative to varying the degree of volumetric smoothing of the surface mesh forming the three-dimensional surface representation 136 ′ can include receiving identification of the one or more anchor portions 142 on the three-dimensional data structure 134 .
  • Each anchor portion 142 can correspond to a predetermined number of voxels 135 of the three-dimensional data structure 134 . Additionally, or alternatively, each anchor portion 142 can correspond to one or more nodes or points in the three-dimensional data structure 134 .
  • each anchor portion 142 can include information regarding, for example, an orientation of the surface 133 of the anatomic structure 132 . Additionally, or alternatively, each anchor portion 142 can include information regarding, for example, a degree of confidence in the location of the anchor portions 142 and/or the associated information.
  • Identification of each anchor portion 142 can be based on input received from a physician (e.g., as a tag), input received from the tip section 124 in the anatomic structure 132 , or a combination thereof. More generally, it should be appreciated that the identification of each anchor portion 142 can be based on observations made or prior knowledge regarding the anatomic structure 132 and can be independent of parameters used to form the three-dimensional surface representation 136 ′.
  • Each anchor portion 142 can be represented on the graphical user interface 110 as visual indicia.
  • visual indicia can be useful, for example, as a visualization tool for the physician to assess how the three-dimensional surface representation 136 ′ will be modified as it is constrained to pass near a position relative to the anchor portion 142 .
  • the physician can reposition the anchor portion 142 .
  • the one or more anchor portions 142 can represent a position of an anatomic feature of the anatomic structure 132 .
  • the one or more anchor portions 142 can correspond to locations at which contact between the tip section 124 and the surface 133 of the anatomic structure 132 is detected such that the one or more anchor portions 142 represent a location known, optionally with some confidence, to lie on the surface 133 of the anatomic structure 132 .
  • the one or more anchor portions 142 can include information regarding, for example, a direction of contact corresponding to an orientation of the surface 133 of the anatomic structure 132 .
  • Contact detection forming the basis of the one or more anchor portions 142 can be based on any of various different forms and combinations of contact detection described herein or otherwise known in the art.
  • contact detection can be based on an intracardiac electrogram indicative of contact between the sensor 125 of the tip section 124 and the surface 133 .
  • contact detection can also, or instead, be based on a force signal (e.g., including magnitude, direction, or both) in implementations in which the sensor 125 is a force sensor.
  • contact detection can also, or instead, be based on detecting deformation (e.g., using the sensor 125 ) of at least a portion of the tip section 124 as the tip section 124 contacts the surface 133 .
  • contact detection can also, or instead, be based on detecting a change in impedance (e.g. using the sensor 125 ).
  • the one or more anchor portions 142 can be based on input from a physician.
  • the input from the physician can be based on the unconstrained three-dimensional surface representation 136 ′ and the physician's knowledge of anatomy. That is, the global shape of the unconstrained three-dimensional surface representation 136 ′ may represent the overall shape of surface 133 of the anatomic structure 132 , albeit with local inaccuracies with respect to some anatomic features.
  • the physician can provide an indication of one or more anchor portions 142 on the three-dimensional data structure 134 and/or on the unconstrained three-dimensional surface representation 136 ′ based on the physician's knowledge of anatomy (e.g., knowledge of the position of a carina in implementations directed to a heart cavity).
  • the physician can provide an indication of one or more anchor portions 142 on the three-dimensional data structure 134 and/or on the unconstrained three-dimensional surface representation 136 ′ based on observation of the tip section 124 and/or the shaft 122 . That is, in certain instances, the physician can push the tip section 124 gently and observe, optionally through fluoroscopy or other similar imaging modality, whether the tip section 124 ( FIG. 3 ) advances in response to the push. If the tip section 124 ( FIG. 3 ) does not advance in response to the push, the physician can manually tag the point as one of the one or more anchor portions 142 .
  • the physician can identify the one or more anchor portions 142 on the three-dimensional data structure 134 by providing inputs to the interface unit 108 ( FIG. 1 ) (e.g., through a keyboard, a mouse, or other input associated with the interface unit 108 ).
  • the three-dimensional surface representation 136 can be generated as a mesh (e.g., a continuous polygonal mesh) of a surface extracted from the three-dimensional data structure 134 and constrained relative to the one or more anchor portions 142 .
  • constraining the three-dimensional surface representation 136 relative to the one or more anchor portions 142 can include any one or more of various different methods that, as compared to the absence of a constraint, reduces the distance between the three-dimensional surface representation 136 and the one or more anchor portions 142 .
  • constraining the three-dimensional surface representation 136 relative to the one or more anchor portions 142 can include constraining the three-dimensional surface representation 136 to pass through the one or more anchor portions 142 .
  • constraining the three-dimensional surface representation 136 relative to the one or more anchor portions 142 can include constraining the three-dimensional surface representation 136 to be at or within a distance (e.g., a fixed distance) relative to the one or more anchor portions 142 .
  • the three-dimensional surface representation 136 can be constrained by a penalty (e.g., cost) function that penalizes for distance from the one or more anchor portions 142 but does not strictly restrict the three-dimensional surface representation 136 to pass within a specified distance of the one or more anchor portions 142 .
  • constraining the three-dimensional surface representation 136 relative to the one or more anchor portions can include setting a fixed value of one or more nodes associated with the one or more anchor portions 142 in the three-dimensional data structure 134 , determining a scalar function on at least a portion of the three-dimensional data structure 134 , and applying an algorithm to extract an isosurface based in part on the scalar function.
  • An indicator function is a well-known example of such a scalar function.
  • An embedding function is another well-known example of such a scalar function.
  • the isosurface extracted based at least in part on the scalar function can include any of various different isosurfaces known in the art and, thus, among various examples, can include an implicit surface.
  • constraining the three-dimensional surface representation 136 can include a clamped signed distance function based only on space carving information, as is well known in the art.
  • a first fixed value can be assigned to nodes, in the three-dimensional data structure 134 , corresponding to locations within the anatomic structure 132 ( FIG. 3 ) visited by the tip section 124 of the medical device 104 ( FIG. 2 ). Because the tip section 124 of the medical device 104 ( FIG. 2 ) can only move through blood, the nodes with the first fixed value in the three-dimensional data structure 134 imply the presence of blood at those positions. However, because the tip section 124 of the medical device 104 ( FIG. 2 ) does not generally move through every location within the anatomic structure 132 ( FIG.
  • the nodes that have not been set to the first fixed value can correspond to one of two physical states. That is, the nodes that have not been set to the first fixed value can correspond to i) locations in the anatomic structure 132 ( FIG. 3 ) that have not been visited by the tip section of the medical device 104 ( FIG. 2 ) or ii) locations that are outside of the volume of the anatomic structure 132 in which the tip section 124 of the medical device 104 ( FIG. 2 ) is moving.
  • inaccuracies in the three-dimensional surface representation 136 can arise from mischaracterization of these two categories of nodes that have not been set to the first fixed value.
  • the one or more anchor portions 142 can provide a constraint that is useful for more accurately characterizing nodes that that have not been set to the first fixed value.
  • the one or more anchor portions 142 can be set to a second fixed value, different from the first fixed value. It should be appreciated that the second fixed value can be one of two values of a signed clamped distance function, and the anchor portions 142 can correspond to surfaces acquired from a range scan.
  • nodes corresponding to the first fixed value can correspond to the known locations of blood while nodes corresponding to the second fixed value can correspond to the known locations of the surface 133 ( FIG. 3 ) and, thus, the blood-tissue boundary.
  • the values of the remaining nodes in the three-dimensional data structure 134 i.e., the nodes that have been assigned neither the first fixed value nor the second fixed value are variable.
  • a volumetric smoothing algorithm e.g., a three-dimensional smoothing kernel
  • the values of these variable nodes can take on values that are a function of neighboring nodes.
  • the values of the variable nodes can take on values that are a function of neighboring nodes according to a thermal diffusion model.
  • the one or more anchor portions 142 can modify the values of neighboring variable nodes as the three-dimensional data structure 134 undergoes volumetric smoothing.
  • a spatial convolution can be applied one or more times to calculate new values for only the variable nodes.
  • the three-dimensional surface representation 136 can be based on the three-dimensional data structure 134 having fixed-value nodes as described above.
  • the three-dimensional surface representation 136 can correspond to an isosurface (e.g., a level set, an implicit surface, etc.) extracted from the three-dimensional data structure 134 having fixed-value nodes.
  • a “marching cubes” algorithm can be applied to the volumetrically smoothed three-dimensional data structure 134 having fixed-value nodes to extract an isosurface corresponding to a fixed value (e.g., a value between the first fixed value associated with known locations of the tip section 124 of the medical device 104 ( FIG. 2 ) and the second fixed value associated with the one or more anchor portions 142 ).
  • the one or more anchor portions 142 can have the effect of constraining, or otherwise limiting, the position of the resulting three-dimensional surface representation 136 extracted from the three-dimensional data structure 134 .
  • a “flying edges” algorithm can be applied to the volumetrically smoothed three-dimensional data structure 134 to extract an isosurface.
  • the three-dimensional surface representation 136 can be based on any of various different algorithms well known in the art for extracting a mesh of a surface from the three-dimensional data structure 134 .
  • the three-dimensional surface representation 136 can be a polygonal mesh extracted from the three-dimensional data structure 134 based on a “marching cubes” algorithm and constrained relative to the one or more anchor points 142 according to any of the various different methods described herein.
  • a “marching cubes” algorithm can include any one or more algorithms in which a polygonal mesh of an isosurface is extracted from the three-dimensional data structure 134 based on analysis of node values in the three-dimensional data structure 134 .
  • the three-dimensional surface representation 136 can be extracted from the three-dimensional data structure 134 according to any one or more computational algorithms known in the art for volumetrically smoothing three-dimensional representations of objects including a “ball-pivoting” algorithm, a “power crust” algorithm, and other similar algorithms.
  • an adaptive ball-pivoting algorithm can constrain the three-dimensional surface representation 136 relative to the anchor portions 142 . That is, away from the anchor portions 142 , the ball forming the basis of the ball-pivoting algorithm can be a fixed global size that yields volumetric smoothing where specific information about anatomical features is not available.
  • the fixed global ball can have a diameter between about 10 mm and 30 mm (e.g. about 15 mm). Close to the anchor portions 142 , the size of the ball can be reduced to facilitate passing the surface representation 136 closer to the anchor portions 142 . For example, as compared to the fixed global size of the ball away from the anchor portions 142 , the size of the ball can be reduced closer to the anchor portions 142 .
  • the three-dimensional surface representation 136 can pass closer to the anchor portions 142 .
  • the size of the ball at a surface location closest to a respective one of the anchor portions 142 can be chosen such that the three-dimensional surface representation 136 lies within a predetermined minimum distance to the respective anchor portion 142 .
  • the size of the ball can vary between the reduced ball size and the global ball size as a function (e.g., a linear function) of distance from the ball to the one or more anchor portions 142 until the ball size reaches the global ball size, beyond which distance the global ball size can remain at the fixed global ball size.
  • the three-dimensional surface representation 136 can be volumetrically smoothed and, optionally, surface smoothed.
  • volumetric smoothing of the three-dimensional surface representation 136 can be accomplished using any one or more of various different volumetric smoothing techniques that are well known in the art and described herein.
  • Surface smoothing can additionally, or alternatively, be accomplished using any one or more of various different surface smoothing techniques well known in the art.
  • An example of such a surface smoothing technique is Laplacian smoothing and variations thereof.
  • three-dimensional surface representation 136 may no longer satisfy the constraints previously applied as part of the volumetric smoothing process. For example, as shown in FIG.
  • the surface-smoothed three-dimensional surface representation 136 may contain fewer than all of the locations visited by the medical device. Further, or instead, the distance between the three-dimensional surface representation 136 and the anchor portions 142 may change as the three-dimensional surface representation 136 is subjected to surface smoothing.
  • the three-dimensional surface representation 136 differs from the unconstrained three-dimensional surface representation 136 ′ near the one or more anchor portions 142 .
  • the three-dimensional surface representation 136 depicts anatomic features that are not readily apparent in the unconstrained three-dimensional surface representation 136 ′.
  • the one or more anchor portions 142 can facilitate efficiently generating an accurate representation of anatomic features of the anatomic structure 132 ( FIG. 3 ).
  • the display of the three-dimensional data structure 134 of FIG. 6C can be hidden (e.g., made translucent) such that the three-dimensional surface representation 136 can be displayed by itself.
  • the display of the three-dimensional surface representation 136 by itself and, optionally, in smoothed form can be useful, for example, for facilitating perception by the physician.
  • FIGS. 6A-6D have been shown and described as occurring in sequence for the sake of clarity of explanation. It should be appreciated, however, that in addition to, or as an alternative, any one or more of the steps shown in FIGS. 6A-6D can be combined, performed in parallel, and/or varied in order.
  • the computer executable instructions stored on the storage medium 111 can cause the processing unit 109 ( FIG. 1 ) to generate the three-dimensional surface representation 136 according to one or more of the following exemplary methods. Unless otherwise indicated or made clear from the context, each of the following exemplary methods can be implemented using the system 100 ( FIG. 1 ) and/or one or more components thereof.
  • FIG. 7 is a flowchart of an exemplary method 160 of representing a surface of an anatomic structure.
  • the exemplary method 160 can include receiving 162 a plurality of location signals of a medical device, forming 164 a three-dimensional data structure representing volumes, within the anatomic structure, occupied by the medical device at the locations corresponding to the plurality of location signals, receiving 166 one or more anchor portions representing locations relative to the anatomic structure, and generating 167 a three-dimensional surface representation of the anatomic structure.
  • the generated 167 three-dimensional surface representation can be constrained relative to the one or more anchor portions and to contain at least a portion of the three-dimensional data structure. That is, in general, the one or more anchor portions can modify the behavior of an algorithm in a way that constrains the resulting three-dimensional surface representation relative to the anchor portions.
  • receiving 162 the plurality of location signals of the medical device can include receiving one or more signals indicative of the location of the medical device according to any one or more of the methods described herein.
  • the received 162 plurality of location signals can be, for example, a plurality of signals received from a single sensor over a period of time.
  • the received 162 plurality of location signals can be one or more signals from a magnetic position sensor (such as the magnetic position sensor 130 described above with respect to FIG. 2 ).
  • the received 162 plurality of location signals can be a plurality of signals received from multiple, different types of sensors that individually, or in combination, provide information regarding the location of the medical device in the anatomic structure.
  • receiving 162 the plurality of location signals of the medical device can include receiving one or more signals from a magnetic position sensor and from a sensor providing a signal indicative of a blood-tissue boundary of the anatomic structure.
  • the received 162 plurality of location signals of the medical device generally correspond to locations visited by a tip section of the medical device (e.g., the tip section 124 of the medical device 104 of FIG. 2 ) in the anatomic structure. More generally, however, it should be understood that the plurality of location signals of the medical device can correspond to locations visited by any portion of the medical device that can be sensed or otherwise determined in the anatomic structure.
  • Forming 164 the three-dimensional data structure representing locations, within the anatomic structure, visited by the medical device can include forming a three-dimensional grid, with locations in the three-dimensional grid reflecting locations visited by the medical device in the anatomic structure.
  • a boundary of the anatomic structure can be approximated through analysis of neighbor locations and/or node values of the three-dimensional grid such as, for example, through application of one or more of a “marching cubes” algorithm, a “ball-pivoting” algorithm, and a “power crust” algorithm, with the algorithm or algorithms extracting a surface from the three-dimensional grid.
  • the one or more anchor portions e.g., the anchor portions 142 in FIGS. 5A-5C
  • receiving 166 the one or more anchor portions representing locations relative to the anatomic structure can include receiving input from one or more sources.
  • receiving 166 the one or more anchor portions can be based on input received from a physician, input received from one or more sensors on the medical device, or a combination thereof.
  • receiving 166 the one or more anchor portions is based on a combination of input from the physician and from one or more sensors on the medical device, it can be advantageous to have a hierarchy of input such that, for example, the input from the physician can override the input from the one or more sensors on the medical device.
  • Receiving 166 the one or more anchor portions representing locations relative to the anatomic structure can include receiving, from one or more sensors disposed on the medical device, a signal indicative of contact between the medical device and tissue of the anatomic structure.
  • Sensed contact e.g., one or more of location, direction, force, consistency, and/or duration of contact
  • one or more anchor portions can be identified at the location of the sensed contact to ensure that the three-dimensional surface representation is constrained relative to the sensed contact, which is known to represent the blood-tissue boundary.
  • the signal indicative of contact between the medical device and tissue of the anatomic chamber can include any one or more of the signals indicative of contact described herein.
  • the signal indicative of contact between the medical device and tissue of the anatomic chamber can include an impedance signal (e.g., a change in impedance) from one or more impedance sensors (e.g., the sensor 125 in FIG. 2 ) disposed on the medical device.
  • signals indicative of contact between the medical device and the surface of the anatomic structure of the patient can include one or more of: a change in an electrical signal (e.g., electrogram or impedance) in one or more electrodes of the medical device; a force detected by a force sensor of the medical device; an ultrasound signal of an ultrasound sensor on the medical device; and a deformation of at least a portion of the medical device.
  • a signal indicative of contact between the medical device and the surface of the anatomic structure of the patient can include an amplitude derived from an electrogram detected by one or more electrodes of the medical device.
  • Receiving 166 the one or more anchor portions can include identification of a subset of the plurality of received location signals.
  • Identification of the subset of the plurality of received location signals can, for example, include an input command from the physician identifying one or more portions of the three-dimensional data structure as corresponding one or more anchor portions.
  • the input command can be received from any of various, different input devices such as a keyboard, a mouse, a touchscreen, etc. and, additionally, or alternatively, can include voice commands.
  • the physician can provide input through one or more input devices to identify the subset as one or more voxels of the three-dimensional grid, as displayed on a graphical user interface.
  • receiving 166 the one or more anchor portions can include receiving a respective confidence level associated with the one or more anchor portions.
  • a confidence level can increase substantially monotonically with a measured indication of contact (e.g. electrogram amplitude, impedance, force, deformation, and/or proximity).
  • constraining the three-dimensional surface representation relative to the one or more anchor portions can be based on the respective confidence level associated with each of the one or more anchor portions.
  • the confidence levels can form a basis for certain of the one or more anchor portions acting as stronger or weaker anchor portions relative to other anchor portions. That is, an anchor portion corresponding to a higher confidence level can act as a stronger anchor as compared to an anchor portion corresponding to a weaker confidence level.
  • an anchor portion identified with contact in a known direction can constrain the normal direction of the resulting surface using any of various different techniques known in the art.
  • generating 167 the three-dimensional surface representation of the anatomic structure can include any one or more of the methods described herein for forming a three-dimensional surface.
  • generating 167 the three-dimensional surface representation can include extracting a surface from the three-dimensional data structure according to an algorithm, such as one or more of a “marching cubes” algorithm, a “ball-pivoting” algorithm, and a “power crust” algorithm, in which the three-dimensional surface representation is constrained relative to the one or more anchor portions according to any one or more of the various different methods of constraint described herein.
  • the three-dimensional surface representation can include a mesh (e.g., a continuous mesh).
  • the mesh can be formed of, for example, a plurality of polygons (e.g., triangles) combined together to represent contours of the surface of the anatomic structure.
  • the generated 167 three-dimensional surface representation can be smoothed according to any of various different smoothing techniques known in the art to provide a more realistic representation of the surface of the anatomic structure.
  • the exemplary method 160 can optionally include representing 168 , on a graphical user interface, any of various different combinations of the three-dimensional surface representation, the anchor portions, and the three-dimensional data structure.
  • the graphical user interface can be, for example, a two-dimensional graphical user interface such as the graphical user interface 110 ( FIG. 1 ). Accordingly, the exemplary method 160 can include representing 168 , on the graphical user interface, a two-dimensional projection of the three-dimensional surface representation. In addition, or in the alternative, the exemplary method 160 can include representing 168 , on the graphical user interface, a two-dimensional projection of the three-dimensional data structure.
  • the exemplary method 160 can further include representing 168 the one or more anchor portions on the graphical user interface.
  • the one or more anchor portions can be represented 168 on the graphical user interface on a projection of the three-dimensional data structure, on a projection of the three-dimensional surface representation, or both.
  • the one or more anchor portions can be represented 168 on the graphical user interface separately from the three-dimensional data structure and/or the three-dimensional surface. It should be appreciated that representing 168 the one or more anchor portions on the graphical user interface can, for example, facilitate modification of the one or more anchor portions in certain instances.
  • the one or more anchor portions can be represented 168 , on the graphical user interface, as annotations on the three-dimensional surface representation of the anatomic structure.
  • the annotations can include, for example, tags of corresponding anatomic features, tags corresponding to locations for application of treatment (e.g., ablation), or combinations thereof.
  • the annotations can constrain the three-dimensional surface representation to remain unchanged as other anchor portions are added.
  • the three-dimensional surface representation can be constrained to pass through a portion of the three-dimensional data structure nearest to the annotation.
  • the exemplary method 160 can optionally include determining 169 whether the one or more anchor portions have been modified. If the one or more anchor portions are determined 169 to be modified, the generating step 167 can be repeated. Thus, in general, the exemplary method 160 can be iterative. That is, in response to the generated 167 three-dimensional surface representation, the physician can continue to make modifications as necessary. These modifications can be based on one or more inputs received from any one or more of various input devices known in the art and described herein. Accordingly, modifying the one or more anchor portions can be based on one or more inputs from a keyboard, a mouse, a touchscreen, the medical device, or combinations thereof.
  • Modifying the one or more anchor portions can include removing at least one of the one or more anchor portions. Such removal can be useful, in certain instances, for adjusting the three-dimensional surface representation (e.g., after the three-dimensional surface representation has been generated 167 ) to achieve a shape that is more accurate. Additionally, or alternatively, removal of at least one of the one or more anchor portions can correct an incorrectly identified anchor portion. It should be appreciated, therefore, that removal of at least one of the one or more anchor portions can serve as an “undo” function such that correction of an incorrectly identified anchor portion does not require the physician to engage, for example, in a complex editing process.
  • modifying the identified one or more anchor portions and repeating the generating step 167 as part of the iterative process described herein can facilitate efficient and accurate generation of the three-dimensional surface representation of the anatomic structure, as compared to tools that allow a user to selectively delete subvolumes. That is, selecting a subvolume on a two-dimensional graphical user interface commonly requires multiple selection steps from different views, which can be time consuming and subject to inaccuracies and can often require complex user interaction.
  • FIG. 8 is a flowchart of an exemplary method 170 of representing a surface of a heart cavity of a patient.
  • the exemplary method 170 can include forming 172 a three-dimensional data structure based on received locations of a tip section of a cardiac catheter in a heart cavity of a patient, receiving 174 identification of one or more anchor portions representing locations within the heart cavity, and generating 176 a three-dimensional surface representation of the heart cavity of the patient.
  • the three-dimensional surface representation can be generated using information from the three-dimensional data structure and can be constrained relative to the one or more anchor portions.
  • the tip section of the catheter can be, for example, the tip section 124 described with respect to FIGS. 2 and 3 .
  • a heart cavity is an example of the anatomic structure 132 .
  • the tip section can interact with the heart cavity in any manner and form described herein with respect to the interaction of the tip section 124 and the anatomic structure 132 .
  • Forming 172 the three-dimensional data structure can include any one or more of the various different methods of forming a three-dimensional data structure disclosed herein.
  • forming 172 the three-dimensional data structure can be analogous to forming 164 the three-dimensional data structure as described with respect to FIG. 7 .
  • forming 172 the three-dimensional data structure can be based on locations visited by the catheter in the heart cavity.
  • voxels corresponding to visited locations of the catheter can be set to a different state than voxels corresponding to locations that have not been visited by the catheter.
  • receiving 174 the one or more anchor portions on the three-dimensional data structure can be analogous to receiving 166 the one or more anchor portions on the three-dimensional data structure, as described with respect to FIG. 7 .
  • receiving 174 the one or more anchor portions on the three-dimensional data structure can include receiving an input command from a user interface (e.g., an input device such as a keyboard, a mouse, a touchscreen, and the like) corresponding to a location of an anatomic feature and/or receiving a signal indicative of contact between the catheter and tissue in the heart cavity.
  • a user interface e.g., an input device such as a keyboard, a mouse, a touchscreen, and the like
  • the one or more anchor portions can correspond, for example, to the location of one or more carina associated with the heart cavity.
  • the one or more anchor portions can correspond to a carina between the left atrial appendage (LAA) and the left superior pulmonary vein (LSPV).
  • LAA left atrial appendage
  • LSPV left superior
  • receiving 174 the one or more anchor portions can include receiving one or more signals corresponding to one or more respective locations of the cardiac catheter in the heart cavity.
  • the one or more signals corresponding to one or more respective locations of the cardiac catheter in the heart cavity can correspond to a blood-tissue boundary of the heart cavity.
  • Such signals can include, for example, to one or more of: a change in an electric signal (e.g., electrogram or impedance) detected by one or more electrodes of the catheter, a force detected by a force sensor of the catheter, an ultrasound signal of an ultrasound sensor of the catheter, and a deformation of at least a portion of the catheter.
  • such signals can correspond to an amplitude derived from an electrogram detected by one or more electrodes of the medical device.
  • Generating 176 the three-dimensional surface representation of the heart cavity of the patient can be, in general, analogous to generating 167 the three-dimensional surface representation of the anatomic structure as described with respect to FIG. 7 . Accordingly, it should be understood that generating 176 the three-dimensional surface representation of the heart cavity can be based, for example, on one or more of a “marching cubes” algorithm, a “ball-pivoting” algorithm, and a “power crust algorithm,” with the algorithm or algorithms constrained relative to the one or more anchor portions.
  • generating 176 the three-dimensional surface representation of the heart cavity of the patient can include an undo function such that one or more of the anchor portions can be removed or repositioned, and the three-dimensional surface representation of the heart cavity can be regenerated based on the updated position of the one or more anchor portions.
  • the three-dimensional surface representation of the heart cavity can be surface smoothed. It should be appreciated that such surface smoothing can produce changes to the three-dimensional surface representation and, in certain instances, can cause the three-dimensional surface representation to no longer fully contain the visited locations. Additionally, or alternatively, surface smoothing can produce changes to the three-dimensional surface representation that can cause the three-dimensional surface representation to no longer pass directly through the anchor portions. Surface-smoothing the three-dimensional surface representation can result in any one or more of the various different advantages described herein. For example, surface smoothing the three-dimensional surface representation can facilitate visualization of the position of the catheter relative to the three-dimensional surface representation, which can be useful for positioning the catheter during an ablation treatment applied to a surface of the heart cavity.
  • the exemplary method 170 can further include representing 177 , on a graphical user interface, at least one of a two-dimensional projection of the three-dimensional data structure, the one or more anchor portions, and a two-dimensional projection of the three-dimensional surface representation. Representing 177 the two-dimensional projection of the three-dimensional surface representation and generating 176 the three-dimensional surface representation can be part of an iterative process, such as an iterative process analogous to the iterative process described with respect to the exemplary method 160 .
  • the graphical user interface can be, for example, the graphical user interface 110 described with respect to FIG. 1 .
  • the one or more anchor portions can be represented 177 , on the graphical user interface, as annotations or other similar visual indicia on the three-dimensional surface representation of the heart cavity, on the three-dimensional data structure, or both. Additionally, or alternatively, the one or more anchor portions can be represented 177 on the graphical user interface independently of the three-dimensional surface representation, the three-dimensional data structure, or both.
  • the annotations can, for example, be tags of certain anatomic features and/or tags related to the position of a treatment (such as a lesion created through tissue ablation).
  • any one or more of the graphical user interfaces described herein can additionally, or alternatively, include a three-dimensional display.
  • a three-dimensional display examples include an augmented reality environment, a virtual reality environment, and combinations thereof.
  • the above systems, devices, methods, processes, and the like may be realized in hardware, software, or any combination of these suitable for a particular application.
  • the hardware may include a general-purpose computer and/or dedicated computing device. This includes realization in one or more microprocessors, microcontrollers, embedded microcontrollers, programmable digital signal processors or other programmable devices or processing circuitry, along with internal and/or external memory. This may also, or instead, include one or more application specific integrated circuits, programmable gate arrays, programmable array logic components, or any other device or devices that may be configured to process electronic signals.
  • a realization of the processes or devices described above may include computer-executable code created using a structured programming language such as C, an object oriented programming language such as C++, or any other high-level or low level programming language (including assembly languages, hardware description languages, and database programming languages and technologies) that may be stored, compiled or interpreted to run on one of the above devices, as well as heterogeneous combinations of processors, processor architectures, or combinations of different hardware and software.
  • the methods may be embodied in systems that perform the steps thereof, and may be distributed across devices in a number of ways. At the same time, processing may be distributed across devices such as the various systems described above, or all of the functionality may be integrated into a dedicated, standalone device or other hardware.
  • means for performing the steps associated with the processes described above may include any of the hardware and/or software described above. All such permutations and combinations are intended to fall within the scope of the present disclosure.
  • Embodiments disclosed herein may include computer program products comprising computer-executable code or computer-usable code that, when executing on one or more computing devices, performs any and/or all of the steps thereof.
  • the code may be stored in a non-transitory fashion in a computer memory, which may be a memory from which the program executes (such as random access memory associated with a processor), or a storage device such as a disk drive, flash memory or any other optical, electromagnetic, magnetic, infrared or other device or combination of devices.
  • any of the systems and methods described above may be embodied in any suitable transmission or propagation medium carrying computer-executable code and/or any inputs or outputs from same.
  • performing the step of X includes any suitable method for causing another party such as a remote user, a remote processing resource (e.g., a server or cloud computer) or a machine to perform the step of X.
  • performing steps X, Y and Z may include any method of directing or controlling any combination of such other individuals or resources to perform steps X, Y and Z to obtain the benefit of such steps.

Landscapes

  • Health & Medical Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Surgery (AREA)
  • Medical Informatics (AREA)
  • Public Health (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Veterinary Medicine (AREA)
  • Animal Behavior & Ethology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Pathology (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Computer Graphics (AREA)
  • Robotics (AREA)
  • Geometry (AREA)
  • Biophysics (AREA)
  • Cardiology (AREA)
  • Epidemiology (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Primary Health Care (AREA)
  • Computer Hardware Design (AREA)
  • Plasma & Fusion (AREA)
  • General Engineering & Computer Science (AREA)
  • Otolaryngology (AREA)
  • Architecture (AREA)
  • Human Computer Interaction (AREA)
  • Ultra Sonic Daignosis Equipment (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)
  • Measuring And Recording Apparatus For Diagnosis (AREA)

Abstract

Devices, systems, and methods of the present disclosure are directed to generating three-dimensional surface representations of an anatomic structure such as a heart cavity. More specifically, a three-dimensional surface representation of the anatomic structure is constrained relative to one or more anchor portions corresponding to received input regarding the location of anatomic features of the anatomic structure. The resulting three-dimensional surface representation includes salient features of the anatomic structure and, therefore, can be useful as visualization tool during any of various different medical procedures, including, for example, cardiac ablation.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. application Ser. No. 16/945,785, filed Jul. 31, 2020, now pending, which is a continuation of U.S. application Ser. No. 16/525,363, filed Jul. 29, 2019, now U.S. Pat. No. 10,765,481, which is a continuation of U.S. application Ser. No. 15/592,815, filed May 11, 2017, now U.S. Pat. No. 10,376,320, which claims the benefit under 35 U.S.C. § 119(e) of U.S. Prov. App. No. 62/334,577, filed May 11, 2016, U.S. Prov. App. No. 62/338,105, filed May 18, 2016, and U.S. Prov. App. No. 62/393,876, filed Sep. 13, 2016, with the entire contents of each of these applications hereby incorporated herein by reference.
  • BACKGROUND
  • Three-dimensional models can be used to assist in the placement or use of a device when such placement or use is not easily observable or practical. For example, in medical procedures, three-dimensional models are used to assist in the placement and use of medical devices for diagnosis or treatment of patients. An example of such a medical procedure carried out with the assistance of a three-dimensional model is the use of a catheter to deliver radio frequency (“RF”) ablation to form lesions that interrupt abnormal conduction in cardiac tissue, thus terminating certain arrhythmias in the heart.
  • SUMMARY
  • The present disclosure is directed to devices, systems, and methods of generating an accurate three-dimensional model of an anatomic structure of a patient to facilitate, for example, moving a medical device through the anatomic structure during a medical procedure in which the three-dimensional model is used to visualize the medical device in the anatomic structure. For example, the systems and methods of the present disclosure can be used to generate a three-dimensional model based on input (e.g., from a physician) of anchor portions corresponding to the position of anatomic features of the anatomic structure. As a more specific example, the systems and methods of the present disclosure can be used to generate a three-dimensional surface representation of the anatomic structure, with the three-dimensional surface representation constrained relative to one or more anchor portions identified on a three-dimensional data structure. Because the constraint imposed by the one or more anchor portions can change the shape of the three-dimensional surface representation and, thus, can have the appearance of changing the position of tissue in a visual representation of the three-dimensional surface representation, the constraint imposed by the anchor portions is sometimes referred to herein as “pinch.” More generally, the devices, systems, and methods of the present disclosure can provide a physician with a greater amount of control over a three-dimensional model of an anatomic structure and, additionally or alternatively, can represent salient features of the anatomic structure in greater detail than is typically achievable in a three-dimensional model built based solely on an incomplete or uncertain data set of known locations of a medical device in an anatomic structure.
  • According to one aspect, a method includes receiving a plurality of location signals, each received location signal indicative of a respective location of a medical device in an anatomic structure of a patient, forming a three-dimensional data structure representing locations, within the anatomic structure, visited by the medical device at the locations corresponding to the plurality of location signals, receiving one or more anchor portions representing locations relative to the anatomic structure, and generating a three-dimensional surface representation of the anatomic structure of the patient, the three-dimensional surface representation of the anatomic structure of the patient constrained relative to the one or more anchor portions and to contain at least a portion of the three-dimensional data structure.
  • In certain implementations, the method can further include displaying, on a graphical user interface, at least one of a two-dimensional projection of the three-dimensional data structure, the one or more anchor portions, and a two-dimensional projection of the three-dimensional surface representation.
  • In some implementations, receiving the one or more anchor portions representing locations relative to the anatomic structure can include receiving, from one or more sensors disposed on the medical device, a signal indicative of contact between the medical device and tissue of the anatomic structure. The signal indicative of contact can be, for example, indicative of a blood-tissue boundary of the anatomic structure of the patient. Additionally, or alternatively, the signal indicative of contact can include one or more of: a change in impedance detected by one or more electrodes of the medical device, a force detected by a force sensor of the medical device, an ultrasound signal of an ultrasound sensor of the medical device, a deformation of at least a portion of the medical device, and an amplitude derived from an electrogram detected by one or more electrodes of the medical device.
  • In certain implementations, receiving the one or more anchor portions can include receiving an input command from a user.
  • In some implementations, receiving the one or more anchor portions can include identifying a subset of the three-dimensional data structure.
  • In certain implementations, receiving the one or more anchor portions can include receiving a respective confidence level associated each of the one or more anchor portions, and constraining the three-dimensional surface representation relative to the one or more anchor portions is based on the respective confidence level associated with each of the one or more anchor portions.
  • In some implementations, the method can further include representing, on a graphical user interface, the one or more anchor portions as annotations on the three-dimensional surface representation of the anatomic structure.
  • In certain implementations, the method can further include representing, on a graphical user interface, the one or more anchor portions as annotations on the three-dimensional data structure.
  • In some implementations, the method can further include determining whether the one or more anchor portions have been modified and, based on whether the one or more anchor portions have been modified, repeating the generating step. Determining whether the one or more anchor portions have been modified can include, for example, determining whether one or more of previously identified anchor portions have been removed.
  • In certain implementations, the three-dimensional surface representation of the anatomic structure can be a continuous mesh.
  • According to another aspect, a method includes forming a three-dimensional data structure based on received locations of a tip section of a cardiac catheter in a heart cavity of a patient, receiving one or more anchor portions representing locations relative to the heart cavity, and generating a three-dimensional surface representation of the heart cavity of the patient, the surface representation of the heart cavity of the patient constrained relative to the anchor portions and to contain at least a portion of the three-dimensional data structure.
  • In some implementations, the method can further include displaying, on a graphical user interface, at least one of a two-dimensional projection of the three-dimensional data structure, the one or more anchor portions, and a two-dimensional projection of the generated three-dimensional surface representation.
  • In certain implementations, receiving the one or more anchor portions on the three-dimensional data structure can include receiving one or more location signals indicative of one or more respective locations of the cardiac catheter in the heart cavity.
  • In some implementations, receiving the one or more anchor portions can include receiving, from a sensor disposed on the cardiac catheter, a signal indicative of a blood-tissue boundary of the heart cavity of the patient. The signal corresponding to the blood-tissue boundary can include one or more of: a change in impedance detected by one or more electrodes of the cardiac catheter, a force detected by a force sensor of the cardiac catheter, an ultrasound signal of an ultrasound sensor of the cardiac catheter, and a deformation of at least a portion of the cardiac catheter, and an amplitude derived from an electrogram detected by one or more electrodes of the cardiac catheter.
  • In certain implementations, receiving the one or more anchor portions on the three-dimensional data structure can include receiving an input command from a user interface.
  • In some implementations, the method can further include representing, on a graphical user interface, the one or more anchor portions as annotations on the three-dimensional surface representation of the heart cavity.
  • According to still another aspect, a non-transitory, computer-readable storage medium has stored thereon computer executable instructions for causing one or more processors to: receive a plurality of location signals, each received location signal indicative of a respective location of a medical device in an anatomic structure of a patient; form a three-dimensional data structure representing volumes, within the anatomic structure, occupied by the medical device at the locations corresponding to the plurality of location signals; receive one or more anchor portions representing locations relative to the anatomic structure; and generate a three-dimensional surface representation of the anatomic structure of the patient, the three-dimensional surface representation of the anatomic structure of the patient constrained relative to the one or more anchor portions and containing at least a portion of the three-dimensional data structure.
  • Implementations can include one or more of the following advantages.
  • In certain implementations, a three-dimensional surface representation of an anatomic structure can be based on one or more anchor portions. For example, the three-dimensional surface representation of the anatomic structure can be constrained to pass near the one or more anchor portions and/or to pass near a fixed position relative to the one or more anchor portions. By imposing such conditions, the three-dimensional surface representation can accurately represent an anatomic structure such as, for example, an anatomic structure with local concavities along a generally convex shape. Such accurate representation can be particularly advantageous for proper manipulation of a catheter for the accurate placement of lesions in the heart, such as placement of lesions in the carina between pulmonary veins. Further, constraining the three-dimensional representation relative to one or more anchor portions can facilitate generating an accurate representation of the anatomic structure based on relatively few data points (e.g., an incomplete or uncertain data set) while still providing a useful representation of salient anatomic features of the anatomic structure. Thus, for example, constraining the three-dimensional surface representation relative to one or more anchor portions can facilitate building a useful depiction of the anatomic structure in less time than would ordinarily be required to achieve the same level of detail in a model built based on catheter position alone. Additionally, or alternatively, constraining the three-dimensional surface representation relative to one or more anchor portions can facilitate shaping the three-dimensional surface representation independently of other parameters used to form the three-dimensional surface representation (e.g., tightness of a surface mesh).
  • In some implementations, one or more anchor portions can be based on feedback from one or more sensors on a medical device within an anatomic structure. In certain implementations, the one or more anchor portions can be based on input from the physician. Thus, for example, the one or more anchor portions can be based on a useful combination of physician input and feedback from one or more sensors of a medical device positioned within the anatomic structure. Such a combination can be useful for providing insights into the shape of the anatomic structure, while providing the physician with the ability to verify and, if necessary, override feedback from the one or more sensors.
  • Other aspects, features, and advantages will be apparent from the description and drawings, and from the claims.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a schematic representation of a system during a medical procedure.
  • FIG. 2 is a perspective view of an exemplary medical device of the system of FIG. 1.
  • FIG. 3 is a schematic representation of a tip section of the medical device of FIG. 2 shown in an anatomic structure.
  • FIG. 4 is a schematic depiction of a projection of a three-dimensional data structure and a three-dimensional surface representation of the anatomic structure projected to a graphical user interface of the system of FIG. 1.
  • FIG. 5A is a schematic representation of a surface of the anatomic structure having superimposed thereon a point cloud, corresponding to known locations of a medical device in an anatomic structure, and a volumetrically smoothed three-dimensional surface representation with a high degree of volumetric smoothing.
  • FIG. 5B is a schematic representation of a surface of the anatomic structure having superimposed thereon the point cloud of FIG. 5A and a three-dimensional surface representation with a low degree of volumetric smoothing.
  • FIG. 5C is a schematic representation of a surface of the anatomic structure having superimposed thereon the point cloud of FIG. 5A and a three-dimensional surface representation constrained relative to one or more anchor portions.
  • FIG. 6A is a schematic depiction of a three-dimensional data structure of FIG. 4 and an unconstrained three-dimensional surface representation displayed on the graphical user interface of the system of FIG. 1.
  • FIG. 6B is a schematic depiction of the three-dimensional data structure and the unconstrained three-dimensional surface representation of FIG. 6A displayed on the graphical user interface of the system of FIG. 1 with an anchor portion identified on the three-dimensional data structure.
  • FIG. 6C is a schematic depiction of the three-dimensional data structure of FIG. 6A and a three-dimensional surface representation constrained relative to the anchor portion of FIG. 6B, the three-dimensional data structure and the three-dimensional surface representation displayed on the graphical user interface of the system of FIG. 1, with the three-dimensional surface representation surface smoothed.
  • FIG. 6D is a schematic depiction of the three-dimensional surface representation of FIG. 6C displayed on the graphical user interface of the system of FIG. 1.
  • FIG. 7 is a flowchart of an exemplary method of representing a surface of an anatomic structure.
  • FIG. 8 is a flowchart of an exemplary method of representing a surface of a heart cavity of a patient.
  • Like reference symbols in the various drawings indicate like elements.
  • DETAILED DESCRIPTION
  • The present disclosure is generally directed to devices, systems, and methods of generating a three-dimensional surface representation of an anatomic structure of a patient. More specifically, the three-dimensional surface representation can accurately represent local anatomic features of the anatomic structure, while being based on an incomplete or uncertain data set, by constraining (e.g., pinching) the three-dimensional surface representation relative to the one or more anchor portions. For at least this reason, the three-dimensional surface representations generated according to the devices, systems and methods of the present disclosure can be generated efficiently and, in use, can be useful for facilitating visualization of a position of a medical device (e.g., a catheter) during a medical procedure (e.g., diagnosis and/or treatment) being performed on the anatomic structure.
  • It should be appreciated that, unless otherwise specified or made clear from the context, the systems and methods of the present disclosure can be used for any of various different medical procedures, such as procedures performed on a hollow anatomic structure of a patient, and, more specifically, in a hollow anatomic structure, in which direct visual access to the medical procedure is impractical and/or is improved by the use of a model of the anatomic structure. Thus, for example, the systems and methods of the present disclosure can be used to facilitate visualization of a catheter inserted into a heart cavity as part of a medical treatment associated with diagnosis, treatment, or both of a cardiac condition (e.g., cardiac arrhythmia). Additionally, or alternatively, the systems and methods of the present disclosure can be used in one or more medical procedures associated with interventional pulmonology, brain surgery, or sinus surgery (e.g., sinuplasty).
  • As used herein, the term “physician” shall be understood to include any type of medical personnel who may be performing or assisting a medical procedure and, thus, is inclusive of a doctor, a nurse, a medical technician, other similar personnel, and any combination thereof. Additionally, or alternatively, as used herein, the term “medical procedure” shall be understood to include any manner and form of diagnosis, treatment, or both, inclusive of any preparation activities associated with such diagnosis, treatment, or both. Thus, for example, the term “medical procedure” shall be understood to be inclusive of any manner and form of movement or positioning of a medical device in an anatomic chamber.
  • As used herein, the term “patient” should be considered to include any mammal, including a human, upon which a medical procedure is being performed.
  • FIG. 1 is a schematic representation of a system 100 during a medical procedure performed in an anatomic structure of a patient 102. The system 100 can include a medical device 104 connected, via an extension cable 106, to an interface unit 108. The interface unit 108 can include a processing unit 109 (e.g., one or more processors), a graphical user interface 110, and a storage medium 111. The graphical user interface 110 and the storage medium 111 can be in electrical communication (e.g., wired communication, wireless communication, or both) with the processing unit 109.
  • In use, the medical device 104 can be moved within the anatomic structure (e.g., as part of a medical procedure) such that the processing unit 109 can receive a plurality of location signals of the medical device 104 in the anatomic structure. As described in greater detail below, the processing unit 109 can construct a three-dimensional surface representation of the anatomic structure based on a three-dimensional data structure representing locations, within the anatomic structure, visited by the medical device 104. To the extent the medical device 104 has not visited each location within the anatomic structure, a corresponding three-dimensional data structure can be an incomplete or uncertain data set. To account for such an incomplete or uncertain data set, it can be useful to volumetrically smooth the three-dimensional surface representation generated based on the three-dimensional data structure. As a result of such volumetric smoothing, however, certain portions of the three-dimensional surface representation may not pass close to the visited locations of the medical device 104 along some areas of the three-dimensional data structure. To account for such unintended distortions of the three-dimensional surface representation, as also described in greater detail below, the processing unit 109 can receive one or more inputs corresponding to one or more anchor portions for advantageously constraining a three-dimensional surface representation of the anatomic structure. For example, without modifying other parameters of a surface mesh, the three-dimensional surface representation can be constrained to include details of the anatomic structure that would not otherwise be represented in a three-dimensional surface representation based on catheter location alone.
  • In general, the three-dimensional surface representation of the anatomic structure formed using system 100 according to any one or more of the methods described herein can be shown on the graphical user interface 110, and the three-dimensional surface representation can be used to facilitate performance of a medical procedure by a physician. For example, as described in greater detail below, the three-dimensional surface representation of the anatomic structure and the position of the medical device 110 can be shown on the graphical user interface 110 and used as a visual guidance tool (e.g., as an analog) for movement of the medical device 104 in the anatomic structure. It should be appreciated, therefore, that the details provided in the three-dimensional surface representation generated as described herein can facilitate fine movement of the medical device 104 relative to the anatomic structure. As an example, as compared to a three-dimensional surface representation based on catheter position alone, the three-dimensional surface representation generated according to any one or more of the methods described herein can more accurately represent anatomic features or landmarks that are useful for positioning the medical device 104 relative to targeted tissue. Further, or in the alternative, as compared to a three-dimensional surface representation constructed solely from interpolation or approximation between points in a data set of known positions of the medical device 104 in an anatomic structure, the three-dimensional surface representation generated according to any one or more of the methods described herein is less likely to be unintentionally distorted in areas in which there are significant spatial gaps in position data of the medical device 104.
  • Referring now to FIGS. 1-2, the medical device 104 can be any of various different medical devices known in the art for use with respect to an anatomic structure and includes, therefore, any manner and form of medical devices useful for diagnosis, treatment, and combinations thereof. For the sake of explanation, and not by way of limitation, the medical device 104 is described herein as a catheter insertable into an anatomic structure. Thus, the medical device 104 can include a handle 120, a shaft 122, and a tip section 124. The shaft 122 can include a proximal portion 126 secured to the handle 120, and a distal portion 128 coupled to the tip section 124.
  • The tip section 124 generally includes any portion of the medical device 104 that directly or indirectly engages tissue for the purpose of treatment, diagnosis, or both and, therefore, can include any one or more of all manner and type of contact and/or non-contact interaction with tissue known in the art. For example, the tip section 124 can include one or more of contact and non-contact interaction with tissue in the form of energy interaction (e.g., electrical energy, ultrasound energy, light energy, cooling and any combinations thereof), chemical interaction with tissue, or both. Thus, for example, the tip section 124 can deliver energy (e.g., electrical energy) to tissue in the anatomic structure as part of any number of medical procedures.
  • In certain implementations, it is desirable to deliver energy (e.g., RF energy) from the tip section 124 to targeted portions of tissue in the anatomic structure to ablate tissue at some depth relative to a surface of the anatomic structure. In implementations in which the anatomic structure is a heart cavity, such ablations created by the tip section 124 along a surface of the anatomic structure can, for example, treat cardiac arrhythmia in patients with this condition. The effectiveness of the ablations created using the tip section 124 in such a cardiac ablation procedure, however, can be dependent upon the location of the ablations. It should be appreciated, therefore, that accurate representation of anatomic features or landmarks in the three-dimensional surface representation used to guide placement of the catheter can be advantageous for accurately delivering such targeted ablation energy to tissue in cardiac ablation procedures or other similar procedures in which there is a benefit derived from targeted energy delivery.
  • The medical device 104 can include a sensor 125 disposed, for example, along the tip section 124 and in electrical communication with the interface unit 108 (e.g., in communication with the processing unit 109). The sensor 125 can be any of various different types of sensors suitable for sensing contact with tissue of an anatomic structure and, therefore, can be useful for providing feedback to the interface unit 108 regarding the location of a blood-tissue boundary. In general, it should be understood that any one or more forms of feedback provided by the sensor 125 can form the basis for generating anchor portions for constraining a three-dimensional surface representation of the anatomic structure. Examples of these forms of feedback provided by the sensor 125 and useful as the basis for generating anchor portions are described below. While the sensor 125 is described herein as a single sensor for the sake of clarity of explanation, the sensor 125 can include an array of any one or more of the sensors described herein, including, for example, any combination of the sensors described herein.
  • As an example, the sensor 125 can include a sensing electrode such that changes to an electrical signal measured between the sensor 125 and another electrode (e.g., another sensor carried on the tip section 124) can be detected as an indication of the presence of viable tissue in contact with the sensor 125. As used herein, viable tissue is tissue that conducts an electrical signal and, thus, includes tissue that has not yet been ablated (e.g., is not scar tissue) as well as tissue that is not otherwise diseased such that conduction of the electrical signal is impaired. The detection of viable tissue in contact with the sensor 125 can include observation of the electrical signal by the physician. Also, or in the alternative, the detection of viable tissue in contact with the sensor can be based on a comparison of the electrical signal relative to a predetermined threshold (e.g., for a bipolar electrogram, a threshold of above about 0.1 mV). More generally, any of the various different devices, systems, and methods described herein can be advantageously used in combination with detecting changes to an electrical signal measured between the sensor 125 and another electrode to detect or confirm contact with tissue.
  • Additionally, or alternatively, the sensor 125 can include a force sensor to detect a magnitude and, optionally or additionally, a direction of force exerted on the sensor 125 through contact with a surface of the anatomic structure. Such a force sensor can include any of various different force sensors responsive to contact between the sensor 125 and tissue of the anatomic structure. Such responsiveness can be independent, or at least substantially independent, of whether or not the contacted tissue is viable. Thus, for example, the sensor 125 can be a force sensor including optical fibers, transmitting or sensing coils, and the like, for sensing force. Contact between the sensor 125 and tissue of the anatomic structure can result in an increase in measured force. For example, a contact force greater than 5 g (e.g. greater than 10 g) can be indicative of contact between the sensor 125 and tissue. The measured force can, in addition or in the alternative, be related to the degree of contact between the sensor 125 and the tissue of the anatomic structure. Additionally, or alternatively, contact between the sensor 125 and tissue of the anatomic structure can result in a measured force in a direction normal to the tissue at a point of contact.
  • As an additional or alternative example, the sensor 125 can include an ultrasound sensor such that the sensor 125 can detect features of an anatomic structure based on any of various different ultrasound techniques that are known in the art. As a specific example, the sensor 125 can include an ultrasound transducer such that ultrasound reflections can be measured with respect to distance along an axis of the sensor 125. Continuing with this example, contact or proximity between the sensor 125 and tissue in the anatomic structure can result in ultrasound reflections at distances corresponding a distance between the sensor 125 and the tissue.
  • As yet another additional or alternative example, the sensor 125 can include a deformation sensor to detect deformation (e.g., magnitude, direction, or both) of the tip section 124 as a result of contact between the tip section 124 and a surface of the anatomic structure. For example, the measured deformation can be a substantially monotonic function of the degree of contact between the sensor 125 and the tissue of the anatomic structure. Additionally, or alternatively, contact between the sensor 125 and tissue of the anatomic structure can result in deformation primarily in a direction normal to the tissue at the point of contact.
  • As yet another additional or alternative example, the sensor 125 can include an impedance sensor to detect a change in an electrical impedance as a result of contact between the tip section 124 and tissue of the anatomic structure. For example, in some implementations, contact between the sensor 125 and tissue in the anatomic structure can be detected as an increase in a measured impedance. Continuing with this example, an increase in a measured impedance larger than the expected variation in the impedance when the sensor 125 is not in contact with tissue (e.g. an increase greater than 100 ohms) can be indicative of contact between the sensor 125 and tissue in the anatomic structure. Additionally, or alternatively, the measured impedance can be a substantially monotonic function of the degree of contact between the sensor 125 and the tissue.
  • In addition to, or instead of, feedback provided by the sensor 125, contact between the tip section 124 and tissue of the anatomic structure can be based on one or more imaging modalities. The use of one such imaging modality can include observation of one or both of the tip section 124 and the shaft 122 by the physician using fluoroscopy. An additional, or alternative, modality can include observation of one or both of the tip section 124 and the shaft by the physician using intracardiac ultrasound in implementations in which the anatomic structure is a heart cavity. In some instances, based on information determined according to any one or more imaging modality, the physician can tag the location of contact with tissue, and the tag can form a basis of the one or more anchor portions near which the three-dimensional surface representation is constrained to pass. Additionally, or alternatively, information determined automatically from an image can provide an indication of contact between the tip section 124 and tissue of the anatomic structure.
  • While contact with tissue that forms a basis for the anchor portions can be based on feedback provided by sensors 125, it should be appreciated that anchor portions can be additionally, or alternatively, based on other types of feedback. For example, anchor portions can be placed (e.g., through tags applied by the physician) in locations in which a physician detects a resistance to movement (e.g., rotation, articulation, advancement, or a combination thereof), with resistance being indicative of contact between the medical device 104 and the tissue.
  • The medical device 104 can further, or instead, include a magnetic position sensor 130 along the distal portion 128 of the shaft 122. It should be appreciated that the magnetic position sensor 130 can be any of various magnetic position sensors well known in the art and can be positioned at any point along the distal portion 128. The magnetic position sensor 130 can, for example, include one or more coils that detect signals emanating from magnetic field generators. One or more coils for determining position with five or six degrees of freedom can be used.
  • The magnetic field detected by the magnetic position sensor 130 can be used to determine the position of the distal portion 128 of the catheter shaft 122 according to one or more methods commonly known in the art such as, for example, methods based on using a sensor, such as the magnetic position sensor 130, to sense magnetic fields indicative of the position of the magnetic position sensor 130 and using a look-up table to determine location of the magnetic position sensor 130. Accordingly, because the tip section 124 is coupled to the distal portion 128 of the shaft 122 in a known, fixed relationship to the magnetic position sensor 130, the magnetic position sensor 130 also provides the location of the tip section 124. While the location of the tip section 124 is described as being determined based on magnetic position sensing, other position sensing methods can additionally or alternatively be used. For example, the location of the tip section 124 can be additionally, or alternatively, based on impedance, ultrasound, and/or imaging (e.g., real time MRI or fluoroscopy).
  • Referring now to FIGS. 1-4, the tip section 124 of the medical device 104 can be moved in an anatomic structure 132 (e.g., prior to application of an ablation treatment or other type of treatment). If the tip section 124 of the medical device 104 is movable in blood in the anatomic structure 132 and obstructed only by a surface 133 of the anatomic cavity 132, the known positions of the tip section 124 of the medical device 104 can be taken together to provide an indication of the size and shape of a volume defined by the surface 133 of the anatomic structure 132 and can form a basis for a three-dimensional data structure 134 corresponding to a volume defined by the anatomic structure 132. The three-dimensional data structure 134 can include any of various different data structures known in the art. As an example, the three-dimensional data structure 134 can include an occupancy grid. As another example, the three-dimensional data structure 134 can include an indicator function. Additionally, or alternatively, the three-dimensional data structure 134 can include a segmentation mask. Further, or instead, the three-dimensional data structure 134 can include space carving.
  • Because it is often difficult or impractical to pass the tip section 124 through each portion of the volume defined by the surface 133 of the anatomic structure 132, the three-dimensional data structure 134 may be an incomplete or uncertain data set. As described in greater detail below, forming the three-dimensional data structure 134 based on such an incomplete or uncertain data set can have implications for accurately representing anatomic features of the surface 133 of the anatomic structure 132. For example, the three-dimensional data structure 134 can include a probabilistic model as a function of location, and an incomplete or uncertain data set of locations can result in regions of high uncertainty in the probabilistic model. Also, or instead, the three-dimensional data structure 134 can include a model with two or more discrete states (e.g., a “blood” state, an “unknown” state, a “tissue” state, a “boundary” state, and combinations thereof) as a function of location, and an incomplete or uncertain data set of locations can result in uncertain states (e.g. “unknown” states).
  • While the three-dimensional data structure 134 can be based on known positions of the tip section 124 of the medical device 104 in the anatomic structure 132, it should be appreciated that other methods of determining the three-dimensional data structure 134 are additionally or alternatively possible. For example, the three-dimensional data structure 134 can be based on images of the surface 133 of the anatomic structure 132 acquired prior to or during the medical procedure, particularly in use cases in which the images of the surface 133 of the anatomic structure 132 may not be complete. Such images can correspond to any of various, different imaging modalities including, for example, x-ray.
  • The three-dimensional data structure 134 can include, for example, a three-dimensional grid of voxels 135. In general, it should be appreciated that the three-dimensional grid of voxels 135 can be any one or more of various different types three-dimensional grids well known in the art. By way of example, the three-dimensional grid of voxels 135 can include one or more of an occupancy grid and an occupancy field. By way of further, non-exclusive example, the three-dimensional grid of voxels 135 can include a volumetric grid representation.
  • Each voxel 135 can be a discrete element of volume. Together, the voxels 135 can form the three-dimensional data structure 134 which, in general, is a three-dimensional notational space. As described in greater detail below, a three-dimensional surface representation 136 can be formed in relation to the three-dimensional data structure 134 such that the three-dimensional surface representation 136 can represent the surface 133 of the anatomic structure 132. In general, the three-dimensional surface representation 136 can be any one or more of the various different types well-known in the art and, thus, by way of non-exclusive example can include any one or more of the following: a “level set”; a “separating surface”; and an “implicit surface”.
  • In certain medical procedures, it can be impractical (e.g., due to time constraints) or impossible (e.g., due to shape) to visit each location of the anatomic structure 132 with the tip section 124. Accordingly, the three-dimensional data structure 134 and/or the three-dimensional surface representation 136 can be necessarily based on certain inferences between data points. These inferences, while serving as a useful expedient for generation of the three-dimensional surface representation 136, can result in discrepancies between the three-dimensional surface representation 136 and the surface 133 of the anatomic structure 132 upon which the three-dimensional surface representation 136 is based. Such discrepancies can, for example, result in one or more anatomic features of the surface 133 of the anatomic structure 132 being obscured, or at least distorted, in the resulting three-dimensional surface representation 136.
  • The inferences used to form the three-dimensional surface representation 136 can include any manner and form of volumetric smoothing known in the art. For example, volumetric smoothing the three-dimensional surface representation 136 can be based on surface tension methods. As an additional or alternative example, volumetric smoothing the three-dimensional surface representation 136 can be based on hole filling methods. As a further or alternative example, volumetric smoothing the three-dimensional surface representation 136 can be based on interpolation. As still a further or alternative example, volumetric smoothing of the three-dimensional surface representation 136 can be based on ball-pivoting.
  • In general, discrepancies between the three-dimensional surface representation 136 and the surface 133 of the anatomic structure 132 can arise as a result of a trade-off between the degree of volumetric smoothing used to form the three-dimensional surface representation 136 and incompleteness or uncertainty of the data set forming the three-dimensional data structure 134. That is, it can be desirable to specify a low degree of volumetric smoothing to achieve resolution of certain anatomic features in the three-dimensional surface representation 136. Such resolution in one area, however, can have the unintended consequence of creating distortions in areas in which the data set is incomplete or uncertain, with such incompleteness or uncertainty being common in implementations in which the data set is based on locations in the anatomic structure 132 visited by the tip section 124. For example, the three-dimensional surface representation 136 can appear to include numerous distortions (e.g., undulations or, more specifically, invaginations) that are not representative of the surface 133 of the anatomic structure 132 when a low degree of volumetric smoothing is applied to an incomplete or uncertain data set. However, increasing the amount of volumetric smoothing of the three-dimensional surface representation 136 to remove such distortions resulting from incompleteness or uncertainty of the data set can have the unintended consequence of obscuring or distorting certain anatomic features.
  • FIGS. 5A-5C are, collectively, a schematic depiction of a generalized example of distortions that can occur as a result of volumetric smoothing a three-dimensional surface representation based solely on surface tension, hole filling, interpolation, ball pivoting, or other similar implicit or explicit surface reconstruction methods applied to an incomplete or uncertain data set. In each of FIGS. 5A-5C, a point cloud 139 is shown superimposed on the surface 133 of the anatomic structure 132 (FIG. 3). The point cloud 139 can correspond to, for example, known locations of the tip section 124 of the medical device 104 in the anatomic structure 132 (FIG. 3). As used, herein, the point cloud 139 should be understood to include any of various different point clouds well known in the art of surface reconstruction. In general, it is desirable to generate a volumetrically smoothed surface mesh from the point cloud 139 to create a three-dimensional surface representation that accurately represents the surface 133 of the anatomic structure 132 (FIG. 3). However, as described in greater detail below, distortions can occur when the point cloud 139 is an incomplete or uncertain data set of an anatomic structure and it is, nevertheless, desirable to represent accurately a geometric feature of the surface 133.
  • FIG. 5A is a schematic representation of a three-dimensional surface representation 136 a of the point cloud 139, with the three-dimensional surface representation 136 a volumetrically smoothed to a high degree. As shown in FIG. 5A, a high degree of volumetric smoothing of the three-dimensional surface representation 136 a can result in significant deviations between the three-dimensional surface representation 136 a and the surface 133 (i.e., the intended shape). For example, with a high degree of volumetric smoothing, the three-dimensional surface representation 136 a can be based on an interpolation between data points of the point cloud 139 that are not close to one another, and details of the shape between the data points may be lost.
  • FIG. 5B is a schematic representation of a three-dimensional surface representation 136 b of the point cloud 139, with the three-dimensional surface representation 136 b smoothed to a low degree of volumetric smoothing. Accordingly, the three-dimensional surface representation 136 b is based on interpolation between data points of the point cloud 139 that are close to one another, as compared to the interpolation associated with FIG. 5A. As shown in FIG. 5B, the interpolation between data points in the point cloud 139 that are relatively close to one another can result in a shape that shows features that are not actually present on the surface 133.
  • In general, it should be appreciated that inaccuracies, such as those shown in FIG. 5A and FIG. 5B, can be particularly prevalent in areas in which the surface 133 is represented by relatively few data points of the point cloud 139. However, it is often impractical or impossible to obtain a complete data set upon which the three-dimensional surface representation 136 a or 136 b can be based. For example, the time associated with obtaining such a complete data set may be impermissibly long. As described in greater detail below, a useful solution to this trade-off between accuracy and challenges associated with data acquisition is to constrain volumetric smoothing in areas in which the shape of the surface 133 is known or known with some level of confidence.
  • Referring now to FIG. 5C, one or more anchor portions 142 can be added to the data set to identify one or more known positions on the surface 133. As described in greater detail below, the one or more anchor portions 142 can constrain the volumetric smoothing of the three-dimensional surface representation 136 c. Because the three-dimensional surface representation 136 c is constrained relative to the one or more anchor portions 142, distortions or other types of inaccuracies associated with a high degree of volumetric smoothing (FIG. 5A) and a low degree of volumetric smoothing (FIG. 5B) are less prevalent in the three-dimensional surface representation 136 c (FIG. 5C). More generally, the three-dimensional surface representation 136 c is a more accurate representation of the surface 133 than would be obtained without some form of constraint relative to the one or more anchor portions 142, which represent corresponding positions known, optionally with some degree of confidence, to lie on the surface 133.
  • Referring again to FIGS. 1-4, the inaccuracies in the three-dimensional surface representation 136 that can result from volumetric smoothing can be particularly prevalent in locally concave areas of the surface 133 along a generally convex portion of the surface 133 of the anatomic structure 132. According to the convention used herein, the concavity of the surface 133 of the anatomic structure 132 is expressed with respect to the shape of a blood-tissue boundary formed by the surface 133 of the anatomic structure 132 (FIG. 3) around the blood. For example, as shown in FIG. 3, the anatomic structure 132 can include carina 137 a, 137 b, 137 c between anatomic elements 131 a, 131 b, 131 c, 131 d, 131 e of the anatomic structure 132. The surface 133 should be understood to be locally concave along each carina 137 a, 137 b, 137 c. In such instances, volumetric smoothing the three-dimensional surface representation 136 can result in an inaccurate depiction of the surface 133 of the anatomic structure 132 in the region of the carina 137 a, 137 b, 137 c. Such an inaccurate depiction in the three-dimensional surface representation 136 in the vicinity of one or more of the carina 137 a, 137 b, 137 c can be problematic, for example, in instances in which it may be desirable to apply lesions, and thus position accurately the tip section 124, in the vicinity of one or more of the carina 137 a, 137 b, 137 c. For example, in FIG. 3, in instances in which the anatomic structure 132 is the left atrium, it may be desirable to apply lesions in the carina 137 c between the left superior pulmonary vein (LSPV) and the left inferior pulmonary vein (LIPV), represented as 131 d and 131 e, respectively. It should be appreciated that the anatomic structure 132 is depicted as the left atrium in FIG. 3 by way of example and not limitation, and, thus, it should be appreciated that the number and orientation of anatomic elements and corresponding carina (or other similar locally concavities) present in the anatomic structure 132 can depend on the type of anatomic structure.
  • To overcome the trade-off that can exist between global and local resolution of the three-dimensional surface representation 136 as a result of incomplete or uncertain data regarding the surface 133 of the anatomic structure 132, the present disclosure is generally directed to constraining the three-dimensional surface representation 136 according to one or more of the methods described herein. For example, constraining the three-dimensional surface representation 136 as described herein can facilitate accurate representation of anatomic features of the surface 133 of the anatomic structure 132 while allowing the three-dimensional surface representation 136 to be generated efficiently (e.g., based on an incomplete or uncertain data set of locations in the anatomic structure 132).
  • The three-dimensional data structure 134 and the three-dimensional surface representation 136 can be stored on the storage medium 111, along with instructions executable by the processing unit 109 to display the three-dimensional data structure 134, the three-dimensional surface representation 136, or both on the graphical user interface 110, as described in greater detail below. The instructions stored on the storage medium 111 and executable by the processing unit 109 to display one or both of the three-dimensional data structure 134 and the three-dimensional surface representation can be, for example, an application built using Visualization Toolkit, an open-source 3D computer graphics toolkit, available at www.vtk.org.
  • FIG. 4 is a schematic representation of the three-dimensional data structure 134 and the three-dimensional surface representation 136 projected onto a viewing window 138 of an image plane 140. While the three-dimensional data structure 134 and the three-dimensional surface representation 136 are described as both being projected onto the viewing window 138, it should be understood that the three-dimensional data structure 134 and the three-dimensional surface representation 136 can be individually projected to the viewing window 138. For example, it may be desirable to project both the three-dimensional data structure 134 and the three-dimensional surface representation 136 onto the viewing window 138 while the three-dimensional surface representation 136 is being built. Additionally, or alternatively, it may be desirable to project only the three-dimensional surface representation 136 (e.g., by making the three-dimensional data structure 134 at least partially translucent) onto the viewing window 138 while the medical device 104 (FIG. 3) is being used to diagnose and/or apply a treatment to the anatomic structure 132 (FIG. 3).
  • The graphical user interface 110 can be two-dimensional such that the image plane 140 corresponds to a plane of the two-dimensional display of the graphical user interface 110, and the viewing window 138 corresponds to a field of view of the two-dimensional display of the graphical user interface 110. Accordingly, the image formed by projecting one or both of the three-dimensional data structure 134 and the three-dimensional surface representation 136 onto the viewing window 138 can be displayed on the graphical user interface 110. As described in greater detail below, a physician can, in certain instances, interact with the projection of the three-dimensional data structure 134 on the graphical user interface 110 to identify one or more locations of anatomic features on the three-dimensional data structure 134 such that the three-dimensional surface representation 136, generated based on the three-dimensional data structure 134, reflects these identified anatomic features.
  • FIGS. 6A-6D collectively represent a sequence of images formed on the viewing window 138 and displayed on the graphical user interface 110. The sequence illustrates, in general, the generation of the three-dimensional surface representation 136.
  • Referring now to FIG. 6A, the three-dimensional data structure 134 and an unconstrained three-dimensional surface representation 136′ can be displayed on the graphical user interface 110. In FIG. 6A, the tricuspid valve, the inferior vena cava, and the superior vena cava have each been cut in the unconstrained three-dimensional surface representation 136′ to facilitate visualization. These cuts do not constrain the surface extracted from the three-dimensional data structure 134.
  • The unconstrained three-dimensional surface representation 136′ can represent a surface extracted from the three-dimensional data structure 134 without specific input regarding the position of anatomic features in the three-dimensional data structure 134. Thus, in certain instances, anatomic features of the surface 133 of the anatomic structure 132 (FIG. 3) may be obscured, distorted, or otherwise misrepresented in the unconstrained three-dimensional surface representation 136′. For example, the unconstrained three-dimensional surface representation 136′ may not accurately represent anatomic features of the surface 133 of the anatomic structure 132 (FIG. 3) if the three-dimensional data structure 134 is based on an incomplete or uncertain data set. That is, attempts to form a surface mesh based on the three-dimensional data structure 134 can result in an obscured and/or distorted representation of the surface 133 of the anatomic structure 132 (FIG. 3), particularly along locally concave portions of the surface 133 of the anatomic structure 132 (FIG. 3).
  • The volumetric smoothing required to create the three-dimensional surface representation 136′ based on an incomplete or uncertain data set can obscure or distort areas that are concave because the difference between missing data and an actual concave region is often not discernible by a smoothing algorithm, such as a surface tension algorithm, a hole filling algorithm, an interpolation algorithm, a ball-pivoting algorithm, or other similar algorithms. As a result, volumetric smoothing required to form the three-dimensional surface representation 136′ based on an incomplete or uncertain data set can have the unintended consequence of covering over concave regions. Conversely, while it may be possible to capture aspects of local detail by decreasing the degree of volumetric smoothing used to form the three-dimensional surface representation 136′, a low degree of volumetric smoothing can create distortions along other portions of the three-dimensional surface representation 136′, such as where the data set is incomplete or uncertain. Accordingly, while the degree of volumetric smoothing can be adjusted to capture aspects of local detail, such adjustments can have a negative impact on the accuracy of the overall shape of the three-dimensional surface representation 136′.
  • Referring now to FIGS. 2,3, and FIG. 6B, an advantageous alternative to varying the degree of volumetric smoothing of the surface mesh forming the three-dimensional surface representation 136′ can include receiving identification of the one or more anchor portions 142 on the three-dimensional data structure 134. Each anchor portion 142 can correspond to a predetermined number of voxels 135 of the three-dimensional data structure 134. Additionally, or alternatively, each anchor portion 142 can correspond to one or more nodes or points in the three-dimensional data structure 134. In certain implementations, each anchor portion 142 can include information regarding, for example, an orientation of the surface 133 of the anatomic structure 132. Additionally, or alternatively, each anchor portion 142 can include information regarding, for example, a degree of confidence in the location of the anchor portions 142 and/or the associated information.
  • Identification of each anchor portion 142 can be based on input received from a physician (e.g., as a tag), input received from the tip section 124 in the anatomic structure 132, or a combination thereof. More generally, it should be appreciated that the identification of each anchor portion 142 can be based on observations made or prior knowledge regarding the anatomic structure 132 and can be independent of parameters used to form the three-dimensional surface representation 136′.
  • Each anchor portion 142 can be represented on the graphical user interface 110 as visual indicia. Such visual indicia can be useful, for example, as a visualization tool for the physician to assess how the three-dimensional surface representation 136′ will be modified as it is constrained to pass near a position relative to the anchor portion 142. For example, based on observation of the visual indicia representing the anchor portion 142 on the graphical user interface 110, the physician can reposition the anchor portion 142.
  • The one or more anchor portions 142 can represent a position of an anatomic feature of the anatomic structure 132. For example, the one or more anchor portions 142 can correspond to locations at which contact between the tip section 124 and the surface 133 of the anatomic structure 132 is detected such that the one or more anchor portions 142 represent a location known, optionally with some confidence, to lie on the surface 133 of the anatomic structure 132. In some instances, the one or more anchor portions 142 can include information regarding, for example, a direction of contact corresponding to an orientation of the surface 133 of the anatomic structure 132. Contact detection forming the basis of the one or more anchor portions 142 can be based on any of various different forms and combinations of contact detection described herein or otherwise known in the art. For example, in cardiac applications, contact detection can be based on an intracardiac electrogram indicative of contact between the sensor 125 of the tip section 124 and the surface 133. As another, non-exclusive example, contact detection can also, or instead, be based on a force signal (e.g., including magnitude, direction, or both) in implementations in which the sensor 125 is a force sensor. As yet another, non-exclusive example, contact detection can also, or instead, be based on detecting deformation (e.g., using the sensor 125) of at least a portion of the tip section 124 as the tip section 124 contacts the surface 133. As still another, non-exclusive example, contact detection can also, or instead, be based on detecting a change in impedance (e.g. using the sensor 125).
  • Additionally, or alternatively, the one or more anchor portions 142 can be based on input from a physician. For example, the input from the physician can be based on the unconstrained three-dimensional surface representation 136′ and the physician's knowledge of anatomy. That is, the global shape of the unconstrained three-dimensional surface representation 136′ may represent the overall shape of surface 133 of the anatomic structure 132, albeit with local inaccuracies with respect to some anatomic features. In such instances, the physician can provide an indication of one or more anchor portions 142 on the three-dimensional data structure 134 and/or on the unconstrained three-dimensional surface representation 136′ based on the physician's knowledge of anatomy (e.g., knowledge of the position of a carina in implementations directed to a heart cavity). As another or alternative example, the physician can provide an indication of one or more anchor portions 142 on the three-dimensional data structure 134 and/or on the unconstrained three-dimensional surface representation 136′ based on observation of the tip section 124 and/or the shaft 122. That is, in certain instances, the physician can push the tip section 124 gently and observe, optionally through fluoroscopy or other similar imaging modality, whether the tip section 124 (FIG. 3) advances in response to the push. If the tip section 124 (FIG. 3) does not advance in response to the push, the physician can manually tag the point as one of the one or more anchor portions 142.
  • The physician can identify the one or more anchor portions 142 on the three-dimensional data structure 134 by providing inputs to the interface unit 108 (FIG. 1) (e.g., through a keyboard, a mouse, or other input associated with the interface unit 108).
  • Referring now to FIG. 6C, the three-dimensional surface representation 136 can be generated as a mesh (e.g., a continuous polygonal mesh) of a surface extracted from the three-dimensional data structure 134 and constrained relative to the one or more anchor portions 142. As used herein, constraining the three-dimensional surface representation 136 relative to the one or more anchor portions 142 can include any one or more of various different methods that, as compared to the absence of a constraint, reduces the distance between the three-dimensional surface representation 136 and the one or more anchor portions 142. For example, constraining the three-dimensional surface representation 136 relative to the one or more anchor portions 142 can include constraining the three-dimensional surface representation 136 to pass through the one or more anchor portions 142. Additionally, or alternatively, constraining the three-dimensional surface representation 136 relative to the one or more anchor portions 142 can include constraining the three-dimensional surface representation 136 to be at or within a distance (e.g., a fixed distance) relative to the one or more anchor portions 142. As a further or alternative example, the three-dimensional surface representation 136 can be constrained by a penalty (e.g., cost) function that penalizes for distance from the one or more anchor portions 142 but does not strictly restrict the three-dimensional surface representation 136 to pass within a specified distance of the one or more anchor portions 142.
  • Referring to FIGS. 6A-6C, in certain implementations, constraining the three-dimensional surface representation 136 relative to the one or more anchor portions can include setting a fixed value of one or more nodes associated with the one or more anchor portions 142 in the three-dimensional data structure 134, determining a scalar function on at least a portion of the three-dimensional data structure 134, and applying an algorithm to extract an isosurface based in part on the scalar function. An indicator function is a well-known example of such a scalar function. An embedding function is another well-known example of such a scalar function. The isosurface extracted based at least in part on the scalar function can include any of various different isosurfaces known in the art and, thus, among various examples, can include an implicit surface.
  • As an example, constraining the three-dimensional surface representation 136 can include a clamped signed distance function based only on space carving information, as is well known in the art. Specifically, a first fixed value can be assigned to nodes, in the three-dimensional data structure 134, corresponding to locations within the anatomic structure 132 (FIG. 3) visited by the tip section 124 of the medical device 104 (FIG. 2). Because the tip section 124 of the medical device 104 (FIG. 2) can only move through blood, the nodes with the first fixed value in the three-dimensional data structure 134 imply the presence of blood at those positions. However, because the tip section 124 of the medical device 104 (FIG. 2) does not generally move through every location within the anatomic structure 132 (FIG. 3), the nodes that have not been set to the first fixed value can correspond to one of two physical states. That is, the nodes that have not been set to the first fixed value can correspond to i) locations in the anatomic structure 132 (FIG. 3) that have not been visited by the tip section of the medical device 104 (FIG. 2) or ii) locations that are outside of the volume of the anatomic structure 132 in which the tip section 124 of the medical device 104 (FIG. 2) is moving. In general, inaccuracies in the three-dimensional surface representation 136 (such as the inaccuracies shown in FIGS. 5A and 5B) can arise from mischaracterization of these two categories of nodes that have not been set to the first fixed value.
  • Continuing with this example, because the one or more anchor portions 142 correspond to one or more locations on the surface 133 of the anatomic structure 132 (FIG. 3), the one or more anchor portions 142 can provide a constraint that is useful for more accurately characterizing nodes that that have not been set to the first fixed value. In some implementations, the one or more anchor portions 142 can be set to a second fixed value, different from the first fixed value. It should be appreciated that the second fixed value can be one of two values of a signed clamped distance function, and the anchor portions 142 can correspond to surfaces acquired from a range scan. In this example, therefore, nodes corresponding to the first fixed value can correspond to the known locations of blood while nodes corresponding to the second fixed value can correspond to the known locations of the surface 133 (FIG. 3) and, thus, the blood-tissue boundary. The values of the remaining nodes in the three-dimensional data structure 134 (i.e., the nodes that have been assigned neither the first fixed value nor the second fixed value) are variable.
  • As is known in the art, a volumetric smoothing algorithm (e.g., a three-dimensional smoothing kernel) can be applied to the three-dimensional data structure 134, and the values of these variable nodes can take on values that are a function of neighboring nodes. For example, the values of the variable nodes can take on values that are a function of neighboring nodes according to a thermal diffusion model. It should be appreciated that, because the one or more anchor portions 142 have a fixed value in this example, the one or more anchor portions 142 can modify the values of neighboring variable nodes as the three-dimensional data structure 134 undergoes volumetric smoothing. For example, a spatial convolution can be applied one or more times to calculate new values for only the variable nodes.
  • The three-dimensional surface representation 136 can be based on the three-dimensional data structure 134 having fixed-value nodes as described above. For example, the three-dimensional surface representation 136 can correspond to an isosurface (e.g., a level set, an implicit surface, etc.) extracted from the three-dimensional data structure 134 having fixed-value nodes. In certain implementations, a “marching cubes” algorithm can be applied to the volumetrically smoothed three-dimensional data structure 134 having fixed-value nodes to extract an isosurface corresponding to a fixed value (e.g., a value between the first fixed value associated with known locations of the tip section 124 of the medical device 104 (FIG. 2) and the second fixed value associated with the one or more anchor portions 142). In this way, the one or more anchor portions 142 can have the effect of constraining, or otherwise limiting, the position of the resulting three-dimensional surface representation 136 extracted from the three-dimensional data structure 134. Additionally, or alternatively, a “flying edges” algorithm can be applied to the volumetrically smoothed three-dimensional data structure 134 to extract an isosurface.
  • Further, or instead, the three-dimensional surface representation 136 can be based on any of various different algorithms well known in the art for extracting a mesh of a surface from the three-dimensional data structure 134. Thus, for example, the three-dimensional surface representation 136 can be a polygonal mesh extracted from the three-dimensional data structure 134 based on a “marching cubes” algorithm and constrained relative to the one or more anchor points 142 according to any of the various different methods described herein. As used herein, a “marching cubes” algorithm can include any one or more algorithms in which a polygonal mesh of an isosurface is extracted from the three-dimensional data structure 134 based on analysis of node values in the three-dimensional data structure 134. More generally, the three-dimensional surface representation 136 can be extracted from the three-dimensional data structure 134 according to any one or more computational algorithms known in the art for volumetrically smoothing three-dimensional representations of objects including a “ball-pivoting” algorithm, a “power crust” algorithm, and other similar algorithms.
  • As an example, an adaptive ball-pivoting algorithm can constrain the three-dimensional surface representation 136 relative to the anchor portions 142. That is, away from the anchor portions 142, the ball forming the basis of the ball-pivoting algorithm can be a fixed global size that yields volumetric smoothing where specific information about anatomical features is not available. For example, the fixed global ball can have a diameter between about 10 mm and 30 mm (e.g. about 15 mm). Close to the anchor portions 142, the size of the ball can be reduced to facilitate passing the surface representation 136 closer to the anchor portions 142. For example, as compared to the fixed global size of the ball away from the anchor portions 142, the size of the ball can be reduced closer to the anchor portions 142. With such a reduced ball size, as compared to the fixed global size, the three-dimensional surface representation 136 can pass closer to the anchor portions 142. As a more specific example, the size of the ball at a surface location closest to a respective one of the anchor portions 142 can be chosen such that the three-dimensional surface representation 136 lies within a predetermined minimum distance to the respective anchor portion 142. In certain implementations, the size of the ball can vary between the reduced ball size and the global ball size as a function (e.g., a linear function) of distance from the ball to the one or more anchor portions 142 until the ball size reaches the global ball size, beyond which distance the global ball size can remain at the fixed global ball size.
  • As shown in FIG. 6C, the three-dimensional surface representation 136 can be volumetrically smoothed and, optionally, surface smoothed. For example, volumetric smoothing of the three-dimensional surface representation 136 can be accomplished using any one or more of various different volumetric smoothing techniques that are well known in the art and described herein. Surface smoothing can additionally, or alternatively, be accomplished using any one or more of various different surface smoothing techniques well known in the art. An example of such a surface smoothing technique is Laplacian smoothing and variations thereof. In certain instances, through surface smoothing, three-dimensional surface representation 136 may no longer satisfy the constraints previously applied as part of the volumetric smoothing process. For example, as shown in FIG. 6C, the surface-smoothed three-dimensional surface representation 136 may contain fewer than all of the locations visited by the medical device. Further, or instead, the distance between the three-dimensional surface representation 136 and the anchor portions 142 may change as the three-dimensional surface representation 136 is subjected to surface smoothing.
  • Comparing FIG. 6B to FIG. 6C, it should be appreciated that the three-dimensional surface representation 136 differs from the unconstrained three-dimensional surface representation 136′ near the one or more anchor portions 142. In particular, because the three-dimensional surface representation 136 is constrained relative to the one or more anchor portions 142, the three-dimensional surface representation 136 depicts anatomic features that are not readily apparent in the unconstrained three-dimensional surface representation 136′. Accordingly, it should be further appreciated that the one or more anchor portions 142 can facilitate efficiently generating an accurate representation of anatomic features of the anatomic structure 132 (FIG. 3).
  • Referring now to FIG. 6D, the display of the three-dimensional data structure 134 of FIG. 6C can be hidden (e.g., made translucent) such that the three-dimensional surface representation 136 can be displayed by itself. The display of the three-dimensional surface representation 136 by itself and, optionally, in smoothed form can be useful, for example, for facilitating perception by the physician.
  • The steps shown in FIGS. 6A-6D have been shown and described as occurring in sequence for the sake of clarity of explanation. It should be appreciated, however, that in addition to, or as an alternative, any one or more of the steps shown in FIGS. 6A-6D can be combined, performed in parallel, and/or varied in order.
  • The computer executable instructions stored on the storage medium 111 (FIG. 1) can cause the processing unit 109 (FIG. 1) to generate the three-dimensional surface representation 136 according to one or more of the following exemplary methods. Unless otherwise indicated or made clear from the context, each of the following exemplary methods can be implemented using the system 100 (FIG. 1) and/or one or more components thereof.
  • FIG. 7 is a flowchart of an exemplary method 160 of representing a surface of an anatomic structure. The exemplary method 160 can include receiving 162 a plurality of location signals of a medical device, forming 164 a three-dimensional data structure representing volumes, within the anatomic structure, occupied by the medical device at the locations corresponding to the plurality of location signals, receiving 166 one or more anchor portions representing locations relative to the anatomic structure, and generating 167 a three-dimensional surface representation of the anatomic structure. The generated 167 three-dimensional surface representation can be constrained relative to the one or more anchor portions and to contain at least a portion of the three-dimensional data structure. That is, in general, the one or more anchor portions can modify the behavior of an algorithm in a way that constrains the resulting three-dimensional surface representation relative to the anchor portions.
  • In general, receiving 162 the plurality of location signals of the medical device can include receiving one or more signals indicative of the location of the medical device according to any one or more of the methods described herein. The received 162 plurality of location signals can be, for example, a plurality of signals received from a single sensor over a period of time. For example, the received 162 plurality of location signals can be one or more signals from a magnetic position sensor (such as the magnetic position sensor 130 described above with respect to FIG. 2). Additionally, or alternatively, the received 162 plurality of location signals can be a plurality of signals received from multiple, different types of sensors that individually, or in combination, provide information regarding the location of the medical device in the anatomic structure. As an example, greater details of which are described below, receiving 162 the plurality of location signals of the medical device can include receiving one or more signals from a magnetic position sensor and from a sensor providing a signal indicative of a blood-tissue boundary of the anatomic structure.
  • As used herein, the received 162 plurality of location signals of the medical device generally correspond to locations visited by a tip section of the medical device (e.g., the tip section 124 of the medical device 104 of FIG. 2) in the anatomic structure. More generally, however, it should be understood that the plurality of location signals of the medical device can correspond to locations visited by any portion of the medical device that can be sensed or otherwise determined in the anatomic structure.
  • Forming 164 the three-dimensional data structure representing locations, within the anatomic structure, visited by the medical device can include forming a three-dimensional grid, with locations in the three-dimensional grid reflecting locations visited by the medical device in the anatomic structure. In such implementations, a boundary of the anatomic structure can be approximated through analysis of neighbor locations and/or node values of the three-dimensional grid such as, for example, through application of one or more of a “marching cubes” algorithm, a “ball-pivoting” algorithm, and a “power crust” algorithm, with the algorithm or algorithms extracting a surface from the three-dimensional grid. In some implementations, the one or more anchor portions (e.g., the anchor portions 142 in FIGS. 5A-5C) can be represented by respective constrained values (or combinations thereof) on the three-dimensional grid or, more generally, on any type of three-dimensional data.
  • In general, receiving 166 the one or more anchor portions representing locations relative to the anatomic structure can include receiving input from one or more sources. For example, receiving 166 the one or more anchor portions can be based on input received from a physician, input received from one or more sensors on the medical device, or a combination thereof. In instances in which receiving 166 the one or more anchor portions is based on a combination of input from the physician and from one or more sensors on the medical device, it can be advantageous to have a hierarchy of input such that, for example, the input from the physician can override the input from the one or more sensors on the medical device.
  • Receiving 166 the one or more anchor portions representing locations relative to the anatomic structure can include receiving, from one or more sensors disposed on the medical device, a signal indicative of contact between the medical device and tissue of the anatomic structure. Sensed contact (e.g., one or more of location, direction, force, consistency, and/or duration of contact) between the medical device and the surface of the anatomic structure can be indicative of a blood-tissue boundary of the anatomic structure of the patient. Accordingly, one or more anchor portions can be identified at the location of the sensed contact to ensure that the three-dimensional surface representation is constrained relative to the sensed contact, which is known to represent the blood-tissue boundary.
  • It should be appreciated that such a signal indicative of contact between the medical device and tissue of the anatomic chamber can include any one or more of the signals indicative of contact described herein. Thus, for example, the signal indicative of contact between the medical device and tissue of the anatomic chamber can include an impedance signal (e.g., a change in impedance) from one or more impedance sensors (e.g., the sensor 125 in FIG. 2) disposed on the medical device. Additional or alternative examples of signals indicative of contact between the medical device and the surface of the anatomic structure of the patient can include one or more of: a change in an electrical signal (e.g., electrogram or impedance) in one or more electrodes of the medical device; a force detected by a force sensor of the medical device; an ultrasound signal of an ultrasound sensor on the medical device; and a deformation of at least a portion of the medical device. As a more specific example, a signal indicative of contact between the medical device and the surface of the anatomic structure of the patient can include an amplitude derived from an electrogram detected by one or more electrodes of the medical device.
  • Receiving 166 the one or more anchor portions can include identification of a subset of the plurality of received location signals. Identification of the subset of the plurality of received location signals can, for example, include an input command from the physician identifying one or more portions of the three-dimensional data structure as corresponding one or more anchor portions. The input command can be received from any of various, different input devices such as a keyboard, a mouse, a touchscreen, etc. and, additionally, or alternatively, can include voice commands. Thus, in implementations in which the data structure includes a three-dimensional grid, the physician can provide input through one or more input devices to identify the subset as one or more voxels of the three-dimensional grid, as displayed on a graphical user interface.
  • In certain implementations, receiving 166 the one or more anchor portions can include receiving a respective confidence level associated with the one or more anchor portions. For example, a confidence level can increase substantially monotonically with a measured indication of contact (e.g. electrogram amplitude, impedance, force, deformation, and/or proximity). In such implementations, constraining the three-dimensional surface representation relative to the one or more anchor portions can be based on the respective confidence level associated with each of the one or more anchor portions. For example, the confidence levels can form a basis for certain of the one or more anchor portions acting as stronger or weaker anchor portions relative to other anchor portions. That is, an anchor portion corresponding to a higher confidence level can act as a stronger anchor as compared to an anchor portion corresponding to a weaker confidence level. Additionally, or alternatively, an anchor portion identified with contact in a known direction can constrain the normal direction of the resulting surface using any of various different techniques known in the art.
  • In general, generating 167 the three-dimensional surface representation of the anatomic structure can include any one or more of the methods described herein for forming a three-dimensional surface. Thus, for example, generating 167 the three-dimensional surface representation can include extracting a surface from the three-dimensional data structure according to an algorithm, such as one or more of a “marching cubes” algorithm, a “ball-pivoting” algorithm, and a “power crust” algorithm, in which the three-dimensional surface representation is constrained relative to the one or more anchor portions according to any one or more of the various different methods of constraint described herein.
  • Additionally, or alternatively, the three-dimensional surface representation can include a mesh (e.g., a continuous mesh). The mesh can be formed of, for example, a plurality of polygons (e.g., triangles) combined together to represent contours of the surface of the anatomic structure.
  • In some implementations, the generated 167 three-dimensional surface representation can be smoothed according to any of various different smoothing techniques known in the art to provide a more realistic representation of the surface of the anatomic structure.
  • The exemplary method 160 can optionally include representing 168, on a graphical user interface, any of various different combinations of the three-dimensional surface representation, the anchor portions, and the three-dimensional data structure. The graphical user interface can be, for example, a two-dimensional graphical user interface such as the graphical user interface 110 (FIG. 1). Accordingly, the exemplary method 160 can include representing 168, on the graphical user interface, a two-dimensional projection of the three-dimensional surface representation. In addition, or in the alternative, the exemplary method 160 can include representing 168, on the graphical user interface, a two-dimensional projection of the three-dimensional data structure.
  • In certain implementations, the exemplary method 160 can further include representing 168 the one or more anchor portions on the graphical user interface. For example, the one or more anchor portions can be represented 168 on the graphical user interface on a projection of the three-dimensional data structure, on a projection of the three-dimensional surface representation, or both. Additionally, or alternatively, the one or more anchor portions can be represented 168 on the graphical user interface separately from the three-dimensional data structure and/or the three-dimensional surface. It should be appreciated that representing 168 the one or more anchor portions on the graphical user interface can, for example, facilitate modification of the one or more anchor portions in certain instances. Additionally, or alternatively, the one or more anchor portions can be represented 168, on the graphical user interface, as annotations on the three-dimensional surface representation of the anatomic structure. The annotations can include, for example, tags of corresponding anatomic features, tags corresponding to locations for application of treatment (e.g., ablation), or combinations thereof. By way of example, the annotations can constrain the three-dimensional surface representation to remain unchanged as other anchor portions are added. As a further or alternative example, the three-dimensional surface representation can be constrained to pass through a portion of the three-dimensional data structure nearest to the annotation.
  • In certain implementations, the exemplary method 160 can optionally include determining 169 whether the one or more anchor portions have been modified. If the one or more anchor portions are determined 169 to be modified, the generating step 167 can be repeated. Thus, in general, the exemplary method 160 can be iterative. That is, in response to the generated 167 three-dimensional surface representation, the physician can continue to make modifications as necessary. These modifications can be based on one or more inputs received from any one or more of various input devices known in the art and described herein. Accordingly, modifying the one or more anchor portions can be based on one or more inputs from a keyboard, a mouse, a touchscreen, the medical device, or combinations thereof.
  • Modifying the one or more anchor portions can include removing at least one of the one or more anchor portions. Such removal can be useful, in certain instances, for adjusting the three-dimensional surface representation (e.g., after the three-dimensional surface representation has been generated 167) to achieve a shape that is more accurate. Additionally, or alternatively, removal of at least one of the one or more anchor portions can correct an incorrectly identified anchor portion. It should be appreciated, therefore, that removal of at least one of the one or more anchor portions can serve as an “undo” function such that correction of an incorrectly identified anchor portion does not require the physician to engage, for example, in a complex editing process. More generally, modifying the identified one or more anchor portions and repeating the generating step 167 as part of the iterative process described herein can facilitate efficient and accurate generation of the three-dimensional surface representation of the anatomic structure, as compared to tools that allow a user to selectively delete subvolumes. That is, selecting a subvolume on a two-dimensional graphical user interface commonly requires multiple selection steps from different views, which can be time consuming and subject to inaccuracies and can often require complex user interaction.
  • FIG. 8 is a flowchart of an exemplary method 170 of representing a surface of a heart cavity of a patient. The exemplary method 170 can include forming 172 a three-dimensional data structure based on received locations of a tip section of a cardiac catheter in a heart cavity of a patient, receiving 174 identification of one or more anchor portions representing locations within the heart cavity, and generating 176 a three-dimensional surface representation of the heart cavity of the patient. The three-dimensional surface representation can be generated using information from the three-dimensional data structure and can be constrained relative to the one or more anchor portions. The tip section of the catheter can be, for example, the tip section 124 described with respect to FIGS. 2 and 3. Additionally, or alternatively, it should be appreciated that a heart cavity is an example of the anatomic structure 132. Accordingly, the tip section can interact with the heart cavity in any manner and form described herein with respect to the interaction of the tip section 124 and the anatomic structure 132.
  • Forming 172 the three-dimensional data structure can include any one or more of the various different methods of forming a three-dimensional data structure disclosed herein. For example, forming 172 the three-dimensional data structure can be analogous to forming 164 the three-dimensional data structure as described with respect to FIG. 7. Accordingly, forming 172 the three-dimensional data structure can be based on locations visited by the catheter in the heart cavity. Thus, in implementations in which the three-dimensional data structure includes a three-dimensional grid, voxels corresponding to visited locations of the catheter can be set to a different state than voxels corresponding to locations that have not been visited by the catheter.
  • In general, receiving 174 the one or more anchor portions on the three-dimensional data structure can be analogous to receiving 166 the one or more anchor portions on the three-dimensional data structure, as described with respect to FIG. 7. Thus, for example, receiving 174 the one or more anchor portions on the three-dimensional data structure can include receiving an input command from a user interface (e.g., an input device such as a keyboard, a mouse, a touchscreen, and the like) corresponding to a location of an anatomic feature and/or receiving a signal indicative of contact between the catheter and tissue in the heart cavity. In the case of the heart cavity, the one or more anchor portions can correspond, for example, to the location of one or more carina associated with the heart cavity. As a more specific example, the one or more anchor portions can correspond to a carina between the left atrial appendage (LAA) and the left superior pulmonary vein (LSPV).
  • In the alternative, or in addition, receiving 174 the one or more anchor portions can include receiving one or more signals corresponding to one or more respective locations of the cardiac catheter in the heart cavity. For example, the one or more signals corresponding to one or more respective locations of the cardiac catheter in the heart cavity can correspond to a blood-tissue boundary of the heart cavity. Such signals can include, for example, to one or more of: a change in an electric signal (e.g., electrogram or impedance) detected by one or more electrodes of the catheter, a force detected by a force sensor of the catheter, an ultrasound signal of an ultrasound sensor of the catheter, and a deformation of at least a portion of the catheter. For example, such signals can correspond to an amplitude derived from an electrogram detected by one or more electrodes of the medical device.
  • Generating 176 the three-dimensional surface representation of the heart cavity of the patient can be, in general, analogous to generating 167 the three-dimensional surface representation of the anatomic structure as described with respect to FIG. 7. Accordingly, it should be understood that generating 176 the three-dimensional surface representation of the heart cavity can be based, for example, on one or more of a “marching cubes” algorithm, a “ball-pivoting” algorithm, and a “power crust algorithm,” with the algorithm or algorithms constrained relative to the one or more anchor portions. Further, or instead, generating 176 the three-dimensional surface representation of the heart cavity of the patient can include an undo function such that one or more of the anchor portions can be removed or repositioned, and the three-dimensional surface representation of the heart cavity can be regenerated based on the updated position of the one or more anchor portions.
  • In certain implementations, the three-dimensional surface representation of the heart cavity can be surface smoothed. It should be appreciated that such surface smoothing can produce changes to the three-dimensional surface representation and, in certain instances, can cause the three-dimensional surface representation to no longer fully contain the visited locations. Additionally, or alternatively, surface smoothing can produce changes to the three-dimensional surface representation that can cause the three-dimensional surface representation to no longer pass directly through the anchor portions. Surface-smoothing the three-dimensional surface representation can result in any one or more of the various different advantages described herein. For example, surface smoothing the three-dimensional surface representation can facilitate visualization of the position of the catheter relative to the three-dimensional surface representation, which can be useful for positioning the catheter during an ablation treatment applied to a surface of the heart cavity.
  • In some implementations, the exemplary method 170 can further include representing 177, on a graphical user interface, at least one of a two-dimensional projection of the three-dimensional data structure, the one or more anchor portions, and a two-dimensional projection of the three-dimensional surface representation. Representing 177 the two-dimensional projection of the three-dimensional surface representation and generating 176 the three-dimensional surface representation can be part of an iterative process, such as an iterative process analogous to the iterative process described with respect to the exemplary method 160.
  • The graphical user interface can be, for example, the graphical user interface 110 described with respect to FIG. 1. In certain implementations, the one or more anchor portions can be represented 177, on the graphical user interface, as annotations or other similar visual indicia on the three-dimensional surface representation of the heart cavity, on the three-dimensional data structure, or both. Additionally, or alternatively, the one or more anchor portions can be represented 177 on the graphical user interface independently of the three-dimensional surface representation, the three-dimensional data structure, or both. The annotations can, for example, be tags of certain anatomic features and/or tags related to the position of a treatment (such as a lesion created through tissue ablation).
  • While certain implementations have been described, other implementations are additionally or alternatively possible.
  • For example, while graphical user interfaces have been described as including a two-dimensional display, any one or more of the graphical user interfaces described herein can additionally, or alternatively, include a three-dimensional display. Examples of such a three-dimensional display include an augmented reality environment, a virtual reality environment, and combinations thereof.
  • The above systems, devices, methods, processes, and the like may be realized in hardware, software, or any combination of these suitable for a particular application. The hardware may include a general-purpose computer and/or dedicated computing device. This includes realization in one or more microprocessors, microcontrollers, embedded microcontrollers, programmable digital signal processors or other programmable devices or processing circuitry, along with internal and/or external memory. This may also, or instead, include one or more application specific integrated circuits, programmable gate arrays, programmable array logic components, or any other device or devices that may be configured to process electronic signals.
  • It will further be appreciated that a realization of the processes or devices described above may include computer-executable code created using a structured programming language such as C, an object oriented programming language such as C++, or any other high-level or low level programming language (including assembly languages, hardware description languages, and database programming languages and technologies) that may be stored, compiled or interpreted to run on one of the above devices, as well as heterogeneous combinations of processors, processor architectures, or combinations of different hardware and software. In another aspect, the methods may be embodied in systems that perform the steps thereof, and may be distributed across devices in a number of ways. At the same time, processing may be distributed across devices such as the various systems described above, or all of the functionality may be integrated into a dedicated, standalone device or other hardware. In another aspect, means for performing the steps associated with the processes described above may include any of the hardware and/or software described above. All such permutations and combinations are intended to fall within the scope of the present disclosure.
  • Embodiments disclosed herein may include computer program products comprising computer-executable code or computer-usable code that, when executing on one or more computing devices, performs any and/or all of the steps thereof. The code may be stored in a non-transitory fashion in a computer memory, which may be a memory from which the program executes (such as random access memory associated with a processor), or a storage device such as a disk drive, flash memory or any other optical, electromagnetic, magnetic, infrared or other device or combination of devices.
  • In another aspect, any of the systems and methods described above may be embodied in any suitable transmission or propagation medium carrying computer-executable code and/or any inputs or outputs from same.
  • The method steps of the implementations described herein are intended to include any suitable method of causing such method steps to be performed, consistent with the patentability of the following claims, unless a different meaning is expressly provided or otherwise clear from the context. So for example performing the step of X includes any suitable method for causing another party such as a remote user, a remote processing resource (e.g., a server or cloud computer) or a machine to perform the step of X. Similarly, performing steps X, Y and Z may include any method of directing or controlling any combination of such other individuals or resources to perform steps X, Y and Z to obtain the benefit of such steps. Thus method steps of the implementations described herein are intended to include any suitable method of causing one or more other parties or entities to perform the steps, consistent with the patentability of the following claims, unless a different meaning is expressly provided or otherwise clear from the context. Such parties or entities need not be under the direction or control of any other party or entity, and need not be located within a particular jurisdiction.
  • It will be appreciated that the methods and systems described above are set forth by way of example and not of limitation. Numerous variations, additions, omissions, and other modifications will be apparent to one of ordinary skill in the art. In addition, the order or presentation of method steps in the description and drawings above is not intended to require this order of performing the recited steps unless a particular order is expressly required or otherwise clear from the context. Thus, while particular embodiments have been shown and described, it will be apparent to those skilled in the art that various changes and modifications in form and details may be made therein without departing from the spirit and scope of this disclosure and are intended to form a part of the invention as defined by the following claims.

Claims (20)

What is claimed is:
1. A method comprising:
receiving a plurality of location signals, each received location signal indicative of a respective location of a medical device in an anatomic structure of a patient;
forming a three-dimensional data structure representing locations, within the anatomic structure, visited by the medical device at the locations corresponding to the plurality of location signals;
receiving one or more anchor portions representing locations relative to the anatomic structure; and
generating a three-dimensional surface representation of the anatomic structure of the patient, the three-dimensional surface representation of the anatomic structure of the patient constrained relative to the one or more anchor portions and to contain at least a portion of the three-dimensional data structure.
2. The method of claim 1, further comprising displaying, on a graphical user interface, at least one of a two-dimensional projection of the three-dimensional data structure, the one or more anchor portions, and a two-dimensional projection of the three-dimensional surface representation.
3. The method of claim 1, wherein receiving the one or more anchor portions representing locations relative to the anatomic structure includes receiving, from one or more sensors disposed on the medical device, a signal indicative of contact between the medical device and tissue of the anatomic structure.
4. The method of claim 3, wherein the signal indicative of contact is indicative of a blood-tissue boundary of the anatomic structure of the patient.
5. The method of claim 3, wherein the signal indicative of contact includes one or more of: a change in impedance detected by one or more electrodes of the medical device, a force detected by a force sensor of the medical device, an ultrasound signal of an ultrasound sensor of the medical device, a deformation of at least a portion of the medical device, and an amplitude derived from an electrogram detected by one or more electrodes of the medical device.
6. The method of claim 1, wherein receiving the one or more anchor portions includes receiving an input command from a user.
7. The method of claim 1, wherein receiving the one or more anchor portions includes identifying a subset of the three-dimensional data structure.
8. The method of claim 1, wherein receiving the one or more anchor portions includes receiving a respective confidence level associated each of the one or more anchor portions, and constraining the three-dimensional surface representation relative to the one or more anchor portions is based on the respective confidence level associated with each of the one or more anchor portions.
9. The method of claim 1, further comprising representing, on a graphical user interface, the one or more anchor portions as annotations on the three-dimensional surface representation of the anatomic structure.
10. The method of claim 1, further comprising representing, on a graphical user interface, the one or more anchor portions as annotations on the three-dimensional data structure.
11. The method of claim 1, further comprising determining whether the one or more anchor portions have been modified and, based on whether the one or more anchor portions have been modified, repeating the generating step.
12. The method of claim 11, wherein determining whether the one or more anchor portions have been modified includes determining whether one or more of previously identified anchor portions have been removed.
13. The method of claim 1, wherein the three-dimensional surface representation of the anatomic structure is a continuous mesh.
14. A method comprising:
forming a three-dimensional data structure based on received locations of a tip section of a cardiac catheter in a heart cavity of a patient;
receiving one or more anchor portions representing locations relative to the heart cavity; and
generating a three-dimensional surface representation of the heart cavity of the patient, the surface representation of the heart cavity of the patient constrained relative to the anchor portions and to contain at least a portion of the three-dimensional data structure.
15. The method of claim 14, further comprising displaying, on a graphical user interface, at least one of a two-dimensional projection of the three-dimensional data structure, the one or more anchor portions, and a two-dimensional projection of the generated three-dimensional surface representation.
16. The method of claim 14, wherein receiving the one or more anchor portions on the three-dimensional data structure includes receiving one or more location signals indicative of one or more respective locations of the cardiac catheter in the heart cavity.
17. The method of claim 14, wherein receiving the one or more anchor portions includes receiving, from a sensor disposed on the cardiac catheter, a signal indicative of a blood-tissue boundary of the heart cavity of the patient.
18. The method of claim 17, wherein the signal indicative of the blood-tissue boundary includes one or more of: a change in impedance detected by one or more electrodes of the cardiac catheter, a force detected by a force sensor of the cardiac catheter, an ultrasound signal of an ultrasound sensor of the cardiac catheter, and a deformation of at least a portion of the cardiac catheter, and an amplitude derived from an electrogram detected by one or more electrodes of the cardiac catheter.
19. The method of claim 14, wherein receiving the one or more anchor portions on the three-dimensional data structure includes receiving an input command from a user interface.
20. A non-transitory, computer-readable storage medium having stored thereon computer executable instructions for causing one or more processors to:
receive a plurality of location signals, each received location signal indicative of a respective location of a medical device in an anatomic structure of a patient;
form a three-dimensional data structure representing volumes, within the anatomic structure, occupied by the medical device at the locations corresponding to the plurality of location signals;
receive one or more anchor portions representing locations relative to the anatomic structure; and
generate a three-dimensional surface representation of the anatomic structure of the patient, the three-dimensional surface representation of the anatomic structure of the patient constrained relative to the one or more anchor portions and containing at least a portion of the three-dimensional data structure.
US17/195,482 2016-05-11 2021-03-08 Anatomical model generation Abandoned US20210401503A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US17/195,482 US20210401503A1 (en) 2016-05-11 2021-03-08 Anatomical model generation
US17/658,252 US20230013302A1 (en) 2016-05-11 2022-04-06 Anatomical model generation

Applications Claiming Priority (7)

Application Number Priority Date Filing Date Title
US201662334577P 2016-05-11 2016-05-11
US201662338105P 2016-05-18 2016-05-18
US201662393876P 2016-09-13 2016-09-13
US15/592,815 US10376320B2 (en) 2016-05-11 2017-05-11 Anatomical model generation
US16/525,363 US10765481B2 (en) 2016-05-11 2019-07-29 Anatomical model generation
US202016945785A 2020-07-31 2020-07-31
US17/195,482 US20210401503A1 (en) 2016-05-11 2021-03-08 Anatomical model generation

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US202016945785A Continuation 2016-05-11 2020-07-31

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US202117477326A Continuation 2016-05-11 2021-09-16

Publications (1)

Publication Number Publication Date
US20210401503A1 true US20210401503A1 (en) 2021-12-30

Family

ID=58765952

Family Applications (4)

Application Number Title Priority Date Filing Date
US15/592,815 Active US10376320B2 (en) 2016-05-11 2017-05-11 Anatomical model generation
US16/525,363 Active US10765481B2 (en) 2016-05-11 2019-07-29 Anatomical model generation
US17/195,482 Abandoned US20210401503A1 (en) 2016-05-11 2021-03-08 Anatomical model generation
US17/658,252 Pending US20230013302A1 (en) 2016-05-11 2022-04-06 Anatomical model generation

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US15/592,815 Active US10376320B2 (en) 2016-05-11 2017-05-11 Anatomical model generation
US16/525,363 Active US10765481B2 (en) 2016-05-11 2019-07-29 Anatomical model generation

Family Applications After (1)

Application Number Title Priority Date Filing Date
US17/658,252 Pending US20230013302A1 (en) 2016-05-11 2022-04-06 Anatomical model generation

Country Status (3)

Country Link
US (4) US10376320B2 (en)
EP (1) EP3454734B1 (en)
WO (1) WO2017197114A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024127164A1 (en) * 2022-12-14 2024-06-20 Biosense Webster (Israel) Ltd. Using signed distance functions to visualize pulsed field ablation (pfa) tags

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017197114A1 (en) 2016-05-11 2017-11-16 Affera, Inc. Anatomical model generation
WO2017197294A1 (en) 2016-05-12 2017-11-16 Affera, Inc. Three-dimensional cardiac representation
US11317966B2 (en) * 2017-07-19 2022-05-03 Biosense Webster (Israel) Ltd. Impedance-based position tracking performance using scattered interpolant
US11596477B2 (en) * 2017-11-20 2023-03-07 Covidien Lp Systems and methods for generating energy delivery heat maps
EP3745976A4 (en) * 2018-02-02 2021-10-06 Centerline Biomedical, Inc. Graphical user interface for marking anatomic structures
US11219488B2 (en) 2018-04-25 2022-01-11 Biosense Webster (Israel) Ltd. Determining catheter touch location using force-vector information
US11854217B2 (en) * 2019-05-03 2023-12-26 Koninklijke Philips N.V. Co-registration of cardiac images
US11443425B2 (en) 2020-04-20 2022-09-13 Biosense Webster (Israel) Ltd. Fast anatomical mapping (FAM) reconstruction using surface update restrictions
US11113899B1 (en) * 2020-08-31 2021-09-07 Biosense Webster (Israel) Ltd. Correcting anatomical maps
US11544847B1 (en) * 2020-10-16 2023-01-03 Neutrace, Inc. System and method for generating point cloud data for electro-anatomical mapping
US11911167B2 (en) 2021-01-19 2024-02-27 Biosense Webster (Israel) Ltd. Automatic mesh reshaping of an anatomical map to expose internal points of interest
CN118805204A (en) 2021-12-28 2024-10-18 伯恩森斯韦伯斯特(以色列)有限责任公司 Automatic shaving of anatomical maps during ablation to expose internal points of interest
US11900524B2 (en) 2022-03-03 2024-02-13 Biosense Webster (Israel) Ltd. Constructing topography of lumen wall in 4D ultrasound image with virtual ellipsoid or polyhedron
US20240050017A1 (en) 2022-08-10 2024-02-15 Biosense Webster (Israel) Ltd. Visualizing and Clustering Multiple Electrodes of a High-Definition Catheter Projected on Tissue Surface

Family Cites Families (148)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4734690A (en) 1984-07-20 1988-03-29 Tektronix, Inc. Method and apparatus for spherical panning
US5276785A (en) 1990-08-02 1994-01-04 Xerox Corporation Moving viewpoint with respect to a target in a three-dimensional workspace
US5133336A (en) 1990-10-22 1992-07-28 Endoscopy Support Services, Inc. Disposable liquid supply system for use in an endoscope
US5687737A (en) 1992-10-09 1997-11-18 Washington University Computerized three-dimensional cardiac mapping with interactive visual displays
US5433198A (en) 1993-03-11 1995-07-18 Desai; Jawahar M. Apparatus and method for cardiac ablation
US5364395A (en) 1993-05-14 1994-11-15 West Jr Hugh S Arthroscopic surgical instrument with cauterizing capability
US5391199A (en) 1993-07-20 1995-02-21 Biosense, Inc. Apparatus and method for treating cardiac arrhythmias
US5447529A (en) 1994-01-28 1995-09-05 Philadelphia Heart Institute Method of using endocardial impedance for determining electrode-tissue contact, appropriate sites for arrhythmia ablation and tissue heating during ablation
JPH07262412A (en) 1994-03-16 1995-10-13 Fujitsu Ltd Device and system for indicating cross section of three-dimensional model
US5797849A (en) 1995-03-28 1998-08-25 Sonometrics Corporation Method for carrying out a medical procedure using a three-dimensional tracking and imaging system
US5655535A (en) 1996-03-29 1997-08-12 Siemens Medical Systems, Inc. 3-Dimensional compound ultrasound field of view
US5889524A (en) 1995-09-11 1999-03-30 University Of Washington Reconstruction of three-dimensional objects using labeled piecewise smooth subdivision surfaces
US6175655B1 (en) 1996-09-19 2001-01-16 Integrated Medical Systems, Inc. Medical imaging system for displaying, manipulating and analyzing three-dimensional images
US5820568A (en) 1996-10-15 1998-10-13 Cardiac Pathways Corporation Apparatus and method for aiding in the positioning of a catheter
IL120867A0 (en) 1997-05-20 1997-09-30 Cadent Ltd Computer user interface for orthodontic use
JP3183632B2 (en) 1997-06-13 2001-07-09 株式会社ナムコ Information storage medium and image generation device
JPH1132986A (en) 1997-07-16 1999-02-09 Olympus Optical Co Ltd Endoscope system
US6490474B1 (en) 1997-08-01 2002-12-03 Cardiac Pathways Corporation System and method for electrode localization using ultrasound
US7092773B1 (en) 1997-09-24 2006-08-15 Sony Corporation Method and system for providing enhanced editing capabilities
CA2333583C (en) 1997-11-24 2005-11-08 Everette C. Burdette Real time brachytherapy spatial registration and visualization system
US6037937A (en) 1997-12-04 2000-03-14 Nortel Networks Corporation Navigation tool for graphical user interface
US6377865B1 (en) 1998-02-11 2002-04-23 Raindrop Geomagic, Inc. Methods of generating three-dimensional digital models of objects by wrapping point cloud data points
US7734656B2 (en) 1998-02-24 2010-06-08 Luc Bessette System and method for electronically managing medical data files in order to facilitate genetic research
US6271856B1 (en) 1998-11-19 2001-08-07 Paraform, Inc. Creating and modifying parameterizations of surfaces
US6256038B1 (en) 1998-12-10 2001-07-03 The Board Of Trustees Of The Leland Stanford Junior University Parameterized surface fitting technique having independent control of fitting and parameterization
US7155042B1 (en) * 1999-04-21 2006-12-26 Auckland Uniservices Limited Method and system of measuring characteristics of an organ
US6443894B1 (en) 1999-09-29 2002-09-03 Acuson Corporation Medical diagnostic ultrasound system and method for mapping surface data for three dimensional imaging
US6556206B1 (en) 1999-12-09 2003-04-29 Siemens Corporate Research, Inc. Automated viewpoint selection for 3D scenes
US6968299B1 (en) 2000-04-14 2005-11-22 International Business Machines Corporation Method and apparatus for reconstructing a surface using a ball-pivoting algorithm
JP3854062B2 (en) 2000-04-28 2006-12-06 株式会社モリタ製作所 Tomographic image display method, display device, and recording medium storing program for realizing the display method
JP3579646B2 (en) 2000-11-21 2004-10-20 ペンタックス株式会社 Ultrasound endoscope
JP3579651B2 (en) 2000-11-21 2004-10-20 ペンタックス株式会社 Ultrasound endoscope
ATE551955T1 (en) 2001-04-20 2012-04-15 Tyco Healthcare SURGICAL DEVICE HAVING BIPOLAR OR ULTRASONIC FEATURES
US7450749B2 (en) 2001-07-06 2008-11-11 Koninklijke Electronics N.V. Image processing method for interacting with a 3-D surface represented in a 3-D image
US6764442B2 (en) 2001-08-10 2004-07-20 Pentax Corporation Liquid and gas supply apparatus and portable endoscope with the same
US7114500B2 (en) 2001-08-28 2006-10-03 Marctec, Llc Surgical draping system
US7023432B2 (en) 2001-09-24 2006-04-04 Geomagic, Inc. Methods, apparatus and computer program products that reconstruct surfaces from data point sets
EP1453430A4 (en) 2001-11-09 2009-02-18 Cardio Optics Inc Direct, real-time imaging guidance of cardiac catheterization
US7285117B2 (en) 2002-03-15 2007-10-23 Boston Scientific Scimed, Inc. Medical device control systems
GB2387519B (en) 2002-04-08 2005-06-22 Canon Europa Nv Viewing controller for three-dimensional computer graphics
TW558689B (en) 2002-08-30 2003-10-21 Univ Taipei Medical Three-dimensional surgery simulation system and method
JP2004105256A (en) 2002-09-13 2004-04-08 Fuji Photo Film Co Ltd Image display device
JP4516957B2 (en) 2003-01-25 2010-08-04 パーデュー リサーチ ファンデーション Method, system and data structure for searching for 3D objects
US7343058B2 (en) 2003-04-22 2008-03-11 Intel Corporation Efficient light coupler from off-chip to on-chip waveguides
WO2004109603A1 (en) 2003-06-11 2004-12-16 Koninklijke Philips Electronics, N.V. User control of 3d volume plane crop
US9082220B2 (en) 2003-08-26 2015-07-14 Siemens Product Lifecycle Management Software Inc. System, method, and computer program product for smoothing
DE10339979B4 (en) 2003-08-29 2005-11-17 Tomtec Imaging Systems Gmbh Method and device for displaying a predeterminable area in multi-dimensional data sets
JP3847744B2 (en) 2003-11-04 2006-11-22 オリンパス株式会社 Insertion support system
US7317955B2 (en) 2003-12-12 2008-01-08 Conmed Corporation Virtual operating room integration
WO2005063125A1 (en) 2003-12-22 2005-07-14 Koninklijke Philips Electronics N.V. System for guiding a medical instrument in a patient body
US7715604B2 (en) 2005-01-18 2010-05-11 Siemens Medical Solutions Usa, Inc. System and method for automatically registering three dimensional cardiac images with electro-anatomical cardiac mapping data
US20060241445A1 (en) 2005-04-26 2006-10-26 Altmann Andres C Three-dimensional cardial imaging using ultrasound contour reconstruction
US8721634B2 (en) 2005-07-21 2014-05-13 Covidien Lp Apparatus and method for ensuring thermal treatment of a hollow anatomical structure
US7365745B2 (en) 2005-09-15 2008-04-29 St. Jude Medical, Atrial Fibrillation Division, Inc. Method of rendering a surface from a solid graphical image
US8229545B2 (en) 2005-09-15 2012-07-24 St. Jude Medical, Atrial Fibrillation Division, Inc. System and method for mapping complex fractionated electrogram information
US7981038B2 (en) 2005-10-11 2011-07-19 Carnegie Mellon University Sensor guided catheter navigation system
DE602006016221D1 (en) 2005-12-05 2010-09-30 Siemens Corp Method and device for discrete network rounding and rounding by means of ball-pivoting
US20090177111A1 (en) 2006-12-06 2009-07-09 Miller Stephan P System and method for displaying contact between a catheter and tissue
US8930834B2 (en) 2006-03-20 2015-01-06 Microsoft Corporation Variable orientation user interface
US7855723B2 (en) 2006-03-21 2010-12-21 Biosense Webster, Inc. Image registration using locally-weighted fitting
EP2004060A1 (en) 2006-04-03 2008-12-24 Koninklijke Philips Electronics N.V. Determining tissue surrounding an object being inserted into a patient
US7515954B2 (en) 2006-06-13 2009-04-07 Rhythmia Medical, Inc. Non-contact cardiac mapping, including moving catheter and multi-beat integration
US7729752B2 (en) 2006-06-13 2010-06-01 Rhythmia Medical, Inc. Non-contact cardiac mapping, including resolution map
US7505810B2 (en) 2006-06-13 2009-03-17 Rhythmia Medical, Inc. Non-contact cardiac mapping, including preprocessing
DE602007006520D1 (en) 2006-08-15 2010-06-24 Spectracure Ab SYSTEM AND METHOD FOR CONTROLLING AND ADJUSTING INTERSTITIAL PHOTODYNAMIC LIGHT THERAPY PARAMETERS
US8014561B2 (en) 2006-09-07 2011-09-06 University Of Louisville Research Foundation, Inc. Virtual fly over of complex tubular anatomical structures
US8007437B2 (en) 2006-11-08 2011-08-30 Siemens Aktiengesellschaft Method and apparatus for interactive 4-dimensional (4D) virtual endoscopy
US9220439B2 (en) 2006-12-29 2015-12-29 St. Jude Medical, Atrial Fibrillation Division, Inc. Navigational reference dislodgement detection method and system
JP5639764B2 (en) 2007-03-08 2014-12-10 シンク−アールエックス,リミティド Imaging and tools for use with moving organs
US9549689B2 (en) 2007-03-09 2017-01-24 St. Jude Medical, Atrial Fibrillation Division, Inc. System and method for correction of inhomogeneous fields
US10433929B2 (en) 2007-03-09 2019-10-08 St. Jude Medical, Atrial Fibrillation Division, Inc. System and method for local deformable registration of a catheter navigation system to image data or a model
US7777130B2 (en) 2007-06-18 2010-08-17 Vivant Medical, Inc. Microwave cable cooling
US7894663B2 (en) 2007-06-30 2011-02-22 General Electric Company Method and system for multiple view volume rendering
US8057394B2 (en) * 2007-06-30 2011-11-15 St. Jude Medical, Atrial Fibrillation Division, Inc. Ultrasound image processing to render three-dimensional images from two-dimensional images
US20090076476A1 (en) 2007-08-15 2009-03-19 Hansen Medical, Inc. Systems and methods employing force sensing for mapping intra-body tissue
US8357152B2 (en) 2007-10-08 2013-01-22 Biosense Webster (Israel), Ltd. Catheter with pressure sensing
EP2206093B1 (en) 2007-11-02 2013-06-26 Koninklijke Philips Electronics N.V. Automatic movie fly-path calculation
US8103327B2 (en) 2007-12-28 2012-01-24 Rhythmia Medical, Inc. Cardiac mapping catheter
US9211160B2 (en) 2008-01-16 2015-12-15 Luiz Geraldo Pivotto Remotely controlled catheter insertion system with automatic control system
WO2009120953A2 (en) 2008-03-27 2009-10-01 Mayo Foundation For Medical Education And Research Navigation and tissue capture systems and methods
US9161817B2 (en) 2008-03-27 2015-10-20 St. Jude Medical, Atrial Fibrillation Division, Inc. Robotic catheter system
US9241768B2 (en) 2008-03-27 2016-01-26 St. Jude Medical, Atrial Fibrillation Division, Inc. Intelligent input device controller for a robotic catheter system
US8494608B2 (en) 2008-04-18 2013-07-23 Medtronic, Inc. Method and apparatus for mapping a structure
US8786594B2 (en) 2008-05-30 2014-07-22 Siemens Aktiengesellschaft Method and system for rendering a medical image
US8803910B2 (en) * 2008-08-28 2014-08-12 Tomotherapy Incorporated System and method of contouring a target area
US9245382B2 (en) 2008-10-04 2016-01-26 Microsoft Technology Licensing, Llc User-guided surface reconstruction
EP2455038B1 (en) 2008-10-21 2015-04-01 Brainlab AG Integration of surgical instrument and display device for supporting image led surgery
US8137343B2 (en) 2008-10-27 2012-03-20 Rhythmia Medical, Inc. Tracking system using field mapping
WO2010054409A1 (en) 2008-11-10 2010-05-14 Cardioinsight Technologies, Inc. Visualization of electrophysiology data
US8745536B1 (en) 2008-11-25 2014-06-03 Perceptive Pixel Inc. Volumetric data exploration using multi-point input controls
US8700129B2 (en) 2008-12-31 2014-04-15 St. Jude Medical, Atrial Fibrillation Division, Inc. Devices and methods for catheter localization
CN101836862B (en) 2009-03-16 2014-03-26 上海微创医疗器械(集团)有限公司 Three-dimensional mapping method of human chamber inner wall and equipment and system thereof
US20100305427A1 (en) 2009-06-01 2010-12-02 General Electric Company Long-range planar sensor array for use in a surgical navigation system
EP3542713A1 (en) 2009-06-12 2019-09-25 Bard Access Systems, Inc. Adapter for a catheter tip positioning device
US9439736B2 (en) 2009-07-22 2016-09-13 St. Jude Medical, Atrial Fibrillation Division, Inc. System and method for controlling a remote medical device guidance system in three-dimensions using gestures
WO2011123669A1 (en) 2010-03-31 2011-10-06 St. Jude Medical, Atrial Fibrillation Division, Inc. Intuitive user interface control for remote catheter navigation and 3d mapping and visualization systems
US10624553B2 (en) 2009-12-08 2020-04-21 Biosense Webster (Israel), Ltd. Probe data mapping using contact information
US10835207B2 (en) 2009-12-23 2020-11-17 Biosense Webster (Israel) Ltd. Fast anatomical mapping using ultrasound images
US9521376B2 (en) 2010-01-20 2016-12-13 Olympus Corporation Endoscope apparatus
US10580325B2 (en) 2010-03-24 2020-03-03 Simbionix Ltd. System and method for performing a computerized simulation of a medical procedure
CN102934006A (en) 2010-04-01 2013-02-13 奥林巴斯株式会社 Endoscope device, and connecting unit for endoscope device
US8811699B2 (en) 2010-09-22 2014-08-19 Siemens Aktiengesellschaft Detection of landmarks and key-frames in cardiac perfusion MRI using a joint spatial-temporal context model
US8340766B2 (en) 2010-10-07 2012-12-25 St. Jude Medical, Atrial Fibrillation Division, Inc. Method and system for identifying cardiac arrhythmia driver sites
US9769912B2 (en) 2010-10-20 2017-09-19 Medtronic Navigation, Inc. Gated image acquisition and patient model construction
US20120123404A1 (en) 2010-11-16 2012-05-17 Tyco Healthcare Group Lp Fingertip Electrosurgical Instruments for Use in Hand-Assisted Surgery and Systems Including Same
US9547752B2 (en) 2010-12-31 2017-01-17 St. Jude Medical, Atrial Fibrillation Division, Inc. Automated catheter guidance system
US9047395B2 (en) 2011-01-05 2015-06-02 Olympus Corporation Endoscope apparatus and method for storing endoscopic image
US20120191086A1 (en) 2011-01-20 2012-07-26 Hansen Medical, Inc. System and method for endoluminal and translumenal therapy
US8606597B2 (en) 2011-02-24 2013-12-10 Olympus Corporation Endoscope inspection report creating apparatus, creating method of endoscope inspection report and storage medium
US20120245465A1 (en) 2011-03-25 2012-09-27 Joger Hansegard Method and system for displaying intersection information on a volumetric ultrasound image
US20130002715A1 (en) 2011-06-28 2013-01-03 Tidman James M Image Sequence Reconstruction based on Overlapping Measurement Subsets
US20130030285A1 (en) 2011-07-26 2013-01-31 General Electric Company System and method for integrating multiple data sources into an x-ray image referential
US8817076B2 (en) 2011-08-03 2014-08-26 General Electric Company Method and system for cropping a 3-dimensional medical dataset
US9298351B2 (en) 2011-08-03 2016-03-29 Olympus Corporation Inspection image display apparatus, inspection image display method and storage medium
US8724881B2 (en) * 2011-11-09 2014-05-13 Siemens Aktiengesellschaft Method and system for precise segmentation of the left atrium in C-arm computed tomography volumes
US8920368B2 (en) 2011-12-22 2014-12-30 St. Jude Medical, Atrial Fibrillation Division, Inc. Multi-user touch-based control of a remote catheter guidance system (RCGS)
US9311744B2 (en) 2012-01-09 2016-04-12 Fca Us Llc System and method for generating an outer layer representation of an object
US20130241929A1 (en) 2012-03-13 2013-09-19 Fady Massarwa Selectably transparent electrophysiology map
WO2013162735A1 (en) 2012-04-25 2013-10-31 University Of Southern California 3d body modeling from one or more depth cameras in the presence of articulated motion
US9846960B2 (en) 2012-05-31 2017-12-19 Microsoft Technology Licensing, Llc Automated camera array calibration
US10070827B2 (en) 2012-10-05 2018-09-11 Volcano Corporation Automatic image playback
JP2016507304A (en) * 2013-02-11 2016-03-10 アンジオメトリックス コーポレーション System for detecting and tracking and superimposing objects
US11278353B2 (en) 2016-03-16 2022-03-22 Synaptive Medical Inc. Trajectory alignment system and methods
EP4086854A1 (en) 2013-05-02 2022-11-09 Smith & Nephew, Inc. Surface and image integration for model evaluation and landmark determination
US9576107B2 (en) * 2013-07-09 2017-02-21 Biosense Webster (Israel) Ltd. Model based reconstruction of the heart from sparse samples
CA2917478A1 (en) 2013-07-10 2015-01-15 Real View Imaging Ltd. Three dimensional user interface
US9460538B2 (en) 2013-08-07 2016-10-04 Siemens Medical Solutions Usa, Inc. Animation for conveying spatial relationships in multi-planar reconstruction
US11324419B2 (en) 2013-08-20 2022-05-10 Biosense Webster (Israel) Ltd. Graphical user interface for medical imaging system
WO2015066113A1 (en) 2013-10-30 2015-05-07 St. Jude Medical, Cardiology Division, Inc. Cardiac mapping system and method for bi-directional activation detection of electrograms
US11278231B2 (en) 2014-03-25 2022-03-22 Acutus Medical, Inc. Cardiac analysis user interface system and method
US20150324114A1 (en) 2014-05-06 2015-11-12 Conceptualiz Inc. System and method for interactive 3d surgical planning and modelling of surgical implants
US20160000300A1 (en) 2014-07-07 2016-01-07 Integrated Medical Systems International, Inc. System and Method for Wirelessly Transmitting Operational Data From an Endoscope to a Remote Device
US10755369B2 (en) * 2014-07-16 2020-08-25 Parkland Center For Clinical Innovation Client management tool system and method
CN107075165A (en) * 2014-07-24 2017-08-18 亚瑟罗凯尔公司 Many sugarwhips of elasticity and application thereof
CN105631936A (en) 2014-10-31 2016-06-01 富泰华工业(深圳)有限公司 Point cloud repair method and system
US11026750B2 (en) * 2015-01-23 2021-06-08 Queen's University At Kingston Real-time surgical navigation
US10019784B2 (en) * 2015-03-18 2018-07-10 Toshiba Medical Systems Corporation Medical image processing apparatus and method
US10929508B2 (en) 2015-04-30 2021-02-23 Merge Healthcare Solutions Inc. Database systems and interactive user interfaces for dynamic interaction with, and indications of, digital medical image data
US10311566B2 (en) 2015-06-12 2019-06-04 International Business Machines Corporation Methods and systems for automatically determining image characteristics serving as a basis for a diagnosis associated with an image study type
CN107771055B (en) * 2015-06-19 2021-02-26 圣犹达医疗用品心脏病学部门有限公司 Electromagnetic dynamic registration for device navigation
KR102035993B1 (en) 2015-09-03 2019-10-25 지멘스 메디컬 솔루션즈 유에스에이, 인크. Ultrasound system and method for generating elastic image
CA3105936C (en) 2015-10-19 2023-08-01 Icu Medical, Inc. Hemodynamic monitoring system with detachable display unit
JP6227200B1 (en) 2015-12-16 2017-11-08 オリンパス株式会社 MEDICAL MANIPULATOR SYSTEM AND MEDICAL MANIPULATOR CONTROL METHOD
US10698505B2 (en) 2016-01-13 2020-06-30 Hewlett-Packard Development Company, L.P. Executing multiple pen inputs
JP6666454B2 (en) 2016-01-26 2020-03-13 セント・ジュード・メディカル・インターナショナル・ホールディング・エスエーアールエルSt. Jude Medical International Holding S.a,r.l. Detection and correction of magnetic field distortion in magnetic localization systems
WO2017192781A1 (en) 2016-05-03 2017-11-09 Affera, Inc. Anatomical model displaying
WO2017197114A1 (en) 2016-05-11 2017-11-16 Affera, Inc. Anatomical model generation
WO2017197294A1 (en) 2016-05-12 2017-11-16 Affera, Inc. Three-dimensional cardiac representation

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024127164A1 (en) * 2022-12-14 2024-06-20 Biosense Webster (Israel) Ltd. Using signed distance functions to visualize pulsed field ablation (pfa) tags

Also Published As

Publication number Publication date
EP3454734A1 (en) 2019-03-20
US20230013302A1 (en) 2023-01-19
US10765481B2 (en) 2020-09-08
EP3454734B1 (en) 2024-04-17
US20200197095A1 (en) 2020-06-25
US10376320B2 (en) 2019-08-13
US20170325891A1 (en) 2017-11-16
WO2017197114A1 (en) 2017-11-16

Similar Documents

Publication Publication Date Title
US10765481B2 (en) Anatomical model generation
US11954815B2 (en) Anatomical model displaying
US20230335262A1 (en) Three-dimensional cardiac representation
CN109589170B (en) Left atrial appendage closure guidance in medical imaging
US10925511B2 (en) System and method for cardiac ablation
JP6868712B2 (en) Systems to assist in the display of cardiac information and methods of representing cardiac information
JP7532089B2 (en) Determining the occurrence of focal and/or rotor proarrhythmic activity within a cardiac tissue region
US20220156933A1 (en) Flattened view for intra-lumenal navigation
JP2022111102A (en) Automatic mesh reshaping of anatomical map to expose internal points of interest
JP2023008979A (en) Ultrasound slice enhancement
WO2019063161A1 (en) Flattened view for intra-lumenal navigation
KR102084256B1 (en) Image registration apparatus and method using multiple candidate points
JP2023031301A (en) Anatomically correct reconstruction of atrium
KR102184001B1 (en) Medical image registration apparatus and method for surgical navigation
GB2576489A (en) Flattened view for intra-lumenal navigation
US12127794B2 (en) Method for planning tissue ablation based on deep learning
US20220233242A1 (en) Method for planning tissue ablation based on deep learning
JP2024151459A (en) Information processing device, information processing method, and program
JP2023513383A (en) Medical imaging systems, devices, and methods for visualizing deployment of internal therapeutic devices

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION