EP1929956B1 - Coloring electroanatomical maps to indicate ultrasound data acquisiton - Google Patents

Coloring electroanatomical maps to indicate ultrasound data acquisiton Download PDF

Info

Publication number
EP1929956B1
EP1929956B1 EP07254746.6A EP07254746A EP1929956B1 EP 1929956 B1 EP1929956 B1 EP 1929956B1 EP 07254746 A EP07254746 A EP 07254746A EP 1929956 B1 EP1929956 B1 EP 1929956B1
Authority
EP
European Patent Office
Prior art keywords
dimensional
image
images
map
ultrasound
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
EP07254746.6A
Other languages
German (de)
French (fr)
Other versions
EP1929956A2 (en
EP1929956A3 (en
Inventor
Andres Claudio Altmann
Assaf Govari
Dina Kirshenbaum
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Biosense Webster Inc
Original Assignee
Biosense Webster Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Biosense Webster Inc filed Critical Biosense Webster Inc
Priority to EP11157122.0A priority Critical patent/EP2359749B1/en
Publication of EP1929956A2 publication Critical patent/EP1929956A2/en
Publication of EP1929956A3 publication Critical patent/EP1929956A3/en
Application granted granted Critical
Publication of EP1929956B1 publication Critical patent/EP1929956B1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/46Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
    • A61B8/461Displaying means of special interest
    • A61B8/466Displaying means of special interest adapted to display 3D data
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/24Detecting, measuring or recording bioelectric or biomagnetic signals of the body or parts thereof
    • A61B5/316Modalities, i.e. specific diagnostic methods
    • A61B5/318Heart-related electrical modalities, e.g. electrocardiography [ECG]
    • A61B5/339Displays specially adapted therefor
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/74Details of notification to user or communication with user or patient ; user input means
    • A61B5/742Details of notification to user or communication with user or patient ; user input means using visual displays
    • A61B5/743Displaying an image simultaneously with additional graphical information, e.g. symbols, charts, function plots
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/08Auxiliary means for directing the radiation beam to a particular spot, e.g. using light beams
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/46Arrangements for interfacing with the operator or the patient
    • A61B6/461Displaying means of special interest
    • A61B6/463Displaying means of special interest characterised by displaying multiple images or images and diagnostic data on one display
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • A61B6/5211Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
    • A61B6/5229Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • A61B6/5211Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
    • A61B6/5229Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
    • A61B6/5235Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from the same or different ionising radiation imaging techniques, e.g. PET and CT
    • A61B6/5241Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from the same or different ionising radiation imaging techniques, e.g. PET and CT combining overlapping images of the same imaging modality, e.g. by stitching
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/12Diagnosis using ultrasonic, sonic or infrasonic waves in body cavities or body tracts, e.g. by using catheters
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/13Tomography
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/44Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
    • A61B8/4483Constructional features of the ultrasonic, sonic or infrasonic diagnostic device characterised by features of the ultrasound transducer
    • A61B8/4488Constructional features of the ultrasonic, sonic or infrasonic diagnostic device characterised by features of the ultrasound transducer the transducer being a phased array
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/46Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
    • A61B8/461Displaying means of special interest
    • A61B8/463Displaying means of special interest characterised by displaying multiple images or images and diagnostic data on one display
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/48Diagnostic techniques
    • A61B8/483Diagnostic techniques involving the acquisition of a 3D volume of data
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/52Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/5215Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
    • A61B8/5238Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for combining image data of patient, e.g. merging several images from different acquisition modes into one image
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/101Computer-aided simulation of surgical operations
    • A61B2034/105Modelling of the patient, e.g. for ligaments or bones
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • A61B2034/2046Tracking techniques
    • A61B2034/2051Electromagnetic tracking systems
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B2090/364Correlation of different images or relation of image positions in respect to the body
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B2090/364Correlation of different images or relation of image positions in respect to the body
    • A61B2090/365Correlation of different images or relation of image positions in respect to the body augmented reality, i.e. correlating a live optical image with another image
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B2090/364Correlation of different images or relation of image positions in respect to the body
    • A61B2090/367Correlation of different images or relation of image positions in respect to the body creating a 3D dataset from 2D images using position information
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B90/00Instruments, implements or accessories specially adapted for surgery or diagnosis and not covered by any of the groups A61B1/00 - A61B50/00, e.g. for luxation treatment or for protecting wound edges
    • A61B90/36Image-producing devices or illumination devices not otherwise provided for
    • A61B90/37Surgical systems with images on a monitor during operation
    • A61B2090/378Surgical systems with images on a monitor during operation using ultrasound
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/72Signal processing specially adapted for physiological signals or for diagnostic purposes
    • A61B5/7271Specific aspects of physiological measurement analysis
    • A61B5/7285Specific aspects of physiological measurement analysis for synchronising or triggering a physiological measurement or image acquisition with a physiological event or waveform, e.g. an ECG signal
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B8/00Diagnosis using ultrasonic, sonic or infrasonic waves
    • A61B8/54Control of the diagnostic device
    • A61B8/543Control of the diagnostic device involving acquisition triggered by a physiological signal

Definitions

  • This invention relates generally to mapping and reconstruction of body organs. More particularly, this invention relates to imaging internal body organs, such as the heart.
  • Ultrasound imaging is now well established as a modality for imaging the heart.
  • U.S. Patent No. 6,066,096 whose disclosure is incorporated herein by reference, describes an imaging probe for volumetric intraluminal ultrasound imaging.
  • the probe configured to be placed inside a patient's body, includes an elongated body having proximal and distal ends.
  • An ultrasonic transducer phased array is connected to and positioned on the distal end of the elongated body.
  • the ultrasonic transducer phased array is positioned to emit and receive ultrasonic energy for volumetric forward scanning from the distal end of the elongated body.
  • the ultrasonic transducer phased array includes a plurality of sites occupied by ultrasonic transducer elements.
  • EP-A-1 720 038 discloses a system and method for imaging a target in a patient's body that uses a pre-acquired image of the target and a catheter having a position sensor and an ultrasonic imaging sensor.
  • the catheter is placed in the patient's body and positional information of a portion of the catheter in the patient's body is determined using the position sensor.
  • the catheter is used to generate an ultrasonic image of the target using the ultrasonic imaging sensor.
  • An image processor is used for determining positional information for any pixel of the ultrasonic image of the target and registering the pre-acquired image with the ultrasonic image; and a display is used for displaying the registered pre-acquired image and ultrasonic image.
  • WO 99/60921 A discloses a method and apparatus for three-dimensional imaging and treatment of a patient's body.
  • the method and apparatus utilize a system for developing a therapy plan for treatment of an organ of the patient, a device for generating ultrasound image data from a treatment region, a device for providing a translucent volume image of a portion of a patient's body and a separate translucent image of the patient organ, and a three-dimensional viewing device to superimpose a translucent article image to enable viewing of the article image simultaneously with the patient organ and a portion of the patient's body.
  • US2003231789 A1 discloses an apparatus for producing images of a living subject comprising: a display; a memory for storing a three-dimensional topological map of a structure in a body of said subject; a two-dimensional imaging module operative for acquiring a sequence of two-dimensional anatomic images of at least a portion of said structure, said two-dimensional anatomic images having respective image planes; and a processor linked to said memory and to said two-dimensional imaging module, said processor operative for marking said three-dimensional topological map on said display to show an intersection of said image plane with said topological map.
  • a three-dimensional image of the heart may be rendered during acquisition.
  • this blocks the user's view of the heart chamber by other tissue reflection, e.g., from other chambers or organs. Therefore, it is difficult for the user to determine when adequate image data have been acquired or whether details are still missing.
  • a three-dimensional representation of the structure such as a functional map, e.g., an electroanatomical map
  • a functional map e.g., an electroanatomical map
  • pseudocolor e.g., the planes of intersection of successive ultrasound two-dimensional fans that are acquired
  • This display enables the operator to determine regions where sufficient ultrasound data have been captured, and guides the operator to areas of the heart chamber where additional data collection is still needed.
  • Various color schemes are used to indicate the relative sufficiency of data collection.
  • An example of the disclosure provides a computer-assisted method for producing images of a living subject, which is carried out by displaying a three-dimensional model of a surface of a structure in a body of the subject, acquiring a sequence of two-dimensional anatomic images of at least a portion of the structure, and while acquiring the sequence, marking the three-dimensional model to show respective intersections of the image planes with the surface.
  • the three-dimensional model may be a computed tomographic image or a magnetic resonance image, which is automatically registered with the image planes.
  • Another aspect of the method includes displaying the three-dimensional model and the respective intersections of the image planes with the surface on the three-dimensional model.
  • a pseudocolor is displayed on the respective intersections of the image planes with the surface.
  • Yet another aspect of the method includes interpolating areas of the three-dimensional model between the respective intersections, marking the interpolated areas, and displaying the interpolated areas.
  • Another aspect of the method includes reconstructing a three-dimensional anatomic image of the structure from the two-dimensional anatomic images, and displaying at least a portion of the three-dimensional anatomic image with the three-dimensional model.
  • the displayed portion of the three-dimensional anatomic image does not extend beyond a predefined distance from a surface of the three-dimensional model.
  • the structure is a heart and the three-dimensional model is an anatomical map.
  • the two-dimensional anatomic images can be acquired by realtime three-dimensional ultrasound imaging, realtime computed tomographic imaging, or realtime magnetic resonance imaging.
  • the invention provides an apparatus for carrying out the above-described method.
  • FIG. 1 is an illustration of a system 20 for imaging and generating electrical activation maps of a heart 24 of a patient, and which is suitable for performing diagnostic or therapeutic procedures involving the heart 24, in accordance with an embodiment of the present invention.
  • the system 20 comprises a catheter 28, which is percutaneously inserted by a physician into a chamber or vascular structure of the heart.
  • the catheter 28 typically comprises a handle 29 for operation of the catheter by the physician. Suitable controls on the handle 29 enable the physician to steer, position and orient the distal end of the catheter as desired.
  • the system 20 enables the physician to perform a variety of mapping and imaging procedures. These procedures comprise, for example, the following techniques, which are described in further detail in copending, commonly assigned Application Nos. 11/115,002 and 11/262,217 , the disclosures of which are herein incorporated by reference:
  • a parametric map such as an electrophysiological information map or an electro-anatomical map on the reconstructed three-dimensional model
  • the system 20 comprises a positioning subsystem that measures three-dimensional location information and orientation coordinates of the catheter 28 with up to six degrees of freedom.
  • the positioning subsystem may comprise a magnetic position tracking system that determines the position and orientation of the catheter 28.
  • the positioning subsystem generates magnetic fields in a predefined working volume its vicinity and senses these fields at the catheter.
  • the positioning subsystem typically comprises a set of external radiators, such as field generating coils 30, which are located in fixed, known positions external to the patient.
  • the coils 30 generate fields, typically electromagnetic fields, in the vicinity of the heart 24.
  • a radiator in the catheter such as a coil, generates electromagnetic fields, which are received by sensors (not shown) outside the patient's body.
  • the position sensor transmits, in response to the sensed fields, position-related electrical signals over cables 33 running through the catheter to a console 34. Alternatively, the position sensor may transmit signals to the console 34 over a wireless link.
  • the console 34 comprises a positioning processor 36 that calculates the location and orientation of the catheter 28 based on the signals sent by a location sensor 46.
  • the positioning processor 36 typically receives, amplifies, filters, digitizes, and otherwise processes signals from the catheter 28. Images produced by the system 20 are displayed on a monitor 44.
  • the system 20 may employ the catheters disclosed in U.S. Patent Nos. 6,716,166 and 6,773,402 , whose disclosures are herein incorporated by reference, in order to acquire ultrasound images for display in near realtime ultrasound images concurrently with an image or representation of the position of a deployment catheter in the same or different sessions, and in many different combinations.
  • catheters have acoustic transducers that are adapted for emitting sound waves, and receiving reflections from echogenic interfaces in the heart. The reflections are then analyzed to construct two-dimensional and three-dimensional images of the heart.
  • the system 20 comprises an ultrasound driver 39 that drives the ultrasound transducers of the catheter 28 when it functions as an ultrasound imaging catheter.
  • a suitable ultrasound driver that can be used for this purpose is an AN2300TM ultrasound system produced by Analogic Corporation, 8 Centennial Drive, Peabody, MA 01960.
  • the ultrasound driver 39 may support different imaging modes such as B-mode, M-mode, CW Doppler and color flow Doppler, as are known in the art.
  • the catheter 28 and another catheter 48 are both incorporated in the system 20 and inserted concurrently into the heart via different vascular approaches.
  • the catheter 28 functions as a mapping catheter
  • the catheter 48 functions as an ultrasound imaging catheter, using an array of acoustic transducers 50.
  • Each has an instance of the location sensor 46 that is used to determine the position and orientation of the catheter within the body.
  • the system 20 contains electronic circuitry for generation of an electrical activation map, and can be used in conjunction with many specialized mapping catheters.
  • a suitable mapping catheter for use as the catheter 28 is described in commonly assigned U.S. Patent No. 6,892,091 , whose disclosure is herein incorporated by reference.
  • the distal end of the mapping catheter includes a distally placed mapping electrode 52 for measuring the electrical properties of the heart tissue.
  • the distal end of the mapping catheter further also includes an array of non-contact electrodes 54 for measuring far field electrical signals in the heart chamber.
  • mapping catheter is introduced first, and an electrical activation map generated from its data. Afterward, an ultrasound imaging catheter is introduced.
  • the two catheters may be introduced via the same or different vascular approaches.
  • a hybrid catheter capable of both data acquisition suitable for electrical activation map generation, and also having ultrasound imaging functions can be used.
  • catheters are described, for example, in U.S. Patents Nos. 6,773,402 , 6,788,967 , and 6,645,145 . Use of such catheters may permit the medical procedure to be shortened. In this alternative, only one catheter need be inserted.
  • the electrical activation map is usually acquired first, and then applied to the ultrasound images to assist in the interpretation of the latter. Suitable image registration techniques for coordinating the two modalities are disclosed in U.S. Patent No. 6,650,927 and in co-pending Application No. 11/215,435 , both of common assignee herewith, and herein incorporated by reference.
  • Fig. 2 is a block diagram illustrating further details of the system 20 ( Fig. 1 ).
  • many elements of the system 20 can be realized as a general purpose or specialized computer that includes a processor and a memory that contains objects corresponding to the functional blocks depicted in Fig. 2 .
  • the positioning processor 36 is linked to location sensors that are placed near the distal tip of the cardiac catheter and performs location tracking.
  • the ultrasound driver 39 which drives the transducers 50 ( Fig. 1 ) is cooperative with ultrasound circuitry 56, and produces two-dimensional ultrasound images.
  • An image processor 60 is linked to the mapping circuitry 58, the positioning processor 36, and the ultrasound circuitry 56.
  • the image processor 60 can perform three-dimensional ultrasound image reconstruction, and is specialized for the automatic identification of cardiac topological features on the ultrasound images. In some embodiments, the image processor 60 may augment automatic identification of topologic features on the electrical activation map by the mapping circuitry 58, without operator assistance.
  • the image processor 60 also performs image registration functions. Its operation is mediated via a user input 62. Its output is sent to a display 64.
  • a commercial unit suitable for use in the system 20, which is capable of generating an electrical activation map, is the CARTO XP EP Navigation and Ablation System, available from Biosense Webster, Inc., 3333 Diamond Canyon Road, Diamond Bar, CA 91765. Images acquired using different modalities can be registered for display using the CartoMergeTM image integration module, which is adapted for operation with the CARTO XP EP Navigation and Ablation System. In particular, it is possible to register a three-dimensional anatomical map or electroanatomical map with a three-dimensional ultrasound image with this module. Furthermore, the ultrasound fan image produced by two-dimensional ultrasound imaging shares the coordinate system of the anatomical or electroanatomical map. The system is able to automatically compute the intersection of the fan image and the three-dimensional image, as well as interpolate between adjacent intersections of different fan images.
  • FIG. 3 is a flow chart of a general method of marking a three-dimensional model of an internal structure of the body to indicate progress in acquiring a plurality of two-dimensional images of the structure in accordance with a disclosed embodiment of the invention.
  • a three-dimensional model of the structure is acquired and displayed.
  • This can be an image of the heart, obtained with a system such as the above-noted CARTO XP EP Navigation and Ablation System.
  • any three-dimensional model can be used, for example a tomographic image. It is important to display the topography of the heart or other structure, and the functional data, for example electrical potentials that may be shown on the model are incidental.
  • a two-dimensional image of a portion of the structure is acquired.
  • This may be an ultrasound image.
  • the two-dimensional image could be a two-dimensional functional image, obtained by techniques such as magnetic resonance imaging or computed tomographic imaging.
  • step 84 the two-dimensional image acquired in step 82 is automatically registered or otherwise coordinated with the three-dimensional model produced in initial step 80. This step allows topographic features of the three-dimensional model to be related to the structures imaged in step 82.
  • step 86 the intersection of the plane of the two-dimensional image with the three-dimensional model is marked on the display.
  • This step can be performed by applying a pseudocolor to the display.
  • many other graphical techniques can be used to indicate the intersection, e.g., flashing effects, bolding emphasis.
  • pseudocolor may be applied in order to display areas on the three-dimensional model located between adjacent intersections of different fan images. Such areas are identified by interpolation.
  • the operator can identify topographical features of the structure that were obtained on the current two-dimensional image by reference to the display and the markings on the three-dimensional model.
  • the operator may annotate the display by textual descriptive information relating to the current two-dimensional image.
  • control proceeds to final step 90, and the procedure ends.
  • Fig. 4 is a detailed flow chart of a method of coloring an electro-anatomical map or other functional map to indicate ultrasound data acquisition in accordance with an alternate embodiment.
  • coloring also referred to herein as the application of pseudocolor, denotes a computing task and involves modifications to memory in which image data is stored. The results of the operation may be visualized on a computer monitor as a colored display.
  • the method is discussed with reference to an electroanatomical map by way of example. However, the method is applicable to other functional images of the heart, so long as the topology of the heart is shown and can be related to the location of the ultrasound data.
  • a mapping catheter is introduced into a subject using well-known techniques.
  • An ultrasound imaging catheter is also introduced into the heart.
  • the mapping catheter is navigated within the heart, and electrical data obtained.
  • a functional image is generated.
  • an electroanatomical map is generated, for example, using the above-mentioned CARTO XP EP Navigation and Ablation System.
  • the image is generated using the mapping catheter by determining spatial coordinates of different locations in the heart to define a three-dimensional space.
  • a functional model is prepared, which is a three-dimensional map of the heart in the three-dimensional space, in which the map displays functional information, i.e., electrical potentials at multiple points of the heart.
  • At step 70 at least one two-dimensional ultrasound image is acquired. Generally, this is a gated image. Position information provided by location sensors on the ultrasound imaging catheter are processed by the positioning subsystem to establish coordinates of different points on the ultrasound image. Typically, the electroanatomical map and the two-dimensional ultrasound image are obtained during the same session. However, this is not necessary, and alternatively, the electroanatomical map may be pre-acquired and registered with the two-dimensional ultrasound image.
  • the area of the electro-anatomical map or other functional image corresponding to the ultrasound image acquired in the last iteration of step 70 is identified by application of pseudocolor.
  • One pseudocolor may be used, at different intensities as the sufficiency of the image improves.
  • multiple pseudocolors can be used and combined in order to indicate current image quality in many different schemes.
  • other graphical indications of image quality may be displayed in this step, for example flashing effects.
  • the relevant portion of the electroanatomical map is determined by computing the plane of intersection of the ultrasound fan image on the electro-anatomical map.
  • FIG. 5 is a display of multimodal images of the heart in accordance with a disclosed embodiment.
  • An image 92, at the left side of Fig. 5 is a topological map of a heart chamber generated by the above-noted CARTO XP EP Navigation and Ablation System.
  • the map is partly colored to show an area 96 of the chamber wall where ultrasound data have been collected.
  • the plane of intersection of each successive ultrasound two-dimensional fan that is acquired may be marked on the image 94 as a colored region on the map surface.
  • the plane of intersection may be marked as a colored line.
  • the image 94 may be colored to mark every data voxel where the ultrasound beam plane intersected the electroanatomical map.
  • the display enables the operator to see where sufficient ultrasound data have been captured and is useful to guide the operator to areas of the heart chamber where additional data collection is still needed.
  • An image 98 on the right of Fig. 5 shows a reconstruction of a three-dimensional ultrasound image 100 superimposed on the image 98, which is here referenced as an area 102.
  • the image 98 and the area 102 are based on the collected ultrasound data.
  • two-dimensional ultrasound images are projected without reconstructing a solid three-dimensional model. This technique is described in the above-noted Application Nos. 11/115,002 and 11/262,217 .
  • successive two-dimensional ultrasound images can be acquired in iterations of step 70 ( Fig. 4 ), and contours-of-interest tagged. The images can then be oriented and projected in three-dimensional space.
  • Fig. 6 shows a skeleton model 88 of the right ventricle of a heart, in accordance with a disclosed embodiment.
  • the system 20 ( Fig. 1 ) can automatically trace and reconstruct contours 90, 92 from untagged ultrasound images and can automatically reconstruct contours 94 from two-dimensional physician-labeled counterparts.
  • Fig. 7 is an exemplary composite image 96 in which a skeleton model of a three-dimensional ultrasound image 98 of the heart is superimposed on an electro-anatomical map 100 of the right ventricle, in accordance with a disclosed embodiment
  • the skeleton model is similar to the skeleton model 88 ( Fig. 6 ), having a plurality of contours 102, 104 outlining the right ventricle and left ventricle, respectively.
  • the contours 102 are overlaid on the electro-anatomical map. Different electrical potential values are indicated by different shading patterns.
  • portions of the map 100 may be automatically marked using pseudocolor to indicate adequate ultrasound data collection. For example, pseudocolor has been applied to an area 105, represented by a diagonally hatched pattern in Fig. 7 .
  • the electro-anatomical map, and optionally vessels which may be shown diagrammatically on the electroanatomical map as contours or cylindrical structures, are progressively colored to indicate the areas that were imaged, as shown on the image 94 ( Fig. 5 ).
  • the map may start with a gray color, as on the image 92 ( Fig. 5 ), and the color may then change from gray to red at every point on the map that corresponds to points where ultrasound image data were acquired. In this manner, the operator receives a clear indication of the current data coverage.
  • the ultrasound images acquired in iterations of step 70 are superimposed on the electroanatomical map, such that the two are seen in registration on a display.
  • the ultrasound catheter includes both a location sensor and an ultrasound transducer in one unit.
  • the system after appropriate calibration, can automatically correlate any point seen on the ultrasound image with its corresponding point in three-dimensional space of the electroanatomic map.
  • the image registration is typically established by correlating the coordinates during the generation of the electroanatomic map with position information and coordinates on the ultrasound image that were obtained in step 70.
  • External anatomic markers may be used to provide a common frame of reference in order to couple the data from the two modalities.
  • the ultrasound image is a three-dimensional ultrasound image that is reconstructed from multiple two-dimensional ultrasound images.
  • two-dimensional fan images are superimposed as lines on the electroanatomical map.
  • step 75 the ultrasound images and the electroanatomical map are displayed separately.
  • This option has the advantages that multimodal image registration issues are avoided in the display. Furthermore, one image is not obscured by the other.
  • step 75 at least a portion of the three-dimensional image is displayed inside the three-dimensional model, and the three-dimensional image does not extend more than a predefined distance from the surface of the three-dimensional model. The result is that a three-dimensional space is segmented according to the proportion of the three-dimensional image that is displayed. Segmentation techniques suitable for this operation are disclosed in the above-noted Application No. 11/215,435 .
  • the ultrasound image 98 and the electro-anatomical map 100 can be acquired using different equipment.
  • propagation delays between the source equipment and the processor 36 necessitate careful attention to synchronization of the two components of the composite image 96.
  • synchronization issues occur generally, in different embodiments of the system 20 ( Fig. 1 ). Solutions for this problem are taught in the above-noted Application No. 11/262,217 .
  • a constant pre-defined offset which can be a temporal offset, is established between the electroanatomical data and the anatomic image gating.
  • This offset compensates for system delays caused by image processing and image transfer from the source of the anatomic images to the image processor, which as noted above, generates an electroanatomical map from the electroanatomical data.
  • the operator may identify anatomical features and mark them on the display, using a graphical user interface.
  • the operator may start the imaging procedure with contour mapping of the left and right atria, marking relevant structures, such as the pulmonary veins, aorta and fossa ovalis.
  • the pulmonary veins and aorta can be shown as vessels with adjustable radii defined by the ultrasound contours.
  • control proceeds to final step 81.
  • the catheters are withdrawn, and the procedure ends.
  • This embodiment is similar to alternate embodiment 1, except that an inverse display mode can be used for displaying a three-dimensional image, e.g., the image 100 ( Fig. 5 ) in steps 72, 75 ( Fig. 4 ).
  • the data acquisition for the ultrasound images is essentially the same, but instead of showing high gray scale levels for tissue, the three-dimensional ultrasound image indicates the blood in the chamber or vessel, and is an indicator of the chamber or vessel blood volume.
  • a chemical sensor may be used to identify areas of the heart with low NADPH levels, indicative of ischemia. Such areas may be registered with corresponding areas observed on ultrasound images.
  • the technique described in the article Quantitative Measurements of Cardiac Phosphorus Metabolites in Coronary Artery Disease by 31P Magnetic Resonance Spectroscopy, Takahiro Yabe et al., Circulation. 1995;92:15-23 is suitable for displaying such areas.
  • step 70 is performed using a modality other than two-dimensional ultrasound imaging to acquire realtime data as a series of image "slices" through the target structure.
  • Step 70 can be performed using a realtime three-dimensional ultrasound imaging probe, realtime computed tomographic imaging, realtime magnetic resonance imaging or other realtime imaging modality from which three-dimensional images can be generated and co-displayed with a functional image to which pseudocolor is applied to indicate sufficiency of data imaging in particular areas.
  • steps 72, 75 ( Fig. 4 )
  • additional indications are shown on the map display to guide the operator during data acquisition.
  • the fill ratio, the ratio of colored area to total target area on the electroanatomical map or other functional map, can be displayed to quantitatively indicate the extent of completion of the session.
  • pseudocolor itself can be modified according to the gray scale level of each voxel using a corresponding lookup table. This enables the user to see if the acquired data corresponds to a wall tissue or to a vessel or valve opening in the chamber.

Landscapes

  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Engineering & Computer Science (AREA)
  • Medical Informatics (AREA)
  • Surgery (AREA)
  • General Health & Medical Sciences (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Biomedical Technology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Animal Behavior & Ethology (AREA)
  • Molecular Biology (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Pathology (AREA)
  • Biophysics (AREA)
  • Physics & Mathematics (AREA)
  • Radiology & Medical Imaging (AREA)
  • Optics & Photonics (AREA)
  • High Energy & Nuclear Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Robotics (AREA)
  • Gynecology & Obstetrics (AREA)
  • Cardiology (AREA)
  • Computer Graphics (AREA)
  • General Engineering & Computer Science (AREA)
  • Ultra Sonic Daignosis Equipment (AREA)
  • Magnetic Resonance Imaging Apparatus (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Measurement And Recording Of Electrical Phenomena And Electrical Characteristics Of The Living Body (AREA)

Description

    BACKGROUND OF THE INVENTION Field of the Invention
  • This invention relates generally to mapping and reconstruction of body organs. More particularly, this invention relates to imaging internal body organs, such as the heart.
  • Description of the Related Art
  • Ultrasound imaging is now well established as a modality for imaging the heart. For example, U.S. Patent No. 6,066,096 , whose disclosure is incorporated herein by reference, describes an imaging probe for volumetric intraluminal ultrasound imaging. The probe, configured to be placed inside a patient's body, includes an elongated body having proximal and distal ends. An ultrasonic transducer phased array is connected to and positioned on the distal end of the elongated body. The ultrasonic transducer phased array is positioned to emit and receive ultrasonic energy for volumetric forward scanning from the distal end of the elongated body. The ultrasonic transducer phased array includes a plurality of sites occupied by ultrasonic transducer elements.
  • However, many physicians find it difficult to interpret ultrasound images, which typically appear as a two-dimensional fan-shaped pattern. Although the physician knows what anatomical features should appear in a display produced by an ultrasound catheter, he may not be able to match these features with the bright and dark areas of the fan.
  • It has been proposed to improve medical image interpretation by superimposing images acquired by different modalities in registration. For example, U.S. Patent No. 6,556,695, issued to Packer et al. , suggests that a magnetic resonance image can be acquired, and then registered with a subsequently acquired electrical activation map or ultrasound image.
  • EP-A-1 720 038 discloses a system and method for imaging a target in a patient's body that uses a pre-acquired image of the target and a catheter having a position sensor and an ultrasonic imaging sensor. The catheter is placed in the patient's body and positional information of a portion of the catheter in the patient's body is determined using the position sensor. The catheter is used to generate an ultrasonic image of the target using the ultrasonic imaging sensor. An image processor is used for determining positional information for any pixel of the ultrasonic image of the target and registering the pre-acquired image with the ultrasonic image; and a display is used for displaying the registered pre-acquired image and ultrasonic image.
  • WO 99/60921 A discloses a method and apparatus for three-dimensional imaging and treatment of a patient's body. The method and apparatus utilize a system for developing a therapy plan for treatment of an organ of the patient, a device for generating ultrasound image data from a treatment region, a device for providing a translucent volume image of a portion of a patient's body and a separate translucent image of the patient organ, and a three-dimensional viewing device to superimpose a translucent article image to enable viewing of the article image simultaneously with the patient organ and a portion of the patient's body.
  • US2003231789 A1 discloses an apparatus for producing images of a living subject comprising: a display; a memory for storing a three-dimensional topological map of a structure in a body of said subject; a two-dimensional imaging module operative for acquiring a sequence of two-dimensional anatomic images of at least a portion of said structure, said two-dimensional anatomic images having respective image planes; and a processor linked to said memory and to said two-dimensional imaging module, said processor operative for marking said three-dimensional topological map on said display to show an intersection of said image plane with said topological map.
  • SUMMARY OF THE INVENTION AND DISCLOSURE
  • In order to assist the physician in performing a realtime cardiac imaging procedure, a three-dimensional image of the heart may be rendered during acquisition. However, this blocks the user's view of the heart chamber by other tissue reflection, e.g., from other chambers or organs. Therefore, it is difficult for the user to determine when adequate image data have been acquired or whether details are still missing.
  • According to disclosed embodiments of the invention, a three-dimensional representation of the structure, such as a functional map, e.g., an electroanatomical map, is displayed and marked, typically by application of pseudocolor, during acquisition of ultrasound data in order to show the progress of data acquisition. For example, the planes of intersection of successive ultrasound two-dimensional fans that are acquired may be marked on an electroanatomical map as lines or colored regions on the map surface. This display enables the operator to determine regions where sufficient ultrasound data have been captured, and guides the operator to areas of the heart chamber where additional data collection is still needed. Various color schemes are used to indicate the relative sufficiency of data collection.
  • An example of the disclosure provides a computer-assisted method for producing images of a living subject, which is carried out by displaying a three-dimensional model of a surface of a structure in a body of the subject, acquiring a sequence of two-dimensional anatomic images of at least a portion of the structure, and while acquiring the sequence, marking the three-dimensional model to show respective intersections of the image planes with the surface.
  • In an aspect of the method, the three-dimensional model may be a computed tomographic image or a magnetic resonance image, which is automatically registered with the image planes.
  • Another aspect of the method includes displaying the three-dimensional model and the respective intersections of the image planes with the surface on the three-dimensional model.
  • According to an additional aspect of the method, a pseudocolor is displayed on the respective intersections of the image planes with the surface.
  • Yet another aspect of the method includes interpolating areas of the three-dimensional model between the respective intersections, marking the interpolated areas, and displaying the interpolated areas.
  • Another aspect of the method includes reconstructing a three-dimensional anatomic image of the structure from the two-dimensional anatomic images, and displaying at least a portion of the three-dimensional anatomic image with the three-dimensional model.
  • According to still another aspect of the method, the displayed portion of the three-dimensional anatomic image does not extend beyond a predefined distance from a surface of the three-dimensional model.
  • According to one aspect of the method, the structure is a heart and the three-dimensional model is an anatomical map.
  • In other aspects of the method, the two-dimensional anatomic images can be acquired by realtime three-dimensional ultrasound imaging, realtime computed tomographic imaging, or realtime magnetic resonance imaging.
  • The invention provides an apparatus for carrying out the above-described method.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • For a better understanding of the present invention, reference is made to the detailed description of the invention, by way of example, which is to be read in conjunction with the following drawings, wherein like elements are given like reference numerals, and wherein:
    • Fig. 1 illustrates a system for imaging and mapping a heart of a patient in accordance with a disclosed embodiment of the invention;
    • Fig. 2 is a block diagram illustrating further details of the system shown in Fig. 1 in accordance with a disclosed embodiment of the invention;
    • Fig. 3 is a flow chart of a general method of marking a three-dimensional model of an internal structure of the body to indicate progress in acquiring a plurality of two-dimensional images of the structure in accordance with a disclosed embodiment of the invention;
    • Fig. 4 is a detailed flow chart of a method of coloring a functional map to indicate ultrasound data acquisition in accordance with an alternate embodiment of the invention;
    • Fig. 5 is a display of multimodal images of the heart in accordance with a disclosed embodiment of the invention;
    • Fig. 6 shows a skeleton model of the right ventricle of a heart, which is prepared in accordance with a disclosed embodiment of the invention; and
    • Fig. 7 is a composite image, in which a skeleton model representing of a three-dimensional ultrasound cardiac image of the heart is superimposed on an electro-anatomical map of the right ventricle, in accordance with a disclosed embodiment of the invention.
    DETAILED DESCRIPTION OF THE INVENTION
  • In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be apparent to one skilled in the art, however, that the present invention may be practiced without these specific details. In other instances, well-known circuits, control logic, and the details of computer program instructions for conventional algorithms and processes have not been shown in detail in order not to obscure the present invention unnecessarily.
  • System Overview
  • Turning now to the drawings, reference is initially made to Fig. 1, which is an illustration of a system 20 for imaging and generating electrical activation maps of a heart 24 of a patient, and which is suitable for performing diagnostic or therapeutic procedures involving the heart 24, in accordance with an embodiment of the present invention.
  • While the principles of the invention are disclosed with reference to cardiac imaging, the techniques described may be adapted for use for imaging other organs using a manually or automatically controlled probe, particularly hollow organs, such as the bladder, which may be imaged using an ultrasound catheter.
  • The system 20 comprises a catheter 28, which is percutaneously inserted by a physician into a chamber or vascular structure of the heart. The catheter 28 typically comprises a handle 29 for operation of the catheter by the physician. Suitable controls on the handle 29 enable the physician to steer, position and orient the distal end of the catheter as desired.
  • The system 20 enables the physician to perform a variety of mapping and imaging procedures. These procedures comprise, for example, the following techniques, which are described in further detail in copending, commonly assigned Application Nos. 11/115,002 and 11/262,217 , the disclosures of which are herein incorporated by reference:
  • display real-time or near real-time two-dimensional images, e.g. ultrasound images;
  • reconstruct three-dimensional models of a target structure in the patient's body, based on two-dimensional ultrasound images;
  • register, overlay and display a parametric map, such as an electrophysiological information map or an electro-anatomical map on the reconstructed three-dimensional model;
  • register, overlay and display a three-dimensional image acquired from an external system on the reconstructed three-dimensional model; and
  • register and display two-dimensional ultrasound images on a three-dimensional image acquired from an external system.
  • The system 20 comprises a positioning subsystem that measures three-dimensional location information and orientation coordinates of the catheter 28 with up to six degrees of freedom. The positioning subsystem may comprise a magnetic position tracking system that determines the position and orientation of the catheter 28. The positioning subsystem generates magnetic fields in a predefined working volume its vicinity and senses these fields at the catheter. The positioning subsystem typically comprises a set of external radiators, such as field generating coils 30, which are located in fixed, known positions external to the patient. The coils 30 generate fields, typically electromagnetic fields, in the vicinity of the heart 24.
  • In an alternative embodiment, a radiator in the catheter, such as a coil, generates electromagnetic fields, which are received by sensors (not shown) outside the patient's body.
  • The position sensor transmits, in response to the sensed fields, position-related electrical signals over cables 33 running through the catheter to a console 34. Alternatively, the position sensor may transmit signals to the console 34 over a wireless link. The console 34 comprises a positioning processor 36 that calculates the location and orientation of the catheter 28 based on the signals sent by a location sensor 46. The positioning processor 36 typically receives, amplifies, filters, digitizes, and otherwise processes signals from the catheter 28. Images produced by the system 20 are displayed on a monitor 44.
  • Some position tracking systems that may be used for this purpose are described, for example, in U.S. Patents 6,690,963 , 6,618,612 and 6, 332, 089 , and U.S. Patent Application Publications 2004/0147920 , and 2004/0068178 , whose disclosures are incorporated herein by reference. Although the positioning subsystem shown in Fig. 1 uses magnetic fields, the methods described below may be implemented using any other suitable positioning subsystem, such as systems based on acoustic or ultrasonic measurements.
  • For ultrasound image generation, the system 20 may employ the catheters disclosed in U.S. Patent Nos. 6,716,166 and 6,773,402 , whose disclosures are herein incorporated by reference, in order to acquire ultrasound images for display in near realtime ultrasound images concurrently with an image or representation of the position of a deployment catheter in the same or different sessions, and in many different combinations. Such catheters have acoustic transducers that are adapted for emitting sound waves, and receiving reflections from echogenic interfaces in the heart. The reflections are then analyzed to construct two-dimensional and three-dimensional images of the heart.
  • The system 20 comprises an ultrasound driver 39 that drives the ultrasound transducers of the catheter 28 when it functions as an ultrasound imaging catheter. One example of a suitable ultrasound driver that can be used for this purpose is an AN2300™ ultrasound system produced by Analogic Corporation, 8 Centennial Drive, Peabody, MA 01960. The ultrasound driver 39 may support different imaging modes such as B-mode, M-mode, CW Doppler and color flow Doppler, as are known in the art.
  • Optionally, the catheter 28 and another catheter 48 are both incorporated in the system 20 and inserted concurrently into the heart via different vascular approaches. In this example, the catheter 28 functions as a mapping catheter, and the catheter 48 functions as an ultrasound imaging catheter, using an array of acoustic transducers 50. Each has an instance of the location sensor 46 that is used to determine the position and orientation of the catheter within the body.
  • The system 20 contains electronic circuitry for generation of an electrical activation map, and can be used in conjunction with many specialized mapping catheters. A suitable mapping catheter for use as the catheter 28 is described in commonly assigned U.S. Patent No. 6,892,091 , whose disclosure is herein incorporated by reference. Briefly, the distal end of the mapping catheter includes a distally placed mapping electrode 52 for measuring the electrical properties of the heart tissue. The distal end of the mapping catheter further also includes an array of non-contact electrodes 54 for measuring far field electrical signals in the heart chamber.
  • Typically, the mapping catheter is introduced first, and an electrical activation map generated from its data. Afterward, an ultrasound imaging catheter is introduced. The two catheters may be introduced via the same or different vascular approaches.
  • In yet another alternative, a hybrid catheter, capable of both data acquisition suitable for electrical activation map generation, and also having ultrasound imaging functions can be used. Such catheters are described, for example, in U.S. Patents Nos. 6,773,402 , 6,788,967 , and 6,645,145 . Use of such catheters may permit the medical procedure to be shortened. In this alternative, only one catheter need be inserted. In all the alternatives, as explained in further detail below, the electrical activation map is usually acquired first, and then applied to the ultrasound images to assist in the interpretation of the latter. Suitable image registration techniques for coordinating the two modalities are disclosed in U.S. Patent No. 6,650,927 and in co-pending Application No. 11/215,435 , both of common assignee herewith, and herein incorporated by reference.
  • Reference is now made to Fig. 2, which is a block diagram illustrating further details of the system 20 (Fig. 1). As noted above, many elements of the system 20 can be realized as a general purpose or specialized computer that includes a processor and a memory that contains objects corresponding to the functional blocks depicted in Fig. 2. The positioning processor 36 is linked to location sensors that are placed near the distal tip of the cardiac catheter and performs location tracking.
  • The ultrasound driver 39, which drives the transducers 50 (Fig. 1) is cooperative with ultrasound circuitry 56, and produces two-dimensional ultrasound images.
  • An image processor 60 is linked to the mapping circuitry 58, the positioning processor 36, and the ultrasound circuitry 56. The image processor 60 can perform three-dimensional ultrasound image reconstruction, and is specialized for the automatic identification of cardiac topological features on the ultrasound images. In some embodiments, the image processor 60 may augment automatic identification of topologic features on the electrical activation map by the mapping circuitry 58, without operator assistance. The image processor 60 also performs image registration functions. Its operation is mediated via a user input 62. Its output is sent to a display 64.
  • A commercial unit suitable for use in the system 20, which is capable of generating an electrical activation map, is the CARTO XP EP Navigation and Ablation System, available from Biosense Webster, Inc., 3333 Diamond Canyon Road, Diamond Bar, CA 91765. Images acquired using different modalities can be registered for display using the CartoMerge™ image integration module, which is adapted for operation with the CARTO XP EP Navigation and Ablation System. In particular, it is possible to register a three-dimensional anatomical map or electroanatomical map with a three-dimensional ultrasound image with this module. Furthermore, the ultrasound fan image produced by two-dimensional ultrasound imaging shares the coordinate system of the anatomical or electroanatomical map. The system is able to automatically compute the intersection of the fan image and the three-dimensional image, as well as interpolate between adjacent intersections of different fan images.
  • Operation
  • Reference is now made to Fig. 3, which is a flow chart of a general method of marking a three-dimensional model of an internal structure of the body to indicate progress in acquiring a plurality of two-dimensional images of the structure in accordance with a disclosed embodiment of the invention.
  • At initial step 80, a three-dimensional model of the structure is acquired and displayed. This can be an image of the heart, obtained with a system such as the above-noted CARTO XP EP Navigation and Ablation System. However any three-dimensional model can be used, for example a tomographic image. It is important to display the topography of the heart or other structure, and the functional data, for example electrical potentials that may be shown on the model are incidental.
  • Next, at step 82 a two-dimensional image of a portion of the structure is acquired. This may be an ultrasound image. Alternatively, the two-dimensional image could be a two-dimensional functional image, obtained by techniques such as magnetic resonance imaging or computed tomographic imaging.
  • Next, at step 84, the two-dimensional image acquired in step 82 is automatically registered or otherwise coordinated with the three-dimensional model produced in initial step 80. This step allows topographic features of the three-dimensional model to be related to the structures imaged in step 82.
  • Next, at step 86, the intersection of the plane of the two-dimensional image with the three-dimensional model is marked on the display. This step can be performed by applying a pseudocolor to the display. Alternatively, many other graphical techniques can be used to indicate the intersection, e.g., flashing effects, bolding emphasis. Additionally, as explained below, pseudocolor may be applied in order to display areas on the three-dimensional model located between adjacent intersections of different fan images. Such areas are identified by interpolation. In any case, the operator can identify topographical features of the structure that were obtained on the current two-dimensional image by reference to the display and the markings on the three-dimensional model. Optionally, the operator may annotate the display by textual descriptive information relating to the current two-dimensional image.
  • Control now proceeds to decision step 88, where it is determined if more images are required to complete the imaging study. If the determination at decision step 88 is affirmative, then control returns to step 82 for another iteration.
  • If the determination at decision step 88 is negative, then control proceeds to final step 90, and the procedure ends.
  • Alternate Embodiment 1
  • Reference is now made to Fig. 4, which is a detailed flow chart of a method of coloring an electro-anatomical map or other functional map to indicate ultrasound data acquisition in accordance with an alternate embodiment. It will be understood that "coloring", also referred to herein as the application of pseudocolor, denotes a computing task and involves modifications to memory in which image data is stored. The results of the operation may be visualized on a computer monitor as a colored display. The method is discussed with reference to an electroanatomical map by way of example. However, the method is applicable to other functional images of the heart, so long as the topology of the heart is shown and can be related to the location of the ultrasound data. In initial step 66, using instrumentation described above with reference to Fig. 1 and Fig. 2, a mapping catheter is introduced into a subject using well-known techniques. An ultrasound imaging catheter is also introduced into the heart.
  • Next, at step 68, the mapping catheter is navigated within the heart, and electrical data obtained. A functional image is generated. In one embodiment, an electroanatomical map is generated, for example, using the above-mentioned CARTO XP EP Navigation and Ablation System. The image is generated using the mapping catheter by determining spatial coordinates of different locations in the heart to define a three-dimensional space. Then a functional model is prepared, which is a three-dimensional map of the heart in the three-dimensional space, in which the map displays functional information, i.e., electrical potentials at multiple points of the heart.
  • Concurrently with step 68, at step 70, at least one two-dimensional ultrasound image is acquired. Generally, this is a gated image. Position information provided by location sensors on the ultrasound imaging catheter are processed by the positioning subsystem to establish coordinates of different points on the ultrasound image. Typically, the electroanatomical map and the two-dimensional ultrasound image are obtained during the same session. However, this is not necessary, and alternatively, the electroanatomical map may be pre-acquired and registered with the two-dimensional ultrasound image.
  • Next, at step 69, the area of the electro-anatomical map or other functional image corresponding to the ultrasound image acquired in the last iteration of step 70 is identified by application of pseudocolor. One pseudocolor may be used, at different intensities as the sufficiency of the image improves. Alternatively, multiple pseudocolors can be used and combined in order to indicate current image quality in many different schemes. Additionally or alternatively, other graphical indications of image quality may be displayed in this step, for example flashing effects. In one embodiment, the relevant portion of the electroanatomical map is determined by computing the plane of intersection of the ultrasound fan image on the electro-anatomical map.
  • Reference is now made to Fig. 5, which is a display of multimodal images of the heart in accordance with a disclosed embodiment. An image 92, at the left side of Fig. 5, is a topological map of a heart chamber generated by the above-noted CARTO XP EP Navigation and Ablation System.
  • In a central image 94, the map is partly colored to show an area 96 of the chamber wall where ultrasound data have been collected. For example, the plane of intersection of each successive ultrasound two-dimensional fan that is acquired may be marked on the image 94 as a colored region on the map surface. Alternatively, the plane of intersection may be marked as a colored line. Further alternatively, the image 94 may be colored to mark every data voxel where the ultrasound beam plane intersected the electroanatomical map. In any case, the display enables the operator to see where sufficient ultrasound data have been captured and is useful to guide the operator to areas of the heart chamber where additional data collection is still needed.
  • An image 98 on the right of Fig. 5 shows a reconstruction of a three-dimensional ultrasound image 100 superimposed on the image 98, which is here referenced as an area 102. The image 98 and the area 102 are based on the collected ultrasound data.
  • In one embodiment, two-dimensional ultrasound images are projected without reconstructing a solid three-dimensional model. This technique is described in the above-noted Application Nos. 11/115,002 and 11/262,217 . For example, successive two-dimensional ultrasound images can be acquired in iterations of step 70 (Fig. 4), and contours-of-interest tagged. The images can then be oriented and projected in three-dimensional space.
  • Reference is now made to Fig. 6, which shows a skeleton model 88 of the right ventricle of a heart, in accordance with a disclosed embodiment. The system 20 (Fig. 1) can automatically trace and reconstruct contours 90, 92 from untagged ultrasound images and can automatically reconstruct contours 94 from two-dimensional physician-labeled counterparts.
  • Reference is now made to Fig. 7, which is an exemplary composite image 96 in which a skeleton model of a three-dimensional ultrasound image 98 of the heart is superimposed on an electro-anatomical map 100 of the right ventricle, in accordance with a disclosed embodiment The skeleton model is similar to the skeleton model 88 (Fig. 6), having a plurality of contours 102, 104 outlining the right ventricle and left ventricle, respectively. The contours 102 are overlaid on the electro-anatomical map. Different electrical potential values are indicated by different shading patterns. Superimposing the skeletal model on the electroanatomical map in step 72 (Fig. 4) results in less interference on the display than using a fully reproduced three-dimensional model, as can be appreciated by a comparison of Fig. 7 with the image 98 (Fig. 5). As in Fig. 5, portions of the map 100 may be automatically marked using pseudocolor to indicate adequate ultrasound data collection. For example, pseudocolor has been applied to an area 105, represented by a diagonally hatched pattern in Fig. 7.
  • Referring again to Fig. 4, as the data are acquired in successive iterations of step 70, the electro-anatomical map, and optionally vessels, which may be shown diagrammatically on the electroanatomical map as contours or cylindrical structures, are progressively colored to indicate the areas that were imaged, as shown on the image 94 (Fig. 5). For example, the map may start with a gray color, as on the image 92 (Fig. 5), and the color may then change from gray to red at every point on the map that corresponds to points where ultrasound image data were acquired. In this manner, the operator receives a clear indication of the current data coverage.
  • Next, at step 72, the ultrasound images acquired in iterations of step 70 are superimposed on the electroanatomical map, such that the two are seen in registration on a display. This is carried out automatically, using methods of synchronization, and registration of the reconstructed image with the electroanatomical map, as noted above. Briefly, the ultrasound catheter includes both a location sensor and an ultrasound transducer in one unit. The system, after appropriate calibration, can automatically correlate any point seen on the ultrasound image with its corresponding point in three-dimensional space of the electroanatomic map. The image registration is typically established by correlating the coordinates during the generation of the electroanatomic map with position information and coordinates on the ultrasound image that were obtained in step 70. External anatomic markers may be used to provide a common frame of reference in order to couple the data from the two modalities. In some embodiments, the ultrasound image is a three-dimensional ultrasound image that is reconstructed from multiple two-dimensional ultrasound images. Alternatively, two-dimensional fan images are superimposed as lines on the electroanatomical map.
  • Optionally, as shown in step 75, the ultrasound images and the electroanatomical map are displayed separately. This option has the advantages that multimodal image registration issues are avoided in the display. Furthermore, one image is not obscured by the other. In a variation of step 75, at least a portion of the three-dimensional image is displayed inside the three-dimensional model, and the three-dimensional image does not extend more than a predefined distance from the surface of the three-dimensional model. The result is that a three-dimensional space is segmented according to the proportion of the three-dimensional image that is displayed. Segmentation techniques suitable for this operation are disclosed in the above-noted Application No. 11/215,435 .
  • In either of steps 72, 75 synchronization between the two modalities required, of course. Referring again to Fig. 7, the ultrasound image 98 and the electro-anatomical map 100 can be acquired using different equipment. When one or both of the images are being tracked in near-real time, and particularly when different equipment is used for the two modalities, propagation delays between the source equipment and the processor 36 (Fig. 1) necessitate careful attention to synchronization of the two components of the composite image 96. Indeed, synchronization issues occur generally, in different embodiments of the system 20 (Fig. 1). Solutions for this problem are taught in the above-noted Application No. 11/262,217 . Briefly, when near real-time electro-anatomical data are acquired and superimposed upon previously acquired anatomic images or models, a constant pre-defined offset, which can be a temporal offset, is established between the electroanatomical data and the anatomic image gating. This offset compensates for system delays caused by image processing and image transfer from the source of the anatomic images to the image processor, which as noted above, generates an electroanatomical map from the electroanatomical data.
  • After performing either of steps 72, 75, the operator may identify anatomical features and mark them on the display, using a graphical user interface.
  • Control next proceeds to decision step 79, where it is determined if more two-dimensional ultrasound images are necessary to complete the examination. This decision is normally made by the operator, but he may be prompted by the system, which can automatically determine if the examination is complete. If the determination at decision step 79 is affirmative, then control returns to step 70. When imaging the heart, the operator may start the imaging procedure with contour mapping of the left and right atria, marking relevant structures, such as the pulmonary veins, aorta and fossa ovalis. The pulmonary veins and aorta can be shown as vessels with adjustable radii defined by the ultrasound contours.
  • If the determination at decision step 79 is negative, then control proceeds to final step 81. The catheters are withdrawn, and the procedure ends.
  • Alternate Embodiment 2
  • This embodiment is similar to alternate embodiment 1, except that an inverse display mode can be used for displaying a three-dimensional image, e.g., the image 100 (Fig. 5) in steps 72, 75 (Fig. 4). The data acquisition for the ultrasound images is essentially the same, but instead of showing high gray scale levels for tissue, the three-dimensional ultrasound image indicates the blood in the chamber or vessel, and is an indicator of the chamber or vessel blood volume.
  • Alternate Embodiment 3
  • Other physiological data that may be mapped for co-display in steps 72, 75 (Fig. 4) with ultrasound images and pseudocolor applied colored to indicate sufficiency of ultrasound data collection as described above. Volumetric intraluminal ultrasound imaging as described by the above-noted U.S. Patent No. 6,066,096 can be used. Other physiological parameters that can be mapped include temperature, blood flow rate, chemical properties and mechanical activity, e.g., regional wall motion. For example, areas of high-speed flow detected by an ultrasound catheters, as disclosed, e.g., in the above-noted U.S. Patent Nos. 6,716,166 and 6,773,402 , may be identified in a Doppler image and registered with stenoses in blood vessels observed in a three-dimensional ultrasound image. As another example, a chemical sensor may be used to identify areas of the heart with low NADPH levels, indicative of ischemia. Such areas may be registered with corresponding areas observed on ultrasound images. The technique described in the article Quantitative Measurements of Cardiac Phosphorus Metabolites in Coronary Artery Disease by 31P Magnetic Resonance Spectroscopy, Takahiro Yabe et al., Circulation. 1995;92:15-23 is suitable for displaying such areas.
  • Alternate Embodiment 4
  • In this embodiment, step 70 (Fig. 4) is performed using a modality other than two-dimensional ultrasound imaging to acquire realtime data as a series of image "slices" through the target structure. Step 70 can be performed using a realtime three-dimensional ultrasound imaging probe, realtime computed tomographic imaging, realtime magnetic resonance imaging or other realtime imaging modality from which three-dimensional images can be generated and co-displayed with a functional image to which pseudocolor is applied to indicate sufficiency of data imaging in particular areas.
  • Alternate Embodiment 5
  • This variation can be employed additionally to any of the preceding embodiments. In steps 72, 75 (Fig. 4), additional indications are shown on the map display to guide the operator during data acquisition. For example, the fill ratio, the ratio of colored area to total target area on the electroanatomical map or other functional map, can be displayed to quantitatively indicate the extent of completion of the session.
  • In additional application of pseudocolor itself can be modified according to the gray scale level of each voxel using a corresponding lookup table. This enables the user to see if the acquired data corresponds to a wall tissue or to a vessel or valve opening in the chamber.
  • It will be appreciated by persons skilled in the art that the present invention is not limited to what has been particularly shown and described hereinabove.

Claims (4)

  1. An apparatus (20) for producing images of a living subject comprising:
    a display (44);
    a memory for storing a three-dimensional topological map of a structure in a body of said subject;
    a two-dimensional imaging module operative for acquiring a sequence of two-dimensional anatomic images of at least a portion of said structure, said two-dimensional anatomic images having respective image planes; and
    a processor (60) linked to said memory and to said two-dimensional imaging module, said processor (60) operative for marking said three-dimensional topological map on said display to show successive respective intersections of said image planes with said topological map where imaging data have been collected, wherein the topological map is an electroanatomical map or other functional map.
  2. The apparatus (20) according to claim 1, wherein said two-dimensional anatomic images are selected from the group consisting of a computed tomographic image and a magnetic resonance image, and said processor (60) is operative for automatically registering said image planes with said three-dimensional topological map.
  3. The apparatus (20) according to claim 1, wherein said processor (60) is operative for:
    reconstructing a three-dimensional anatomic image (100) of said structure from said two-dimensional anatomic images; and
    displaying said three-dimensional anatomic image (100) with said three-dimensional topological map.
  4. The apparatus (20) according to claim 1, wherein said two-dimensional imaging module generates two-dimensional anatomic images selected from the group consisting of ultrasound images, computed tomographic images and magnetic resonance images, said structure is a heart and said three-dimensional topological map is an electroanatomical map.
EP07254746.6A 2006-12-08 2007-12-07 Coloring electroanatomical maps to indicate ultrasound data acquisiton Active EP1929956B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
EP11157122.0A EP2359749B1 (en) 2006-12-08 2007-12-07 Coloring electroanatomical maps to indicate ultrasound data acquisition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/608,506 US7831076B2 (en) 2006-12-08 2006-12-08 Coloring electroanatomical maps to indicate ultrasound data acquisition

Related Child Applications (2)

Application Number Title Priority Date Filing Date
EP11157122.0A Division EP2359749B1 (en) 2006-12-08 2007-12-07 Coloring electroanatomical maps to indicate ultrasound data acquisition
EP11157122.0A Division-Into EP2359749B1 (en) 2006-12-08 2007-12-07 Coloring electroanatomical maps to indicate ultrasound data acquisition

Publications (3)

Publication Number Publication Date
EP1929956A2 EP1929956A2 (en) 2008-06-11
EP1929956A3 EP1929956A3 (en) 2008-09-10
EP1929956B1 true EP1929956B1 (en) 2016-06-29

Family

ID=39256380

Family Applications (2)

Application Number Title Priority Date Filing Date
EP07254746.6A Active EP1929956B1 (en) 2006-12-08 2007-12-07 Coloring electroanatomical maps to indicate ultrasound data acquisiton
EP11157122.0A Active EP2359749B1 (en) 2006-12-08 2007-12-07 Coloring electroanatomical maps to indicate ultrasound data acquisition

Family Applications After (1)

Application Number Title Priority Date Filing Date
EP11157122.0A Active EP2359749B1 (en) 2006-12-08 2007-12-07 Coloring electroanatomical maps to indicate ultrasound data acquisition

Country Status (10)

Country Link
US (1) US7831076B2 (en)
EP (2) EP1929956B1 (en)
JP (1) JP5460955B2 (en)
KR (1) KR20080053224A (en)
CN (1) CN101219061B (en)
AU (1) AU2007237321B2 (en)
BR (1) BRPI0704524A (en)
CA (1) CA2614033C (en)
IL (1) IL187783A (en)
MX (1) MX2007015573A (en)

Families Citing this family (138)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8784336B2 (en) 2005-08-24 2014-07-22 C. R. Bard, Inc. Stylet apparatuses and methods of manufacture
US8075486B2 (en) * 2006-05-03 2011-12-13 Biosense Webster, Inc. Enhanced ultrasound image display
US8187260B1 (en) * 2006-12-29 2012-05-29 Endocare, Inc. Variable cryosurgical probe planning system
US7925068B2 (en) * 2007-02-01 2011-04-12 General Electric Company Method and apparatus for forming a guide image for an ultrasound image scanner
US9375164B2 (en) 2007-03-08 2016-06-28 Sync-Rx, Ltd. Co-use of endoluminal data and extraluminal imaging
EP2129284A4 (en) 2007-03-08 2012-11-28 Sync Rx Ltd Imaging and tools for use with moving organs
US11064964B2 (en) 2007-03-08 2021-07-20 Sync-Rx, Ltd Determining a characteristic of a lumen by measuring velocity of a contrast agent
US10716528B2 (en) 2007-03-08 2020-07-21 Sync-Rx, Ltd. Automatic display of previously-acquired endoluminal images
US8542900B2 (en) 2007-03-08 2013-09-24 Sync-Rx Ltd. Automatic reduction of interfering elements from an image stream of a moving organ
US8700130B2 (en) 2007-03-08 2014-04-15 Sync-Rx, Ltd. Stepwise advancement of a medical tool
US11197651B2 (en) 2007-03-08 2021-12-14 Sync-Rx, Ltd. Identification and presentation of device-to-vessel relative motion
US9629571B2 (en) 2007-03-08 2017-04-25 Sync-Rx, Ltd. Co-use of endoluminal data and extraluminal imaging
US9968256B2 (en) 2007-03-08 2018-05-15 Sync-Rx Ltd. Automatic identification of a tool
WO2012176191A1 (en) 2011-06-23 2012-12-27 Sync-Rx, Ltd. Luminal background cleaning
US20090088628A1 (en) * 2007-09-27 2009-04-02 Klaus Klingenbeck-Regn Efficient workflow for afib treatment in the ep lab
US9649048B2 (en) 2007-11-26 2017-05-16 C. R. Bard, Inc. Systems and methods for breaching a sterile field for intravascular placement of a catheter
US10449330B2 (en) 2007-11-26 2019-10-22 C. R. Bard, Inc. Magnetic element-equipped needle assemblies
US10751509B2 (en) 2007-11-26 2020-08-25 C. R. Bard, Inc. Iconic representations for guidance of an indwelling medical device
EP3202318B1 (en) 2007-11-26 2020-10-21 C.R. Bard, Inc. Integrated system for intravascular placement of a catheter
US8781555B2 (en) 2007-11-26 2014-07-15 C. R. Bard, Inc. System for placement of a catheter including a signal-generating stylet
US9521961B2 (en) 2007-11-26 2016-12-20 C. R. Bard, Inc. Systems and methods for guiding a medical instrument
US10524691B2 (en) 2007-11-26 2020-01-07 C. R. Bard, Inc. Needle assembly including an aligned magnetic element
US9575140B2 (en) 2008-04-03 2017-02-21 Covidien Lp Magnetic interference detection system and method
US8218846B2 (en) 2008-05-15 2012-07-10 Superdimension, Ltd. Automatic pathway and waypoint generation and navigation method
WO2009147671A1 (en) 2008-06-03 2009-12-10 Superdimension Ltd. Feature-based registration method
US8218847B2 (en) 2008-06-06 2012-07-10 Superdimension, Ltd. Hybrid registration method
WO2010022370A1 (en) 2008-08-22 2010-02-25 C.R. Bard, Inc. Catheter assembly including ecg sensor and magnetic assemblies
US20100061611A1 (en) * 2008-09-11 2010-03-11 Siemens Corporate Research, Inc. Co-registration of coronary artery computed tomography and fluoroscopic sequence
KR100978476B1 (en) * 2008-11-03 2010-08-30 주식회사 메디슨 Ultrasound system and methdo for processing volume data
US9101286B2 (en) 2008-11-18 2015-08-11 Sync-Rx, Ltd. Apparatus and methods for determining a dimension of a portion of a stack of endoluminal data points
US9095313B2 (en) 2008-11-18 2015-08-04 Sync-Rx, Ltd. Accounting for non-uniform longitudinal motion during movement of an endoluminal imaging probe
US10362962B2 (en) 2008-11-18 2019-07-30 Synx-Rx, Ltd. Accounting for skipped imaging locations during movement of an endoluminal imaging probe
US8855744B2 (en) 2008-11-18 2014-10-07 Sync-Rx, Ltd. Displaying a device within an endoluminal image stack
US11064903B2 (en) 2008-11-18 2021-07-20 Sync-Rx, Ltd Apparatus and methods for mapping a sequence of images to a roadmap image
US9974509B2 (en) 2008-11-18 2018-05-22 Sync-Rx Ltd. Image super enhancement
US9144394B2 (en) 2008-11-18 2015-09-29 Sync-Rx, Ltd. Apparatus and methods for determining a plurality of local calibration factors for an image
US20100168557A1 (en) * 2008-12-30 2010-07-01 Deno D Curtis Multi-electrode ablation sensing catheter and system
US8900150B2 (en) 2008-12-30 2014-12-02 St. Jude Medical, Atrial Fibrillation Division, Inc. Intracardiac imaging system utilizing a multipurpose catheter
US8948476B2 (en) * 2010-12-20 2015-02-03 St. Jude Medical, Atrial Fibrillation Division, Inc. Determination of cardiac geometry responsive to doppler based imaging of blood flow characteristics
US9610118B2 (en) 2008-12-31 2017-04-04 St. Jude Medical, Atrial Fibrillation Division, Inc. Method and apparatus for the cancellation of motion artifacts in medical interventional navigation
US20100249589A1 (en) * 2009-03-25 2010-09-30 Peter Lysyansky System and method for functional ultrasound imaging
CN102365653B (en) * 2009-03-27 2015-02-25 皇家飞利浦电子股份有限公司 Improvements to medical imaging
EP3427687A1 (en) * 2009-05-14 2019-01-16 Covidien LP Automatic registration technique
US9532724B2 (en) 2009-06-12 2017-01-03 Bard Access Systems, Inc. Apparatus and method for catheter navigation using endovascular energy mapping
KR101121353B1 (en) 2009-08-03 2012-03-09 한국과학기술원 System and method for providing 2-dimensional ct image corresponding to 2-dimensional ultrasound image
WO2011019760A2 (en) 2009-08-10 2011-02-17 Romedex International Srl Devices and methods for endovascular electrography
KR101120744B1 (en) * 2009-08-10 2012-03-23 삼성메디슨 주식회사 Ultrasound system and method for performing vessel labeling
US10835207B2 (en) 2009-12-23 2020-11-17 Biosense Webster (Israel) Ltd. Fast anatomical mapping using ultrasound images
AU2011210257B2 (en) 2010-02-01 2013-12-19 Covidien Lp Region-growing algorithm
CN102821679B (en) 2010-02-02 2016-04-27 C·R·巴德股份有限公司 For the apparatus and method that catheter navigation and end are located
EP2912999B1 (en) 2010-05-28 2022-06-29 C. R. Bard, Inc. Apparatus for use with needle insertion guidance system
ES2864665T3 (en) 2010-05-28 2021-10-14 Bard Inc C R Apparatus for use with needle insertion guidance system
WO2012012664A2 (en) * 2010-07-21 2012-01-26 Moehrle Armin E Image reporting method
KR101856267B1 (en) 2010-08-20 2018-05-09 씨. 알. 바드, 인크. Reconfirmation of ecg-assisted catheter tip placement
US9717478B2 (en) * 2010-12-27 2017-08-01 St. Jude Medical, Atrial Fibrillation Division, Inc. Refinement of an anatomical model using ultrasound
US20120165671A1 (en) * 2010-12-27 2012-06-28 Hill Anthony D Identification of objects in ultrasound
US9218687B2 (en) 2010-12-30 2015-12-22 St. Jude Medical, Atrial Fibrillation Division, Inc. Display of medical device position information in a volumetric rendering
US20120172724A1 (en) * 2010-12-31 2012-07-05 Hill Anthony D Automatic identification of intracardiac devices and structures in an intracardiac echo catheter image
US9002442B2 (en) 2011-01-13 2015-04-07 Rhythmia Medical, Inc. Beat alignment and selection for cardiac mapping
US8948837B2 (en) 2011-01-13 2015-02-03 Rhythmia Medical, Inc. Electroanatomical mapping
US10039502B2 (en) 2011-04-12 2018-08-07 Medtronic Ablation Frontiers Llc Electrophysiological signal processing and utilization
US9713435B2 (en) * 2011-07-27 2017-07-25 Biosense Webster (Israel) Ltd. Cardiac mapping using non-gated MRI
US8620417B2 (en) * 2011-09-22 2013-12-31 Biosense Webster (Israel), Ltd. Graphic user interface for physical parameter mapping
CN102622775B (en) * 2012-01-12 2016-03-09 北京理工大学 A kind of real-time dynamic reconstruction technology of heart compensated based on model interpolation
EP2822472B1 (en) 2012-03-07 2022-09-28 Ziteo, Inc. Systems for tracking and guiding sensors and instruments
US20130241929A1 (en) * 2012-03-13 2013-09-19 Fady Massarwa Selectably transparent electrophysiology map
EP2863802B1 (en) 2012-06-26 2020-11-04 Sync-RX, Ltd. Flow-related image processing in luminal organs
US9091628B2 (en) 2012-12-21 2015-07-28 L-3 Communications Security And Detection Systems, Inc. 3D mapping with two orthogonal imaging views
US8903149B1 (en) * 2013-01-18 2014-12-02 Anritsu Company System and method of communicating information about an object concealed by a scanned surface
US9916691B2 (en) 2013-02-14 2018-03-13 Seiko Epson Corporation Head mounted display and control method for head mounted display
EP2992513B1 (en) 2013-05-02 2022-08-10 Smith & Nephew, Inc. Surface and image integration for model evaluation and landmark determination
KR102135153B1 (en) 2013-05-07 2020-07-20 삼성전자주식회사 Ultrasonic imaging apparatus and image display method thereof
EP2807978A1 (en) 2013-05-28 2014-12-03 Universität Bern Method and system for 3D acquisition of ultrasound images
FR3008806B1 (en) * 2013-07-22 2017-07-07 Centre Nat De La Rech Scient - Cnrs - METHOD AND DEVICE FOR ACOUSTOELECTRIC IMAGING
US9775578B2 (en) * 2013-08-12 2017-10-03 Biosense Webster (Israel) Ltd. Unmapped region visualization
US11324419B2 (en) * 2013-08-20 2022-05-10 Biosense Webster (Israel) Ltd. Graphical user interface for medical imaging system
CN105658164B (en) 2013-09-06 2018-08-21 柯惠有限合伙公司 microwave ablation catheter, handle and system
US10201265B2 (en) 2013-09-06 2019-02-12 Covidien Lp Microwave ablation catheter, handle, and system
US10448862B2 (en) 2013-09-06 2019-10-22 Covidien Lp System and method for light based lung visualization
US10098566B2 (en) 2013-09-06 2018-10-16 Covidien Lp System and method for lung visualization using ultrasound
EP3073910B1 (en) 2014-02-06 2020-07-15 C.R. Bard, Inc. Systems for guidance and placement of an intravascular device
US9754372B2 (en) * 2014-08-15 2017-09-05 Biosense Webster (Israel) Ltd. Marking of fluoroscope field-of-view
US10617401B2 (en) 2014-11-14 2020-04-14 Ziteo, Inc. Systems for localization of targets inside a body
EP3220828B1 (en) 2014-11-18 2021-12-22 C.R. Bard, Inc. Ultrasound imaging system having automatic image presentation
CN107106124B (en) 2014-11-18 2021-01-08 C·R·巴德公司 Ultrasound imaging system with automatic image rendering
US10973584B2 (en) 2015-01-19 2021-04-13 Bard Access Systems, Inc. Device and method for vascular access
CA2978378A1 (en) * 2015-03-09 2016-09-15 Planmeca Oy Tracking a motion of a jaw
US9433363B1 (en) * 2015-06-18 2016-09-06 Genetesis Llc Method and system for high throughput evaluation of functional cardiac electrophysiology
US10349890B2 (en) 2015-06-26 2019-07-16 C. R. Bard, Inc. Connector interface for ECG-based catheter positioning system
WO2017053921A1 (en) 2015-09-26 2017-03-30 Boston Scientific Scimed Inc. Intracardiac egm signals for beat matching and acceptance
WO2017102369A1 (en) * 2015-12-16 2017-06-22 Koninklijke Philips N.V. Interventional device recognition
US11000207B2 (en) 2016-01-29 2021-05-11 C. R. Bard, Inc. Multiple coil system for tracking a medical device
US10327624B2 (en) * 2016-03-11 2019-06-25 Sony Corporation System and method for image processing to generate three-dimensional (3D) view of an anatomical portion
US20170307755A1 (en) 2016-04-20 2017-10-26 YoR Labs Method and System for Determining Signal Direction
CN106097428B (en) * 2016-06-03 2022-07-12 青岛海信医疗设备股份有限公司 Method and device for labeling three-dimensional model measurement information
US10376221B2 (en) * 2016-07-06 2019-08-13 Biosense Webster (Israel) Ltd. Automatic creation of multiple electroanatomic maps
US10905402B2 (en) 2016-07-27 2021-02-02 Canon Medical Systems Corporation Diagnostic guidance systems and methods
US10984533B2 (en) * 2016-10-25 2021-04-20 Koninklijke Philips N.V. Method and apparatus for segmenting a two-dimensional image of an anatomical structure
US10792106B2 (en) 2016-10-28 2020-10-06 Covidien Lp System for calibrating an electromagnetic navigation system
US10722311B2 (en) 2016-10-28 2020-07-28 Covidien Lp System and method for identifying a location and/or an orientation of an electromagnetic sensor based on a map
US10751126B2 (en) 2016-10-28 2020-08-25 Covidien Lp System and method for generating a map for electromagnetic navigation
US10517505B2 (en) 2016-10-28 2019-12-31 Covidien Lp Systems, methods, and computer-readable media for optimizing an electromagnetic navigation system
US10446931B2 (en) 2016-10-28 2019-10-15 Covidien Lp Electromagnetic navigation antenna assembly and electromagnetic navigation system including the same
US10638952B2 (en) 2016-10-28 2020-05-05 Covidien Lp Methods, systems, and computer-readable media for calibrating an electromagnetic navigation system
US10418705B2 (en) 2016-10-28 2019-09-17 Covidien Lp Electromagnetic navigation antenna assembly and electromagnetic navigation system including the same
US10615500B2 (en) 2016-10-28 2020-04-07 Covidien Lp System and method for designing electromagnetic navigation antenna assemblies
US10814128B2 (en) 2016-11-21 2020-10-27 Covidien Lp Electroporation catheter
US11129574B2 (en) * 2016-12-12 2021-09-28 Biosense Webster (Israel) Ltd. Real time electroanatomical coloring of the heart
CN106725598B (en) * 2016-12-28 2023-09-12 苏州科技城医院 Heart ultrasonic system based on multiple percutaneous ultrasonic transducers and imaging method
US11730395B2 (en) 2017-01-12 2023-08-22 Navix International Limited Reconstruction of an anatomical structure from intrabody measurements
US11311204B2 (en) 2017-01-12 2022-04-26 Navix International Limited Systems and methods for reconstruction of intrabody electrical readings to anatomical structure
CN110573089A (en) * 2017-03-31 2019-12-13 皇家飞利浦有限公司 Intravascular flow and pressure measurements
EP3629911A4 (en) 2017-05-22 2021-01-20 Genetesis LLC Machine differentiation of abnormalities in bioelectromagnetic fields
US10499879B2 (en) 2017-05-31 2019-12-10 General Electric Company Systems and methods for displaying intersections on ultrasound images
US20180360342A1 (en) 2017-06-16 2018-12-20 Biosense Webster (Israel) Ltd. Renal ablation and visualization system and method with composite anatomical display image
US11134877B2 (en) 2017-08-09 2021-10-05 Genetesis, Inc. Biomagnetic detection
CN111050641B (en) * 2017-08-17 2023-06-09 纳维斯国际有限公司 Remote imaging based on field gradients
US11006852B2 (en) 2017-12-11 2021-05-18 Covidien Lp Systems, methods, and computer-readable media of estimating thoracic cavity movement during respiration
US11109792B2 (en) 2018-01-31 2021-09-07 Biosense Webster (Israel) Ltd. Signal color morphology
US10665338B2 (en) * 2018-02-22 2020-05-26 Biosense Webster (Israel) Ltd. Automatic identification of multiple activation pathways
US10682098B2 (en) * 2018-03-22 2020-06-16 Shenzhen Mindray Bio-Medical Electronics Co., Ltd. Predictive use of quantitative imaging
US20190388060A1 (en) * 2018-06-22 2019-12-26 General Electric Company Imaging system and method with live examination completeness monitor
US11941754B2 (en) * 2018-09-12 2024-03-26 St. Jude Medical, Cardiology Division, Inc. System and method for generating three dimensional geometric models of anatomical regions
US10992079B2 (en) 2018-10-16 2021-04-27 Bard Access Systems, Inc. Safety-equipped connection systems and methods thereof for establishing electrical connections
IL272254B2 (en) 2019-02-15 2023-04-01 Biosense Webster Israel Ltd Transesophageal catheter with carbon dioxide delivery system for thermal protection of esophagus
US11439358B2 (en) 2019-04-09 2022-09-13 Ziteo, Inc. Methods and systems for high performance and versatile molecular imaging
WO2020239518A1 (en) 2019-05-24 2020-12-03 Koninklijke Philips N.V. Devices, systems, and methods for guilding repeated ultrasound exams for serial monitoring
US20210186642A1 (en) 2019-12-23 2021-06-24 Ethicon, Inc. Esophageal Protection Pathways
US20210186601A1 (en) 2019-12-23 2021-06-24 Ethicon, Inc. Transesophageal Catheter for Thermal Protection of the Esophagus
US20210187242A1 (en) 2019-12-23 2021-06-24 Ethicon, Inc. Fluid Delivery System for Creating Separation Between Biological Surfaces
US11998391B1 (en) 2020-04-02 2024-06-04 yoR Labs, Inc. Method and apparatus for composition of ultrasound images with integration of “thick-slice” 3-dimensional ultrasound imaging zone(s) and 2-dimensional ultrasound zone(s) utilizing a multi-zone, multi-frequency ultrasound image reconstruction scheme with sub-zone blending
US11832991B2 (en) 2020-08-25 2023-12-05 yoR Labs, Inc. Automatic ultrasound feature detection
US20220079462A1 (en) * 2020-09-16 2022-03-17 Biosense Webster (Israel) Ltd. Systems and methods for cardiac chamber visualization
US11751850B2 (en) 2020-11-19 2023-09-12 yoR Labs, Inc. Ultrasound unified contrast and time gain compensation control
WO2022264011A1 (en) 2021-06-14 2022-12-22 Ethicon, Inc. Catheter with carbon dioxide delivery system and methods
US20230240790A1 (en) * 2022-02-03 2023-08-03 Medtronic Navigation, Inc. Systems, methods, and devices for providing an augmented display
US12004821B2 (en) 2022-02-03 2024-06-11 Medtronic Navigation, Inc. Systems, methods, and devices for generating a hybrid image
US20230329678A1 (en) * 2022-04-14 2023-10-19 Biosense Webster (Israel) Ltd. Augmented ultrasonic images

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001070269A (en) * 1999-09-06 2001-03-21 Toshiba Corp Electrophysiologic mapping device

Family Cites Families (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2607769C (en) * 1994-08-19 2012-04-24 Biosense, Inc. Medical diagnosis, treatment and imaging systems
US6690963B2 (en) * 1995-01-24 2004-02-10 Biosense, Inc. System for determining the location and orientation of an invasive medical instrument
WO1997029685A1 (en) * 1996-02-15 1997-08-21 Biosense, Inc. Independently positionable transducers for location system
CA2246287C (en) * 1996-02-15 2006-10-24 Biosense, Inc. Medical procedures and apparatus using intrabody probes
JP3569608B2 (en) * 1997-04-25 2004-09-22 アロカ株式会社 Ultrasound diagnostic equipment
AU4318499A (en) 1997-11-24 1999-12-13 Burdette Medical Systems, Inc. Real time brachytherapy spatial registration and visualization system
US6066096A (en) * 1998-05-08 2000-05-23 Duke University Imaging probes and catheters for volumetric intraluminal ultrasound imaging and related systems
ATE237810T1 (en) * 1998-06-04 2003-05-15 Ge Medical Systems Kretztechni METHOD FOR EXAMINING OBJECTS USING ULTRASOUND
JP4068234B2 (en) * 1998-10-05 2008-03-26 株式会社東芝 Ultrasonic diagnostic equipment
US6701176B1 (en) * 1998-11-04 2004-03-02 Johns Hopkins University School Of Medicine Magnetic-resonance-guided imaging, electrophysiology, and ablation
US6352509B1 (en) * 1998-11-16 2002-03-05 Kabushiki Kaisha Toshiba Three-dimensional ultrasonic diagnosis apparatus
US6645145B1 (en) * 1998-11-19 2003-11-11 Siemens Medical Solutions Usa, Inc. Diagnostic medical ultrasound systems and transducers utilizing micro-mechanical components
JP3645727B2 (en) * 1999-01-28 2005-05-11 株式会社日立製作所 Ultrasonic diagnostic apparatus, program for synthesizing panoramic image, and recording medium thereof
US6556695B1 (en) * 1999-02-05 2003-04-29 Mayo Foundation For Medical Education And Research Method for producing high resolution real-time images, of structure and function during medical procedures
US6892091B1 (en) * 2000-02-18 2005-05-10 Biosense, Inc. Catheter, method and apparatus for generating an electrical map of a chamber of the heart
US6716166B2 (en) * 2000-08-18 2004-04-06 Biosense, Inc. Three-dimensional reconstruction using ultrasound
US6650927B1 (en) * 2000-08-18 2003-11-18 Biosense, Inc. Rendering of diagnostic imaging data on a three-dimensional map
WO2002082375A2 (en) * 2001-04-06 2002-10-17 Stephen Solomon Cardiological mapping and navigation system
WO2002083016A1 (en) * 2001-04-13 2002-10-24 Surgi-Vision, Inc. Systems and methods for magnetic-resonance-guided interventional procedures
JP2003038492A (en) * 2001-07-30 2003-02-12 Pentax Corp Ultrasonic endoscopic device
US6773402B2 (en) * 2001-07-10 2004-08-10 Biosense, Inc. Location sensing with real-time ultrasound imaging
US7477763B2 (en) * 2002-06-18 2009-01-13 Boston Scientific Scimed, Inc. Computer generated representation of the imaging pattern of an imaging device
US20040068178A1 (en) * 2002-09-17 2004-04-08 Assaf Govari High-gradient recursive locating system
US7306593B2 (en) * 2002-10-21 2007-12-11 Biosense, Inc. Prediction and assessment of ablation of cardiac tissue
JP4343592B2 (en) * 2003-06-16 2009-10-14 オリンパス株式会社 Ultrasonic diagnostic equipment
US20060025679A1 (en) * 2004-06-04 2006-02-02 Viswanathan Raju R User interface for remote control of medical devices
US20060239527A1 (en) * 2005-04-25 2006-10-26 Sriram Krishnan Three-dimensional cardiac border delineation in medical imaging
US10143398B2 (en) * 2005-04-26 2018-12-04 Biosense Webster, Inc. Registration of ultrasound data with pre-acquired image
AU2006201646B2 (en) * 2005-04-26 2011-01-06 Biosense Webster, Inc. Display of catheter tip with beam direction for ultrasound system
US8870779B2 (en) * 2005-04-26 2014-10-28 Biosense Webster, Inc. Display of two-dimensional ultrasound fan
US7517318B2 (en) * 2005-04-26 2009-04-14 Biosense Webster, Inc. Registration of electro-anatomical map with pre-acquired image using ultrasound
US20060241445A1 (en) 2005-04-26 2006-10-26 Altmann Andres C Three-dimensional cardial imaging using ultrasound contour reconstruction
US20070049817A1 (en) 2005-08-30 2007-03-01 Assaf Preiss Segmentation and registration of multimodal images using physiological data
US7918793B2 (en) 2005-10-28 2011-04-05 Biosense Webster, Inc. Synchronization of ultrasound imaging data with electrical mapping
US8303505B2 (en) * 2005-12-02 2012-11-06 Abbott Cardiovascular Systems Inc. Methods and apparatuses for image guided medical procedures
JP5738507B2 (en) * 2006-01-19 2015-06-24 東芝メディカルシステムズ株式会社 Ultrasonic probe trajectory expression device and ultrasonic diagnostic device
JP5394622B2 (en) * 2007-07-31 2014-01-22 オリンパスメディカルシステムズ株式会社 Medical guide system
JP7500386B2 (en) 2020-10-12 2024-06-17 株式会社ジャパンディスプレイ Display device

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001070269A (en) * 1999-09-06 2001-03-21 Toshiba Corp Electrophysiologic mapping device

Also Published As

Publication number Publication date
EP1929956A2 (en) 2008-06-11
JP5460955B2 (en) 2014-04-02
EP1929956A3 (en) 2008-09-10
JP2008183398A (en) 2008-08-14
MX2007015573A (en) 2009-02-16
BRPI0704524A (en) 2008-07-29
CA2614033A1 (en) 2008-06-08
US7831076B2 (en) 2010-11-09
AU2007237321B2 (en) 2013-09-19
IL187783A (en) 2011-07-31
CN101219061B (en) 2013-02-06
IL187783A0 (en) 2009-02-11
US20080137927A1 (en) 2008-06-12
AU2007237321A1 (en) 2008-06-26
KR20080053224A (en) 2008-06-12
EP2359749B1 (en) 2016-10-26
EP2359749A1 (en) 2011-08-24
CA2614033C (en) 2013-03-26
CN101219061A (en) 2008-07-16

Similar Documents

Publication Publication Date Title
EP1929956B1 (en) Coloring electroanatomical maps to indicate ultrasound data acquisiton
EP1852824B1 (en) Enhanced ultrasound image display
JP4795099B2 (en) Superposition of electroanatomical map and pre-acquired image using ultrasound
JP5345275B2 (en) Superposition of ultrasonic data and pre-acquired image
JP5265091B2 (en) Display of 2D fan-shaped ultrasonic image
JP5622995B2 (en) Display of catheter tip using beam direction for ultrasound system
JP2006305358A (en) Three-dimensional cardiac imaging using ultrasound contour reconstruction
JP2006305359A (en) Software product for three-dimensional cardiac imaging using ultrasound contour reconstruction
JP2009160397A (en) Three-dimensional image reconstruction using doppler ultrasound
AU2013251245B2 (en) Coloring electroanatomical maps to indicate ultrasound data acquisition

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC MT NL PL PT RO SE SI SK TR

AX Request for extension of the european patent

Extension state: AL BA HR MK RS

PUAL Search report despatched

Free format text: ORIGINAL CODE: 0009013

AK Designated contracting states

Kind code of ref document: A3

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC MT NL PL PT RO SE SI SK TR

AX Request for extension of the european patent

Extension state: AL BA HR MK RS

REG Reference to a national code

Ref country code: HK

Ref legal event code: DE

Ref document number: 1118436

Country of ref document: HK

17P Request for examination filed

Effective date: 20090224

17Q First examination report despatched

Effective date: 20090331

AKX Designation fees paid

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC MT NL PL PT RO SE SI SK TR

REG Reference to a national code

Ref country code: HK

Ref legal event code: WD

Ref document number: 1118436

Country of ref document: HK

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

RIC1 Information provided on ipc code assigned before grant

Ipc: A61B 8/13 20060101AFI20160114BHEP

Ipc: A61B 5/055 20060101ALN20160114BHEP

Ipc: A61B 5/044 20060101ALI20160114BHEP

Ipc: A61B 6/08 20060101ALI20160114BHEP

Ipc: A61B 8/12 20060101ALI20160114BHEP

RIC1 Information provided on ipc code assigned before grant

Ipc: A61B 8/13 20060101AFI20160121BHEP

Ipc: A61B 5/044 20060101ALI20160121BHEP

Ipc: A61B 8/12 20060101ALI20160121BHEP

Ipc: A61B 6/08 20060101ALI20160121BHEP

Ipc: A61B 5/055 20060101ALN20160121BHEP

INTG Intention to grant announced

Effective date: 20160205

GRAR Information related to intention to grant a patent recorded

Free format text: ORIGINAL CODE: EPIDOSNIGR71

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

RIC1 Information provided on ipc code assigned before grant

Ipc: A61B 6/08 20060101ALI20160421BHEP

Ipc: A61B 5/055 20060101ALN20160421BHEP

Ipc: A61B 5/044 20060101ALI20160421BHEP

Ipc: A61B 8/13 20060101AFI20160421BHEP

Ipc: A61B 8/12 20060101ALI20160421BHEP

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

RIC1 Information provided on ipc code assigned before grant

Ipc: A61B 8/13 20060101AFI20160422BHEP

Ipc: A61B 5/055 20060101ALN20160422BHEP

Ipc: A61B 6/08 20060101ALI20160422BHEP

Ipc: A61B 8/12 20060101ALI20160422BHEP

Ipc: A61B 5/044 20060101ALI20160422BHEP

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LI LT LU LV MC MT NL PL PT RO SE SI SK TR

INTG Intention to grant announced

Effective date: 20160523

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: AT

Ref legal event code: REF

Ref document number: 808437

Country of ref document: AT

Kind code of ref document: T

Effective date: 20160715

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602007046779

Country of ref document: DE

REG Reference to a national code

Ref country code: NL

Ref legal event code: FP

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG4D

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 10

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160930

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 808437

Country of ref document: AT

Kind code of ref document: T

Effective date: 20160629

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20161029

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20161031

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602007046779

Country of ref document: DE

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

26N No opposition filed

Effective date: 20170330

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160929

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

REG Reference to a national code

Ref country code: IE

Ref legal event code: MM4A

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20161231

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20161231

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20161207

REG Reference to a national code

Ref country code: FR

Ref legal event code: PLFP

Year of fee payment: 11

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20161207

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

Ref country code: HU

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO

Effective date: 20071207

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20160629

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MT

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20161207

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: NL

Payment date: 20221114

Year of fee payment: 16

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: BE

Payment date: 20221118

Year of fee payment: 16

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20231102

Year of fee payment: 17

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: IT

Payment date: 20231110

Year of fee payment: 17

Ref country code: FR

Payment date: 20231108

Year of fee payment: 17

Ref country code: DE

Payment date: 20231031

Year of fee payment: 17

REG Reference to a national code

Ref country code: NL

Ref legal event code: MM

Effective date: 20240101

REG Reference to a national code

Ref country code: BE

Ref legal event code: MM

Effective date: 20231231

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20240101

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20240101

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20231231