US20180214129A1 - Medical imaging apparatus - Google Patents
Medical imaging apparatus Download PDFInfo
- Publication number
- US20180214129A1 US20180214129A1 US15/505,626 US201515505626A US2018214129A1 US 20180214129 A1 US20180214129 A1 US 20180214129A1 US 201515505626 A US201515505626 A US 201515505626A US 2018214129 A1 US2018214129 A1 US 2018214129A1
- Authority
- US
- United States
- Prior art keywords
- image data
- ultrasound
- segmentation
- data
- medical image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000002059 diagnostic imaging Methods 0.000 title claims abstract description 29
- 238000002604 ultrasonography Methods 0.000 claims abstract description 164
- 230000011218 segmentation Effects 0.000 claims abstract description 83
- 238000003709 image segmentation Methods 0.000 claims abstract description 38
- 239000000523 sample Substances 0.000 claims abstract description 34
- 230000004927 fusion Effects 0.000 claims description 23
- 238000000034 method Methods 0.000 claims description 9
- 238000011156 evaluation Methods 0.000 claims description 7
- 238000004590 computer program Methods 0.000 claims description 6
- 230000000875 corresponding effect Effects 0.000 description 7
- 238000004458 analytical method Methods 0.000 description 5
- 230000002596 correlated effect Effects 0.000 description 5
- 238000003384 imaging method Methods 0.000 description 4
- 238000004364 calculation method Methods 0.000 description 3
- 230000001419 dependent effect Effects 0.000 description 3
- 210000004185 liver Anatomy 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 210000003240 portal vein Anatomy 0.000 description 2
- 238000002591 computed tomography Methods 0.000 description 1
- 229940079593 drug Drugs 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000001356 surgical procedure Methods 0.000 description 1
- 238000003325 tomography Methods 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
- A61B8/461—Displaying means of special interest
- A61B8/466—Displaying means of special interest adapted to display 3D data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/055—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves involving electronic [EMR] or nuclear [NMR] magnetic resonance, e.g. magnetic resonance imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computed tomography [CT]
- A61B6/032—Transmission computed tomography [CT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/08—Detecting organic movements or changes, e.g. tumours, cysts, swellings
- A61B8/0833—Detecting organic movements or changes, e.g. tumours, cysts, swellings involving detecting or locating foreign bodies or organic structures
- A61B8/085—Detecting organic movements or changes, e.g. tumours, cysts, swellings involving detecting or locating foreign bodies or organic structures for locating body or organic structures, e.g. tumours, calculi, blood vessels, nodules
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/42—Details of probe positioning or probe attachment to the patient
- A61B8/4245—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/48—Diagnostic techniques
- A61B8/483—Diagnostic techniques involving the acquisition of a 3D volume of data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/52—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/5215—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
- A61B8/5238—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for combining image data of patient, e.g. merging several images from different acquisition modes into one image
- A61B8/5261—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for combining image data of patient, e.g. merging several images from different acquisition modes into one image combining images from different diagnostic modalities, e.g. ultrasound and X-ray
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/12—Edge-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/30—Determination of transform parameters for the alignment of images, i.e. image registration
- G06T7/33—Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/40—Arrangements for generating radiation specially adapted for radiation diagnosis
- A61B6/4064—Arrangements for generating radiation specially adapted for radiation diagnosis specially adapted for producing a particular type of beam
- A61B6/4085—Cone-beams
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/46—Arrangements for interfacing with the operator or the patient
- A61B6/467—Arrangements for interfacing with the operator or the patient characterised by special input means
- A61B6/469—Arrangements for interfacing with the operator or the patient characterised by special input means for selecting a region of interest [ROI]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/50—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment specially adapted for specific body parts; specially adapted for specific clinical applications
- A61B6/504—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment specially adapted for specific body parts; specially adapted for specific clinical applications for diagnosis of blood vessels, e.g. by angiography
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5211—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
- A61B6/5229—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
- A61B6/5247—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from an ionising-radiation diagnostic technique and a non-ionising radiation diagnostic technique, e.g. X-ray and ultrasound
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/42—Details of probe positioning or probe attachment to the patient
- A61B8/4245—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient
- A61B8/4254—Details of probe positioning or probe attachment to the patient involving determining the position of the probe, e.g. with respect to an external reference frame or to the patient using sensors mounted on the probe
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/44—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
- A61B8/4416—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device related to combined acquisition of different diagnostic modalities, e.g. combination of ultrasound and X-ray acquisitions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/44—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
- A61B8/4444—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device related to the probe
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/44—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
- A61B8/4483—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device characterised by features of the ultrasound transducer
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
- A61B8/461—Displaying means of special interest
- A61B8/463—Displaying means of special interest characterised by displaying multiple images or images and diagnostic data on one display
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
- A61B8/467—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient characterised by special input means
- A61B8/469—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient characterised by special input means for selection of a region of interest
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10132—Ultrasound image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10132—Ultrasound image
- G06T2207/10136—3D ultrasound image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20092—Interactive image processing based on input by user
- G06T2207/20101—Interactive definition of point of interest, landmark or seed
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30101—Blood vessel; Artery; Vein; Vascular
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30172—Centreline of tubular or elongated structure
Definitions
- the present invention relates to a medical imaging apparatus for evaluating medical image data.
- the present invention further relates to a medical image evaluation method for evaluating medical image data and to a computer program comprising program code means for causing a computer to carry out the steps of the method for evaluating medical image data.
- ultrasound systems which combine ultrasound images and preoperative image data of a patient derived from a different analytic system like MRT or CT.
- a position tracking system is usually utilized to spatially align the different image data.
- the position tracking systems rely on a calibration e.g. based on artificial markers which can be identified in the preoperative and the ultrasound data and which can be correlated to each other so that the alignment of the data can be determined.
- the alignment of the different image data can be based on automatic registration of anatomical features like vessels identified in the different image data, however, the automatic registration is complex, involves large technical effort and is not reliable for any case.
- a corresponding system for automatically correlating images from different imaging systems is e.g. known from US 2013/0053679 A1.
- the position tracking system may also be calibrated on the basis of a user input, wherein a plurality of corresponding positions are identified by the operator in both image data to be aligned.
- this method needs an expert as an operator to calibrate the position tracking system, so that this system is cumbersome.
- a medical imaging apparatus for evaluating medical image data, comprising:
- a medical image evaluation method for evaluating medical image data, comprising the steps of:
- segmentation of the anatomical features in the ultrasound data and/or the 3D medical image data is initiated on the basis of the position identified by the user
- a computer program comprising program code means for causing a computer to carry out the steps of the medical image evaluation method according to the present invention, when said computer program is carried out on a computer.
- the present invention is based on the idea that a position in the 3D medical image data and/or the ultrasound image data is identified by the user via a user input interface in order to provide the information to the system which anatomical features are considered to be advantageous for a segmentation and for a registration of the different image data.
- the segmentation of the anatomical features performed by the medical image segmentation unit and/or the ultrasound segmentation unit are initiated on the basis of the identified position so that the segmentation unit does not need to perform a segmentation of the whole image data and the technical effort and the calculation time is significantly reduced.
- the registration unit can register the ultrasound image data and the 3D medical image data so that a tracking of the ultrasound probe and/or or a fusion of the different image data can be performed with high accuracy. Since the position identified by the user initiates the position of the segmentation, the effort for segmenting the ultrasound image data and/or the 3D medical image data can be significantly reduced and the reliability of the registration can be improved since the most significant anatomical features of the image data can be easily identified by the user input and the influence of artifacts can be reduced.
- the present invention achieves an evaluation of medical image data with improved reliability and which is comfortable for the user.
- the position identified by the user is a point in the 3D medical image data and/or the ultrasound image data. In a further preferred embodiment, the position identified by the user corresponds to a voxel of the 3D medical image data and/or a voxel or a pixel of the ultrasound image data.
- the medical imaging apparatus further comprises a position determining unit attached to the ultrasound probe for determining the position of the ultrasound probe, wherein the position determining unit includes a calibration unit for calibrating the position of the ultrasound probe on the basis of the correlation of the segmentation data received from the registration unit.
- the medical imaging apparatus further comprises a fusion unit for fusion of the ultrasound image data and the 3D medical image data on the basis of the position of the ultrasound probe determined by the position determining unit. This is a possibility to provide a continuously fused medical image on the basis of the ultrasound image data and the 3D medical image data.
- the fusion unit may also alternatively or additionally to the fusion of the ultrasound image data and the 3D medical image data on the basis of the position of the ultrasound probe adapted to fuse the ultrasound image data and the 3D medical image data on the basis of the correlation of the ultrasound segmentation data and the medical image segmentation data provided by the a registration unit. This is a possibility to improve the fusion of the ultrasound image data and the 3D medical image data.
- the fusion of the ultrasound image data and the 3D medical image data is performed by the fusion unit continuously during the acquisition of the ultrasound image data so that a fused image based on the combination of the ultrasound image data and the 3D medical image data can be provided in real time.
- the 3D medical image segmentation unit is adapted to segment anatomical features in the 3D medical image data adjacent to or surrounding the position defined in the 3D medical image data. This is a possibility to utilize segmentation data of certain anatomical features which can be easily identified so that the reliability of the registration can be improved.
- the ultrasound segmentation unit is adapted to segment anatomical features in the ultrasound image data adjacent to or surrounding the position identified in the ultrasound image data. This is a possibility to initiate the segmentation of certain anatomical features which can be easily identified in the ultrasound image data so that the reliability of the registration can be improved.
- the anatomical features are surfaces in the vicinity of the identified position. This is a possibility to improve the accuracy of the registration, since surfaces in the image data can be easily identified by means of the segmentation unit.
- the anatomical features are vessels of the patient. This is a possibility to further improve the registration, since the shape of the vessels can be easily identified by the segmentation unit and the unique shape of the vessels can be easily registered so that the reliability of the correlation of the different image data can be improved.
- the ultrasound segmentation unit and a medical image segmentation unit are adapted to determine centre lines and/or bifurcations of the vessels and wherein the registration unit is adapted to register the ultrasound image data and the 3D medical image data on the basis of the determined centre lines and/or bifurcations of the vessels.
- the user input interface comprises a display unit for displaying the 3D medical image data and/or the ultrasound image data and wherein the user interface comprises an input device for identifying the position of the 3D medical image data and/or the ultrasound image data at the display unit. This is a possibility to easily identify the position in the image data so that the user input is more comfortable.
- the input device is adapted to control a position of an indicator displayed at the display unit within the displayed image data and to identify the position in the displayed image on the basis of the position of the indicator and a user input.
- the indicator is a mouse pointer or the like and the input device comprises an input unit like a mouse or the like, wherein the position can be identified by a single mouse click within the displayed image data. This is a possibility to further improve the comfort of the user input and to reduce the effort for the user.
- the display unit comprises a contact sensitive surface for identifying the position in the displayed image by a user input.
- the display unit is formed as a touchscreen, wherein the position in the image data is identified by a single touch at the corresponding position displayed at the display unit. This is a possibility to further improve the accuracy of the identification of the position and to reduce the effort for the user.
- the 3D medical image data is previously acquired and the image data is stored in a memory device. This is a possibility to combine medical image data of different analysis methods which can be captured of the patient prior to the ultrasound analysis so that the examination time can be reduced.
- the 3D medical image data is MR image data, CT image data, cone-beam CT image data or ultrasound image data.
- the present invention can improve the reliability of the registration, since the segmentation is based on the position identified by the user or initiated on the basis of the identified position and the technical effort, and in particular the calculation effort can be reduced, since the system does not need to provide segmentation data of the whole image data since the region of interest is identified by the user input. Further, since the operator merely needs to identify one position in the image data and does not need to identify corresponding positions in the different image data, no expert knowledge is necessary and the handling is more comfortable for the user.
- FIG. 1 shows a schematic representation of a medical imaging apparatus in use to scan a volume of a patient's body
- FIG. 2 a, b show an ultrasound image and a CT image of a certain site of the patient's body to be correlated
- FIG. 3 a, b show the images of FIG. 2 a, b partially segmented to register the image data
- FIG. 4 a, b show segmentation data of the vessels of the image data shown in FIG. 3 a, b;
- FIG. 5 a, b show centre lines and bifurcations derived from the segmentation data shown in FIG. 4 a, b;
- FIG. 6 shows a correlation of the centre lines and the bifurcations identified in the segmentation data
- FIG. 7 a, b show the initial ultrasound image shown in FIG. 2 a and the fused ultrasound and CT image fused on the basis of the segmentation and registration procedure;
- FIG. 8 shows a flow diagram of a method for evaluating medical image data.
- FIG. 1 shows a schematic illustration of a medical imaging apparatus generally denoted by 10 .
- the medical imaging apparatus 10 is applied to inspect a volume of an anatomical site, in particular an anatomical site of a patient 12 .
- the medical imaging apparatus 10 comprises an ultrasound probe 14 having at least one transducer array including a multitude of transducer elements for transmitting and receiving ultrasound waves.
- the transducer elements are preferably arranged in a 2D array, in particular for providing multi-dimensional image data.
- a medical imaging apparatus 10 comprises in general an image processing apparatus 16 connected to the ultrasound probe 14 for evaluating the ultrasound data received from the ultrasound probe 14 and for combining or correlating the ultrasound images with preoperative images of the patient 12 .
- the imaging processing apparatus 16 comprises an image interface 18 for receiving the preoperative 3D medical image data from a database 20 or an external analysis and imaging apparatus 20 .
- the preoperative image data is preferably computer tomography image data (CT), magnetic resonance tomography image data (MRT), cone-beam CT image data or preoperative 3D ultrasound image data.
- the image processing apparatus 16 comprises an image processing unit 22 connected to the ultrasound probe 14 and to the image interface 18 for evaluating the ultrasound data and for providing ultrasound image data from the volume or object of the patient 12 which is analyzed by the ultrasound probe 14 and for evaluating the preoperative 3D medical image data received from the image interface 18 .
- the image processing apparatus 16 further comprises an ultrasound segmentation unit 24 for segmenting anatomical features of the patient in the ultrasound image data and for providing a corresponding ultrasound segmentation data to the image processing unit 22 .
- the image processing apparatus 16 further comprises a medical image segmentation unit 26 for segmenting the 3D medical image data received from the database 20 via the interface 18 and for providing medical image segmentation data to the image processing unit 22 .
- the medical imaging apparatus 10 further comprises a position determining unit 28 attached to the ultrasound probe 14 for determining a position of the ultrasound probe 14 .
- the position determining unit 28 determines the relative position of the ultrasound probe, e.g. by means of electromagnetic tracking in order to determine a movement of the ultrasound probe 14 with respect to an initial or a calibrated position.
- the initial position is calibrated by means of a calibration unit 30 .
- the calibration unit 30 is connected to the image processing unit 22 in order to correlate the ultrasound data captured by the ultrasound probe 14 , the position of the ultrasound probe 14 received from the position determining unit 28 and the 3D medical image data on the basis of the ultrasound segmentation data and medical image segmentation data received from the ultrasound segmentation unit 24 and the medical image segmentation unit 26 as described in the following.
- the so determined position of the ultrasound probe 14 with respect to the 3D medical image data is used as a reference position or used as calibrated position of the ultrasound probe 14 . If the ultrasound probe 14 is moved with respect to the calibrated position, the position determining unit 28 detects the distance and the direction with respect to the calibrated position and provides a so determined current position of the ultrasound probe 14 .
- the image processing unit 22 further comprises a registration unit 32 for correlating the ultrasound segmentation data and the medical image segmentation data.
- the calibration unit 30 calibrates the position of the ultrasound probe 14 with the respective ultrasound data and the 3D medical image data of the patient 12 on the basis of the correlation of the ultrasound segmentation data and the medical image segmentation data received from the registration unit 32 .
- the image processing unit 22 further comprises a fusion unit 34 for fusion of the ultrasound image data and the 3D medical image data on the basis of the position of the ultrasound probe 14 determined by the position determining unit 28 .
- the fusion unit 34 may also utilize the correlation of the ultrasound segmentation data and the medical image segmentation data received from the registration unit 32 in order to fuse the ultrasound image data and the 3D medical image data. This is a possibility to further improve the fusion of the different data.
- the medical imaging apparatus 10 further comprises a display unit 36 for displaying image data received from the image processing apparatus 16 .
- the display unit 36 receives the image data in general from the image processing unit 22 and is adapted to display the ultrasound image data and the 3D medical image data and also the respective segmentation data.
- the medical imaging apparatus 10 further comprises an input device 38 which may be connected to the display unit 36 or to the image processing apparatus 16 in order to control the image acquisition and to identify a position in the 3D medical image data and/or in the ultrasound image data displayed on the display unit 36 .
- the input device 38 may comprise a keyboard or a mouse or the like or may be formed as a touchscreen of the display unit 36 to identify or indicate a certain anatomical feature or a position within the displayed ultrasound image data and/or the 3D medical image data.
- the image processing unit 22 is adapted to receive the position identified in the image data by means of the input device 38 by the user.
- the image processing unit 22 initiates on the basis of the position identified in image data the ultrasound segmentation unit 24 or the medical image segmentation unit 26 to perform a segmentation of the respective image data at the identified position and in the vicinity of the identified position and/or surrounding the identified position.
- the imaging processing unit 22 and in particular the registration unit 32 comprised in the image processing unit 22 correlates the ultrasound segmentation data and the 3D medical image segmentation data and the fusion unit 34 combines the respective ultrasound image data and the 3D medical image data to provide a composed medical image and provides the composed medical image to the display unit 36 .
- the respective segmentation unit 24 , 26 performs the segmentation at a certain anatomical feature which can be easily identified by the segmentation unit so that the technical effort and the calculation time for the segmentation is reduced and the anatomical features for the correlation can be identified faster and with an improved reliability.
- the spatial alignment of the ultrasound image data and the 3D medical image data is performed by the fusion unit 34 on the basis of the correlation received from the registration unit 32 .
- the fusion of the ultrasound image data and the 3D medical image data is performed by the fusion unit 34 continuously during the ultrasound scan.
- the fused image based on the combination of the ultrasound image data and the 3D medical image data can therefore be provided in real time during the ultrasound scan.
- FIG. 2 a shows an ultrasound image 40 on the basis of the ultrasound image data received from the ultrasound probe and captured from the patient 12 .
- FIG. 2 b shows a sectional medical image 42 on the basis of the 3D medical image data of the patient 12 .
- FIG. 2 a and FIG. 2 b show the liver of the patient 12 , wherein the ultrasound image 40 and the sectional medical image 42 are not yet spatially aligned or correlated to each other.
- the user identifies a position in the sectional medical image 42 by means of the input device 38 as a user input interface.
- the position is identified by an indicator 44 movable within the sectional medical image 42 .
- the indicator 44 shows the position identified by the user, which is in this particular case the portal vein of the liver which is also visible in the field of view of the ultrasound image 40 .
- the segmentation of the 3D medical image data is initiated as shown in FIG. 3 b and also the segmentation of the vessels is performed in the ultrasound image data as shown in FIG. 3 a. Since the portal vein is identified by the user input and the indicator 44 , the segmentation of this anatomical feature can be performed faster and with a higher reliability so that the overall reliability of the registration and correlation of the respective image data is improved. It shall be understood that the position of a certain anatomical feature can be identified within the sectional medical image 42 and/or in the ultrasound image 40 so that the segmentation in general can be performed faster and with a higher reliability.
- the anatomical features surrounding the identified position are segmented, wherein the anatomical features may be surrounding surfaces like the vessels or other anatomical surfaces within the patient's body.
- the ultrasound segmentation data is in FIG. 3 a denoted by 46 and the medical image segmentation data is in FIG. 3 b denoted by 48 .
- FIG. 4 a shows the ultrasound segmentation data 46 of the vessels derived from the ultrasound image data by means of the ultrasound segmentation unit 24 .
- FIG. 4 b shows the medical image segmentation data derived from the 3D medical image data by means of the medical image segmentation unit 26 .
- the ultrasound segmentation unit 24 determines centre line 50 and bifurcations 52 from the ultrasound segmentation data 46 and the medical image segmentation unit 26 determines from the medical image segmentation data 48 centre lines 54 and bifurcations 56 as shown in FIG. 5 b.
- the registration unit 32 correlates the centre lines 50 , 54 and the bifurcations 52 , 56 of the segmentation data 46 , 48 as shown in FIG. 6 and the fusion unit 34 combines the ultrasound image data and the 3D medical image data on the basis of the correlation received from the registration unit 32 and/or the position of the ultrasound probe 14 determined by the position determining unit 28 .
- FIG. 7 a shows the ultrasound image 40 shown in FIG. 2 a and FIG. 7 b shows an image spatially aligned by the fusion unit 34 of the image processing unit 22 on the basis of the correlation received from the registration unit 32 .
- the correlation can be easily performed by the user input, since the segmentation effort is reduced and the reliability of the identification of significant anatomical features within the image data can be improved.
- FIG. 8 a flow diagram of a method for evaluating medical image data is shown and generally denoted by 60 .
- the method 60 starts with acquiring ultrasound data of the patient 12 by means of the ultrasound probe 14 as shown at step 62 and with receiving 3D medical image data of the patient 12 from the external database 20 , which is usually MRT or CT data previously acquired from the patient 12 as shown at step 64 .
- a position is identified in the 3D medical image data and/or the ultrasound image data by the user via the input device 38 as shown at step 66 .
- the anatomical features of the patient 12 are segmented in the ultrasound data and corresponding segmentation data of the anatomical features are provided as shown at step 68 . Further, anatomical features in the 3D medical image data are segmented and the medical image segmentation data 48 is provided as shown at the step 70 .
- the segmentation of the anatomical feature in the ultrasound data and/or in the 3D medical image data are based on the identified position, wherein the respective segmentation in the ultrasound data and/or the medical image data are initiated on the basis of the identified position.
- the anatomical features surrounding the identified position are segmented in order to segment only those anatomical features which are relevant and which are identified by the user. If the position is identified in the 3D medical image data, the segmentation of the ultrasound data may be performed in general or if the position is identified in the ultrasound image data, the segmentation of the medical image data may be performed in general. In a certain embodiment, the position to be segmented is identified in the 3D medical image data as well as the ultrasound image data.
- the ultrasound segmentation data and the medical image segmentation data are provided to the registration unit 32 , wherein the ultrasound segmentation data 46 and the medical image segmentation data 48 are correlated at step 72 .
- the calibration of the position determining unit 28 can be performed and the fusion of the ultrasound image data and the 3D medical image data can be performed by the fusion unit 34 .
- a computer program may be stored/distributed on a suitable medium, such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the Internet or other wired or wireless telecommunication systems.
- a suitable medium such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the Internet or other wired or wireless telecommunication systems.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Animal Behavior & Ethology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Pathology (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Radiology & Medical Imaging (AREA)
- Biophysics (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Vascular Medicine (AREA)
- High Energy & Nuclear Physics (AREA)
- Computer Graphics (AREA)
- General Engineering & Computer Science (AREA)
- Pulmonology (AREA)
- Optics & Photonics (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
Abstract
A medical imaging apparatus (10) for evaluating medical image data is disclosed. The medical imaging apparatus comprises an ultrasound acquisition unit including an ultrasound probe (14) for acquiring ultrasound image data of a patient (12) and an ultrasound segmentation unit (24) for segmenting anatomical features of the patient in the ultrasound image data and for providing ultrasound segmentation data (46). The apparatus comprises an image data interface (18) for receiving 3D medical image data of the patient and a medical image segmentation unit (26) for segmenting the 3D medical image data and for providing medical image segmentation data (48). A user input interface (38) is provided for identifying a position (44) by the user in the 3D medical image data and/or in the ultrasound image data in order to initiate the segmentation of anatomical features by the medical image segmentation unit and/or the ultrasound segmentation unit on the basis of the position identified by the user, wherein a registration unit (32) correlates the ultrasound segmentation data and the medical image segmentation data.
Description
- The present invention relates to a medical imaging apparatus for evaluating medical image data. The present invention further relates to a medical image evaluation method for evaluating medical image data and to a computer program comprising program code means for causing a computer to carry out the steps of the method for evaluating medical image data.
- In the field of medical imaging systems, it is generally known to combine different images of a patient acquired by different medical analysis systems in order to improve the diagnostic possibilities. In particular, ultrasound systems are known which combine ultrasound images and preoperative image data of a patient derived from a different analytic system like MRT or CT. To enable the fusion of live ultrasound images of a patient with the preoperative volume data of the same patient, a position tracking system is usually utilized to spatially align the different image data.
- The position tracking systems rely on a calibration e.g. based on artificial markers which can be identified in the preoperative and the ultrasound data and which can be correlated to each other so that the alignment of the data can be determined.
- Further, the alignment of the different image data can be based on automatic registration of anatomical features like vessels identified in the different image data, however, the automatic registration is complex, involves large technical effort and is not reliable for any case. A corresponding system for automatically correlating images from different imaging systems is e.g. known from US 2013/0053679 A1.
- The position tracking system may also be calibrated on the basis of a user input, wherein a plurality of corresponding positions are identified by the operator in both image data to be aligned. However, this method needs an expert as an operator to calibrate the position tracking system, so that this system is cumbersome.
- Physics in Medicine and Biology, vol. 57, no 1, 29 Nov. 2011, pages 81-91, discloses an automatic registration between 3D intra-operative ultrasound and pre-operative CT images of the liver.
- It is therefore an object of the invention to provide an improved medical imaging apparatus and a corresponding improved medical imaging evaluation method for evaluating medical image data, which is more reliable and less complicated for the user.
- According to one aspect of the present invention, a medical imaging apparatus is provided for evaluating medical image data, comprising:
-
- an ultrasound acquisition unit including an ultrasound probe for acquiring ultrasound image data of a patient,
- an ultrasound segmentation unit for segmenting anatomical features of the patient in the ultrasound image data and for providing ultrasound segmentation data,
- an image data interface for receiving 3D medical image data of the patient, a medical image segmentation unit for segmenting the 3D medical image data and for providing medical image segmentation data,
- a user input interface for identifying a position by the user in the 3D medical image data and/or in the ultrasound image data in order to initiate the segmentation of anatomical features by the medical image segmentation unit and/or the ultrasound segmentation unit on the basis of the position identified by the user,
- a registration unit for correlating the ultrasound segmentation data and the medical image segmentation data.
- According to another aspect of the present invention, a medical image evaluation method is provided for evaluating medical image data, comprising the steps of:
-
- acquiring ultrasound data of a patient by means of an ultrasound probe,
- receiving 3D medical image data of the patient,
- identifying a position in the 3D medical image data and/or in the ultrasound image data by a user via a user input interface,
- segmenting anatomical features of the patient in the ultrasound data and providing ultrasound segmentation data of the anatomical features,
- segmenting anatomical features the 3D medical image data and providing medical image segmentation data,
- wherein the segmentation of the anatomical features in the ultrasound data and/or the 3D medical image data is initiated on the basis of the position identified by the user, and
-
- correlating the ultrasound segmentation data and the medical image segmentation data.
- According to still another aspect of the present invention, a computer program is provided comprising program code means for causing a computer to carry out the steps of the medical image evaluation method according to the present invention, when said computer program is carried out on a computer.
- Preferred embodiments of the invention are defined in the dependent claims. It shall be understood that the claimed method has similar and/or identical preferred embodiments as the claimed device and as defined in the dependent claims.
- The present invention is based on the idea that a position in the 3D medical image data and/or the ultrasound image data is identified by the user via a user input interface in order to provide the information to the system which anatomical features are considered to be advantageous for a segmentation and for a registration of the different image data. The segmentation of the anatomical features performed by the medical image segmentation unit and/or the ultrasound segmentation unit are initiated on the basis of the identified position so that the segmentation unit does not need to perform a segmentation of the whole image data and the technical effort and the calculation time is significantly reduced. On the basis of the segmentation data which is calculated on the basis of the identified position, the registration unit can register the ultrasound image data and the 3D medical image data so that a tracking of the ultrasound probe and/or or a fusion of the different image data can be performed with high accuracy. Since the position identified by the user initiates the position of the segmentation, the effort for segmenting the ultrasound image data and/or the 3D medical image data can be significantly reduced and the reliability of the registration can be improved since the most significant anatomical features of the image data can be easily identified by the user input and the influence of artifacts can be reduced.
- Consequently, the present invention achieves an evaluation of medical image data with improved reliability and which is comfortable for the user.
- In a preferred embodiment, the position identified by the user is a point in the 3D medical image data and/or the ultrasound image data. In a further preferred embodiment, the position identified by the user corresponds to a voxel of the 3D medical image data and/or a voxel or a pixel of the ultrasound image data.
- In a preferred embodiment, the medical imaging apparatus further comprises a position determining unit attached to the ultrasound probe for determining the position of the ultrasound probe, wherein the position determining unit includes a calibration unit for calibrating the position of the ultrasound probe on the basis of the correlation of the segmentation data received from the registration unit. This is a possibility to reduce the evaluation effort during a surgery, since the position determining unit can further improve the registration if it is calibrated on the basis of the correlation of the segmentation data.
- In a further preferred embodiment, the medical imaging apparatus further comprises a fusion unit for fusion of the ultrasound image data and the 3D medical image data on the basis of the position of the ultrasound probe determined by the position determining unit. This is a possibility to provide a continuously fused medical image on the basis of the ultrasound image data and the 3D medical image data.
- The fusion unit may also alternatively or additionally to the fusion of the ultrasound image data and the 3D medical image data on the basis of the position of the ultrasound probe adapted to fuse the ultrasound image data and the 3D medical image data on the basis of the correlation of the ultrasound segmentation data and the medical image segmentation data provided by the a registration unit. This is a possibility to improve the fusion of the ultrasound image data and the 3D medical image data.
- The fusion of the ultrasound image data and the 3D medical image data is performed by the fusion unit continuously during the acquisition of the ultrasound image data so that a fused image based on the combination of the ultrasound image data and the 3D medical image data can be provided in real time.
- In a preferred embodiment, the 3D medical image segmentation unit is adapted to segment anatomical features in the 3D medical image data adjacent to or surrounding the position defined in the 3D medical image data. This is a possibility to utilize segmentation data of certain anatomical features which can be easily identified so that the reliability of the registration can be improved.
- In a preferred embodiment, the ultrasound segmentation unit is adapted to segment anatomical features in the ultrasound image data adjacent to or surrounding the position identified in the ultrasound image data. This is a possibility to initiate the segmentation of certain anatomical features which can be easily identified in the ultrasound image data so that the reliability of the registration can be improved.
- In a preferred embodiment, the anatomical features are surfaces in the vicinity of the identified position. This is a possibility to improve the accuracy of the registration, since surfaces in the image data can be easily identified by means of the segmentation unit.
- In a preferred embodiment, the anatomical features are vessels of the patient. This is a possibility to further improve the registration, since the shape of the vessels can be easily identified by the segmentation unit and the unique shape of the vessels can be easily registered so that the reliability of the correlation of the different image data can be improved.
- In a preferred embodiment, the ultrasound segmentation unit and a medical image segmentation unit are adapted to determine centre lines and/or bifurcations of the vessels and wherein the registration unit is adapted to register the ultrasound image data and the 3D medical image data on the basis of the determined centre lines and/or bifurcations of the vessels. This is a possibility to reduce the technical effort for the registration of the image data and to improve the accuracy of the registration, since the centre lines and the bifurcations can be easily derived from the segmentation data and the so derived data can be registered with high accuracy and low technical effort.
- In a preferred embodiment, the user input interface comprises a display unit for displaying the 3D medical image data and/or the ultrasound image data and wherein the user interface comprises an input device for identifying the position of the 3D medical image data and/or the ultrasound image data at the display unit. This is a possibility to easily identify the position in the image data so that the user input is more comfortable.
- In a preferred embodiment, the input device is adapted to control a position of an indicator displayed at the display unit within the displayed image data and to identify the position in the displayed image on the basis of the position of the indicator and a user input. This is a further possibility to identify the position with high precision and low effort for the user, since the indicator is displayed at the display unit within the displayed image data. In a further preferred embodiment, the indicator is a mouse pointer or the like and the input device comprises an input unit like a mouse or the like, wherein the position can be identified by a single mouse click within the displayed image data. This is a possibility to further improve the comfort of the user input and to reduce the effort for the user.
- In a further preferred embodiment, the display unit comprises a contact sensitive surface for identifying the position in the displayed image by a user input. In other words, the display unit is formed as a touchscreen, wherein the position in the image data is identified by a single touch at the corresponding position displayed at the display unit. This is a possibility to further improve the accuracy of the identification of the position and to reduce the effort for the user.
- In a preferred embodiment, the 3D medical image data is previously acquired and the image data is stored in a memory device. This is a possibility to combine medical image data of different analysis methods which can be captured of the patient prior to the ultrasound analysis so that the examination time can be reduced.
- In a preferred embodiment, the 3D medical image data is MR image data, CT image data, cone-beam CT image data or ultrasound image data. These are possibilities to improve the diagnostic possibilities, since the different analysis methods have different contrasts and different identification techniques and the amount of information of the anatomical features can be improved.
- As mentioned above, the present invention can improve the reliability of the registration, since the segmentation is based on the position identified by the user or initiated on the basis of the identified position and the technical effort, and in particular the calculation effort can be reduced, since the system does not need to provide segmentation data of the whole image data since the region of interest is identified by the user input. Further, since the operator merely needs to identify one position in the image data and does not need to identify corresponding positions in the different image data, no expert knowledge is necessary and the handling is more comfortable for the user.
- These and other aspects of the invention will be apparent from and elucidated with reference to the embodiment(s) described hereinafter. In the following drawings
-
FIG. 1 shows a schematic representation of a medical imaging apparatus in use to scan a volume of a patient's body; -
FIG. 2a, b show an ultrasound image and a CT image of a certain site of the patient's body to be correlated; -
FIG. 3a, b show the images ofFIG. 2a, b partially segmented to register the image data; -
FIG. 4a, b show segmentation data of the vessels of the image data shown inFIG. 3 a, b; -
FIG. 5a, b show centre lines and bifurcations derived from the segmentation data shown inFIG. 4 a, b; -
FIG. 6 shows a correlation of the centre lines and the bifurcations identified in the segmentation data; -
FIG. 7a, b show the initial ultrasound image shown inFIG. 2a and the fused ultrasound and CT image fused on the basis of the segmentation and registration procedure; and -
FIG. 8 shows a flow diagram of a method for evaluating medical image data. -
FIG. 1 shows a schematic illustration of a medical imaging apparatus generally denoted by 10. Themedical imaging apparatus 10 is applied to inspect a volume of an anatomical site, in particular an anatomical site of apatient 12. Themedical imaging apparatus 10 comprises anultrasound probe 14 having at least one transducer array including a multitude of transducer elements for transmitting and receiving ultrasound waves. The transducer elements are preferably arranged in a 2D array, in particular for providing multi-dimensional image data. - A
medical imaging apparatus 10 comprises in general animage processing apparatus 16 connected to theultrasound probe 14 for evaluating the ultrasound data received from theultrasound probe 14 and for combining or correlating the ultrasound images with preoperative images of thepatient 12. Theimaging processing apparatus 16 comprises animage interface 18 for receiving the preoperative 3D medical image data from adatabase 20 or an external analysis andimaging apparatus 20. The preoperative image data is preferably computer tomography image data (CT), magnetic resonance tomography image data (MRT), cone-beam CT image data or preoperative 3D ultrasound image data. Theimage processing apparatus 16 comprises animage processing unit 22 connected to theultrasound probe 14 and to theimage interface 18 for evaluating the ultrasound data and for providing ultrasound image data from the volume or object of the patient 12 which is analyzed by theultrasound probe 14 and for evaluating the preoperative 3D medical image data received from theimage interface 18. - The
image processing apparatus 16 further comprises anultrasound segmentation unit 24 for segmenting anatomical features of the patient in the ultrasound image data and for providing a corresponding ultrasound segmentation data to theimage processing unit 22. Theimage processing apparatus 16 further comprises a medicalimage segmentation unit 26 for segmenting the 3D medical image data received from thedatabase 20 via theinterface 18 and for providing medical image segmentation data to theimage processing unit 22. - The
medical imaging apparatus 10 further comprises aposition determining unit 28 attached to theultrasound probe 14 for determining a position of theultrasound probe 14. Theposition determining unit 28 determines the relative position of the ultrasound probe, e.g. by means of electromagnetic tracking in order to determine a movement of theultrasound probe 14 with respect to an initial or a calibrated position. The initial position is calibrated by means of acalibration unit 30. Thecalibration unit 30 is connected to theimage processing unit 22 in order to correlate the ultrasound data captured by theultrasound probe 14, the position of theultrasound probe 14 received from theposition determining unit 28 and the 3D medical image data on the basis of the ultrasound segmentation data and medical image segmentation data received from theultrasound segmentation unit 24 and the medicalimage segmentation unit 26 as described in the following. The so determined position of theultrasound probe 14 with respect to the 3D medical image data is used as a reference position or used as calibrated position of theultrasound probe 14. If theultrasound probe 14 is moved with respect to the calibrated position, theposition determining unit 28 detects the distance and the direction with respect to the calibrated position and provides a so determined current position of theultrasound probe 14. - The
image processing unit 22 further comprises aregistration unit 32 for correlating the ultrasound segmentation data and the medical image segmentation data. Thecalibration unit 30 calibrates the position of theultrasound probe 14 with the respective ultrasound data and the 3D medical image data of the patient 12 on the basis of the correlation of the ultrasound segmentation data and the medical image segmentation data received from theregistration unit 32. - The
image processing unit 22 further comprises afusion unit 34 for fusion of the ultrasound image data and the 3D medical image data on the basis of the position of theultrasound probe 14 determined by theposition determining unit 28. - The
fusion unit 34 may also utilize the correlation of the ultrasound segmentation data and the medical image segmentation data received from theregistration unit 32 in order to fuse the ultrasound image data and the 3D medical image data. This is a possibility to further improve the fusion of the different data. - The
medical imaging apparatus 10 further comprises adisplay unit 36 for displaying image data received from theimage processing apparatus 16. Thedisplay unit 36 receives the image data in general from theimage processing unit 22 and is adapted to display the ultrasound image data and the 3D medical image data and also the respective segmentation data. Themedical imaging apparatus 10 further comprises aninput device 38 which may be connected to thedisplay unit 36 or to theimage processing apparatus 16 in order to control the image acquisition and to identify a position in the 3D medical image data and/or in the ultrasound image data displayed on thedisplay unit 36. Theinput device 38 may comprise a keyboard or a mouse or the like or may be formed as a touchscreen of thedisplay unit 36 to identify or indicate a certain anatomical feature or a position within the displayed ultrasound image data and/or the 3D medical image data. - The
image processing unit 22 is adapted to receive the position identified in the image data by means of theinput device 38 by the user. Theimage processing unit 22 initiates on the basis of the position identified in image data theultrasound segmentation unit 24 or the medicalimage segmentation unit 26 to perform a segmentation of the respective image data at the identified position and in the vicinity of the identified position and/or surrounding the identified position. - The
imaging processing unit 22 and in particular theregistration unit 32 comprised in theimage processing unit 22 correlates the ultrasound segmentation data and the 3D medical image segmentation data and thefusion unit 34 combines the respective ultrasound image data and the 3D medical image data to provide a composed medical image and provides the composed medical image to thedisplay unit 36. - Since the segmentation of an anatomical feature of the
patient 12 is initiated by the position determined by the user within the displayed ultrasound image data and/or the 3D medical image data, therespective segmentation unit - The spatial alignment of the ultrasound image data and the 3D medical image data is performed by the
fusion unit 34 on the basis of the correlation received from theregistration unit 32 . - The fusion of the ultrasound image data and the 3D medical image data is performed by the
fusion unit 34 continuously during the ultrasound scan. The fused image based on the combination of the ultrasound image data and the 3D medical image data can therefore be provided in real time during the ultrasound scan. -
FIG. 2a shows anultrasound image 40 on the basis of the ultrasound image data received from the ultrasound probe and captured from thepatient 12.FIG. 2b shows a sectionalmedical image 42 on the basis of the 3D medical image data of thepatient 12. In this particular case,FIG. 2a andFIG. 2b show the liver of thepatient 12, wherein theultrasound image 40 and the sectionalmedical image 42 are not yet spatially aligned or correlated to each other. - To initiate the segmentation in the sectional
medical image 42, the user identifies a position in the sectionalmedical image 42 by means of theinput device 38 as a user input interface. InFIG. 2 b, the position is identified by anindicator 44 movable within the sectionalmedical image 42. Theindicator 44 shows the position identified by the user, which is in this particular case the portal vein of the liver which is also visible in the field of view of theultrasound image 40. - On the basis of the identified position, the segmentation of the 3D medical image data is initiated as shown in
FIG. 3b and also the segmentation of the vessels is performed in the ultrasound image data as shown inFIG. 3 a. Since the portal vein is identified by the user input and theindicator 44, the segmentation of this anatomical feature can be performed faster and with a higher reliability so that the overall reliability of the registration and correlation of the respective image data is improved. It shall be understood that the position of a certain anatomical feature can be identified within the sectionalmedical image 42 and/or in theultrasound image 40 so that the segmentation in general can be performed faster and with a higher reliability. - The anatomical features surrounding the identified position are segmented, wherein the anatomical features may be surrounding surfaces like the vessels or other anatomical surfaces within the patient's body. The ultrasound segmentation data is in
FIG. 3a denoted by 46 and the medical image segmentation data is inFIG. 3b denoted by 48. -
FIG. 4a shows theultrasound segmentation data 46 of the vessels derived from the ultrasound image data by means of theultrasound segmentation unit 24.FIG. 4b shows the medical image segmentation data derived from the 3D medical image data by means of the medicalimage segmentation unit 26. - The
ultrasound segmentation unit 24 determinescentre line 50 andbifurcations 52 from theultrasound segmentation data 46 and the medicalimage segmentation unit 26 determines from the medicalimage segmentation data 48centre lines 54 andbifurcations 56 as shown inFIG. 5 b. - The
registration unit 32 correlates thecentre lines bifurcations segmentation data FIG. 6 and thefusion unit 34 combines the ultrasound image data and the 3D medical image data on the basis of the correlation received from theregistration unit 32 and/or the position of theultrasound probe 14 determined by theposition determining unit 28. -
FIG. 7a shows theultrasound image 40 shown inFIG. 2a andFIG. 7b shows an image spatially aligned by thefusion unit 34 of theimage processing unit 22 on the basis of the correlation received from theregistration unit 32 . The correlation can be easily performed by the user input, since the segmentation effort is reduced and the reliability of the identification of significant anatomical features within the image data can be improved. - In
FIG. 8 a flow diagram of a method for evaluating medical image data is shown and generally denoted by 60. - The
method 60 starts with acquiring ultrasound data of the patient 12 by means of theultrasound probe 14 as shown atstep 62 and with receiving 3D medical image data of the patient 12 from theexternal database 20, which is usually MRT or CT data previously acquired from the patient 12 as shown atstep 64. At step 66 a position is identified in the 3D medical image data and/or the ultrasound image data by the user via theinput device 38 as shown atstep 66. - The anatomical features of the patient 12 are segmented in the ultrasound data and corresponding segmentation data of the anatomical features are provided as shown at
step 68. Further, anatomical features in the 3D medical image data are segmented and the medicalimage segmentation data 48 is provided as shown at thestep 70. The segmentation of the anatomical feature in the ultrasound data and/or in the 3D medical image data are based on the identified position, wherein the respective segmentation in the ultrasound data and/or the medical image data are initiated on the basis of the identified position. - Preferably, the anatomical features surrounding the identified position are segmented in order to segment only those anatomical features which are relevant and which are identified by the user. If the position is identified in the 3D medical image data, the segmentation of the ultrasound data may be performed in general or if the position is identified in the ultrasound image data, the segmentation of the medical image data may be performed in general. In a certain embodiment, the position to be segmented is identified in the 3D medical image data as well as the ultrasound image data.
- The ultrasound segmentation data and the medical image segmentation data are provided to the
registration unit 32, wherein theultrasound segmentation data 46 and the medicalimage segmentation data 48 are correlated atstep 72. - On the basis of the so correlated segmentation data, the calibration of the
position determining unit 28 can be performed and the fusion of the ultrasound image data and the 3D medical image data can be performed by thefusion unit 34. - While the invention has been illustrated and described in detail in the drawings and foregoing description, such illustration and description are to be considered illustrative or exemplary and not restrictive; the invention is not limited to the disclosed embodiments. Other variations to the disclosed embodiments can be understood and effected by those skilled in the art in practicing the claimed invention, from a study of the drawings, the disclosure, and the appended claims.
- In the claims, the word “comprising” does not exclude other elements or steps, and the indefinite article “a” or “an” does not exclude a plurality. A single element or other unit may fulfill the functions of several items recited in the claims. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.
- A computer program may be stored/distributed on a suitable medium, such as an optical storage medium or a solid-state medium supplied together with or as part of other hardware, but may also be distributed in other forms, such as via the Internet or other wired or wireless telecommunication systems.
- Any reference signs in the claims should not be construed as limiting the scope.
Claims (15)
1. A medical imaging apparatus for evaluating medical image data, comprising:
an ultrasound acquisition unit including an ultrasound probe for acquiring ultrasound image data of a patient,
an ultrasound segmentation unit for segmenting anatomical features of the patient in the ultrasound image data and for providing ultrasound segmentation data,
an image data interface for receiving 3D medical image data of the patient,
a medical image segmentation unit for segmenting the 3D medical image data and for providing medical image segmentation data,
a user input interface arranged to identify a position input by the user in the 3D medical image data and/or in the ultrasound image data in order to initiate the segmentation of anatomical features by the medical image segmentation unit and/or the ultrasound segmentation unit on the basis of the position identified by the user, wherein said segmentation of anatomical features is performed in the data adjacent to or surrounding the identified position in the respective image data,
a registration unit for correlating the ultrasound segmentation data and the medical image segmentation data.
2. The medical imaging apparatus as claimed in claim 1 , further comprising a position determining unit attached to the ultrasound probe for determining the position of the ultrasound probe, and a calibration unit for calibrating the position of the ultrasound probe on the basis of the correlation of the segmentation data received from the registration unit.
3. The medical imaging apparatus as claimed in claim 2 , further comprising a fusion unit for fusion of the ultrasound image data and the 3D medical image data on the basis of the position of the ultrasound probe determined by the position determining unit.
4. (canceled)
5. (canceled)
6. The medical imaging apparatus as claimed in claim 1 , wherein the anatomical features are surfaces in the vicinity of the identified position.
7. The medical imaging apparatus as claimed in claim 6 , wherein the anatomical features are vessels of the patient.
8. The medical imaging apparatus as claimed in claim 7 , wherein the ultrasound segmentation unit and the medical image segmentation unit are adapted to determine centre lines and/or bifurcations of the vessels and wherein the registration unit is adapted to register the ultrasound image data and the 3D medical image data on the basis of the determined centre lines and/or bifurcations of the vessels.
9. The medical imaging apparatus as claimed in claim 1 , wherein the user input interface comprises a display unit for displaying the 3D medical image data and/or the ultrasound image data and wherein the user interface comprises an input device for identifying the position in the 3D medical image data and/or the ultrasound image data at the display unit.
10. The medical imaging apparatus as claimed in claim 9 , wherein the input device is adapted to control a position of an indicator displayed at the display unit within the displayed image data and to identify the position in the displayed image data on the basis of the position of the indicator and a user input.
11. The medical imaging apparatus as claimed in claim 9 , wherein the display unit comprises a contact sensitive surface for identifying the position the displayed image data by a user input.
12. The medical imaging apparatus as claimed in claim 1 , wherein the 3D medical image data is previously acquired image data stored in a memory device.
13. The medical imaging apparatus as claimed in claim 1 , wherein the 3D medical image data is MR image data, CT image data, cone-beam CT image data or ultrasound image data.
14. A medical image evaluation method for evaluating medical image data, comprising the steps of:
acquiring ultrasound data of a patient,
receiving 3D medical image data of the patient,
identifying a position in the 3D medical image data and/or in the ultrasound image data by a user via a user input interface,
segmenting anatomical features of the patient in the ultrasound data and providing ultrasound segmentation data of the anatomical features,
segmenting anatomical features the 3D medical image data and providing medical image segmentation data,
wherein the segmentation of the anatomical features in the ultrasound data and/or the 3D medical image data is initiated on the basis of the position identified by the user and performed in the data adjacent to or surrounding the identified position in the respective image data, and
correlating the ultrasound segmentation data and the medical image segmentation data.
15. A computer program comprising program code means for causing a computer to carry out the steps of the method as claimed in claim 14 , when said computer program is carried out on a computer.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP14306376.6 | 2014-09-08 | ||
EP14306376 | 2014-09-08 | ||
PCT/EP2015/070362 WO2016037969A1 (en) | 2014-09-08 | 2015-09-07 | Medical imaging apparatus |
Publications (1)
Publication Number | Publication Date |
---|---|
US20180214129A1 true US20180214129A1 (en) | 2018-08-02 |
Family
ID=51582329
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/505,626 Abandoned US20180214129A1 (en) | 2014-09-08 | 2015-09-07 | Medical imaging apparatus |
Country Status (6)
Country | Link |
---|---|
US (1) | US20180214129A1 (en) |
EP (1) | EP3190973A1 (en) |
JP (1) | JP2017526440A (en) |
CN (1) | CN106687048A (en) |
RU (1) | RU2017111807A (en) |
WO (1) | WO2016037969A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020037668A1 (en) * | 2018-08-24 | 2020-02-27 | 深圳迈瑞生物医疗电子股份有限公司 | Ultrasound image processing device and method, and computer-readable storage medium |
US20220317294A1 (en) * | 2021-03-30 | 2022-10-06 | GE Precision Healthcare LLC | System And Method For Anatomically Aligned Multi-Planar Reconstruction Views For Ultrasound Imaging |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107548294B (en) * | 2015-03-31 | 2021-11-09 | 皇家飞利浦有限公司 | Medical imaging apparatus |
JP7270331B2 (en) * | 2017-06-15 | 2023-05-10 | キヤノンメディカルシステムズ株式会社 | Medical image diagnosis device and image processing device |
CN108833230A (en) * | 2018-06-27 | 2018-11-16 | 梧州井儿铺贸易有限公司 | A kind of smart home system |
CN112971982B (en) * | 2019-12-12 | 2022-08-19 | 珠海横乐医学科技有限公司 | Operation navigation system based on intrahepatic vascular registration |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7574026B2 (en) * | 2003-02-12 | 2009-08-11 | Koninklijke Philips Electronics N.V. | Method for the 3d modeling of a tubular structure |
US8411919B2 (en) * | 2008-07-07 | 2013-04-02 | Siemens Aktiengesellschaft | Fluid dynamics approach to image segmentation |
EP2225724A1 (en) * | 2007-12-18 | 2010-09-08 | Koninklijke Philips Electronics N.V. | System for multimodality fusion of imaging data based on statistical models of anatomy |
KR101121396B1 (en) * | 2009-07-31 | 2012-03-05 | 한국과학기술원 | System and method for providing 2-dimensional ct image corresponding to 2-dimensional ultrasound image |
US8751961B2 (en) * | 2012-01-30 | 2014-06-10 | Kabushiki Kaisha Toshiba | Selection of presets for the visualization of image data sets |
WO2013140315A1 (en) * | 2012-03-23 | 2013-09-26 | Koninklijke Philips N.V. | Calibration of tracked interventional ultrasound |
US9375195B2 (en) * | 2012-05-31 | 2016-06-28 | Siemens Medical Solutions Usa, Inc. | System and method for real-time ultrasound guided prostate needle biopsy based on biomechanical model of the prostate from magnetic resonance imaging data |
CN105025803B (en) * | 2013-02-28 | 2018-02-23 | 皇家飞利浦有限公司 | Segmentation from multiple 3-D views to blob |
-
2015
- 2015-09-07 EP EP15766425.1A patent/EP3190973A1/en not_active Withdrawn
- 2015-09-07 CN CN201580048332.XA patent/CN106687048A/en active Pending
- 2015-09-07 RU RU2017111807A patent/RU2017111807A/en not_active Application Discontinuation
- 2015-09-07 US US15/505,626 patent/US20180214129A1/en not_active Abandoned
- 2015-09-07 JP JP2017511345A patent/JP2017526440A/en active Pending
- 2015-09-07 WO PCT/EP2015/070362 patent/WO2016037969A1/en active Application Filing
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020037668A1 (en) * | 2018-08-24 | 2020-02-27 | 深圳迈瑞生物医疗电子股份有限公司 | Ultrasound image processing device and method, and computer-readable storage medium |
US12102484B2 (en) | 2018-08-24 | 2024-10-01 | Shenzhen Mindray Bio-Medical Electronics Co., Ltd. | Ultrasound image processing device and method, and computer-readable storage medium |
US20220317294A1 (en) * | 2021-03-30 | 2022-10-06 | GE Precision Healthcare LLC | System And Method For Anatomically Aligned Multi-Planar Reconstruction Views For Ultrasound Imaging |
Also Published As
Publication number | Publication date |
---|---|
CN106687048A (en) | 2017-05-17 |
RU2017111807A3 (en) | 2019-03-14 |
RU2017111807A (en) | 2018-10-10 |
WO2016037969A1 (en) | 2016-03-17 |
EP3190973A1 (en) | 2017-07-19 |
JP2017526440A (en) | 2017-09-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20180214129A1 (en) | Medical imaging apparatus | |
CN106456084B (en) | Ultrasonic imaging apparatus | |
US20170251988A1 (en) | Ultrasound imaging apparatus | |
US20100063387A1 (en) | Pointing device for medical imaging | |
EP3393367B1 (en) | Medical imaging apparatus and medical imaging method for inspecting a volume of a subject | |
US20170360396A1 (en) | Ultrasound imaging apparatus and method for segmenting anatomical objects | |
JP6293619B2 (en) | Image processing method, apparatus, and program | |
CN105684040B (en) | Method of supporting tumor response measurement | |
US20210015465A1 (en) | Medical imaging apparatus | |
EP2710553B1 (en) | Determination of a physically-varying anatomical structure | |
WO2016039763A1 (en) | Image registration fiducials | |
WO2017101990A1 (en) | Determination of registration accuracy | |
EP4128145B1 (en) | Combining angiographic information with fluoroscopic images | |
CN109727297B (en) | Medical image reconstruction method, system, readable storage medium and apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: KONINKLIJKE PHILIPS N.V., NETHERLANDS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DUFOUR, CECILE;MORY, BENOIT JEAN-DOMINIQUE BERTRAND MAURICE;NY, GARY CHENG-HOW;SIGNING DATES FROM 20150907 TO 20170216;REEL/FRAME:041329/0467 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |