WO2010134512A1 - 医用画像診断装置とその関心領域設定方法 - Google Patents
医用画像診断装置とその関心領域設定方法 Download PDFInfo
- Publication number
- WO2010134512A1 WO2010134512A1 PCT/JP2010/058335 JP2010058335W WO2010134512A1 WO 2010134512 A1 WO2010134512 A1 WO 2010134512A1 JP 2010058335 W JP2010058335 W JP 2010058335W WO 2010134512 A1 WO2010134512 A1 WO 2010134512A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- region
- unit
- dimensional
- cross
- Prior art date
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/08—Detecting organic movements or changes, e.g. tumours, cysts, swellings
- A61B8/0883—Detecting organic movements or changes, e.g. tumours, cysts, swellings for diagnosis of the heart
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
- A61B8/461—Displaying means of special interest
- A61B8/466—Displaying means of special interest adapted to display 3D data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
- A61B8/467—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient characterised by special input means
- A61B8/469—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient characterised by special input means for selection of a region of interest
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/48—Diagnostic techniques
- A61B8/483—Diagnostic techniques involving the acquisition of a 3D volume of data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/52—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/52017—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
- G01S7/52053—Display arrangements
- G01S7/52057—Cathode ray tube displays
- G01S7/5206—Two-dimensional coordinated display of distance and direction; B-scan display
- G01S7/52063—Sector scan display
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/52017—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
- G01S7/52053—Display arrangements
- G01S7/52057—Cathode ray tube displays
- G01S7/5206—Two-dimensional coordinated display of distance and direction; B-scan display
- G01S7/52066—Time-position or time-motion displays
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/89—Sonar systems specially adapted for specific applications for mapping or imaging
- G01S15/8906—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
- G01S15/8993—Three dimensional imaging systems
Definitions
- the present invention relates to a medical image diagnostic apparatus capable of improving operability for setting a region of interest (ROI) in a medical image and a region of interest setting method thereof.
- ROI region of interest
- Medical image diagnostic devices are used by examiners such as doctors and clinical technologists.
- the examiner sets an ROI for a part of the diagnostic region of the image.
- the ROI is set when the examiner traces a part of the image displayed on the display screen using a pointing device.
- the tracing operation performed by the examiner is called manual setting of ROI.
- manual setting of the ROI is a complicated operation for the examiner.
- the ROI set for a moving organ such as the heart has regularity in the size and displacement of the ROI based on the movement of the heart and the like. Therefore, it is possible to set an ROI in a moving organ by creating a computer image processing program according to the regularity described above and executing the image processing program.
- extraction of ROI by an image processing program for image processing is referred to as “ROI automatic setting”.
- Patent Document 1 The method for automatically setting ROI is proposed in Patent Document 1, for example.
- an automatic contour tracking (ACT) method is used to extract the boundary between the heart chamber and the myocardium based on the density gradient on the ultrasound image, and the computer (CPU ).
- Patent Document 1 the image to be handled is a two-dimensional image, and there is no suggestion about automatic ROI setting for a three-dimensional image.
- an object of the present invention is to provide a medical image diagnostic apparatus capable of automatically setting an ROI for a three-dimensional image of a moving organ and a region of interest setting method thereof.
- the present invention generates a reference cross-sectional image from a three-dimensional image of a moving organ, and further divides the generated reference cross-sectional image into a plurality of regions based on a region division criterion.
- a region having a different motion state among a plurality of regions is specified, and a region of interest is set in a region on the medical image including the specified region.
- the medical image diagnostic apparatus of the present invention includes a medical image acquisition unit that acquires a medical image, a 3D image configuration unit that configures a 3D image including a moving organ region in the medical image, and the tertiary A cross-sectional image generating unit that generates a two-dimensional cross-sectional image as a reference image from the original image, a region dividing unit that divides the reference image into a plurality of regions based on a region dividing criterion, and movement of the plurality of regions
- a region-of-interest setting unit that calculates a state, specifies at least one region of the plurality of regions based on the calculated motion state, and sets a region of the medical image including the specified region as a region of interest; , Provided.
- a medical image is acquired by a medical image acquisition unit, a 3D image including a moving organ region in the medical image is configured by a 3D image configuration unit, and a cross-sectional image image generation unit To generate a two-dimensional cross-sectional image as a reference image from the three-dimensional image, divide the reference image into a plurality of regions based on a region division criterion, and the region of interest setting unit to the plurality of regions Calculating at least one of the plurality of regions based on the calculated motion state, and setting the region of the medical image including the specified region as a region of interest, Divide a 3D image of a moving organ into multiple regions based on a predetermined region division criterion, calculate the motion state of each region in each region, and set the region where the motion state is different from other regions as ROI Kill.
- the region-of-interest setting method of the present invention includes a first step of acquiring a medical image by a medical image acquisition unit, and a three-dimensional image including a moving organ region in the medical image by a three-dimensional image configuration unit.
- a second step a third step of generating a two-dimensional cross-sectional image as a reference image from the three-dimensional image by the cross-sectional image image generating unit, and a plurality of the reference images based on a region division criterion by the region dividing unit
- the first step acquires a medical image by the medical image acquisition unit
- the second step includes a three-dimensional image including a moving organ region in the medical image by the three-dimensional image configuration unit.
- An image is constructed, and a third step generates a two-dimensional cross-sectional image as a reference image from the three-dimensional image by the cross-sectional image image generation unit, and a fourth step performs region division of the reference image by the region division unit.
- the fifth step calculates the motion state of the plurality of regions by the region of interest setting unit, and at least one region of the plurality of regions based on the calculated motion state And the region of the medical image including the specified region is set as a region of interest, thereby dividing the three-dimensional image of the moving organ into a plurality of regions based on a predetermined region division criterion.
- the dynamic state is calculated in each region can be set a region where motion state is different from the other regions as ROI.
- the present invention divides a three-dimensional image of a moving organ into a plurality of regions based on a predetermined region division criterion, calculates a motion state of each of the plurality of regions, and determines a region where the motion state is different from other regions.
- FIG. 1 is an example of a system configuration diagram of an ultrasonic diagnostic imaging apparatus according to Embodiment 1 of the present invention.
- the figure which shows an example of the setting of the outline of FIG. Display example of measurement processing of the ultrasonic diagnostic imaging apparatus of Embodiment 1 of the present invention
- Flowchart of measurement processing of ultrasonic diagnostic imaging apparatus of embodiment 2 of the present invention The figure which shows an example of the setting of the outline of FIG.
- Examples of the medical image diagnostic apparatus of the present invention include an ultrasonic diagnostic apparatus, an X-ray CT apparatus, and an MRI apparatus.
- an ultrasonic diagnostic apparatus is illustrated as an example of the medical image diagnostic apparatus.
- Example 1 a three-dimensional image including a moving organ of a subject is acquired by an ultrasonic diagnostic apparatus, and a two-dimensional reference image is extracted from the acquired three-dimensional image by a CPU mounted on the ultrasonic diagnostic apparatus, A case where the ROI is automatically set by the CPU mounted on the ultrasonic diagnostic apparatus from the extracted two-dimensional reference image will be described.
- FIG. 1 is a block diagram illustrating an outline of an ultrasonic diagnostic apparatus according to the present embodiment.
- the ultrasonic diagnostic apparatus 1 includes an ultrasonic signal generation unit 2, an ultrasonic image generation unit 3, a calculation unit 4, a storage unit 5, a setting unit 6, a display unit 7, and a control unit 8. ing.
- solid arrows indicate control, and white arrows indicate the flow of image signal data.
- the ultrasonic signal generation unit 2 includes an ultrasonic probe 21 and an ultrasonic signal transmission / reception unit 23.
- the ultrasonic probe 21 transmits ultrasonic waves to the subject 9 and receives a reception signal from the subject 9.
- the ultrasonic signal transmission / reception unit 23 passes the reception signal received by the ultrasonic probe 21 through a phasing addition circuit (not shown) to obtain a three-dimensional ultrasonic signal.
- the types of the ultrasonic probe 21 are classified according to the arrangement direction of a plurality of transducers. Specifically, there are a two-dimensional ultrasonic probe in which a plurality of transducer elements are two-dimensionally arranged, and a one-dimensional ultrasonic probe in which a plurality of transducers are arranged one-dimensionally.
- the two-dimensional ultrasonic probe can transmit and receive ultrasonic waves toward a three-dimensional space, and can directly obtain a three-dimensional ultrasonic signal. Therefore, the two-dimensional ultrasonic probe is suitable as an ultrasonic probe used in the present invention. .
- a two-dimensional ultrasonic signal of the subject can be obtained.
- a method of obtaining a three-dimensional ultrasonic signal with a one-dimensional ultrasonic probe is obtained by sequentially obtaining a two-dimensional ultrasonic signal of a subject in an orthogonal direction orthogonal to the arrangement direction of the transducers, and storing the ultrasonic signal transmitting / receiving unit 23. 5, the phasing addition circuit sequentially calculates the two-dimensional ultrasonic signals of the subject obtained in the orthogonal direction in the orthogonal direction, and performs an operation for forming a three-dimensional ultrasonic signal.
- the ultrasonic image generation unit 3 generates a three-dimensional image composed of voxel data from the three-dimensional ultrasonic signal input from the ultrasonic signal generation unit 2 based on the conditions set in the setting unit 6.
- the calculation unit 4 includes a two-dimensional section extraction unit 41, a two-dimensional contour extraction unit 43, and an ROI / measurement value calculation unit 45.
- the two-dimensional section extraction unit 41 performs a calculation for extracting a signal of a specific section from the three-dimensional ultrasonic signal.
- the specific cross section refers to a reference image acquired by echocardiography, an image of the apex 2 cavities (A2C), and an image of the 4 apex cavities (A4C).
- the A2C image and the A4C image are in a positional relationship orthogonal to each other.
- the classification of each image is performed by a known image recognition technique such as a block matching method. Specifically, A2C and A4C templates stored in the database are compared with 3D ultrasound signals, and the result of the comparison operation is the 2D formed by the 3D ultrasound signal with the highest similarity.
- the images are A2C images and A4C images.
- the two-dimensional contour extraction unit 43 extracts the endocardial and epicardial contours of the heart on A2C and A4C.
- ASE fractionation method a fractionation method recommended by American Society Echo Cardiography (ASE) (referred to as “ASE fractionation method”) is used as the predetermined area division criterion.
- region which fractionated the myocardium is called myocardial fractionation.
- the ROI / measurement value calculation unit 45 measures the ROI and the size and movement of the moving organ using the extracted endocardial and outer membrane contours and myocardial fraction.
- the myocardial fraction is obtained, for example, by the ASE fractionation method.
- the ASE fractionation method is exemplified as a method of local region division.
- the local region division method labels (extracts) a region where pixels are connected by assigning a label (number) to each pixel.
- a known region division method such as a K-average method for classifying the number of clusters into a given number K using the average of clusters may be adopted.
- the region division by the ASE fractionation method is performed by an outline / fraction position setting unit 63 of the setting unit 6 described later.
- the contour line / fraction position setting unit 63 traces the organ region drawn on the ultrasonic image displayed on the display unit 7.
- an example of an organ is a heart.
- the contour line / fraction position setting unit 63 performs an operation of tracing the position of the endocardium and epicardium on the image where the heart is depicted on the image.
- the endocardial and epicardial positional information is indicated by double thick lines as shown in the cross-sectional images 301a and 301b in FIG.
- the outer thick line 302 indicates the epicardium, and the inner thick line 303 indicates the endocardium.
- the position information of the endocardium and the epicardium means a position where the lumen portion and the myocardial portion of the heart, which is the target of heart volume measurement, are separated.
- Manual operation is an operation method in which the examiner manually traces all endocardial and epicardial positional information using a pointing device. Specifically, the examiner traces the boundary of the region corresponding to the endocardium and the epicardium while referring to the image of the heart region of the ultrasonic image displayed on the display unit 7, Position information of endocardium and epicardium is input.
- the control unit 8 temporarily stores the input endocardial and epicardial positional information in the storage unit 5.
- the examiner inputs a plurality of points on the boundary of the endocardial or epicardial region using a pointing device, and the endocardium or heart This is an operation method for extracting the boundary of the outer membrane region.
- the examiner refers to the image of the heart region of the ultrasound image displayed on the display unit 7 while the examiner corresponds to the endocardial and epicardial regions, and the adjacent region. Enter multiple boundary points.
- the control unit 8 receives a plurality of input boundary points, connects the boundary points, and performs interpolation calculation such as spline interpolation to obtain the boundary line of the region as endocardial and epicardial position information. To do.
- the control unit 8 temporarily stores the input endocardial and epicardial positional information in the storage unit 5.
- the examiner inputs a pixel point in the endocardial or epicardial region using a pointing device, and the CPU inputs the pixel point in the endocardial or epicardial region.
- This is an operation method for extracting a boundary.
- the examiner refers to the image of the heart region of the ultrasonic image displayed on the display unit 7 while the examiner specifies one point in order to identify the regions corresponding to the endocardium and the epicardium. Make input.
- the input 1 point becomes a seed in the region growing method.
- the control unit 8 causes the calculation unit 4 to perform region extraction calculation by the region growing method using the seed so as to obtain the boundary line of the region as position information of the endocardium and the epicardium.
- the control unit 8 temporarily stores the obtained endocardial and epicardial positional information in the storage unit 5.
- An example of a predetermined segmentation index is the ASE myocardium by the 16 fraction method or the 17 fraction method.
- the 17-fraction method is becoming the industry standard for measuring myocardium with medical diagnostic imaging equipment.
- the application of the 17 fraction method to the myocardium is performed by the examiner directly setting the position of the 17 fraction to the myocardium on the screen while referring to the image on the image display unit 71 and inputting the position of the fraction. Done.
- the storage unit 5 includes a program storage unit 51, a database unit 53, and an ROI / measurement value storage unit 55.
- Specific hardware of the storage unit 5 is a storage medium such as a semiconductor memory, a hard disk, and an optical disk.
- the program storage unit 51 stores a program describing an algorithm such as contour extraction processing and measurement calculation in the calculation unit 4 and a program for controlling each unit.
- the database unit 53 stores the local position information of the heart including the information of the two-dimensional cross-sectional position and the fractional position information of the myocardial fraction, and the contour data of the two-dimensional contour shape when contour extraction using a contour model is applied. Has been.
- the ROI / measurement value storage unit 55 stores the measurement value calculated by the ROI / measurement value calculation unit 45.
- the setting unit 6 includes a measurement condition setting unit 61 and an outline / fraction position setting unit 63.
- the setting unit 6 is a user interface, and specific hardware is an information input device including a keyboard, a trackball, and a switch.
- the measurement condition setting unit 61 is used when the examiner manually sets parameters, and the set parameters are transmitted to the control unit 8.
- the contour line / fraction position setting unit 63 is used to finely adjust the position manually when the contour or fraction position extracted from the two-dimensional image is not accurately set. It is done.
- the display unit 7 includes an image display unit 71 and an ROI / measurement value display unit 73.
- the hardware of the display unit 7 is a display device such as a CRT display, a liquid crystal display, a plasma display, or an organic EL display.
- the image display unit 71 selectively displays a three-dimensional image on which a three-dimensional contour surface is superimposed and a two-dimensional cross-sectional image on which a two-dimensional contour line is superimposed.
- the ROI / measurement value display unit 73 displays the measurement values calculated by the ROI / measurement value calculation unit 45 in a graph or table format together with the image group displayed by the image display unit 71.
- the control unit 8 is connected to each component of the ultrasonic signal generation unit 2, the ultrasonic image generation unit 3, the calculation unit 4, the storage unit 5, the setting unit 6, and the display unit 7 so that each component functions. To oversee and control.
- Specific hardware of the control unit 8 is a CPU of a computer system.
- FIG. 2 is a flowchart of the measurement process of the ultrasonic diagnostic imaging apparatus according to Embodiment 1 of the present invention.
- FIG. 3 is a diagram showing an example of setting of the contour line in FIG.
- FIG. 4 is a display example of measurement processing of the ultrasonic diagnostic imaging apparatus according to Embodiment 1 of the present invention.
- FIG. 2 (a) is a flowchart for explaining from the generation of the long-axis image and the short-axis image model to the registration of each model and contour in the database unit 53 (referred to as “database unit registration process”).
- FIG. 2 (b) is a flowchart for explaining from the extraction of the long axis image and the short axis image to the display of the measurement result on the ROI / measurement value display unit 73 (referred to as “ROI automatic setting process”).
- the database part registration process is executed by the following procedure shown in FIG. (Step 201)
- the control unit 8 stores a long axis image such as the A2C image 301b and the A4C image 301a and a model indicating the shape of the short axis image orthogonal to the long axis image in the database unit 53.
- Examples of short-axis images include Apex (short-axis image apex partial image), Mid (short-axis image papillary muscle partial image), and Base (short-axis image center base partial image).
- Apex, Mid, and Base are provided at levels 309a, 309b, and 309c at different positions in the long axis direction in the left ventricle.
- the control unit 8 generates a long-axis image contour line based on the ASE fractionation method and stores it in the storage unit 5.
- the ASE fractionation method mainly shows that the left ventricular myocardium is classified.
- the myocardium exists between the epicardium and endocardium, indicated by bold lines 302 and 303.
- the left ventricle is divided into seven myocardial regions a to g by a fractional boundary 308 (illustrated by dotted lines) as shown in the upper left A4C image 301a in FIG. . That is, regions a to g are myocardial fractions. Contour points are provided on the contour lines indicated by bold lines 302 and 303.
- an A2C image having a positional relationship orthogonal to the A4C image is shown.
- the A2C image is also divided into seven myocardial regions a to g as a myocardial fraction by a segmentation boundary 308 (illustrated by dotted lines).
- a new reference cross-section type of a long-axis image such as A3C (apex apex long-axis image) may be defined.
- the long-axis image contour line is set at the boundary portion of the myocardial region as shown in the images 301a and 301b.
- Step 205 The control unit 8 generates a short-axis image contour line based on the ASE fractionation method and stores it in the storage unit 5. Also in the short axis image, the myocardium exists between the epicardium and endocardium indicated by the thick lines 306 and 307, as in the case of the long axis image.
- the short-axis image is divided into six myocardial regions using the positional relationship between the A2C image 301b orthogonal to the A4C image 301a of the long-axis image. Specifically, description will be made using a coordinate system in which the positional relationship between the A4C image and the A2C image of the long-axis image in the lower right of FIG. 3 is shown. In the lower right coordinate system of FIG.
- the position of the A4C image is on the vertical axis
- the position of the A2C image is on the horizontal axis.
- the outline of the short axis image indicated by the thick lines 306 and 307 intersects the vertical axis and the horizontal axis with 8 points indicated by black circles on the lower right coordinate system of FIG. Using the relative positions with respect to these eight points, the myocardial region of the short-axis image is divided into six.
- the short axis image contour line is set at the boundary portion of the myocardial region as indicated by the thick lines 306 and 307.
- Step 207 The control unit 8 associates the model of the long axis image and the short axis image created in Step 201 with the contour line of the long axis image created in Step 203 and the contour line of the short axis image created in Step 205 to the database unit 53. And registered as contour models 304 and 305.
- the examiner manually operates the measurement condition setting unit 61 and sets parameters for acquiring the ultrasonic signal in the control unit 8.
- the control unit 8 receives the set parameters and drives the ultrasonic probe 21 to the ultrasonic signal transmitting / receiving unit 23.
- the ultrasonic probe 21 is switched between transmission of ultrasonic waves and reception of reflected signals from the subject, that is, transmission and reception periods.
- the ultrasonic probe 21 transmits ultrasonic waves to a diagnosis site (eg, heart) of the subject during the transmission period.
- the ultrasonic probe 21 receives the reflected signal from the subject during the reception period.
- the ultrasonic signal transmitting / receiving unit 23 adjusts the phase of the received reflected signal to obtain a three-dimensional ultrasonic signal.
- a medical image is acquired by a medical image acquisition unit.
- the step of obtaining a three-dimensional ultrasonic signal is an example of a step of acquiring a medical image by a medical image acquisition unit.
- the control unit 8 transmits, to the ultrasonic image generation unit 3, a three-dimensional image composed of voxel data based on the conditions set in the setting unit 6, from the three-dimensional ultrasonic signal input from the ultrasonic signal transmission / reception unit 23. Generate.
- the three-dimensional image is generated by a three-dimensional image constructing unit that constitutes a three-dimensional image including a moving organ region in the medical image.
- the step of generating a three-dimensional image is a step of forming a three-dimensional image including a moving organ region in the medical image by the three-dimensional image forming unit.
- the control unit 8 causes the two-dimensional section extraction unit 41 to perform an operation of extracting an A2C image and an A4C image from the three-dimensional image.
- a cross-sectional image generation unit generates a two-dimensional cross-sectional image that is a reference image such as A2C or A4C from the three-dimensional image.
- the step of generating a reference image is a step of generating a two-dimensional cross-sectional image that becomes a reference image from the three-dimensional image by the cross-sectional image image generating unit.
- the short axis images Apex, Mid, and Base are in a positional relationship orthogonal to the long axis images such as A2C images and A4C images, and are provided at different positions in the long axis direction in the left ventricle.
- the short axis images Apex, Mid, and Base are extracted from the apex side with their respective positional relationships.
- Step 212 The examiner can manually operate the measurement condition setting unit 61 and set position information for fine adjustment of the position of the myocardium in the control unit 8.
- the control unit 8 receives the set position information and resets the initial position of the myocardial contour model.
- the accuracy of the extracted contour after the contour deformation can be improved by first specifying the rough position of the myocardium in the image.
- the manual operation in step 212 is not essential, and the myocardial position may be extracted by the control unit 8 using a known image recognition technique.
- the control unit 8 causes the two-dimensional contour extraction unit 43 to extract the endocardial and epicardial contours on the A2C image and the A4C image.
- the contour line extraction method a method using edge detection processing for detecting a change in the image luminance value of the film surface, template matching, and a contour model can be applied.
- a method using a contour model will be described as an example.
- the contour model is a representation of the contour shape of the object to be extracted and the law of luminance value in a generalized form.
- the contour of the heart can be extracted while adaptively deforming the contour according to the shape of the contour model, that is, the actual shape of the heart.
- a method of generating contour models by learning contour data extracted in the past can also be used.
- the upper right of FIG. 3 is an example of a contour model stored in the database unit 53, which is a long-axis image contour model 304 and a short-axis image contour model 305.
- a contour model stored in the database unit 53 which is a long-axis image contour model 304 and a short-axis image contour model 305.
- the endocardium since the endocardium has relatively high brightness of the myocardium and the heart chamber, it can be said that the endocardium is extracted in comparison with the epicardium, and the endocardial extraction accuracy is higher than that of the epicardium.
- the contour model is stored as a contour model that associates the endocardium with the epicardium, so that the epicardial contour extraction is complemented with the endocardial contour extraction data while the epicardial extraction is performed. The accuracy is improved.
- Step 215 The control unit 8 causes the two-dimensional contour extraction unit 43 to extract the endocardial and epicardial contour on the short-axis image.
- the contour model stores the fractional boundary 308 and the short-axis image level positions 309a to 309c (in this example, three stages of Apex, Mid, and Base) along with the contour line. Has been.
- the contour model deforms according to the deformation of the left ventricle of the image.
- the contour on the A4C image 301a and the contour on the A2C image 301b are extracted, and at the same time, the positions 309a to 309c of the short axis image level and the fractional boundary 308 are also determined. Is done.
- the determined separation boundary 308 serves as a reference for region division into a plurality of regions, and the reference image of the A4C image is divided into a plurality of regions along the region division reference.
- Steps 212 to 215. are an example of a process in which the reference image is divided into a plurality of regions based on the reference for region division by the region dividing unit.
- Step 217) The control unit 8 causes the display unit 7 to display a long axis image and a short axis image.
- the long-axis images (A2C image, A4C image) are displayed on the display screen 401 of the ultrasonic diagnostic apparatus in FIG.
- Short-axis images (Apex, Mid, and Base images) are displayed as reference numerals 404, 405, and 406, respectively, on the display screen 401 of the ultrasonic diagnostic apparatus in FIG.
- a three-dimensional image may be displayed on the display screen 401 as a reference numeral 407.
- Step 2128 The examiner can manually operate the measurement condition setting unit 61 and set position information for fine adjustment of the contour position or the myocardial fraction position in the control unit 8.
- the control unit 8 receives the set position information and finely adjusts the contour position or the myocardial fraction position.
- the manual operation in step 218 is not essential, and execution can be omitted if fine adjustment of the contour position or myocardial fraction position is unnecessary.
- the control unit 8 causes the ROI / measurement value calculation unit 45 to measure the area defined by the contour surface, the contour line, and the fractionation position.
- the ROI / measurement value calculation unit 45 measures the movement for each myocardial fraction, and automatically sets the ROI based on the myocardial fraction that exhibits extremely fast or slow movement compared to the surrounding myocardial fraction.
- the movement for each myocardial fraction can be measured by the myocardial tracking method described below.
- the myocardial tracking method is a method for extracting feature points appearing on an image frame.
- the ROI / measurement value calculation unit 45 performs the detection of the feature point for each frame and tracks the movement of the feature point. For example, in the case of the heart, there is a large difference in the intensity (amplitude) of the echo signal between the myocardial tissue and the blood flow part inside the heart, so by setting a threshold for the echo signal, the heart that is the boundary between these two parts The position of the intima can be detected as a feature point.
- the ROI / measurement value calculation unit 45 tracks feature points of the subject tissue and outputs tissue displacement information.
- the ROI / measurement value calculation unit 45 calculates tissue displacement information such as tissue motion speed using the inter-frame movement amount for each of a plurality of myocardial fractions divided by the fraction boundary 308. Furthermore, the ROI / measurement value calculation unit 45 calculates statistical values such as an average value, a variance value, and a median value from the motion speeds of a plurality of myocardial fractions. Calculate the specific region position information of the fast or slow myocardial region. The ROI / measurement value calculation unit 45 automatically sets a region corresponding to the singular region information position information as the ROI. Further, the area set as the ROI is not limited to one, but may be plural.
- the ROI / measurement value calculation unit 45 can calculate the distance between the endocardium and the epicardium, that is, the thickness of the myocardium because the position of the myocardium is specified by the endocardium and the epicardium.
- the ROI / measurement value calculation unit 45 calculates the myocardial weight by subtracting the volume surrounded by the endocardium from the volume of the region surrounded by the epicardium and multiplying the subtracted volume by the specific gravity of the known myocardium. Is also possible.
- the ROI / measurement value calculation unit 45 can also calculate the various measurement values at a specific fraction position.
- the ROI / measurement value calculation unit 45 can be applied to measurement with a two-dimensional contour line. As a result, it is possible to make a detailed diagnosis with reference to the three-dimensional measurement at the same time while performing a two-dimensional diagnosis established conventionally.
- the ROI / measurement value calculation unit 45 can apply a method of calculating the movement amount by causing the contour surface or the contour line to follow the motion of the heart.
- the movement amount may be calculated using a tracking calculation such as speckle tracking that has been conventionally proposed.
- speckle tracking a tracking calculation such as speckle tracking that has been conventionally proposed.
- the time change of a measured value can be measured.
- indices such as volume change, strain, ejection fraction.
- the region-of-interest setting unit calculates the motion state of the plurality of regions, specifies at least one region of the plurality of regions based on the calculated motion state, and is specified.
- a region of the medical image including a region is set as a region of interest.
- the region-of-interest setting unit calculates the motion state of the plurality of regions, specifies at least one region of the plurality of regions based on the calculated motion state, and is specified It is an example of the process of setting the area
- the control unit 8 displays the ROI and measurement values on the display unit 7 in accordance with the long axis image and the short axis image.
- the ROI is indicated by reference numeral 409 in FIG.
- the ROI 409 is a region indicated by a dotted circle including the myocardial region f.
- the display example of the ROI 409 is indicated by a dotted line, but may be a color when the background image is black and white, or may be a line segment such as a solid line or a one-dot chain line instead of a dotted line.
- the shape of the ROI 409 is not limited to a circle, but may be a rectangle or a shape along or approximate to the extracted contour by extracting the contour of an organ or an organ by a separate method.
- the measurement value may be displayed as a time change rate of the volume by a graph display as indicated by reference numeral 40A, or may be displayed as various numerical values of the volume, area, myocardial mass, and cardiac ejection ratio as indicated by reference numeral 40C. good.
- a biological signal such as an electrocardiographic waveform indicated by reference numeral 40B may be displayed together with these.
- FIG. 5 is a flowchart of the measurement process of the ultrasonic image diagnostic apparatus according to the second embodiment of the present invention.
- FIG. 6 is a diagram showing an example of setting of the contour line in FIG.
- Step 511 The control unit 8 extracts a long-axis image (A2C image and A4C image) from the three-dimensional ultrasonic signal by known image recognition.
- the control unit 8 displays the extracted long axis image on the display unit 7.
- the examiner manually sets the endocardial and epicardial contours on the A2C image and the A4C image using the measurement condition setting unit 61 for the long-axis image displayed on the display unit 7. Further, the examiner sets the fraction boundary of the myocardial fraction (reference numeral 308 in FIG. 4 of Example 1) using the measurement condition setting unit 61.
- Step 513 The examiner sets the position of the short axis image (Apex, Mid, Base) with respect to the long axis image displayed on the display unit 7 by using the measurement condition setting unit 61.
- the control unit 8 displays a short axis image of the set position on the display unit 7.
- the examiner extracts the endocardial and epicardial contour on each short axis image using the measurement condition setting unit 61 with respect to the short axis image displayed on the display unit 7.
- the endocardial and epicardial contours of the A2C image and the A4C image obtained in step 511 intersect with the short-axis cross section 601 at eight intersections as shown in the lower left of FIG.
- the examiner sets contour points so that the short axis image displayed on the display unit 7 passes 8 points using the measurement condition setting unit 61.
- the fraction boundary 308 of the myocardial fraction is also set on the short axis image as shown in the lower right of FIG.
- Step 515 (Step 515)-(Step 51B)
- step 511 or step 513 may refer to a model stored in the database unit 53.
- a three-dimensional image of a moving organ can be divided into a plurality of regions, and a unique divided region among the plurality of regions can be set as an ROI. Further, the examiner can select reference or non-reference of the database unit 53, and the examiner's degree of freedom of operation is widened.
- the positional relationship between the long-axis images or the short-axis images is not necessarily orthogonal, and if the angle has a specific relationship, the type of the reference cross section can be freely determined regardless of whether the angle is orthogonal or orthogonal. . Since the difference between the first embodiment and the third embodiment is only the positional relationship between orthogonal and non-orthogonal, only the positional relationship will be described.
- FIG. 7 is a diagram illustrating an example of setting of the contour line according to the third embodiment of the present invention. For example, when a three-dimensional contour surface is cut along a long-axis cross section that obliquely intersects A4C as shown in the lower left of FIG. 7, an oblique coordinate axis as shown in the lower right of FIG. If the relative position between the intersection 707 between the oblique coordinate axis and the short-axis image contour line and the division boundary is stored in advance, the myocardial region of the short-axis image can be divided using the relative position.
- the first embodiment an example in which two long-axis images are displayed has been described. However, it is not always necessary to display two long-axis images, and a short-axis image can be set if at least one is displayed. The difference between the first embodiment and the fourth embodiment is only whether to display two long-axis images or one long-axis image.
- FIG. 8 is a display example of measurement processing of the ultrasonic diagnostic imaging apparatus according to Embodiment 4 of the present invention.
- FIG. 8 shows an example in which ROI 809 is displayed in the A4C image as a result of obtaining ROI.
- a three-dimensional image of a moving organ can be divided into a plurality of regions, and a unique divided region among the plurality of regions can be set as an ROI.
- Example 1 describes an example in which two long-axis images are displayed
- Example 4 describes an example in which one long-axis image is displayed.
- the short-axis image is automatically set if the geometric settings such as dividing the left ventricle into four equal parts as in Example 1 are set in advance. it can.
- the only difference between the first embodiment and the fifth embodiment is whether to display two long-axis images or not to display a long-axis image.
- FIG. 9 is a display example of measurement processing of the ultrasonic diagnostic imaging apparatus according to Embodiment 5 of the present invention.
- a button “ROI automatic setting” is prepared in the operation unit 6, and the examiner operates the “ROI automatic setting” button.
- FIG. 9 shows an example in which ROI 909 is displayed in the A4C image as a result of obtaining ROI.
- the heart has been described as an example of a moving organ.
- an organ that moves with the movement of a moving organ without moving itself, or an organ or an organ that moves with the movement of respiratory motion is also included. Shall be.
- the present invention can be used for various medical image diagnostic apparatuses such as an ultrasonic diagnostic apparatus, an X-ray CT apparatus, and an MRI apparatus. In addition, it can also be used for information devices that can perform image processing on images obtained from medical image diagnostic apparatuses such as computers and various portable terminals.
- ultrasonic diagnostic device 1 ultrasonic diagnostic device 2 ultrasonic signal generator 3 ultrasonic image generator 4 arithmetic unit 5 storage unit 6 setting unit 7 display unit 8 control unit
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Biomedical Technology (AREA)
- Molecular Biology (AREA)
- Veterinary Medicine (AREA)
- Public Health (AREA)
- Biophysics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Pathology (AREA)
- Radiology & Medical Imaging (AREA)
- General Health & Medical Sciences (AREA)
- Heart & Thoracic Surgery (AREA)
- Medical Informatics (AREA)
- Animal Behavior & Ethology (AREA)
- Surgery (AREA)
- Remote Sensing (AREA)
- Computer Networks & Wireless Communication (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Computer Graphics (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Cardiology (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Description
本発明の医用画像診断装置は、超音波診断装置、X線CT装置、MRI装置などがある。本発明の実施の形態では、医用画像診断装置のうち超音波診断装置を例示する。
超音波診断装置1は、図1に示されるように、超音波信号生成部2、超音波画像生成部3、演算部4、記憶部5、設定部6、表示部7及び制御部8を備えている。図中の実線矢印は制御を、白抜き矢印は画像信号データの流れをそれぞれ示している。
超音波探触子21は、被検体9へ超音波を送信し、被検体9からの受信信号を受信する。超音波信号送受信部23は、超音波探触子21によって受信された受信信号を整相加算回路(図示しない)に通して、三次元超音波信号を得る。
本明細書では、所定の領域分割基準として、例えば、アメリカン・ソサエティ・エコーカーディオグラフィ(ASE)が推奨するような分画方法 (「ASE分画法」と称している)を用いる。心筋を分画した領域を心筋分画と呼ぶ。
図2は本発明の実施例1の超音波画像診断装置の計測処理のフローチャートである。
図3は図2の輪郭線の設定の一例を示す図である。図4は本発明の実施例1の超音波画像診断装置の計測処理の表示例である。
(ステップ201)
制御部8は、A2Cの画像301b、A4Cの画像301aなどの長軸像と、長軸像と直交する短軸像の形状を示すモデルをデータベース部53に記憶する。短軸像の例としてはApex(短軸像心尖部分像)、Mid(短軸像乳頭筋部分像)、Base(短軸像心基部部分像)が挙げられる。Apex、Mid、Baseは左心室において長軸方向の異なる位置のレベル309a、309b、309cに設けられるとする。
制御部8は、ASE分画法に基づいて長軸像輪郭線を生成し、記憶部5に記憶する。ASE分画法は主に左心室の心筋について区分することを示している。心筋は太線302と303で示される心外膜と心内膜の間に存在する。17分画法によれば左心室は、図3の左上方のA4Cの画像301aに示されるように、分画境界308(点線で図示)により7個の心筋の領域a~gに分割される。すなわち領域a~gが心筋分画となる。太線302と303で示した輪郭線上には輪郭点が設けられる。A4Cの画像の右隣には、A4Cの画像と直交する位置関係にあるA2Cの画像が示されている。A2Cの画像も分画境界308(点線で図示)により心筋分画として7個の心筋の領域a~gに分割される。また、図示しないが、A3C(心尖部長軸像)など長軸像の新たな基準断面の種類を定義してもよい。以上説明したように長軸像輪郭線は、画像301a及び301bに示されるように、心筋領域の境界部分に設定される。
制御部8は、ASE分画法に基づいて短軸像輪郭線を生成し、記憶部5に記憶する。短軸像においても心筋は、長軸画像の場合と同様に、太線306と307で示される心外膜と心内膜の間に存在する。短軸像は、長軸像のA4Cの画像301aと直交するA2Cの画像301bの位置関係を利用して6個の心筋の領域に分割される。具体的には、図3の右下の長軸像のA4Cの画像とA2Cの画像の位置関係が示される座標系を用いて説明する。図3の右下の座標系はA4Cの画像の位置を縦軸に、A2Cの画像の位置を横軸としている。太線306と307で示した短軸像の輪郭線は縦軸及び横軸と、図3の右下の座標系上の黒丸で示した8点と交わる。この8点に対する相対位置を利用して、短軸像の心筋領域は6個に分割される。
制御部8は、ステップ201で作成した長軸像及び短軸像のモデルと、ステップ203で作成した長軸像の輪郭線、ステップ205で作成した短軸像の輪郭線をデータベース部53に関連づけて、輪郭モデル304及び305として登録する。
(ステップ211)
検者は、計測条件設定部61を手動操作して超音波信号取得のためのパラメータを制御部8に設定する。制御部8は、設定されたパラメータを受けて、超音波探触子21を超音波信号送受信部23に駆動させる。超音波探触子21は超音波の送信、被検体からの反射信号の受信、すなわち送受信の期間が切替られる。超音波探触子21は送信期間に被検体の診断部位(例えば心臓など)に超音波を送信する。超音波探触子21は受信期間に被検体からの反射信号を受信する。超音波信号送受信部23は受信された反射信号を整相し、三次元超音波信号を得る。本ステップは医用画像取得部により医用画像を取得する例が開示される。また、三次元超音波信号を得る工程は、医用画像取得部により医用画像を取得する工程の一例である。
検者は、計測条件設定部61を手動操作して心筋の位置の微調整のための位置情報を制御部8に設定することができる。制御部8は、設定された位置情報を受けて、心筋の輪郭モデルの初期位置を再設定する。輪郭変形した抽出輪郭の精度は、最初に画像内における心筋の大まかな位置を特定することで向上できる。ステップ212の手動操作は必須でなく、制御部8に心筋の位置を公知の画像認識技術で抽出させてもよい。
制御部8は、A2Cの画像、A4Cの画像上で心内外膜輪郭線を二次元輪郭線抽出部43に抽出させる。輪郭線抽出方法は、膜面の画像輝度値の変化を検出するエッジ検出処理、テンプレートマッチング、輪郭モデルを利用した方法を適用できる。ここでは、輪郭モデルを利用した方法を例に説明する。輪郭モデルとは、抽出したい物体の輪郭の形状や輝度値の法則を一般化した形式で表現したものである。輪郭モデルの形状、すなわち、実際の心臓の形状に応じて適応的に輪郭を変形させながら、心臓の輪郭を抽出することができる。また、輪郭モデルでは、過去に抽出した輪郭データを学習させて輪郭モデルを生成する方法も用いることができる。
制御部8は短軸像上で心内外膜輪郭線を二次元輪郭線抽出部43に抽出させる。図3に示すように、輪郭モデルには、分画境界308と短軸像レベルの位置309a~309c(ここでは、Apex、Mid、Baseの3段階を例とする)が輪郭線と一緒に記憶されている。輪郭モデルは画像の左心室の変形に合せて変形する。
制御部8は長軸像と短軸像を表示部7に表示させる。具体的には、長軸像(A2Cの画像、A4Cの画像)は、図4の超音波診断装置の表示画面401上にそれぞれ符号402、符号403として表示される。短軸像(Apex、Mid、Baseの各画像)は、図4の超音波診断装置の表示画面401上にそれぞれ符号404、符号405、符号406として表示される。さらに三次元画像が符号407として表示画面401に表示されても良い。
検者は、計測条件設定部61を手動操作して輪郭位置あるいは心筋分画位置の微調整のための位置情報を制御部8に設定することができる。制御部8は、設定された位置情報を受けて、輪郭位置あるいは心筋分画位置を微調整する。ステップ218の手動操作は必須でなく、輪郭位置あるいは心筋分画位置の微調整が不要であれば実行を省略することができる。
制御部8は、前記輪郭面、輪郭線、分画位置で定義される領域の計測をROI・計測値演算部45に行わせる。ROI・計測値演算部45は、心筋分画毎の動きを計測し、周辺の心筋分画と比較して極端に動きが速い又は遅い挙動を示す心筋分画に基づいてROIを自動設定する。心筋分画毎の動きは次に説明する心筋トラッキング法により計測することができる。
制御部8は長軸像と短軸像に合せて、ROIや計測値を表示部7に表示させる。
ROIは図4の符号409で示される。ROI409は、心筋の領域fを含む点線の円形で示される領域である。ROI409の表示例は点線で示しているが、背景画像が白黒である場合はカラーであってもよいし、点線でなくとも実線や一点鎖線等の線分でもよい。また、ROI409の形状は円形に限られず、矩形、あるいは別途の手法で臓器や器官の輪郭を抽出し、抽出された輪郭に沿ったあるいは近似した形状であってもよい。
図5は本発明の実施例2の超音波画像診断装置の計測処理のフローチャートである。図6は図5の輪郭線の設定の一例を示す図である。
制御部8は、三次元超音波信号から公知の画像認識によって長軸像(A2Cの画像とA4Cの画像)を抽出する。制御部8は、抽出された長軸像を表示部7に表示する。検者は、表示部7に表示された長軸像に対し、計測条件設定部61を用いてA2Cの画像とA4Cの画像上で心内外膜輪郭を手動設定する。さらに、検者は、計測条件設定部61を用いて心筋分画の分画境界(実施例1の図4の符号308)を設定する。
検者は、表示部7に表示された長軸像に対し、計測条件設定部61を用いて短軸像(Apex、Mid、Base)の位置を設定する。制御部8は、設定された位置の短軸像を表示部7に表示する。次に、検者は、表示部7に表示された短軸像に対し、計測条件設定部61を用いて各短軸像上の心内外膜輪郭を抽出する。ステップ511によるA2Cの画像およびA4Cの画像の心内外膜輪郭線は、図6左下のように短軸断面601に対して8点の交点をもって交差する。検者は、表示部7に表示された短軸像に対し、計測条件設定部61を用いて8点を通るように輪郭点を設定する。また、図6右下のように短軸像上においても心筋分画の分画境界308を設定する。
上記ステップは、実施例1で説明したステップ215~ステップ51Bと同じであるため、説明を省略する。
また、ステップ511又はステップ513は、何れか一方はデータベース部53に記憶されているモデルを参照してもよい。
しかし、長軸像又は短軸像同士の位置関係は必ずしも直交である必要はなく、角度が特定関係を持っていれば、直交する、直交しないに拘らず、基準断面の種類を自由に決められる。実施例1と実施例3の相違点は直交するか、直交しないかの位置関係だけであるので、位置関係の違いだけを説明する。
例えば、図7左下のように三次元の輪郭面を、A4Cに対し斜めに交わる長軸断面で切断すると、図7右下のような斜めの座標軸となる。斜めの座標軸と短軸像輪郭線との交点707と、分割境界との相対位置をあらかじめ記憶しておけば、前記相対位置を利用して短軸像の心筋領域を分割することができる。
しかし、長軸像は必ず2つ表示する必要はなく、少なくとも一方が表示されれば、短軸像が設定できる。実施例1と実施例4の相違点は2つの長軸画像を表示するか、1つの長軸画像を表示するかだけである。
図8では、例えば、A2Cの画像片方だけを手動指定する場合には、画面上にA2Cの画像のみを表示しておいて、A4Cの画像は表示しないようにしてもよい。そして、図8では、ROIを求めた結果、A4Cの画像内にROI809が表示されている例を示している。
図9では、例えば、操作部6に「ROI自動設定」というボタンを用意し、検者が「ROI自動設定」のボタンを操作する。そして、図9では、ROIを求めた結果、A4Cの画像内にROI909が表示されている例を示している。
Claims (18)
- 医用画像を取得する医用画像取得部と、
前記医用画像中の運動臓器領域を含む三次元画像を構成する三次元画像構成部と、
前記三次元画像から基準画像となる二次元断面像画像を生成する断面像画像生成部と、
前記基準画像を領域分割の基準に基づいて複数の領域に分割する領域分割部と、
前記複数の領域の運動状態を算出し、算出された運動状態に基づき前記複数の領域のうちの少なくとも一つの領域を特定し、特定された領域が含まれる前記医用画像の領域を関心領域として設定する関心領域設定部と、
を備えたことを特徴とする医用画像診断装置。 - 前記断面像画像生成部は、二次元断面像画像を複数生成し、複数の二次元断面像画像の少なくとも2つの画像は互いに直交する位置関係である請求項1に記載の医用画像診断装置。
- 前記二次元断面像画像を生成するための比較モデルが登録されたデータベース部を有し、前記断面像画像生成部は前記データベースに登録された比較モデルを参照して二次元断面像画像を生成する請求項1に記載の医用画像診断装置。
- 前記複数の領域に分割するための比較モデルが登録されたデータベース部を有し、前記領域分割部は前記データベースに登録された比較モデルを参照して複数の領域に分割する請求項1に記載の医用画像診断装置。
- 前記二次元断面像画像を生成するために前記表示部に表示された画像に対して位置情報を入力する設定部を有し、前記断面像画像生成部は前記設定部に入力された位置情報により二次元断面像画像を生成する請求項1に記載の医用画像診断装置。
- 前記複数の領域に分割するために前記表示部に表示された画像に対して位置情報を入力する設定部を有し、前記領域分割部は前記設定部に入力された位置情報により複数の領域に分割する請求項1に記載の医用画像診断装置。
- 前記断面像画像生成部は、二次元断面像画像を複数生成し、複数の二次元断面像画像の少なくとも2つの画像は互いに特定の角度を有する位置関係である請求項1に記載の医用画像診断装置。
- 前記断面像画像生成部は、二次元断面像画像を少なくとも一つ生成し、前記領域分割部は、一つの二次元断面像画像に対し領域分割の基準により複数の領域に分割する請求項1に記載の医用画像診断装置。
- 前記三次元画像構成部から前記関心領域自動設定部までの一連の構成部を起動する起動部をさらに備え、前記起動部によって前記三次元画像構成部、前記断面像画像生成部、前記領域分割部及び前記関心領域自動設定部を起動する請求項1に記載の医用画像診断装置。
- 前記関心領域設定部は前記複数の領域毎に算出され運動状態を用いて統計値を演算し、前記統計値を閾値として用いて領域を特定する請求項1に記載の医用画像診断装置。
- 前記断面像画像生成部は心内膜と心外膜を有する輪郭モデルを用いて前記二次元断面像画像を生成する請求項1に記載の医用画像診断装置。
- 医用画像取得部により医用画像を取得する第1のステップと、
三次元画像構成部により前記医用画像中の運動臓器領域を含む三次元画像を構成する第2のステップと、
断面像画像生成部により前記三次元画像から基準画像となる二次元断面像画像を生成する第3のステップと、
領域分割部により前記基準画像を領域分割の基準に基づいて複数の領域に分割する第4のステップと、
関心領域設定部により前記複数の領域の運動状態を算出し、算出された運動状態に基づき前記複数の領域のうちの少なくとも一つの領域を特定し、特定された領域が含まれる前記医用画像の領域を関心領域として設定する第5のステップと、
を備えたことを特徴とする医用画像診断装置の関心領域設定方法。 - 前記第3のステップは、前記断面像画像生成部により二次元断面像画像を複数生成し、複数の二次元断面像画像の少なくとも2つの画像は互いに直交する位置関係である請求項12に記載の医用画像診断装置の関心領域設定方法。
- 設定部により前記二次元断面像画像を生成するために前記表示部に表示された画像に対して位置情報を入力する第6のステップを含み、前記第3のステップは、前記断面像画像生成部により前記設定部に入力された位置情報により二次元断面像画像を生成する請求項12に記載の医用画像診断装置の関心領域設定方法。
- 設定部により前記複数の領域に分割するために前記表示部に表示された画像に対して位置情報を入力する第7のステップを含み、前記第3のステップは、前記断面像画像生成部により前記設定部に入力された位置情報により複数の領域に分割する請求項12に記載の医用画像診断装置の関心領域設定方法。
- 前記第3のステップは、前記断面像画像生成部により二次元断面像画像を複数生成し、複数の二次元断面像画像の少なくとも2つの画像は互いに特定の角度を有する位置関係である請求項12に記載の医用画像診断装置の関心領域設定方法。
- 前記第3のステップは、前記断面像画像生成部により二次元断面像画像を少なくとも一つ生成し、前記第4のステップは、前記領域分割部により一つの二次元断面像画像に対し領域分割の基準により複数の領域に分割する請求項12に記載の医用画像診断装置の関心領域設定方法。
- 起動部により前記三次元画像構成部から前記関心領域自動設定部までの一連の構成部を起動する第8のステップをさらに備え、第8のステップによって前記三次元画像構成部、前記断面像画像生成部、前記領域分割部及び前記関心領域自動設定部を前記起動部により起動する請求項12に記載の医用画像診断装置の関心領域設定方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/321,319 US20120065499A1 (en) | 2009-05-20 | 2010-05-18 | Medical image diagnosis device and region-of-interest setting method therefore |
JP2011514415A JP5670324B2 (ja) | 2009-05-20 | 2010-05-18 | 医用画像診断装置 |
CN201080020359.5A CN102421373B (zh) | 2009-05-20 | 2010-05-18 | 医用图像诊断装置及其感兴趣区域设定方法 |
EP10777745.0A EP2433567A4 (en) | 2009-05-20 | 2010-05-18 | DEVICE FOR DIAGNOSING MEDICAL IMAGES AND ASSOCIATED METHOD OF DETERMINING A ZONE INVESTIGATED |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2009121508 | 2009-05-20 | ||
JP2009-121508 | 2009-05-20 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010134512A1 true WO2010134512A1 (ja) | 2010-11-25 |
Family
ID=43126189
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2010/058335 WO2010134512A1 (ja) | 2009-05-20 | 2010-05-18 | 医用画像診断装置とその関心領域設定方法 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20120065499A1 (ja) |
EP (1) | EP2433567A4 (ja) |
JP (1) | JP5670324B2 (ja) |
CN (1) | CN102421373B (ja) |
WO (1) | WO2010134512A1 (ja) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2012235926A (ja) * | 2011-05-12 | 2012-12-06 | Fujifilm Corp | 医療装置システム及び生体情報監視方法 |
JP2013017716A (ja) * | 2011-07-13 | 2013-01-31 | Hitachi Aloka Medical Ltd | 超音波診断装置 |
JP2014000393A (ja) * | 2012-05-22 | 2014-01-09 | Toshiba Corp | 医用画像診断装置及び画像表示装置 |
CN104105441A (zh) * | 2012-02-13 | 2014-10-15 | 株式会社日立制作所 | 区域提取处理系统 |
WO2015147595A1 (ko) * | 2014-03-28 | 2015-10-01 | 재단법인 아산사회복지재단 | 의료 영상에서 아티팩트와 병변을 구분하는 방법 |
JP2016067559A (ja) * | 2014-09-29 | 2016-05-09 | 株式会社東芝 | 医用画像診断装置、画像処理装置、画像処理方法及び画像処理プログラム |
Families Citing this family (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8638989B2 (en) | 2012-01-17 | 2014-01-28 | Leap Motion, Inc. | Systems and methods for capturing motion in three-dimensional space |
US9679215B2 (en) | 2012-01-17 | 2017-06-13 | Leap Motion, Inc. | Systems and methods for machine control |
US9501152B2 (en) | 2013-01-15 | 2016-11-22 | Leap Motion, Inc. | Free-space user interface and control using virtual constructs |
US10691219B2 (en) | 2012-01-17 | 2020-06-23 | Ultrahaptics IP Two Limited | Systems and methods for machine control |
US9070019B2 (en) | 2012-01-17 | 2015-06-30 | Leap Motion, Inc. | Systems and methods for capturing motion in three-dimensional space |
US11493998B2 (en) | 2012-01-17 | 2022-11-08 | Ultrahaptics IP Two Limited | Systems and methods for machine control |
US8693731B2 (en) | 2012-01-17 | 2014-04-08 | Leap Motion, Inc. | Enhanced contrast for object detection and characterization by optical imaging |
WO2013111813A1 (ja) * | 2012-01-27 | 2013-08-01 | 株式会社 東芝 | 医用画像処理装置 |
US9285893B2 (en) | 2012-11-08 | 2016-03-15 | Leap Motion, Inc. | Object detection and tracking with variable-field illumination devices |
US10609285B2 (en) | 2013-01-07 | 2020-03-31 | Ultrahaptics IP Two Limited | Power consumption in motion-capture systems |
US9465461B2 (en) | 2013-01-08 | 2016-10-11 | Leap Motion, Inc. | Object detection and tracking with audio and optical signals |
US9459697B2 (en) | 2013-01-15 | 2016-10-04 | Leap Motion, Inc. | Dynamic, free-space user interactions for machine control |
US9632658B2 (en) | 2013-01-15 | 2017-04-25 | Leap Motion, Inc. | Dynamic user interactions for display control and scaling responsiveness of display objects |
WO2014200589A2 (en) | 2013-03-15 | 2014-12-18 | Leap Motion, Inc. | Determining positional information for an object in space |
JP6228969B2 (ja) * | 2013-03-29 | 2017-11-08 | 株式会社日立製作所 | 医療用診断装置およびその計測方法 |
US10620709B2 (en) | 2013-04-05 | 2020-04-14 | Ultrahaptics IP Two Limited | Customized gesture interpretation |
US9916009B2 (en) | 2013-04-26 | 2018-03-13 | Leap Motion, Inc. | Non-tactile interface systems and methods |
US9747696B2 (en) | 2013-05-17 | 2017-08-29 | Leap Motion, Inc. | Systems and methods for providing normalized parameters of motions of objects in three-dimensional space |
US10281987B1 (en) | 2013-08-09 | 2019-05-07 | Leap Motion, Inc. | Systems and methods of free-space gestural interaction |
KR102122793B1 (ko) * | 2013-08-22 | 2020-06-15 | 삼성전자주식회사 | 전자 디바이스 및 전자 디바이스에서 이미지 처리 방법 |
US9721383B1 (en) | 2013-08-29 | 2017-08-01 | Leap Motion, Inc. | Predictive information for free space gesture control and communication |
US9632572B2 (en) | 2013-10-03 | 2017-04-25 | Leap Motion, Inc. | Enhanced field of view to augment three-dimensional (3D) sensory space for free-space gesture interpretation |
US9996638B1 (en) | 2013-10-31 | 2018-06-12 | Leap Motion, Inc. | Predictive information for free space gesture control and communication |
JP6253053B2 (ja) * | 2013-11-29 | 2017-12-27 | 富士通株式会社 | データ探索装置、データ探索装置の制御方法およびデータ探索装置の制御プログラム |
US9613262B2 (en) | 2014-01-15 | 2017-04-04 | Leap Motion, Inc. | Object detection and tracking for providing a virtual device experience |
DE202014103729U1 (de) | 2014-08-08 | 2014-09-09 | Leap Motion, Inc. | Augmented-Reality mit Bewegungserfassung |
CN112716521B (zh) | 2014-11-18 | 2024-03-01 | C·R·巴德公司 | 具有自动图像呈现的超声成像系统 |
CN106999146B (zh) | 2014-11-18 | 2020-11-10 | C·R·巴德公司 | 具有自动图像呈现的超声成像系统 |
US20170238907A1 (en) * | 2016-02-22 | 2017-08-24 | General Electric Company | Methods and systems for generating an ultrasound image |
WO2017206023A1 (zh) * | 2016-05-30 | 2017-12-07 | 深圳迈瑞生物医疗电子股份有限公司 | 一种心脏容积识别分析系统和方法 |
KR102655299B1 (ko) * | 2016-11-15 | 2024-04-08 | 삼성메디슨 주식회사 | 초음파 영상장치 및 그 제어방법 |
KR102002279B1 (ko) * | 2017-04-06 | 2019-07-23 | 한국한의학연구원 | 3차원 안면 진단 장치 |
JP7024954B2 (ja) * | 2017-04-12 | 2022-02-24 | キヤノンメディカルシステムズ株式会社 | 医用画像処理装置及びx線ct装置 |
US11875012B2 (en) | 2018-05-25 | 2024-01-16 | Ultrahaptics IP Two Limited | Throwable interface for augmented reality and virtual reality environments |
US10685439B2 (en) * | 2018-06-27 | 2020-06-16 | General Electric Company | Imaging system and method providing scalable resolution in multi-dimensional image data |
JP7363675B2 (ja) * | 2020-06-15 | 2023-10-18 | 株式会社島津製作所 | イメージング質量分析装置、及びイメージング質量分析方法 |
CN112656445B (zh) * | 2020-12-18 | 2023-04-07 | 青岛海信医疗设备股份有限公司 | 一种超声设备、超声图像处理方法及存储介质 |
JP2022158712A (ja) * | 2021-04-02 | 2022-10-17 | キヤノンメディカルシステムズ株式会社 | 超音波診断装置、画像処理装置、及び画像処理プログラム |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11155862A (ja) | 1997-11-28 | 1999-06-15 | Toshiba Corp | 超音波診断装置及び超音波画像処理装置 |
JP2003250804A (ja) * | 2002-03-05 | 2003-09-09 | Toshiba Corp | 画像処理装置及び超音波診断装置 |
JP2006314790A (ja) * | 2005-05-13 | 2006-11-24 | Tomtec Imaging Syst Gmbh | 2次元断面画像を再構築するための方法及び装置 |
JP2008073423A (ja) * | 2006-09-25 | 2008-04-03 | Toshiba Corp | 超音波診断装置、診断パラメータ計測装置及び診断パラメータ計測方法 |
JP2008301920A (ja) * | 2007-06-06 | 2008-12-18 | Toshiba Corp | 超音波診断装置、超音波画像処理装置、及び超音波画像処理プログラム |
JP2009011468A (ja) * | 2007-07-03 | 2009-01-22 | Aloka Co Ltd | 超音波診断装置 |
JP2009078122A (ja) * | 2007-09-04 | 2009-04-16 | Toshiba Corp | 超音波診断装置、超音波画像処理装置、及び超音波画像処理プログラム |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7450746B2 (en) * | 2002-06-07 | 2008-11-11 | Verathon Inc. | System and method for cardiac imaging |
CN1729484A (zh) * | 2002-12-04 | 2006-02-01 | 康复米斯公司 | 用于mri中的各向同性成像及使用各向同性或近似各向同性成像的定量图像分析中的多个成像平面的融合 |
WO2004112568A2 (ja) * | 2003-06-25 | 2004-12-29 | Matsushita Electric Ind Co Ltd | 超音波診断装置 |
JP4299189B2 (ja) * | 2004-05-27 | 2009-07-22 | アロカ株式会社 | 超音波診断装置及び画像処理方法 |
JP2006087827A (ja) * | 2004-09-27 | 2006-04-06 | Toshiba Corp | 画像診断装置、画像処理システム、画像処理プログラムおよび画像処理方法 |
US7512284B2 (en) * | 2005-03-29 | 2009-03-31 | General Electric Company | Volumetric image enhancement system and method |
US7551807B2 (en) * | 2005-11-14 | 2009-06-23 | Korea Institute Of Industrial Technology | Apparatus and method for searching for 3-dimensional shapes |
JP2008068032A (ja) * | 2006-09-15 | 2008-03-27 | Toshiba Corp | 画像表示装置 |
JP2009018115A (ja) * | 2007-07-13 | 2009-01-29 | Toshiba Corp | 3次元超音波診断装置 |
JP5394620B2 (ja) * | 2007-07-23 | 2014-01-22 | ジーイー・メディカル・システムズ・グローバル・テクノロジー・カンパニー・エルエルシー | 超音波撮像装置および画像処理装置 |
WO2009044316A1 (en) * | 2007-10-03 | 2009-04-09 | Koninklijke Philips Electronics N.V. | System and method for real-time multi-slice acquisition and display of medical ultrasound images |
-
2010
- 2010-05-18 WO PCT/JP2010/058335 patent/WO2010134512A1/ja active Application Filing
- 2010-05-18 US US13/321,319 patent/US20120065499A1/en not_active Abandoned
- 2010-05-18 JP JP2011514415A patent/JP5670324B2/ja active Active
- 2010-05-18 CN CN201080020359.5A patent/CN102421373B/zh active Active
- 2010-05-18 EP EP10777745.0A patent/EP2433567A4/en not_active Withdrawn
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11155862A (ja) | 1997-11-28 | 1999-06-15 | Toshiba Corp | 超音波診断装置及び超音波画像処理装置 |
JP2003250804A (ja) * | 2002-03-05 | 2003-09-09 | Toshiba Corp | 画像処理装置及び超音波診断装置 |
JP2006314790A (ja) * | 2005-05-13 | 2006-11-24 | Tomtec Imaging Syst Gmbh | 2次元断面画像を再構築するための方法及び装置 |
JP2008073423A (ja) * | 2006-09-25 | 2008-04-03 | Toshiba Corp | 超音波診断装置、診断パラメータ計測装置及び診断パラメータ計測方法 |
JP2008301920A (ja) * | 2007-06-06 | 2008-12-18 | Toshiba Corp | 超音波診断装置、超音波画像処理装置、及び超音波画像処理プログラム |
JP2009011468A (ja) * | 2007-07-03 | 2009-01-22 | Aloka Co Ltd | 超音波診断装置 |
JP2009078122A (ja) * | 2007-09-04 | 2009-04-16 | Toshiba Corp | 超音波診断装置、超音波画像処理装置、及び超音波画像処理プログラム |
Non-Patent Citations (1)
Title |
---|
See also references of EP2433567A4 |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2012235926A (ja) * | 2011-05-12 | 2012-12-06 | Fujifilm Corp | 医療装置システム及び生体情報監視方法 |
US8965474B2 (en) | 2011-05-12 | 2015-02-24 | Fujifilm Corporation | Tissue imaging system and in vivo monitoring method |
JP2013017716A (ja) * | 2011-07-13 | 2013-01-31 | Hitachi Aloka Medical Ltd | 超音波診断装置 |
CN104105441A (zh) * | 2012-02-13 | 2014-10-15 | 株式会社日立制作所 | 区域提取处理系统 |
JP2014000393A (ja) * | 2012-05-22 | 2014-01-09 | Toshiba Corp | 医用画像診断装置及び画像表示装置 |
WO2015147595A1 (ko) * | 2014-03-28 | 2015-10-01 | 재단법인 아산사회복지재단 | 의료 영상에서 아티팩트와 병변을 구분하는 방법 |
JP2016067559A (ja) * | 2014-09-29 | 2016-05-09 | 株式会社東芝 | 医用画像診断装置、画像処理装置、画像処理方法及び画像処理プログラム |
Also Published As
Publication number | Publication date |
---|---|
EP2433567A1 (en) | 2012-03-28 |
EP2433567A4 (en) | 2013-10-16 |
JP5670324B2 (ja) | 2015-02-18 |
CN102421373B (zh) | 2014-07-16 |
JPWO2010134512A1 (ja) | 2012-11-12 |
US20120065499A1 (en) | 2012-03-15 |
CN102421373A (zh) | 2012-04-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5670324B2 (ja) | 医用画像診断装置 | |
JP6640922B2 (ja) | 超音波診断装置及び画像処理装置 | |
US11715202B2 (en) | Analyzing apparatus and analyzing method | |
JP4060615B2 (ja) | 画像処理装置及び超音波診断装置 | |
JP5438002B2 (ja) | 医用画像処理装置及び医用画像処理方法 | |
RU2708792C2 (ru) | Ультразвуковая диагностика работы сердца с использованием сегментации модели сердечной камеры под контролем пользователя | |
KR101625256B1 (ko) | 심장 m-모드 뷰들의 자동 분석 | |
US9881410B2 (en) | Medical diagnostic imaging apparatus, image processing apparatus, and image generating method | |
US8715189B2 (en) | Ultrasonic diagnosis apparatus for setting a 3D ROI using voxel values and opacity | |
WO2014080833A1 (ja) | 超音波診断装置、画像処理装置及び画像処理方法 | |
US11191520B2 (en) | Ultrasonic diagnostic apparatus, image processing apparatus, and image processing method | |
US9888905B2 (en) | Medical diagnosis apparatus, image processing apparatus, and method for image processing | |
US11931201B2 (en) | Device and method for obtaining anatomical measurements from an ultrasound image | |
US11278259B2 (en) | Thrombus detection during scanning | |
JP6863774B2 (ja) | 超音波診断装置、画像処理装置及び画像処理プログラム | |
RU2708317C2 (ru) | Ультразвуковая диагностика работы сердца посредством сегментации камеры с одной степенью свободы | |
US20210093300A1 (en) | Ultrasonic diagnostic apparatus and ultrasonic diagnostic method | |
JP7132996B2 (ja) | 単一自由度の心腔セグメント化による心臓性能の超音波診断 | |
US11413019B2 (en) | Method and apparatus for displaying ultrasound image of target object | |
CN113939234A (zh) | 超声波诊断装置、医用图像处理装置以及医用图像处理方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080020359.5 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10777745 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2011514415 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13321319 Country of ref document: US Ref document number: 2010777745 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |