WO2002034138A1 - Dispositif et procede de traitement d'images - Google Patents
Dispositif et procede de traitement d'images Download PDFInfo
- Publication number
- WO2002034138A1 WO2002034138A1 PCT/JP2001/009341 JP0109341W WO0234138A1 WO 2002034138 A1 WO2002034138 A1 WO 2002034138A1 JP 0109341 W JP0109341 W JP 0109341W WO 0234138 A1 WO0234138 A1 WO 0234138A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image processing
- image
- dimensional images
- images
- dimensional
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims abstract description 208
- 238000003672 processing method Methods 0.000 title claims description 15
- 238000003384 imaging method Methods 0.000 claims abstract description 19
- 230000015572 biosynthetic process Effects 0.000 claims abstract description 8
- 238000003786 synthesis reaction Methods 0.000 claims abstract description 8
- 238000000034 method Methods 0.000 claims description 34
- 230000008859 change Effects 0.000 claims description 30
- 238000006243 chemical reaction Methods 0.000 claims description 20
- 238000003745 diagnosis Methods 0.000 claims description 18
- 230000003287 optical effect Effects 0.000 claims description 18
- 230000008569 process Effects 0.000 claims description 18
- 238000009877 rendering Methods 0.000 claims description 9
- 230000002194 synthesizing effect Effects 0.000 claims description 4
- 239000002131 composite material Substances 0.000 claims description 3
- 239000003086 colorant Substances 0.000 abstract description 3
- 230000006870 function Effects 0.000 description 22
- 238000012937 correction Methods 0.000 description 13
- 230000000052 comparative effect Effects 0.000 description 7
- 238000010586 diagram Methods 0.000 description 5
- 201000010099 disease Diseases 0.000 description 5
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 5
- 238000002595 magnetic resonance imaging Methods 0.000 description 4
- 238000003825 pressing Methods 0.000 description 4
- 230000009467 reduction Effects 0.000 description 4
- 230000004044 response Effects 0.000 description 3
- 238000006073 displacement reaction Methods 0.000 description 2
- 238000003780 insertion Methods 0.000 description 2
- 230000037431 insertion Effects 0.000 description 2
- 238000007689 inspection Methods 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000009206 nuclear medicine Methods 0.000 description 2
- PXFBZOLANLWPMH-UHFFFAOYSA-N 16-Epiaffinine Natural products C1C(C2=CC=CC=C2N2)=C2C(=O)CC2C(=CC)CN(C)C1C2CO PXFBZOLANLWPMH-UHFFFAOYSA-N 0.000 description 1
- 241000287463 Phalacrocorax Species 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 238000004587 chromatography analysis Methods 0.000 description 1
- 238000013170 computed tomography imaging Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000002405 diagnostic procedure Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000035876 healing Effects 0.000 description 1
- 230000001678 irradiating effect Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000001936 parietal effect Effects 0.000 description 1
- 238000002310 reflectometry Methods 0.000 description 1
- 230000008054 signal transmission Effects 0.000 description 1
- 230000007480 spreading Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computed tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computed tomography [CT]
- A61B6/032—Transmission computed tomography [CT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5211—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
- A61B6/5229—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
- A61B6/5235—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from the same or different ionising radiation imaging techniques, e.g. PET and CT
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5211—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
- A61B6/5229—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
- A61B6/5247—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from an ionising-radiation diagnostic technique and a non-ionising radiation diagnostic technique, e.g. X-ray and ultrasound
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/52—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/5215—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data
- A61B8/5238—Devices using data or image processing specially adapted for diagnosis using ultrasonic, sonic or infrasonic waves involving processing of medical diagnostic data for combining image data of patient, e.g. merging several images from different acquisition modes into one image
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/055—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves involving electronic [EMR] or nuclear [NMR] magnetic resonance, e.g. magnetic resonance imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2219/00—Indexing scheme for manipulating 3D models or images for computer graphics
- G06T2219/028—Multiple view windows (top-side-front-sagittal-orthogonal)
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2219/00—Indexing scheme for manipulating 3D models or images for computer graphics
- G06T2219/20—Indexing scheme for editing of 3D models
- G06T2219/2004—Aligning objects, relative positioning of parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2219/00—Indexing scheme for manipulating 3D models or images for computer graphics
- G06T2219/20—Indexing scheme for editing of 3D models
- G06T2219/2016—Rotation, translation, scaling
Definitions
- the present invention performs image processing on medical images obtained by medical image diagnostic apparatuses such as an ultrasonic diagnostic apparatus, an X-ray CT apparatus, a magnetic resonance imaging (MRI) apparatus, and a nuclear medicine diagnostic apparatus in the medical field.
- medical image diagnostic apparatuses such as an ultrasonic diagnostic apparatus, an X-ray CT apparatus, a magnetic resonance imaging (MRI) apparatus, and a nuclear medicine diagnostic apparatus in the medical field.
- the present invention relates to an image processing device for displaying and an image processing method.
- two different images may be simultaneously displayed and compared for observation.
- the images can be easily compared by setting the same observation angle, magnification, and the like.
- FIG. 6 is a diagram for explaining an operation in a case where two different three-dimensional images are simultaneously displayed in a conventional image processing apparatus and a comparative observation is performed.
- Fig. 6 when observing a plurality (in this case, two) of three-dimensional images displayed on the image monitor 18, the observation angle between the images and the enlargement are required to facilitate comparative observation.
- large-scale, display image processing, etc. are set to the same conditions. ( Many operations for setting these conditions to the same are performed by the user's manual operation in the conventional image processing device. It has been.
- comparative reading is a useful diagnostic / diagnostic method in clinical medicine is widely recognized in comparative reading of two-dimensional images such as tomographic images such as X-rays and CT and MRI. From this fact, it is easy to imagine that comparative reading of 3D images is useful.
- the advent of an image processing device capable of comparing and observing three-dimensional images with simple operations has been desired in the past.
- the present invention provides a method for simultaneously displaying a plurality of three-dimensional images for comparative observation
- a user friendly display that can automatically or automatically display the observation angles between images according to conditions, and can also display necessary diagnostic information at the same time. It is intended to provide a simple image processing apparatus and an image processing method.
- the present invention employs the following means to achieve the above object.
- a first viewpoint of the present invention is that an observation angle is unified with an image processing unit that performs image processing that matches observation angles of a plurality of three-dimensional images. And a display device for displaying the plurality of three-dimensional images simultaneously. .
- an image processing unit for performing image processing for unifying magnifications of a plurality of three-dimensional images, and a display device for simultaneously displaying the plurality of three-dimensional images having a unified magnification. It is an image processing device provided.
- a third viewpoint of the present invention is an image comprising: an image processing unit for performing image processing on a plurality of three-dimensional images under the same conditions; and a display device for simultaneously displaying the plurality of three-dimensional images. It is a processing device.
- an image processing unit that performs at least one image processing of unifying observation angles, magnifications, and image processing of a plurality of three-dimensional images
- An image processing apparatus comprising: an image synthesis unit that generates a synthesized image obtained by synthesizing the plurality of three-dimensional images to which the processing is applied; and a display device that displays the synthesized image.
- a fifth viewpoint of the present invention is to generate a plurality of volume data by performing an image reconstruction process using a spatial filter under the same condition on each projection image group corresponding to each of a plurality of three-dimensional images.
- An image processing unit comprising: an image reconstruction unit to generate, an image processing unit to generate a plurality of three-dimensional images from each of the plurality of pieces of data, and a display device that simultaneously displays the plurality of three-dimensional images.
- image processing for unifying at least one of the observation angles, magnifications, and image processing of a plurality of three-dimensional images.
- An image processing unit to be performed a database for storing diagnostic information on past diagnoses, and a diagnostic information management unit for reading diagnostic information related to at least one of the plurality of three-dimensional images from the database.
- a display device for simultaneously displaying the plurality of three-dimensional images after image processing and the read diagnostic information.
- a seventh aspect of the present invention is a display device that displays a three-dimensional image or diagnostic information, a database that stores diagnostic information or a three-dimensional image related to past diagnosis, the displayed three-dimensional image or the diagnostic information, A control device for determining whether there is a difference between the diagnostic information or the three-dimensional image related to the past diagnosis stored in the database, and when the control device determines that there is a difference,
- the display device is an image processing device that reads the information or image indicating that there is a difference or information or an image that has a difference from the database and displays the read information or image.
- An eighth viewpoint of the present invention displays a plurality of three-dimensional images, and displays the desired one image based on a viewing direction change instruction input for a desired one of the plurality of three-dimensional images. Performing image processing of changing the observation angle of the plurality of three-dimensional images and the observation angle of the remaining images of the plurality of three-dimensional images so as to match the conditions, and simultaneously displaying the plurality of three-dimensional images after the change of the observation angle. It is an image processing method to be provided.
- a plurality of three-dimensional images are displayed, and the desired one of the plurality of three-dimensional images is changed based on a magnification change instruction input for a desired one of the plurality of three-dimensional images.
- magnification of Performing image processing for changing the enlargement ratios of the remaining images of the plurality of three-dimensional images so as to match the conditions, and simultaneously displaying the plurality of three-dimensional images after the change of the enlargement ratios.
- an image input device for displaying a plurality of three-dimensional images and inputting a desired one of the plurality of three-dimensional images.
- Image processing of the one image and image processing of the remaining images of the plurality of three-dimensional images are executed under condition matching, and the plurality of three-dimensional images after the image processing are simultaneously displayed.
- an image processing method including the following.
- FIG. 1 is a schematic diagram illustrating a configuration of an image processing apparatus according to an embodiment of the present invention.
- FIG. 2 is a schematic diagram for explaining the configuration of the image processing device according to the embodiment of the present invention.
- FIG. 3 is a flowchart showing an example of a procedure of a condition matching process for conditionally matching an image A and an image B which are simultaneously displayed.
- FIG. 4 shows a flowchart of an image processing apparatus according to an embodiment of the present invention. An example of such a screen display is shown below.
- FIG. 5 is a flowchart showing another example of the procedure of the condition matching process for conditionally matching the image A and the image B displayed simultaneously.
- FIG. 6 shows an example of a screen display by a conventional image processing apparatus.
- FIG. 1 is a schematic diagram of a diagnostic system including an image processing device according to a first embodiment of the present invention.
- the diagnostic system includes an image diagnostic device such as a first image processing device 2, a second image processing device 3, a CT device 4, an MRI device 5, an ultrasonic device 6, an X-ray device 7, or a nuclear medicine device, and an image database. 1 is provided.
- Each device is connected by a network 8 so that data can be exchanged.
- a three-dimensional image obtained by three-dimensionally imaging the internal structure of the patient is reconstructed, and the generated image is stored in the image database 1.
- the image database 1 stores not only three-dimensional images but also patient / photographing information attached to the images, information indicating the orientation of the images, and the like.
- a diagnostic system having two image processing apparatuses as shown in FIG. 1 is taken as an example.
- the number of image processing devices provided in the system is not limited.
- the first image processing apparatus 2 or the second image processing apparatus first Remind as in c Figure 2 is a schematic diagram for explaining the internal structure of the 3 or the second shown in Fig. 1
- Image processing devices include a CPU 14, an input device 13, an image memory 9, a three-dimensional image processing unit 10, a three-dimensional image processing unit 11, and a LUT (Look Up Table) 1. 2, a display unit (monitor) 15, and a signal node 16.
- the function of each component is, for example, as follows (
- the CPU 14 controls the processing of the entire image processing apparatus.
- the input device 13 sends the instruction information from the operator to the device. It is a device for inputting.
- the input device 13 is provided with, for example, buttons, a keyboard, a dash ball, and a joystick for setting various display conditions. In an image condition matching process to be described later, a display angle changing operation or the like is instructed by the input device 13 for at least one of a plurality of images simultaneously displayed.
- the image memory 9 stores three-dimensional image data, projection image data before reconstruction, and the like.
- the three-dimensional affinity processing unit 10 is a conversion unit that performs parallel movement, enlargement, reduction, rotation, and the like (ie, affinity conversion) on the display image.
- the three-dimensional image processing unit 11 performs a three-dimensional image processing such as a volume rendering processing ′ and a surface rendering processing to configure a three-dimensional display image.
- LUT (LookUpTable) 12 converts the gradation of the three-dimensional display image based on a conversion correspondence table between input values and output values in advance.
- the display unit 15 is a monitor that displays a three-dimensional display image.
- the display unit 15 has a plurality of display areas, and can display a plurality of three-dimensional display images at a time.
- the structure may be a multi-window display, or may be composed of a plurality of monitors.
- the signal bus 16 performs signal transmission between components and data transmission / reception.
- the diagnostic database 22 stores the diagnostic images obtained from the image database 1 or various diagnostic devices via the network 8. (2D or 3D images), patient data, examination data, etc. are linked and saved. The information stored in the diagnostic database 22 can be searched for by items such as “site”, “disease name”, and “comment (history leading to diagnosis)”.
- the diagnostic database 22 has a dictionary function that automatically manages data for each “region” and “disease name”.
- the image reconstruction unit 23 performs an image reconstruction process based on the projected image data stored in the memory 9.
- the processes executed by the three-dimensional affect process 10 and the three-dimensional image process 11 may be configured to be executed by the CPU 14.
- condition matching display processing executed by the first image processing device 2 or the second image processing device 3 will be described.
- FIG. 3 is a flowchart showing a procedure of a condition matching process for condition-matching the image A and the image B simultaneously displayed. The following describes an example in which the display angles of the image A and the image B are matched.
- the three-dimensional images A and B registered in the image database 1 are simultaneously displayed on the image processing device (step S1).
- the three-dimensional images A and B may be images taken with the same modality or images taken with different modalities. Usually, at this stage, the displayed images are different from each other unless the displayed images are usually intentionally matched.
- Step S2 the condition matching switch is turned on by the operator.
- the rotation conversion affine conversion
- Step S3 This rotation is not merely for matching the amount of change from the initial display angle, but is based on various information, taking into account the orientation of the image. It is performed so that it is observed from the same angle.
- image orientation information the information for matching the anatomical observation angles of the image B and the image A is referred to as “image orientation information.”.
- the image orientation information of the CT device 4 includes a patient insertion direction, an image observation direction, a patient orientation, and the like.
- the patient entry direction is image orientation information that indicates whether the patient is to be inserted into the CT apparatus 4 from the head or from the feet.
- This patient insertion direction determines the arrangement order of the reconstructed three-dimensional image, and Top First (TF: images are constructed in order from the parietal direction) ZFotFirst (FF: images are constructed in order from the foot and tail direction).
- TF images are constructed in order from the parietal direction
- ZFotFirst ZFotFirst
- the image viewing direction indicates the direction in which the image is viewed, and is expressed as Viewfrom Top (VFT: head-top force, viewed image) ZV iewfrom Foot (VFF: image viewed from the foot and tail direction) Is done.
- VFT head-top force, viewed image
- VFF image viewed from the foot and tail direction
- the patient position is, for example, image orientation information indicating whether the patient was facing up, down, left, or right at the time of imaging.
- image A and image B Is Both are the initial display angles, and ⁇ ) image orientation information of image A is VFT, TF, upward, ⁇ ) image ⁇ image orientation information is VFF, FF, upward Assume a match.
- step S3 by performing an Ain transform for inverting the image B up and down, the observation angles of the image A and the image B can be matched.
- the principle is the same even in the case of the MRI apparatus.
- the MR 1 apparatus can set the tomographic plane freely for the patient, it is preferable to use more detailed image orientation information.
- step S3 The image A and the image B whose observation angles have been matched in step S3 are displayed on the display unit 15 under conditions that match each other (step S4).
- the displacement correction has the following function.
- the misalignment between images can be corrected by the following two functions as necessary.
- One function is to identify at least two anatomically corresponding points, for example, three points, on both images A and B, and to correct them to correct the gap between the images. For example, the identified point on image A (a A, b A. CA) and (a B , b B , c B ) on the image B. A A and a B, b A and b B, and, respectively Re c A and c B pixel in that coincide anatomically to is found.
- Ni straight line a A b A and a B b B when the call Let 's become the same direction skills, and Ri by the and the child that is a A c A and a B c B is calculated to jar'll become the same direction In addition, it is possible to correct a deviation between images.
- the other is a function to correct the deviation between images by manual operation of the operator.
- the rotation process is performed on only one image.
- the specific processing operation is the same as the content of step S3.
- step S5 If it is determined in step S5 that there is a shift correction instruction, the shift is corrected by at least one of the above two types of correction (step S6), and the conditions of the image A and the image B again match. It is displayed (step S7).
- step S6 the shift is corrected by at least one of the above two types of correction
- step S7 the condition match be displayed automatically when the SHIFT key is released.
- step S6 When the deviation is corrected in step S6, the amount of change from the first condition matching state is stored as an error, and the subsequent condition matching display is executed after correcting the error.
- step S5 if it is determined in step S5 that there is no instruction for deviation correction, the image A and the image B are displayed as they are under conditions that match (step S7). ⁇
- FIG. 4 shows an example of a three-dimensional image processed by the image processing device according to the first embodiment and displayed on the display device 15.
- the three-dimensional images A and B of the observation target site are conditionally displayed.
- the other three-dimensional image display will be changed to match the conditions. In this way, multiple 3D images can always be compared under the same conditions.
- condition match switch By setting the condition match switch to OFF at an arbitrary timing, the condition match display process ends, and the condition match state is released (step S8).
- the present image processing apparatus and the image processing apparatuses according to the respective embodiments described later can also perform, for example, the following operations.
- the present invention is not limited to the number of images, and for example, three or more images can be displayed.
- the image processing device, the inspection device, and the image database 1 are all described as separate devices, but the present invention is not limited to the configuration, and the configuration of these devices is not limited. It is feasible even if two or all are configured in one device.
- image condition matching centered on image orientation information was performed. But for example three
- appropriate image orientation information cannot be adopted, such as when the two-dimensional images A and B are captured by different modalities, at least three corresponding points are selected in each image, and based on the corresponding points.
- the condition may be matched.
- a second embodiment of the present invention will be described with reference to FIG.
- the enlargement ratio condition matching processing will be described in more detail.
- the description of the same configuration as that of the first embodiment is omitted.
- three-dimensional images A and B registered in the image database 1 are three-dimensionally displayed on the image processing device (step S1).
- the images A and B may be images taken with the same modality or images taken with different modalities. Normally, each display image has a different display angle and enlargement ratio unless intentionally matched.
- step S2 the condition matching switch is turned ON by the operator.
- the affinity conversion is performed so that the three-dimensional images A and B have the same observation angle and the same magnification (step S3), and the automatic conversion is performed.
- the condition is consistently displayed (step S4). Therefore, the operator can easily observe the condition-matching image without bothering them.
- step S3 for matching the enlargement ratio conditions.
- the conversion for matching the viewing angle condition is the same as in the first embodiment. Omit the explanation.
- Enlarging the condition of the magnification refers to displaying the same length physically on the display device at the same length. For example, if the current image A is a controllable image, the display angle and magnification of image B are rotated to match the display angle of image A. This is simply the initial display angle and initial magnification. Instead of matching the amount of change from the original, the same physical length is displayed so that it can be observed anatomically from the same angle considering the orientation of the image-pixel pitch For example, in the case of a CT device, the pixel pitch on the tomographic plane is determined by various imaging conditions, that is, the imaging area and matrix size.
- the pixel pitch in the body axis direction is determined by the moving width (the moving speed of the bed in the case of the helical scanning method) and the collimator width. Assume that the pixel pitch of the pixel A in the tomographic plane and the body axis direction are both 0.5 mm, and the pixel pitch of the image B in the tomographic plane and the body axis direction is 0.7 mm. In this case, even if the same part is photographed, image A is displayed at a magnification of 1.4 times that of image B in the initial state.
- movement is only a relative change. In other words, if one image is moved 5 mm in the z-axis direction, the other image The image also moves 5 mm in the z-axis direction.
- step S4 The image A and the image B, whose magnifications have been matched in step S3, are displayed on the display unit 15 under condition matching (step S4).
- condition matching switch is ON
- processing to be applied to one image for example, rotation, movement, enlargement / reduction, etc. is also applied to the other image. Is done.
- step S5 it is determined whether or not an instruction to correct a deviation between the three-dimensional image A and the three-dimensional image B has been given (step S5).
- an instruction to correct a deviation between the three-dimensional image A and the three-dimensional image B has been given (step S5).
- the image processing apparatus has two types of functions for correcting such a deviation.
- One is a function that identifies two anatomically identical points on both images A and B, and based on this, detects the gap between the images.
- two of the points specified to identify the angle error, (a A , b A) on image A, (a B , b ⁇ ) on image B, or use an angle error such have as when it is determined in the above only two points images ⁇ of (a a, b a), on the image B (a B, b B) may be specified.
- Straight line a A b A and a B b B is calculate the expansion rate in the jar by the same length of time of this.
- the other is a method in which the function is manually operated.In the case of an operation while pressing a button to which a predetermined function is assigned, for example, an enlargement operation while pressing the SHIFT key on the keyboard, the processing is performed. This is done for only one image.
- step S5 it is determined that there is a shift correction instruction, and in this case, the shift is corrected by at least one of the above two types of correction (step S6), and the image A and the image B are again set. And are matched with the condition (step S7).
- step S6 it is determined that there is a shift correction instruction, and in this case, the shift is corrected by at least one of the above two types of correction (step S6), and the image A and the image B are again set. And are matched with the condition (step S7).
- step S6 the misalignment is corrected by manual operation, it is desirable that the condition is automatically displayed when the SHIFT key is released.
- step S6 When the deviation is corrected in step S6, the amount of change from the first condition matching state is stored as an error, and the subsequent condition matching display (step S7) corrects the error and corrects the error. Be executed.
- step S5 determines whether there is no instruction for deviation correction. If it is determined in step S5 that there is no instruction for deviation correction, the image A and the image B are displayed as they are (step S7).
- condition matching state is released by setting the condition matching switch to OFF (step S8).
- condition matching processing operation of the image processing apparatus according to the second embodiment has been described above.
- the amount of change is stored as an error from the first condition matching state, and thereafter, The condition match can be displayed after correcting the error. Also, without being limited to the number of images, it is possible to simultaneously process three or more images.
- the present embodiment is an example in which image processing is performed under the same condition for each image to be displayed under condition matching.
- the configuration of the image processing apparatus according to the third embodiment is substantially the same as the image processing apparatus described in the first embodiment.
- the description of the parts overlapping with the first and second embodiments described above will be omitted.
- step S1 similarly to the first embodiment, the A image and the B image are simultaneously displayed (step S1), and the condition matching switch is turned on (step S2).
- step S2 In response to the ON operation in step S2, for example, assuming that the current image A is a controllable image, the display angle of image B and the image processing for display are performed based on the display angle and display image of image A. Rotation conversion etc. so as to match the image processing. This is not merely to match the initial display angle and the amount of change from the initial display image processing, but it is anatomically observed from the same angle considering the orientation of the image. It means to change to.
- image processing is performed on at least one of the image A and the image B so that the display image processing of the image A and the image B is completely the same (step S3 ').
- image processing is performed on at least one of the image A and the image B so that the display image processing of the image A and the image B is completely the same (step S3 ').
- a threshold is set and the threshold is set.
- the area that falls within the value range is set as the target area, and the target is displayed by performing a pseudo operation as if the target were illuminated from any direction and calculating the reflected light.
- An image is calculated.
- processing is performed so that the threshold, the position of the light source, the intensity, the target color, and the like at this time match.
- a function for converting the number of pixels into an optical parameter such as a reflectance or a refractive index is defined.
- the display image data is calculated by irradiating the light from above and calculating the reflected light.
- the optical conversion function depending on the definition of the optical conversion function, the internal structure can be visualized, for example, so that the object appears to be blurred in the fog. Therefore, unlike the surface rendering method, the surface of the object can be clearly defined, and information hidden behind the surface can be imaged.
- processing is performed so that the optical conversion function, the position and intensity of the light source, the color of the target, and the like match.
- the A image and the B image are subjected to misalignment correction as needed (steps S5 and S6), and are displayed in a condition-matched manner (step S4, step S4).
- Step S7 The condition match display is canceled by switch-off (step S8), and the condition match display process is completed.
- an image processing apparatus according to a fourth embodiment will be described.
- This embodiment is a modified example in which the same image processing is performed.
- the image processing apparatus according to the fourth embodiment Is substantially the same as that of the image processing apparatus described in the first embodiment.
- description of portions that are the same as in the above-described embodiments will be omitted.
- the condition matching processing of each image processing apparatus according to the fourth embodiment will be described with reference to FIG.
- step S1 the A image and the B image are simultaneously displayed (step S1), and the condition matching switch is turned on (step S2).
- step S2 In response to the ON operation in step S2, for example, if the current image A is a controllable image, the display angle of image B and the image processing for display are changed to the display angle of image A and the display angle of image A. Rotation conversion etc. so as to match the image processing. This is not simply to match the initial display angle and the amount of change from the initial 'display image processing', but to observe the image at the same size from the same anatomical angle in consideration of the orientation of the image. Then, image processing is performed on at least one of the images A and B so that the display image processing of the images A and B is completely the same. (Step S3-1). For example, the image processing of image A and image B are made the same by correcting the optical conversion function of image B by pixel pitch as follows.
- the pixel pitch of image B is m and the pixel pitch of image A is n, and the optical conversion function of image B is corrected by the following function.
- x indicates the passing distance of the target.
- step S3 When image processing is performed by the optical conversion function in step S3, the A image and the B image are subjected to misalignment correction as necessary (step S5), as in the first embodiment. , S 6), and the condition is displayed (step S 4, step S 7).
- step S8 The condition match display is canceled by switch-off (step S8), and the condition match display process is completed.
- the device according to the present embodiment can display the three-dimensional images A and B displayed simultaneously by synthesizing (fusion). This synthesis can be performed at any time.
- the switches for instructing the combination are pressed to combine the three-dimensional images A and B.
- a series of processings such as correcting the designated points in each image so as to coincide with each other, and then performing three-dimensional synthesis are performed.
- the angle deviation, position deviation, and magnification ratio deviation information between images are saved in the first condition matching display, and when the composite switch is pressed, the display angle, position, and magnification ratio match.
- the composition may be such that the composition is performed after such correction.
- volume rendering method it is assumed that pixel values are converted into optical parameters such as reflectivity and refractive index, and light is applied to an object composed of these optical parameters and parameters. Then, the displayed image is calculated.
- the synthesis described in the fifth embodiment is performed by matching alternate pixels every other pixel to one image.
- the optical parameter menu chromatography data of the composite image may be calculated cormorants good follows.
- f indicates an arbitrary function, and is used, for example, in the following two types.
- the images may be processed as separate images and then synthesized on the display image.
- the images are synthesized.
- the distinction between the two can be clarified.
- a range of pixel values is specified to determine a region included in the range, and the color of an overlapping portion of the regions and a color of a non-overlapping portion are changed. Good This makes it possible to clarify the distinction between the two at the region of interest.
- the present image processing apparatus it is also possible to compositely display three or more images. For example, when combining three images A, B, and C, if the optical parameter of each image is ⁇ A'B. ⁇ C, then
- the optical parameter ⁇ required by can be adopted.
- an image processing apparatus capable of providing a three-dimensional image more useful for diagnosis by performing image reconstruction on a plurality of images to be displayed simultaneously under the same condition is described. .
- a three-dimensional image is generated through acquisition processing of a plurality of projection images related to a plurality of cross sections, image reconstruction processing based on the plurality of projection images, image processing of the reconstructed image, and the like.
- image reconstruction processing edge enhancement processing using a spatial filter is also performed.
- This spatial filter is implemented by a convolution operation. Therefore, when simultaneously displaying and observing the three-dimensional images ⁇ and ⁇ separated by the timing of the reconstruction, a difference may appear between the two images based on the difference in the parameters of the spatial filter. is there.
- the image processing apparatus can provide a three-dimensional image that can be easily compared by performing reconstruction processing under the same conditions.
- the image reconstructing unit 23 performs image reconstruction on the projection image corresponding to the image B under the same conditions as the image A. After reconstructing this image, it is possible to display a plurality of three-dimensional images edge-enhanced under the same conditions by performing an affinity conversion again to match the display conditions. it can.
- an image processing apparatus capable of providing effective medical information when performing a diagnosis using a three-dimensional image.
- the first or second image processing apparatus 2 or 3 is a diagnostic image processing apparatus. (2D or 3D images), patient data, inspection data, etc. are linked and stored in the diagnostic database 22.
- the following diagnosis can be performed with a three-dimensional image using the diagnostic database 22 constructed in this manner.
- new diagnostic information such as a three-dimensional image and a test result is obtained in the diagnosis
- past data in the diagnostic database 22 is searched based on the patient name, part, disease name, etc. relating to the diagnostic information. be able to.
- the currently displayed image or diagnostic information is compared with information in the diagnostic database 22 by the CPU 14 as appropriate. If there is a difference in this comparison, the information in the database 22 is read out and displayed on the display unit 15. Specifically, it is used in the following situations.
- diagnosis date”, “diagnosis person”, “case”, “site”, etc. are presented on the display unit 15 in a pop-up window. If you want to observe the details further, you can press the detail button in the window to check the details of the examination of the image, etc., and the results at that time (for example, how the diagnosis was made). Read from diagnostic database 22. Therefore, the operator can easily read out the past diagnostic information related to the three-dimensional image used during the diagnosis and can refer to it.
- the operator when the operator inputs a comment different from the comment added to the past image of the same patient in the database 22 to the currently displayed 3D image.
- the information is automatically displayed on the display unit 15 in a pop-up window II format or the like. If the operator wants to know more about the content of the difference, the operator can display the information in the database 22 by a predetermined operation.
- CAD Computer Aided Diagnost
- the CPU 14 discriminates the difference and suggests it to the operator on the display unit 15. The operator is assisted in the comparison work by this function, and can perform a high-quality diagnosis work more efficiently.
- the image processing apparatus by turning on the condition matching switch, the observation angle, the enlargement ratio, and the image processing for display can be controlled by the imaging conditions unique to the imaging apparatus. Since the display can be performed in a condition-matching manner based on the display parameters, the comparison of the three-dimensional images is facilitated, and the change over time in the affected part and the effect of the treatment are confirmed.
- condition matching switch By setting the condition matching switch to ON, it is possible to match the conditions of the observation angle based on the imaging angle information unique to the imaging device. '' In addition, the observation angle and the magnification can be matched based on the imaging information unique to the imaging device, and the observation angle, magnification, and display image processing can be adjusted based on the imaging conditions and display parameters unique to the imaging device They can be displayed consistently.
- past diagnostic information related to the diagnostic image can be easily provided. As a result, workability can be improved and the quality of diagnosis can be improved.
- the observation angles and the like between the images can be displayed automatically or by a simple operation in accordance with conditions, and the like.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Medical Informatics (AREA)
- Physics & Mathematics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- General Health & Medical Sciences (AREA)
- Surgery (AREA)
- Molecular Biology (AREA)
- Animal Behavior & Ethology (AREA)
- Public Health (AREA)
- Biophysics (AREA)
- Veterinary Medicine (AREA)
- Pathology (AREA)
- Radiology & Medical Imaging (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- High Energy & Nuclear Physics (AREA)
- Optics & Photonics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Theoretical Computer Science (AREA)
- Pulmonology (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Computer Graphics (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Image Processing (AREA)
- Processing Or Creating Images (AREA)
- Image Generation (AREA)
- Image Analysis (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP01978888A EP1336376B1 (en) | 2000-10-24 | 2001-10-24 | Image processing device and image processing method |
KR1020037005631A KR100723737B1 (ko) | 2000-10-24 | 2001-10-24 | 화상처리장치 및 화상처리방법 |
DE60135143T DE60135143D1 (de) | 2000-10-24 | 2001-10-24 | Bildverarbeitungsvorrichtung und bildverarbeitungsverfahren |
US10/421,926 US6990229B2 (en) | 2000-10-24 | 2003-04-24 | Image processing device and image processing method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2000324411A JP5361103B2 (ja) | 2000-10-24 | 2000-10-24 | 画像処理装置 |
JP2000-324411 | 2000-10-24 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/421,926 Continuation US6990229B2 (en) | 2000-10-24 | 2003-04-24 | Image processing device and image processing method |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2002034138A1 true WO2002034138A1 (fr) | 2002-05-02 |
Family
ID=18801948
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2001/009341 WO2002034138A1 (fr) | 2000-10-24 | 2001-10-24 | Dispositif et procede de traitement d'images |
Country Status (7)
Country | Link |
---|---|
US (1) | US6990229B2 (ja) |
EP (1) | EP1336376B1 (ja) |
JP (1) | JP5361103B2 (ja) |
KR (1) | KR100723737B1 (ja) |
CN (2) | CN100546546C (ja) |
DE (1) | DE60135143D1 (ja) |
WO (1) | WO2002034138A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1568322A1 (en) * | 2002-12-03 | 2005-08-31 | Kabushiki Kaisha Toshiba | Computer-aided diagnostic apparatus |
Families Citing this family (66)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7130457B2 (en) * | 2001-07-17 | 2006-10-31 | Accuimage Diagnostics Corp. | Systems and graphical user interface for analyzing body images |
US7433507B2 (en) * | 2003-07-03 | 2008-10-07 | Ge Medical Systems Global Technology Co. | Imaging chain for digital tomosynthesis on a flat panel detector |
JP2005114713A (ja) * | 2003-09-19 | 2005-04-28 | Keyence Corp | 拡大観察装置、拡大画像観察方法、拡大観察用操作プログラムおよびコンピュータで読み取り可能な記録媒体 |
JP4675633B2 (ja) * | 2004-03-09 | 2011-04-27 | 株式会社東芝 | 放射線レポートシステム |
US7664299B2 (en) * | 2004-04-02 | 2010-02-16 | Kabushiki Kaisha Toshiba | Apparatus that prepares information relating to image data |
US8160314B2 (en) * | 2004-06-18 | 2012-04-17 | Siemens Aktiengesellschaft | System and method for linking VOIs across timepoints for analysis of disease progression or response to therapy |
CN100367706C (zh) * | 2004-10-28 | 2008-02-06 | 上海交通大学 | 基于网络服务资源框架的图像网格处理系统 |
EP1817744A2 (en) * | 2004-11-22 | 2007-08-15 | Koninklijke Philips Electronics N.V. | Improved data representation for rtp |
KR100702148B1 (ko) * | 2004-12-30 | 2007-03-30 | 한국전기연구원 | 단층영상과 입체 표면영상을 동시에 얻을 수 있는 엑스선단층 촬영장치 |
JP4703193B2 (ja) * | 2005-01-14 | 2011-06-15 | 株式会社東芝 | 画像処理装置 |
JP4649236B2 (ja) * | 2005-03-04 | 2011-03-09 | 株式会社東芝 | 3次元画像処理装置、x線診断装置および3次元画像処理プログラム |
JP4744926B2 (ja) * | 2005-05-16 | 2011-08-10 | 株式会社東芝 | 医用画像表示装置及び医用画像表示方法 |
KR100828358B1 (ko) * | 2005-06-14 | 2008-05-08 | 삼성전자주식회사 | 영상 디스플레이 모드 전환 방법, 장치, 및 그 방법을 실행하기 위한 프로그램을 기록한 컴퓨터로 읽을 수 있는 기록매체 |
CN100411009C (zh) * | 2005-08-26 | 2008-08-13 | 致伸科技股份有限公司 | 图像显示系统及方法 |
US7518619B2 (en) * | 2005-11-07 | 2009-04-14 | General Electric Company | Method and apparatus for integrating three-dimensional and two-dimensional monitors with medical diagnostic imaging workstations |
US20070127791A1 (en) * | 2005-11-15 | 2007-06-07 | Sectra Ab | Automated synchronization of 3-D medical images, related methods and computer products |
CN101057786B (zh) * | 2006-04-19 | 2010-11-17 | 陈兆秋 | Ct、mr图像融合体外控制点用的模块 |
US8179396B2 (en) * | 2006-08-02 | 2012-05-15 | General Electric Company | System and methods for rule-based volume rendition and navigation |
DE102006039389A1 (de) * | 2006-08-22 | 2008-02-28 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zur Reduzierung von Übergangsartefakten in einem Gesamtbild, das sich aus Teilbildern zusammensetzt |
US8090166B2 (en) * | 2006-09-21 | 2012-01-03 | Surgix Ltd. | Medical image analysis |
JP2008104798A (ja) * | 2006-10-27 | 2008-05-08 | Ziosoft Inc | 画像処理方法 |
US8160395B2 (en) * | 2006-11-22 | 2012-04-17 | General Electric Company | Method and apparatus for synchronizing corresponding landmarks among a plurality of images |
IL179582A0 (en) | 2006-11-26 | 2007-05-15 | Algotec Systems Ltd | Comparison workflow automation by registration |
IL184151A0 (en) | 2007-06-21 | 2007-10-31 | Diagnostica Imaging Software Ltd | X-ray measurement method |
WO2009065079A2 (en) * | 2007-11-14 | 2009-05-22 | The Regents Of The University Of California | Longitudinal registration of anatomy in magnetic resonance imaging |
CN101178816B (zh) * | 2007-12-07 | 2010-06-16 | 桂林电子科技大学 | 基于面采样的体绘制可视化方法 |
WO2009104583A1 (ja) * | 2008-02-21 | 2009-08-27 | コニカミノルタエムジー株式会社 | 画像表示システム及び画像処理プログラム |
CN102124320A (zh) * | 2008-06-18 | 2011-07-13 | 苏尔吉克斯有限公司 | 用于将多个图像拼接成全景图像的方法和系统 |
US20100055657A1 (en) * | 2008-08-27 | 2010-03-04 | Warren Goble | Radiographic and ultrasound simulators |
JP5242472B2 (ja) * | 2009-03-23 | 2013-07-24 | 株式会社ニデック | 眼科観察装置 |
EP2233065B8 (en) | 2009-03-23 | 2015-11-25 | Nidek Co., Ltd. | Ophthalmic observation apparatus |
JP5601609B2 (ja) * | 2009-03-23 | 2014-10-08 | 株式会社ニデック | 眼科観察プログラム及び眼科観察装置 |
KR100970899B1 (ko) * | 2009-05-14 | 2010-07-16 | (주)간연사 | 다목적 간판프레임 |
US9792012B2 (en) | 2009-10-01 | 2017-10-17 | Mobile Imaging In Sweden Ab | Method relating to digital images |
US8934686B2 (en) * | 2009-11-26 | 2015-01-13 | Algotec Systems Ltd. | User interface for selecting paths in an image |
JP5689662B2 (ja) * | 2009-12-09 | 2015-03-25 | 株式会社東芝 | 超音波診断装置、超音波画像処理装置、超音波画像処理プログラム、医用画像診断装置、医用画像処理装置及び医用画像処理プログラム |
JP2011224086A (ja) * | 2010-04-16 | 2011-11-10 | Morita Mfg Co Ltd | 画像処理装置、x線撮影装置、画像表示方法、画像比較方法、および画像表示プログラム |
JP4920771B2 (ja) * | 2010-06-15 | 2012-04-18 | 株式会社東芝 | 医用画像表示装置 |
EP2603834B1 (en) * | 2010-09-20 | 2020-12-09 | Nokia Technologies Oy | Method for forming images |
KR101783000B1 (ko) | 2011-07-19 | 2017-09-28 | 삼성전자주식회사 | 복수의 3차원 볼륨 영상들을 이용하여 3차원 볼륨 파노라마 영상 생성 방법 및 장치 |
KR101842043B1 (ko) | 2011-07-22 | 2018-03-26 | 삼성전자주식회사 | 초음파 영상 분석 장치 및 방법 |
JP4997341B2 (ja) * | 2011-11-21 | 2012-08-08 | 株式会社東芝 | 医用画像表示装置 |
CN102508671B (zh) * | 2011-11-22 | 2015-01-21 | 云南电力试验研究院(集团)有限公司电力研究院 | 一种x射线数字图片结合图谱数据的集成软件系统 |
CN102508921B (zh) * | 2011-11-22 | 2013-08-14 | 云南电力试验研究院(集团)有限公司电力研究院 | X射线数字图片结合局部放电图谱的图像数据库系统 |
JP5863435B2 (ja) * | 2011-12-15 | 2016-02-16 | Hoya株式会社 | 画像信号処理装置 |
EP2814398B1 (en) * | 2012-02-13 | 2017-06-28 | Koninklijke Philips N.V. | Simultaneous ultrasonic viewing of 3d volume from multiple directions |
JP2013169359A (ja) * | 2012-02-21 | 2013-09-02 | Toshiba Corp | X線ct装置 |
CN103764038A (zh) | 2012-02-21 | 2014-04-30 | 株式会社东芝 | X射线ct装置、图像显示装置、图像显示方法 |
JP6025456B2 (ja) * | 2012-08-28 | 2016-11-16 | キヤノン株式会社 | 被検体情報取得装置、表示方法、及びプログラム |
US10016181B2 (en) * | 2012-09-26 | 2018-07-10 | Hitachi, Ltd. | Ultrasound diagnostic apparatus and ultrasound three-dimensional image creation method |
JP5632446B2 (ja) * | 2012-12-03 | 2014-11-26 | 株式会社東芝 | X線診断装置 |
US9091628B2 (en) | 2012-12-21 | 2015-07-28 | L-3 Communications Security And Detection Systems, Inc. | 3D mapping with two orthogonal imaging views |
US9386936B2 (en) * | 2013-03-13 | 2016-07-12 | Ellumen, Inc. | Distributed microwave image processing system and method |
JP2013198822A (ja) * | 2013-07-11 | 2013-10-03 | Toshiba Corp | 画像処理装置 |
KR102273831B1 (ko) * | 2014-01-07 | 2021-07-07 | 삼성메디슨 주식회사 | 의료 영상을 디스플레이 하는 방법 및 그 의료 영상 장치 |
US10216762B2 (en) | 2014-06-04 | 2019-02-26 | Panasonic Corporation | Control method and non-transitory computer-readable recording medium for comparing medical images |
JP6514724B2 (ja) | 2014-07-02 | 2019-05-15 | コヴィディエン リミテッド パートナーシップ | 患者の肺の複数のctスキャンのための統合された座標システム |
JP6476041B2 (ja) * | 2015-03-31 | 2019-02-27 | 株式会社Aze | 医用画像診断装置、その制御方法、及びプログラム |
EP3178380A1 (en) * | 2015-12-09 | 2017-06-14 | Canon Kabushiki Kaisha | Photoacoustic apparatus, display control method, and program |
JP6797557B2 (ja) * | 2016-05-17 | 2020-12-09 | キヤノンメディカルシステムズ株式会社 | 医用画像診断装置、医用画像処理装置および画像表示プログラム |
JP7022584B2 (ja) * | 2017-12-27 | 2022-02-18 | キヤノン株式会社 | 放射線撮影装置、画像処理装置及び画像判定方法 |
JP6896145B2 (ja) * | 2018-03-09 | 2021-06-30 | 富士フイルム株式会社 | 画像処理装置、画像処理方法、および画像処理プログラム |
JP7128030B2 (ja) * | 2018-05-21 | 2022-08-30 | エスペック株式会社 | 環境形成装置 |
CN109700528B (zh) * | 2019-02-27 | 2021-02-02 | 江苏霆升科技有限公司 | 一种实时构建心脏三维模型方法及装置 |
JP7408361B2 (ja) * | 2019-11-29 | 2024-01-05 | 富士フイルムヘルスケア株式会社 | 医用画像診断支援システム及び医用画像処理装置、医用画像処理方法 |
JP7388244B2 (ja) * | 2020-03-04 | 2023-11-29 | 株式会社ニデック | 眼科観察装置、および眼科観察プログラム |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08212325A (ja) * | 1995-02-02 | 1996-08-20 | Toshiba Corp | 3次元医用画像処理装置 |
Family Cites Families (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4263916A (en) * | 1978-03-27 | 1981-04-28 | University Of Southern California | Image averaging for angiography by registration and combination of serial images |
JPH0197441A (ja) * | 1987-10-09 | 1989-04-14 | Hitachi Medical Corp | 超音波診断装置 |
JP2620946B2 (ja) * | 1987-11-13 | 1997-06-18 | 株式会社日立メディコ | 超音波診断装置 |
FR2662813B1 (fr) * | 1990-05-29 | 1992-08-14 | Traitement Synthese Image | Procede d'acquisition d'images d'echographie. |
DE69332042T2 (de) * | 1992-12-18 | 2003-01-02 | Koninklijke Philips Electronics N.V., Eindhoven | Ortungszurückstellung von relativ elastisch verformten räumlichen Bildern durch übereinstimmende Flächen |
US5456255A (en) * | 1993-07-12 | 1995-10-10 | Kabushiki Kaisha Toshiba | Ultrasonic diagnosis apparatus |
JPH08131403A (ja) * | 1994-11-09 | 1996-05-28 | Toshiba Medical Eng Co Ltd | 医用画像処理装置 |
US5579360A (en) * | 1994-12-30 | 1996-11-26 | Philips Electronics North America Corporation | Mass detection by computer using digital mammograms of the same breast taken from different viewing directions |
JP3639030B2 (ja) * | 1995-02-28 | 2005-04-13 | 株式会社東芝 | 画像表示システム及びそのシステムを用いた画像表示方法 |
JPH08280684A (ja) * | 1995-04-18 | 1996-10-29 | Fujitsu Ltd | 超音波診断装置 |
US5911691A (en) * | 1996-05-21 | 1999-06-15 | Aloka Co., Ltd. | Ultrasound image processing apparatus and method of forming and displaying ultrasound images by the apparatus |
JPH105213A (ja) * | 1996-06-24 | 1998-01-13 | Hitachi Medical Corp | X線造影検査装置 |
JP3878259B2 (ja) * | 1996-11-13 | 2007-02-07 | 東芝医用システムエンジニアリング株式会社 | 医用画像処理装置 |
JPH10286251A (ja) * | 1997-04-14 | 1998-10-27 | Eiichi Masai | 透視下操作装置 |
JP3117665B2 (ja) * | 1997-08-26 | 2000-12-18 | ジーイー横河メディカルシステム株式会社 | 画像表示方法および画像表示装置 |
US6106464A (en) * | 1999-02-22 | 2000-08-22 | Vanderbilt University | Apparatus and method for bone surface-based registration of physical space with tomographic images and for guiding an instrument relative to anatomical sites in the image |
US6266453B1 (en) * | 1999-07-26 | 2001-07-24 | Computerized Medical Systems, Inc. | Automated image fusion/alignment system and method |
US6826297B2 (en) * | 2001-05-18 | 2004-11-30 | Terarecon, Inc. | Displaying three-dimensional medical images |
-
2000
- 2000-10-24 JP JP2000324411A patent/JP5361103B2/ja not_active Expired - Lifetime
-
2001
- 2001-10-24 CN CNB2007100072264A patent/CN100546546C/zh not_active Expired - Fee Related
- 2001-10-24 KR KR1020037005631A patent/KR100723737B1/ko active IP Right Grant
- 2001-10-24 WO PCT/JP2001/009341 patent/WO2002034138A1/ja active IP Right Grant
- 2001-10-24 EP EP01978888A patent/EP1336376B1/en not_active Expired - Lifetime
- 2001-10-24 CN CNB018179223A patent/CN100346748C/zh not_active Expired - Fee Related
- 2001-10-24 DE DE60135143T patent/DE60135143D1/de not_active Expired - Lifetime
-
2003
- 2003-04-24 US US10/421,926 patent/US6990229B2/en not_active Expired - Lifetime
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08212325A (ja) * | 1995-02-02 | 1996-08-20 | Toshiba Corp | 3次元医用画像処理装置 |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1568322A1 (en) * | 2002-12-03 | 2005-08-31 | Kabushiki Kaisha Toshiba | Computer-aided diagnostic apparatus |
EP1568322A4 (en) * | 2002-12-03 | 2009-04-29 | Toshiba Kk | COMPUTERGEST TZTES DIAGNOSED |
US9168007B2 (en) | 2002-12-03 | 2015-10-27 | Kabushiki Kaisha Toshiba | Computer-aided diagnostic apparatus |
Also Published As
Publication number | Publication date |
---|---|
KR100723737B1 (ko) | 2007-05-30 |
JP5361103B2 (ja) | 2013-12-04 |
CN1471375A (zh) | 2004-01-28 |
DE60135143D1 (de) | 2008-09-11 |
KR20030045839A (ko) | 2003-06-11 |
EP1336376A1 (en) | 2003-08-20 |
CN100546546C (zh) | 2009-10-07 |
CN100346748C (zh) | 2007-11-07 |
CN100998509A (zh) | 2007-07-18 |
EP1336376A4 (en) | 2006-02-22 |
JP2002125937A (ja) | 2002-05-08 |
US20030185426A1 (en) | 2003-10-02 |
EP1336376B1 (en) | 2008-07-30 |
US6990229B2 (en) | 2006-01-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2002034138A1 (fr) | Dispositif et procede de traitement d'images | |
JP4937261B2 (ja) | 2dのx線画像及び3d超音波画像を選択的に混合するためのシステム及び方法 | |
JP4901531B2 (ja) | X線診断装置 | |
JP5551957B2 (ja) | 投影画像生成装置およびその作動方法、並びに投影画像生成プログラム | |
CN106485691B (zh) | 信息处理装置、信息处理系统和信息处理方法 | |
CN107847274B (zh) | 用于在机器人外科手术期间提供更新后的患者图像的方法和装置 | |
JP5775244B2 (ja) | 医学的イメージング・ボリュームの3dグラフィカル・プリスクリプションのためのシステム及び方法 | |
JP2011125568A (ja) | 画像処理装置、画像処理方法、プログラム及び画像処理システム | |
JP2006320721A (ja) | 対象領域のボリューム撮影の表示範囲の拡張方法 | |
KR20130018168A (ko) | 의료 이미지 데이터 세트의 정합 품질의 시각화 방법 및 장치 | |
WO2006107801A2 (en) | System and method for reducing artifacts in motion corrected dynamic image sequences | |
JP2017205217A (ja) | 医用画像診断装置、医用画像処理装置および画像表示プログラム | |
JP5575620B2 (ja) | 画像処理装置 | |
JP3989896B2 (ja) | 医用画像処理装置、関心領域抽出方法、ならびに、プログラム | |
JP2019126654A (ja) | 画像処理装置、画像処理方法、及びプログラム | |
JP2017099531A (ja) | 画像処理装置、画像処理方法およびx線診断装置 | |
JP4794993B2 (ja) | 画像診断支援装置及び画像表示方法 | |
JP7127212B2 (ja) | 画像の向き設定装置、方法およびプログラム | |
JP5883378B2 (ja) | 画像処理装置 | |
JP2007125102A (ja) | 医用画像表示装置及び医用画像診断装置 | |
JP2002301063A (ja) | 医療用画像の作成方法およびそれに用いられる医療用画像表示装置 | |
JP2012000477A (ja) | 画像処理装置 | |
JP2007275295A (ja) | 乳房画像表示方法および乳房画像表示装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 1020037005631 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 10421926 Country of ref document: US Ref document number: 018179223 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2001978888 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1020037005631 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 2001978888 Country of ref document: EP |
|
WWR | Wipo information: refused in national office |
Ref document number: 1020037005631 Country of ref document: KR |
|
WWG | Wipo information: grant in national office |
Ref document number: 2001978888 Country of ref document: EP |