WO2001097174A1 - Dispositif et procede d'entree de points pour images tridimensionnelles - Google Patents
Dispositif et procede d'entree de points pour images tridimensionnelles Download PDFInfo
- Publication number
- WO2001097174A1 WO2001097174A1 PCT/JP2001/005167 JP0105167W WO0197174A1 WO 2001097174 A1 WO2001097174 A1 WO 2001097174A1 JP 0105167 W JP0105167 W JP 0105167W WO 0197174 A1 WO0197174 A1 WO 0197174A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- point
- display surface
- dimensional image
- orientation
- display
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/08—Volume rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/21—Collision detection, intersection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2219/00—Indexing scheme for manipulating 3D models or images for computer graphics
- G06T2219/008—Cut plane or projection plane definition
Definitions
- the present invention relates to an apparatus and a method for designating and inputting an arbitrary point on a three-dimensional image displayed on a display surface.
- the present invention relates to a device and method for inputting an arbitrary designated point to a three-dimensional image used in the medical imaging field.
- the methods for constructing medical three-dimensional images can be broadly classified into two types: surface rendering and volume rendering.
- the surface rendering method is a method of constructing the outline and surface shape of the object to be displayed (in terms of algorithm, surface patches / tiles represented by triangular patches).
- the volume rendering method is a method based on a ray casting algorithm, and is a method of assigning opacity and color information to each of the poxels constituting a three-dimensional object so as to make the contents of the object visible.
- the present invention has been made under such a background, and a main object of the present invention is to provide a point input device and a method capable of designating an arbitrary point of a three-dimensional image by a simple operation.
- the present invention also provides a point input device that can specify an arbitrary point on a three-dimensional image regardless of a three-dimensional construction method such as surface rendering or volume rendering, and has a simple processing operation and a short processing time. And methods.
- the invention according to claim 1 is a display unit for displaying a three-dimensional image on a display surface, a display control unit for setting a direction of the three-dimensional image displayed on the display surface to a desired direction, and a display unit for displaying a three-dimensional image on the display surface.
- a method of creating a vertical straight line passing through the point and on the display surface In response to the specification of an arbitrary point of the three-dimensional image displayed in the second orientation different from the first orientation, a method of creating a vertical straight line passing through the point and on the display surface.
- a point input device for a three-dimensional image comprising: a step; and means for obtaining coordinates of an intersection between the created surface and the straight line, and holding the intersection as a designated point.
- the designated point is displayed on a three-dimensional image, or the data value of the designated point is displayed.
- a designated point related processing means such as to extract the peripheral portion around the designated point, an input device points for the three-dimensional image of claim 1, wherein.
- the invention according to claim 3 is a method for designating an arbitrary point on a three-dimensional image displayed on a display surface, wherein the direction of the three-dimensional image displayed on the display surface is determined by a user.
- the display is controlled so as to be in the first orientation according to the operation, and in response to an arbitrary point of the three-dimensional image displayed in the first orientation on the display surface being designated by the user, Passing the point, creating a plane in the vertical direction on the display surface; and changing the orientation of the three-dimensional image displayed on the display surface to a second orientation different from the first orientation in accordance with a user operation.
- a step of creating a straight line in the vertical direction on the display surface In response to the fact that an arbitrary point of the three-dimensional image displayed in the second orientation is designated by the user on the display surface, A step of creating a straight line in the vertical direction on the display surface, and a step of obtaining intersection coordinates of the created surface and the straight line and holding the coordinates as a designated point.
- This is an arbitrary point input method.
- the invention according to claim 4 is a display unit for displaying a three-dimensional image on a display surface, a display control unit for setting a direction of the three-dimensional image displayed on the display surface to a desired direction, and a first display unit on the display surface.
- the invention according to claim 5 is a method for specifying an arbitrary point in a three-dimensional image displayed on a display surface
- the display of the three-dimensional image displayed on the display surface is controlled in the first direction by controlling the display so that the orientation of the three-dimensional image displayed on the display surface becomes the first direction according to the operation of the user.
- the user specifying any point in, create two straight lines through the point and perpendicular to the display surface or a single line perpendicular to the display surface,
- the orientation of the three-dimensional image displayed on the display surface is controlled to be a second orientation different from the first orientation according to a user operation, and the second orientation is displayed on the display surface.
- a plane is created through the point and perpendicular to the display surface,
- This is a method for inputting an arbitrary point in a three-dimensional image, characterized in that the point of the obtained intersection coordinates is held as a designated point.
- a human input method for an arbitrary point on a three-dimensional image wherein the processing of (1) and the processing of (2) are reversed. is there.
- the invention according to claim 7 is a program for designating an arbitrary point on a three-dimensional image displayed on a display surface
- the display of the three-dimensional image displayed on the display surface is controlled in the first direction by controlling the display so that the orientation of the three-dimensional image displayed on the display surface becomes the first direction according to the operation of the user.
- the orientation of the three-dimensional image displayed on the display surface is controlled so as to be a second orientation different from the first orientation in accordance with a user operation, and is displayed in the second orientation on the display surface.
- the invention according to claim 8 is a method for designating an arbitrary point on a three-dimensional image displayed on a display surface, wherein a direction of the three-dimensional image displayed on the display surface is changed by a user operation.
- the display is controlled to be in the first orientation according to the image, and in response to an arbitrary point of the three-dimensional image displayed in the first orientation on the display surface being designated by the user.
- a step of creating a cross-sectional image passing through the point in the vertical direction on the display surface, a step of displaying the created cross-sectional image on the display surface, and an arbitrary point on the cross-sectional image displayed on the display surface Step that stores the point coordinates in response to the designation by And a step of converting the stored point coordinates on the cross-sectional image into three-dimensional coordinates and holding the three-dimensional coordinates, and an arbitrary point in the three-dimensional image.
- the user can specify an arbitrary point appearing in the three-dimensional image while checking the three-dimensional image displayed on the display surface.
- the designation of an arbitrary point in the three-dimensional image can be performed while checking the three-dimensional image displayed in the first orientation and the three-dimensional image displayed in the second orientation. Can be easily specified.
- the three-dimensional image passes through the first designated point and is perpendicular to the display surface, and the second designated point is perpendicular to the display surface. Since the specified point in the 3D image is specified by the coordinates of the intersection with the straight line in the direction, the specified point can be specified correctly and only one point can be specified.
- the designated point in addition to the method for determining the designated point based on an intersection between a plane and a straight line, the designated point is designated based on a three-dimensional image displayed in the first orientation as described in claim 4. Based on the three-dimensional image displayed in the second orientation, passing through a point and perpendicular to the display surface, and based on the three-dimensional image displayed in the second orientation, passing through the specified point and passing through a specified plane and perpendicular to the display surface. Only one point can be specified by the coordinates of the intersection of the planes.
- the 3D image is displayed in the second orientation. You can create a straight line or one plane at that time, then display the 3D image in the first orientation, and then create one or two planes. The order of creation is not fixed.
- any point in the three-dimensional image can be correctly specified by a simple operation.
- FIG. 1 is a block diagram showing a hardware configuration of a point input device for a three-dimensional image according to an embodiment of the present invention.
- FIG. 2 is a diagram illustrating an example of a three-dimensional image displayed on the display surface of the display.
- FIG. 3 is a diagram showing the created surface (7j plane).
- FIG. 4 is a diagram showing a display example of a state where the three-dimensional rooster image of FIG. 2 is viewed from another viewpoint.
- FIG. 5 is a diagram showing a display example of a designated point displayed on a three-dimensional image.
- FIG. 6 is a diagram showing another example of the three-dimensional image displayed on the display surface of the display.
- FIG. 7 is a diagram showing the created surface (vertical section).
- FIG. 8 is a diagram showing a display example in a state where the three-dimensional image of FIG. 6 is viewed from another viewpoint.
- FIG. 9 is an explanatory diagram showing the processing shown in FIGS. 6 to 8 in a plan view.
- FIG. 10 is a flowchart showing the processing operation of the control unit.
- FIG. 11 is a flowchart showing the processing operation of the control unit.
- FIG. 12 is an explanatory diagram of a case where a point designation process according to the present invention is performed on an image of a display method that performs three-dimensional projection display from an arbitrary direction.
- FIG. 1 is a block diagram showing a hardware configuration of a point input device for a three-dimensional image according to an embodiment of the present invention.
- This device has a control unit 1 including a microcomputer and the like, and the control unit 1 includes a memory 2 for storing an operation program and various data.
- a display 3 for displaying a three-dimensional image is connected to the control unit 1.
- the display 3 can be constituted by a display using a cathode ray tube, a liquid crystal display, or the like.
- An operation unit 4 is also connected to the control unit 1.
- the operation unit 4 includes a keyboard, a mouse, and other input devices for signal input and output.
- the control unit 1 can read information recorded on a CD-ROM, an optical disc, or another information recording medium 5, and record the read information in the memory 2 as needed. Also, a modem 6 for communication is connected to the control unit 1 so that necessary information and data can be fetched from an information network or the like via this modem 6 or predetermined data can be transmitted to the information network. It is possible. '' The features of the point input device according to this embodiment are displayed on the display surface of the display 3. The user can specify a desired position with respect to the three-dimensional image, and is stored and displayed in association with the specified point force three-dimensional image.
- Such processing is executed according to a program described below, which is performed by the control unit 1.
- the present invention is also characterized by a method of designating an arbitrary point on a three-dimensional image executed by the program.
- a program for realizing the present invention may be stored in the memory 2 of the control unit 1 in advance, or may be recorded on the information recording medium 5.
- the program recorded on the information recording medium 5 is stored in the memory 2 of the control unit 1 by being installed in the control unit 1.
- the program according to the present invention may be downloaded to the control unit 1 from an external information network via the modem 6.
- the operation executed by the control unit 1, that is, the features of the program according to this embodiment will be specifically described.
- the control unit 1 can install the program recorded on the information recording medium 5 and the program received via the modem 6 into the memory 2 as described above, as well as tomographic images captured by MRI, CT, or the like.
- Image data can be captured and stored in memory 2.
- data recorded on the information recording medium 5 may be imported, or tomographic image data sent via the modem 6 may be imported.
- the present invention is not limited to the case where the tomographic image data is provided to the control unit 1 and a medical three-dimensional image is constructed in the control unit 1, and the medical three-dimensional image data itself may be provided to the control unit 1.
- FIG. 2 is a view showing an example of the three-dimensional image 8 displayed on the display surface 7 of the display 3.
- the three-dimensional image 8 is an image constructed three-dimensionally based on tomographic images such as MRI and CT given to the control unit 1. Alternatively, it is a display of the three-dimensional image data itself given to the control unit 1.
- the user rotates the three-dimensional image 8 displayed on the display surface 7 so as to be in a desired direction, enlarges a desired part, or obtains a tomographic image of a desired part.
- the direction of the three-dimensional image 8 displayed on the display surface 7 can be set to a desired direction, and an arbitrary point appearing in the three-dimensional image 8 can be designated. For this point specification, for example, move the cursor to the specified part of the three-dimensional image 8 with the mouse. This can be done by clicking the mouse.
- the control unit 1 When an arbitrary point on the three-dimensional image 8 is designated, the control unit 1 responds to the designated point, creates and stores a vertical plane, for example, a horizontal plane on the display surface 7 through the designated point.
- Figure 3 shows the horizontal plane created and stored.
- the horizontal plane shown in FIG. 3 need only be created in the control unit 1 and stored in the memory, and need not be displayed on the display 3.
- the user operates the three-dimensional image 8 displayed on the display surface 7 of the display 3 so as to be in a different direction from the direction displayed in FIG.
- the direction of the three-dimensional image 8 shown in FIG. 2 and the direction of the ⁇ -dimensional image 8 shown in FIG. 4 are not limited as long as the user specifies an arbitrary point in the three-dimensional image and the point is visible. Good.
- the user specifies a desired point. This point designation can also be performed using a mouse or the like as in the above-described operation.
- the control unit 1 passes through the designated point and a straight line in the vertical direction on the display surface 7. Create and memorize. The created line does not need to be displayed on the display 3. Then, the control unit 1 obtains the intersection coordinates of the horizontal plane created and stored when the display screen of FIG. 2 is displayed and the straight line created when the display screen of FIG. A point at the intersection coordinates is displayed on the three-dimensional image as a designated point. For example, the designated point 9 is displayed as shown in FIG.
- the user can directly specify an arbitrary point of the three-dimensional image 8 while viewing the three-dimensional image. Since the point designation is performed based on the three-dimensional image displayed in the first orientation displayed on the display surface 7 and the three-dimensional image displayed in the second orientation, The designated point can be specified easily.
- a three-dimensional image 8 is displayed on the display surface 7 of the display 3 in an arbitrary direction, and a part of the specified image is designated.
- the designated point is indicated by an X.
- a vertical cross section is created on the display surface 7 through the specified point.
- Figure 7 shows the created vertical section.
- the user rotates and moves the three-dimensional image displayed on the display surface 7 in FIG. 6, for example, in the direction shown in FIG.
- the part to be specified is the part that also appeared in Fig. 6.
- FIG. 9 is an explanatory diagram showing the processing of FIGS. 6 to 8 in a plan view.
- 9A is the display surface 7 in FIG. 6, and 7B is the display surface 7 in FIG.
- the vertical section perpendicular to the display surface 7 formed in FIG. 7 is represented as VS in FIG.
- a line passing through the point specified in FIG. 8 and perpendicular to the display surface 7B is represented by VL in FIG.
- VL a line passing through the point specified in FIG. 8 and perpendicular to the display surface 7B.
- FIGS. 10 and 11 are flowcharts showing the processing operation of the control unit 1 shown in FIG. Next, a description will be given according to the flow of FIGS. 10 and 11 with reference to FIGS. 1 to 9.
- a three-dimensional image is displayed on the display 3 (step S1).
- the displayed three-dimensional image is an image that is three-dimensionally constructed based on tomographic images such as MRI and CT.
- Such a program for three-dimensional construction is installed in the control unit 1 in advance, and the control unit 1 displays a tomographic image as a three-dimensional image based on the three-dimensional image construction program.
- the three-dimensional image may be an image constructed by a surface rendering method or an image constructed by a volume rendering method. Alternatively, it may be a three-dimensional image based on data such as a pseudo projection method.
- the controller 1 determines whether or not the mode is the three-dimensional point input mode (step S2). When performing three-dimensional point input, the device is set to that mode. If the mode is not the three-dimensional point input mode, the control unit 1 performs another process (step S18).
- the control unit 1 performs a process of displaying the three-dimensional image 3 displayed on the display surface 7 of the display 3 in a direction desired by the user (step S3).
- the direction desired by the user is a direction in which the user can visually recognize a point to be designated in the three-dimensional image.
- the user cannot visually recognize the point that the user wants to specify, so the user rotates the three-dimensional image so that the point can be visually recognized and displays it. Therefore, it is possible to specify a point to be specified while checking it on the three-dimensional image.
- the user specifies a point on the three-dimensional image 8 displayed on the display surface 7. As described above, this point can be specified by moving the cursor to a position to be specified using a mouse or the like.
- control unit 1 determines that the first designated point has been input (step S 4), it passes through the input point, is parallel to the viewpoint direction, and is perpendicular to the display surface 7 (perpendicular to the display surface 7).
- the direction is a vertical or horizontal cross section with respect to the display surface 7.) is created and stored in the memory 2 (step S5).
- an instruction to input the second designated point is displayed on the display 3.
- the user operates the image displayed on the display surface 7 so as to be different from the first direction.
- the three-dimensional image 8 on the display surface 7 is displayed in the second direction (step S6).
- Step S7 determines that there is a designated point input at the point (step S7), and draws a line passing through the designated input point and parallel to the viewpoint direction (a straight line perpendicular to the display surface 7). Create and memorize (Step S8).
- step S8 the coordinates of the intersection of the plane (cross section) created in step S5 and the straight line created in step S7 are calculated (step S8), and the intersection coordinates are stored in a memory as a designated point, and The point specified by the coordinates is displayed on the three-dimensional image 8 on the display surface 7 (step S9).
- step S5 the plane created in step S5 and the straight line created in step S8 may be displayed together with the designated point, or the display of these planes and straight lines may be omitted. . If the displayed designated point is at a position different from the desired point, the user can input that cancellation is to be performed (step S10). If there is a cancel input, the control unit 1 invalidates the points specified in steps S4, S7, and S9 (step S11), and the process returns to step 'S3.
- the user can then change the shape, size and color of the mark representing the designated point. If the user wants to change the shape, size, and color of the mark representing the designated point, the user inputs that fact through the operation unit 4 and the control unit 1 performs the change processing (step S13).
- a mark representing a designated point is displayed on the 3D image due to the relationship with the 3D image to be displayed, if there is a problem such as difficulty in viewing the 3D image, the mark representing the designated point is replaced with the 3D image. It does not have to be displayed above. In this case, if you want to recall a specified point that has already been specified, for example, if you want to extract the area around the specified point and display it in an enlarged view, the edge of the display surface (image display An icon for specifying the designated point may be displayed outside the area. By displaying such an icon at the end of the display surface, it is possible to call up a point on the specified 3D image once, to extract the surrounding area of the point on the specified 3D image, and to enlarge and display it. If you want to create a tomographic image including the specified point, instead of picking the specified point on the 3D image, select the existing specified point by picking the icon displayed in the column outside the image Operation can be simplified.
- CT image is the value of the radiation reflected by human tissue when irradiated with radiation, expressed as the relative value of the X-ray absorption coefficient.
- the relative value of the X-ray absorption coefficient of human tissue when the coefficient is set to 0.For example, it is about 80 to 100 for bone and lime, and about 20 to 70 for blood and organ.
- the control unit 1 performs necessary extraction processing (step S17), and displays the extracted portion, for example, in an enlarged manner.
- the three-dimensional image is displayed in the first orientation, a plane perpendicular to the display surface at that time is created, and then the three-dimensional image is oriented in the second orientation, and the display at that time is performed.
- a plane perpendicular to the display surface at that time is created, and then the three-dimensional image is oriented in the second orientation, and the display at that time is performed.
- two planes perpendicular to the display plane for example, Direction and horizontal direction
- one plane perpendicular to the display surface is created, and those three planes are created. May be determined and that point may be set as the designated point.
- first create a straight line then create a surface, and specify the intersection as a specified point.
- Any method may be used in which an arbitrary point is specified, a plane or a straight line passing through the specified point and perpendicular to the display surface is created, and the specified point is specified from the intersection.
- the image displayed on the display surface is described as being a three-dimensionally constructed image.
- the image displayed on the display surface is not limited to the three-dimensionally constructed image, and may be M a X-IF, It may be a three-dimensional projection display image from any direction such as Min_IF, Rayum, or the like.
- an image of a display method that performs three-dimensional projection display from an arbitrary direction such as a MIP (Maximum Intensity Projection) method or a Laysum method.
- Images in this notation are displayed as images like ordinary radiographs, but have three-dimensional information. Therefore, while rotating the image in an arbitrary direction, a projected image in that state can be created and displayed as needed.
- the present invention can be applied to an image of the relevant display method.
- FIG. 12 when a point is specified on the projection display screen in one direction, the point is specified.
- a plane perpendicular to the screen is created.
- the created plane is displayed as a straight line on the projection screen, but it is displayed on the image as a straight line because the plane perpendicular to the screen is projected from the viewpoint direction.
- the computer has three-dimensional information.
- a projection display screen in another direction is displayed. That is, a projection display screen with a changed viewpoint direction is displayed.
- a projection display screen with a changed viewpoint direction is displayed.
- you display the surface created by specifying the point first it will be displayed as a surface instead of a straight line.
- a second point is designated from this viewpoint, the three-dimensional coordinates of the intersection of a straight line passing through the point and perpendicular to the screen and the plane can be calculated in the control unit 1 and designated. Point coordinates can be retained. Also, the designated point can be displayed on the image.
- the present invention can be applied to a point input in a medical image three-dimensionally projected and displayed from an arbitrary direction.
- an arbitrary point in a medical three-dimensional image can be specified as follows.
- display control is performed such that the orientation of the three-dimensional image or the three-dimensional projected image displayed on the display surface is set to an arbitrary first orientation according to a user operation.
- a cross-sectional image passing through the point and perpendicular to the display surface is created. Then, the created cross-sectional image is displayed on the display surface.
- the user can specify an arbitrary point on the displayed cross-sectional image. Since the specified point is a point on the cross-sectional image, the three-dimensional coordinates of the point are calculated by the control unit 1 and held. Thereby, any point in the three-dimensional image can be specified.
- condition setting for creating the cross-sectional image passes through the point specified in the first direction and is specified as a cross-section perpendicular to the display surface, the cross-sectional image of the position, Even if you do not make complicated settings for creation, simply specify one point in the first direction and a section perpendicular to the display surface will be created automatically through that point, so section creation is Easy.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- Architecture (AREA)
- Apparatus For Radiation Diagnosis (AREA)
- Processing Or Creating Images (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
Description
Claims
Priority Applications (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE60141834T DE60141834D1 (de) | 2000-06-16 | 2001-06-18 | Punkteingabeeinrichtung und verfahren f r dreidimensionale bilder |
EP01938717A EP1308903B1 (en) | 2000-06-16 | 2001-06-18 | Point inputting device and method for three-dimensional images |
CA002411193A CA2411193C (en) | 2000-06-16 | 2001-06-18 | Point inputting device and method for three-dimensional images |
JP2002511295A JP4141253B2 (ja) | 2000-06-16 | 2001-06-18 | 三次元画像のための点入力装置および方法 |
KR1020027016155A KR100770248B1 (ko) | 2000-06-16 | 2001-06-18 | 삼차원 화상을 위한 점입력장치 및 방법, 및 그 방법을 실현하기 위한 프로그램을 기록한 컴퓨터로 판독가능한 기록매체 |
AT01938717T ATE464623T1 (de) | 2000-06-16 | 2001-06-18 | Punkteingabeeinrichtung und verfahren f r dreidimensionale bilder |
US10/296,717 US7215325B2 (en) | 2000-06-16 | 2001-06-18 | Point inputting device and method for three-dimensional images |
CNB018112927A CN1248168C (zh) | 2000-06-16 | 2001-06-18 | 用于三维图象的点输入装置和方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2000-182069 | 2000-06-16 | ||
JP2000182069 | 2000-06-16 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2001097174A1 true WO2001097174A1 (fr) | 2001-12-20 |
Family
ID=18682907
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2001/005167 WO2001097174A1 (fr) | 2000-06-16 | 2001-06-18 | Dispositif et procede d'entree de points pour images tridimensionnelles |
Country Status (9)
Country | Link |
---|---|
US (1) | US7215325B2 (ja) |
EP (1) | EP1308903B1 (ja) |
JP (1) | JP4141253B2 (ja) |
KR (1) | KR100770248B1 (ja) |
CN (1) | CN1248168C (ja) |
AT (1) | ATE464623T1 (ja) |
CA (1) | CA2411193C (ja) |
DE (1) | DE60141834D1 (ja) |
WO (1) | WO2001097174A1 (ja) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007130240A (ja) * | 2005-11-10 | 2007-05-31 | Imagunooshisu Kk | 医用三次元画像の表示方向修正装置、修正方法および修正用プログラム |
JP2008520312A (ja) * | 2004-11-23 | 2008-06-19 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | インターベンション手順の間の画像表示用の画像処理システム及び方法 |
JP2010162340A (ja) * | 2008-12-15 | 2010-07-29 | Fujifilm Corp | 画像処理装置および方法並びにプログラム |
JP5495357B2 (ja) * | 2005-10-07 | 2014-05-21 | 株式会社日立メディコ | 画像表示方法及び医用画像診断システム |
US9330496B2 (en) | 2009-12-28 | 2016-05-03 | Fukuoka University | Three dimensional coordinate location device, method for same, and program |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1449481A4 (en) * | 2001-10-31 | 2010-06-23 | Imagnosis Inc | MEDICAL PICTURE PROCESSING DEVICE, METHOD AND PROGRAM |
JP4319156B2 (ja) * | 2005-03-02 | 2009-08-26 | 任天堂株式会社 | 情報処理プログラムおよび情報処理装置 |
JP4732925B2 (ja) * | 2006-03-09 | 2011-07-27 | イマグノーシス株式会社 | 医用画像の表示方法およびそのプログラム |
CA2645539A1 (en) * | 2006-03-09 | 2007-09-13 | Imagnosis Inc. | Medical 3-dimensional image display control program and medical 3-dimensional image display method |
US20100328304A1 (en) * | 2006-12-14 | 2010-12-30 | Imagnosis Inc. | Display direction correcting device, correcting method, and correction program for medical 3d image |
JP5224451B2 (ja) * | 2008-06-03 | 2013-07-03 | 富士フイルム株式会社 | 投影画像作成装置、方法およびプログラム |
ATE529841T1 (de) | 2008-11-28 | 2011-11-15 | Agfa Healthcare Nv | Verfahren und vorrichtung zur bestimmung einer position in einem bild, insbesondere einem medizinischen bild |
KR101941644B1 (ko) * | 2011-07-19 | 2019-01-23 | 삼성전자 주식회사 | 휴대 단말기의 피드백 제공 방법 및 장치 |
JP6265027B2 (ja) * | 2014-04-22 | 2018-01-24 | 富士通株式会社 | 表示装置、位置特定プログラム、および位置特定方法 |
CN111772664B (zh) * | 2020-06-24 | 2024-02-13 | 北京朗视仪器股份有限公司 | 一种快速头影测量装置及方法 |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0916814A (ja) * | 1995-06-30 | 1997-01-17 | Shimadzu Corp | 医用画像表示装置 |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0198084A (ja) * | 1987-10-09 | 1989-04-17 | Toshiba Corp | 3次元画像処理装置 |
JPH03231643A (ja) * | 1990-02-08 | 1991-10-15 | Toshiba Corp | 画像表示装置 |
US5602564A (en) | 1991-11-14 | 1997-02-11 | Hitachi, Ltd. | Graphic data processing system |
JPH06337920A (ja) * | 1993-05-28 | 1994-12-06 | Toshiba Medical Eng Co Ltd | 画像処理装置 |
US6052100A (en) * | 1994-03-16 | 2000-04-18 | The United States Of America Represented By The Secertary Of The Navy | Computer controlled three-dimensional volumetric display |
JP3788847B2 (ja) * | 1997-06-23 | 2006-06-21 | 東芝医用システムエンジニアリング株式会社 | 画像処理装置 |
JPH1186039A (ja) * | 1997-09-04 | 1999-03-30 | Toshiba Iyou Syst Eng Kk | 画像処理支援装置 |
US6273896B1 (en) * | 1998-04-21 | 2001-08-14 | Neutar, Llc | Removable frames for stereotactic localization |
US6546277B1 (en) * | 1998-04-21 | 2003-04-08 | Neutar L.L.C. | Instrument guidance system for spinal and other surgery |
JP4290273B2 (ja) | 1999-01-13 | 2009-07-01 | 株式会社日立メディコ | 画像表示装置 |
-
2001
- 2001-06-18 CN CNB018112927A patent/CN1248168C/zh not_active Expired - Fee Related
- 2001-06-18 DE DE60141834T patent/DE60141834D1/de not_active Expired - Lifetime
- 2001-06-18 AT AT01938717T patent/ATE464623T1/de not_active IP Right Cessation
- 2001-06-18 CA CA002411193A patent/CA2411193C/en not_active Expired - Fee Related
- 2001-06-18 KR KR1020027016155A patent/KR100770248B1/ko not_active IP Right Cessation
- 2001-06-18 WO PCT/JP2001/005167 patent/WO2001097174A1/ja active Application Filing
- 2001-06-18 EP EP01938717A patent/EP1308903B1/en not_active Expired - Lifetime
- 2001-06-18 JP JP2002511295A patent/JP4141253B2/ja not_active Expired - Fee Related
- 2001-06-18 US US10/296,717 patent/US7215325B2/en not_active Expired - Fee Related
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0916814A (ja) * | 1995-06-30 | 1997-01-17 | Shimadzu Corp | 医用画像表示装置 |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2008520312A (ja) * | 2004-11-23 | 2008-06-19 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | インターベンション手順の間の画像表示用の画像処理システム及び方法 |
JP5495357B2 (ja) * | 2005-10-07 | 2014-05-21 | 株式会社日立メディコ | 画像表示方法及び医用画像診断システム |
US8747319B2 (en) | 2005-10-07 | 2014-06-10 | Hitachi Medical Corporation | Image displaying method and medical image diagnostic system |
JP2007130240A (ja) * | 2005-11-10 | 2007-05-31 | Imagunooshisu Kk | 医用三次元画像の表示方向修正装置、修正方法および修正用プログラム |
JP4638330B2 (ja) * | 2005-11-10 | 2011-02-23 | イマグノーシス株式会社 | 医用三次元画像の表示方向修正装置、修正方法および修正用プログラム |
JP2010162340A (ja) * | 2008-12-15 | 2010-07-29 | Fujifilm Corp | 画像処理装置および方法並びにプログラム |
US9330496B2 (en) | 2009-12-28 | 2016-05-03 | Fukuoka University | Three dimensional coordinate location device, method for same, and program |
Also Published As
Publication number | Publication date |
---|---|
KR100770248B1 (ko) | 2007-10-25 |
KR20030013432A (ko) | 2003-02-14 |
CA2411193A1 (en) | 2002-11-25 |
CN1248168C (zh) | 2006-03-29 |
US7215325B2 (en) | 2007-05-08 |
CA2411193C (en) | 2009-02-03 |
ATE464623T1 (de) | 2010-04-15 |
EP1308903A4 (en) | 2009-05-13 |
DE60141834D1 (de) | 2010-05-27 |
CN1443342A (zh) | 2003-09-17 |
EP1308903B1 (en) | 2010-04-14 |
US20030132933A1 (en) | 2003-07-17 |
JP4141253B2 (ja) | 2008-08-27 |
EP1308903A1 (en) | 2003-05-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9014438B2 (en) | Method and apparatus featuring simple click style interactions according to a clinical task workflow | |
US10420533B2 (en) | Ultrasound diagnosis apparatus and control method thereof | |
US7889227B2 (en) | Intuitive user interface for endoscopic view visualization | |
JP5427179B2 (ja) | 解剖学的データの視覚化 | |
WO2001097174A1 (fr) | Dispositif et procede d'entree de points pour images tridimensionnelles | |
US20070279436A1 (en) | Method and system for selective visualization and interaction with 3D image data, in a tunnel viewer | |
JPH1011614A (ja) | 三次元画像構成法における視点位置及び視線方向の設定方法及び装置 | |
JPH08107875A (ja) | 内視鏡形状検出装置 | |
JP2006513503A (ja) | 三次元表示における複数箇所を管理する装置及び方法 | |
KR20020041290A (ko) | 3차원 슬랩 렌더링 시스템 및 그 방법 | |
CN113645896A (zh) | 手术计划、手术导航和成像用系统 | |
US20140055448A1 (en) | 3D Image Navigation Method | |
KR20130089645A (ko) | 정보 비주얼화를 위한 방법, 장치 및 어레인지먼트 | |
KR101611484B1 (ko) | 의료영상의 제공방법 | |
JP4159396B2 (ja) | 内視鏡形状検出装置 | |
JP2001195610A (ja) | 画像処理装置 | |
JP2004089599A (ja) | 画像表示装置 | |
JP2001087228A (ja) | 読影支援装置 | |
JPH11283018A (ja) | 医用画像処理装置 | |
JP4536176B2 (ja) | カラー画像表示装置およびその方法 | |
JP2001101449A (ja) | 三次元画像表示装置 | |
JP3301654B2 (ja) | 医用画像処理装置 | |
JP3441123B2 (ja) | 画像表示装置 | |
JPH07146954A (ja) | 医用画像診断装置 | |
JPH06149967A (ja) | 三次元画像診断装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): CA CN JP KR US |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2001938717 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2411193 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 10296717 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020027016155 Country of ref document: KR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 018112927 Country of ref document: CN |
|
WWP | Wipo information: published in national office |
Ref document number: 1020027016155 Country of ref document: KR |
|
WWP | Wipo information: published in national office |
Ref document number: 2001938717 Country of ref document: EP |