WO1997020288A1 - Systeme d'imagerie en trois dimensions - Google Patents
Systeme d'imagerie en trois dimensions Download PDFInfo
- Publication number
- WO1997020288A1 WO1997020288A1 PCT/CA1996/000777 CA9600777W WO9720288A1 WO 1997020288 A1 WO1997020288 A1 WO 1997020288A1 CA 9600777 W CA9600777 W CA 9600777W WO 9720288 A1 WO9720288 A1 WO 9720288A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- dimensional
- image
- image data
- dimensional images
- ultrasound
- Prior art date
Links
- QWHNJUXXYKPLQM-UHFFFAOYSA-N CC1(C)CCCC1 Chemical compound CC1(C)CCCC1 QWHNJUXXYKPLQM-UHFFFAOYSA-N 0.000 description 1
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/42—Details of probe positioning or probe attachment to the patient
- A61B8/4209—Details of probe positioning or probe attachment to the patient by using holders, e.g. positioning frames
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/89—Sonar systems specially adapted for specific applications for mapping or imaging
- G01S15/8906—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
- G01S15/8934—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques using a dynamic transducer configuration
- G01S15/8945—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques using a dynamic transducer configuration using transducers mounted for linear mechanical movement
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S15/00—Systems using the reflection or reradiation of acoustic waves, e.g. sonar systems
- G01S15/88—Sonar systems specially adapted for specific applications
- G01S15/89—Sonar systems specially adapted for specific applications for mapping or imaging
- G01S15/8906—Short-range imaging systems; Acoustic microscope systems using pulse-echo techniques
- G01S15/8993—Three dimensional imaging systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/08—Volume rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/12—Diagnosis using ultrasonic, sonic or infrasonic waves in body cavities or body tracts, e.g. by using catheters
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/13—Tomography
- A61B8/14—Echo-tomography
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/48—Diagnostic techniques
- A61B8/483—Diagnostic techniques involving the acquisition of a 3D volume of data
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/52017—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
- G01S7/52023—Details of receivers
- G01S7/52025—Details of receivers for pulse systems
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/52017—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
- G01S7/52023—Details of receivers
- G01S7/52034—Data rate converters
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/52017—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
- G01S7/52053—Display arrangements
- G01S7/52057—Cathode ray tube displays
- G01S7/5206—Two-dimensional coordinated display of distance and direction; B-scan display
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/52—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00
- G01S7/52017—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S15/00 particularly adapted to short-range imaging
- G01S7/52053—Display arrangements
- G01S7/52057—Cathode ray tube displays
- G01S7/52071—Multicolour displays; using colour coding; Optimising colour or information content in displays, e.g. parametric imaging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/41—Medical
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2219/00—Indexing scheme for manipulating 3D models or images for computer graphics
- G06T2219/008—Cut plane or projection plane definition
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S128/00—Surgery
- Y10S128/916—Ultrasound 3-D imaging
Definitions
- the present invention relates to medical diagnostics and in particular to a method and system for constructing and displaying three-dimensional images.
- TRUS transrectal ultrasound
- Conventional ultrasound diagnostic equipment typically comprise an ultrasound probe for transmitting ultrasound signals into the subject and receiving reflected ultrasound signals therefrom.
- the reflected ultrasound signals received by the ultrasound probe are processed and a two-dimensional image of the target under examination is formed.
- this conventional equipment produces two-dimensional images even though the target under examination is three-dimensional. Also, the two- dimensional images represent a single thin plane taken at an arbitrary angle to the target making it very difficult to localize the image plane in the target and very difficult to reproduce an image at a particular location at a later time.
- a three- dimensional imaging system for acquiring a succession of two-dimensional images of a target volume represented by an array of pixels I(x,y,z) into a three-dimensional image represented by a volumetric image array V(x,y,z) comprising: scanning means to scan said target volume along a predetermined geometric scanning path and generate a succession of digitized two-dimensional images thereof representing cross-sections of said target volume on a plurality of planes spaced along said scanning path; memory storing said succession of digitized two-dimensional images together with other related image data defining the location of said two-dimensional images in said memory and defining interpretation information relating to the relative position of pixels within said two-dimensional images and to the relative position of pixels in adjacent two-dimensional images within said target volume; transformation means receiving said digitized two-dimensional images and said other related image data and transfo- ⁇ r-ing said two-dimensional images and said other related image data into a volumetric image array; and display means to generate a three-dimensional image of said
- the geometric scanning path is linear.
- the cross-sections are tilted with respect to an axis normal to the linear scanning path.
- the transformation means transforms the two-dimensional images and the other related image data into the volumetric image array using a shear transformation.
- the other related image data includes an address pointer indicating the location of the memory of which the two-dimensional image data begins together with data representing the number of pixels along x and y axis of each two-dimensional image.
- the other related image data also includes physical distance values between adjacent pixels in each of the two-dimensional images as well as the distance between corresponding pixels in adjacent two-dimensional images together with the tilt angle of the cross-sections.
- Figure 1 is a perspective view of a three-dimensional ultrasound imaging system
- Figure 2 is a perspective view of an ultrasound probe actuating assembly forming part of the system illustrated in Figure 1;
- Figure 3 is a side view of the probe actuating assembly illustrated in Figure 2 supporting an ultrasound probe;
- Figure 3a is a perspective view of the geometry of the ultrasound probe movement when driven by the ultrasound probe actuating assembly
- Figure 3b is an illustration of a number of tilted two-dimensional images
- Figure 3c is an illustration of a shear transformation
- Figure 4 is a block diagram showing various hardware and software modules of the computer forming part of the system illustrated in Figure 1;
- Figure 5a is a flowchart showing some of the operational steps of the system illustrated in Figure 1;
- Figure 5b is a flowchart showing additional operational steps of the system illustrated in Figure 1; and
- Figure 6 shows a flowchart of steps performed by the display module during initialization
- Figures 7a to 7d show flowcharts of steps performed by the user interface and display modules when manipulating a displayed three-dimensional image
- Figures 8a to 8c show a three-dimensional image and model undergoing a rotation about a vertical axis
- Figures 9a to 9c show a three-dimensional image and model undergoing a rotation in a direction from top-left towards bottom right about an axis, angled at about 30° to the horizontal and sloping up and to the right;
- Figures 10a to 10c show a three-dimensional image and model in which a plane of the model is translated towards the geometric center of the model;
- Figures Ila to l ie show a three-dimensional image and model in which a plane of the model is rotated about an axis, angled at about 30° to the horizontal and sloping up and to the right;
- Figures 12a to 12d show a three-dimensional image and model in which a plane of the model is translated away from the geometric center of the model until it disappears;
- Figure 13 shows a typical full screen display including a main display window and a control display window
- Figure 14 shows another full screen display further including a measure control window and an animation control window
- Figures 15a to 15c show full screen displays further including an orientation view window. DESCRIPTION OF THE PREFERRED EMBODIMENTS
- three-dimensional ultrasound imaging systems are disclosed.
- the two-dimensional images are stored as a stack to form an image data array.
- the image data array Before a three-dimensional image of the scanned target volume can be created, the image data array must be reconstructed to form a volumetric image array.
- the process of reconstructing the image data array to the volumetric image data array is an unconsuming process which results in significant delays before acquired two-dimensional images can be used to generate a three- dimensional image.
- the present invention relates to a three-dimensional ultrasound imaging system which overcomes this problem by allowing a three-dimensional image to be generated using the two-dimensional image data array without requiring the image data array to undergo reconstruction.
- a three-dimensional ultrasound imaging system in accordance with the present invention is shown and is generally indicated by reference numeral 20.
- the system 20 is capable of generating a three-dimensional ultrasound image of a target volume of a subject under examination from a succession of two- dimensional ultrasound images of the target volume and allow the generated three- dimensional image to be manipulated.
- the subject under examination may be inanimate or animate.
- the system 20 may be used in both medical and veterinary environments and may be used as a diagnostic tool or during surgery to provide updated images of the target volume of the subject undergoing surgery.
- the system 20 includes an ultrasound probe actuating assembly 22 for removably retaining an ultrasound probe 24.
- the probe actuating assembly 22 is designed to move the ultrasound probe along a linear scanning path Z so that the succession of two-dimensional images of the target volume can be taken.
- the ultrasound probe 24 is connected to a clinical ultrasound machine 28 via a commumcation line 30.
- the ultrasound machine 28 in turn is connected to a computer 32 via communication line 34.
- the computer 32 includes a keyboard (not shown), a monitor 36 with a display screen 36a and a graphical input device 38 such as a single button mouse. It should however be realized that many other graphical input devices can be used to allow a user to input commands to the computer.
- the computer 32 provides output signals to a controller 40 via communication line 42 which in turn provides control signals to the probe actuating assembly 22 via communication line 44 to control the scanning motion of the ultrasound probe 24.
- the probe actuating assembly 22 includes a housing 22a accommodating a motor 22b coupled to a threaded output shaft 22c by way of a pair of reduction wheels 22d and 22e and a belt 22 f.
- An I-block 22g is mounted to the threaded shaft 22c via an internal threaded hole so that the I-block 22g moves along the linear scanning path Z in either direction as represented by arrow B when the threaded shaft 22c is rotated via the motor 22b.
- a probe holder 22h is secured to the I-block 22g by way of screws 22i and 22j and supports the ultrasound probe 24.
- the probe holder 22h is positioned relative to the threaded shaft 22c so that the longitudinal axis of the ultrasound probe 24 forms an angle x with an axis normal to the longitudinal axis of the threaded shaft 22c.
- the angle of inclination of the probe holder 22h can be adjusted by loosening the screws 22i and 22j, rotating the probe holder 22h to the desired position and tightening the screws.
- the probe actuating assembly 22 is particularly suited to applications where ultrasound images of a subject's internal organs or lesions such as breast tumours within the trunk of a subject P are to be taken.
- a layer of coupling gel 22k is typically placed on the subject P between the subject and the ultrasound probe 24.
- the motor 22b is operated so that it rotates at a constant velocity.
- the threaded shaft 22c is rotated at a constant rate causing the probe holder 22h to move the ultrasound probe 24 at a constant rate along the linear scanning path Z.
- the ultrasound probe 24 moves along the linear scanning path Z, it transmits ultrasound signals at specific predefined intervals which impinge on the target volume.
- Reflected ultrasound signals from the target volume are also received by the probe 24 and are converted into analog signals by a crystal (not shown) in the ultrasound probe 24. These analog signals are conveyed to the clinical ultrasound machine 28 where a succession of two-dimensional analog images of the target volume are generated. Each two-dimensional image represents a cross-section of the target volume having x and y dimensions (see Figure 3a).
- the operation of the ultrasound probe 24 and clinical ultrasound machine 28 is well known to those of skill in the art and therefore, will not be described in any further detail herein.
- the relative position of the two-dimensional analog images of the target volume along the z-axis can be readily determined.
- the two-dimensional analog images generated by the ultrasound machine 28 are conveyed to the computer 32 via commumcation line 34.
- the computer 32 in turn digitizes the two-dimensional images and stores the digitized image data together with other related image data in a manner which allows a three-dimensional image of the target volume to be displayed virtually without delay as will be described. Once displayed, the computer 32 allows the image to be manipulated as will also be described.
- the computer 32 includes a frame grabber module 80, such as for example, an IMAXX
- the frame grabber module 80 captures and digitizes the succession of two-dimensional analog images. Once digitized, the succession of two-dimensional images is stored in local physical memory 82.
- the computer 32 also includes a user interface module 84 for interpreting input commands received via the graphical input device 38.
- the user interface module 84 controls and co-ordinates the operation of the other modules of system 20 in response to input from the graphical input device 38 allowing the user to control the system as desired.
- the digitized information can be processed in a number of ways depending on the input commands received by the user interface module 84 from the graphical input device 38. Specifically, the digitized information can be transferred to an external file storage memory 88. The digitized information, whether stored in local physical memory 82 or in the external file storage memory 88, may be processed by a display module 92 in response to input received from graphical input device 38 so that a three-dimensional image of the target volume can be displayed on the screen 36a of the momtor 36 and manipulated as will be described further herein.
- the computer 32 also includes a probe scanning control module 98 which provides output signals to controller 40 to actuate the probe actuating assembly 22 and sweep the ultrasound probe 24 along the linear scanning path Z as desired.
- the probe scanning control module 98 also receives input from the user interface module
- the ultrasound images taken by the ultrasound machine are cross-sections (or image slices) of the target volume on parallel planes tilted by the angle ⁇ (see Figure 3b).
- Shear transformations which are well known to those of skill in the art of computer graphics, are linear.
- the linearity of both the scanning path Z and the shear transformation allows the display module 92 to use the shear transformation as the viewing transformation to convert pixel co-ordinates into corresponding voxel co- ordinates.
- the display module 92 can act directly on the acquired digitized two-dimensional ultrasound image data and generate a three-dimensional image without requiring the acquired ultrasound image data to undergo the reconstruction processes described in Applicant's U.S. Patent No. 5,454,371 and patent application serial no. 08/419,049. This significantly increases the speed by which a three- dimensional image may be displayed.
- the display module 92 In order for the display module 92 to be able to act directly on the acquired digitized ultrasound image data, the display module 92 requires other related image data which includes: 1.) an address pointer defining the address of the location in the physical memory 82 at which the acquired ultrasound image data starts;
- the extent of the acquired two-dimensional image data including the number of pixels along the x and y axis of each two-dimensional image as well as the number of two-dimensional images taken in the z-direction; 3.) the physical distance between the centers of adjacent pixels in both the x and y directions in each two-dimensional image as well as the physical distance between corresponding pixels in adjacent two-dimensional images (the z-distance); and
- the other related image data referred two at points 1 and 2 above defines the location of the acquired digitized ultrasound image data in the memory. This information allows the display module 92 to locate the acquired ultrasound image data and associate sections of the image data with the proper image slices of the target volume.
- the other related image data referred to at points 3 and 4 above defines image data interpretation information which allows the display module 92 to interpret the acquired ultrasound image data correctly.
- the frame grabber module 80 is programmed to write the acquired digitized ultrasound image data to the appropriate locations of the physical memory 82 and to generate the address pointer which in turn is stored in a calibration file in physical memory 82.
- the system 20 is calibrated. Specifically, during calibration, the velocity of the ultrasound probe 24 along the linear scanning path Z and the ultrasound signal transmit interval are determined so that the number of two-dimensional image slices of the target volume to be taken are known. Also, the number of pixels along the x and y axis of the two-dimensional images are measured. The number of pixels in the x and y directions defines the edges of each two-dimensional ultrasound image. Thus, these numbers are used to locate the ultrasound image data within the scanned target volume. Once the numbers are determined they are also stored in the calibration file.
- the center to center distance between two pixels in the same line of an ultrasound image ie. in the x direction
- the center to center distance between adjacent pixels in two different lines of the ultrasound image are determined
- the center to center distance between corresponding pixels in two adjacent ultrasound images are determined.
- the distance values are stored in the calibration file.
- the tilt angle of the ultrasound probe 24 is also measured and is stored in the calibration file.
- the ultrasound probe transmits ultrasound signals to scan a cross- section of the target volume.
- the reflected ultrasound signals received from the target volume are conveyed to the clinical ultrasound machine 28 wherein a two- dimensional analog image slice of the target volume upon which the ultrasound signals impinged, is created.
- the two-dimensional analog image is then conveyed to the computer 32 via communication line 34 wherein it is captured and digitized via frame grabber module 80.
- the digitized two-dimensional image is then stored in the memory 82.
- a copy of the digitized two-dimensional image is then conveyed to the user interface module 84 and the frame is drawn on the screen 36a of the momtor 36
- block 104 The user then manually moves the probe 24 along the linear scanning path Z while it is transmitting ultrasound signals so that two-dimensional analog images generated by the clinical ultrasound machine 28 are captured and digitized by the frame grabber module 80. These two-dimensional images are also then drawn on the screen 36a of monitor 36 via user interface module 84 (block 106). Next, the user is prompted to confirm that the ultrasound signals are properly directed at the target volume after having viewed the frames drawn on the screen 36a of the momtor (block 108). If the target volume is outside of the drawn frames, then operation returns to block 104. Otherwise, the user provides input to the user interface module 84 using the graphical input device 38 to signify that the target volume is within the drawn frames.
- the system 20 is ready for proper operation.
- a start command such as for example by using an extemal foot or hand switch (not shown)
- the user interface module 84 signals the probe scanning module 98.
- icon refers to any graphical element displayed on the screen 36a which can be selected using graphical input device 38.
- the probe scanning module 98 conveys control signals to the probe actuating assembly 22 via controller 40 so that the ultrasound probe 24 is moved along the linear scanning path Z at a constant velocity. While this occurs, the ultrasound probe 24 is conditioned to transmit ultrasound signals at the predefined intervals so that the entire target volume is scanned. As the ultrasound probe receives reflected ultrasound signals, it conveys analog information to the clinical ultrasound machine 28 which in turn generates two-dimensional analog images. In this manner, a succession of two-dimensional analog images of the target volume representing a volume image are generated by the clinical ultrasound machine 28 in response to the output of the ultrasound probe 24 (block 112). The succession of two-dimensional analog images generated by the clinical ultrasound machine 28 are captured and digitized by the frame grabber module 80.
- the digitized two-dimensional images are then conveyed to the physical memory 82 and stored as a stack to form an array of two-dimensional images I(x,y,z) with the pixels in the array I(x,y,z) representing pixels of the digitized two-dimensional images. Because the computer 32 controls the movement of the probe actuating assembly 22 and the operation of the ultrasound probe 24, the spatial orientation of the individual two-dimensional images relative to the target volume is known.
- the two-dimensional images are considered to be grayscale images.
- the present invention does not depend on the "color" of the two-dimensional images to function properly.
- a grayscale pixel is associated with a gray-level having a value between 0 and (2 n - 1) inclusively, with n being the number of bits required for storing the gray-levels.
- the gray-level 0 is usually used as a "background color” and is said to be Black.
- the user interface module 84 generates a prompt to signify that this stage of the image capturing has been completed. At this time, the user may review the acquired frames individually in the manner described previously
- the user can condition the system 20 to return to block 102. If the two- dimensional images have been acquired correctly, the user interface module 84 generates a prompt to determine if the acquired two-dimensional images together with the other related image data in the calibration file are to be saved in the extemal file storage memory 88 (block 120) together as a sheared volumetric image array
- V(x,y,z) If the user selects the prompt, the acquired two-dimensional images together with the other related image data are saved as the sheared volumetric image array in the external file storage memory 88 (block 124).
- the user is prompted to decide whether the three- dimensional image is to be displayed on the screen 36a of the monitor 36 (block 126).
- the display module 92 retrieves the sheared volumetric image array from the extemal file storage memory. If the image data has not been stored in the external file storage memory 88, the display module 92 retrieves the image data stored in the memory 82 and retrieves the calibration file from the external file storage memory 88. In either case, once the display module 92 receives the image data and other related image data, it uses the data in conjunction with the shear transformation to display a three-dimensional image of the target volume on the screen 36a (block 128) with virtually no delay.
- the displayed image can be manipulated by the user as will be described. During image manipulation, the user can store displayed views in the memory 82 or in the extemal file storage memory
- the user can decide whether the three-dimensional image is to be displayed on the screen of monitor 36 using the graphical input device 38 (block 126).
- the display module 92 enters an initialization routine (see Figure 6).
- the display module 92 sets a number of parameters to their default values as will be explained and uses this info ⁇ nation when an image is to be displayed on the screen 36a (blocks 302 to 306).
- the parameters which are set to their default value are mouse-drag sensitivity, successive refinement sequence and pseudo-color mapping.
- the display module 92 determines whether the image data has been stored with the calibration file as a sheared volumetric image array or resides in the physical memory 82. If the image data has been stored as a sheared volumetric image array, the display module 92 retrieves a copy of the sheared volumetric image array from extemal file storage memory 88 (block 308). If the image data resides in the physical memory 82, the display module retrieves the image data from the memory 82 and retrieves the other related image data in the calibration file from the memory 88.
- the display module 92 checks to see if the image data has been assigned a model (block 310).
- the model is in the form of a convex polyhedron having a plurality of planar faces defined in the same space as the three-dimensional image to be displayed.
- the polyhedron may take a variety of shapes as will be described herein.
- the dimensions of the image data are examined and a model in the form of a parallelepiped, which encloses substantially all of the pixels in the image data array is created and assigned to the image data automatically (block 312).
- the parallelepiped has two interior angles at right angles and a third interior angle equal to the angle « .
- the assigned model is then saved in memory with the image data array.
- the image data array is examined to determine if any or all preferred Views A to C have been assigned to the image data array (block 314). If some or all of the preferred Views have not been assigned, the preferred Views are created automatically and saved in the memory 82 with the image array (block 315).
- FIG. 13 illustrates the screen 36a of monitor 36 on which the model and three-dimensional image are displayed within the main display window MDW. It should have been realized that other projection transformations such as a perspective projection may be used to project the model onto the screen. Only the visible faces of the model are displayed on the screen 36a, i.e., hidden-surface elimination is performed so that the displayed model has an opaque appearance. The perimeter lines of the model faces are depicted on the screen by white lines. Each visible model face is projected onto the screen within a polygonal area. Each visible face's polygonal area is divided into an internal area and an external area, the latter being that part of the total area immediately adjacent to the displayed boundary of the face and represents a small portion of the total area.
- each visible model face is accomplished as follows.
- Each screen pixel within the polygonal area of the displayed face has an associated two- dimensional cartesian coordinate pair (x',y'). With precise specifications of the model surface, this can be extended to a three-dimensional coordinate triple (x',y',z').
- the pixel co-ordinates (x',y',z') may be converted to corresponding voxel coordinates (x,y,z), to select a voxel value within the volumetric image array V(x,y,z).
- the extracted voxel value is indexed into the pseudo-color mapping table to yield a gray-level or color.
- the gray-level or color in rum is used to illuminate the screen pixel. This process is repeated for all screen pixels located within the displayed model faces (block 318). This technique of display is called “texture mapping", and is known to those of skill in the art.
- Each pixel on the display screen is identified by its associated cartesian coordinates (x',y'), which are usually integers.
- Each voxel in the volumetric image array is identified by its associated array indices (x,y,z), which are also usually integers.
- x',y' For pixels within the perimeter of a displayed model face, it is possible to compute from the pixel coordinates (x',y'), a value z' representing the distance from the plane of the screen 36a to the point on the model face which is projected onto that pixel.
- Other techniques would be required for non-planar model faces, but in general, the problem remains one of solving an equation.
- the correspondence between display coordinates (x',y',z') and volumetric image coordinates (x,y,z) is given by the shear transformation.
- the particulars of the shear transformation are re-computed whenever the user decides to change one or more view parameters such as angle of view, display scale, etc.
- the viewing transformation is a simple linear mapping.
- the viewing transformation may be more complex.
- the image coordinates (x,y,z) computed from display coordinates (x'.y'.z') via the shear transformation will not generally be integers, and hence, will not correspond to individual image voxels. In such cases, a voxel value must be computed by interpolation from the nearest available image voxels. This process is called re-sampling, and is known to those of skill in the art. Those of skill in the art will furthermore be aware that a variety of interpolation techniques or "re-sampling methods", are known, and will be aware of their relative advantages and drawbacks.
- the preferred embodiment of the present invention performs the display process in multiple passes, using computationally inexpensive re-sampling methods in earlier passes and progressing to slower, more accurate methods in later passes. It also permits the user to enable or disable selectively, individual passes to choose a satisfactory trade-off between rapidity of response and image fidelity. Furthermore, it is preferred that later passes be interruptible so that if the user requests rapid change of the displayed view, only the earliest passes are performed until such time as there is a pause in user input. At this point, the later passes are performed on the final view only. This technique is called successive refinement and is known to those of skill in the art.
- the initial view is saved in the memory 82 with the image data array and is indexed as the "Reset" view. Therefore, after initialization, four saved views of the image data array exist and can be retrieved from memory and displayed when an associated icon is selected as will be described. It should be apparent that the number of saved views is arbitrary and that fewer or more saved views may be created and saved.
- the display module 92 begins to momtor continuously the graphical input device 38 to detect input commands representing desired manipulations to be performed on the displayed image (see Figures 7a to 7d). When input commands are detected by the display module 92, the display module manipulates the displayed image in accordance with the received input commands.
- the display module 92 has been described as retrieving the two- dimensional ultrasound image data from the physical memory 82 and the other related image data from the memory 88 after the entire target volume has been scanned, it should be appreciated by those of skill in the art that the calibration file may be retrieved by the display module 92 prior to acquisition of the two-dimensional ultrasound image data.
- the other related image data in the calibration file does not describe acquired two-dimensional ultrasound image data per se but rather describes the memory location and characteristics of the structure in physical memory 82 into which the two-dimensional ultrasound images will be captured.
- the display module 92 is able to present a three-dimensional image of the target volume virtually simultaneously as the target volume is being scanned.
- All manipulations of the displayed image can be executed via three actions using the graphical input device 38. These actions are termed “point”, “click” and “drag”. To “point” is to move the graphical input device 38 so that the cursor is positioned on the screen 36a at a desired region without depressing its button. To “click” is to press down on the button of the graphical input device while to “drag” is to move the graphical input device while the button is depressed.
- point- click-drag sequence is used to denote the complete action of moving the cursor into a desired region on the screen via the graphical input device, depressing the button on the graphical input device, moving the graphical input device to another desired region with the button depressed and then releasing the button. Specific manipulations of the image are carried out according to the region in which the click aspect of the point-click-drag sequences occur.
- the graphical input device is monitored to determine whether the user wishes to manipulate the displayed image.
- the manipulations supported by the display module 92 are rotation of the entire model and three-dimensional image about an arbitrary axis, translation of a selected plane of the model and rotation of a selected plane of the model about an arbitrary axis.
- the manner in which the display module 92 interprets movement of the graphical input device 38 and manipulates the displayed model and image in response to the graphical input device 38 will now be described.
- the display module 92 continuously monitors the graphical input device 38 to determine the position of the cursor on the screen and to determine if a click has occurred. Specifically, the display module 92 determines if the cursor is located within the exterior area of a visible model face (block 324). If the cursor is positioned within such an exterior area, the display module 92 colors the perimeter lines of that model face blue (block 326). Otherwise, the display module 92 changes the color of any perimeter lines which may have been blue to their previous color (block 328). The display module 92 also determines if a click has occurred (block 330). If a click has not occurred, the movement of die graphical input device 38 is tracked.
- the display module 92 moves the cursor on the screen 36a to follow the movement of the graphical input device (blocks 332 and 334). As the cursor is moved, the display module 92 reverts back to block 324 so that the perimeter lines of the visible faces change colors depending on the location of the cursor as just described.
- the display module 92 determines if the cursor is positioned over an option icon (block 338). If the cursor is not over an option icon, the click is ignored and the graphical input device 38 is monitored until the button is released (block 339). At that time, the display module 92 reverts back to block 324. If however, the cursor is positioned over an option icon, then the display module 92 executes a routine associated with the selected option icon as will be described.
- the display module 92 determines whether the cursor is located within the interior area or exterior area of a displayed model face or in the background (block 340). If the cursor is located in the background, the display module 92 determines that the user wishes to rotate the entire model and three-dimensional image. In this case, after the button has been depressed, the drag direction and drag distance of the graphical input device 38 is monitored
- the fixed point of rotation for the three-dimensional image is constrained to be the geometric center of the initial model.
- movement of the graphical input device vertically on the screen 36a through the center of the displayed image causes the displayed image to rotate about a horizontal axis
- movement of the graphical input device horizontally through the center 34 of the image causes the displayed image to rotate about a vertical axis.
- Figures 8a to 8c show the model and three-dimensional image within the main window display undergoing a rotation about a vertical axis as the graphical input device 38 is moved to drag the cursor across the main display window from mid-right to mid-left.
- Figures 9a to 9c show the model and three-dimensional image undergoing a rotation about an axis, angled at about 30° to the horizontal and sloping up and to the right, as the graphical input device 38 is moved to drag the cursor across die main display window from top-left to bottom-right.
- this operation gives d e user the sense of taking hold of the displayed image and pulling it around. Further details of this image manipulation can be found in the above-mentioned publication.
- a convex polyhedron can be characterized as the intersection of a set of half-spaces defined by at least four planes, herein called bounding planes. Each face of the polyhedron is a convex polygon embedded in the corresponding bounding plane.
- the number of bounding planes may also be changed. Specifically, new bounding planes may be added and existing planes removed from the mathematical specification of the model. The result is that the model polyhedron gains or loses faces.
- the display module 92 supports two primary manipulations of bounding plane coefficients, namely translation (change of coefficient D, which essentially specifies the perpendicular distance from the plane to the coordinate origin) and rotation (change of coefficients A,B,C, which collectively specify the orientation of the plane relative to the coordinate axes).
- translation change of coefficient D
- rotation change of coefficients A,B,C, which collectively specify the orientation of the plane relative to the coordinate axes.
- the choice of which bounding plane (and hence which corresponding model face) is to be affected, and whether to perform translation or rotation is determined by contextual interpretation of point-click-drag sequences relative to the displayed model.
- the display module 92 also provides means to add and delete bounding planes from the model specification, as will also be described below.
- original bounding planes are aspects of the model assigned to the volumetric image array V(x,y,z) when it is first loaded into memory (blocks 310 and 312 in Figure 6) and planes added in response to user input.
- Model faces corresponding to original bounding planes have their perimeter lines displayed as white lines, while faces corresponding to added planes are indicated using another color, typically yellow or green. Only added planes may be translated, rotated or deleted.
- the original planes represent the boundaries of the volumetric image and, provide the means to support the addition of new planes.
- the display module 92 determines that it is desired to translate the corresponding bounding plane. Thereafter, the display module 92 examines the corresponding plane to determine whether the plane is an original plane i.e. one denoted by white lines (block 354). If the plane is an original plane, a new plane is created and added to the model (block 356). Initially, the added plane is congruent to the original bounding plane.
- the perimeter lines denoting die plane i.e. the perimeter of the conesponding displayed model face
- all perimeter lines of other added planes are colored green (block 357).
- the drag distance and direction of the graphical input device are then monitored (block 358).
- the display module 92 next determines me direction of translation of the added plane. To do this, the display module 92 calculates the dot product of the drag vector and the projection onto the screen 36a of a normal vector to the plane is computed. If the dot product is positive, the plane is translated in the direction of the normal vector; if negative, it is translated in the opposite direction.
- the model specifications are such that all boundary plane normal vectors point away from the interior of the model polyhedron. Hence, movement of the graphical input device 38 which results in a positive dot product pulls the plane outward from the center of the model, while movement of the graphical input device 38 which results in a negative dot product pushes it in (block 360).
- the display module 92 checks to ensure the translation does not result in the collapse of the model (block 362). If it does, the display model 92 reverts back to block 358 without updating the displayed model and three-dimensional image. Otherwise, translation of the added plane occurs, and the points in the volumetric image array V(x,y,z) which correspond to those on the translated plane are texture-mapped onto the plane
- the display module 92 checks to see if the translation can result in the plane disappearing (block 366). If not, the display module 92 proceeds to block 364 and updates the displayed model and three-dimensional image. Otherwise, the display module 92 notes that the translation may result in the disappearance of the plane before proceeding to block 364 (block 368).
- various cross- sections of die image can be viewed as the translated plane slices through the volumetric image array V(x,y,z).
- the display module checks to see if the plane was deleted (block 370). If not, the perimeter lines of the translated plane are colored yellow and all other perimeter lines are colored green (block 372). Otherwise, the perimeter lines of the next most recendy modified plane are colored yellow and all other perimeter lines are colored green (block 374). When the button is released, the display module 92 considers the translation manipulation complete and reverts back to block 324 (block 376).
- FIGS. 10a to 10c show the model and image within the main display window, wherein a plane of the model is translated towards the geometric center of the model.
- the display module 92 determines that the corresponding bounding plane is to be rotated. The display module then examines the plane to determine whether the plane is an original plane (block 380). If the plane is an original plane, a new plane congruent to the original plane is created and added to the model (block 382). The perimeter lines of the added plane are colored yellow. Once this has been done or if at block 380, the plane is not an original plane, the perimeter lines of all other added planes are colored green (block 384).
- the drag distance and direction of the graphical input device 38 are monitored and the rotation axis and direction of the graphical input device are computed by the display module 92 using Shoemake' s technique referred to previously (block 386).
- the display module 92 determines whether the rotation would cause the plane to disappear (block 388). If so, the display module 92 reverts back to block 386 without updating the displayed model and three- dimensional image. Otherwise, the display module 92 rotates the added plane about the geometric center of the initial model by the calculated amount. As die rotation is occurring, the display module 92 updates the image on the screen (block 390). This allows different cross-sections of the image to be viewed as the rotated plane slices through the volumetric image array V(x,y,z).
- Figures Ila to lie show the model and three-dimensional image within the main display window, wherein a plane of the model is rotated about an axis, angled at about 30° to the horizontal and sloping up and to the right. It is through this manipulation that new oblique planes may be added to the model.
- the display module 92 reverts to block 324 (block 392).
- FIGS 12a to 12d show die model and three-dimensional image witiiin the main window display wherein an oblique plane of the model is translated away from the geometric center of the model until it disappears. Altiiough it would seem that if an added plane is translated sufficiendy into the model, the entire model can collapse, the display module 92 does not permit an added plane to be translated so far as to collapse the model (see block 362).
- the display module 92 While die graphical input device 38 is being moved to effect changes in the displayed view and die display is updated showing intermediate positions and orientations of die affected plane or planes, the display module 92 must re-sample the volumetric image array V(x,y,z) and complete the texture mapping process, a process of discrete approximation. In this embodiment, a number of re-sampling methods are available, each of which offers a different compromise between computational speed and image quality.
- the display may be computed (rendered) at less than the full resolution of the monitor screen and/or a simpler interpolation technique may be employed in the re-sampling process.
- the display is computed in as many as three rendering passes, the first and third of which may be disabled if die user so wishes, by selecting the appropriate option icon via the graphical input device 38.
- the enabled/disabled status of each pass is what is actually set to a default state at block 324 during initialization.
- the first enabled pass in the sequence is uninterruptible, i.e. while die graphical input device is being moved, the first enabled pass is performed in its entirety, yielding a succession of complete views on the screen.
- Subsequent enabled passes are automatically interrupted by graphical input device movement, the visible result being that the displayed view is only replaced by a higher quality view (computed by an interruptible rendering pass) when there is a sufficient pause in graphical input device movement.
- the three rendering passes supported by the present embodiment are:
- the display module 92 determines whether an option icon has been selected.
- the available option icons allow the user to select parameters different from die default values, to enhance image display and to execute special feature routines. These option icons include “Reset”, “Views A to C”, “Remember”, “Snapshot”, “Animation”, “Indicator”, “Orientation”, “Fast”, “Smooth”, “Win”, “Lev”, “Magnify” and “Measure”.
- Figure 13 illustrates most of these option icons in a control display window positioned beside the main display window.
- the preferred Views A to C have been labelled by die user as “Sagittal", “Coronal” and “Axial”. The available options which can be selected via an option icon will now be described.
- the Reset icon is selected, the original view of the image and model stored with the volumetric image array V(x,y,z) is recomputed and displayed on the screen, likewise, if one of View A to C icons is selected, the corresponding preferred view is recomputed and displayed.
- the user can substitute the displayed view for the stored view.
- the present embodiment permits the user to activate a distinct window in which the View icon labels (eg. Sagittal, Coronal, Axial etc. in Figure 13) are displayed and to edit the labels as desired. Changes made to the labels persist only for the current session, unless the user elects to save the changes in memory 88 using an option icon provided for that purpose, in which case any preferred Views associated with the data file are overwritten.
- the View icon labels eg. Sagittal, Coronal, Axial etc. in Figure 13
- the current view on the screen is stored in memory 82 overwriting the "Reset" view for the current session only.
- the "Reset" view associated with the current data file in memory 88 is not changed, only d e copy in memory 82. This view may be recalled to the screen at any time by selecting the
- the image currendy displayed in the main display window is stored as a file in memory 88 in an industry-standard image file format, in order that it be may subsequendy be used with other software.
- the present embodiment uses a tagged image file format ("TIFF"). It should be realized that adding support for other file formats can be achieved in a straightforward manner by following published format specifications.
- Animation icon animated sequences of displayed views can be created and saved into memory 82 in an industry-standard image format as just described.
- the display module 92 determines whether a view of the image has been saved using the Remember icon and retrieves it. If no view has been saved using the Remember icon, the original view is retrieved. While this is occurring, an animation display window appears on the screen 36a. The display window allows die user to select the number of intermediate views of the displayed image which are to be computed and displayed (see Figure
- the animation display window also allows the user to adjust the image size, assign an identifier to the animation sequence and preview the animation to ensure that the selected parameters are satisfactory.
- the display module 92 computes the view orientation, position and orientation of each plane of the intermediate views by simultaneous interpolation between the saved and current views.
- the model is displayed with axis indicators to indicate standard directions of the image, such as front F, back B, left L etc.
- axis indicators to indicate standard directions of the image, such as front F, back B, left L etc.
- These symbols may be changed by die user to suit die application.
- the symbols can represent the standard ocular axes, namely superior S, inferior I, nasal N and temporal T.
- These indicators float as die view of die displayed image changes. To avoid cluttering, it is preferred that axis indicators disappear when the curcent view orientation would place then behind the displayed model.
- a representation of the stmcture which has been imaged In some applications such as in medical imaging, it is desired to show the current spatial relationship of the model to a representation of the stmcture which has been imaged. This can be achieved when the Orientation icon is selected. When this icon is selected, d e representation of the stmcture being imaged is selected from a list of stored stmctures. The stmcture is modelled using conventional three- dimensional computer graphics techniques. The stmcture is then displayed as a semi- transparent solid object intersecting the model polyhedron, either in the main display window or in a second display window elsewhere on the monitor screen. This requires use of a rendering algorithm which makes it apparent how the stmcture and model polyhedron intersect.
- the position, size and spatial orientation of the stmcture relative to the model polyhedron which may be expressed as a 4 x 4 transformation matrix, must be determined.
- the stmcture display window is updated as the user manipulates the displayed view, so that the two displays are always oriented in d e same manner.
- This approach which works best when the stmcture is displayed in a window distinct from the main window, is used in the present embodiment.
- Figures 15a to 15c show the model and three-dimensional image within the main window display as well as the structure of an eye in a second display window beside the main window display.
- die second display window is small and is positioned below the control display window while in Figure 15c, the second display window has been increased in size.
- the Fast and Smooth icons may be individually selected or deselected to enable or disable the first and third rendering passes described previously (die second pass is always enabled).
- the initial state of these icons is established during initialization at block 324. It should be realized that this general scheme can be altered slighdy, e.g. by addition of a fourth pass with a conesponding option icon to selectively enable or disable it if desired.
- Each displayed point of the image array V(x,y,z) is converted to a pixel brightness or color by pseudo-color mapping.
- the domain of the pseudo-color mapping is the range of values in the volumetric image array V(x,y,z).
- the pseudo ⁇ color mapping can be adjusted by a user via the window and level slide controls (labelled “Win” and “Lev” in Figures 13 to 15) to allow the contrast, brightness etc. of the display to be enhanced.
- the terms “window” and "level” and their interpretation have become standardized in the medical imaging field. The present embodiment is consistent with established medical imaging practice in this regard.
- a magnifying window appears superimposed upon die main display window and can be moved over die displayed view.
- Cross-hairs are located at the center of the window and can be positioned over a certain area of the displayed view.
- the user can use the graphical input device to adjust the magnification of the area at which the cross-hairs are located.
- a measure display window appears on the screen (see Figure 14).
- the user can use the graphical input device to measure distances and areas of the three-dimensional image within the most recendy moved plane (i.e. the one denoted by yellow lines). If the user wishes to measure a distance, the user simply needs to use the graphical input device 38 to indicate the two end points over which the distance is to be measured. If an area is to be measured, the user must identify at least three points. When the cursor is moved over the most recently moved plane, it changes to cross-hairs to facilitate placement of the points on the image.
- the display module 92 in this mode connects adjacent points by straight line segments and computes both the overall line length and the area bounded by the lines joining the points using an appropriate scale.
- the user can depress a specified key on the keyboard and this will adjust the scaling factor for as long as the key is depressed. This allows for greater or smaller mouse movements for a given change in the displayed image. This of course can also be achieved by increasing or decreasing die size of the main display window.
- the cine loop technique may also be applied to visualize four dimensional image arrays.
- the four-dimensional image arrays are a series of three-dimensional images of the same target volume, acquired at different times.
- the three-dimensional images will all have the same spatial interpretation and hence, it is possible to impose a single bounding model upon all of them simultaneously.
- the intersection of the model polyhedron with each volume image yields a two-dimensional image.
- the images can be computed and displayed in time order as a cine loop.
- die system has been described as including a single button mouse to allow a user to input commands, it should be apparent to those of skill in the art that other input/output devices such as a multi-button mouse, a digitizer, a light pen, a trackball, a keyboard or the like or any combination of the above can be used. When other input/output devices are used, different inputs can be chosen to represent different commands or to select the various option icons.
- a graphical input device otiier than the single button mouse
- manipulation of an image by rotation of a plane of the model can be enhanced.
- the fixed point of rotation of the plane can be determined by the position of d e mouse when it is clicked and the keyboard input signifying rotation of the plane is selected. This allows the plane to be rotated about a point different from the geometric center of the initial model.
- buttons can be assigned to signify translation of a plane and the other button can be used to signify rotation of a plane. This avoids the need to examine the position of the cursor to determine whether it is within interior or exterior areas of the selected model face.
- the three-dimensional image display technique has been described in an ultrasound three-dimensional imaging system, it should be apparent that the image display technique may be used in different environments where a three- dimensional image needs to be manipulated to allow different views of the image to be visualized.
- the present display technique can be used to display any volumetric image array V(x,y,z) which represents a discretely sampled function defined over a three-dimensional space. This allows the display technique to be used in other environments such as, for example, magnetic resonance imaging (MRI) and x-ray computed tomography (CT).
- MRI magnetic resonance imaging
- CT x-ray computed tomography
- system 20 has been described as including a clinical ultrasound machine 28 and a computer 32, it is contemplated that a single machine be used to perform the functions of both of these components. Variations and modifications of the present invention should be apparent to those of skill in the art without departing from the scope of the present invention as defined by the dependant claims.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- General Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Theoretical Computer Science (AREA)
- Computer Graphics (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Networks & Wireless Communication (AREA)
- Biomedical Technology (AREA)
- Software Systems (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Molecular Biology (AREA)
- Medical Informatics (AREA)
- Heart & Thoracic Surgery (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Surgery (AREA)
- Radiology & Medical Imaging (AREA)
- Pathology (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Biophysics (AREA)
- Computing Systems (AREA)
- Geometry (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
- Image Processing (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Instruments For Viewing The Inside Of Hollow Bodies (AREA)
- Endoscopes (AREA)
Abstract
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AU76154/96A AU7615496A (en) | 1995-11-24 | 1996-11-25 | Three-dimensional imaging system |
DE69619262T DE69619262T2 (de) | 1995-11-24 | 1996-11-25 | Dreidimensionales bildverarbeitungssystem |
AT96938879T ATE213351T1 (de) | 1995-11-24 | 1996-11-25 | Dreidimensionales bildverarbeitungssystem |
CA002238184A CA2238184C (fr) | 1995-11-24 | 1996-11-25 | Systeme d'imagerie en trois dimensions |
JP9520027A JP2000500679A (ja) | 1995-11-24 | 1996-11-25 | 三次元画像形成システム |
EP96938879A EP0862766B1 (fr) | 1995-11-24 | 1996-11-25 | Systeme d'imagerie en trois dimensions |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US08/562,590 US5842473A (en) | 1993-11-29 | 1995-11-24 | Three-dimensional imaging system |
US08/562,590 | 1995-11-24 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO1997020288A1 true WO1997020288A1 (fr) | 1997-06-05 |
Family
ID=24246904
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CA1996/000777 WO1997020288A1 (fr) | 1995-11-24 | 1996-11-25 | Systeme d'imagerie en trois dimensions |
Country Status (8)
Country | Link |
---|---|
US (3) | US5842473A (fr) |
EP (1) | EP0862766B1 (fr) |
JP (1) | JP2000500679A (fr) |
AT (1) | ATE213351T1 (fr) |
AU (1) | AU7615496A (fr) |
CA (1) | CA2238184C (fr) |
DE (1) | DE69619262T2 (fr) |
WO (1) | WO1997020288A1 (fr) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1998043109A1 (fr) * | 1997-03-21 | 1998-10-01 | Life Imaging Systems Inc. | Systeme d'imagerie tridimensionnelle |
WO1999000675A1 (fr) * | 1997-06-25 | 1999-01-07 | Life Imaging Systems Inc. | Systeme et procede de presentation dynamique de donnees d'images en 3d |
EP0952457A2 (fr) * | 1998-04-23 | 1999-10-27 | General Electric Company | Procédé et dispositif à ultrasons d'affichage des données en trois dimensions |
EP1001379A2 (fr) * | 1998-11-12 | 2000-05-17 | Mitsubishi Denki Kabushiki Kaisha | Région de plan de coupe calculée incrémentalement pour voir une portion d'un ensemble de données de volume en temps réel |
WO2000058754A1 (fr) * | 1999-03-31 | 2000-10-05 | General Electric Company | Afficheur de donnees ultrasonores en trois dimensions, dans lequel des plans de coupe multiples sont utilises |
EP1073965A1 (fr) * | 1998-04-29 | 2001-02-07 | Sensormatic Electronics Corporation | Enregistreur video numerique possedant des champs pairs et impairs |
FR2807549A1 (fr) * | 2000-04-06 | 2001-10-12 | Ge Med Sys Global Tech Co Llc | Procede de traitement d'une image et dispositif associe |
JP2001319229A (ja) * | 2000-05-10 | 2001-11-16 | Toyota Central Res & Dev Lab Inc | 画像における相関演算方法 |
WO2002041256A2 (fr) * | 2000-11-20 | 2002-05-23 | Koninklijke Philips Electronics N.V. | Dispositif d'analyse medicale |
WO2002043009A2 (fr) * | 2000-10-23 | 2002-05-30 | Siemens Corporate Research, Inc. | Procede permettant d'accelerer la production et l'affichage de vues eclatees de rendus de volume d'images en trois dimensions |
EP1857051A1 (fr) * | 2006-05-09 | 2007-11-21 | Kabushiki Kaisha Toshiba | Appareil d'imagerie à ultrasons et procédé d'affichage d'images à ultrasons |
EP2193747A1 (fr) | 2008-12-02 | 2010-06-09 | Medison Co., Ltd. | Système à ultrasons et procédé pour la fourniture d'une vue d'aide à l'orientation |
Families Citing this family (182)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5842473A (en) * | 1993-11-29 | 1998-12-01 | Life Imaging Systems | Three-dimensional imaging system |
JP3296193B2 (ja) * | 1996-06-26 | 2002-06-24 | 富士写真光機株式会社 | 超音波画像生成装置 |
US6167296A (en) | 1996-06-28 | 2000-12-26 | The Board Of Trustees Of The Leland Stanford Junior University | Method for volumetric image navigation |
US6334847B1 (en) * | 1996-11-29 | 2002-01-01 | Life Imaging Systems Inc. | Enhanced image processing for a three-dimensional imaging system |
US6360116B1 (en) | 1998-02-27 | 2002-03-19 | Varian Medical Systems, Inc. | Brachytherapy system for prostate cancer treatment with computer implemented systems and processes to facilitate pre-operative planning and post-operative evaluations |
US6327490B1 (en) | 1998-02-27 | 2001-12-04 | Varian Medical Systems, Inc. | Brachytherapy system for prostate cancer treatment with computer implemented systems and processes to facilitate pre-implantation planning and post-implantation evaluations with storage of multiple plan variations for a single patient |
US6099474A (en) * | 1998-05-27 | 2000-08-08 | Solek; Roman | Ultrasound system for displaying real time simultaneous multiplane image |
US6279399B1 (en) * | 1998-08-03 | 2001-08-28 | Vingmed Sound A/S | Multi-dimensional transducer array apparatus |
US6947584B1 (en) | 1998-08-25 | 2005-09-20 | General Electric Company | Volume imaging system |
US6574499B1 (en) * | 1998-11-25 | 2003-06-03 | Xdata Corporation | Mammography method and apparatus |
US6839762B1 (en) * | 1998-12-31 | 2005-01-04 | U-Systems, Inc. | Ultrasound information processing system and ultrasound information exchange protocol therefor |
US6174285B1 (en) * | 1999-02-02 | 2001-01-16 | Agilent Technologies, Inc. | 3-D ultrasound imaging system with pre-set, user-selectable anatomical images |
US8944070B2 (en) | 1999-04-07 | 2015-02-03 | Intuitive Surgical Operations, Inc. | Non-force reflecting method for providing tool force information to a user of a telesurgical system |
CA2267519A1 (fr) | 1999-04-13 | 2000-10-13 | Inspeck Inc. | Numeriseur optique 3d de tout le corps humain |
FR2792441B1 (fr) * | 1999-04-14 | 2002-07-26 | Iodp | Systeme d'imagerie medicale |
US7263667B1 (en) * | 1999-06-09 | 2007-08-28 | Microsoft Corporation | Methods, apparatus and data structures for providing a user interface which facilitates decision making |
US6778690B1 (en) | 1999-08-13 | 2004-08-17 | Hanif M. Ladak | Prostate boundary segmentation from 2D and 3D ultrasound images |
US6198956B1 (en) | 1999-09-30 | 2001-03-06 | Oti Ophthalmic Technologies Inc. | High speed sector scanning apparatus having digital electronic control |
US6610013B1 (en) | 1999-10-01 | 2003-08-26 | Life Imaging Systems, Inc. | 3D ultrasound-guided intraoperative prostate brachytherapy |
EP1100047A3 (fr) * | 1999-11-09 | 2002-07-24 | Konica Corporation | Méthode et appareil d'affichage de données médicales |
US6690371B1 (en) | 2000-05-03 | 2004-02-10 | Ge Medical Systems Global Technology, Llc | Relevant image data extraction from a medical image data volume |
WO2001090875A1 (fr) * | 2000-05-24 | 2001-11-29 | Koninklijke Philips Electronics N.V. | Commande de souris immediate pour mesurer les fonctionnalites d'images medicales |
IL136373A0 (en) * | 2000-05-25 | 2001-06-14 | Itpen Ltd Partnership Israel | Representation of three-dimensional bodies on computer screens and games involving such representations |
US6983063B1 (en) * | 2000-06-29 | 2006-01-03 | Siemens Corporate Research, Inc. | Computer-aided diagnosis method for aiding diagnosis of three dimensional digital image data |
US6524246B1 (en) | 2000-10-13 | 2003-02-25 | Sonocine, Inc. | Ultrasonic cellular tissue screening tool |
DE10052535A1 (de) * | 2000-10-23 | 2002-05-08 | Siemens Ag | Diagnostikeinrichtung zur Wiedergabe von Bildern |
US7556602B2 (en) * | 2000-11-24 | 2009-07-07 | U-Systems, Inc. | Breast cancer screening with adjunctive ultrasound mammography |
US7103205B2 (en) * | 2000-11-24 | 2006-09-05 | U-Systems, Inc. | Breast cancer screening with ultrasound image overlays |
US7597663B2 (en) | 2000-11-24 | 2009-10-06 | U-Systems, Inc. | Adjunctive ultrasound processing and display for breast cancer screening |
US7615008B2 (en) * | 2000-11-24 | 2009-11-10 | U-Systems, Inc. | Processing and displaying breast ultrasound information |
US7940966B2 (en) | 2000-11-24 | 2011-05-10 | U-Systems, Inc. | Full-field breast image data processing and archiving |
US6626834B2 (en) | 2001-01-25 | 2003-09-30 | Shane Dunne | Spiral scanner with electronic control |
US7262797B2 (en) * | 2001-02-22 | 2007-08-28 | Ge Inspection Technologies Lp | Method and system for storing calibration data within image files |
CA2936404C (fr) * | 2001-04-18 | 2018-06-12 | Landmark Graphics Corporation, A Halliburton Company | Unite de rendu de corps volumiques |
US20050114831A1 (en) * | 2001-04-18 | 2005-05-26 | Andres Callegari | Volume body renderer |
US8137279B2 (en) * | 2001-10-16 | 2012-03-20 | Envisioneering, Llc | Scanning probe |
US6709397B2 (en) * | 2001-10-16 | 2004-03-23 | Envisioneering, L.L.C. | Scanning probe |
EP1448098A1 (fr) * | 2001-11-20 | 2004-08-25 | Mia Norman Andersen | Procede et produit pour la mise en image du coeur |
DE10157268A1 (de) * | 2001-11-22 | 2003-06-12 | Philips Intellectual Property | Verfahren und Vorrichtung zur simultanten Darstellung von beliebig wählbaren komplementären Schnittbildern |
AU2003232881A1 (en) * | 2002-01-16 | 2003-09-09 | Mayo Foundation For Medical Education And Research | Method and apparatus for image-guided therapy |
US7477763B2 (en) * | 2002-06-18 | 2009-01-13 | Boston Scientific Scimed, Inc. | Computer generated representation of the imaging pattern of an imaging device |
KR100490564B1 (ko) * | 2002-07-23 | 2005-05-19 | 주식회사 메디슨 | 초음파 영상 신호로부터 장기를 인식하는 장치 및 방법 |
US6825838B2 (en) | 2002-10-11 | 2004-11-30 | Sonocine, Inc. | 3D modeling system |
US7424133B2 (en) * | 2002-11-08 | 2008-09-09 | Pictometry International Corporation | Method and apparatus for capturing, geolocating and measuring oblique images |
US6719700B1 (en) | 2002-12-13 | 2004-04-13 | Scimed Life Systems, Inc. | Ultrasound ranging for localization of imaging transducer |
US6975232B1 (en) | 2003-03-10 | 2005-12-13 | Mckenna Lou | Apparatus and method for “seeing” foot inside of shoe to determine the proper fit of the shoe |
US20060241450A1 (en) * | 2003-03-17 | 2006-10-26 | Biotelligent Inc. | Ultrasound guided tissue measurement system |
JP2004283490A (ja) * | 2003-03-25 | 2004-10-14 | Fuji Photo Film Co Ltd | 超音波送受信装置 |
US7314448B2 (en) * | 2003-03-28 | 2008-01-01 | Scimed Life Systems, Inc. | Imaging transducer assembly |
MXPA05011120A (es) * | 2003-04-16 | 2005-12-15 | Eastern Viriginai Medical Scho | Sistema y metodo para generar imagenes de ultrasonido independientes del operador. |
US8083678B2 (en) * | 2003-04-16 | 2011-12-27 | Eastern Virginia Medical School | System, method and medium for acquiring and generating standardized operator independent ultrasound images of fetal, neonatal and adult organs |
JP4245976B2 (ja) * | 2003-05-16 | 2009-04-02 | オリンパス株式会社 | 超音波画像処理装置 |
US6896657B2 (en) * | 2003-05-23 | 2005-05-24 | Scimed Life Systems, Inc. | Method and system for registering ultrasound image in three-dimensional coordinate system |
DE20308606U1 (de) * | 2003-05-30 | 2004-10-14 | Siemens Ag | Vorrichtung zur Halterung wenigstens eines Ultraschallkopfes |
WO2004106970A1 (fr) * | 2003-06-03 | 2004-12-09 | Koninklijke Philips Electronics, N.V. | Synchronisation d'une image ultrasonore tridimensionnelle pivotante avec un objet oscillant |
DE10339979B4 (de) * | 2003-08-29 | 2005-11-17 | Tomtec Imaging Systems Gmbh | Verfahren und Vorrichtung zur Darstellung eines vorbestimmbaren Bereichs in mehrdimensionalen Datensätzen |
EP1524011B1 (fr) | 2003-10-14 | 2011-05-25 | Nucletron B.V. | Procédé et dispositif pour le repérage relatif d'un instrument chirurgical dans un volume cible d'un corps humain ou animal |
US20060036162A1 (en) * | 2004-02-02 | 2006-02-16 | Ramin Shahidi | Method and apparatus for guiding a medical instrument to a subsurface target site in a patient |
US8235909B2 (en) * | 2004-05-12 | 2012-08-07 | Guided Therapy Systems, L.L.C. | Method and system for controlled scanning, imaging and/or therapy |
US7604595B2 (en) * | 2004-06-22 | 2009-10-20 | General Electric Company | Method and system for performing real time navigation of ultrasound volumetric data |
US20060058651A1 (en) * | 2004-08-13 | 2006-03-16 | Chiao Richard Y | Method and apparatus for extending an ultrasound image field of view |
JP4130428B2 (ja) * | 2004-09-02 | 2008-08-06 | ザイオソフト株式会社 | 画像処理方法及び画像処理プログラム |
US10864385B2 (en) | 2004-09-24 | 2020-12-15 | Guided Therapy Systems, Llc | Rejuvenating skin by heating tissue for cosmetic treatment of the face and body |
US8444562B2 (en) | 2004-10-06 | 2013-05-21 | Guided Therapy Systems, Llc | System and method for treating muscle, tendon, ligament and cartilage tissue |
US8535228B2 (en) | 2004-10-06 | 2013-09-17 | Guided Therapy Systems, Llc | Method and system for noninvasive face lifts and deep tissue tightening |
US9694212B2 (en) | 2004-10-06 | 2017-07-04 | Guided Therapy Systems, Llc | Method and system for ultrasound treatment of skin |
US9827449B2 (en) | 2004-10-06 | 2017-11-28 | Guided Therapy Systems, L.L.C. | Systems for treating skin laxity |
US11883688B2 (en) | 2004-10-06 | 2024-01-30 | Guided Therapy Systems, Llc | Energy based fat reduction |
CA2583600A1 (fr) | 2004-10-06 | 2006-04-20 | Guided Therapy Systems, L.L.C. | Methode et systeme de chirurgie esthetique non invasif |
US11235179B2 (en) | 2004-10-06 | 2022-02-01 | Guided Therapy Systems, Llc | Energy based skin gland treatment |
US8133180B2 (en) | 2004-10-06 | 2012-03-13 | Guided Therapy Systems, L.L.C. | Method and system for treating cellulite |
US20060111744A1 (en) | 2004-10-13 | 2006-05-25 | Guided Therapy Systems, L.L.C. | Method and system for treatment of sweat glands |
US8690779B2 (en) | 2004-10-06 | 2014-04-08 | Guided Therapy Systems, Llc | Noninvasive aesthetic treatment for tightening tissue |
PL2409728T3 (pl) | 2004-10-06 | 2018-01-31 | Guided Therapy Systems Llc | Układ do ultradźwiękowego leczenia tkanki |
US11207548B2 (en) | 2004-10-07 | 2021-12-28 | Guided Therapy Systems, L.L.C. | Ultrasound probe for treating skin laxity |
US11724133B2 (en) | 2004-10-07 | 2023-08-15 | Guided Therapy Systems, Llc | Ultrasound probe for treatment of skin |
US7516417B2 (en) * | 2004-11-29 | 2009-04-07 | Canon U.S.A. | Display parameter adjustment |
JP4785149B2 (ja) * | 2005-02-09 | 2011-10-05 | 株式会社日立メディコ | 超音波診断装置とその作動方法 |
DE102005007571A1 (de) * | 2005-02-18 | 2006-11-09 | Siemens Ag | Verfahren zur Visualisierung dreidimensionaler, seitens einer Datenverarbeitungseinrichtung vorliegender und/oder empfangener vektorieller Größen mit farbcodierter Richtungsinformation sowie zugehörige Vorrichtung |
CN103142251B (zh) * | 2005-04-14 | 2015-04-29 | 维拉声学公司 | 利用面向像素处理的超声成像系统 |
US9789608B2 (en) | 2006-06-29 | 2017-10-17 | Intuitive Surgical Operations, Inc. | Synthetic representation of a surgical robot |
EP2289452A3 (fr) * | 2005-06-06 | 2015-12-30 | Intuitive Surgical Operations, Inc. | Système chirurgical robotique ultrasonore laparoscopique |
US8398541B2 (en) | 2006-06-06 | 2013-03-19 | Intuitive Surgical Operations, Inc. | Interactive user interfaces for robotic minimally invasive surgical systems |
US11259870B2 (en) | 2005-06-06 | 2022-03-01 | Intuitive Surgical Operations, Inc. | Interactive user interfaces for minimally invasive telesurgical systems |
KR101320379B1 (ko) | 2005-10-20 | 2013-10-22 | 인튜어티브 서지컬 인코포레이티드 | 의료용 로봇 시스템의 컴퓨터 표시부 상에 보조 이미지의표시 및 조작 |
DE102005051783A1 (de) * | 2005-10-28 | 2007-05-03 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Verfahren und Vorrichtung zur bildgebenden Ultraschallprüfung an einem dreidimensionalen Werkstück |
US8105239B2 (en) | 2006-02-06 | 2012-01-31 | Maui Imaging, Inc. | Method and apparatus to visualize the coronary arteries using ultrasound |
EP3756537B1 (fr) † | 2006-02-22 | 2023-08-02 | DexCom, Inc. | Capteur d'analytes |
US20070259158A1 (en) * | 2006-05-05 | 2007-11-08 | General Electric Company | User interface and method for displaying information in an ultrasound system |
US8425418B2 (en) * | 2006-05-18 | 2013-04-23 | Eigen, Llc | Method of ultrasonic imaging and biopsy of the prostate |
CN104688327B (zh) | 2006-06-13 | 2017-06-09 | 直观外科手术操作公司 | 微创手术系统 |
US20080008371A1 (en) * | 2006-06-13 | 2008-01-10 | Kevin Woods | Considerations when colon segmentation differs between CAD processing and visualization |
US20090192523A1 (en) | 2006-06-29 | 2009-07-30 | Intuitive Surgical, Inc. | Synthetic representation of a surgical instrument |
US10258425B2 (en) | 2008-06-27 | 2019-04-16 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxiliary view of articulatable instruments extending out of a distal end of an entry guide |
US9718190B2 (en) | 2006-06-29 | 2017-08-01 | Intuitive Surgical Operations, Inc. | Tool position and identification indicator displayed in a boundary area of a computer display screen |
US10008017B2 (en) | 2006-06-29 | 2018-06-26 | Intuitive Surgical Operations, Inc. | Rendering tool information as graphic overlays on displayed images of tools |
US9119582B2 (en) | 2006-06-30 | 2015-09-01 | Abbott Diabetes Care, Inc. | Integrated analyte sensor and infusion device and methods therefor |
DE102006048834A1 (de) * | 2006-10-16 | 2008-04-17 | Rust, Georg-Friedemann, Dr. | Bildliche Darstellung dreidimensionaler Datensätze |
US8064664B2 (en) * | 2006-10-18 | 2011-11-22 | Eigen, Inc. | Alignment method for registering medical images |
WO2008051639A2 (fr) | 2006-10-25 | 2008-05-02 | Maui Imaging, Inc. | Procédé et appareil de production d'images ultrasonores au moyen d'une pluralité d'orifices |
US7804989B2 (en) * | 2006-10-30 | 2010-09-28 | Eigen, Inc. | Object recognition system for medical imaging |
US8834373B2 (en) * | 2006-12-01 | 2014-09-16 | Konica Minolta, Inc. | Ultrasonographic device with weighted addition of data |
US8213676B2 (en) | 2006-12-20 | 2012-07-03 | Ge Inspection Technologies Lp | Inspection apparatus method and apparatus comprising motion responsive control |
US8810636B2 (en) | 2006-12-20 | 2014-08-19 | Ge Inspection Technologies, Lp | Inspection apparatus method and apparatus comprising selective frame output |
US9633426B2 (en) | 2014-05-30 | 2017-04-25 | General Electric Company | Remote visual inspection image capture system and method |
US20080161687A1 (en) * | 2006-12-29 | 2008-07-03 | Suri Jasjit S | Repeat biopsy system |
US8175350B2 (en) * | 2007-01-15 | 2012-05-08 | Eigen, Inc. | Method for tissue culture extraction |
US20080186378A1 (en) * | 2007-02-06 | 2008-08-07 | Feimo Shen | Method and apparatus for guiding towards targets during motion |
US7856130B2 (en) * | 2007-03-28 | 2010-12-21 | Eigen, Inc. | Object recognition system for medical imaging |
US9089256B2 (en) | 2008-06-27 | 2015-07-28 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxiliary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide |
US8620473B2 (en) | 2007-06-13 | 2013-12-31 | Intuitive Surgical Operations, Inc. | Medical robotic system with coupled control modes |
US9084623B2 (en) | 2009-08-15 | 2015-07-21 | Intuitive Surgical Operations, Inc. | Controller assisted reconfiguration of an articulated instrument during movement into and out of an entry guide |
US9469034B2 (en) | 2007-06-13 | 2016-10-18 | Intuitive Surgical Operations, Inc. | Method and system for switching modes of a robotic system |
US9138129B2 (en) | 2007-06-13 | 2015-09-22 | Intuitive Surgical Operations, Inc. | Method and system for moving a plurality of articulated instruments in tandem back towards an entry guide |
US9883818B2 (en) * | 2007-06-19 | 2018-02-06 | Accuray Incorporated | Fiducial localization |
US20090003528A1 (en) | 2007-06-19 | 2009-01-01 | Sankaralingam Ramraj | Target location by tracking of imaging device |
US8864675B2 (en) | 2007-06-28 | 2014-10-21 | W. L. Gore & Associates, Inc. | Catheter |
US8852112B2 (en) | 2007-06-28 | 2014-10-07 | W. L. Gore & Associates, Inc. | Catheter with deflectable imaging device and bendable electrical conductor |
US8285362B2 (en) | 2007-06-28 | 2012-10-09 | W. L. Gore & Associates, Inc. | Catheter with deflectable imaging device |
US20090048515A1 (en) * | 2007-08-14 | 2009-02-19 | Suri Jasjit S | Biopsy planning system |
US9788813B2 (en) | 2010-10-13 | 2017-10-17 | Maui Imaging, Inc. | Multiple aperture probe internal apparatus and cable assemblies |
US9282945B2 (en) * | 2009-04-14 | 2016-03-15 | Maui Imaging, Inc. | Calibration of ultrasound probes |
US8571277B2 (en) * | 2007-10-18 | 2013-10-29 | Eigen, Llc | Image interpolation for medical imaging |
US7942829B2 (en) * | 2007-11-06 | 2011-05-17 | Eigen, Inc. | Biopsy planning and display apparatus |
US20090227874A1 (en) * | 2007-11-09 | 2009-09-10 | Eigen, Inc. | Holder assembly for a medical imaging instrument |
US20090324041A1 (en) * | 2008-01-23 | 2009-12-31 | Eigen, Llc | Apparatus for real-time 3d biopsy |
US20100001996A1 (en) * | 2008-02-28 | 2010-01-07 | Eigen, Llc | Apparatus for guiding towards targets during motion using gpu processing |
WO2009109205A1 (fr) * | 2008-03-07 | 2009-09-11 | Georg-Friedemann Rust | Représentation par images dans l'endoscopie virtuelle |
WO2009146458A2 (fr) | 2008-05-30 | 2009-12-03 | Gore Enterprise Holdings, Inc. | Sonde de cathéter ultrasonore en temps réel |
US8506490B2 (en) * | 2008-05-30 | 2013-08-13 | W.L. Gore & Associates, Inc. | Real time ultrasound probe |
KR102479936B1 (ko) | 2008-06-06 | 2022-12-22 | 얼테라, 인크 | 초음파 치료 시스템 |
EP2303131B1 (fr) * | 2008-06-26 | 2015-04-22 | Verasonics, Inc. | Imagerie d'écoulement de doppler quantitative à fréquence de trame élevée utilisant des faisceaux de transmission non focalisés |
US8864652B2 (en) | 2008-06-27 | 2014-10-21 | Intuitive Surgical Operations, Inc. | Medical robotic system providing computer generated auxiliary views of a camera instrument for controlling the positioning and orienting of its tip |
US7875039B2 (en) * | 2008-06-30 | 2011-01-25 | Eigen, Inc. | Support assembly for a tracking assembly and mounted transrectal ultrasound probe |
JP5666446B2 (ja) | 2008-08-08 | 2015-02-12 | マウイ イマギング,インコーポレーテッド | マルチアパーチャ方式の医用超音波技術を用いた画像形成方法及びアドオンシステムの同期方法 |
EP2350999A4 (fr) * | 2008-09-25 | 2017-04-05 | CAE Healthcare Canada Inc. | Simulation d imagerie médicale |
US9554770B2 (en) * | 2008-09-29 | 2017-01-31 | Siemens Medical Solutions Usa, Inc. | High pulse repetition frequency for detection of tissue mechanical property with ultrasound |
KR20110101204A (ko) | 2008-12-24 | 2011-09-15 | 가이디드 테라피 시스템스, 엘.엘.씨. | 지방 감소 및/또는 셀룰라이트 치료 방법 및 시스템 |
EP2419022B1 (fr) | 2009-04-14 | 2019-11-06 | Maui Imaging, Inc. | Accessoire d'alignement de réseau ultrasonique à ouvertures multiples |
US8918211B2 (en) | 2010-02-12 | 2014-12-23 | Intuitive Surgical Operations, Inc. | Medical robotic system providing sensory feedback indicating a difference between a commanded state and a preferred pose of an articulated instrument |
US9492927B2 (en) | 2009-08-15 | 2016-11-15 | Intuitive Surgical Operations, Inc. | Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose |
JP5683213B2 (ja) | 2009-11-17 | 2015-03-11 | キヤノン株式会社 | 画像形成装置及び画像形成方法 |
US20120014578A1 (en) * | 2010-07-19 | 2012-01-19 | Qview Medical, Inc. | Computer Aided Detection Of Abnormalities In Volumetric Breast Ultrasound Scans And User Interface |
EP2536339B1 (fr) | 2010-02-18 | 2024-05-15 | Maui Imaging, Inc. | Transmission à source ponctuelle et correction de la vitesse du son à l'aide d'une imagerie par ultrasons à ouvertures multiples |
US9668714B2 (en) | 2010-04-14 | 2017-06-06 | Maui Imaging, Inc. | Systems and methods for improving ultrasound image quality by applying weighting factors |
US9053562B1 (en) | 2010-06-24 | 2015-06-09 | Gregory S. Rabin | Two dimensional to three dimensional moving image converter |
US8758256B2 (en) | 2010-07-12 | 2014-06-24 | Best Medical International, Inc. | Apparatus for brachytherapy that uses a scanning probe for treatment of malignant tissue |
US9044216B2 (en) | 2010-07-12 | 2015-06-02 | Best Medical International, Inc. | Biopsy needle assembly |
KR101906838B1 (ko) | 2010-10-13 | 2018-10-11 | 마우이 이미징, 인코포레이티드 | 오목한 초음파 트랜스듀서들 및 3d 어레이들 |
JP5627436B2 (ja) * | 2010-12-13 | 2014-11-19 | キヤノン株式会社 | 超音波診断装置 |
EP2785253B1 (fr) | 2011-12-01 | 2023-11-15 | Maui Imaging, Inc. | Détection de mouvement par utilisation d'un doppler à ouverture basé sur l'impulsion acoustique émise et à ouvertures multiples |
JP2015503404A (ja) | 2011-12-29 | 2015-02-02 | マウイ イマギング,インコーポレーテッド | 任意経路のmモード超音波イメージング |
CN104135937B (zh) | 2012-02-21 | 2017-03-29 | 毛伊图像公司 | 使用多孔超声确定材料刚度 |
US9498182B2 (en) | 2012-05-22 | 2016-11-22 | Covidien Lp | Systems and methods for planning and navigation |
US8750568B2 (en) | 2012-05-22 | 2014-06-10 | Covidien Lp | System and method for conformal ablation planning |
US9439627B2 (en) | 2012-05-22 | 2016-09-13 | Covidien Lp | Planning system and navigation system for an ablation procedure |
US9439623B2 (en) | 2012-05-22 | 2016-09-13 | Covidien Lp | Surgical planning system and navigation system |
US9439622B2 (en) | 2012-05-22 | 2016-09-13 | Covidien Lp | Surgical navigation system |
US20130324850A1 (en) * | 2012-05-31 | 2013-12-05 | Mindray Ds Usa, Inc. | Systems and methods for interfacing with an ultrasound system |
WO2014026185A1 (fr) | 2012-08-10 | 2014-02-13 | Maui Imaging, Inc. | Étalonnage de sondes à ultrasons à ouvertures multiples |
CN104582582B (zh) | 2012-08-21 | 2017-12-15 | 毛伊图像公司 | 超声成像系统存储器架构 |
US9510802B2 (en) | 2012-09-21 | 2016-12-06 | Guided Therapy Systems, Llc | Reflective ultrasound technology for dermatological treatments |
US9075933B2 (en) * | 2012-10-11 | 2015-07-07 | Adobe Systems Incorporated | 3D transformation of objects using 2D controls projected in 3D space and contextual face selections of a three dimensional bounding box |
US9652589B2 (en) * | 2012-12-27 | 2017-05-16 | General Electric Company | Systems and methods for using a touch-sensitive display unit to analyze a medical image |
US10507066B2 (en) | 2013-02-15 | 2019-12-17 | Intuitive Surgical Operations, Inc. | Providing information of tools by filtering image areas adjacent to or on displayed images of the tools |
CN113648551A (zh) | 2013-03-08 | 2021-11-16 | 奥赛拉公司 | 用于多焦点超声治疗的装置和方法 |
US9510806B2 (en) | 2013-03-13 | 2016-12-06 | Maui Imaging, Inc. | Alignment of ultrasound transducer arrays and multiple aperture probe assembly |
US9992021B1 (en) | 2013-03-14 | 2018-06-05 | GoTenna, Inc. | System and method for private and point-to-point communication between computing devices |
WO2014210430A1 (fr) | 2013-06-27 | 2014-12-31 | Tractus Corporation | Systèmes et procédés pour la cartographie de tissu |
US9883848B2 (en) | 2013-09-13 | 2018-02-06 | Maui Imaging, Inc. | Ultrasound imaging using apparent point-source transmit transducer |
SG11201608691YA (en) | 2014-04-18 | 2016-11-29 | Ulthera Inc | Band transducer ultrasound therapy |
US10401493B2 (en) | 2014-08-18 | 2019-09-03 | Maui Imaging, Inc. | Network-based ultrasound imaging system |
CN104644215B (zh) * | 2014-12-05 | 2017-09-26 | 广州丰谱信息技术有限公司 | 一种多模三维超声探测与成像装置及方法 |
US10716544B2 (en) | 2015-10-08 | 2020-07-21 | Zmk Medical Technologies Inc. | System for 3D multi-parametric ultrasound imaging |
PT3405294T (pt) | 2016-01-18 | 2023-03-03 | Ulthera Inc | Dispositivo de ultrassom compacto com matriz anular de ultrassom ligada eletricamente de modo periférico a placa de circuito impresso flexível |
CN113729764A (zh) | 2016-01-27 | 2021-12-03 | 毛伊图像公司 | 具有稀疏阵列探测器的超声成像 |
EP3981466B9 (fr) | 2016-08-16 | 2023-10-04 | Ulthera, Inc. | Systèmes et méthodes de traitement cosmétique par ultrasons de la peau |
EP3520083A4 (fr) | 2016-09-30 | 2020-05-06 | University Hospitals Cleveland Medical Center | Appareil et procédé de construction d'un modèle 3d virtuel à partir d'une vidéo ultrasonore 2d |
TW202327520A (zh) | 2018-01-26 | 2023-07-16 | 美商奧賽拉公司 | 用於多個維度中的同時多聚焦超音治療的系統和方法 |
US11944849B2 (en) | 2018-02-20 | 2024-04-02 | Ulthera, Inc. | Systems and methods for combined cosmetic treatment of cellulite with ultrasound |
CN112566581B (zh) | 2018-08-10 | 2024-03-19 | 柯惠有限合伙公司 | 用于消融可视化的系统 |
CN116433476B (zh) * | 2023-06-09 | 2023-09-08 | 有方(合肥)医疗科技有限公司 | Ct图像处理方法及装置 |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4989142A (en) * | 1988-10-24 | 1991-01-29 | General Electric Company | Three-dimensional images obtained from tomographic slices with gantry tilt |
US5396890A (en) * | 1993-09-30 | 1995-03-14 | Siemens Medical Systems, Inc. | Three-dimensional scan converter for ultrasound imaging |
Family Cites Families (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR430506A (fr) * | 1910-05-10 | 1911-10-18 | United Shoe Machinery Co Fr | Perfectionnements dans les porte-ouvrage |
DE569790C (de) | 1930-03-01 | 1933-02-08 | Siemens Ag | Verfahren zur Geheimhaltung von elektrisch uebermittelten Nachrichten |
US4028934A (en) * | 1975-11-04 | 1977-06-14 | Yeda Research & Development Co. Ltd. | Ultrasonic stereoscopic imaging device |
US4070707A (en) * | 1976-07-12 | 1978-01-24 | General Electric Company | Reduction of offsets in data acquisition systems |
US4271706A (en) * | 1978-05-03 | 1981-06-09 | Georgetown University | Ultrasonic scanner |
US4341120A (en) * | 1979-11-09 | 1982-07-27 | Diasonics Cardio/Imaging, Inc. | Ultrasonic volume measuring system |
US4564018A (en) * | 1982-10-28 | 1986-01-14 | Storz Instrument Company | Ultrasonic system for obtaining ocular measurements |
US4562540A (en) * | 1982-11-12 | 1985-12-31 | Schlumberger Technology Corporation | Diffraction tomography system and methods |
US4747411A (en) * | 1984-03-28 | 1988-05-31 | National Biochemical Research Foundation | Three-dimensional imaging system |
US4866614A (en) * | 1985-12-26 | 1989-09-12 | General Electric Company | Ultrasound characterization of 3-dimensional flaws |
US4763661A (en) * | 1986-02-11 | 1988-08-16 | Stanford University | Filtered ultrasonic wave method and apparatus for detecting diseased tissue |
DE3610439A1 (de) * | 1986-03-27 | 1987-10-01 | Helmut Prof Dr Ing Ermert | Ultraschallgeraet zur doppler-bilddarstellung und verfahren zum betrieb desselben |
JPS6417634A (en) * | 1987-07-14 | 1989-01-20 | Yokogawa Medical Syst | Pulse doppler mti apparatus |
US4819650A (en) * | 1987-10-30 | 1989-04-11 | Wayne State University | Biplane probe including centerline highlighting |
US4932414A (en) * | 1987-11-02 | 1990-06-12 | Cornell Research Foundation, Inc. | System of therapeutic ultrasound and real-time ultrasonic scanning |
US4945478A (en) * | 1987-11-06 | 1990-07-31 | Center For Innovative Technology | Noninvasive medical imaging system and method for the identification and 3-D display of atherosclerosis and the like |
GB2212267B (en) * | 1987-11-11 | 1992-07-29 | Circulation Res Ltd | Methods and apparatus for the examination and treatment of internal organs |
US5170347A (en) * | 1987-11-27 | 1992-12-08 | Picker International, Inc. | System to reformat images for three-dimensional display using unique spatial encoding and non-planar bisectioning |
FR2624634B1 (fr) * | 1987-12-09 | 1994-04-29 | Schlumberger Etienne | Procede et dispositif permettant de reconstituer la forme et la position d'objets dans l'espace |
US4841979A (en) * | 1988-01-25 | 1989-06-27 | Capistrano Labs, Inc. | Ultrasonic prostate probe assembly |
US4858613A (en) * | 1988-03-02 | 1989-08-22 | Laboratory Equipment, Corp. | Localization and therapy system for treatment of spatially oriented focal disease |
US5036855A (en) * | 1988-03-02 | 1991-08-06 | Laboratory Equipment, Corp. | Localization and therapy system for treatment of spatially oriented focal disease |
US4955365A (en) * | 1988-03-02 | 1990-09-11 | Laboratory Equipment, Corp. | Localization and therapy system for treatment of spatially oriented focal disease |
US4934370A (en) * | 1988-09-09 | 1990-06-19 | Allergan Humphrey | Pinhole focused optics for locating visual axis of the eye for ultrasonic interior measurement |
US5159931A (en) * | 1988-11-25 | 1992-11-03 | Riccardo Pini | Apparatus for obtaining a three-dimensional reconstruction of anatomic structures through the acquisition of echographic images |
US5315512A (en) * | 1989-09-01 | 1994-05-24 | Montefiore Medical Center | Apparatus and method for generating image representations of a body utilizing an ultrasonic imaging subsystem and a three-dimensional digitizer subsystem |
US5170346A (en) * | 1989-11-22 | 1992-12-08 | General Electric Company | Method for reducing patient translation artifacts in tomographic imaging |
CA2032204C (fr) * | 1989-12-14 | 1995-03-14 | Takashi Mochizuki | Appareil d'echographie tridimensionnelle |
US5157931A (en) * | 1990-04-06 | 1992-10-27 | Alsenz Richard H | Refrigeration method and apparatus utilizing an expansion engine |
DE69020723T2 (de) * | 1990-04-20 | 1996-01-18 | Hiroshi Furuhata | Vorrichtung zur Ultrachalldiagnose. |
US5201035A (en) * | 1990-07-09 | 1993-04-06 | The United States Of America As Represented By The Secretary Of The Air Force | Dynamic algorithm selection for volume rendering, isocontour and body extraction within a multiple-instruction, multiple-data multiprocessor |
US5181514A (en) * | 1991-05-21 | 1993-01-26 | Hewlett-Packard Company | Transducer positioning system |
US5282471A (en) * | 1991-07-31 | 1994-02-01 | Kabushiki Kaisha Toshiba | Ultrasonic imaging system capable of displaying 3-dimensional angiogram in real time mode |
JP3187148B2 (ja) * | 1991-08-26 | 2001-07-11 | 株式会社東芝 | 超音波診断装置 |
US5361768A (en) * | 1992-06-30 | 1994-11-08 | Cardiovascular Imaging Systems, Inc. | Automated longitudinal position translator for ultrasonic imaging probes, and methods of using same |
CA2110148C (fr) * | 1992-12-24 | 1999-10-05 | Aaron Fenster | Systeme d'echographie tridimensionnelle |
US5454371A (en) * | 1993-11-29 | 1995-10-03 | London Health Association | Method and system for constructing and displaying three-dimensional images |
US5842473A (en) * | 1993-11-29 | 1998-12-01 | Life Imaging Systems | Three-dimensional imaging system |
-
1995
- 1995-11-24 US US08/562,590 patent/US5842473A/en not_active Expired - Lifetime
-
1996
- 1996-11-25 WO PCT/CA1996/000777 patent/WO1997020288A1/fr active IP Right Grant
- 1996-11-25 DE DE69619262T patent/DE69619262T2/de not_active Expired - Lifetime
- 1996-11-25 CA CA002238184A patent/CA2238184C/fr not_active Expired - Lifetime
- 1996-11-25 JP JP9520027A patent/JP2000500679A/ja active Pending
- 1996-11-25 EP EP96938879A patent/EP0862766B1/fr not_active Expired - Lifetime
- 1996-11-25 AT AT96938879T patent/ATE213351T1/de not_active IP Right Cessation
- 1996-11-25 AU AU76154/96A patent/AU7615496A/en not_active Abandoned
-
1998
- 1998-09-09 US US09/149,443 patent/US5964707A/en not_active Expired - Lifetime
-
1999
- 1999-08-27 US US09/384,317 patent/US6461298B1/en not_active Expired - Lifetime
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4989142A (en) * | 1988-10-24 | 1991-01-29 | General Electric Company | Three-dimensional images obtained from tomographic slices with gantry tilt |
US5396890A (en) * | 1993-09-30 | 1995-03-14 | Siemens Medical Systems, Inc. | Three-dimensional scan converter for ultrasound imaging |
Non-Patent Citations (1)
Title |
---|
COHEN-OR D ET AL: "An incremental alignment algorithm for parallel volume rendering", EUROGRAPHICS '95, MAASTRICHT, NETHERLANDS, 18 AUG.-1 SEPT. 1995, vol. 14, no. 3, ISSN 0167-7055, COMPUTER GRAPHICS FORUM, 1995, UK, pages C/123 - 33, XP000616948 * |
Cited By (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1998043109A1 (fr) * | 1997-03-21 | 1998-10-01 | Life Imaging Systems Inc. | Systeme d'imagerie tridimensionnelle |
US6342891B1 (en) * | 1997-06-25 | 2002-01-29 | Life Imaging Systems Inc. | System and method for the dynamic display of three-dimensional image data |
WO1999000675A1 (fr) * | 1997-06-25 | 1999-01-07 | Life Imaging Systems Inc. | Systeme et procede de presentation dynamique de donnees d'images en 3d |
EP0952457A2 (fr) * | 1998-04-23 | 1999-10-27 | General Electric Company | Procédé et dispositif à ultrasons d'affichage des données en trois dimensions |
EP0952457A3 (fr) * | 1998-04-23 | 2002-03-27 | General Electric Company | Procédé et dispositif à ultrasons d'affichage des données en trois dimensions |
EP1073965A4 (fr) * | 1998-04-29 | 2009-03-18 | Sensormatic Electronics Corp | Enregistreur video numerique possedant des champs pairs et impairs |
EP1073965A1 (fr) * | 1998-04-29 | 2001-02-07 | Sensormatic Electronics Corporation | Enregistreur video numerique possedant des champs pairs et impairs |
EP1001379A2 (fr) * | 1998-11-12 | 2000-05-17 | Mitsubishi Denki Kabushiki Kaisha | Région de plan de coupe calculée incrémentalement pour voir une portion d'un ensemble de données de volume en temps réel |
EP1001379A3 (fr) * | 1998-11-12 | 2002-07-10 | TeraRecon, Inc., A Delaware Corporation | Région de plan de coupe calculée incrémentalement pour voir une portion d'un ensemble de données de volume en temps réel |
WO2000058754A1 (fr) * | 1999-03-31 | 2000-10-05 | General Electric Company | Afficheur de donnees ultrasonores en trois dimensions, dans lequel des plans de coupe multiples sont utilises |
FR2807549A1 (fr) * | 2000-04-06 | 2001-10-12 | Ge Med Sys Global Tech Co Llc | Procede de traitement d'une image et dispositif associe |
JP2001319229A (ja) * | 2000-05-10 | 2001-11-16 | Toyota Central Res & Dev Lab Inc | 画像における相関演算方法 |
JP4517449B2 (ja) * | 2000-05-10 | 2010-08-04 | 株式会社豊田中央研究所 | 画像における相関演算方法 |
US6573891B1 (en) | 2000-10-23 | 2003-06-03 | Siemens Corporate Research, Inc. | Method for accelerating the generation and display of volume-rendered cut-away-views of three-dimensional images |
WO2002043009A3 (fr) * | 2000-10-23 | 2002-08-29 | Siemens Corp Res Inc | Procede permettant d'accelerer la production et l'affichage de vues eclatees de rendus de volume d'images en trois dimensions |
WO2002043009A2 (fr) * | 2000-10-23 | 2002-05-30 | Siemens Corporate Research, Inc. | Procede permettant d'accelerer la production et l'affichage de vues eclatees de rendus de volume d'images en trois dimensions |
WO2002041256A3 (fr) * | 2000-11-20 | 2002-09-12 | Koninkl Philips Electronics Nv | Dispositif d'analyse medicale |
WO2002041256A2 (fr) * | 2000-11-20 | 2002-05-23 | Koninklijke Philips Electronics N.V. | Dispositif d'analyse medicale |
EP1857051A1 (fr) * | 2006-05-09 | 2007-11-21 | Kabushiki Kaisha Toshiba | Appareil d'imagerie à ultrasons et procédé d'affichage d'images à ultrasons |
EP2193747A1 (fr) | 2008-12-02 | 2010-06-09 | Medison Co., Ltd. | Système à ultrasons et procédé pour la fourniture d'une vue d'aide à l'orientation |
Also Published As
Publication number | Publication date |
---|---|
CA2238184C (fr) | 2004-11-23 |
DE69619262T2 (de) | 2002-10-17 |
AU7615496A (en) | 1997-06-19 |
US5964707A (en) | 1999-10-12 |
US5842473A (en) | 1998-12-01 |
JP2000500679A (ja) | 2000-01-25 |
US6461298B1 (en) | 2002-10-08 |
EP0862766A1 (fr) | 1998-09-09 |
CA2238184A1 (fr) | 1997-06-05 |
DE69619262D1 (de) | 2002-03-21 |
EP0862766B1 (fr) | 2002-02-13 |
ATE213351T1 (de) | 2002-02-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5842473A (en) | Three-dimensional imaging system | |
EP0941521B1 (fr) | Traitement d'image rehaussee destine a un systeme d'imagerie tridimensionnelle | |
EP0766857B1 (fr) | Procede et systeme de reconstitution et de presentation d'une image en trois dimensions | |
WO1998024058A9 (fr) | Traitement d'image rehaussee destine a un systeme d'imagerie tridimensionnelle | |
CA2386050C (fr) | Procede et systeme de stockage d'une vue choisie d'une image tridimensionnelle | |
JP4324274B2 (ja) | 物体ボリュ―ムをイメ―ジングするシステム | |
US6049622A (en) | Graphic navigational guides for accurate image orientation and navigation | |
US10235497B2 (en) | Volumetric ultrasound image data reformatted as an image plane sequence | |
US9196092B2 (en) | Multiple volume renderings in three-dimensional medical imaging | |
US7302092B1 (en) | Three-dimensional imaging system | |
EP2162862A2 (fr) | Systèmes et procédés pour étiqueter des images de volume 3d sur un dispositif d'affichage 2d d'un système d'imagerie ultrasonore | |
US6256037B1 (en) | Method and system for selecting at least one optimal view of a three dimensional image | |
JPH1128214A (ja) | 三次元イメージング・システムおよび方法 | |
US20110055148A1 (en) | System and method for reducing ultrasound information storage requirements | |
JP3752921B2 (ja) | 超音波像の立体パノラマ画像合成装置 | |
EP0912905B1 (fr) | Systeme d'imagerie tridimensionnelle | |
CA2277723C (fr) | Procede et systeme de reconstitution et de presentation d'une image en trois dimensions | |
CA2254939C (fr) | Systeme d'imagerie tridimensionnelle | |
WO2000072273A1 (fr) | Systemes, procedes, et produits pour programmes d'ordinateur servant a la presentation de plans d'images tomographiques en trois dimensions | |
JPH07103994A (ja) | 超音波診断装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AL AM AT AU AZ BA BB BG BR BY CA CH CN CZ DE DK EE ES FI GB GE HU IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK TJ TM TR TT UA UG US UZ VN AM AZ BY KG KZ MD RU TJ TM |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): KE LS MW SD SZ UG AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI |
|
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
ENP | Entry into the national phase |
Ref document number: 2238184 Country of ref document: CA Kind code of ref document: A Ref document number: 2238184 Country of ref document: CA |
|
ENP | Entry into the national phase |
Ref document number: 1997 520027 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1996938879 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1996938879 Country of ref document: EP |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |
|
WWG | Wipo information: grant in national office |
Ref document number: 1996938879 Country of ref document: EP |