US11484279B2 - Systems to assess projection data inconsistency - Google Patents
Systems to assess projection data inconsistency Download PDFInfo
- Publication number
- US11484279B2 US11484279B2 US16/409,967 US201916409967A US11484279B2 US 11484279 B2 US11484279 B2 US 11484279B2 US 201916409967 A US201916409967 A US 201916409967A US 11484279 B2 US11484279 B2 US 11484279B2
- Authority
- US
- United States
- Prior art keywords
- projection images
- center
- measure
- local fluctuation
- projection
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5205—Devices using data or image processing specially adapted for radiation diagnosis involving processing of raw data to produce diagnostic data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/42—Arrangements for detecting radiation specially adapted for radiation diagnosis
- A61B6/4208—Arrangements for detecting radiation specially adapted for radiation diagnosis characterised by using a particular type of detector
- A61B6/4258—Arrangements for detecting radiation specially adapted for radiation diagnosis characterised by using a particular type of detector for detecting non x-ray radiation, e.g. gamma radiation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/42—Arrangements for detecting radiation specially adapted for radiation diagnosis
- A61B6/4266—Arrangements for detecting radiation specially adapted for radiation diagnosis characterised by using a plurality of detector units
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/223—Analysis of motion using block-matching
- G06T7/238—Analysis of motion using block-matching using non-full search, e.g. three-step search
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computed tomography [CT]
- A61B6/037—Emission tomography
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5258—Devices using data or image processing specially adapted for radiation diagnosis involving detection or reduction of artifacts or noise
- A61B6/5264—Devices using data or image processing specially adapted for radiation diagnosis involving detection or reduction of artifacts or noise due to motion
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/54—Control of apparatus or devices for radiation diagnosis
- A61B6/542—Control of apparatus or devices for radiation diagnosis involving control of exposure
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/049—Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10108—Single photon emission computed tomography [SPECT]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
Definitions
- detectors rotate around a patient to acquire a set of projection images based on radiation emitted from the patient.
- Three-dimensional images are reconstructed from the set of projection images using known techniques.
- the three-dimensional images may be used for diagnosis and/or treatment planning.
- Acquisition of the projection images may be performed over a timespan of a few minutes to several hours, during which the patient is likely to move. Patient motion during projection image acquisition may result in data inconsistency between the projection images. A three-dimensional image reconstructed from these projection images may exhibit poor quality and thereby hinder accurate diagnosis and treatment. If this three-dimensional image is reconstructed after the patient has been released from the scanning theater, the patient may have to return for another scan at a later time.
- a technician may visually assess data inconsistency among projection images prior to releasing the patient.
- This approach is subjective and technician-dependent, and does not provide consistent or reliably repeatable results.
- Systems are desired to efficiently quantify data inconsistency among a set of projection images, preferably prior to image reconstruction.
- FIG. 1 is a block diagram of a system to generate a quality measure based on projection images according to some embodiments
- FIG. 2 is a flow diagram of a process to evaluate the quality of projection images prior to image reconstruction according to some embodiments
- FIG. 3 a is a representation of a center-of-light locations within regions of a projection image according to some embodiments
- FIG. 3 b is a representation of a center-of-light locations within regions of a projection image according to some embodiments
- FIG. 4 includes two graphs depicting changes in center-of-light locations for a set of projection images depicting low levels of target motion according to some embodiments
- FIG. 5 includes two graphs depicting changes in center-of-light locations for a set of projection images depicting moderate levels of target motion according to some embodiments
- FIG. 6 illustrates a set of projection image images associated with respective center-of-light locations according to some embodiments
- FIG. 7 depicts determination of a local fluctuation measure according to some embodiments.
- FIG. 8 is a block diagram of a system to generate a fluctuation measure-to-quality measure mapping according to some embodiments
- FIG. 9 is a block diagram of a system to train an artificial neural network to generate a quality measure based on an input fluctuation measure according to some embodiments.
- FIG. 10 illustrates a SPECT imaging system according to some embodiments.
- Embodiments may provide technical improvements over existing systems by efficiently evaluating data inconsistency between projection images. Consequently, implementations may avoid reconstruction of images based on inconsistent data and/or signal the desirability of re-scanning a patient prior to releasing the patient from the scanning theater.
- Some embodiments provide a system to assess the data consistency of projection images, for example those acquired for purposes of reconstructing a three-dimensional image therefrom.
- the coordinates of one or more center-of-light (CL) locations are determined for each projection image. Fluctuations of the coordinates between successive images are determined and a measure of data inconsistency (e.g., represented by a measure of data quality) is determined based on the determined fluctuations.
- CL center-of-light
- a mapping is used to determine a measure of data quality based on the determined fluctuations.
- the mapping may be generated based on determinations of local fluctuations within historical sets of projection images depicting many different types of motion, and in human-generated characterizations of the quality of images reconstructed from each of the historical sets of projection images.
- Some embodiments may alternatively use the determined local fluctuations of the historical sets of projection images and corresponding human-generated characterizations to train an artificial neural network to generate a quality measure based on one or more input fluctuation measures.
- the mapping maps a first small set of data (i.e., fluctuation measures) representing a dense and complex set of data (i.e., a set of projection images) to another small set of data (i.e., the human-generated characterizations) representing the dense and complex set of data.
- a first small set of data i.e., fluctuation measures
- a dense and complex set of data i.e., a set of projection images
- the human-generated characterizations representing the dense and complex set of data.
- Embodiments may be used to analyze sets of projection images of any suitable type and acquired by any suitable means.
- a set of projection images according to some embodiments represents n-dimensional space using k (n ⁇ 1)-dimensional images, where k is the number of different projection angles from which each (n ⁇ 1)-dimensional image is acquired.
- FIG. 1 is a block diagram of system 100 according to some embodiments.
- System 100 receives a set of projection images 110 , each of which depicts a subject from a respective projection angle.
- Projection images 110 may include any number of projection images.
- embodiments are described herein with respect to projection images generated based on radioactive emissions, embodiments may be implemented in conjunction with transmission images (e.g., computed tomography, X-ray) and any other type of images including pixels associated with brightness values, regardless of imaging modality.
- transmission images e.g., computed tomography, X-ray
- Quality determination unit 120 receives projection images 110 of a patient, for example.
- Quality determination unit 120 includes center-of-light determination component 122 , local fluctuation determination component 124 and local fluctuation-to-data quality mapping component 126 .
- center-of-light determination component 122 determines the coordinates of one or more CL locations for each of projection images 110 .
- Local fluctuation determination component 124 determines one or more measures of local fluctuation coordinates between successive ones of projection images 110 based on their respective CL location coordinates.
- local fluctuation-to-data quality mapping component 126 maps the one or more measures of local fluctuation coordinates to a data quality measure.
- the data quality measure may be used to determine whether to re-scan the patient or to proceed with image reconstruction based on projection images 110 .
- Each unit, component or other functional system described herein may be implemented by one or more computing devices (e.g., computer servers), storage devices (e.g., hard or solid-state disk drives), and other hardware as is known in the art. These components may be located remote from one another and may be elements of one or more cloud computing platforms, including but not limited to a Software-as-a-Service, a Platform-as-a-Service, and an Infrastructure-as-a-Service platform.
- computing devices e.g., computer servers
- storage devices e.g., hard or solid-state disk drives
- These components may be located remote from one another and may be elements of one or more cloud computing platforms, including but not limited to a Software-as-a-Service, a Platform-as-a-Service, and an Infrastructure-as-a-Service platform.
- FIG. 2 is a flow diagram of a process according to some embodiments.
- Process 200 and the other processes described herein may be performed using any suitable combination of hardware and software.
- Software program code embodying these processes may be stored by any non-transitory tangible medium, including a fixed disk, a volatile or non-volatile random access memory, a CD, a DVD, a Flash drive, or a magnetic tape. Embodiments are not limited to the examples described below.
- a plurality of projection images are acquired at S 210 .
- Each projection image is associated with a respective projection angle.
- a SPECT detector may rotate around a patient and receive gamma ray emissions therefrom at each of a plurality of projection angles. The emissions received at a projection angle are used to generate an image associated with the projection angle.
- an X-ray source transmits X-rays through the patient toward a detector at each projection angle.
- the projection images may be acquired at S 210 by an imaging system which also performs the remaining steps of process 200 .
- the projection images were acquired by an imaging system in the past and are acquired at S 210 (e.g., via a Flash drive or network connection) by a computing system which performs the remaining steps of process 200 .
- the projection images Prior to performing the remaining steps, the projection images may be subjected to preprocessing such as image denoising. Image denoising may be particularly desirable if the count densities of the images are low.
- a CL location is determined at S 220 for each of one or more image regions.
- the CL location of an image region is determined based on weighted values of each image pixel in the region, wherein the weight of an image pixel is based on its associated brightness (or other visual characteristic) value.
- the weighted distribution of image pixels in the region should be roughly similar in all directions around the CL location in some embodiments.
- the CL location may be therefore be considered roughly analogous to the center of mass of a physical object.
- a first “global” CL location is determined for an entire projection image, four image regions are determined based on the global CL location, and a local CL location is determined for each of the four image regions.
- FIGS. 3 a and 3 b illustrate the determination of CL locations for projection image 310 and projection image 320 according to some embodiments.
- the global CL location is defined by coordinate (x 1o , y 1o ).
- image 310 is logically divided into four image regions (i.e., Top Right, Top Left, Bottom Right and Bottom Left) based on the x- and y-coordinates of the global CL location.
- a local CL location is determined within each region based on the pixel values of the region.
- the coordinates of the local CL locations of each of the image regions are (x 1TL , y 1TL ), (x 1TR , y 1TR ), (x 1BL , y 1BL ), and (x 1BR , y 1BR ).
- the global CL location of projection image 320 of FIG. 3 is defined by coordinate (x 2o , y 2o ).
- Image 320 is also logically divided into four image regions (i.e., Top Right, Top Left, Bottom Right and Bottom Left) based on the x- and y-coordinates of the global CL location.
- the sizes of the regions differ because the coordinate (x 2o , y 2o ) of the global CL location is not the physical center of image 320 .
- the coordinates of the local CL locations determined for each of the image regions are (x 2TL , y 2TL ), (x 2TR , y 2TR ), (x 2BL , y 2BL ), and (x 2BR , y 2BR ).
- Embodiments are not limited to four image regions or to image regions defined as described above.
- a projection image may be divided into regions by two or more angular, curved and/or circular boundaries.
- FIG. 4 shows plot 410 of X and Y coordinates of CL locations over a series of projection images. As expected, the CL locations change due to the changing angular perspectives of the projection images. Plot 420 shows second derivatives of the curves of plot 410 . Plot 420 shows that the rate of change of the CL location changes is fairly constant, indicating a lack of substantial subject motion.
- FIG. 5 shows plot 510 of X and Y coordinates of CL locations over a series of projection images, and plot 520 shows second derivatives of the curves of plot 510 .
- Plot 520 shows fluctuations in the second derivative and thereby suggests subject motion during imaging.
- a local fluctuation measure is determined at S 230 for each set of three adjacent projection images based on the determined center of light locations.
- the determined local fluctuation measure may comprise two or more local fluctuation measures or may be a composite of two or more local fluctuation measures according to some embodiments.
- FIG. 6 illustrates projection images 1 through n acquired at S 210 according to one example.
- Each of projection images 1 through n is associated with a different projection angle. It will be assumed that the projection images are numbered in accordance with their order of acquisition. That is, projection image 2 was acquired after projection image 1 and after projection image 3 , and is associated with a projection angle between the projection angles associated with projection images 1 and 3 .
- a global CL location and four local CL locations have been determined for each of projection images 1 through n at S 220 .
- the regions of the local CL locations may differ among projection images 1 through n due to difference in the coordinates of their associated global CL locations.
- a local fluctuation measure is determined for each region for which CL locations were determined at S 220 .
- a local fluctuation measure is determined based on coordinates (x 1o , y 1o ), (x 2o , y 2o ), . . . , (x no , y no )
- a local fluctuation measure is determined based on coordinates (x 1TL , y 1TL ), (x 2TL , y 2TL ), . . . , (x nTL , y nTL ), and so on for the coordinates of each CL location determined for regions TR, BL and BR.
- FIG. 7 illustrates determination of a local fluctuation measure for a single region according to some embodiments.
- the local fluctuation measure ⁇ k for a region is the distance of the CL location of the k th view to a line connecting the CL locations of the two neighboring views.
- the determination of measure ⁇ k utilizes three coordinates (x k ⁇ 1 , y k ⁇ 1 ), (x k , y k ) and (x k+1 , y k+1 ).
- a series of local fluctuation measures may be determined at S 230 for each region.
- a single local fluctuation measure is determined for each region by summing the series of measures determined for the region. Any other suitable systems to determine a single local fluctuation measure for each region based on a series thereof may be employed.
- the single local fluctuation measures for each region may be combined (e.g., summed) to determine a single overall local fluctuation measure.
- a quality measure is determined based on the local fluctuation measure(s) at S 240 .
- the quality measure may be determined based on a mapping between local fluctuation measure(s) and image quality measures.
- FIG. 8 illustrates system 800 to generate such a mapping according to some embodiments.
- Q sets 810 of historical projection images 1-k preferably represent different types of motion which may be encountered during subsequent use of the generated mapping as described above. Accordingly, it is desirable that each of Q sets 810 of historical projection images 1-k represents some degree of translational or rotational target motion (i.e., from no motion to extreme motion) along any one or more axes.
- unit 820 determines CL locations ( 822 ) and local fluctuation measures ( 824 ), resulting in associated local fluctuation measures 1-Q .
- Each of local fluctuation measures 1-Q may comprise more than one local fluctuation measure (e.g., one fluctuation measure for each projection image region), and is associated with a single set of historical projection images 1-k .
- Each set Q of historical projection images 1-k is also subjected to reconstruction ( 830 ) to generate an associated one of reconstructed volumes 1-Q .
- Human observer 840 inspects each of reconstructed volumes 1-Q and determines a quality measure for each reconstructed volume 1-Q .
- more than one human observer 840 inspects each reconstructed volume and/or each of reconstructed volumes 1-Q is inspected by one of any number of human observers 840 , each of whom follows a specified protocol for determining a quality measure based on a reconstructed volume.
- each set Q of historical projection images 1-k is associated with one of local fluctuation measures 1-Q and one of quality measures 1-Q .
- These associations are used to generate a mapping ( 850 ) between the two data types. Accordingly, such a mapping may be used at S 240 to determine a quality measure based on a local fluctuation measure.
- FIG. 9 illustrates system 900 according to some embodiments.
- local fluctuation measures 1-Q and associated quality measures 1-Q are generated based on Q sets of historical projection images 1-n as described above.
- Local fluctuation measures 1-Q and associated quality measures 1-Q are input to training system 910 as training data for artificial neural network 920 .
- system 910 may train system 920 such that system 920 substantially outputs one of quality measures 1-Q upon receiving its associated local fluctuation measure 1-Q as input.
- training system 910 uses CL locations and associated quality measures 1-Q as training data, eliminating a need to determine local fluctuations.
- Artificial neural network 920 may comprise a network of neurons which receive input, change internal state according to that input, and produce output depending on the input and internal state. The output of certain neurons is connected to the input of other neurons to form a directed and weighted graph. The weights as well as the functions that compute the internal state can be modified by a training process based on ground truth data. Artificial neural network 920 may comprise any one or more types of artificial neural network that are or become known, including but not limited to convolutional neural networks, recurrent neural networks, long short-term memory networks, deep reservoir computing and deep echo state networks, deep belief networks, and deep stacking networks.
- the training of network 920 generates parameter values for kernels of a fully convolutional network.
- Another fully convolutional network comprising thusly-parameterized kernels may be efficiently incorporated into a system to determine a quality measure at S 240 based on a local fluctuation measure.
- the quality measure is acceptable.
- the determined quality measure may be compared to a prespecified threshold. The threshold may differ based on the intended usage of the projection images, time constraints, and any other suitable factor. If the quality measure is not acceptable, flow returns to S 210 to acquire another set of projection images. As described above, it may be advantageous to determine the unacceptable quality and acquire the other set of images before releasing the patient.
- an image is reconstructed based on the set of projection images. Any suitable reconstructions algorithm may be employed at S 260 . The reconstructed images may then be used as intended.
- FIG. 10 illustrates SPECT system 1000 which may implement process 200 as described above.
- System 1000 includes gantry 1002 to which two or more gamma cameras 1004 a , 1004 b are attached, although any number of gamma cameras can be used.
- a detector within each gamma camera detects gamma photons (i.e., emission data) emitted by a radioisotope within the body of a patient 1006 lying on a bed 1008 .
- gamma photons i.e., emission data
- Bed 1008 is capable of moving patient 1006 along axis A and/or axis B. At respective bed positions (i.e., imaging positions), a portion of the body of patient 1006 is positioned between gamma cameras 1004 a , 1004 b in order to capture emission data from that body portion.
- Gamma cameras 1004 a , 1004 b may include multi-focal cone-beam collimators or parallel-hole collimators as is known in the art.
- Control system 1020 may comprise any general-purpose or dedicated computing system. Accordingly, control system 1020 includes one or more processing units 1022 configured to execute processor-executable program code to cause system 1020 to operate as described herein, and storage device 1030 for storing the program code. Storage device 1030 may comprise one or more fixed disks, solid-state random access memory, and/or removable media (e.g., a thumb drive) mounted in a corresponding interface (e.g., a USB port).
- processing units 1022 configured to execute processor-executable program code to cause system 1020 to operate as described herein
- storage device 1030 for storing the program code.
- Storage device 1030 may comprise one or more fixed disks, solid-state random access memory, and/or removable media (e.g., a thumb drive) mounted in a corresponding interface (e.g., a USB port).
- Storage device 1030 stores program code of system control program 1032 .
- One or more processing units 1022 may execute system control program 1032 , in conjunction with SPECT system interface 1040 , to control motors, servos, and encoders to cause gamma cameras 1004 a , 1004 b to rotate along gantry 1002 and to acquire two-dimensional emission data (i.e. projection images) at defined imaging positions during the rotation.
- the acquired data 1034 may be stored in memory 1030 .
- Control program 1032 may also be executed to reconstruct volumes 1036 from emission data 1034 as is known.
- Control program 1032 may also be executed to cause control system 1020 to perform process 200 , including determinations of CL locations and of local fluctuation measures.
- Quality measures may be determined based on fluctuation-to-quality measure mapping 1038 , which may comprise convolution kernel parameters of a trained neural network.
- Terminal 1050 may comprise a display device and an input device coupled to system 1020 .
- Terminal 1050 may display any of two-dimensional emission data 1034 , reconstructed volumes 1036 , and local fluctuation measures, and may receive user input for controlling display of the data, operation of imaging system 1000 , and/or the processing described herein.
- terminal 1050 is a separate computing device such as, but not limited to, a desktop computer, a laptop computer, a tablet computer, and a smartphone.
- Each of component of system 1000 may include other elements which are necessary for the operation thereof, as well as additional elements for providing functions other than those described herein.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Medical Informatics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- General Physics & Mathematics (AREA)
- Heart & Thoracic Surgery (AREA)
- Surgery (AREA)
- Public Health (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Pathology (AREA)
- Animal Behavior & Ethology (AREA)
- Optics & Photonics (AREA)
- High Energy & Nuclear Physics (AREA)
- Veterinary Medicine (AREA)
- Data Mining & Analysis (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- Evolutionary Computation (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Nuclear Medicine (AREA)
Abstract
Description
=[x min , x max ]×[y min , y max], where
A={I(x, y):x, y are positive integers with (x, y)∈}
TR =[x o , x max ]×[y min , y o]
TL =[x min , x o ]×[y min , y o]
BL =[x min , x o ]×[y o , y min]
BR =[x o , x min ]×[y o , y max]
(x 1 , y 1):=(x CL( RT), y CL( TR))
(x 2 , y 2):=(x CL( TL), y CL( TL))
(x 3 , y 3):=(x CL( BL), y CL( BL))
(x 4 , y 4):=(x CL( BR), y CL( BR))
Claims (20)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/409,967 US11484279B2 (en) | 2018-09-24 | 2019-05-13 | Systems to assess projection data inconsistency |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201862735431P | 2018-09-24 | 2018-09-24 | |
| US16/409,967 US11484279B2 (en) | 2018-09-24 | 2019-05-13 | Systems to assess projection data inconsistency |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20200093454A1 US20200093454A1 (en) | 2020-03-26 |
| US11484279B2 true US11484279B2 (en) | 2022-11-01 |
Family
ID=69885665
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/409,967 Active 2041-06-27 US11484279B2 (en) | 2018-09-24 | 2019-05-13 | Systems to assess projection data inconsistency |
Country Status (1)
| Country | Link |
|---|---|
| US (1) | US11484279B2 (en) |
Families Citing this family (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11494955B2 (en) | 2020-06-10 | 2022-11-08 | Siemens Medical Solutions Usa, Inc. | Data driven reconstruction in emission tomography |
| CN112581595B (en) * | 2020-12-02 | 2023-12-19 | 中国人民解放军战略支援部队航天工程大学 | A multi-view satellite image consistency analysis method |
Citations (33)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN1408102A (en) | 1999-07-26 | 2003-04-02 | 计算机化医疗系统公司 | Automated image fusion/alignment system and method |
| US20030128801A1 (en) * | 2002-01-07 | 2003-07-10 | Multi-Dimensional Imaging, Inc. | Multi-modality apparatus for dynamic anatomical, physiological and molecular imaging |
| US20070048326A1 (en) | 2001-06-20 | 2007-03-01 | Genentech, Inc. | Compositions and Methods for the Diagnosis and Treatment of Tumor |
| US20070237290A1 (en) * | 2006-04-07 | 2007-10-11 | Varian Medical Systems Technologies, Inc. | Patient setup using tomosynthesis techniques |
| CN101167662A (en) | 2006-10-25 | 2008-04-30 | 美国西门子医疗解决公司 | System and method for lymph node imaging by cooperation of CT and MR |
| CN101408976A (en) | 2007-10-11 | 2009-04-15 | 通用电气公司 | Enhanced system and method for volume based registration |
| US20100232645A1 (en) | 2007-05-10 | 2010-09-16 | Koninklijke Philips Electronics N.V. | Model-based spect heart orientation estimation |
| US20100289813A1 (en) | 2009-05-13 | 2010-11-18 | Kabushiki Kaisha Toshiba | Nuclear medical imaging apparatus, image processing apparatus, and image processing method |
| US20110142314A1 (en) | 2009-12-15 | 2011-06-16 | Jiang Hsieh | System and method of increasing temporal resolution of an x-ray image |
| US20110228897A1 (en) | 2008-03-07 | 2011-09-22 | Aloka Co., Ltd. | X-ray ct scanner and control program thereof |
| US20120169333A1 (en) | 2009-09-08 | 2012-07-05 | Koninklijke Philips Electronics N.V. | Motion monitoring system for monitoring motion within a region of interest |
| CN102846326A (en) | 2011-06-03 | 2013-01-02 | 通用电气公司 | Method and system for processing gated image data |
| US20130034286A1 (en) | 2011-08-03 | 2013-02-07 | Siemens Medical Solutions Usa, Inc. | Registration of Emission Tomography and Computed Tomography |
| CN102982510A (en) | 2011-06-21 | 2013-03-20 | 通用电气公司 | Motion correction of SPECT images |
| CN103136734A (en) | 2013-02-27 | 2013-06-05 | 北京工业大学 | Restraining method on edge Halo effects during process of resetting projections onto convex sets (POCS) super-resolution image |
| US20130315459A1 (en) | 2012-05-22 | 2013-11-28 | General Electric Company | Method and apparatus for motion correcting medical images |
| US20140072194A1 (en) | 2011-05-12 | 2014-03-13 | Koninklijke Philips N.V. | Motion compensated imaging |
| CN104169969A (en) | 2012-03-15 | 2014-11-26 | 皇家飞利浦有限公司 | Multi-modality deformable registration |
| US20140355855A1 (en) | 2013-05-30 | 2014-12-04 | Siemens Aktiengesellschaft | System and Method for Magnetic Resonance Imaging Based Respiratory Motion Correction for PET/MRI |
| US20150302616A1 (en) | 2012-12-06 | 2015-10-22 | Koninklijke Philips N.V. | Local artifact reduction with insignificant side effects |
| US20160095565A1 (en) | 2014-10-01 | 2016-04-07 | Siemens Aktiengesellschaft | Method and imaging system for compensating for location assignment errors in pet data occurring due to a cyclical motion of a patient |
| US20160125605A1 (en) | 2014-10-29 | 2016-05-05 | Samsung Electronics Co., Ltd. | Image processing apparatus, image processing method, and ultrasound imaging apparatus having image processing unit |
| US20160203609A1 (en) | 2013-09-27 | 2016-07-14 | Fujifilm Corporation | Image alignment device, method, and program, and method for generating 3-d deformation model |
| US20160220311A1 (en) | 2015-01-29 | 2016-08-04 | Siemens Healthcare Gmbh | Planning and guidance of electrophysiology therapies |
| US20170193159A1 (en) | 2014-06-13 | 2017-07-06 | Siemens Healthcare Gmbh | One Gate Reconstruction |
| US20170209112A1 (en) | 2016-01-25 | 2017-07-27 | Samsung Electronics Co., Ltd | Tomography apparatus and method of reconstructing tomography image |
| US20170219674A1 (en) | 2014-04-18 | 2017-08-03 | The General Hospital Corporation | System and method for magnetic resonance imaging with prospective motion control |
| US20170278280A1 (en) | 2014-07-30 | 2017-09-28 | Navix International Limited | Registering nuclear medicine data |
| US20170360325A1 (en) | 2014-12-11 | 2017-12-21 | Elekta, Inc. | Motion management in mri-guided linac |
| US20180025512A1 (en) | 2016-07-20 | 2018-01-25 | Shanghai United Imaging Healthcare Co., Ltd. | System and method for segmenting medical image |
| US20180033166A1 (en) | 2014-06-13 | 2018-02-01 | Siemens Healthcare Gmbh | Intra Reconstruction Motion Correction |
| US20180056091A1 (en) | 2016-08-29 | 2018-03-01 | Accuray Incorporated | Offline angle selection in rotational imaging and tracking systems |
| US20180336677A1 (en) | 2017-05-18 | 2018-11-22 | Toshiba Medical Systems Corporation | Apparatus and method for medical image processing |
-
2019
- 2019-05-13 US US16/409,967 patent/US11484279B2/en active Active
Patent Citations (34)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN1408102A (en) | 1999-07-26 | 2003-04-02 | 计算机化医疗系统公司 | Automated image fusion/alignment system and method |
| US20070048326A1 (en) | 2001-06-20 | 2007-03-01 | Genentech, Inc. | Compositions and Methods for the Diagnosis and Treatment of Tumor |
| US20030128801A1 (en) * | 2002-01-07 | 2003-07-10 | Multi-Dimensional Imaging, Inc. | Multi-modality apparatus for dynamic anatomical, physiological and molecular imaging |
| US20070237290A1 (en) * | 2006-04-07 | 2007-10-11 | Varian Medical Systems Technologies, Inc. | Patient setup using tomosynthesis techniques |
| CN101167662A (en) | 2006-10-25 | 2008-04-30 | 美国西门子医疗解决公司 | System and method for lymph node imaging by cooperation of CT and MR |
| US20100232645A1 (en) | 2007-05-10 | 2010-09-16 | Koninklijke Philips Electronics N.V. | Model-based spect heart orientation estimation |
| CN101408976A (en) | 2007-10-11 | 2009-04-15 | 通用电气公司 | Enhanced system and method for volume based registration |
| US20110228897A1 (en) | 2008-03-07 | 2011-09-22 | Aloka Co., Ltd. | X-ray ct scanner and control program thereof |
| US20100289813A1 (en) | 2009-05-13 | 2010-11-18 | Kabushiki Kaisha Toshiba | Nuclear medical imaging apparatus, image processing apparatus, and image processing method |
| US20120169333A1 (en) | 2009-09-08 | 2012-07-05 | Koninklijke Philips Electronics N.V. | Motion monitoring system for monitoring motion within a region of interest |
| US20110142314A1 (en) | 2009-12-15 | 2011-06-16 | Jiang Hsieh | System and method of increasing temporal resolution of an x-ray image |
| US20140072194A1 (en) | 2011-05-12 | 2014-03-13 | Koninklijke Philips N.V. | Motion compensated imaging |
| CN102846326A (en) | 2011-06-03 | 2013-01-02 | 通用电气公司 | Method and system for processing gated image data |
| CN102982510A (en) | 2011-06-21 | 2013-03-20 | 通用电气公司 | Motion correction of SPECT images |
| US20130034286A1 (en) | 2011-08-03 | 2013-02-07 | Siemens Medical Solutions Usa, Inc. | Registration of Emission Tomography and Computed Tomography |
| CN104169969A (en) | 2012-03-15 | 2014-11-26 | 皇家飞利浦有限公司 | Multi-modality deformable registration |
| CN103417234A (en) | 2012-05-22 | 2013-12-04 | 通用电气公司 | Method and apparatus for motion correcting medical images |
| US20130315459A1 (en) | 2012-05-22 | 2013-11-28 | General Electric Company | Method and apparatus for motion correcting medical images |
| US20150302616A1 (en) | 2012-12-06 | 2015-10-22 | Koninklijke Philips N.V. | Local artifact reduction with insignificant side effects |
| CN103136734A (en) | 2013-02-27 | 2013-06-05 | 北京工业大学 | Restraining method on edge Halo effects during process of resetting projections onto convex sets (POCS) super-resolution image |
| US20140355855A1 (en) | 2013-05-30 | 2014-12-04 | Siemens Aktiengesellschaft | System and Method for Magnetic Resonance Imaging Based Respiratory Motion Correction for PET/MRI |
| US20160203609A1 (en) | 2013-09-27 | 2016-07-14 | Fujifilm Corporation | Image alignment device, method, and program, and method for generating 3-d deformation model |
| US20170219674A1 (en) | 2014-04-18 | 2017-08-03 | The General Hospital Corporation | System and method for magnetic resonance imaging with prospective motion control |
| US20180033166A1 (en) | 2014-06-13 | 2018-02-01 | Siemens Healthcare Gmbh | Intra Reconstruction Motion Correction |
| US20170193159A1 (en) | 2014-06-13 | 2017-07-06 | Siemens Healthcare Gmbh | One Gate Reconstruction |
| US20170278280A1 (en) | 2014-07-30 | 2017-09-28 | Navix International Limited | Registering nuclear medicine data |
| US20160095565A1 (en) | 2014-10-01 | 2016-04-07 | Siemens Aktiengesellschaft | Method and imaging system for compensating for location assignment errors in pet data occurring due to a cyclical motion of a patient |
| US20160125605A1 (en) | 2014-10-29 | 2016-05-05 | Samsung Electronics Co., Ltd. | Image processing apparatus, image processing method, and ultrasound imaging apparatus having image processing unit |
| US20170360325A1 (en) | 2014-12-11 | 2017-12-21 | Elekta, Inc. | Motion management in mri-guided linac |
| US20160220311A1 (en) | 2015-01-29 | 2016-08-04 | Siemens Healthcare Gmbh | Planning and guidance of electrophysiology therapies |
| US20170209112A1 (en) | 2016-01-25 | 2017-07-27 | Samsung Electronics Co., Ltd | Tomography apparatus and method of reconstructing tomography image |
| US20180025512A1 (en) | 2016-07-20 | 2018-01-25 | Shanghai United Imaging Healthcare Co., Ltd. | System and method for segmenting medical image |
| US20180056091A1 (en) | 2016-08-29 | 2018-03-01 | Accuray Incorporated | Offline angle selection in rotational imaging and tracking systems |
| US20180336677A1 (en) | 2017-05-18 | 2018-11-22 | Toshiba Medical Systems Corporation | Apparatus and method for medical image processing |
Also Published As
| Publication number | Publication date |
|---|---|
| US20200093454A1 (en) | 2020-03-26 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10593071B2 (en) | Network training and architecture for medical imaging | |
| US11769277B2 (en) | Deep learning based scatter correction | |
| US11164344B2 (en) | PET image reconstruction using TOF data and neural network | |
| US20200311490A1 (en) | Apparatus and method for sinogram restoration in computed tomography (ct) using adaptive filtering with deep learning (dl) | |
| US9848844B2 (en) | Iterative reconstruction process | |
| US10475538B2 (en) | System for predicting computed tomography data from surface data | |
| US11874411B2 (en) | Estimation of partially missing attenuation in time-of-flight positron emission tomography | |
| US11484279B2 (en) | Systems to assess projection data inconsistency | |
| US8755586B2 (en) | Image reconstruction including shift-variant blur compensation | |
| US11334987B2 (en) | Physics-constrained network and training thereof | |
| CN113643394B (en) | Scattering correction method, device, computer equipment and storage medium | |
| US11810228B2 (en) | Network determination of limited-angle reconstruction | |
| EP4062836A1 (en) | Neural network-assisted dose assessment based on activity image | |
| US11335040B2 (en) | Multi-focal non-parallel collimator-based imaging | |
| EP4148680A1 (en) | Attenuation correction-based weighting for tomographic inconsistency detection | |
| US11574184B2 (en) | Multi-modal reconstruction network | |
| US10687766B2 (en) | System to detect features using multiple reconstructions | |
| US20250259349A1 (en) | Ai-driven motion correction of pet data | |
| CN120635243B (en) | Spiral CT image reconstruction method based on nerve radiation field | |
| US20260000377A1 (en) | Imaging with scatter correction with the aid of noisy scatter estimations and scatter interpolation | |
| US20210312675A1 (en) | Evaluation of Image Formation System Components | |
| CN119908737A (en) | Provides adapted recording parameters | |
| Tirunelveli | Comparison of square and hexagonal pixels for image processing and alternative solutions to linear equations for Computed Tomography |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC., PENNSYLVANIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:VIJA, ALEXANDER HANS;DING, XINHONG;SIGNING DATES FROM 20190522 TO 20190528;REEL/FRAME:049315/0958 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |