US11484279B2 - Systems to assess projection data inconsistency - Google Patents

Systems to assess projection data inconsistency Download PDF

Info

Publication number
US11484279B2
US11484279B2 US16/409,967 US201916409967A US11484279B2 US 11484279 B2 US11484279 B2 US 11484279B2 US 201916409967 A US201916409967 A US 201916409967A US 11484279 B2 US11484279 B2 US 11484279B2
Authority
US
United States
Prior art keywords
projection images
center
measure
local fluctuation
projection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US16/409,967
Other versions
US20200093454A1 (en
Inventor
Alexander Hans Vija
Xinhong Ding
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Siemens Medical Solutions USA Inc
Original Assignee
Siemens Medical Solutions USA Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Siemens Medical Solutions USA Inc filed Critical Siemens Medical Solutions USA Inc
Priority to US16/409,967 priority Critical patent/US11484279B2/en
Assigned to SIEMENS MEDICAL SOLUTIONS USA, INC. reassignment SIEMENS MEDICAL SOLUTIONS USA, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: DING, XINHONG, VIJA, ALEXANDER HANS
Publication of US20200093454A1 publication Critical patent/US20200093454A1/en
Application granted granted Critical
Publication of US11484279B2 publication Critical patent/US11484279B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • A61B6/5205Devices using data or image processing specially adapted for radiation diagnosis involving processing of raw data to produce diagnostic data
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/42Arrangements for detecting radiation specially adapted for radiation diagnosis
    • A61B6/4208Arrangements for detecting radiation specially adapted for radiation diagnosis characterised by using a particular type of detector
    • A61B6/4258Arrangements for detecting radiation specially adapted for radiation diagnosis characterised by using a particular type of detector for detecting non x-ray radiation, e.g. gamma radiation
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/42Arrangements for detecting radiation specially adapted for radiation diagnosis
    • A61B6/4266Arrangements for detecting radiation specially adapted for radiation diagnosis characterised by using a plurality of detector units
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/09Supervised learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/223Analysis of motion using block-matching
    • G06T7/238Analysis of motion using block-matching using non-full search, e.g. three-step search
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/02Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
    • A61B6/03Computed tomography [CT]
    • A61B6/037Emission tomography
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • A61B6/5258Devices using data or image processing specially adapted for radiation diagnosis involving detection or reduction of artifacts or noise
    • A61B6/5264Devices using data or image processing specially adapted for radiation diagnosis involving detection or reduction of artifacts or noise due to motion
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/54Control of apparatus or devices for radiation diagnosis
    • A61B6/542Control of apparatus or devices for radiation diagnosis involving control of exposure
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10108Single photon emission computed tomography [SPECT]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning

Definitions

  • detectors rotate around a patient to acquire a set of projection images based on radiation emitted from the patient.
  • Three-dimensional images are reconstructed from the set of projection images using known techniques.
  • the three-dimensional images may be used for diagnosis and/or treatment planning.
  • Acquisition of the projection images may be performed over a timespan of a few minutes to several hours, during which the patient is likely to move. Patient motion during projection image acquisition may result in data inconsistency between the projection images. A three-dimensional image reconstructed from these projection images may exhibit poor quality and thereby hinder accurate diagnosis and treatment. If this three-dimensional image is reconstructed after the patient has been released from the scanning theater, the patient may have to return for another scan at a later time.
  • a technician may visually assess data inconsistency among projection images prior to releasing the patient.
  • This approach is subjective and technician-dependent, and does not provide consistent or reliably repeatable results.
  • Systems are desired to efficiently quantify data inconsistency among a set of projection images, preferably prior to image reconstruction.
  • FIG. 1 is a block diagram of a system to generate a quality measure based on projection images according to some embodiments
  • FIG. 2 is a flow diagram of a process to evaluate the quality of projection images prior to image reconstruction according to some embodiments
  • FIG. 3 a is a representation of a center-of-light locations within regions of a projection image according to some embodiments
  • FIG. 3 b is a representation of a center-of-light locations within regions of a projection image according to some embodiments
  • FIG. 4 includes two graphs depicting changes in center-of-light locations for a set of projection images depicting low levels of target motion according to some embodiments
  • FIG. 5 includes two graphs depicting changes in center-of-light locations for a set of projection images depicting moderate levels of target motion according to some embodiments
  • FIG. 6 illustrates a set of projection image images associated with respective center-of-light locations according to some embodiments
  • FIG. 7 depicts determination of a local fluctuation measure according to some embodiments.
  • FIG. 8 is a block diagram of a system to generate a fluctuation measure-to-quality measure mapping according to some embodiments
  • FIG. 9 is a block diagram of a system to train an artificial neural network to generate a quality measure based on an input fluctuation measure according to some embodiments.
  • FIG. 10 illustrates a SPECT imaging system according to some embodiments.
  • Embodiments may provide technical improvements over existing systems by efficiently evaluating data inconsistency between projection images. Consequently, implementations may avoid reconstruction of images based on inconsistent data and/or signal the desirability of re-scanning a patient prior to releasing the patient from the scanning theater.
  • Some embodiments provide a system to assess the data consistency of projection images, for example those acquired for purposes of reconstructing a three-dimensional image therefrom.
  • the coordinates of one or more center-of-light (CL) locations are determined for each projection image. Fluctuations of the coordinates between successive images are determined and a measure of data inconsistency (e.g., represented by a measure of data quality) is determined based on the determined fluctuations.
  • CL center-of-light
  • a mapping is used to determine a measure of data quality based on the determined fluctuations.
  • the mapping may be generated based on determinations of local fluctuations within historical sets of projection images depicting many different types of motion, and in human-generated characterizations of the quality of images reconstructed from each of the historical sets of projection images.
  • Some embodiments may alternatively use the determined local fluctuations of the historical sets of projection images and corresponding human-generated characterizations to train an artificial neural network to generate a quality measure based on one or more input fluctuation measures.
  • the mapping maps a first small set of data (i.e., fluctuation measures) representing a dense and complex set of data (i.e., a set of projection images) to another small set of data (i.e., the human-generated characterizations) representing the dense and complex set of data.
  • a first small set of data i.e., fluctuation measures
  • a dense and complex set of data i.e., a set of projection images
  • the human-generated characterizations representing the dense and complex set of data.
  • Embodiments may be used to analyze sets of projection images of any suitable type and acquired by any suitable means.
  • a set of projection images according to some embodiments represents n-dimensional space using k (n ⁇ 1)-dimensional images, where k is the number of different projection angles from which each (n ⁇ 1)-dimensional image is acquired.
  • FIG. 1 is a block diagram of system 100 according to some embodiments.
  • System 100 receives a set of projection images 110 , each of which depicts a subject from a respective projection angle.
  • Projection images 110 may include any number of projection images.
  • embodiments are described herein with respect to projection images generated based on radioactive emissions, embodiments may be implemented in conjunction with transmission images (e.g., computed tomography, X-ray) and any other type of images including pixels associated with brightness values, regardless of imaging modality.
  • transmission images e.g., computed tomography, X-ray
  • Quality determination unit 120 receives projection images 110 of a patient, for example.
  • Quality determination unit 120 includes center-of-light determination component 122 , local fluctuation determination component 124 and local fluctuation-to-data quality mapping component 126 .
  • center-of-light determination component 122 determines the coordinates of one or more CL locations for each of projection images 110 .
  • Local fluctuation determination component 124 determines one or more measures of local fluctuation coordinates between successive ones of projection images 110 based on their respective CL location coordinates.
  • local fluctuation-to-data quality mapping component 126 maps the one or more measures of local fluctuation coordinates to a data quality measure.
  • the data quality measure may be used to determine whether to re-scan the patient or to proceed with image reconstruction based on projection images 110 .
  • Each unit, component or other functional system described herein may be implemented by one or more computing devices (e.g., computer servers), storage devices (e.g., hard or solid-state disk drives), and other hardware as is known in the art. These components may be located remote from one another and may be elements of one or more cloud computing platforms, including but not limited to a Software-as-a-Service, a Platform-as-a-Service, and an Infrastructure-as-a-Service platform.
  • computing devices e.g., computer servers
  • storage devices e.g., hard or solid-state disk drives
  • These components may be located remote from one another and may be elements of one or more cloud computing platforms, including but not limited to a Software-as-a-Service, a Platform-as-a-Service, and an Infrastructure-as-a-Service platform.
  • FIG. 2 is a flow diagram of a process according to some embodiments.
  • Process 200 and the other processes described herein may be performed using any suitable combination of hardware and software.
  • Software program code embodying these processes may be stored by any non-transitory tangible medium, including a fixed disk, a volatile or non-volatile random access memory, a CD, a DVD, a Flash drive, or a magnetic tape. Embodiments are not limited to the examples described below.
  • a plurality of projection images are acquired at S 210 .
  • Each projection image is associated with a respective projection angle.
  • a SPECT detector may rotate around a patient and receive gamma ray emissions therefrom at each of a plurality of projection angles. The emissions received at a projection angle are used to generate an image associated with the projection angle.
  • an X-ray source transmits X-rays through the patient toward a detector at each projection angle.
  • the projection images may be acquired at S 210 by an imaging system which also performs the remaining steps of process 200 .
  • the projection images were acquired by an imaging system in the past and are acquired at S 210 (e.g., via a Flash drive or network connection) by a computing system which performs the remaining steps of process 200 .
  • the projection images Prior to performing the remaining steps, the projection images may be subjected to preprocessing such as image denoising. Image denoising may be particularly desirable if the count densities of the images are low.
  • a CL location is determined at S 220 for each of one or more image regions.
  • the CL location of an image region is determined based on weighted values of each image pixel in the region, wherein the weight of an image pixel is based on its associated brightness (or other visual characteristic) value.
  • the weighted distribution of image pixels in the region should be roughly similar in all directions around the CL location in some embodiments.
  • the CL location may be therefore be considered roughly analogous to the center of mass of a physical object.
  • a first “global” CL location is determined for an entire projection image, four image regions are determined based on the global CL location, and a local CL location is determined for each of the four image regions.
  • FIGS. 3 a and 3 b illustrate the determination of CL locations for projection image 310 and projection image 320 according to some embodiments.
  • the global CL location is defined by coordinate (x 1o , y 1o ).
  • image 310 is logically divided into four image regions (i.e., Top Right, Top Left, Bottom Right and Bottom Left) based on the x- and y-coordinates of the global CL location.
  • a local CL location is determined within each region based on the pixel values of the region.
  • the coordinates of the local CL locations of each of the image regions are (x 1TL , y 1TL ), (x 1TR , y 1TR ), (x 1BL , y 1BL ), and (x 1BR , y 1BR ).
  • the global CL location of projection image 320 of FIG. 3 is defined by coordinate (x 2o , y 2o ).
  • Image 320 is also logically divided into four image regions (i.e., Top Right, Top Left, Bottom Right and Bottom Left) based on the x- and y-coordinates of the global CL location.
  • the sizes of the regions differ because the coordinate (x 2o , y 2o ) of the global CL location is not the physical center of image 320 .
  • the coordinates of the local CL locations determined for each of the image regions are (x 2TL , y 2TL ), (x 2TR , y 2TR ), (x 2BL , y 2BL ), and (x 2BR , y 2BR ).
  • Embodiments are not limited to four image regions or to image regions defined as described above.
  • a projection image may be divided into regions by two or more angular, curved and/or circular boundaries.
  • FIG. 4 shows plot 410 of X and Y coordinates of CL locations over a series of projection images. As expected, the CL locations change due to the changing angular perspectives of the projection images. Plot 420 shows second derivatives of the curves of plot 410 . Plot 420 shows that the rate of change of the CL location changes is fairly constant, indicating a lack of substantial subject motion.
  • FIG. 5 shows plot 510 of X and Y coordinates of CL locations over a series of projection images, and plot 520 shows second derivatives of the curves of plot 510 .
  • Plot 520 shows fluctuations in the second derivative and thereby suggests subject motion during imaging.
  • a local fluctuation measure is determined at S 230 for each set of three adjacent projection images based on the determined center of light locations.
  • the determined local fluctuation measure may comprise two or more local fluctuation measures or may be a composite of two or more local fluctuation measures according to some embodiments.
  • FIG. 6 illustrates projection images 1 through n acquired at S 210 according to one example.
  • Each of projection images 1 through n is associated with a different projection angle. It will be assumed that the projection images are numbered in accordance with their order of acquisition. That is, projection image 2 was acquired after projection image 1 and after projection image 3 , and is associated with a projection angle between the projection angles associated with projection images 1 and 3 .
  • a global CL location and four local CL locations have been determined for each of projection images 1 through n at S 220 .
  • the regions of the local CL locations may differ among projection images 1 through n due to difference in the coordinates of their associated global CL locations.
  • a local fluctuation measure is determined for each region for which CL locations were determined at S 220 .
  • a local fluctuation measure is determined based on coordinates (x 1o , y 1o ), (x 2o , y 2o ), . . . , (x no , y no )
  • a local fluctuation measure is determined based on coordinates (x 1TL , y 1TL ), (x 2TL , y 2TL ), . . . , (x nTL , y nTL ), and so on for the coordinates of each CL location determined for regions TR, BL and BR.
  • FIG. 7 illustrates determination of a local fluctuation measure for a single region according to some embodiments.
  • the local fluctuation measure ⁇ k for a region is the distance of the CL location of the k th view to a line connecting the CL locations of the two neighboring views.
  • the determination of measure ⁇ k utilizes three coordinates (x k ⁇ 1 , y k ⁇ 1 ), (x k , y k ) and (x k+1 , y k+1 ).
  • a series of local fluctuation measures may be determined at S 230 for each region.
  • a single local fluctuation measure is determined for each region by summing the series of measures determined for the region. Any other suitable systems to determine a single local fluctuation measure for each region based on a series thereof may be employed.
  • the single local fluctuation measures for each region may be combined (e.g., summed) to determine a single overall local fluctuation measure.
  • a quality measure is determined based on the local fluctuation measure(s) at S 240 .
  • the quality measure may be determined based on a mapping between local fluctuation measure(s) and image quality measures.
  • FIG. 8 illustrates system 800 to generate such a mapping according to some embodiments.
  • Q sets 810 of historical projection images 1-k preferably represent different types of motion which may be encountered during subsequent use of the generated mapping as described above. Accordingly, it is desirable that each of Q sets 810 of historical projection images 1-k represents some degree of translational or rotational target motion (i.e., from no motion to extreme motion) along any one or more axes.
  • unit 820 determines CL locations ( 822 ) and local fluctuation measures ( 824 ), resulting in associated local fluctuation measures 1-Q .
  • Each of local fluctuation measures 1-Q may comprise more than one local fluctuation measure (e.g., one fluctuation measure for each projection image region), and is associated with a single set of historical projection images 1-k .
  • Each set Q of historical projection images 1-k is also subjected to reconstruction ( 830 ) to generate an associated one of reconstructed volumes 1-Q .
  • Human observer 840 inspects each of reconstructed volumes 1-Q and determines a quality measure for each reconstructed volume 1-Q .
  • more than one human observer 840 inspects each reconstructed volume and/or each of reconstructed volumes 1-Q is inspected by one of any number of human observers 840 , each of whom follows a specified protocol for determining a quality measure based on a reconstructed volume.
  • each set Q of historical projection images 1-k is associated with one of local fluctuation measures 1-Q and one of quality measures 1-Q .
  • These associations are used to generate a mapping ( 850 ) between the two data types. Accordingly, such a mapping may be used at S 240 to determine a quality measure based on a local fluctuation measure.
  • FIG. 9 illustrates system 900 according to some embodiments.
  • local fluctuation measures 1-Q and associated quality measures 1-Q are generated based on Q sets of historical projection images 1-n as described above.
  • Local fluctuation measures 1-Q and associated quality measures 1-Q are input to training system 910 as training data for artificial neural network 920 .
  • system 910 may train system 920 such that system 920 substantially outputs one of quality measures 1-Q upon receiving its associated local fluctuation measure 1-Q as input.
  • training system 910 uses CL locations and associated quality measures 1-Q as training data, eliminating a need to determine local fluctuations.
  • Artificial neural network 920 may comprise a network of neurons which receive input, change internal state according to that input, and produce output depending on the input and internal state. The output of certain neurons is connected to the input of other neurons to form a directed and weighted graph. The weights as well as the functions that compute the internal state can be modified by a training process based on ground truth data. Artificial neural network 920 may comprise any one or more types of artificial neural network that are or become known, including but not limited to convolutional neural networks, recurrent neural networks, long short-term memory networks, deep reservoir computing and deep echo state networks, deep belief networks, and deep stacking networks.
  • the training of network 920 generates parameter values for kernels of a fully convolutional network.
  • Another fully convolutional network comprising thusly-parameterized kernels may be efficiently incorporated into a system to determine a quality measure at S 240 based on a local fluctuation measure.
  • the quality measure is acceptable.
  • the determined quality measure may be compared to a prespecified threshold. The threshold may differ based on the intended usage of the projection images, time constraints, and any other suitable factor. If the quality measure is not acceptable, flow returns to S 210 to acquire another set of projection images. As described above, it may be advantageous to determine the unacceptable quality and acquire the other set of images before releasing the patient.
  • an image is reconstructed based on the set of projection images. Any suitable reconstructions algorithm may be employed at S 260 . The reconstructed images may then be used as intended.
  • FIG. 10 illustrates SPECT system 1000 which may implement process 200 as described above.
  • System 1000 includes gantry 1002 to which two or more gamma cameras 1004 a , 1004 b are attached, although any number of gamma cameras can be used.
  • a detector within each gamma camera detects gamma photons (i.e., emission data) emitted by a radioisotope within the body of a patient 1006 lying on a bed 1008 .
  • gamma photons i.e., emission data
  • Bed 1008 is capable of moving patient 1006 along axis A and/or axis B. At respective bed positions (i.e., imaging positions), a portion of the body of patient 1006 is positioned between gamma cameras 1004 a , 1004 b in order to capture emission data from that body portion.
  • Gamma cameras 1004 a , 1004 b may include multi-focal cone-beam collimators or parallel-hole collimators as is known in the art.
  • Control system 1020 may comprise any general-purpose or dedicated computing system. Accordingly, control system 1020 includes one or more processing units 1022 configured to execute processor-executable program code to cause system 1020 to operate as described herein, and storage device 1030 for storing the program code. Storage device 1030 may comprise one or more fixed disks, solid-state random access memory, and/or removable media (e.g., a thumb drive) mounted in a corresponding interface (e.g., a USB port).
  • processing units 1022 configured to execute processor-executable program code to cause system 1020 to operate as described herein
  • storage device 1030 for storing the program code.
  • Storage device 1030 may comprise one or more fixed disks, solid-state random access memory, and/or removable media (e.g., a thumb drive) mounted in a corresponding interface (e.g., a USB port).
  • Storage device 1030 stores program code of system control program 1032 .
  • One or more processing units 1022 may execute system control program 1032 , in conjunction with SPECT system interface 1040 , to control motors, servos, and encoders to cause gamma cameras 1004 a , 1004 b to rotate along gantry 1002 and to acquire two-dimensional emission data (i.e. projection images) at defined imaging positions during the rotation.
  • the acquired data 1034 may be stored in memory 1030 .
  • Control program 1032 may also be executed to reconstruct volumes 1036 from emission data 1034 as is known.
  • Control program 1032 may also be executed to cause control system 1020 to perform process 200 , including determinations of CL locations and of local fluctuation measures.
  • Quality measures may be determined based on fluctuation-to-quality measure mapping 1038 , which may comprise convolution kernel parameters of a trained neural network.
  • Terminal 1050 may comprise a display device and an input device coupled to system 1020 .
  • Terminal 1050 may display any of two-dimensional emission data 1034 , reconstructed volumes 1036 , and local fluctuation measures, and may receive user input for controlling display of the data, operation of imaging system 1000 , and/or the processing described herein.
  • terminal 1050 is a separate computing device such as, but not limited to, a desktop computer, a laptop computer, a tablet computer, and a smartphone.
  • Each of component of system 1000 may include other elements which are necessary for the operation thereof, as well as additional elements for providing functions other than those described herein.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Medical Informatics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • General Physics & Mathematics (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Surgery (AREA)
  • Public Health (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Pathology (AREA)
  • Animal Behavior & Ethology (AREA)
  • Optics & Photonics (AREA)
  • High Energy & Nuclear Physics (AREA)
  • Veterinary Medicine (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Nuclear Medicine (AREA)

Abstract

A system and method include acquisition of a plurality of projection images of a subject, each of the projection images associated with a respective projection angle, determination, for each of the projection images, of a center-of-light location in a first image region, determination of a local fluctuation measure based on the determined center-of-light locations, and determination of a quality measure associated with the plurality of projection images based on the local fluctuation measure.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
The present application is related to and claims priority to U.S. Provisional Patent Application No. 62/735,431, filed Sep. 24, 2018, the contents of which are incorporated by reference in their entirety, for all purposes.
BACKGROUND
In emission imaging such as single-photon-emission-computer-tomography (SPECT), detectors rotate around a patient to acquire a set of projection images based on radiation emitted from the patient. Three-dimensional images are reconstructed from the set of projection images using known techniques. The three-dimensional images may be used for diagnosis and/or treatment planning.
Acquisition of the projection images may be performed over a timespan of a few minutes to several hours, during which the patient is likely to move. Patient motion during projection image acquisition may result in data inconsistency between the projection images. A three-dimensional image reconstructed from these projection images may exhibit poor quality and thereby hinder accurate diagnosis and treatment. If this three-dimensional image is reconstructed after the patient has been released from the scanning theater, the patient may have to return for another scan at a later time.
Conventionally, a technician may visually assess data inconsistency among projection images prior to releasing the patient. This approach is subjective and technician-dependent, and does not provide consistent or reliably repeatable results. Systems are desired to efficiently quantify data inconsistency among a set of projection images, preferably prior to image reconstruction.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram of a system to generate a quality measure based on projection images according to some embodiments;
FIG. 2 is a flow diagram of a process to evaluate the quality of projection images prior to image reconstruction according to some embodiments;
FIG. 3a is a representation of a center-of-light locations within regions of a projection image according to some embodiments;
FIG. 3b is a representation of a center-of-light locations within regions of a projection image according to some embodiments;
FIG. 4 includes two graphs depicting changes in center-of-light locations for a set of projection images depicting low levels of target motion according to some embodiments;
FIG. 5 includes two graphs depicting changes in center-of-light locations for a set of projection images depicting moderate levels of target motion according to some embodiments;
FIG. 6 illustrates a set of projection image images associated with respective center-of-light locations according to some embodiments;
FIG. 7 depicts determination of a local fluctuation measure according to some embodiments;
FIG. 8 is a block diagram of a system to generate a fluctuation measure-to-quality measure mapping according to some embodiments;
FIG. 9 is a block diagram of a system to train an artificial neural network to generate a quality measure based on an input fluctuation measure according to some embodiments; and
FIG. 10 illustrates a SPECT imaging system according to some embodiments.
DETAILED DESCRIPTION
The following description is provided to enable any person in the art to make and use the described embodiments and sets forth the best mode contemplated for carrying out the described embodiments. Various modifications, however, will remain apparent to those in the art.
Embodiments may provide technical improvements over existing systems by efficiently evaluating data inconsistency between projection images. Consequently, implementations may avoid reconstruction of images based on inconsistent data and/or signal the desirability of re-scanning a patient prior to releasing the patient from the scanning theater.
Some embodiments provide a system to assess the data consistency of projection images, for example those acquired for purposes of reconstructing a three-dimensional image therefrom. Generally, the coordinates of one or more center-of-light (CL) locations are determined for each projection image. Fluctuations of the coordinates between successive images are determined and a measure of data inconsistency (e.g., represented by a measure of data quality) is determined based on the determined fluctuations.
Intuitively, if the set of projection images is acquired with proper count density and with no patient motion, the coordinates of the CL locations should not change drastically among adjacent projection images. Conversely, any patient motion or other device abnormality may cause a non-negligible change in CL location coordinates among some adjacent projection images. Data quality of a dense and complex set of data (i.e., a set of projection images) may therefore be determined based on a much smaller set of data (i.e., coordinates of the one or more CL locations).
According to some embodiments, a mapping is used to determine a measure of data quality based on the determined fluctuations. The mapping may be generated based on determinations of local fluctuations within historical sets of projection images depicting many different types of motion, and in human-generated characterizations of the quality of images reconstructed from each of the historical sets of projection images. Some embodiments may alternatively use the determined local fluctuations of the historical sets of projection images and corresponding human-generated characterizations to train an artificial neural network to generate a quality measure based on one or more input fluctuation measures. Generally, the mapping maps a first small set of data (i.e., fluctuation measures) representing a dense and complex set of data (i.e., a set of projection images) to another small set of data (i.e., the human-generated characterizations) representing the dense and complex set of data.
Embodiments may be used to analyze sets of projection images of any suitable type and acquired by any suitable means. A set of projection images according to some embodiments represents n-dimensional space using k (n−1)-dimensional images, where k is the number of different projection angles from which each (n−1)-dimensional image is acquired.
FIG. 1 is a block diagram of system 100 according to some embodiments. System 100 receives a set of projection images 110, each of which depicts a subject from a respective projection angle. Projection images 110 may include any number of projection images. Although embodiments are described herein with respect to projection images generated based on radioactive emissions, embodiments may be implemented in conjunction with transmission images (e.g., computed tomography, X-ray) and any other type of images including pixels associated with brightness values, regardless of imaging modality.
Quality determination unit 120 receives projection images 110 of a patient, for example. Quality determination unit 120 includes center-of-light determination component 122, local fluctuation determination component 124 and local fluctuation-to-data quality mapping component 126. Briefly, center-of-light determination component 122 determines the coordinates of one or more CL locations for each of projection images 110. Local fluctuation determination component 124 determines one or more measures of local fluctuation coordinates between successive ones of projection images 110 based on their respective CL location coordinates. Next, local fluctuation-to-data quality mapping component 126 maps the one or more measures of local fluctuation coordinates to a data quality measure. The data quality measure may be used to determine whether to re-scan the patient or to proceed with image reconstruction based on projection images 110. Each of the foregoing processes will be described in detail below.
Each unit, component or other functional system described herein may be implemented by one or more computing devices (e.g., computer servers), storage devices (e.g., hard or solid-state disk drives), and other hardware as is known in the art. These components may be located remote from one another and may be elements of one or more cloud computing platforms, including but not limited to a Software-as-a-Service, a Platform-as-a-Service, and an Infrastructure-as-a-Service platform.
FIG. 2 is a flow diagram of a process according to some embodiments. Process 200 and the other processes described herein may be performed using any suitable combination of hardware and software. Software program code embodying these processes may be stored by any non-transitory tangible medium, including a fixed disk, a volatile or non-volatile random access memory, a CD, a DVD, a Flash drive, or a magnetic tape. Embodiments are not limited to the examples described below.
Initially, a plurality of projection images are acquired at S210. Each projection image is associated with a respective projection angle. For example, a SPECT detector may rotate around a patient and receive gamma ray emissions therefrom at each of a plurality of projection angles. The emissions received at a projection angle are used to generate an image associated with the projection angle. In other embodiments, an X-ray source transmits X-rays through the patient toward a detector at each projection angle.
The projection images may be acquired at S210 by an imaging system which also performs the remaining steps of process 200. In some embodiments, the projection images were acquired by an imaging system in the past and are acquired at S210 (e.g., via a Flash drive or network connection) by a computing system which performs the remaining steps of process 200. Prior to performing the remaining steps, the projection images may be subjected to preprocessing such as image denoising. Image denoising may be particularly desirable if the count densities of the images are low.
For each projection image, a CL location is determined at S220 for each of one or more image regions. Generally, the CL location of an image region is determined based on weighted values of each image pixel in the region, wherein the weight of an image pixel is based on its associated brightness (or other visual characteristic) value. The weighted distribution of image pixels in the region should be roughly similar in all directions around the CL location in some embodiments. The CL location may be therefore be considered roughly analogous to the center of mass of a physical object.
According to some embodiments, a first “global” CL location is determined for an entire projection image, four image regions are determined based on the global CL location, and a local CL location is determined for each of the four image regions. FIGS. 3a and 3b illustrate the determination of CL locations for projection image 310 and projection image 320 according to some embodiments. Referring to FIG. 3a , the global CL location is defined by coordinate (x1o, y1o). As shown, image 310 is logically divided into four image regions (i.e., Top Right, Top Left, Bottom Right and Bottom Left) based on the x- and y-coordinates of the global CL location. A local CL location is determined within each region based on the pixel values of the region. As shown, the coordinates of the local CL locations of each of the image regions are (x1TL, y1TL), (x1TR, y1TR), (x1BL, y1BL), and (x1BR, y1BR).
The global CL location of projection image 320 of FIG. 3 is defined by coordinate (x2o, y2o). Image 320 is also logically divided into four image regions (i.e., Top Right, Top Left, Bottom Right and Bottom Left) based on the x- and y-coordinates of the global CL location. However, unlike projection image 310, the sizes of the regions differ because the coordinate (x2o, y2o) of the global CL location is not the physical center of image 320. The coordinates of the local CL locations determined for each of the image regions are (x2TL, y2TL), (x2TR, y2TR), (x2BL, y2BL), and (x2BR, y2BR).
The following is an explanation of the determination of the coordinates of the five CL locations of projection images 310 and 320 according to some embodiments. It is assumed that the two-dimensional image array A is expressed as a function on some closed discrete subset in the X-Y plane:
Figure US11484279-20221101-P00001
=[x min , x max ]×[y min , y max], where
A={I(x, y):x, y are positive integers with (x, y)∈
Figure US11484279-20221101-P00001
}
The center of light of A may be defined in some embodiments as:
x CL ( 𝕍 ) = 1 N ( x , y ) V xI ( x , y ) y CL ( 𝕍 ) = 1 N ( x , y ) V yI ( x , y ) where N = ( x , y ) V I ( x , y )
Using the above notation, the global CL location is:
x o :=x CL(
Figure US11484279-20221101-P00001
), y o :=y CL(
Figure US11484279-20221101-P00001
)
As illustrated in FIGS. 3a and 3b , the global CL location may be used to determine multiple sub-regions Q of the discrete subset V such as:
Figure US11484279-20221101-P00002
TR =[x o , x max ]×[y min , y o]
Figure US11484279-20221101-P00002
TL =[x min , x o ]×[y min , y o]
Figure US11484279-20221101-P00002
BL =[x min , x o ]×[y o , y min]
Figure US11484279-20221101-P00002
BR =[x o , x min ]×[y o , y max]
The CL locations of each of the sub-regions may then be expressed as:
(x 1 , y 1):=(x CL(
Figure US11484279-20221101-P00002
RT), y CL(
Figure US11484279-20221101-P00002
TR))
(x 2 , y 2):=(x CL(
Figure US11484279-20221101-P00002
TL), y CL(
Figure US11484279-20221101-P00002
TL))
(x 3 , y 3):=(x CL(
Figure US11484279-20221101-P00002
BL), y CL(
Figure US11484279-20221101-P00002
BL))
(x 4 , y 4):=(x CL(
Figure US11484279-20221101-P00002
BR), y CL(
Figure US11484279-20221101-P00002
BR))
Embodiments are not limited to four image regions or to image regions defined as described above. For example, a projection image may be divided into regions by two or more angular, curved and/or circular boundaries.
FIG. 4 shows plot 410 of X and Y coordinates of CL locations over a series of projection images. As expected, the CL locations change due to the changing angular perspectives of the projection images. Plot 420 shows second derivatives of the curves of plot 410. Plot 420 shows that the rate of change of the CL location changes is fairly constant, indicating a lack of substantial subject motion.
FIG. 5 shows plot 510 of X and Y coordinates of CL locations over a series of projection images, and plot 520 shows second derivatives of the curves of plot 510. Plot 520 shows fluctuations in the second derivative and thereby suggests subject motion during imaging.
Returning to process 200, a local fluctuation measure is determined at S230 for each set of three adjacent projection images based on the determined center of light locations. The determined local fluctuation measure may comprise two or more local fluctuation measures or may be a composite of two or more local fluctuation measures according to some embodiments.
FIG. 6 illustrates projection images 1 through n acquired at S210 according to one example. Each of projection images 1 through n is associated with a different projection angle. It will be assumed that the projection images are numbered in accordance with their order of acquisition. That is, projection image 2 was acquired after projection image 1 and after projection image 3, and is associated with a projection angle between the projection angles associated with projection images 1 and 3.
A global CL location and four local CL locations have been determined for each of projection images 1 through n at S220. As illustrated, the regions of the local CL locations may differ among projection images 1 through n due to difference in the coordinates of their associated global CL locations.
In one example of S230 with respect to projection images 1 through n of FIG. 6, a local fluctuation measure is determined for each region for which CL locations were determined at S220. Specifically, a local fluctuation measure is determined based on coordinates (x1o, y1o), (x2o, y2o), . . . , (xno, yno), a local fluctuation measure is determined based on coordinates (x1TL, y1TL), (x2TL, y2TL), . . . , (xnTL, ynTL), and so on for the coordinates of each CL location determined for regions TR, BL and BR.
FIG. 7 illustrates determination of a local fluctuation measure for a single region according to some embodiments. For the kth view in a set of projection images, the local fluctuation measure δk for a region is the distance of the CL location of the kth view to a line connecting the CL locations of the two neighboring views. As shown, the determination of measure δk utilizes three coordinates (xk−1, yk−1), (xk, yk) and (xk+1, yk+1).
With reference to FIG. 6, the three coordinates (x1TL, y1TL), (x2TL, y2TL) and (x3TL, y3TL) may be used to determine δ2. More specifically, using the illustrated calculations of FIG. 7, measure δ2 may be calculated for region TL based on the values of x1TL, y1TL, x2TL, y2TL, x3TL, and y3TL. Measure δk may be calculated for each other combination of CL locations of region TL within three successive projection images. Measure δk, k=2 to n−1 may be similarly calculated for each other region, including the global region, of projection images 1 through n.
Therefore, a series of local fluctuation measures may be determined at S230 for each region. According to some embodiments, a single local fluctuation measure is determined for each region by summing the series of measures determined for the region. Any other suitable systems to determine a single local fluctuation measure for each region based on a series thereof may be employed. Moreover, the single local fluctuation measures for each region may be combined (e.g., summed) to determine a single overall local fluctuation measure.
A quality measure is determined based on the local fluctuation measure(s) at S240. The quality measure may be determined based on a mapping between local fluctuation measure(s) and image quality measures.
FIG. 8 illustrates system 800 to generate such a mapping according to some embodiments. Q sets 810 of historical projection images1-k preferably represent different types of motion which may be encountered during subsequent use of the generated mapping as described above. Accordingly, it is desirable that each of Q sets 810 of historical projection images1-k represents some degree of translational or rotational target motion (i.e., from no motion to extreme motion) along any one or more axes.
For each set Q of historical projection images1-k, unit 820 determines CL locations (822) and local fluctuation measures (824), resulting in associated local fluctuation measures1-Q. Each of local fluctuation measures1-Q may comprise more than one local fluctuation measure (e.g., one fluctuation measure for each projection image region), and is associated with a single set of historical projection images1-k.
Each set Q of historical projection images1-k is also subjected to reconstruction (830) to generate an associated one of reconstructed volumes1-Q. Human observer 840 inspects each of reconstructed volumes1-Q and determines a quality measure for each reconstructed volume1-Q. In some examples, more than one human observer 840 inspects each reconstructed volume and/or each of reconstructed volumes1-Q is inspected by one of any number of human observers 840, each of whom follows a specified protocol for determining a quality measure based on a reconstructed volume.
Accordingly, each set Q of historical projection images1-k is associated with one of local fluctuation measures1-Q and one of quality measures1-Q. These associations are used to generate a mapping (850) between the two data types. Accordingly, such a mapping may be used at S240 to determine a quality measure based on a local fluctuation measure.
FIG. 9 illustrates system 900 according to some embodiments. As illustrated, local fluctuation measures1-Q and associated quality measures1-Q are generated based on Q sets of historical projection images1-n as described above. Local fluctuation measures1-Q and associated quality measures1-Q are input to training system 910 as training data for artificial neural network 920. For example, system 910 may train system 920 such that system 920 substantially outputs one of quality measures1-Q upon receiving its associated local fluctuation measure1-Q as input. According to some embodiments, training system 910 uses CL locations and associated quality measures1-Q as training data, eliminating a need to determine local fluctuations.
Artificial neural network 920 may comprise a network of neurons which receive input, change internal state according to that input, and produce output depending on the input and internal state. The output of certain neurons is connected to the input of other neurons to form a directed and weighted graph. The weights as well as the functions that compute the internal state can be modified by a training process based on ground truth data. Artificial neural network 920 may comprise any one or more types of artificial neural network that are or become known, including but not limited to convolutional neural networks, recurrent neural networks, long short-term memory networks, deep reservoir computing and deep echo state networks, deep belief networks, and deep stacking networks.
In one example, the training of network 920 generates parameter values for kernels of a fully convolutional network. Another fully convolutional network comprising thusly-parameterized kernels may be efficiently incorporated into a system to determine a quality measure at S240 based on a local fluctuation measure.
Returning to process 200, it is determined at S250 if the quality measure is acceptable. For example, the determined quality measure may be compared to a prespecified threshold. The threshold may differ based on the intended usage of the projection images, time constraints, and any other suitable factor. If the quality measure is not acceptable, flow returns to S210 to acquire another set of projection images. As described above, it may be advantageous to determine the unacceptable quality and acquire the other set of images before releasing the patient.
Flow proceeds from S250 to S260 is the quality measure is deemed acceptable. At S260, an image is reconstructed based on the set of projection images. Any suitable reconstructions algorithm may be employed at S260. The reconstructed images may then be used as intended.
FIG. 10 illustrates SPECT system 1000 which may implement process 200 as described above. System 1000 includes gantry 1002 to which two or more gamma cameras 1004 a, 1004 b are attached, although any number of gamma cameras can be used. A detector within each gamma camera detects gamma photons (i.e., emission data) emitted by a radioisotope within the body of a patient 1006 lying on a bed 1008.
Bed 1008 is capable of moving patient 1006 along axis A and/or axis B. At respective bed positions (i.e., imaging positions), a portion of the body of patient 1006 is positioned between gamma cameras 1004 a, 1004 b in order to capture emission data from that body portion. Gamma cameras 1004 a, 1004 b may include multi-focal cone-beam collimators or parallel-hole collimators as is known in the art.
Control system 1020 may comprise any general-purpose or dedicated computing system. Accordingly, control system 1020 includes one or more processing units 1022 configured to execute processor-executable program code to cause system 1020 to operate as described herein, and storage device 1030 for storing the program code. Storage device 1030 may comprise one or more fixed disks, solid-state random access memory, and/or removable media (e.g., a thumb drive) mounted in a corresponding interface (e.g., a USB port).
Storage device 1030 stores program code of system control program 1032. One or more processing units 1022 may execute system control program 1032, in conjunction with SPECT system interface 1040, to control motors, servos, and encoders to cause gamma cameras 1004 a, 1004 b to rotate along gantry 1002 and to acquire two-dimensional emission data (i.e. projection images) at defined imaging positions during the rotation. The acquired data 1034 may be stored in memory 1030. Control program 1032 may also be executed to reconstruct volumes 1036 from emission data 1034 as is known.
Control program 1032 may also be executed to cause control system 1020 to perform process 200, including determinations of CL locations and of local fluctuation measures. Quality measures may be determined based on fluctuation-to-quality measure mapping 1038, which may comprise convolution kernel parameters of a trained neural network.
Terminal 1050 may comprise a display device and an input device coupled to system 1020. Terminal 1050 may display any of two-dimensional emission data 1034, reconstructed volumes 1036, and local fluctuation measures, and may receive user input for controlling display of the data, operation of imaging system 1000, and/or the processing described herein. In some embodiments, terminal 1050 is a separate computing device such as, but not limited to, a desktop computer, a laptop computer, a tablet computer, and a smartphone.
Each of component of system 1000 may include other elements which are necessary for the operation thereof, as well as additional elements for providing functions other than those described herein.
Those in the art will appreciate that various adaptations and modifications of the above-described embodiments can be configured without departing from the claims. Therefore, it is to be understood that the claims may be practiced other than as specifically described herein.

Claims (20)

What is claimed is:
1. A system comprising:
a storage device;
a processor to execute processor-executable process steps stored on the storage device to cause the system to:
for each of a plurality of projection images of a subject, each of the projection images associated with a respective projection angle, determine a center-of-light location in a first image region;
determine a local fluctuation measure based on the determined center-of-light locations; and
determine a quality measure associated with the plurality of projection images based on the local fluctuation measure.
2. A system according to claim 1,
wherein determination of the local fluctuation measure is based on differences in the center-of-light locations determined for projection images associated with adjacent projection angles.
3. A system according to claim 1, wherein the processor is to execute processor-executable process steps to cause the system to:
determine, for each of the projection images, a center-of-light location in a second image region,
wherein determination of the local fluctuation measure comprises:
determination of a first local fluctuation measure based on the center-of-light locations determined in the first image region,
determination of a second local fluctuation measure based on the center-of-light locations determined in the second image region, and
determination of the local fluctuation measure based on the first local fluctuation measure and the second local fluctuation measure.
4. A system according to claim 3,
wherein determination of the first local fluctuation measure is based on differences in the center-of-light locations determined in the first image region for projection images associated with adjacent projection angles, and
wherein determination of the second local fluctuation measure is based on differences in the center-of-light locations determined in the second image region for projection images associated with adjacent projection angles.
5. A system according to claim 3, wherein the processor is to execute processor-executable process steps to cause the system to:
determine a first center-of-light location for a first one of the projection images;
determine the first region and the second region of the first one of the projection images based on the determined first center-of-light location;
determine a second center-of-light location for a second one of the projection images; and
determine the first region and the second region of the second one of the projection images based on the determined second center-of-light location.
6. A system according to claim 5,
wherein determination of the first local fluctuation measure is based on differences in the center-of-light locations determined in the first image region for projection images associated with adjacent projection angles, and
wherein determination of the second local fluctuation measure is based on differences in the center-of-light locations determined in the second image region for projection images associated with adjacent projection angles.
7. A system according to claim 1, wherein
when it is determined that the quality measure indicates unsuitable quality, the processor executes processor-executable process steps to cause the system to acquire a second plurality of projection images of the subject, and
when it is determined that the quality measure indicates suitable quality, the processor executes processor-executable process steps to cause the system to reconstruct a three-dimensional image based on the plurality of projection images of the subject.
8. A system according to claim 7, wherein determination of the quality measure associated with the plurality of projection images based on the local fluctuation measure comprises inputting the local fluctuation measure to a trained artificial neural network.
9. A method comprising:
for each of a plurality of projection images of a subject, each of the projection images associated with a respective projection angle, determining a center-of-light location in a first image region;
determining a local fluctuation measure based on the determined center-of-light locations; and
determining a quality measure associated with the plurality of projection images based on the local fluctuation measure.
10. A method according to claim 9, wherein determining the local fluctuation measure is based on differences in the center-of-light locations determined for projection images associated with adjacent projection angles.
11. A method according to claim 9, further comprising:
determining, for each of the projection images, a center-of-light location in a second image region,
wherein determining the local fluctuation measure comprises:
determining a first local fluctuation measure based on the center-of-light locations determined in the first image region,
determining a second local fluctuation measure based on the center-of-light locations determined in the second image region, and
determining the local fluctuation measure based on the first local fluctuation measure and the second local fluctuation measure.
12. A method according to claim 11,
wherein determining the first local fluctuation measure is based on differences in the center-of-light locations determined in the first image region for projection images associated with adjacent projection angles, and
wherein determining the second local fluctuation measure is based on differences in the center-of-light locations determined in the second image region for projection images associated with adjacent projection angles.
13. A method according to claim 11, further comprising:
determining a first center-of-light location for a first one of the projection images;
determining the first region and the second region of the first one of the projection images based on the determined first center-of-light location;
determining a second center-of-light location for a second one of the projection images; and
determining the first region and the second region of the second one of the projection images based on the determined second center-of-light location.
14. A method according to claim 13,
wherein determining the first local fluctuation measure is based on differences in the center-of-light locations determined in the first image region for projection images associated with adjacent projection angles, and
wherein determining the second local fluctuation measure is based on differences in the center-of-light locations determined in the second image region for projection images associated with adjacent projection angles.
15. A method according to claim 9, further comprising:
acquiring a second plurality of projection images of the subject when it is determined that the quality measure indicates unsuitable quality; and
reconstructing a three-dimensional image based on the plurality of projection images of the subject when it is determined that the quality measure indicates suitable quality.
16. A method according to claim 15, wherein determining the quality measure associated with the plurality of projection images based on the local fluctuation measure comprises inputting the local fluctuation measure to a trained artificial neural network.
17. A system comprising:
an imaging system to acquire a plurality of projection images of a subject, each of the projection images acquired from a respective projection angle; and
a control system to:
determine, for each of the projection images, a center-of-light location in a first image region;
determine a local fluctuation measure based on the determined center-of-light locations;
determine a quality measure associated with the plurality of projection images based on the local fluctuation measure; and
acquire a second plurality of projection images of the subject before reconstructing a three-dimensional image based on the plurality of projection images when it is determined that the quality measure indicates unsuitable quality.
18. A system according to claim 17,
wherein determination of the local fluctuation measure is based on differences in the center-of-light locations determined for projection images associated with adjacent projection angles.
19. A system according to claim 17, wherein the processor is to execute processor-executable process steps to cause the system to:
determine, for each of the projection images, a center-of-light location in a second image region,
wherein determination of the local fluctuation measure comprises:
determination of a first local fluctuation measure based on differences in the center-of-light locations determined in the first image region for projection images associated with adjacent projection angles,
determination of a second local fluctuation measure based on differences in the center-of-light locations determined in the second image region for projection images associated with adjacent projection angles, and
determination of the local fluctuation measure based on the first local fluctuation measure and the second local fluctuation measure.
20. A system according to claim 19, wherein the processor is to execute processor-executable process steps to cause the system to:
determine a first center-of-light location for a first one of the projection images;
determine the first region and the second region of the first one of the projection images based on the determined first center-of-light location;
determine a second center-of-light location for a second one of the projection images; and
determine the first region and the second region of the second one of the projection images based on the determined second center-of-light location.
US16/409,967 2018-09-24 2019-05-13 Systems to assess projection data inconsistency Active 2041-06-27 US11484279B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/409,967 US11484279B2 (en) 2018-09-24 2019-05-13 Systems to assess projection data inconsistency

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201862735431P 2018-09-24 2018-09-24
US16/409,967 US11484279B2 (en) 2018-09-24 2019-05-13 Systems to assess projection data inconsistency

Publications (2)

Publication Number Publication Date
US20200093454A1 US20200093454A1 (en) 2020-03-26
US11484279B2 true US11484279B2 (en) 2022-11-01

Family

ID=69885665

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/409,967 Active 2041-06-27 US11484279B2 (en) 2018-09-24 2019-05-13 Systems to assess projection data inconsistency

Country Status (1)

Country Link
US (1) US11484279B2 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11494955B2 (en) 2020-06-10 2022-11-08 Siemens Medical Solutions Usa, Inc. Data driven reconstruction in emission tomography
CN112581595B (en) * 2020-12-02 2023-12-19 中国人民解放军战略支援部队航天工程大学 A multi-view satellite image consistency analysis method

Citations (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1408102A (en) 1999-07-26 2003-04-02 计算机化医疗系统公司 Automated image fusion/alignment system and method
US20030128801A1 (en) * 2002-01-07 2003-07-10 Multi-Dimensional Imaging, Inc. Multi-modality apparatus for dynamic anatomical, physiological and molecular imaging
US20070048326A1 (en) 2001-06-20 2007-03-01 Genentech, Inc. Compositions and Methods for the Diagnosis and Treatment of Tumor
US20070237290A1 (en) * 2006-04-07 2007-10-11 Varian Medical Systems Technologies, Inc. Patient setup using tomosynthesis techniques
CN101167662A (en) 2006-10-25 2008-04-30 美国西门子医疗解决公司 System and method for lymph node imaging by cooperation of CT and MR
CN101408976A (en) 2007-10-11 2009-04-15 通用电气公司 Enhanced system and method for volume based registration
US20100232645A1 (en) 2007-05-10 2010-09-16 Koninklijke Philips Electronics N.V. Model-based spect heart orientation estimation
US20100289813A1 (en) 2009-05-13 2010-11-18 Kabushiki Kaisha Toshiba Nuclear medical imaging apparatus, image processing apparatus, and image processing method
US20110142314A1 (en) 2009-12-15 2011-06-16 Jiang Hsieh System and method of increasing temporal resolution of an x-ray image
US20110228897A1 (en) 2008-03-07 2011-09-22 Aloka Co., Ltd. X-ray ct scanner and control program thereof
US20120169333A1 (en) 2009-09-08 2012-07-05 Koninklijke Philips Electronics N.V. Motion monitoring system for monitoring motion within a region of interest
CN102846326A (en) 2011-06-03 2013-01-02 通用电气公司 Method and system for processing gated image data
US20130034286A1 (en) 2011-08-03 2013-02-07 Siemens Medical Solutions Usa, Inc. Registration of Emission Tomography and Computed Tomography
CN102982510A (en) 2011-06-21 2013-03-20 通用电气公司 Motion correction of SPECT images
CN103136734A (en) 2013-02-27 2013-06-05 北京工业大学 Restraining method on edge Halo effects during process of resetting projections onto convex sets (POCS) super-resolution image
US20130315459A1 (en) 2012-05-22 2013-11-28 General Electric Company Method and apparatus for motion correcting medical images
US20140072194A1 (en) 2011-05-12 2014-03-13 Koninklijke Philips N.V. Motion compensated imaging
CN104169969A (en) 2012-03-15 2014-11-26 皇家飞利浦有限公司 Multi-modality deformable registration
US20140355855A1 (en) 2013-05-30 2014-12-04 Siemens Aktiengesellschaft System and Method for Magnetic Resonance Imaging Based Respiratory Motion Correction for PET/MRI
US20150302616A1 (en) 2012-12-06 2015-10-22 Koninklijke Philips N.V. Local artifact reduction with insignificant side effects
US20160095565A1 (en) 2014-10-01 2016-04-07 Siemens Aktiengesellschaft Method and imaging system for compensating for location assignment errors in pet data occurring due to a cyclical motion of a patient
US20160125605A1 (en) 2014-10-29 2016-05-05 Samsung Electronics Co., Ltd. Image processing apparatus, image processing method, and ultrasound imaging apparatus having image processing unit
US20160203609A1 (en) 2013-09-27 2016-07-14 Fujifilm Corporation Image alignment device, method, and program, and method for generating 3-d deformation model
US20160220311A1 (en) 2015-01-29 2016-08-04 Siemens Healthcare Gmbh Planning and guidance of electrophysiology therapies
US20170193159A1 (en) 2014-06-13 2017-07-06 Siemens Healthcare Gmbh One Gate Reconstruction
US20170209112A1 (en) 2016-01-25 2017-07-27 Samsung Electronics Co., Ltd Tomography apparatus and method of reconstructing tomography image
US20170219674A1 (en) 2014-04-18 2017-08-03 The General Hospital Corporation System and method for magnetic resonance imaging with prospective motion control
US20170278280A1 (en) 2014-07-30 2017-09-28 Navix International Limited Registering nuclear medicine data
US20170360325A1 (en) 2014-12-11 2017-12-21 Elekta, Inc. Motion management in mri-guided linac
US20180025512A1 (en) 2016-07-20 2018-01-25 Shanghai United Imaging Healthcare Co., Ltd. System and method for segmenting medical image
US20180033166A1 (en) 2014-06-13 2018-02-01 Siemens Healthcare Gmbh Intra Reconstruction Motion Correction
US20180056091A1 (en) 2016-08-29 2018-03-01 Accuray Incorporated Offline angle selection in rotational imaging and tracking systems
US20180336677A1 (en) 2017-05-18 2018-11-22 Toshiba Medical Systems Corporation Apparatus and method for medical image processing

Patent Citations (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1408102A (en) 1999-07-26 2003-04-02 计算机化医疗系统公司 Automated image fusion/alignment system and method
US20070048326A1 (en) 2001-06-20 2007-03-01 Genentech, Inc. Compositions and Methods for the Diagnosis and Treatment of Tumor
US20030128801A1 (en) * 2002-01-07 2003-07-10 Multi-Dimensional Imaging, Inc. Multi-modality apparatus for dynamic anatomical, physiological and molecular imaging
US20070237290A1 (en) * 2006-04-07 2007-10-11 Varian Medical Systems Technologies, Inc. Patient setup using tomosynthesis techniques
CN101167662A (en) 2006-10-25 2008-04-30 美国西门子医疗解决公司 System and method for lymph node imaging by cooperation of CT and MR
US20100232645A1 (en) 2007-05-10 2010-09-16 Koninklijke Philips Electronics N.V. Model-based spect heart orientation estimation
CN101408976A (en) 2007-10-11 2009-04-15 通用电气公司 Enhanced system and method for volume based registration
US20110228897A1 (en) 2008-03-07 2011-09-22 Aloka Co., Ltd. X-ray ct scanner and control program thereof
US20100289813A1 (en) 2009-05-13 2010-11-18 Kabushiki Kaisha Toshiba Nuclear medical imaging apparatus, image processing apparatus, and image processing method
US20120169333A1 (en) 2009-09-08 2012-07-05 Koninklijke Philips Electronics N.V. Motion monitoring system for monitoring motion within a region of interest
US20110142314A1 (en) 2009-12-15 2011-06-16 Jiang Hsieh System and method of increasing temporal resolution of an x-ray image
US20140072194A1 (en) 2011-05-12 2014-03-13 Koninklijke Philips N.V. Motion compensated imaging
CN102846326A (en) 2011-06-03 2013-01-02 通用电气公司 Method and system for processing gated image data
CN102982510A (en) 2011-06-21 2013-03-20 通用电气公司 Motion correction of SPECT images
US20130034286A1 (en) 2011-08-03 2013-02-07 Siemens Medical Solutions Usa, Inc. Registration of Emission Tomography and Computed Tomography
CN104169969A (en) 2012-03-15 2014-11-26 皇家飞利浦有限公司 Multi-modality deformable registration
CN103417234A (en) 2012-05-22 2013-12-04 通用电气公司 Method and apparatus for motion correcting medical images
US20130315459A1 (en) 2012-05-22 2013-11-28 General Electric Company Method and apparatus for motion correcting medical images
US20150302616A1 (en) 2012-12-06 2015-10-22 Koninklijke Philips N.V. Local artifact reduction with insignificant side effects
CN103136734A (en) 2013-02-27 2013-06-05 北京工业大学 Restraining method on edge Halo effects during process of resetting projections onto convex sets (POCS) super-resolution image
US20140355855A1 (en) 2013-05-30 2014-12-04 Siemens Aktiengesellschaft System and Method for Magnetic Resonance Imaging Based Respiratory Motion Correction for PET/MRI
US20160203609A1 (en) 2013-09-27 2016-07-14 Fujifilm Corporation Image alignment device, method, and program, and method for generating 3-d deformation model
US20170219674A1 (en) 2014-04-18 2017-08-03 The General Hospital Corporation System and method for magnetic resonance imaging with prospective motion control
US20180033166A1 (en) 2014-06-13 2018-02-01 Siemens Healthcare Gmbh Intra Reconstruction Motion Correction
US20170193159A1 (en) 2014-06-13 2017-07-06 Siemens Healthcare Gmbh One Gate Reconstruction
US20170278280A1 (en) 2014-07-30 2017-09-28 Navix International Limited Registering nuclear medicine data
US20160095565A1 (en) 2014-10-01 2016-04-07 Siemens Aktiengesellschaft Method and imaging system for compensating for location assignment errors in pet data occurring due to a cyclical motion of a patient
US20160125605A1 (en) 2014-10-29 2016-05-05 Samsung Electronics Co., Ltd. Image processing apparatus, image processing method, and ultrasound imaging apparatus having image processing unit
US20170360325A1 (en) 2014-12-11 2017-12-21 Elekta, Inc. Motion management in mri-guided linac
US20160220311A1 (en) 2015-01-29 2016-08-04 Siemens Healthcare Gmbh Planning and guidance of electrophysiology therapies
US20170209112A1 (en) 2016-01-25 2017-07-27 Samsung Electronics Co., Ltd Tomography apparatus and method of reconstructing tomography image
US20180025512A1 (en) 2016-07-20 2018-01-25 Shanghai United Imaging Healthcare Co., Ltd. System and method for segmenting medical image
US20180056091A1 (en) 2016-08-29 2018-03-01 Accuray Incorporated Offline angle selection in rotational imaging and tracking systems
US20180336677A1 (en) 2017-05-18 2018-11-22 Toshiba Medical Systems Corporation Apparatus and method for medical image processing

Also Published As

Publication number Publication date
US20200093454A1 (en) 2020-03-26

Similar Documents

Publication Publication Date Title
US10593071B2 (en) Network training and architecture for medical imaging
US11769277B2 (en) Deep learning based scatter correction
US11164344B2 (en) PET image reconstruction using TOF data and neural network
US20200311490A1 (en) Apparatus and method for sinogram restoration in computed tomography (ct) using adaptive filtering with deep learning (dl)
US9848844B2 (en) Iterative reconstruction process
US10475538B2 (en) System for predicting computed tomography data from surface data
US11874411B2 (en) Estimation of partially missing attenuation in time-of-flight positron emission tomography
US11484279B2 (en) Systems to assess projection data inconsistency
US8755586B2 (en) Image reconstruction including shift-variant blur compensation
US11334987B2 (en) Physics-constrained network and training thereof
CN113643394B (en) Scattering correction method, device, computer equipment and storage medium
US11810228B2 (en) Network determination of limited-angle reconstruction
EP4062836A1 (en) Neural network-assisted dose assessment based on activity image
US11335040B2 (en) Multi-focal non-parallel collimator-based imaging
EP4148680A1 (en) Attenuation correction-based weighting for tomographic inconsistency detection
US11574184B2 (en) Multi-modal reconstruction network
US10687766B2 (en) System to detect features using multiple reconstructions
US20250259349A1 (en) Ai-driven motion correction of pet data
CN120635243B (en) Spiral CT image reconstruction method based on nerve radiation field
US20260000377A1 (en) Imaging with scatter correction with the aid of noisy scatter estimations and scatter interpolation
US20210312675A1 (en) Evaluation of Image Formation System Components
CN119908737A (en) Provides adapted recording parameters
Tirunelveli Comparison of square and hexagonal pixels for image processing and alternative solutions to linear equations for Computed Tomography

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: SIEMENS MEDICAL SOLUTIONS USA, INC., PENNSYLVANIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:VIJA, ALEXANDER HANS;DING, XINHONG;SIGNING DATES FROM 20190522 TO 20190528;REEL/FRAME:049315/0958

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE