US10859657B2 - MRI reconstruction using deep learning, generative adversarial network and acquisition signal model - Google Patents

MRI reconstruction using deep learning, generative adversarial network and acquisition signal model Download PDF

Info

Publication number
US10859657B2
US10859657B2 US16/427,599 US201916427599A US10859657B2 US 10859657 B2 US10859657 B2 US 10859657B2 US 201916427599 A US201916427599 A US 201916427599A US 10859657 B2 US10859657 B2 US 10859657B2
Authority
US
United States
Prior art keywords
network
image
images
data
cost
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US16/427,599
Other versions
US20190369191A1 (en
Inventor
Enhao Gong
Greg Zaharchuk
John M. Pauly
Morteza Mardani Korani
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Leland Stanford Junior University
Original Assignee
Leland Stanford Junior University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Leland Stanford Junior University filed Critical Leland Stanford Junior University
Priority to US16/427,599 priority Critical patent/US10859657B2/en
Assigned to THE BOARD OF TRUSTEES OF THE LELAND STANFORD JUNIOR UNIVERSITY reassignment THE BOARD OF TRUSTEES OF THE LELAND STANFORD JUNIOR UNIVERSITY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MARDANI KORANI, MORTEZA, GONG, Enhao, PAULY, JOHN M., ZAHARCHUK, GREG
Publication of US20190369191A1 publication Critical patent/US20190369191A1/en
Application granted granted Critical
Publication of US10859657B2 publication Critical patent/US10859657B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01RMEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
    • G01R33/00Arrangements or instruments for measuring magnetic variables
    • G01R33/20Arrangements or instruments for measuring magnetic variables involving magnetic resonance
    • G01R33/44Arrangements or instruments for measuring magnetic variables involving magnetic resonance using nuclear magnetic resonance [NMR]
    • G01R33/48NMR imaging systems
    • G01R33/54Signal processing systems, e.g. using pulse sequences ; Generation or control of pulse sequences; Operator console
    • G01R33/56Image enhancement or correction, e.g. subtraction or averaging techniques, e.g. improvement of signal-to-noise ratio and resolution
    • G01R33/5608Data processing and visualization specially adapted for MR, e.g. for feature analysis and pattern recognition on the basis of measured MR data, segmentation of measured MR data, edge contour detection on the basis of measured MR data, for enhancing measured MR data in terms of signal-to-noise ratio by means of noise filtering or apodization, for enhancing measured MR data in terms of resolution by means for deblurring, windowing, zero filling, or generation of gray-scaled images, colour-coded images or images displaying vectors instead of pixels
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/05Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves 
    • A61B5/055Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves  involving electronic [EMR] or nuclear [NMR] magnetic resonance, e.g. magnetic resonance imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01RMEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
    • G01R33/00Arrangements or instruments for measuring magnetic variables
    • G01R33/20Arrangements or instruments for measuring magnetic variables involving magnetic resonance
    • G01R33/44Arrangements or instruments for measuring magnetic variables involving magnetic resonance using nuclear magnetic resonance [NMR]
    • G01R33/48NMR imaging systems
    • G01R33/4818MR characterised by data acquisition along a specific k-space trajectory or by the temporal order of k-space coverage, e.g. centric or segmented coverage of k-space
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01RMEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
    • G01R33/00Arrangements or instruments for measuring magnetic variables
    • G01R33/20Arrangements or instruments for measuring magnetic variables involving magnetic resonance
    • G01R33/44Arrangements or instruments for measuring magnetic variables involving magnetic resonance using nuclear magnetic resonance [NMR]
    • G01R33/48NMR imaging systems
    • G01R33/483NMR imaging systems with selection of signals or spectra from particular regions of the volume, e.g. in vivo spectroscopy
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01RMEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
    • G01R33/00Arrangements or instruments for measuring magnetic variables
    • G01R33/20Arrangements or instruments for measuring magnetic variables involving magnetic resonance
    • G01R33/44Arrangements or instruments for measuring magnetic variables involving magnetic resonance using nuclear magnetic resonance [NMR]
    • G01R33/48NMR imaging systems
    • G01R33/54Signal processing systems, e.g. using pulse sequences ; Generation or control of pulse sequences; Operator console
    • G01R33/56Image enhancement or correction, e.g. subtraction or averaging techniques, e.g. improvement of signal-to-noise ratio and resolution
    • G01R33/561Image enhancement or correction, e.g. subtraction or averaging techniques, e.g. improvement of signal-to-noise ratio and resolution by reduction of the scanning time, i.e. fast acquiring systems, e.g. using echo-planar pulse sequences
    • G01R33/5611Parallel magnetic resonance imaging, e.g. sensitivity encoding [SENSE], simultaneous acquisition of spatial harmonics [SMASH], unaliasing by Fourier encoding of the overlaps using the temporal dimension [UNFOLD], k-t-broad-use linear acquisition speed-up technique [k-t-BLAST], k-t-SENSE
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/40ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/20ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H50/00ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
    • G16H50/50ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for simulation or modelling of medical disorders
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Definitions

  • the present invention relates generally to magnetic resonance imaging (MRI). More specifically, it relates to techniques for MRI reconstruction.
  • MRI magnetic resonance imaging
  • MRI magnetic resonance imaging
  • Real-time MRI acquisition, reconstruction and visualization is of paramount importance for diagnostic and therapeutic guidance.
  • Interventional and image-guided therapies as well as interactive diagnostic tasks need rapid image preparation within a few milliseconds. This is hindered, however, by the slow acquisition process, taking several minutes to acquire clinically acceptable images. Inefficient acquisition becomes more pronounced for high-resolution and volumetric images.
  • One possible solution is to decrease the scan duration through significant undersampling. However, such undersampling leads to a seriously ill-posed linear inverse reconstruction problem.
  • Deep neural networks have been used to learn image prior or sparsifying transform from historical data in order to solve a nonlinear system using iterative optimization algorithms as in the conventional CS methods. While improving the reconstruction performance, these methods incur high computational cost due to several iterations for finding the optimal reconstruction.
  • the present disclosure provides a CS framework that applies generative adversarial networks (GAN) to modeling a (low-dimensional) manifold of high-quality MR images.
  • GAN generative adversarial networks
  • LS least-squares
  • l 1 /l 2 a deep residual network with skip connections
  • a discriminator network which is a multilayer convolutional neural network (CNN) plays the role of a perceptual cost that is then jointly trained based on high quality MR images to score the quality of retrieved images.
  • CNN multilayer convolutional neural network
  • an initial aliased estimate (e.g., simply obtained by zero-filling) is propagated into the trained generator to output the desired reconstructions, that demands very low computational overheads.
  • GANCS generative adversarial networks compressed sensing
  • GANs generative adversarial networks
  • GANs have proven successful in modeling distributions (low-dimensional manifolds) and generating natural images (high-dimensional data) that are perceptually appealing.
  • GANs have not been considered for correcting the aliasing artifacts in biomedical image reconstruction tasks.
  • aliasing artifacts emanate from data undersampling in the frequency domain that globally impacts the entire space domain image.
  • the present approach uses GANs for MRI reconstruction.
  • the approach uses GANs for modeling low-dimensional manifold of high-quality MR images.
  • the present GANCS approach adopts a tandem network of a generator (G), an affine projection operator (A), and a discriminator (D).
  • the generator aims to create gold-standard images from the complex-valued aliased inputs using a deep residual network (ResNet) with skip connections which retain high resolution information.
  • the data-consistency projection builds upon the (known) signal model and performs an affine projection onto the space of data consistent images.
  • the D network is a multilayer convolutional neural network (CNN) that is trained using both the “fake” images created by G, and the corresponding gold-standard ones, and aims to correctly distinguish fake from real.
  • Least-squares GANs Least-squares GANs (LSGANs) is used due to their stability properties.
  • GANs frameworks may be used, where usually the differences are just cost functions, e.g., Wasserstein GAN (W-GAN), Cycle-GAN, BigGAN, or StarGAN.
  • W-GAN Wasserstein GAN
  • Cycle-GAN Cycle-GAN
  • BigGAN BigGAN
  • StarGAN StarGAN
  • the GANCS results have almost similar quality to the gold-standard (fully-sampled) images, and are superior in terms of diagnostic quality relative to the existing alternatives including conventional iterative CS and deep learning based methods that solely adopt the pixel-wise l 2 -based and l 1 -based criteria. Moreover, the reconstruction only takes around 30 ms, which is two orders of magnitude faster than state-of-the-art conventional CS toolboxes.
  • the invention provides a method for diagnostic imaging comprising: measuring undersampled data y with a diagnostic imaging apparatus; linearly transforming the undersampled data y to obtain an initial image estimate ⁇ tilde over (x) ⁇ ; applying the initial image estimate ⁇ tilde over (x) ⁇ as input to a generator network to obtain an aliasing artifact-reduced image x ⁇ as output of the generator network, wherein the aliasing artifact-reduced image x ⁇ is a projection onto a manifold of realistic images of the initial image estimate ⁇ tilde over (x) ⁇ ; and performing an acquisition signal model projection of the aliasing artifact-reduced image x ⁇ onto a space of consistent images to obtain a reconstructed image ⁇ circumflex over (x) ⁇ having suppressed image artifacts.
  • the diagnostic imaging apparatus may be, for example, an MRI scanner, and the undersampled data is k-space data.
  • Linearly transforming the undersampled data y may comprise, for example, zero padding the undersampled data y, or finding an approximate zero-filling reconstruction from the undersampled data y.
  • the generator network is trained to learn the projection onto the manifold of realistic images using a set of training images X and corresponding set of undersampled measurements Y using least-squares generative adversarial network techniques in tandem with a discriminator network to learn texture details and supervised cost function to control high-frequency noise.
  • the supervised cost function may comprise a mixture of smooth l 2 cost and non-smooth l 1 cost.
  • the discriminator network may be a multilayer deep convolutional neural network.
  • the discriminator network may be trained using least-squares cost for a discriminator decision.
  • the generator network may be a deep residual network with skip connections.
  • performing the acquisition signal model projection is implemented as part of the generator network using a soft least-squares penalty during training of the generator network.
  • the reconstructed image ⁇ circumflex over (x) ⁇ is applied to the generator network to obtain a second aliasing artifact-reduced image, and the second aliasing artifact-reduced image is projected onto the space of consistent images to obtain a final reconstructed image.
  • FIG. 1 is a flowchart illustrating a method for image reconstruction according to an embodiment of the invention.
  • FIG. 2 is a schematic block diagram illustrating a Generator ResNet architecture with residual blocks (RB) according to an embodiment of the invention.
  • FIG. 3 is a schematic block diagram illustrating a Discriminator multilayer CNN architecture according to an embodiment of the invention.
  • the reconstruction goal is to quickly recover the image x after collecting the undersampled measurements y.
  • a flowchart illustrating the steps of the m reconstruction method is shown in FIG. 1 ., where step 100 performs an MRI acquisition to obtain the under-sampled k-space measurement data 102 .
  • an initial image estimate 106 denoted ⁇ tilde over (x) ⁇ , that is calculated in step 104 by a linear transform from undersampled measurements y and possibly contains aliasing artifacts.
  • the subsequent reconstruction can then be envisioned as artifact suppression that is modelled as projection onto the manifold of high-quality images. Learning the corresponding manifold is accomplished via generative adversarial networks.
  • the inverse imaging problem is to find a solution at the intersection of a subspace defined by the acquisition model and the image manifold.
  • the technique In order to effectively learn the image manifold from the available (limited number of) training samples, the technique must ensure the trained manifold contains plausible MR images and must ensure the points on the manifold are data consistent, i.e., y ⁇ A(x), ⁇ x ⁇ M.
  • Standard GAN includes a tandem network of G and D networks.
  • the G network then projects ⁇ tilde over (x) ⁇ onto the low-dimensional manifold M containing the high-quality images X.
  • x ⁇ denote the output 110 of G.
  • the G net 108 is trained to learn to project to the low-dimensional manifold and achieve realistic reconstruction.
  • the output 110 of G may not be consistent with the data.
  • C ⁇ x:y ⁇ A(x) ⁇
  • ⁇ circumflex over (x) ⁇ ⁇ circumflex over (x) ⁇ .
  • LS soft least-squares
  • the final reconstructed image 116 passes through the discriminator network 118 that tries to output one if ⁇ circumflex over (x) ⁇ X, and zero otherwise 122 .
  • the G net 108 learns realistic reconstruction, such that D net 118 cannot always perfectly assign the right labels 122 to the real (fully-sampled) image 120 and “fake” (recovered) images 116 .
  • Training the networks amounts to playing a game with conflicting objectives between the generator G and the discriminator D.
  • the D network aims to score one for the real gold-standard images x, and zero for the fake/reconstructed images ⁇ circumflex over (x) ⁇ reconstructed by G.
  • the G network also aims to map the input aliased image ⁇ tilde over (x) ⁇ to a fake image x ⁇ that looks so realistic and plausible that it can fool D.
  • Various strategies to reach the equilibrium mostly differ in terms of the cost function adopted for G and D networks.
  • a standard GAN uses a sigmoid cross-entropy loss that leads to vanishing gradients which renders the training unstable, and as a result it suffers from severe degrees of mode collapse.
  • the standard GAN tends to pull samples away from the decision boundary, which can introduce non-realistic images.
  • Such images can hallucinate image features, and thus are not reliable for diagnostic decisions.
  • the present method adopts an LS cost for the discriminator decision. In essence, the LS cost penalizes the decision variables without any nonlinear transformation, and as a result it tends to pull the generated samples toward the decision boundary.
  • GANs may overemphasize high-frequency texture, and thus ignore important diagnostic image content.
  • the G net is preferably trained using a supervised l 1 /l 2 cost as well.
  • a supervised l 1 /l 2 cost can properly penalize the noise and stabilize the training.
  • the smooth l 2 -cost preserves the main structure and leads to a stable training at the expense of introducing blurring artifacts.
  • the non-smooth l 1 -cost however may not be as stable as l 2 in training, but it can better discard the low-intensity noise and achieve better solutions.
  • a mixture of LSGAN and l 1 /l 2 cost is preferably used to train the generator.
  • the overall procedure aims to jointly minimize the expected discriminator cost min ⁇ d E x [(1 ⁇ D(x; ⁇ d )) 2 ]+E y [D(G( ⁇ tilde over (x) ⁇ ; ⁇ g ); ⁇ d ) 2 ], (P1.1)
  • ⁇ d and ⁇ g are parameters of the discriminator network D, and generator network G, respectively, and the minimum is taken over ⁇ d , and the expected generator cost min ⁇ g E y [ ⁇ y ⁇ AG( ⁇ tilde over (x) ⁇ ; ⁇ g ) ⁇ 2 ]+ ⁇ E x,y [ ⁇ x ⁇ G( ⁇ tilde over (x) ⁇ ; ⁇ g ) ⁇ 1,2 ]+ ⁇ E y [(1 ⁇ D(G( ⁇ tilde over (x) ⁇ ; ⁇ g ); ⁇ d )) 2 ] (P1.2)
  • the parameters ⁇ d and ⁇ g are usually weights in the CNNs and are trained based on the dataset by optimizing the above and below cost functions. Usually back-projection is used.
  • the first LS data fidelity term in (P1.2) is also a soft penalty to ensure the direct output of G network is approximately data consistent as mentioned before. Tuning parameters ⁇ and ⁇ also control the balance between manifold projection, noise suppression, and data consistency.
  • the conventional CS scheme which involves an iterative optimization algorithm to solve for the l 1 /l 2 -regularized LS cost, the optimization only happens in training and the optimized weights in the network can generalize to any future samples.
  • the learned generator can be immediately applied to new test data to retrieve the image in real time. Even in the presence of LS data consistency and l 1 /l 2 penalty, the distribution achieved by G network can coincide with the true data distribution, which ensures the reconstruction is regularized to be as designed for this manifold learning scheme: both data consistent and MRI realistic.
  • the input 200 and output 216 of the generator G are complex-valued images of the same size, where the real and imaginary components are considered as two separate channels.
  • the input image ⁇ tilde over (x) ⁇ is simply an initial estimate obtained, e.g., via zero-filling, which undergoes aliasing artifacts. After convolving the input channels with different kernels, they are added up in the next layer. All network kernels are assumed real-valued.
  • a deep residual network (ResNet) is used for the generator that contains 5 residual blocks 202 through 204 .
  • each block has two convolutional layers with small 3 ⁇ 3 kernels and 128 feature maps that are followed by batch normalization (BN) and rectified linear unit (ReLU) activation.
  • the five residual blocks are followed by three convolutional layers with map size 1 ⁇ 1.
  • the first layer has a convolution 206 and ReLU activation 208 .
  • the second layer has a convolution 210 and ReLU activation 212
  • the last layer has convolution 214 but uses no activation to return two output channels corresponding the real and imaginary image parts.
  • the G network learns the projection onto the manifold of realistic MR images.
  • the manifold dimension is controlled by the number of residual blocks (RB), feature maps, stride size, and the size of discriminator.
  • RB residual blocks
  • n and k refer to number of feature maps and filter size, respectively.
  • the D network takes the magnitude of the complex-valued output of the G net and data consistency projection as an input 300 .
  • it is composed of a series of convolutional layers, where in all layers except the last one, a convolution operation is followed by batch normalization, and subsequently by ReLU activation. No pooling is used.
  • Layer 1 has a convolution 302 followed by batch normalization 304 and ReLU activation 306 .
  • layer 4 has a convolution 308 followed by batch normalization 310 and ReLU activation 312
  • layer 5 has a convolution 314 followed by batch normalization 316 and ReLU activation 318 .
  • Layer 6 is a convolution 320 only, and layer 7 is an average 322 .
  • the number of feature maps is doubled at each successive layer from 8 to 64 , while at the same time convolution with stride 2 is used to reduce the image resolution.
  • Kernel size 3 ⁇ 3 is adopted for the first 4 layers, while layers 5 and 6 use kernel size 1 ⁇ 1.
  • the last layer 7 simply averages out the features to form the decision variable for binary classification. No soft-max operation is used.
  • the variables n, k, and s in the figure refer to number of feature maps, filter size, and stride size, respectively.
  • abdominal image volumes were acquired for 350 pediatric patients after gadolinium-based contrast enhancement.
  • Each 3D volume includes from 150 to 220 axial slices of size 256 ⁇ 128 with voxel resolution 1.07 ⁇ 1.12 ⁇ 2.4 mm.
  • Axial slices are used as input images for training a neural network.
  • 340 patients 50,000 2D slices
  • 10 patients (1,920 2D slices) for testing. All scans were acquired on a 3T MRI scanner (GE MR750).
  • the G cost mainly pertains to the last term which shows how well the G net can fool the D net.
  • GANCS Reconstructing 30 slices per second makes GANCS a suitable choice for real-time imaging.
  • GANCS with l 1 -cost alone achieves the best performance.
  • GANCS with proper l 1 -cost mixing can achieve good performance with a marginally decrease from GANCS with l 1 -cost alone.
  • the perceptual quality of resulting images was confirmed by radiologist opinion scores (ROS).
  • ROS radiologist opinion scores
  • the images retrieved by GANCS attain the highest score that is as good as the gold-standard.
  • GANCS may create hallucinated images
  • two abnormal patients with missing left and right kidneys were scanned and images reconstructed, where the training data does not include patients with similar abnormalities. It was confirmed that GANCS misses/introduces no structures or edges.
  • Prediction (generalization) performance of the deep learning model heavily depends on the amount of training data. This becomes more important when dealing with scarce medical data that are typically not accessible in large scales due to privacy concerns and institutional regulations.
  • To address this question we examined an evaluation scenario to assess the reconstruction performance for a fixed test dataset, for variable number of patients used for training.
  • GANCS captures the signal model, and therefore it can easily accommodate different sampling trajectories. Also note that, if more datasets are available for training, we can further improve the model performance by increasing model complexity. Further study of the number of patients needed for other random sampling schemes and different network models is an important question that is a focus of our current research.
  • the discriminator plays a role like a radiologist that scores the quality of images created by the generator.
  • D learns to correctly discern the real fully-sampled images from the fake ones, where the fake ones become quite realistic as training progresses. It is thus insightful to understand image features that drive the quality score.
  • we compared original images with heat maps of feature maps of D net at hidden convolutional layers. This demonstrated that, after learning from tens of thousands of generated MRI images by the G network together with the corresponding gold-standard ones, where different organs are m present, the D network learns to focus on certain regions of interest that are more susceptible to artifacts.
  • a CS framework leverages the historical data for rapid and high diagnostic-quality image reconstruction from highly undersampled MR measurements.
  • a low-dimensional manifold is learned where the reconstructed images have not only superior sharpness and diagnostic quality, but also consistent with both the real MRI data and the acquisition model.
  • a neural network scheme based on LSGANs and l 1 /l 2 costs is trained, where a generator is used to map a readily obtainable undersampled image to a realistic-looking one consistent with the measurements, while a discriminator network is trained jointly to score the quality of the resulting image.
  • the overall training acts as a game between generator and discriminator that makes them more intelligent at reconstruction and quality evaluation.

Abstract

A method for diagnostic imaging includes measuring undersampled data y with a diagnostic imaging apparatus; linearly transforming the undersampled data y to obtain an initial image estimate {tilde over (x)}; applying the initial image estimate {tilde over (x)} as input to a generator network to obtain an aliasing artifact-reduced image x̆ as output of the generator network, where the aliasing artifact-reduced image x̆ is a projection onto a manifold of realistic images of the initial image estimate {tilde over (x)}; and performing an acquisition signal model projection of the aliasing artifact-reduced x̆ onto a space of consistent images to obtain a reconstructed image {circumflex over (x)} having suppressed image artifacts.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This application claims priority from U.S. Provisional Patent Application 62/678,663 filed May, 31, 2018, which is incorporated herein by reference.
STATEMENT REGARDING FEDERALLY SPONSORED RESEARCH OR DEVELOPMENT
None.
FIELD OF THE INVENTION
The present invention relates generally to magnetic resonance imaging (MRI). More specifically, it relates to techniques for MRI reconstruction.
BACKGROUND OF THE INVENTION
Due to its superb soft tissue contrast, magnetic resonance imaging (MRI) is a major imaging modality in clinical practice. MRI image reconstruction is typically an ill-posed linear inverse problem demanding time and resource intensive computations that substantially trade off accuracy for speed in real-time imaging.
Real-time MRI acquisition, reconstruction and visualization is of paramount importance for diagnostic and therapeutic guidance. Interventional and image-guided therapies as well as interactive diagnostic tasks need rapid image preparation within a few milliseconds. This is hindered, however, by the slow acquisition process, taking several minutes to acquire clinically acceptable images. Inefficient acquisition becomes more pronounced for high-resolution and volumetric images. One possible solution is to decrease the scan duration through significant undersampling. However, such undersampling leads to a seriously ill-posed linear inverse reconstruction problem.
To render the MRI reconstruction well-posed, conventional compressed sensing (CS) incorporates the prior about the inherent low dimensionality of images by means of sparsity regularization in a proper transform domain such as Wavelet (WV), or, finite differences (or Total Variation, TV). This typically demands running iterative algorithms, for solving non-smooth optimization programs, that are time and resource intensive, and thus not affordable for real-time MRI visualization. Moreover, the sparsity assumption is rather universal and although it is useful for certain image types, it is oblivious to the inherent latent structures that are specific to each dataset.
A few attempts have been recently carried out to speed up medical image reconstruction by leveraging historical patient data, e.g., by training a network that learns the relation map between an initial aliased image and the gold-standard one. Although reconstruction speeds up, these techniques suffer from blurring and aliasing artifacts. This is mainly because they adopt a pixel-wise l1/l2 cost for training that is oblivious to structured artifacts and high-frequency texture details. These details, however, are crucial for accurately making diagnostic decisions. In addition, these techniques lack any mechanism that ensures the retrieved images are consistent with the measurements.
Deep neural networks have been used to learn image prior or sparsifying transform from historical data in order to solve a nonlinear system using iterative optimization algorithms as in the conventional CS methods. While improving the reconstruction performance, these methods incur high computational cost due to several iterations for finding the optimal reconstruction.
BRIEF SUMMARY OF THE INVENTION
To cope with these challenges, the present disclosure provides a CS framework that applies generative adversarial networks (GAN) to modeling a (low-dimensional) manifold of high-quality MR images. Leveraging a mixture of least-squares (LS) GANs and pixel-wise l1/l2 cost, a deep residual network with skip connections is trained as the generator that learns to remove the aliasing artifacts by projecting onto the image manifold. This least-squares generative adversarial network (LSGAN) learns the texture details, while l1/l2 suppresses the high-frequency noise. A discriminator network, which is a multilayer convolutional neural network (CNN), plays the role of a perceptual cost that is then jointly trained based on high quality MR images to score the quality of retrieved images.
In the operational phase, an initial aliased estimate (e.g., simply obtained by zero-filling) is propagated into the trained generator to output the desired reconstructions, that demands very low computational overheads.
Extensive evaluations on a large contrast-enhanced MR dataset of images rated by expert radiologists corroborate that this generative adversarial networks compressed sensing (GANCS) technique recovers higher quality images with fine texture details relative to conventional CS schemes as well as pixel-wise training schemes. In addition, GANCS performs reconstruction under a few milliseconds, m which is two orders of magnitude faster than state-of-the-art CS-MRI schemes. Moreover, the perceptual quality metric offered by the trained discriminator network can significantly facilitate the radiologists' quality assurance tasks.
In other contexts, generative adversarial networks (GANs) have proven successful in modeling distributions (low-dimensional manifolds) and generating natural images (high-dimensional data) that are perceptually appealing. Despite the success of GANs for local image restoration such as super-resolution and inpainting, due to fundamental differences, GANs have not been considered for correcting the aliasing artifacts in biomedical image reconstruction tasks. In essence, aliasing artifacts (e.g., in MRI) emanate from data undersampling in the frequency domain that globally impacts the entire space domain image. Nevertheless, the present approach uses GANs for MRI reconstruction. The approach uses GANs for modeling low-dimensional manifold of high-quality MR images. The images lying on the manifold are not however necessarily consistent with the observed (undersampled) data. As a result, the reconstruction deals with modeling the intersection of the image manifold and subspace of data-consistent images; such a space is an affine subspace for linear measurements. To this end, the present GANCS approach adopts a tandem network of a generator (G), an affine projection operator (A), and a discriminator (D). The generator aims to create gold-standard images from the complex-valued aliased inputs using a deep residual network (ResNet) with skip connections which retain high resolution information. The data-consistency projection builds upon the (known) signal model and performs an affine projection onto the space of data consistent images. The D network is a multilayer convolutional neural network (CNN) that is trained using both the “fake” images created by G, and the corresponding gold-standard ones, and aims to correctly distinguish fake from real. Least-squares GANs (LSGANs) is used due to their stability properties. Alternatively, other GANs frameworks may be used, where usually the differences are just cost functions, e.g., Wasserstein GAN (W-GAN), Cycle-GAN, BigGAN, or StarGAN. To control the high-frequency texture details returned by LSGANs, and to further improve the training stability, we partially use the pixel-wise l1 and l2 costs for training the generator.
The GANCS results have almost similar quality to the gold-standard (fully-sampled) images, and are superior in terms of diagnostic quality relative to the existing alternatives including conventional iterative CS and deep learning based methods that solely adopt the pixel-wise l2-based and l1-based criteria. Moreover, the reconstruction only takes around 30 ms, which is two orders of magnitude faster than state-of-the-art conventional CS toolboxes.
In one aspect, the invention provides a method for diagnostic imaging comprising: measuring undersampled data y with a diagnostic imaging apparatus; linearly transforming the undersampled data y to obtain an initial image estimate {tilde over (x)}; applying the initial image estimate {tilde over (x)} as input to a generator network to obtain an aliasing artifact-reduced image x̆ as output of the generator network, wherein the aliasing artifact-reduced image x̆ is a projection onto a manifold of realistic images of the initial image estimate {tilde over (x)}; and performing an acquisition signal model projection of the aliasing artifact-reduced image x̆ onto a space of consistent images to obtain a reconstructed image {circumflex over (x)} having suppressed image artifacts.
The diagnostic imaging apparatus may be, for example, an MRI scanner, and the undersampled data is k-space data.
Linearly transforming the undersampled data y may comprise, for example, zero padding the undersampled data y, or finding an approximate zero-filling reconstruction from the undersampled data y.
Preferably, the generator network is trained to learn the projection onto the manifold of realistic images using a set of training images X and corresponding set of undersampled measurements Y using least-squares generative adversarial network techniques in tandem with a discriminator network to learn texture details and supervised cost function to control high-frequency noise.
The supervised cost function may comprise a mixture of smooth l2 cost and non-smooth l1 cost.
The discriminator network may be a multilayer deep convolutional neural network.
The discriminator network may be trained using least-squares cost for a discriminator decision.
The generator network may be a deep residual network with skip connections.
In some embodiments, performing the acquisition signal model projection is implemented as part of the generator network using a soft least-squares penalty during training of the generator network.
In some embodiments, the reconstructed image {circumflex over (x)} is applied to the generator network to obtain a second aliasing artifact-reduced image, and the second aliasing artifact-reduced image is projected onto the space of consistent images to obtain a final reconstructed image.
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
FIG. 1 is a flowchart illustrating a method for image reconstruction according to an embodiment of the invention.
FIG. 2 is a schematic block diagram illustrating a Generator ResNet architecture with residual blocks (RB) according to an embodiment of the invention.
FIG. 3 is a schematic block diagram illustrating a Discriminator multilayer CNN architecture according to an embodiment of the invention.
DETAILED DESCRIPTION OF THE INVENTION
Consider a generic MRI acquisition model that forms an image x∈RN from k-space projection data y∈RM
y=A(x)+v   (1)
where the (possibly) nonlinear map A: CN→CM encompasses the effects of sampling, coil sensitivities, and the discrete Fourier transform (DFT). The error term v∈RM captures the noise and unmodeled dynamics. We assume the unknown (complex-valued) image x lies in a low-dimensional manifold, M. No information is known about the manifold besides the K training samples X={xk} drawn from it with the corresponding K (possibly) noisy observations Y={yk}. The data {X,Y} can be obtained for instance from the K past patients in the dataset that have been already scanned for a sufficient time, and their high-quality reconstruction is available. Given the training data {X,Y}, the reconstruction goal is to quickly recover the image x after collecting the undersampled measurements y. A flowchart illustrating the steps of the m reconstruction method is shown in FIG. 1., where step 100 performs an MRI acquisition to obtain the under-sampled k-space measurement data 102.
Instead of relying on simple sparsity assumption of X, the approach is to automate the image recovery by learning the nonlinear inversion map x=A−1(y) from the historical training data {X,Y}. To this end, we begin with an initial image estimate 106, denoted {tilde over (x)}, that is calculated in step 104 by a linear transform from undersampled measurements y and possibly contains aliasing artifacts. The initial image estimate {tilde over (x)}=A(y) may be obtained via zero-filling the missing k-space components, which is the least square solution for data-consistency, and running a single iteration of conjugate gradient. The subsequent reconstruction can then be envisioned as artifact suppression that is modelled as projection onto the manifold of high-quality images. Learning the corresponding manifold is accomplished via generative adversarial networks.
The inverse imaging problem is to find a solution at the intersection of a subspace defined by the acquisition model and the image manifold. In order to effectively learn the image manifold from the available (limited number of) training samples, the technique must ensure the trained manifold contains plausible MR images and must ensure the points on the manifold are data consistent, i.e., y≈A(x), ∀x∈M.
Alternating Projection with GANs for Plausible Reconstruction To ensure plausibility of the reconstruction, we use GANs. Standard GAN includes a tandem network of G and D networks. The initial image estimate {tilde over (x)}=A†(y) is applied as the input to the G network 108. The G network then projects {tilde over (x)} onto the low-dimensional manifold M containing the high-quality images X. Let x̆ denote the output 110 of G. As will be clear later, the G net 108 is trained to learn to project to the low-dimensional manifold and achieve realistic reconstruction.
Affine Projection and Soft Penalty for Data-Consistency
The output 110 of G may not be consistent with the data. To tackle this issue, G is followed by another layer 112 that projects the output x̆ of G onto the set of data-consistent images, namely C={x:y≈A(x)} to obtain a reconstructed image 114, denoted {circumflex over (x)}. For Cartesian grid with the linear acquisition model y=Ax, the projection is expressible as {circumflex over (x)}=Ay+PNx̆, where PN=(I−AA) resembles projection onto the null space of A. Alternatively, one can impose data consistency to the output of G through a soft least-squares (LS) penalty when training the G network, as will be seen later.
To further ensure that the reconstructed image {circumflex over (x)} falls in the intersection of the manifold M and the set of data-consistent images C, we can perform multiple back-and-forth projections. The network structure in FIG. 1 can then be extended by repeating the G network and PC(⋅) in serial for a few times, or iterating steps 108 and 112 with feedback of reconstructed image {circumflex over (x)}, as indicated by the dotted line in the figure. For simplicity of exposition, we discuss here a single back-and-forth projection. However, repeating with multiple projections shows significant performance improvement.
During training, the final reconstructed image 116 passes through the discriminator network 118 that tries to output one if {circumflex over (x)}∈X, and zero otherwise 122. The G net 108 learns realistic reconstruction, such that D net 118 cannot always perfectly assign the right labels 122 to the real (fully-sampled) image 120 and “fake” (recovered) images 116.
Least-Squares GANs for Stable Training
Training the networks amounts to playing a game with conflicting objectives between the generator G and the discriminator D. The D network aims to score one for the real gold-standard images x, and zero for the fake/reconstructed images {circumflex over (x)} reconstructed by G. On the other hand, the G network also aims to map the input aliased image {tilde over (x)} to a fake image x̆ that looks so realistic and plausible that it can fool D. Various strategies to reach the equilibrium mostly differ in terms of the cost function adopted for G and D networks. A standard GAN uses a sigmoid cross-entropy loss that leads to vanishing gradients which renders the training unstable, and as a result it suffers from severe degrees of mode collapse. In addition, for the generated images classified as the real with high confidence (i.e., with large decision variable), no penalty is incurred. Hence, the standard GAN tends to pull samples away from the decision boundary, which can introduce non-realistic images. Such images can hallucinate image features, and thus are not reliable for diagnostic decisions. The present method adopts an LS cost for the discriminator decision. In essence, the LS cost penalizes the decision variables without any nonlinear transformation, and as a result it tends to pull the generated samples toward the decision boundary.
Mixed Costs to Avoid High Frequency Noise
One issue with GANs is that they may overemphasize high-frequency texture, and thus ignore important diagnostic image content. In order to discard the high-frequency noise and avoid hallucination, the G net is preferably trained using a supervised l1/l2 cost as well. Such mixture with pixel-wise costs can properly penalize the noise and stabilize the training. In particular, the smooth l2-cost preserves the main structure and leads to a stable training at the expense of introducing blurring artifacts. The non-smooth l1-cost however may not be as stable as l2 in training, but it can better discard the low-intensity noise and achieve better solutions. All in all, to reveal fine texture details while discarding noise, a mixture of LSGAN and l1/l2 cost is preferably used to train the generator. The overall procedure aims to jointly minimize the expected discriminator cost
minΘdEx[(1−D(x;Θd))2]+Ey[D(G({tilde over (x)};Θg);Θd)2],   (P1.1)
where Θd and Θg are parameters of the discriminator network D, and generator network G, respectively, and the minimum is taken over Θd, and the expected generator cost
minΘgEy[∥y−AG({tilde over (x)};Θg2]+ηEx,y[∥x−G({tilde over (x)};Θg)∥1,2]+λEy[(1−D(G({tilde over (x)};Θg);Θd))2]  (P1.2)
where the minimum is taken over Θg, and E[⋅] is the statistical expectation operator, and ∥⋅∥1,2 denotes a convex combination of the element-wise l1-norm and l2-norm with non-negative weights η1 and η2 respectively, such that η12=η. The parameters Θd and Θg are usually weights in the CNNs and are trained based on the dataset by optimizing the above and below cost functions. Usually back-projection is used. The first LS data fidelity term in (P1.2) is also a soft penalty to ensure the direct output of G network is approximately data consistent as mentioned before. Tuning parameters λ and η also control the balance between manifold projection, noise suppression, and data consistency.
Using the cost (P1.2), taking initial estimation {tilde over (x)} as input, the generator reconstructs improved x̆=G({tilde over (x)};Θg) from k-space measurement y using the expected regularized-LS estimator, where the regularization is not based on sparsity but learned from training data via LSGAN and l1-net. Different from the conventional CS scheme, which involves an iterative optimization algorithm to solve for the l1/l2-regularized LS cost, the optimization only happens in training and the optimized weights in the network can generalize to any future samples. The learned generator can be immediately applied to new test data to retrieve the image in real time. Even in the presence of LS data consistency and l1/l2 penalty, the distribution achieved by G network can coincide with the true data distribution, which ensures the reconstruction is regularized to be as designed for this manifold learning scheme: both data consistent and MRI realistic.
Network Architecture for GANCS
The architectures of G and D nets are now described in relation to FIG. 2 and FIG. 3.
Residual Networks for the Generator
The input 200 and output 216 of the generator G are complex-valued images of the same size, where the real and imaginary components are considered as two separate channels. The input image {tilde over (x)} is simply an initial estimate obtained, e.g., via zero-filling, which undergoes aliasing artifacts. After convolving the input channels with different kernels, they are added up in the next layer. All network kernels are assumed real-valued. A deep residual network (ResNet) is used for the generator that contains 5 residual blocks 202 through 204. As shown in the detail 218 for block 202, each block has two convolutional layers with small 3×3 kernels and 128 feature maps that are followed by batch normalization (BN) and rectified linear unit (ReLU) activation. The five residual blocks are followed by three convolutional layers with map size 1×1. The first layer has a convolution 206 and ReLU activation 208. Similarly, the second layer has a convolution 210 and ReLU activation 212, while the last layer has convolution 214 but uses no activation to return two output channels corresponding the real and imaginary image parts.
The G network learns the projection onto the manifold of realistic MR images. The manifold dimension is controlled by the number of residual blocks (RB), feature maps, stride size, and the size of discriminator. In the figure, n and k refer to number of feature maps and filter size, respectively.
Convolutional Neural Networks for Discriminator
The D network takes the magnitude of the complex-valued output of the G net and data consistency projection as an input 300. In a preferred embodiment, it is composed of a series of convolutional layers, where in all layers except the last one, a convolution operation is followed by batch normalization, and subsequently by ReLU activation. No pooling is used. Layer 1 has a convolution 302 followed by batch normalization 304 and ReLU activation 306. Similarly, layer 4 has a convolution 308 followed by batch normalization 310 and ReLU activation 312, and layer 5 has a convolution 314 followed by batch normalization 316 and ReLU activation 318. Layer 6 is a convolution 320 only, and layer 7 is an average 322. For the first four layers, the number of feature maps is doubled at each successive layer from 8 to 64, while at the same time convolution with stride 2 is used to reduce the image resolution. Kernel size 3×3 is adopted for the first 4 layers, while layers 5 and 6 use kernel size 1×1. The last layer 7 simply averages out the features to form the decision variable for binary classification. No soft-max operation is used. The variables n, k, and s in the figure refer to number of feature maps, filter size, and stride size, respectively.
Evaluations
Effectiveness of the GANCS scheme was assessed in this section for a single-coil MR acquisition model with Cartesian sampling. For the n-th patient, the acquired k-space data is denoted yi,j (n)=[F(Xn)]i,j+vi,j (n), where (i,j)∈Ω. We adopt the single coil model for demonstration purposes, but extension to multi-coil MRI acquisition model is straightforward by simply updating the signal model. Sampling set Ω indexes the sampled Fourier coefficients. As it is conventionally performed with CS MRI, we select Ω based on a variable density sampling with radial view ordering that tends to sample more low frequency components from the center of k-space. Different undersampling rates (5 and 10) are chosen throughout the experiment. The input zero-filling (ZF) image {tilde over (x)} is simply generated using inverse 2D FT of the sampled k-space, which is severely contaminated by aliasing artifacts. Input images are normalized to have the maximum magnitude unity per image.
Adam optimizer is used with the momentum parameter β=0.9, mini-batch size 4, and initial learning rate 10-6 that is halved every 10,000 iterations. Training is performed with TensorFlow interface on a NVIDIA Titan X Pascal GPU, 12 GB RAM. We allow 20 epochs that takes around 10 hours for training.
For the dataset, abdominal image volumes were acquired for 350 pediatric patients after gadolinium-based contrast enhancement. Each 3D volume includes from 150 to 220 axial slices of size 256×128 with voxel resolution 1.07×1.12×2.4 mm. Axial slices are used as input images for training a neural network. 340 patients (50,000 2D slices) are considered for training, and 10 patients (1,920 2D slices) for testing. All scans were acquired on a 3T MRI scanner (GE MR750).
Training Convergence
Stable convergence of GANCS was confirmed by considering evolution of different components of G and D costs for training over batches (size 4), with η=0.025 and λ=0.975 as an example to emphasize the GAN loss in training. According to (P1.2), the G cost mainly pertains to the last term which shows how well the G net can fool the D net. The D cost also includes two components based m on (P1.1) associated with the classification performance for both real and fake images. It was confirmed that all cost components decrease, and after about 5,000 batches it reaches the equilibrium cost 0.25. This implies that upon convergence the G-net images become so realistic that the D-net will behave as a flipping coin, i.e., D({circumflex over (x)})=½. In this setting with a hard affine projection layer no data-consistency cost is incurred.
It is also worth mentioning that to improve the convergence stability of GANCS, and to ensure the initial distributions of fake and real images are overlapping, we trained with pure l1 cost (η=1, λ=0) at the beginning and then gradually switch to the mixture loss intended.
Quantitative Image Evaluation and Comparison
For comparison, images were reconstructed by various methods with 5-fold and 10-fold undersampling of k-space. Specifically, the gold-standard image, was compared with images reconstructed by GANCS with l1-cost (η=0.975, λ=0.025), GANCS with l1-cost (η=1, λ=0), GANCS with l2-cost (η=1, λ=0), and CS-WV. For 5-fold undersampling, the ZF reconstruction is also included. CS reconstruction is performed using the Berkeley Advanced Reconstruction Toolbox (BART), where the tuning parameters are optimized for the best performance. These image comparisons confirmed that GANCS with l1 cost (η=0.975, λ=0.025) returns the sharpest images with highest contrast and texture details that can reveal the small anatomical details. Images retrieved by GANCS with l2-cost alone results in overly smooth textures as the l2-cost encourages finding pixel-wise averages of all plausible solutions. Also, images obtained using GANCS with l1 alone look more realistic than the l2 counterpart. The reconstructed images however are not as sharp as the GANCS (η=0.975, λ=0.025) which leverages both l1-net and GANs. We have observed that using m GAN alone (η=0, λ=1), the retrieved images are quite sharp with a high-frequency noise present over the image that can distort the image structure. It turns out that including the l1 cost during training behaves as a low-pass filter to discard the high-frequency noises, while still achieving reasonably sharp images. It is also evident that CS-WV introduces blurring artifacts. We also tested CS-TV, but CS-WV is observed to consistently outperform CS-TV, and thus we choose CS-WV as the representative for CS-MRI.
Reconstructing 30 slices per second makes GANCS a suitable choice for real-time imaging. In terms of SNR and SSIM, GANCS with l1-cost alone achieves the best performance. GANCS with proper l1-cost mixing can achieve good performance with a marginally decrease from GANCS with l1-cost alone.
Diagnostic Quality Assessment
The perceptual quality of resulting images was confirmed by radiologist opinion scores (ROS). The images retrieved by GANCS attain the highest score that is as good as the gold-standard.
Performance on Abnormal Cases
To address the concern that GANCS may create hallucinated images, two abnormal patients with missing left and right kidneys were scanned and images reconstructed, where the training data does not include patients with similar abnormalities. It was confirmed that GANCS misses/introduces no structures or edges.
Number of Patients for Prediction
Prediction (generalization) performance of the deep learning model heavily depends on the amount of training data. This becomes more important when dealing with scarce medical data that are typically not accessible in large scales due to privacy concerns and institutional regulations. To address this question we examined an evaluation scenario to assess the reconstruction performance for a fixed test dataset, for variable number of patients used for training. The test measured SNR versus the number of training patients for the GANCS scheme with η=0.975, λ=0.025. As the number of patients increased from 1 to 130, a noticeable SNR gain was observed. The performance gain then gradually saturates as the number of patients reaches 150. It thus seems with 150 or more patients we can take full advantage of both learning from historical data and the complexity of the networks. Recall that a fixed sampling mask is used for training and testing. GANCS, however, captures the signal model, and therefore it can easily accommodate different sampling trajectories. Also note that, if more datasets are available for training, we can further improve the model performance by increasing model complexity. Further study of the number of patients needed for other random sampling schemes and different network models is an important question that is a focus of our current research.
Discriminator Interpretation
As suggested by the training strategy, the discriminator plays a role like a radiologist that scores the quality of images created by the generator. During adversarial training, D learns to correctly discern the real fully-sampled images from the fake ones, where the fake ones become quite realistic as training progresses. It is thus insightful to understand image features that drive the quality score. To this end, we compared original images with heat maps of feature maps of D net at hidden convolutional layers. This demonstrated that, after learning from tens of thousands of generated MRI images by the G network together with the corresponding gold-standard ones, where different organs are m present, the D network learns to focus on certain regions of interest that are more susceptible to artifacts.
CONCLUSIONS
A CS framework is provided that leverages the historical data for rapid and high diagnostic-quality image reconstruction from highly undersampled MR measurements. A low-dimensional manifold is learned where the reconstructed images have not only superior sharpness and diagnostic quality, but also consistent with both the real MRI data and the acquisition model. To this end, a neural network scheme based on LSGANs and l1/l2 costs is trained, where a generator is used to map a readily obtainable undersampled image to a realistic-looking one consistent with the measurements, while a discriminator network is trained jointly to score the quality of the resulting image. The overall training acts as a game between generator and discriminator that makes them more intelligent at reconstruction and quality evaluation.

Claims (11)

The invention claimed is:
1. A method for diagnostic imaging comprising:
measuring undersampled data y with a diagnostic imaging apparatus;
linearly transforming the undersampled data y to obtain an initial image estimate {tilde over (x)};
applying the initial image estimate {tilde over (x)} as input to a generator network to obtain an aliasing artifact-reduced image x̆ as output of the generator network,
wherein the aliasing artifact-reduced image x̆ is a projection onto a manifold of realistic images of the initial image estimate {tilde over (x)};
performing an acquisition signal model projection of the aliasing artifact-reduced image x̆ onto a space of consistent images to obtain a reconstructed image {circumflex over (x)} having suppressed image artifacts.
2. The method of claim 1 wherein the diagnostic imaging apparatus is an MRI scanner, and wherein the undersampled data is k-space data.
3. The method of claim 1 wherein linearly transforming the undersampled data y comprises zero padding the undersampled data y.
4. The method of claim 1 wherein linearly transforming the undersampled data y comprises finding an approximate zero-filling reconstruction from the undersampled data y.
5. The method of claim 1 wherein the generator network is trained to learn the projection onto the manifold of realistic images using a set of training images X and corresponding set of undersampled measurements Y using least-squares generative adversarial network techniques in tandem with a discriminator network to learn texture details and supervised cost function to control high-frequency noise.
6. The method of claim 5 wherein the supervised cost function comprises a mixture of smooth l2 cost and non-smooth l1 cost.
7. The method of claim 5 wherein the discriminator network is a multilayer deep convolutional neural network.
8. The method of claim 5 wherein the discriminator network is trained using a least squares cost for a discriminator decision.
9. The method of claim 1 wherein the generator network is a deep residual network with skip connections.
10. The method of claim 1 wherein performing an acquisition signal model projection is implemented as part of the generator network using a soft least-squares penalty during training of the generator network.
11. The method of claim 1 wherein the reconstructed image {circumflex over (x)} is applied to the generator network to obtain a second aliasing artifact-reduced image, and the second aliasing artifact-reduced image is projected onto the space of consistent images to obtain a final reconstructed image.
US16/427,599 2018-05-31 2019-05-31 MRI reconstruction using deep learning, generative adversarial network and acquisition signal model Active 2039-07-01 US10859657B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/427,599 US10859657B2 (en) 2018-05-31 2019-05-31 MRI reconstruction using deep learning, generative adversarial network and acquisition signal model

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201862678663P 2018-05-31 2018-05-31
US16/427,599 US10859657B2 (en) 2018-05-31 2019-05-31 MRI reconstruction using deep learning, generative adversarial network and acquisition signal model

Publications (2)

Publication Number Publication Date
US20190369191A1 US20190369191A1 (en) 2019-12-05
US10859657B2 true US10859657B2 (en) 2020-12-08

Family

ID=68692891

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/427,599 Active 2039-07-01 US10859657B2 (en) 2018-05-31 2019-05-31 MRI reconstruction using deep learning, generative adversarial network and acquisition signal model

Country Status (1)

Country Link
US (1) US10859657B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11170543B2 (en) * 2020-01-13 2021-11-09 The Board Of Trustees Of The Leland Stanford Junior University MRI image reconstruction from undersampled data using adversarially trained generative neural network

Families Citing this family (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110249365B (en) * 2017-11-10 2023-05-30 上海联影医疗科技股份有限公司 System and method for image reconstruction
US10956635B1 (en) * 2019-12-04 2021-03-23 David Byron Douglas Radiologist-assisted machine learning with interactive, volume subtending 3D cursor
US11238197B1 (en) * 2018-04-03 2022-02-01 David Byron Douglas Generating a 3D dataset containing a simulated surgical device
US10832451B2 (en) * 2018-07-18 2020-11-10 Siemens Healthcare Gmbh Machine learning in iterative image reconstruction
US10825149B2 (en) * 2018-08-23 2020-11-03 Siemens Healthcare Gmbh Defective pixel correction using adversarial networks
US10997717B2 (en) * 2019-01-31 2021-05-04 Siemens Healthcare Gmbh Method and system for generating a confidence score using deep learning model
US11164021B2 (en) * 2019-05-16 2021-11-02 Arizona Board Of Regents On Behalf Of Arizona State University Methods, systems, and media for discriminating and generating translated images
US11580673B1 (en) * 2019-06-04 2023-02-14 Duke University Methods, systems, and computer readable media for mask embedding for realistic high-resolution image synthesis
US11422217B2 (en) * 2019-06-26 2022-08-23 Siemens Healthcare Gmbh Progressive generative adversarial network in medical image reconstruction
CN111192209B (en) * 2019-12-19 2023-06-30 浙江工业大学 Magnetic induction tomography optimization method based on deep learning
CN110916708A (en) * 2019-12-26 2020-03-27 南京安科医疗科技有限公司 CT scanning projection data artifact correction method and CT image reconstruction method
CN111340758B (en) * 2020-02-17 2023-08-11 天津中科智能识别产业技术研究院有限公司 Novel efficient iris image quality evaluation method based on deep neural network
EP4128135A4 (en) * 2020-04-01 2023-06-07 BOE Technology Group Co., Ltd. Computer-implemented method, apparatus, and computer-program product
CN113570508A (en) * 2020-04-29 2021-10-29 上海耕岩智能科技有限公司 Image restoration method and device, storage medium and terminal
CN111681297B (en) * 2020-05-28 2024-02-27 上海联影智能医疗科技有限公司 Image reconstruction method, computer device, and storage medium
CN111814689A (en) * 2020-07-09 2020-10-23 浙江大华技术股份有限公司 Fire recognition network model training method, fire recognition method and related equipment
CN111815692B (en) * 2020-07-15 2023-12-01 东软教育科技集团有限公司 Artifact-free data, method and system for generating artifact-free data, and storage medium
CN112001978B (en) * 2020-07-28 2023-09-01 中国人民解放军战略支援部队信息工程大学 Method and device for reconstructing image based on dual-energy dual-90-degree CT scanning of generating countermeasure network
CN112116601B (en) * 2020-08-18 2023-04-28 河南大学 Compressed sensing sampling reconstruction method and system based on generation of countermeasure residual error network
CN112230210B (en) * 2020-09-09 2022-07-29 南昌航空大学 HRRP radar target identification method based on improved LSGAN and CNN
CN112085687B (en) * 2020-09-10 2023-12-01 浙江大学 Method for converting T1 to STIR image based on detail enhancement
CN111899844B (en) * 2020-09-28 2021-11-23 平安科技(深圳)有限公司 Sample generation method and device, server and storage medium
US11748921B2 (en) * 2020-10-12 2023-09-05 Siemens Healthcare Gmbh Learnable extrapolation for deep learning-based medical image reconstruction
KR102337031B1 (en) 2020-11-30 2021-12-08 주식회사 코어라인소프트 Medical image reconstruction apparatus and method for screening a plurality of types lung diseases
KR102283673B1 (en) 2020-11-30 2021-08-03 주식회사 코어라인소프트 Medical image reading assistant apparatus and method for adjusting threshold of diagnostic assistant information based on follow-up exam
CN112489154B (en) * 2020-12-07 2022-06-03 重庆邮电大学 MRI motion artifact correction method for generating countermeasure network based on local optimization
KR20220082292A (en) * 2020-12-10 2022-06-17 주식회사 에어스 메디컬 Magnetic resonance image processing apparatus and method using artificial neural network in k-space domain
WO2022126614A1 (en) * 2020-12-18 2022-06-23 中国科学院深圳先进技术研究院 Manifold optimization-based deep learning method for dynamic magnetic resonance imaging
CN112561888B (en) * 2020-12-18 2024-04-16 中国科学院深圳先进技术研究院 Manifold optimization-based deep learning method for magnetic resonance dynamic imaging
CN112697435B (en) * 2021-01-26 2022-09-09 山西三友和智慧信息技术股份有限公司 Rolling bearing fault diagnosis method based on improved SELD-TCN network
US20220381861A1 (en) * 2021-05-19 2022-12-01 Siemens Healthcare Gmbh Method and system for accelerated acquisition and artifact reduction of undersampled mri using a deep learning based 3d generative adversarial network
CN113762349B (en) * 2021-08-11 2024-03-29 同济大学 Marine organism-oriented lightweight aliasing dense network classification method and system
US20230079353A1 (en) * 2021-09-14 2023-03-16 Siemens Healthcare Gmbh Image correction using an invertable network
EP4231036A1 (en) * 2022-02-21 2023-08-23 Koninklijke Philips N.V. Detection of artifical structures in magentic resonance images due to neural networks
CN116071478B (en) * 2023-04-06 2023-06-30 腾讯科技(深圳)有限公司 Training method of image reconstruction model and virtual scene rendering method
CN116563169B (en) * 2023-07-07 2023-09-05 成都理工大学 Ground penetrating radar image abnormal region enhancement method based on hybrid supervised learning
CN116630205B (en) * 2023-07-19 2024-03-12 深圳佳弟子科技有限公司 Image processing method, intelligent terminal and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120262167A1 (en) * 2011-04-18 2012-10-18 General Electric Company Accelerated multispectral data magnetic resonance imaging system and method
US20170309019A1 (en) * 2016-04-22 2017-10-26 New York University System, method and computer-accessible medium for learning an optimized variational network for medical image reconstruction
US20180336680A1 (en) * 2017-05-19 2018-11-22 Siemens Healthcare Gmbh Method and apparatus for functional magnetic resonance imaging

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120262167A1 (en) * 2011-04-18 2012-10-18 General Electric Company Accelerated multispectral data magnetic resonance imaging system and method
US20170309019A1 (en) * 2016-04-22 2017-10-26 New York University System, method and computer-accessible medium for learning an optimized variational network for medical image reconstruction
US20180336680A1 (en) * 2017-05-19 2018-11-22 Siemens Healthcare Gmbh Method and apparatus for functional magnetic resonance imaging

Non-Patent Citations (9)

* Cited by examiner, † Cited by third party
Title
Cham et al. (2017). Deep De-Aliasing for Fast Compressive Sensing MRI. arXiv preprint arXiv:1705.07137.
Hammernik et al. (2017). Learning a Variational Network for Reconstruction of Accelerated MRI Data. arXiv preprint arXiv:1704.00447.
Lee, et al. (2017). Deep artifact learning for compressed sensing and parallel MRI. arXiv preprint arXiv:1703.01120.
Mardani et al., "Deep Generative Adversarial Networks for Compressed Sensing (GANCS) Automates MRI" arXiv:1706.00051v1 [cs.CV], Wed, May 31, 2017.
Schlemper, et al. A Deep Cascade of Convolutional Neural Networks for MR Image Reconstruction. In International Conference on Information Processing in Medical Imaging (pp. 647-658).
Sun et al. (2016). Deep ADMM-net for compressive sensing MRI. In Advances in Neural Information Processing Systems (pp. 10-18).
Sun et al. (2016). Deep ADMM—net for compressive sensing MRI. In Advances in Neural Information Processing Systems (pp. 10-18).
Wang et al. (Apr. 2016). Accelerating magnetic resonance imaging via deep learning. In Biomedical Imaging (ISBI), 2016 IEEE 13th International Symposium on (pp. 514-517). IEEE.
Zhu et al. (2017). Image reconstruction by domain transform manifold learning. arXiv preprint arXiv:1704.08841.

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11170543B2 (en) * 2020-01-13 2021-11-09 The Board Of Trustees Of The Leland Stanford Junior University MRI image reconstruction from undersampled data using adversarially trained generative neural network

Also Published As

Publication number Publication date
US20190369191A1 (en) 2019-12-05

Similar Documents

Publication Publication Date Title
US10859657B2 (en) MRI reconstruction using deep learning, generative adversarial network and acquisition signal model
Dar et al. Prior-guided image reconstruction for accelerated multi-contrast MRI via generative adversarial networks
Mardani et al. Deep generative adversarial networks for compressed sensing automates MRI
US10671939B2 (en) System, method and computer-accessible medium for learning an optimized variational network for medical image reconstruction
US11783451B2 (en) Systems and methods for reducing colored noise in medical images using deep neural network
US20220237748A1 (en) Methods and system for selective removal of streak artifacts and noise from images using deep neural networks
US9542763B2 (en) Systems and methods for fast reconstruction for quantitative susceptibility mapping using magnetic resonance imaging
US11062490B2 (en) Reinforcement learning for online sampling trajectory optimization for magnetic resonance imaging
CN108090871A (en) A kind of more contrast MR image reconstruction methods based on convolutional neural networks
US11346911B2 (en) Magnetic resonance fingerprinting image reconstruction and tissue parameter estimation
Sandino et al. Deep convolutional neural networks for accelerated dynamic magnetic resonance imaging
Kelkar et al. Compressible latent-space invertible networks for generative model-constrained image reconstruction
Hammernik et al. $\Sigma $-net: Systematic Evaluation of Iterative Deep Neural Networks for Fast Parallel MR Image Reconstruction
Kelkar et al. Prior image-constrained reconstruction using style-based generative models
Singh et al. Joint frequency and image space learning for MRI reconstruction and analysis
CN114419176A (en) Reconstruction system and method for propeller magnetic resonance imaging
Belov et al. Towards ultrafast MRI via extreme k-space undersampling and superresolution
Ouchi et al. Reconstruction of compressed-sensing MR imaging using deep residual learning in the image domain
Guan et al. Magnetic resonance imaging reconstruction using a deep energy‐based model
Hou et al. PNCS: Pixel-level non-local method based compressed sensing undersampled MRI image reconstruction
Barbano et al. Steerable conditional diffusion for out-of-distribution adaptation in imaging inverse problems
Malkiel et al. Conditional WGANs with adaptive gradient balancing for sparse MRI reconstruction
Thomaz et al. Undersampled magnetic resonance image reconstructions based on a combination of u-nets and l1, l2, and tv optimizations
Mardani et al. Deep generative adversarial networks for compressed sensing (GANCS) automates MRI
CN114998460A (en) Dynamic undersampling method and device for reconstructing nuclear magnetic resonance image

Legal Events

Date Code Title Description
AS Assignment

Owner name: THE BOARD OF TRUSTEES OF THE LELAND STANFORD JUNIO

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GONG, ENHAO;ZAHARCHUK, GREG;PAULY, JOHN M.;AND OTHERS;SIGNING DATES FROM 20180531 TO 20190531;REEL/FRAME:049329/0281

Owner name: THE BOARD OF TRUSTEES OF THE LELAND STANFORD JUNIOR UNIVERSITY, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GONG, ENHAO;ZAHARCHUK, GREG;PAULY, JOHN M.;AND OTHERS;SIGNING DATES FROM 20180531 TO 20190531;REEL/FRAME:049329/0281

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STCF Information on status: patent grant

Free format text: PATENTED CASE