Machine Learning for Image Reconstruction
Acq/Recon/Analysis Tuesday, 18 May 2021

Oral Session - Machine Learning for Image Reconstruction
Acq/Recon/Analysis
Tuesday, 18 May 2021 14:00 - 16:00
  • Can Un-trained Networks Compete with Trained Ones for Accelerated MRI?
    Mohammad Zalbagi Darestani1 and Reinhard Heckel1,2
    1Electrical and Computer Engineering, Rice University, Houston, TX, United States, 2Electrical and Computer Engineering, Technical University of Munich, Munich, Germany
    Our work shows that a close performance to trained neural networks can be achieved without training using an un-trained network for accelerated MRI. We further show that untrained networks have another advantage, that is, naturally generalizing better to un-seen samples.
    ConvDecoder architecture. It is comprised of up-sampling, convolutional, ReLU, batch normalization, and linear combination layers.
    Sample reconstructions for ConvDecoder, TV, U-net, and the end-to-end variational network (VarNet) for a validation image from multi-coil knee measurements (4x accelerated). The second row represents the zoomed-in version of the first row. ConvDecoder and the end-to-end variational network (VarNet) find the best reconstructions for this image (slightly better than U-net and significantly better than TV). The scores given below are averaged over 200 different mid-slice images from the FastMRI validation set.
  • Learning data consistency for MR dynamic imaging
    Jing Cheng1, Wenqi Huang1, Zhuoxu Cui1, Ziwen Ke1, Leslie Ying2, Haifeng Wang1, Yanjie Zhu1, and Dong Liang1
    1Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China, 2University at Buffalo, The State University of New York, Buffalo, Buffalo, NY, United States
    We propose a new DL-based approach that implicitly learns the data consistency, which is corresponding to the real distribution of system noise. The DC term and the prior knowledge are both embedded in the weights of the networks to provides an utterly implicit learning of reconstruction model. 
    Figure 2. Reconstructions in spatial domain under different acceleration factors with the zoom-in images of the enclosed part and the corresponding error maps.
    Figure 1. Illustration of the proposed Learned DC. The architecture of each iteration is corresponding to Eq. (6), where the k-space sub-network has 2 layers and image domain sub-network has 3 layers. The numbers on the layers indicate the number of filters in that convolutional layer.
  • eRAKI: Fast Robust Artificial neural networks for K‐space Interpolation (RAKI) with Coil Combination and Joint Reconstruction
    Heng Yu1, Zijing Dong2,3, Yamin Arefeen2, Congyu Liao4, Kawin Setsompop4,5, and Berkin Bilgic3,6,7
    1Department of Automation, Tsinghua University, Beijing, China, 2Department of Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, MA, United States, 3Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States, 4Radiological Sciences Laboratory, Stanford University, Stanford, CA, United States, 5Department of Electrical Engineering, Stanford University, Stanford, CA, United States, 6Harvard Medical School, Boston, MA, United States, 7Harvard-MIT Health Sciences and Technology, Massachusetts Institute of Technology, Cambridge, MA, United States
    Robust Artificial neural networks for K‐space Interpolation (RAKI) can be significantly accelerated using coil combination in 3D k-space. We show improvements in 3D imaging with joint reconstruction and elliptical-CAIPI sampling and rapid EPTI reconstruction.
    Comparisons between 2D GRAPPA, 2D RAKI, and eRAKI reconstructions on 3D ME-MPRAGE data using 24×24×256 ACS and 3×3 undersampling. We use Tikhonov regularization in GRAPPA and set parameter λ=1×e-9. Our eRAKI benefits from 3D convolutions and can achieve comparable performance with high-speed reconstruction.
    Comparisons between GRAPPA, RAKI, and eRAKI reconstructions on EPTI data using 32x80x16 ACS. All three methods are implemented in kx-ky-t space and eRAKI can achieve comparable performance with GRAPPA but uses only two models to perform the whole learning/reconstruction process.
  • Compressed Sensing MRI Revisited: Optimizing $$$\ell_{1}$$$-Wavelet Reconstruction with Modern Data Science Tools
    Hongyi Gu1,2, Burhaneddin Yaman2,3, Kamil Ugurbil2, Steen Moeller2, and Mehmet Akcakaya2,3
    1Electrical Engineering, University of Minnesota, Minneapolis, MN, United States, 2Center for magnetic resonance research, Minneapolis, MN, United States, 3University of Minnesota, Minneapolis, MN, United States
    An optimized l1-wavelet compressed sensing method achieves comparable reconstruction quality to physics-guided deep learning, while using much fewer parameters and a linear representation amenable to interpretation. 
    Figure 1. Schematic of the unrolled ADMM for $$$\ell_{1}$$$-wavelet compressed sensing (CS). One unrolled iteration of ADMM with $$$\ell_{1}$$$-wavelet regularizers consists of regularizer (R), data consistency (DC) and dual update (DWT: Discrete wavelet transform). For T = 10 iterations, this leads to a total of 116 trainable parameters, as dual update is not needed for the last iteration.
    Figure 3. Two representative slices from coronal PD–FS knee MRI, reconstructed using CG-SENSE, PG-DL, and the optimized -wavelet CS. CG-SENSE is visibly outperformed by PG-DL and -wavelet CS. Both PG-DL and the optimized -wavelet CS show comparable image quality and quantitative metrics.
  • XPDNet for MRI Reconstruction: an application to the 2020 fastMRI challenge
    Zaccharie Ramzi1,2,3, Jean-Luc Starck2, and Philippe Ciuciu1,3
    1Neurospin, Gif-Sur-Yvette, France, 2Cosmostat team, CEA, Gif-Sur-Yvette, France, 3Parietal team, Inria Saclay, Gif-Sur-Yvette, France
    We introduce the XPDNet, a neural network for MRI reconstruction, which ranked second in the fastMRI 2020 challenge.
    General cross-domain networks architecture. Skip and residual connection are omitted for the sake of clarity. y are the under-sampled measurements,in our case the k-space measurements, Ω is the under-sampling scheme,Fis the measurement operator, in our case the Fourier Transform, and x is the recovered solution.
    Magnitude reconstruction results for the different fastMRI contrasts at acceleration factor 4. The top row represents the ground truth, the middle one represents the reconstruction from a retrospectively under-sampled k-space,and the bottom row represents the absolute error when comparing the two. The average image quantitative metrics are given for 30 validation volumes.
  • Estimating Uncertainty in Deep Learning MRI Reconstruction using a Pixel Classification Image Reconstruction Framework
    Kamlesh Pawar1,2, Gary F Egan1,2,3, and Zhaolin Chen1,4
    1Monash Biomedical Imaging, Monash University, Melbourne, Australia, 2School of Psychological Sciences, Monash University, Melbourne, Australia, 3ARC Centre of Excellence for Integrative Brain Function, Monash University, Melbourne, Australia, 4Department of Electrical and Computer Systems Engineering, Monash University, Melbourne, Australia
    Image reconstruction was modeled as a classification problem and the predicted probability of the reconstructed pixel was used to estimate the uncertainty map. The predicted uncertainties correlate with the actual errors, providing a tool for risk assessment of DL image reconstruction.
    Figure 1: A image obtained from the undersampled k-space data was used as input and a fully sampled image quantized to 8-bit (256 grey levels) was used as target image to the network for training. The DL Unet network classifies each pixel in the reconstructed image and the output of the network was probability for each pixel belonging to one of the 256 distinct quantized grey levels. A weighted linear combination of the predicted probability forms the reconstructed image and standard deviation of a Gaussian fitted curve to the predicted probability distribution from the uncertainty maps.
    Figure 2: (a) Input image from undersampled (factor of 4) k-space data; (b, d) are the predicted probabilities at spatial location as pointed out by the red, blue and green dots in the reconstructed image respectively; (c) Reconstructed image obtained using weighted linear combination of the predicted probabilities; (e) fully sampled reference image; ; (f) error image obtained by subtracting reference and output image; (c) Uncertainty maps obtained from standard deviation of the fitted Gaussian curve to the predicted probabilities.
  • DSLR+: Enhancing deep subspace learning reconstruction for high-dimensional MRI
    Christopher Michael Sandino1, Frank Ong2, Ke Wang3, Michael Lustig3, and Shreyas Vasanawala2
    1Electrical Engineering, Stanford University, Stanford, CA, United States, 2Radiology, Stanford University, Stanford, CA, United States, 3Electrical Engineering and Computer Sciences, UC Berkeley, Berkeley, CA, United States
    A novel deep subspace learning reconstruction (DSLR+) method is proposed to reconstruct compressed, locally low-rank representations of high-dimensional imaging data enabling high-quality reconstructions with relatively small memory footprint.
    Fig. 2: Undersampled k-space data are reconstructed by zero-filling and then converted into blocks, which are decomposed using SVD to initialize $$$L_b$$$ and $$$R_b$$$. These are iteratively processed by DSLR+ by alternating conjugate gradient and CNN updates. Before each CNN, the basis functions are split into real/imaginary parts and concatenated along the featuredimension. For simplicity, 2-D and 1-D ResNets comprised of 6 convolutions each are used. At the end of the network, the basis functions are combined to form the output images.
    Fig. 5: Two prospectively undersampled scans are performed in a patient with premature ventricular contractions (PVC), which manifests as a double heartbeat motion. The first acquisition is 12X accelerated and reconstructed using MoDL, DSLR, and DSLR+ (left to right). A second acquisition is 2X accelerated and reconstructed with SENSE (rightmost). All show remarkable image quality despite severe acceleration, and never having seen a PVC during training. Both MoDL and DSLR+ faithfully demonstrate the double beat motion, although, DSLR+ depicts less residual aliasing (arrows).
  • Anomaly-aware multi-contrast deep learning model for reduced gadolinium dose in contrast-enhanced brain MRI - a feasibility study
    Srivathsa Pasumarthi1, Enhao Gong1, Greg Zaharchuk1, and Tao Zhang1
    1Subtle Medical Inc., Menlo Park, CA, United States
    This work investigates the feasibility of improving the performance of such DL algorithms using multi-contrast MRI data, combined with an unsupervised anomaly detection based attention mechanism.
    GBCA dose-reduction model with T2W image and the generated UAD mask as attention mechanism. The model is trained with a combination of SSIM, perceptual, adversarial and UAD-mask weighted L1-loss.
    Examples of cases where the proposed model has a better tumor/lesion enhancement pattern compared to the T1-only model.
  • Blind Primed Supervised (BLIPS) Learning for MR Image Reconstruction
    Anish Lahiri1, Guanhua Wang2, Saiprasad Ravishankar3, and Jeffrey A. Fessler1
    1Electrical and Computer Engineering, University of Michigan, Ann Arbor, MI, United States, 2Biomedical Engineering, University of Michigan, Ann Arbor, MI, United States, 3Computational Mathematics, Science and Engineering, and Biomedical Engineering, Michigan State University, East Lansing, MI, United States
    Our findings indicate that due to synergy between learned features, there is significant benefit to combining shallow, sparse prior-based blind learning reconstruction with deep-supervised reconstruction in MRI
    Figure 5: Comparison of reconstructions for a knee image using the proposed BLIPS method versus strict supervised learning, blind dictionary learning, and zero-filled reconstruction for the 5-fold undersampling mask depicted in Fig. 2. Metrics listed below each reconstruction correspond to PSNR(in dB)/SSIM/HFEN respectively.
    Figure 1: Proposed pipeline for combining blind dictionary learning and supervised learning-based MR image reconstruction.
  • Joint estimation of coil sensitivities and image content using a deep image prior
    Guanxiong Luo1, Xiaoqing Wang1, Volkert Roeloffs1, Zhengguo Tan1, and Martin Uecker1,2
    1Institute for Diagnostic and Interventional Radiology, University Medical Center Göttingen, Germany, Göttingen, Germany, 2Campus Institute Data Science (CIDAS), University of Göttingen, Germany, Göttingen, Germany
    The nonlinear inversion reconstruction is a calibrationless parallel imaging technique, which jointly estimate coil sensitivities and image content. We demonstrate how to combine such a calibrationless technique with an advanced neural network based image prior for efficient MR imaging. 
    Figure 1. For the case of moderate undersampling, the two reconstructions regularized by log-likelihood and $$$\ell_1$$$ are very close, and the structural similarity index between them is 0.95. The $$$\ell_1$$$ reconstruction has blocky artifacts in Region 1 introduced by the wavelet transform, especially for higher undersampling. Overall, the learned log-likelihood outperforms $$$\ell_1$$$ in noise suppression, especially in Region 2. The reconstructions regularized by the learned log-likelihood also better preserve the boundaries between tissues and have less noise.
    Figure 2. Reconstructed coil sensitivities (grayscale magnitude and color-coded phase) after channel compression.
Back to Top
Digital Poster Session - Machine Learning for Image Reconstruction
Acq/Recon/Analysis
Tuesday, 18 May 2021 15:00 - 16:00
  • Unsupervised Dynamic Image Reconstruction using Deep Generative Adversarial Networks and Total Variation Smoothing
    Elizabeth Cole1, Shreyas Vasanawala1, and John Pauly1
    1Stanford University, Palo Alto, CA, United States
    We test the new method in two scenarios: undersampled cine and abdominal DCE. The method produces higher quality images which reveal vessels and recover more anatomical structure compared to compressed sensing.

    Figure 1. (a) Framework overview example in a supervised setting with a conditional GAN when fully-sampled datasets are available.

    (b) Our proposed framework overview in an unsupervised setting. A sensing matrix comprised of coil sensitivity maps, an FFT and a randomized undersampling mask is applied to the generated image to simulate the imaging process. The discriminator takes simulated and observed measurements as inputs and tries to differentiate between them. The generator’s loss is based on the discriminator as well as reducing spatial and temporal variation.

    Figure 5. (a) Representative DCE images. The leftmost column is the input zero-filled reconstruction, the middle column is our generator’s reconstruction, and the rightmost column is the CS reconstruction. The generator improves the input image quality by recovering sharpness and adding more structure to the input images.

    (b) Comparison of DCE inference time per three-dimensional DCE volume (2D + time) between CS low-rank and our unsupervised GAN. Our method is approximately 2.98 times faster.

  • Deep Manifold Learning for Dynamic MR Imaging
    Ziwen Ke1, Zhuo-Xu Cui1, Jing Cheng1, Leslie Ying2, Xin Liu1, Hairong Zheng1, Yanjie Zhu1, and Dong Liang1
    1Shenzhen Institutes of Advanced Technology, Shenzhen, China, 2University at Buffalo, The State University of New York, Buffalo, NY, United States
    Our work tries to develop a deep optimization model on a nonlinear manifold directly. The comparisons with DC-CNN and CRNN show that our work can achieve improved results. To our knowledge, this work represents the first study applying a deep manifold optimization to dynamic MR images.
    Figure 2. The proposed Manifold-Net for dynamic MR imaging.
    Figure 1. Gradient descent on the low-rank tensor manifold.
  • A Plug-and-play Low-rank Network Module in Dynamic MR Imaging
    Ziwen Ke1, Wenqi Huang1, Jing Cheng1, Leslie Ying2, Xin Liu1, Hairong Zheng1, Yanjie Zhu1, and Dong Liang1
    1Shenzhen Institutes of Advanced Technology, Shenzhen, China, 2University at Buffalo, The State University of New York, Buffalo, NY, United States
    This paper explores a plug-and-play low-rank network module in dynamic MR imaging. It can be easily embedded into any other dynamic MR neural networks. Experimental results show that the proposed plug-and-play low-rank module can improve the reconstruction results. 
    Figure 1. The proposed plug-and-play LR network module. (a) The original ISTA-Net. (b) The original DC-CNN. (c) The original CRNN. (d) ISTA-LR-Net by embedding the LR network module into the original ISTA-Net. (e) DC-CNN-LR by embedding the LR network module into the original DC-CNN.(f) CRNN-LR by embedding the LR network module into the original CRNN. The numbers in the dotted box represent the locations where the LR module can be embedded.
    Figure 2. The reconstruction results of the different methods (DC-CNN, DC-CNN-LR, CRNN, and CRNN-LR) at 8-fold acceleration. From left to right, the first row shows the ground truth, the reconstruction result of these methods. The second row shows the enlarged view of their respective heart regions framed by a yellow box. The third row shows the error map (display ranges [0, 0.07]). The y-t image (extraction of the 124th slice along the y and temporal dimensions) and the error of the y-t image are also given for each signal to show the reconstruction performance in the temporal dimension.
  • Cascaded U-net with Deformable Convolution for Dynamic Magnetic Resonance Imaging
    Zhehong Zhang1, Yuze Li2, and Huijun Chen2
    1Department of Engineering Physics, Tsinghua University, Beijing, China, 2Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
    We present a new method that incorporates deformable 2D convolution kernels into the cascaded U-net. The proposed method leverages motion information of dynamic MRI and thus deformable convolution kernel naturally adapts to image structures.
    Figure 2: An example of the concatenation of U-net blocks which operate in either image domain or k-space domain with data consistency layer in between. The deformable convolution layer, which extracts the motion feature and generates offsets from a series of dynamic MRI, is at the beginning of the U-net block on the image domain.
    Figure 1: The illustration of the generation offset fields and the performance of deformed sampling. The offsets are obtained by applying a convolution layer over the same input feature map. The channel dimension of the offset fields is twice as much as that of the input feature maps, given the 2D offsets.
  • Joint deep learning-based optimization of undersampling pattern and reconstruction for dynamic contrast-enhanced MRI
    Jiaren Zou1,2 and Yue Cao1,2,3
    1Department of Radiation Oncology, University of Michigan, Ann Arbor, MI, United States, 2Department of Biomedical Engineering, University of Michigan, Ann Arbor, MI, United States, 3Department of Radiology, University of Michigan, Ann Arbor, MI, United States
    Joint training of reconstruction network, sampling pattern and data sharing for dynamic contrast-enhanced MRI was investigated. The temporal degree of freedom of the sampling pattern and learned data sharing can improve the reconstruction accuracy.
    Figure 1. The network structure with joint learning of sampling pattern and data sharing. The interconnections in the data sharing stage highlight the different data sharing used by each frame.
    Figure 2. Exemplary reconstruction results of a testing frame. The (a) ground truth, (b, c) with temporal DoF and data sharing (model A) and its error map, (d, e) with temporal DoF and without data sharing (model B) and its error map, (f, g) without temporal DoF and data sharing (model C) and its error map, and (h, i) trained with pseudo golden angle radial sampling and without data sharing (model D) and its error map are shown. Model A shows better qualitative reconstruction quality.
  • A Few-Shot Learning Approach for Accelerated MRI via Fusion of Data-Driven and Subject-Driven Priors
    Salman Ul Hassan Dar1,2, Mahmut Yurt1,2, and Tolga Çukur1,2,3
    1Department of Electrical and Electronics Engineering, Bilkent University, Ankara, Turkey, 2National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara, Turkey, 3Neuroscience Program, Aysel Sabuncu Brain Research Center, Bilkent University, Ankara, Turkey
    Here we proposed a few shot learning approach to address the issue of data scarcity in deep neural networks for accelerated MRI. The proposed approach enables data efficient training of deep neural networks by merging subject-driven priors with data-driven prior.
    Figure. 1 - COMNET consists of an unrolled cascade of sub-networks where each sub-network consists of a calibration consistency (CC) block fused with a network consistency (NC) block, both followed by a data consistency (DC) block.
    Figure. 2 - Average PSNR values of a) cT1, b) T2, and c) FLAIR images of test subjects as a function of number of training subjects (upper x-axis), and training samples (lower x-axis). COMNET requires just a few training samples from a single subject to outperform L1-SPIRiT. On the other hand, DNN on average requires around 90 samples from 9 different subjects to start performing better than L1-SPIRiT.
  • Weakly Supervised MR Image Reconstruction using Untrained Neural Networks
    Beliz Gunel1, Morteza Mardani1, Akshay Chaudhari2, Shreyas Vasanawala2, and John Pauly1
    1Electrical Engineering, Stanford University, Stanford, CA, United States, 2Radiology, Stanford University, Stanford, CA, United States
    Untrained networks to construct weak labels from undersampled MR scans at training time. Use limited supervised and weakly supervised pairs to train an unrolled network with strong reconstruction performance and fast inference time, improving over supervised and self-training baselines.
    Figure 1: Diagram of our two-step weakly supervised framework. Step 1 involves weakly labeled training data creation using either Self-training or ConvDecoder Weak Labeling. In Self-training, a "teacher model" is trained with fully sampled data to run inference on undersampled data, for creating weak labels. Alternatively, ConvDecoder can directly create weakly labeled data from undersampled data. Step 2 involves training a "student model" unrolled neural network using both supervised and weakly labeled pairs, which is then used at inference time.
    Figure 5: Image quality comparison between Supervised Baseline, Self-training, and CD Weak Labeling. All methods use 1 supervised and 5 weakly labeled volumes during training with 4x acceleration. We observe that CD Weak Labeling provides clear improvement in reducing background noise and aliasing artifacts over both Supervised Baseline and Self-training.
  • A Custom Loss Function for Deep Learning-Based Brain MRI Reconstruction
    Abhinav Saksena1, Makarand Parigi1, Nicole Seiberlich2, and Yun Jiang2
    1EECS, University of Michigan, Ann Arbor, MI, United States, 2Department of Radiology, University of Michigan, Ann Arbor, MI, United States

    We investigate both per-pixel (L1) and perceptual based (SSIM) loss functions for deep learning based MRI reconstruction.

    A loss function that linearly combines MSSSIM and L1 (with a higher weightage towards MSSSIM) produces superior image reconstructions and achieves higher SSIM scores.

    Figure 1: Sample reconstructions for the different loss functions with 4-fold acceleration. Each row represents a different image contrast.
    Table 1: 4-fold acceleration results averaged across the 16-coil only FastMRI validation dataset. For PSNR and SSIM, the higher the score the better, and vice-versa for NMSE.
  • A lightweight and efficient convolutional neural network for MR image restoration
    Aowen Liu1, Meiling Ji2, Xiaoqian Huang1, Yawei Zhao2, Renkuan Zhai2, Guobin Li2, Dinggang Shen1, and Shu Liao1
    1United Imaging Intelligence, Shanghai, China, 2United Imaging Healthcare, Shanghai, China
    An efficient and lightweight neural network EEDN with a novel loss function for MRI restoration is proposed, which  can not only greatly improve network efficiency with reduced network parameters and FLOPs, but also help enhance image quality with better visualization. 
    Fig. 1. The architecture of EEDN. FLB represents feature learning sub-module, which is implemented with EAM [5]. Up and Down are the up and down sampling operation, respectively. For FLB modules, multiple inputs of each FLB module are first concatenated and then reduced for channels by 1x1 conv.
    Fig. 2. Comparison of different models using five different metrics.
  • Scalable and Interpretable Neural MRI Reconstruction via Layer-Wise Training
    Batu Ozturkler1, Arda Sahiner1, Mert Pilanci1, Shreyas Vasanawala2, John Pauly1, and Morteza Mardani1
    1Electrical Engineering, Stanford University, Stanford, CA, United States, 2Radiology, Stanford University, Stanford, CA, United States
    An interpretable and scalable training method for MRI reconstruction based on convex layer-wise training is presented. The method is evaluated on the fastMRI knee dataset. Our experiments show it attains on par image quality with end-to-end training with less memory footprint for training.
    The proposed layer-wise training algorithm. At each training step, a two-layer network is trained to fit the ground-truth. After training has finished, the weights of the first layer are frozen, and the second training step is performed. This training algorithm is repeated k times to obtain a deep network with k+1 convolutional layers. L(y,$$$\hat{x}_i$$$) is the loss for ith step where y is ground-truth, and $$$\hat{x}_i$$$ is the prediction of ith layer. In the final network, each layer except the final layer have 7x7 kernel size and 256 channels, and the final layer has 1x1 kernel size.
    Representative knee images from the test set for layer-wise training vs end-to-end training with Poisson disc sampling (top) and Cartesian sampling (bottom). End-to-end training and layer-wise training have similar perceptual quality.
  • Effective Training of 3D Unrolled Neural Networks on Small Databases
    Zilin Deng1,2, Burhaneddin Yaman1,2, Chi Zhang1,2, Steen Moeller2, and Mehmet Akçakaya1,2
    1University of Minnesota, Minneapolis, MN, United States, 2Center for Magnetic Resonance Research, Minneapolis, MN, United States
    Training of 3D unrolled networks for volumetric MRI reconstruction with small databases and limited GPU resources may be facilitated by small-slab processing
    Figure 3. A representative test slice from reconstructions using 2D and 3D unrolled networks. 2D processing suffers from residual artifacts (red arrows), which are suppressed with the 3D processing.
    Figure 2. Schematic of the supervised training on the small slabs generated from the volumetric datasets. Unrolled neural networks comprise a series of data consistency and regularizer units, arising from conventional iterative optimization algorithms. In this work, we use a 3D unrolled network. A ResNet architecture with 5 residual blocks, consisting of 2 convolution layers with former followed by ReLU and latter followed by a scaling layer is used for implicit regularization. All layers of this ResNet use 3×3×3 kernels and 64 channels.
  • Memory-Efficient Learning for High-Dimensional MR Reconstruction
    Ke Wang1, Michael Kellman2, Christopher M. Sandino3, Kevin Zhang1, Shreyas S. Vasanawala4, Jonathan I. Tamir5, Stella X. Yu6, and Michael Lustig1
    1Electrical Engineering and Computer Sciences, University of California, Berkeley, Berkeley, CA, United States, 2Pharmaceutical Chemistry, University of California, San Francisco, Berkeley, CA, United States, 3Electrical Engineering, Stanford University, Stanford, CA, United States, 4Radiology, Stanford University, Stanford, CA, United States, 5Electrical and Computer Engineering, The University of Texas at Austin, Austin, TX, United States, 6International Computer Science Institute, University of California, Berkeley, Berkeley, CA, United States
    We present a memory-efficient learning (MEL) framework for high-dimensional MR reconstruction to enable training larger and deeper unrolled networks on a single GPU. We demonstrate improved image quality with learned reconstruction enabled by MEL for 3D MRI and 2D cardiac cine applications.
    Figure 1. GPU memory limitations for high-dimensional unrolled DL recons: a) Conventionally, in order to reconstruct a 3D volume, each slice in the readout dimension is independently passed through a 2D unrolled network and then re-joined into a 3D volume. In contrast, 3D unrolled networks require more memory but use 3D slabs during the training, which leverages more data redundancy. b) Typically, cardiac cine DL recons are often performed with a small number of unrolls due to memory limitations. More unrolls are able to better learn the finer spatial and temporal textures.
    Figure 2. a) Unrolled networks are formed by unrolling the iterations of an image reconstruction optimization. Each unroll consists of a CNN-based regularization layer and a DC layer. Conventionally, gradients of all layers are computed simultaneously, which requires a significant GPU memory cost. b) Memory-efficient learning procedure for a single layer: 1) Recalculate the layer’s input $$$\mathbf{x}^{(n-1)}$$$, from the known output $$$\mathbf{x}^{(n)}$$$. 2) Reform the AD graph for that layer. 3) Backpropagate gradients $$$q^{(n)}$$$ through the layer’s AD graph.
  • Novel insights on SSA-FARY: Amplitude-based respiratory binning in self-gated cardiac MRI
    Sebastian Rosenzweig1,2 and Martin Uecker1,2
    1Diagnostic and Interventional Radiology, University Medical Center Göttingen, Göttingen, Germany, 2Partner Site Göttingen, German Centre for Cardiovascular Research (DZHK), Göttingen, Germany
    We present novel insights on the quadrature-pair self-gating signals estimated by SSA-FARY: We show that one element of each pair is certain to be in-phase with the motion it represents, as it is the result of a filtering process with a zero-phase filter.
    FIG_3 SSA-FARY-gated compressed sensing recontruction of the heart. All three SMS-slices are depicted. The respiratory motion is resolved into 5 bins using SSA-FARY amplitud-binning. The white lines serve as reference to appreciate the different respiratory states. Bin 1: End-expiration, Bin 5: End-inspiration. The slightly blurred appearance of the fifth bin is a result of the limited number of spokes which were aquired in the typically shorter end-inspiration state.
    FIG_2 Respiratory gating with SSA-FARY. A) The background shows the diaphragm-motion extracted from a real-time reconstruction of the entire time-series. On top, EOF 3 and EOF 4, i.e. the third and fourth column of the SVD-Matrix $$$U$$$, are plotted. B) For each EOF, four representative filters extracted from the$$$V^H$$$-matrix are depicted. Note, that the symmetric filters (green) correspond to EOF 4, which is in-phase with the diaphragm and which can therefore be used for amplitude-binning.
  • Reconstruction of Whole-Heart Cardiac Radial MRI using Neural Network Transfer Learning Approach
    Ibtisam Aslam1,2, Fariha Aamir2, Lindsey A CROWE1, Miklos KASSAI1, Hammad Omer2, and Jean-Paul VALLEE1
    1Service of Radiology, Geneva University Hospitals and Faculty of Medicine, University of Geneva, Geneva, Switzerland, 2Medical Image Processing Research Group (MIPRG), Deptt. of Electrical & Computer Engineering, COMSATS University Islamabad, Islamabad, Pakistan
    Non-Cartesian CMR acquisition helps to lessen the scan time but have artifacts. This work proposes Transfer-learning approach with NUFFT (NUFFT TL-Net) to reconstruct artifact-free whole heart, radial CMR images.
    Figure 2: Schematic illustration of the pre-trained network for the Proposed NUFFT TL-Net framework at AF=7 & 13
    Figure 4: Middle slice (short axis) End-diastole and End-Systole reconstructed images of a patient for whole heart cine Radial MR at acceleration factor 13 with 24 radial lines per image. Reference cine: Fully sampled image: Undersampled Image: NUFFT with iFFT image. NUFFT U-Net: Reconstructed image of the fully trained NUFFT U-Net at AF=13, NUFFT TL-Net: Reconstructed image of NUFFT LT-Net at AF=13. Corresponding edge images provide reconstructed image quality assessment. Arrows show the myocardial wall distortion that is less pronounced in the NUFFT TL-Net.
  • Alignment & joint recovery of multi-slice cine MRI data using deep generative manifold model
    Qing Zou1, Abdul Haseeb Ahmed1, Prashant Nagpal1, Rolf Schulte2, and Mathews Jacob1
    1University of Iowa, Iowa City, IA, United States, 2GE Global Research, Munich, Germany
    This work proposed a scheme for the Alignment & joint recovery of multi-slice cine MRI data using deep generative manifold model. The proposed scheme can significantly reduce the scan time. 
    Fig. 2. Demonstration of the framework of the proposed scheme on the first dataset. We plot the latent variables of 150 frames in time series on the first dataset. We showed four different phases from 4 different slices that are reconstructed in the time series: systole in End-Expiration (E-E), systole in End-Inspiration (E-I), diastole in End-Expiration (E-E) and diastole in End-Inspiration (E-I). The latent vectors corresponding to the four different phases are indicated in the plot of the latent vectors.
    Fig. 3. Illustration of the framework of the proposed scheme on the second dataset. We plot the latent variables of 150 frames in time series on the first dataset. We showed four different phases from 4 different slices that are reconstructed in the time series: systole in End-Expiration (E-E), systole in End-Inspiration (E-I), diastole in End-Expiration (E-E) and diastole in End-Inspiration (E-I). The latent vectors corresponding to the four different phases are indicated in the plot of the latent vectors.
  • Deep image reconstruction for MRI using unregistered measurement pairs without ground truth
    Weijie Gan1, Yu Sun1, Cihat Eldeniz2, Jiaming Liu3, Hongyu An2, and Ulugbek S. Kamilov1,3
    1Department of Computer Science and Engineering, Washington University in St. Louis, St. Louis, MO, United States, 2Mallinckrodt Institute of Radiology, Washington University in St. Louis, St. Louis, MO, United States, 3Department of Electrical and Systems Engineering, Washington University in St. Louis, St. Louis, MO, United States
    An unsupervised deep learning method for MRI that simultaneously addresses the problems of image registration and reconstruction was proposed. The CNN consists of two separate modules that are jointly trained by directly using pairs of unregistered images.
    U-Dream jointly trains image reconstruction and image registration modules. Inputs are unregistered measurement pairs from the same subject. The zero-filled images are passed through the reconstruction CNN to remove artifacts due to noise and undersampling. The output images are then used in the registration CNN to obtain the motion field characterizing the directional mapping between their coordinates. We register one of the reconstructed images to the other by the Spatial Transform Network.
    Illustration of several reconstruction methods on synthetic undersampled brain MRI images. The top-right corner of each image provides the PSNR and SSIM values with respect to the groundtruth image. Unregistered N2N is directly trained on unregistered data, while SyN+N2N trains the CNN on pre-registered images that are corrupted with artifacts. U-Dream achieves a significant improvement relative to both methods by jointly addressing the reconstruction and the registration problems.
  • Adaptive deep image reconstruction using G-SURE
    Hemant Kumar Aggarwal1 and Mathews Jacob1
    1University of Iowa, Iowa City, IA, United States

    We introduce a novel approach to adapt deep learned reconstruction algorithms to a new forward model. The proposed scheme relies on G-SURE loss metric, which accounts for the noise in the measurements. By minimizing the risk of overfitting, this scheme offers improved reconstructions.

    Fig. 1: These are the experiments in 2D settings at 6-fold acceleration with a noise of standard deviation of 0.02. Here, we trained the MoDL only with mask $$$M_0$$$ as shown in (b). During testing, we utilized the same mask $$$M_0$$$ as well as a different mask $$$M_1$$$ (d). The reconstruction results in (c) and (e) show that MoDL architecture is robust to small changes in the sampling mask.
    Fig. 2: (a) The Cartesian mask $$$M_0$$$ used during training. (b) The regridding reconstruction of the test image using $$$M_0$$$. (c) MoDL results in good reconstruction since $$$M_0$$$ is used here during testing. (d) This is a different mask $$$M_1$$$ to test the robustness of MoDL architecture. (e) is the corresponding regridding reconstruction. The reconstruction (f) has lower PSNR than (c), as expected since $$$M_1$$$ was not used during training. (h) is the result of model adaptation on (f) using MSE (i) is the reconstructed image using the G-SURE-based model adaption.
  • Influence of training data on RAKI reconstruction quality in standard 2D imaging
    Peter Dawood1,2, Martin Blaimer3, Peter M. Jakob1, and Johannes Oberberger2
    1Department of Experimental Physics 5, University of Würzburg, Würzburg, Germany, 2Department of Internal Medicine I, University Hospital Würzburg, Würzburg, Germany, 3Magnetic Resonance and X-Ray Imaging Department, Development Center X-ray Technology EZRT, Fraunhofer Institute for Integrated Circuits IIS, Würzburg, Germany
    Compared to standard parallel imaging, deep-learning based RAKI yields superior signal-to-noise ratio but introduces blurring at high accelerations. The contrast in the calibration data should be similar to that of the accelerated scans.

    Fig. 1: The CNN architecture used in this work. The input layer takes in the subsampled, zerofilled k-space data of the coil array with Nc independent choils. Real- and imaginary part of the k-space data are passed to separate channels, resulting in 2 x Nc input-channels. Two hidden layers are assigned 256 and 128 channels, respectively (n2=256 and n3=128). The output layer predicts all missing points across all coils simultaneously, and thus has 2 x (R-1) x Nc channels, with R denoting the undersampling rate.

    Fig. 2: 2D image reconstructions of brain dataset (32 coils, T1-weighted) for retrospective undersampling rates in range 4-6 (denoted as R4-R6). ACS data (30 central phase lines) were re-inserted into reconstructed k-spaces. GRAPPA kernel size was optimized for each undersampling rate: 11 x 4 (R4-5), and 15 x 2 (R6) in read- and phase direction. RAKI performs better than GRAPPA in terms of noise resilience, but suffers from blurring artifacts, which have pronounced appearance at 6-fold undersampling.

  • Non-uniform Fast Fourier Transform via Deep Learning
    Yuze Li1, Zhangxuan Hu2, Haikun Qi3, Guangqi Li1, Dongyue Si1, Haiyan Ding1, Hua Guo1, and Huijun Chen1
    1Center for Biomedical Imaging Research, Medical School, Tsinghua University, Beijing, China, 2MR Research China, GE Healthcare, Beijing, China, 3King’s College London, London, United Kingdom
    A deep learning-based MR reconstruction framework called DLNUFFT was proposed, which can restore the under-sampled non-uniform k-space to fully sampled Cartesian k-space without NUFFT gridding. Results showed DLNUFFT can achieve higher performance than compared methods.
    Figure 1. The network structure of DLNUFFT. (a) Under-sampled k-space data Xu is processed by Block Layer to obtain Xp with multiple patches. Xp goes through the Density Compensation and Reordering Layer, which is initialized using the position map. After Adaptive Interpolation Layer, the fully sampled Cartesian patch data Xkf are obtained, followed by ReBlock Layer to generate the fully sampled k-space data Xf. Finally, the inverse FFT is applied on Xf to get the reconstructed image If. (b) Illustration of real acquired k-space data processed by DLNUFFT and the output after each block.
    Figure 4. Reconstruction results for DLNUFFT and other methods on three datasets with radial trajectory (R=4). DLNUFFT can achieve relatively high PSNR (36.65dB) and the highest SSIM (93.21%).
  • Deep Learning Image Reconstruction from Incomplete Fast Spin Echo MR Data
    Linfang Xiao1,2, Yilong Liu1,2, Yujiao Zhao1,2, Zheyuan Yi1,2,3, Vick Lau1,2, Alex T.L. Leong1,2, and Ed X. Wu1,2
    1Laboratory of Biomedical Imaging and Signal Processing, The University of Hong Kong, Hong Kong, China, 2Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, China, 3Department of Electrical and Electronic Engineering, Southern University of Science and Technology, Shenzhen, China
    We propose a new FSE acquisition and deep learning reconstruction approach to acquire and reconstruct single-channel FSE data with significantly reduced number of shots, which effectively removes aliasing artifacts and recovers high frequency information without noise amplification.
    Figure 1 (A) Illustration of proposed sampling pattern for incomplete single-channel FSE acquisition with partial shots (TRs). It acquires partial shots, e.g., by skipping 7 out of total 20 shots at 1/3/5/14/16/18/20 for each segment. (B) Proposed ResNet for image reconstruction, consisting of 2 convolutional layers with downsampling, 16 residual blocks (RBs), and 2 convolutional layers with upsampling. Each RB contains 2 convolutional layers with a rectified linear unit. The inputs are images directly reconstructed from incomplete k-space data with missing shots zero-filled.
    Figure 2 (A) Demonstration of the MR reconstruction results by the proposed ResNet from the incomplete single-channel FSE data (containing 13 out of total 20 shots with ETL=15). (B) Typical MR reconstruction results with 13/12/11 out of total 20 shots, respectively. Quantitative measures PSNR and SSIM are also shown. These results demonstrated that the trained ResNet model could effectively remove the aliasing artifacts and recover the high frequency information without noticeable blurring at various undersampling levels.
Back to Top
Digital Poster Session - Machine Learning to Reconstruct Accelerated Scans
Acq/Recon/Analysis
Tuesday, 18 May 2021 15:00 - 16:00
  • Using Untrained Convolutional Neural Networks to Accelerate MRI in 2D and 3D
    Dave Van Veen1,2, Arjun Desai1, Reinhard Heckel3,4, and Akshay S. Chaudhari1
    1Stanford University, Stanford, CA, United States, 2University of Texas at Austin, Austin, TX, United States, 3Rice University, Houston, TX, United States, 4Technical University of Munich, Munich, Germany
    We find that untrained convolutional neural networks are comparable to supervised methods for accelerating MRI scans in both 2D and 3D. Further, we demonstrate a method for regularizing the network feature maps using undersampled k-space measurements.
    Figure 4. Comparison of untrained method against fully supervised baseline on a 3D qDESS axial slice. The k-space was downsampled using a 2D Poisson disc at acceleration factors of 4x and 8x for the top and bottom rows, respectively. The image quality of the reconstruction with the same undersampling factors is higher for 3D qDESS than 2D fastMRI data, likely because the undersampling aliasing is spread across two dimensions ($$$k_y$$$ and $$$k_z$$$).
    Figure 1. Network architecture for our untrained method.
  • Zero-shot Learning for Unsupervised Reconstruction of Accelerated MRI Acquisitions
    Yilmaz Korkmaz1,2, Salman Ul Hassan Dar1,2, Mahmut Yurt1,2, and Tolga Çukur1,2,3
    1Department of Electrical and Electronics Engineering, Bilkent University, Ankara, Turkey, 2National Magnetic Resonance Research Center, Bilkent University, Ankara, Turkey, 3Neuroscience Program, Aysel Sabuncu Brain Research Center, Bilkent University, Ankara, Turkey
    We propose a zero-shot learning approach for unsupervised reconstruction of accelerated MRI  without any prior information about reconstruction task. Our approach efficiently recovers undersampled acquisitions, irrespective of the contrast, acceleration rate or undersampling pattern.
    Figure 1: (a) Pretraining of the style-generative model. A fully connected mapper to generate intermediate latent vectors w, a synthesizer to generate images, a discriminator for adversarial training and noise n. w and n are defined for each synthesizer block separately, where block cover resolution from 4x4 to 256x256 pixels. (b) Testing phase of ZSL-Net. w* and n* correspond to optimized latent vector and noise components for the synthesizer. Optimization is performed to minimize partial k-space loss between masked Fourier coefficients of reconstructed and undersampled images.
    Figure 2: Demonstrations of the proposed and competing methods on IXI for T1-contrast image reconstruction when acceleration rate R is 8. Reconstructed images are shown along with the error maps which are absolute differences between reconstructed and reference images. Error map corresponds to ZSL-Net, appears to be darker compared to the competing methods and most of the error concentrated on skull.
  • Multi-Mask Self-Supervised Deep Learning for Highly Accelerated Physics-Guided MRI Reconstruction
    Burhaneddin Yaman1,2, Seyed Amir Hossein Hosseini1,2, Steen Moeller2, Jutta Ellermann2, Kâmil Uğurbil2, and Mehmet Akçakaya1,2
    1University of Minnesota, Minneapolis, MN, United States, 2Center for Magnetic Resonance Research, Minneapolis, MN, United States
    The proposed multi-mask self-supervised physics-guided learning technique significantly improves upon its previously proposed single-mask counterpart for highly accelerated MRI reconstruction.
    Figure 3. A representative test brain MRI slice showing reconstruction results using CG-SENSE, SSDU PG-DL and proposed multi-mask SSDU PG-DL at R=8, as well as CG-SENSE at acquisition acceleration R=2. CG-SENSE suffers from major noise amplification at R=8, whereas SSDU PG-DL at R=8 achieves similar reconstruction quality to CG-SENSE at R=2. The proposed multi-mask SSDU PG-DL further improves the reconstruction quality compared to SSDU PG-DL.
    Figure 2. a) Reconstruction results on a representative test slice at R = 8 using CG-SENSE, supervised PG-DL, SSDU PG-DL and proposed multi-mask SSDU PG-DL. CG-SENSE suffers from major noise amplification and artifacts. SSDU PG-DL also shows residual artifacts (red arrows) at this high acceleration rate. Proposed multi-mask SSDU PG-DL further suppresses these artifacts and achieve artifact-free reconstruction, removing artifacts that are still visible in supervised PG-DL.
  • PIC-GAN: A Parallel Imaging Coupled Generative Adversarial Network for Accelerated Multi-Channel MRI Reconstruction
    Jun Lyu1, Chengyan Wang2, and Guang Yang3,4
    1School of Computer and Control Engineering, Yantai University, Yantai, China, 2Human Phenome Institute, Fudan University, Shanghai, China, 3Cardiovascular Research Centre, Royal Brompton Hospital, London, United Kingdom, 4National Heart and Lung Institute, Imperial College London, London, United Kingdom
    To demonstrate the feasibility of combining parallel imaging (PI) with the generative adversarial network (GAN) for accelerated multi-channel MRI reconstruction. In our proposed PIC-GAN framework, we used a progressive refinement method in both frequency and image domains.
    Figure 1. Schema of the proposed parallel imaging and generative adversarial network (PIC-GAN) reconstruction network.
    Figure 2. Representative reconstructed abdominal images with acceleration AF= 6. The 1st and 2nd rows depict reconstruction results for regular Cartesian sampling, the 3rd and 4th row depict the same for variable density random sampling. The PIC-GAN reconstruction shows reduced artifacts compared to other methods.
  • Adaptive convolutional neural networks for accelerating magnetic resonance imaging via k-space data interpolation
    Tianming Du1, Yuemeng Li1, Honggang Zhang2, Stephen Pickup1, Rong Zhou1, Hee Kwon Song1, and Yong Fan1
    1Radiology, University of Pennsylvania, Philadelphia, PA, United States, 2Beijing University of Posts and Telecommunications, Beijing, China
    A deep learning model with adaptive convolutional neural networks is developed for k-space data interpolation. Ablation and experimental results show that our method achieves better image reconstruction than existing state-of-the-art techniques.
    Figure-1. A residual Encoder-Decoder network of CNNs (top and middle), enhanced by frequency-attention and channel-attention layers (bottom), for image reconstruction from undersampled k-space data.
    Figure-2. Visualization of representative cases of the Stanford knee dataset (top row) and the fastMRI brain dataset (bottom row), including images reconstructed from the fully sampled data and from under-sampled data without CNN processing. The difference images were amplified 5 times for Stanford cases and 10 times for fastMRI dataset. Yellow and red boxes indicate the zoomed-in and difference images, respectively.
  • Accelerated Magnetic Resonance Spectroscopy with Model-inspired Deep Learning
    Zi Wang1, Yihui Huang1, Zhangren Tu2, Di Guo2, Vladislav Orekhov3, and Xiaobo Qu1
    1Department of Electronic Science, National Institute for Data Science in Health and Medicine, Xiamen University, Xiamen, China, 2School of Computer and Information Engineering, Xiamen University of Technology, Xiamen, China, 3Department of Chemistry and Molecular Biology, University of Gothenburg, Gothenburg, Sweden
    The proof-of-concept of the significance of merging the optimization with deep learning, for reliable, high-quality, and ultra-fast accelerated NMR spectroscopy, and provides a relatively explicit understanding of the complex mapping in deep learning.
    Figure 1. MoDern: The proposed Model-inspired Deep Learning framework for NMR spectra reconstruction. The recursive MoDern framework that alternates between the data consistency (DC), which is same to Eq. (1a), and the learnable adaptive soft-thresholding (LS) inspired by Eq. (1b). With the increase of iterations, artifacts are gradually removed, and finally a high-quality reconstructed spectrum can be obtained. Note: “FT” is the Fourier transform. A data consistency followed by a learnable adaptive soft-thresholding constitutes an iteration.
    Figure 2. 2D 1H-15N HSQC spectrum of the cytosolic domain of CD79b protein. (a) The fully sampled spectrum. (b)-(c) are reconstructed spectra using DLNMR and MoDern from 20% data, respectively. (d) is zoomed out 1D 15N traces. (e) is the peak intensity correlation obtained with two methods under different NUS levels. The insets of (b)-(c) show the corresponding peak intensity correlation between fully sampled spectrum and reconstructed spectrum. Note: The average and standard deviations of correlations in (e) are computed over 100 NUS trials.
  • DL2 - Deep Learning + Dictionary Learning-based Regularization for Accelerated 2D Dynamic Cardiac MR Image Reconstruction
    Andreas Kofler1, Tobias Schaeffter1,2,3, and Christoph Kolbitsch1,2
    1Physikalisch-Technische Bundesanstalt, Berlin and Braunschweig, Berlin, Germany, 2School of Imaging Sciences and Biomedical Engineering, King's College London, London, United Kingdom, 3Department of Biomedical Engineering, Technical University of Berlin, Berlin, Germany
    Here, we consider a image reconstruction method for dynamic cardiac MR which combines Convolutional Neural Networks (CNNs) with Dictionary Learning (DL) and Sparse Coding (SC). The combination of CNNs with DL+SC improves the results obtained with CNNs and DL+SC used separately.
    Figure 4: An example of results and corresponding point-wise error-images obtained for an acceleration factor of $$$R=18$$$. From left to right: The initial NUFFT-reconstruction obtained from $$$N_{\theta}=560$$$ radial spokes, the CNN-regularized solution10, the DL+SC-regularized solution7, the proposed method and the $$$kt$$$-SENSE reconstruction obtained from $$$N_{\theta}=3400$$$ radial spokes from which the $$$k$$$-space data was retrospectively simulated.

    Figure 2: Convergence results for the proposed method (red) compared to DL and SC (blue) for the two acceleration factors $$$R=18$$$ and $$$R=9$$$ given by $$$N_{\theta}=560$$$ and $$$N_{\theta}=1130$$$ radial spokes, respectively. The solid lines correspond to the mean value of the respective mesure obtained over the entire test set. The dashed lines correspond to the mean $$$\pm$$$ the standard deviation obtained over the test set.

    Note that the values differ from the ones shown in the Table because here, no masks were used to restrict the calculations to a region of interest.

  • Cardiac Functional Analysis with Cine MRI via Deep Learning Reconstruction
    Eric Z. Chen1, Xiao Chen1, Jingyuan Lyu2, Qi Liu2, Zhongqi Zhang3, Yu Ding2, Shuheng Zhang3, Terrence Chen1, Jian Xu2, and Shanhui Sun1
    1United Imaging Intelligence, Cambridge, MA, United States, 2UIH America, Inc., Houston, TX, United States, 3United Imaging Healthcare, Shanghai, China
    This is the first work to evaluate the cine MRI with deep learning reconstruction for cardiac function analysis. The cardiac functional values obtained from cine MRI with deep learning reconstruction are consistent with values from clinical standard retro-cine MRI.
    Figure 2. Cardiac functional analysis based on DL-cine, CS-cine and retro-cine. Difference with statistical significance (p<0.05) is indicated by the star.
    Figure 1. Examples of reconstructed images from DL-cine, CS-cine and retro-cine MRI. Data were acquired from the same subject but three different scans.
  • Deep Laplacian Pyramid Networks for Fast MRI Reconstruction with Multiscale T1 Priors
    Xiaoxin Li1,2, Xinjie Lou1, Junwei Yang3, Yong Chen4, and Dinggang Shen2,5
    1College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China, 2School of Biomedical Engineering, ShanghaiTech University, Shanghai, China, 3Department of Computer Science and Technology, University of Cambridge, Cambridge, United Kingdom, 4Case Western Reserve University, Cleveland, OH, United States, 5Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China
    To accelerate multimodal Magnetic Resonance Imaging (MRI) acquisitions, we propose a deep Laplacian pyramid MRI reconstruction framework (LapMRI), which performs progressive upsampling while integrating multiscale prior of T1.
    Figure 1: Schematic overview of the proposed LapMRI framework.
    Figure 2: The inputs and outputs of LapMRI(D5C5, Λ=3) at three pyramid levels. For each pyramid level, the two images in the left column are the inputs, and the two images in the right column are the output and the respective ground-truth image, respectively. For visual understanding, the output of each pyramid level is displayed in the image domain, while the respective ground-truth image is framed with a red dotted box.
  • Deep learning-based reconstruction of highly-accelerated 3D MRI MPRAGE images
    Sangtae Ahn1, Uri Wollner2, Graeme McKinnon3, Rafi Brada2, John Huston4, J. Kevin DeMarco5, Robert Y. Shih5,6, Joshua D. Trzasko4, Dan Rettmann7, Isabelle Heukensfeldt Jansen1, Christopher J. Hardy1, and Thomas K. F. Foo1
    1GE Research, Niskayuna, NY, United States, 2GE Research, Herzliya, Israel, 3GE Healthcare, Waukesha, WI, United States, 4Mayo Clinic College of Medicine, Rochester, MN, United States, 5Walter Reed National Military Medical Center, Bethesda, MD, United States, 6Uniformed Services University of the Health Sciences, Bethesda, MD, United States, 7GE Healthcare, Rochester, MN, United States
    Our deep-learning reconstruction, DCI-Net, can accelerate 3D T1-weighted MPRAGE scans by an additional factor of 5 compared to conventional two-fold accelerated parallel acquisition, while maintaining comparable diagnostic image quality.
    Figure 2. Image quality scores averaged over 5 subjects for ARC (net R=2.1) and 3 variants of DCI-Net (net R=10) in 8 scoring categories. The error bars denote the sample standard deviation.
    Figure 3. Example image slices from subject D, showing the cerebellar vermis indicated by the yellow arrow, reconstructed by (A) ARC (net R=2.1), (B) 2D DCI-Net (net R=10), (C) alternating 2D DCI-Net (net R=10), and (D) 3D DCI-Net (net R=10).
  • Improved CNN-based Image reconstruction using regularly under-sampled signal obtained in phase scrambling Fourier transform imaging
    satoshi ITO1 and Shun UEMATSU1
    1Utsunomiya University, Utsunomiya, Japan
    A CNN-based image reconstruction using phase scrambling Fourier transform imaging was proposed and demonstrated. It was shown that proposed method showed that preservation of structure and image contrast were improved compared to standard Fourier transform based CS-CNN.
    Figure 4. Results of reconstructing spatially phase varied images. Figure (a) and (b) show the phase map and magnitude of fully scanned image. Figure (c) and (d) are reconstructed images with EsUS and RaUS, respectively. Enlarged images corresponding to (a) through (d) are shown in (e) through (h).
    Figure 2. Comparison of PSNR and SSIM between PSFT-CS-Net and FT-CS-Net for real-value images. PSNR results using sampling pattern of Fig.1 (a) are shown. Comparison among CNN reconstruction using PSFT or FT imaging and conventional iterative reconstruction were made.
  • Progressive Volumetrization for Data-Efficient Image Recovery in Accelerated Multi-Contrast MRI
    Mahmut Yurt1,2, Muzaffer Ozbey1,2, Salman Ul Hassan Dar1,2, Berk Tinaz1,2,3, Kader Karlı Oğuz2,4, and Tolga Çukur1,2,5
    1Department of Electrical and Electronics Engineering, Bilkent University, Ankara, Turkey, 2National Magnetic Resonance Research Center, Bilkent University, Ankara, Turkey, 3Department of Electrical and Computer Engineering, University of Southern California, Los Angeles, CA, United States, 4Department of Radiology, Hacettepe University, Ankara, Turkey, 5Neuroscience Program, Aysel Sabuncu Brain Research Center, Bilkent University, Ankara, Turkey
    We propose a progressively volumetrized generative model for efficient context learning in 3D multi-contrast MRI accelerated across contrast sets or k-space coefficients. The proposed method decomposes volumetric recovery tasks into a sequence of less complex cross-sectional subtasks.

    Fig. 1: ProvoGAN performs a series of cross-sectional subtasks optimally-ordered across individual rectilinear orientations (Axial$$$\rightarrow$$$Sagittal$$$\rightarrow$$$Coronal illustrated here) to handle the aimed volumetric recovery task. Within a given subtask, source-contrast volume is divided into cross-sections across the longitudinal dimension, and a cross-sectional mapping is learned to recover target cross-sections from source cross-sections, where the previous subtask's (if available) output is further incorporated to leverage contextual priors.

    Fig. 2: ProvoGAN is demonstrated on IXI for T1-weighted image synthesis. Representative results from ProvoGAN, sGAN models (sGAN-A is trained axially, sGAN-C coronally, and sGAN-S sagittally), and vGAN are displayed for all rectilinear orientations (first row: axial, second row: coronal, third row: sagittal) together with reference images.
  • Enhanced Multi-Slice Partial Fourier MRI Reconstruction Using Residual Network
    Linshan Xie1,2, Yilong Liu1,2, Linfang Xiao1,2, Peibei Cao1,2, Alex T. L. Leong1,2, and Ed X. Wu1,2
    1Laboratory of Biomedical Imaging and Signal Processing, The University of Hong Kong, Hong Kong, China, 2Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, China
    A residual network based reconstruction method is proposed for multi-slice partial Fourier acquisition, where adjacent slices are sampled in a complementary way. It enables highly partial Fourier imaging without losing image details or significant noise amplification.
    Figure 1 The flowchart of the proposed EMS-PF reconstruction method, with adjacent slices having complementary sampling patterns. Re and Im denote the real and imaginary parts of the reconstructed slice (Slice 2) and its adjacent slices (Slices 1 and 3); Re2’ and Im2’ denote the real and imaginary parts of the predicted residual image. The ResNet model has 16 RBs and each of them contains 2 convolutional layers followed by Rectified Linear Unit (ReLU) activation function. In each convolutional layer, 64 convolutional kernels of size 3×3 each are included.
    Figure 3 Error maps of the reconstructed images in Figure 2 with enhanced brightness (×10) and corresponding peak signal-to-noise ratio (PSNR) / structural similarity (SSIM) at PF fraction=0.51. It was obvious that the proposed EMS-PF outperformed the other methods in terms of reduced residual error.
  • Joint-ISTA-Net: A model-based deep learning network for multi-contrast CS-MRI reconstruction
    Yuan Lian1, Xinyu Ye1, Yajing Zhang2, and Hua Guo1
    1Center for Biomedical Imaging Research, Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China, 2MR Clinical Science, Philips Healthcare, Suzhou, China
    We design a deep learning model Joint-ISTA-Net, which exploits the group sparsity of multi-contrast MR images in our model to improve the reconstruction quality of Compressed Sensing.
    Fig 1. Structure of proposed Joint-ISTA-Net and a iterative Share Block. In Share Block, three 3x3 CNN layers with Relu stands forward sparse transformation. Figures in transform domain pass through both individual and joint soft threshold function to be denoised, and then decoded back to image domain. Here joint soft threshold function is implemented to exploit the group sparsity property of multi-contrast MR images.
    Fig 2. Comparison of reconstruction method with R=10. Proposed Joint-ISTA-Net has advantages on feature preserving and provides sharper edge.
  • Training- and Database-free Deep Non-Linear Inversion (DNLINV) for Highly Accelerated Parallel Imaging and Calibrationless PI&CS MR Imaging
    Andrew Palmera Leynes1,2 and Peder E.Z. Larson1,2
    1Department of Radiology and Biomedical Imaging, University of California San Francisco, San Francisco, CA, United States, 2UC Berkeley - UC San Francisco Joint Graduate Program in Bioengineering, Berkeley and San Francisco, CA, United States
    We introduce Deep Non-Linear Inversion (DNLINV), a deep image reconstruction approach that may be used with any hardware and acquisition configuration. We demonstrate DNLINV on different anatomies and sampling patterns and show high quality reconstructions at higher acceleration factors.
    Figure 4. Calibrationless parallel imaging and compressed sensing on a T1-weighted brain image. All methods were able to successfully reconstruct the image at R=4.0. However, at R=8.5, only DNLINV was able to reconstruct the image without any loss of structure. Furthermore, DNLINV reconstructions have higher apparent SNR.
    Figure 5. Autocalibrating parallel imaging with CAIPI sampling on a T1-weighted brain image. All methods were able to successfully reconstruct the image at R=16.0 with DNLINV having the highest apparent SNR. At R=25.0, residual aliasing artifacts remain on ESPIRiT and ENLIVE while these are largely suppressed in DNLINV.
  • A Modified Generative Adversarial Network using Spatial and Channel-wise Attention for Compressed Sensing MRI Reconstruction
    Guangyuan Li1, Chengyan Wang2, Weibo Chen3, and Jun Lyu1
    1School of Computer and Control Engineering, Yantai University, Yantai, China, 2Human Phenome Institute, Fudan University, Shanghai, China, 3Philips Healthcare, Shanghai, China
    In order to solve the reconstruction effect of CS-MRI under highly under-sampling,we proposed a modified GAN architecture for accelerating CS-MRI reconstruction, namely RSCA-GAN,and added spatial and channel-wise attention in Generative Adversarial Networks.
    Fig.1.(a) Framework of the proposed method.The generator of the network is connected with two residual autoencoder U-net. The discriminator is composed of 6 layers. (b) Composition of KS-Block.
    Fig.2.The architecture of Residual SCAU-Net.The encode block is indicated by green, and the decode block is indicated by blue. The 4D tensor is used as input, and using the 2D convolution with filter_size of 3x3 and Stride of 2. The number of feature maps is defined as feature_num=64. The residual block is represented by orange, which is used to increase the depth of the network. SCA Block is indicated by yellow composed of spatial attention and channel attention.
  • Compressed sensing MRI via a fusion model based on image and gradient priors
    Yuxiang Dai1, Cheng yan Wang2, and He Wang1
    1Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai, China, 2Human Phenome Institute, Fudan University, Shanghai, China
    We proposed a fusion model based on the optimization method to integrate the image and gradient-based priors into CS-MRI for better reconstruction results via convolutional neural network models. In addition, the proposed fusion model exhibited effective reconstruction performance in MRA.
    Figure 1 The framework of the proposed fusion model in which the above network is MDN and the below network is SRLN. $$$N_f$$$ represents the number of convolutional kernels, DF represents the dilated factor of convolutional kernel. $$$c_n$$$ and $$$res_n$$$ represent n-th convolution layer and residual learning respectively.
    Figure 2 Reconstruction results for 30% radial sampling. The first and third row include groundtruth and reconstruction results of CSMRI methods. The second and fourth row include radial sampling mask and errors. Values of PSNR and SSIM are shown in the upper left corner.
  • Kernel-based Fast EPTI Reconstruction with Neural Network
    Muheng Li1, Jie Xiang2, Fuyixue Wang3,4, Zijing Dong3,5, and Kui Ying2
    1Department of Automation, Tsinghua University, Beijing, China, 2Department of Engineering Phycics, Tsinghua University, Beijing, China, 3A. A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Charlestown, MA, United States, 4Harvard-MIT Health Sciences and Technology, MIT, Cambridge, MA, United States, 5Department of Electrical Engineering and Computer Science, MIT, Cambridge, MA, United States
    Through image reconstruction tests on human brain data set acquired by EPTI, we demonstrated the high efficiency of the kernel-based reconstruction with neural network by shortening the reconstruction time of 216×216×48×32 k-data from over 10 minutes to about 20 seconds.
    Figure2. Process of restoring the missing k-data in the specified kernel. Extract the acquired data in this kernel and all the data in the target region as a 1D vector respectively. The mapping function can be fitted based on the fully sampled calibration data.
    Figure4. Reconstructed images with different (a)loss function: MSE, MAE, Huber (b)number of nodes in each hidden layer (c)multi-contrast and reference images by conventional linear algorithm.
  • Wave-Encoded Model-Based Deep Learning with Joint Reconstruction and Segmentation
    Jaejin Cho1,2, Qiyuan Tian1,2, Robert Frost1,2, Itthi Chatnuntawech3, and Berkin Bilgic1,2,4
    1Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States, 2Havard Medical School, Cambridge, MA, United States, 3National Nanotechnology Center, Pathum Thani, Thailand, 4Harvard/MIT Health Sciences and Technology, Cambridge, MA, United States
    Simultaneously training wave-encoded model-based deep learning (wave-MoDL) with hybrid k- and image-space priors and a U-net enables high-fidelity image reconstruction and segmentation performance at high acceleration rates.
    Figure 1. a. the proposed network architecture for joint MRI reconstruction and segmentation with model-based deep learning and U-net. b. the reconstruction scheme using model-based deep learning for cartesian and wave-encodings.
    Figure 5. Reconstructed images using SENSE, MoDL, wave-CAIPI and wave-MoDL at RyxRz=4x4 on 32-channel HCP data.