Learning-Based Image Reconstruction Learning IR

Learning-Based Image Reconstruction Learning IR

J. Fessler Learning-based image reconstruction Learning IR Jeffrey A. Fessler EECS Department, BME Department, Dept. of Radiology University of Michigan http://web.eecs.umich.edu/~fessler RSNA 2020 Declaration: No relevant financial interests or relationships to disclose 1 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 2 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 3 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret 4 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret Most obvious place for machine learning is in post-processing (image analysis). Numerous special issues and surveys in medical imaging journals, e.g., [1–9]. 4 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret Machine learning for scan design Choose best k-space phase encoding locations based on training images Hot topic in MRI recently [10–15]. Precursor by Yue Cao and David Levin, MRM Sep. 1993 [16–18]. 4 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret Machine learning in medical image reconstruction June 2018 special issue of IEEE Trans. on Medical Imaging [19]. Surveys: [20–27] Possibly easier than diagnosis due to lower bar: • current reconstruction methods based on simplistic image models; • human eyes are better at detection than at solving inverse problems. 4 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret ML-based “magic” A holy grail for machine learning in medical imaging? I CT sinogram to vessel diameter [28, 29] I k-space to ??? 4 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 5 / 43 J. Fessler Generations of medical image reconstruction methods Learning IR 1. 70’s “Analytical” methods (integral equations) FBP for SPECT / PET / X-ray CT, IFFT for MRI, ... 2. 80’s Algebraic methods (as in “linear algebra”) Solve y = Ax 3. 90’s Statistical methods • LS / ML methods based on imaging physics (“model based”) • Bayesian methods (Markov random fields, ...) • regularized methods 4. 00’s Compressed sensing methods (mathematical sparsity models) 5. 10’s Adaptive / data-driven methods machine learning, deep learning, ... 6 / 43 J. Fessler Two important milestones for clinical CT Learning IR • Model-based image reconstruction (MBIR) FDA approved circa 2012 [30] • Deep-learning image reconstruction FDA approved 2019 [31, 32] 7 / 43 J. Fessler Machine-learning approaches to image reconstruction Learning IR I Learn models (sparsifying transform or dictionary) for image patches from training data • interpretable (?) optimization formulations • local prior information only (patch size) • perhaps slower computation due to optimization iterations I Train neural network (aka deep learning) • less interpretable • possibly more global prior information • slow training, but perhaps faster computation after trained 8 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 9 / 43 J. Fessler Patch-based regularization and TV Learning IR Anisotropic discrete TV regularizer: R(x) = kTxk 1 -1 1 where T is finite-differences ≡ patches of size 2 × 1. -1 1 Larger patches provide more context for distinguishing signal from noise. 3 2 cf. CNN approaches 1 Patch-based regularizers: 0 • synthesis models • analysis methods -1 10 / 43 J. Fessler X-ray CT with learned sparsifying transforms Learning IR I Data I Population adaptive methods I Patient adaptive methods I Spatial structure I Patch-based models I Convolutional models I Regularizer formulation I Synthesis (dictionary) approach I Analysis (sparsifying transform) approach 11 / 43 J. Fessler Patch-wise transform sparsity model Learning IR Assumption: if x is a plausible image, then each patch transform TPmx is sparse. I Pmx extracts the mth of M patches from x I T is a (often square) sparsifying transform matrix. What T? 12 / 43 J. Fessler Sparsifying transform learning (population adaptive) Learning IR Given training images x1,..., xL from a representative population, find transform T∗ that best sparsifies their patches: L M X X 2 T∗ = arg min min kTPmxl − zl,mk2 + α kzl,mk0 T unitary {zl,m} l=1 m=1 I Encourage aggregate sparsity, not patch-wise sparsity (cf K-SVD [33]) I Non-convex due to unitary constraint and k·k0 I Efficient alternating minimization algorithm [34] • z update : simple hard thresholding • T update : orthogonal Procrustes problem (SVD) • Subsequence convergence guarantees [34] 13 / 43 J. Fessler Example of learned sparsifying transform Learning IR 3D X-ray training data (XCAT phantom) Parts of learned sparsifier T∗ (2D slices in x-y, x-z, y-z, from 3D image volume) 3 3 8 × 8 × 8 patches =⇒ T∗ is 8 × 8 = 512 × 512 top 8 × 8 slice of 256 of the 512 rows of T ↑ ∗ 14 / 43 J. Fessler Regularizer based on learned sparsifying transform Learning IR Regularized inverse problem [35]: 2 xˆ = arg min kAx − ykW + β R(x) x M X 2 R(x) = min kT∗Pmx − zmk2 + α kzmk0 . {z } m m=1 T∗ adapted to population training data Alternating minimization optimizer: I zm update : simple hard thresholding I x update : quadratic problem (many options) Linearized augmented Lagrangian method (LALM) [36] 15 / 43 J. Fessler Example: low-dose 3D X-ray CT simulation Learning IR X. Zheng, S. Ravishankar, FDK Y. Long, JF: IEEE T-MI, June 2018 [35]. PWLS-EP PWLS-ULTRA 16 / 43 J. Fessler 3D X-ray CT simulation Error maps Learning IR 100 100 100 FDK Error PWLS-EP Error PWLS-ULTRA Error 0 0 0 X-ray Intensity FDK EP ST T∗ ULTRA ULTRA-{τj } RMSE in HU 1 × 104 67.8 34.6 32.1 30.7 29.2 5 × 103 89.0 41.1 37.3 35.7 34.2 I Physics / statistics provides dramatic improvement I Data adaptive regularization further reduces RMSE 17 / 43 J. Fessler Union of Learned TRAnsforms (ULTRA) Learning IR Given training images x1,..., xL from a representative population, find a set of n oK transforms Tˆk that best sparsify image patches: k=1 L M ! n o X X 2 Tˆk = arg min min min kTk Pmxl − zl,mk + α kzl,mk k∈{1,...,K} 2 0 {Tk unitary} {zl,m} l=1 m=1 I Joint unsupervised clustering / sparsification I Further nonconvexity due to clustering I Efficient alternating minimization algorithm [37] 18 / 43 J. Fessler Example: 3D X-ray CT learned set of transforms Learning IR Class 1 Class 2 Class 3 Class 4 Class 5 19 / 43 X. Zheng, S. Ravishankar, Y. Long, JF: IEEE T-MI, June 2018 [35] J. Fessler Example: 3D X-ray CT ULTRA for chest scan Learning IR FDK PWLS-EP PWLS-ULTRA Zheng et al., IEEE T-MI, June 2018 [35] (Special issue on machine learning for image reconstruction) Matlab code: http://web.eecs.umich.edu/~fessler/irt/reproduce/ https://github.com/xuehangzheng/PWLS-ULTRA-for-Low-Dose-3D-CT-Image-Reconstruction 20 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 21 / 43 J. Fessler Deep-learning approaches to image reconstruction Learning IR Overview: I image-domain learning [38–40]... I k-space or data-domain learning e.g., [41], [42], [43] I transform learning (direct from k-space to image) e.g., AUTOMAP [44], [45–47] I hybrid-domain learning (unrolled loop, e.g., variational network) alternate between denoising/dealiasing and reconstruction from k-space e.g., [42, 48–52] ... 22 / 43 J. Fessler DL for IR: image-domain learning Learning IR Figure courtesy of Jong Chul Ye, KAIST University. + simple and fast − aliasing is spatially widespread, requires deep network 23 / 43 J. Fessler Dangers of image-domain learningI Learning IR [53] ISMRM 2020 Workshop on Data Sampling & Image Reconstruction 24 / 43 J. Fessler Dangers of image-domain learningII Learning IR 25 / 43 J. Fessler Image-domain learning variations Learning IR I Use NN output as a “prior” for iterative reconstruction [38, 54]: 2 2 0 −1 0 xˆβ = arg min kAx − yk2 + β kx − xNNk2 = (A A + βI) (A y + βxNN) x I For single-coil Cartesian case: • no iterations are needed (solve with FFTs) • limβ→0 xˆβ replaces missing k-space data with FFT of xNN I Iterations needed for parallel MRI and/or non-Cartesian sampling (PCG) I Learn residual (aliasing artifacts), then subtract [55, 56] 26 / 43 J. Fessler DL for IR: k-space / sinogram domain learning Learning IR Figure courtesy of Jong Chul Ye, KAIST University. + simple and fast (“nonlinear GRAPPA”) + “database-free” : learn from auto-calibration data − perhaps harder to represent local image features? 27 / 43 J. Fessler DL for IR: transform learning Learning IR Figure courtesy of Jong Chul Ye, KAIST University. + in principle, purely data driven; potential to avoid model mismatch − high memory requirement for fully connected layers 28 / 43 J. Fessler DL for IR: hybrid domain learning Learning IR Figure courtesy of Jong Chul Ye, KAIST University. + physics-based use of k-space data & image-domain priors + interpretable connections to optimization approaches − more computation to due to “iterations” (layers) and repeated Ax, A0 r 29 / 43 J. Fessler Convolutional sparsity revisted Learning IR Cost function for convolutional sparsity regularization: 1 2 PK 1 2 arg min 2 kAx − ykW + β min k=1 2 khk ∗ x − ζk k2 + α kζk k1 x ζ Alternating minimization, aka block coordinate descent (BCD), updates: (n+1) (n) Sparse code: ζk = soft hk ∗ x , α Image: x (n+1) = arg min F (x; y, z (n)) x 2 (n) 1 2 PK 1 (n+1) (n+1) F (x; y, z ) kAx − yk + β hk ∗ x − ζ + α ζ , 2 W k=1 2 k 2 k 1 1 2 1 (n) 2 = 2 kAx − ykW + β 2 x − z 2 (quadratic but large =⇒ majorize) (n) (n) PK (n) z = R(z ) = k=1 flip(hk ) ∗ soft{hk ∗ x } (denoise =⇒ learn) 30 / 43 J.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    47 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us