
J. Fessler Learning-based image reconstruction Learning IR Jeffrey A. Fessler EECS Department, BME Department, Dept. of Radiology University of Michigan http://web.eecs.umich.edu/~fessler RSNA 2020 Declaration: No relevant financial interests or relationships to disclose 1 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 2 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 3 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret 4 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret Most obvious place for machine learning is in post-processing (image analysis). Numerous special issues and surveys in medical imaging journals, e.g., [1–9]. 4 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret Machine learning for scan design Choose best k-space phase encoding locations based on training images Hot topic in MRI recently [10–15]. Precursor by Yue Cao and David Levin, MRM Sep. 1993 [16–18]. 4 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret Machine learning in medical image reconstruction June 2018 special issue of IEEE Trans. on Medical Imaging [19]. Surveys: [20–27] Possibly easier than diagnosis due to lower bar: • current reconstruction methods based on simplistic image models; • human eyes are better at detection than at solving inverse problems. 4 / 43 J. Fessler Medical imaging overview Learning IR raw data images y xˆ Design → Acquire −−−→ Reconstruct −−−−→ Process → Diagnose Scan Data Images Images Interpret ML-based “magic” A holy grail for machine learning in medical imaging? I CT sinogram to vessel diameter [28, 29] I k-space to ??? 4 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 5 / 43 J. Fessler Generations of medical image reconstruction methods Learning IR 1. 70’s “Analytical” methods (integral equations) FBP for SPECT / PET / X-ray CT, IFFT for MRI, ... 2. 80’s Algebraic methods (as in “linear algebra”) Solve y = Ax 3. 90’s Statistical methods • LS / ML methods based on imaging physics (“model based”) • Bayesian methods (Markov random fields, ...) • regularized methods 4. 00’s Compressed sensing methods (mathematical sparsity models) 5. 10’s Adaptive / data-driven methods machine learning, deep learning, ... 6 / 43 J. Fessler Two important milestones for clinical CT Learning IR • Model-based image reconstruction (MBIR) FDA approved circa 2012 [30] • Deep-learning image reconstruction FDA approved 2019 [31, 32] 7 / 43 J. Fessler Machine-learning approaches to image reconstruction Learning IR I Learn models (sparsifying transform or dictionary) for image patches from training data • interpretable (?) optimization formulations • local prior information only (patch size) • perhaps slower computation due to optimization iterations I Train neural network (aka deep learning) • less interpretable • possibly more global prior information • slow training, but perhaps faster computation after trained 8 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 9 / 43 J. Fessler Patch-based regularization and TV Learning IR Anisotropic discrete TV regularizer: R(x) = kTxk 1 -1 1 where T is finite-differences ≡ patches of size 2 × 1. -1 1 Larger patches provide more context for distinguishing signal from noise. 3 2 cf. CNN approaches 1 Patch-based regularizers: 0 • synthesis models • analysis methods -1 10 / 43 J. Fessler X-ray CT with learned sparsifying transforms Learning IR I Data I Population adaptive methods I Patient adaptive methods I Spatial structure I Patch-based models I Convolutional models I Regularizer formulation I Synthesis (dictionary) approach I Analysis (sparsifying transform) approach 11 / 43 J. Fessler Patch-wise transform sparsity model Learning IR Assumption: if x is a plausible image, then each patch transform TPmx is sparse. I Pmx extracts the mth of M patches from x I T is a (often square) sparsifying transform matrix. What T? 12 / 43 J. Fessler Sparsifying transform learning (population adaptive) Learning IR Given training images x1,..., xL from a representative population, find transform T∗ that best sparsifies their patches: L M X X 2 T∗ = arg min min kTPmxl − zl,mk2 + α kzl,mk0 T unitary {zl,m} l=1 m=1 I Encourage aggregate sparsity, not patch-wise sparsity (cf K-SVD [33]) I Non-convex due to unitary constraint and k·k0 I Efficient alternating minimization algorithm [34] • z update : simple hard thresholding • T update : orthogonal Procrustes problem (SVD) • Subsequence convergence guarantees [34] 13 / 43 J. Fessler Example of learned sparsifying transform Learning IR 3D X-ray training data (XCAT phantom) Parts of learned sparsifier T∗ (2D slices in x-y, x-z, y-z, from 3D image volume) 3 3 8 × 8 × 8 patches =⇒ T∗ is 8 × 8 = 512 × 512 top 8 × 8 slice of 256 of the 512 rows of T ↑ ∗ 14 / 43 J. Fessler Regularizer based on learned sparsifying transform Learning IR Regularized inverse problem [35]: 2 xˆ = arg min kAx − ykW + β R(x) x M X 2 R(x) = min kT∗Pmx − zmk2 + α kzmk0 . {z } m m=1 T∗ adapted to population training data Alternating minimization optimizer: I zm update : simple hard thresholding I x update : quadratic problem (many options) Linearized augmented Lagrangian method (LALM) [36] 15 / 43 J. Fessler Example: low-dose 3D X-ray CT simulation Learning IR X. Zheng, S. Ravishankar, FDK Y. Long, JF: IEEE T-MI, June 2018 [35]. PWLS-EP PWLS-ULTRA 16 / 43 J. Fessler 3D X-ray CT simulation Error maps Learning IR 100 100 100 FDK Error PWLS-EP Error PWLS-ULTRA Error 0 0 0 X-ray Intensity FDK EP ST T∗ ULTRA ULTRA-{τj } RMSE in HU 1 × 104 67.8 34.6 32.1 30.7 29.2 5 × 103 89.0 41.1 37.3 35.7 34.2 I Physics / statistics provides dramatic improvement I Data adaptive regularization further reduces RMSE 17 / 43 J. Fessler Union of Learned TRAnsforms (ULTRA) Learning IR Given training images x1,..., xL from a representative population, find a set of n oK transforms Tˆk that best sparsify image patches: k=1 L M ! n o X X 2 Tˆk = arg min min min kTk Pmxl − zl,mk + α kzl,mk k∈{1,...,K} 2 0 {Tk unitary} {zl,m} l=1 m=1 I Joint unsupervised clustering / sparsification I Further nonconvexity due to clustering I Efficient alternating minimization algorithm [37] 18 / 43 J. Fessler Example: 3D X-ray CT learned set of transforms Learning IR Class 1 Class 2 Class 3 Class 4 Class 5 19 / 43 X. Zheng, S. Ravishankar, Y. Long, JF: IEEE T-MI, June 2018 [35] J. Fessler Example: 3D X-ray CT ULTRA for chest scan Learning IR FDK PWLS-EP PWLS-ULTRA Zheng et al., IEEE T-MI, June 2018 [35] (Special issue on machine learning for image reconstruction) Matlab code: http://web.eecs.umich.edu/~fessler/irt/reproduce/ https://github.com/xuehangzheng/PWLS-ULTRA-for-Low-Dose-3D-CT-Image-Reconstruction 20 / 43 J. Fessler Outline Learning IR Introduction Image reconstruction Adaptive regularization Deep-learning approaches Summary Bibliography 21 / 43 J. Fessler Deep-learning approaches to image reconstruction Learning IR Overview: I image-domain learning [38–40]... I k-space or data-domain learning e.g., [41], [42], [43] I transform learning (direct from k-space to image) e.g., AUTOMAP [44], [45–47] I hybrid-domain learning (unrolled loop, e.g., variational network) alternate between denoising/dealiasing and reconstruction from k-space e.g., [42, 48–52] ... 22 / 43 J. Fessler DL for IR: image-domain learning Learning IR Figure courtesy of Jong Chul Ye, KAIST University. + simple and fast − aliasing is spatially widespread, requires deep network 23 / 43 J. Fessler Dangers of image-domain learningI Learning IR [53] ISMRM 2020 Workshop on Data Sampling & Image Reconstruction 24 / 43 J. Fessler Dangers of image-domain learningII Learning IR 25 / 43 J. Fessler Image-domain learning variations Learning IR I Use NN output as a “prior” for iterative reconstruction [38, 54]: 2 2 0 −1 0 xˆβ = arg min kAx − yk2 + β kx − xNNk2 = (A A + βI) (A y + βxNN) x I For single-coil Cartesian case: • no iterations are needed (solve with FFTs) • limβ→0 xˆβ replaces missing k-space data with FFT of xNN I Iterations needed for parallel MRI and/or non-Cartesian sampling (PCG) I Learn residual (aliasing artifacts), then subtract [55, 56] 26 / 43 J. Fessler DL for IR: k-space / sinogram domain learning Learning IR Figure courtesy of Jong Chul Ye, KAIST University. + simple and fast (“nonlinear GRAPPA”) + “database-free” : learn from auto-calibration data − perhaps harder to represent local image features? 27 / 43 J. Fessler DL for IR: transform learning Learning IR Figure courtesy of Jong Chul Ye, KAIST University. + in principle, purely data driven; potential to avoid model mismatch − high memory requirement for fully connected layers 28 / 43 J. Fessler DL for IR: hybrid domain learning Learning IR Figure courtesy of Jong Chul Ye, KAIST University. + physics-based use of k-space data & image-domain priors + interpretable connections to optimization approaches − more computation to due to “iterations” (layers) and repeated Ax, A0 r 29 / 43 J. Fessler Convolutional sparsity revisted Learning IR Cost function for convolutional sparsity regularization: 1 2 PK 1 2 arg min 2 kAx − ykW + β min k=1 2 khk ∗ x − ζk k2 + α kζk k1 x ζ Alternating minimization, aka block coordinate descent (BCD), updates: (n+1) (n) Sparse code: ζk = soft hk ∗ x , α Image: x (n+1) = arg min F (x; y, z (n)) x 2 (n) 1 2 PK 1 (n+1) (n+1) F (x; y, z ) kAx − yk + β hk ∗ x − ζ + α ζ , 2 W k=1 2 k 2 k 1 1 2 1 (n) 2 = 2 kAx − ykW + β 2 x − z 2 (quadratic but large =⇒ majorize) (n) (n) PK (n) z = R(z ) = k=1 flip(hk ) ∗ soft{hk ∗ x } (denoise =⇒ learn) 30 / 43 J.
Details
-
File Typepdf
-
Upload Time-
-
Content LanguagesEnglish
-
Upload UserAnonymous/Not logged-in
-
File Pages47 Page
-
File Size-