2014_Bayesian Nonparametric Dictionary Learning for
2014_Bayesian Nonparametric Dictionary Learning for
Abstract—We develop a Bayesian nonparametric model for Motivated by the need to find a sparse domain for repre-
reconstructing magnetic resonance images (MRI) from highly senting the MR signal, a large body of literature now exists
undersampled k-space data. We perform dictionary learning as on reconstructing MRI from significantly undersampled k-
part of the image reconstruction process. To this end, we use
the beta process as a nonparametric dictionary learning prior space data. Existing improvements in CS-MRI mostly focus
arXiv:1302.2712v3 [cs.CV] 26 Jul 2014
for representing an image patch as a sparse combination of on (i) seeking sparse domains for the image, such as con-
dictionary elements. The size of the dictionary and the patch- tourlets [4], [5]; (ii) using approximations of the `0 norm for
specific sparsity pattern are inferred from the data, in addition better reconstruction performance with fewer measurements,
to other dictionary learning variables. Dictionary learning is for example `1 , FOCUSS, `p quasi-norms with 0 < p < 1,
performed directly on the compressed image, and so is tailored
to the MRI being considered. In addition, we investigate a total or using smooth functions to approximate the `0 norm [6],
variation penalty term in combination with the dictionary learn- [7]; and (iii) accelerating image reconstruction through more
ing model, and show how the denoising property of dictionary efficient optimization techniques [8], [10], [29]. In this paper
learning removes dependence on regularization parameters in we present a modeling framework that is similarly motivated.
the noisy setting. We derive a stochastic optimization algorithm CS-MRI reconstruction algorithms tend to fall into two
based on Markov Chain Monte Carlo (MCMC) for the Bayesian
model, and use the alternating direction method of multipliers categories: Those which enforce sparsity directly within some
(ADMM) for efficiently performing total variation minimization. image transform domain [3]–[8], [10], [11], [12], and those
We present empirical results on several MRI, which show that the which enforce sparsity in some underlying latent representa-
proposed regularization framework can improve reconstruction tion of the image, such as an adaptive dictionary-based rep-
accuracy over other methods. resentation [9], [14]. Most CS-MRI reconstruction algorithms
Index Terms—compressed sensing, magnetic resonance imag- belong to the first category. For example Sparse MRI [3], the
ing, Bayesian nonparametrics, dictionary learning leading study in CS-MRI, performs MR image reconstruction
by enforcing sparsity in both the wavelet domain and the total
I. I NTRODUCTION variation (TV) of the reconstructed image. Algorithms with
image-level sparsity constraints such as Sparse MRI typically
Magnetic resonance imaging (MRI) is a widely used tech- employ an off-the-shelf basis, which can usually capture only
nique for visualizing the structure and functioning of the one feature of the image. For example, wavelets recover point-
body. A limitation of MRI is its slow scan speed during like features, while contourlets recover curve-like features.
data acquisition. Therefore, methods for accelerating the MRI Since MR images contain a variety of underlying features,
process have been heavily researched. Recent advances in such as edges and textures, using a basis not adapted to the
signal reconstruction from measurements sampled below the image can be considered a drawback of these algorithms.
Nyquist rate, called compressed sensing (CS) [1], [2], have had Finding a sparse basis that is suited to the image at hand
a major impact on MRI [3]. CS-MRI allows for significant can benefit MR image reconstruction, since CS theory shows
undersampling in the Fourier measurement domain of MR that the required number of measurements is linked to the
images (called k-space), while still outputting a high-quality sparsity of the signal in the selected transform domain. Using
image reconstruction. While image reconstruction using this a standard basis not adapted to the image under consideration
undersampled data is a case of an ill-posed inverse problem, will likely not provide a representation that can compete in
compressed sensing theory has shown that it is possible to sparsity with an adapted basis. To this end, dictionary learning,
reconstruct a signal from significantly fewer measurements which falls in the second group of algorithms, learns a sparse
than mandated by traditional Nyquist sampling if the signal basis on image subregions called patches that is adapted
is sparse in a particular transform domain. to the image class of interest. Recent studies in the image
processing literature have shown that dictionary learning is
This work supported by the National Natural Science Foundation of China
(Nos. 30900328, 61172179, 61103121, 81301278), the Fundamental Research an effective means for finding a sparse, patch-level represen-
Funds for the Central Universities (Nos. 2011121051, 2013121023) and the tation of an image [19], [20], [25]. These algorithms learn
Natural Science Foundation of Fujian Province of China (No. 2012J05160). a patch-level dictionary by exploiting structural similarities
Y. Huang, Q. Lin, X. Ding and X. Fu are with the Department of
Communications Engineering at Xiamen University in Xiamen, Fujian, China. between patches extracted from images within a class of inter-
J. Paisley is with the Department of Electrical Engineering at Columbia est. Among these approaches, adaptive dictionary learning—
University in New York, NY, USA. where the dictionary is learned directly from the image being
X.-P. Zhang is with the Department of Electrical and Computer Engineering
at Ryerson University in Toronto, Canada. considered—based on patch-level sparsity constraints usually
† Equal contributions. ‡ Corresponding author: [email protected] outperforms analytical dictionary approaches in denoising,
2
super-resolution reconstruction, interpolation, inpainting, clas- introduction, the function h can take several forms, but tends
sification and other applications, since the adaptively learned to fall into one of two categories according to whether image-
dictionary suits the signal of interest [19]–[22]. level or patch-level information is considered. We next review
Dictionary learning has previously been applied to CS- these two approaches.
MRI to learn a sparse basis for reconstruction, e.g., [14]. 1) Image-level sparse regularization: CS-MRI with an
With these methods, parameters such as the dictionary size image-level, or global regularization function hg (x) is one in
and patch sparsity are preset, and algorithms are considered which sparsity is enforced within a transform domain defined
that are non-Bayesian. In this paper, we consider a new on the entire image. For example, in Sparse MRI [3] the
dictionary learning algorithm for CS-MRI that is motivated regularization function is
by Bayesian nonparametric statistics. Specifically, we consider
hg (x) = kW xk1 + µ T V (x), (2)
a nonparametric dictionary learning model called BPFA [23]
that uses the beta process to learn the sparse representation where W is the wavelet basis and T V (x) is the total variation
necessary for CS-MRI reconstruction. The beta process is (spatial finite differences) of the image. Regularizing with
an effective prior for nonparametric learning of latent factor this function requires that the image be sparse in the wavelet
models; in this case the latent factors correspond to dictionary domain, as measured by the `1 norm of the wavelet coefficients
elements. While the dictionary size is therefore infinite in kW xk1 , which acts as a surrogate for `0 [1], [2]. The total
principle, through posterior inference the beta process learns a variation term enforces homogeneity within the image by
suitably compact dictionary in which the signal can be sparsely encouraging neighboring pixels to have similar values while
represented. allowing for sudden high frequency jumps at edges. The
We organize the paper as follows. In Section II we review parameter µ > 0 controls the trade-off between the two terms.
CS-MRI inversion methods and the beta process for dictionary A variety of other image-level regularization approaches have
learning. In Section III, we describe the proposed regulariza- been proposed along these lines, e.g., [4], [5], [7].
tion framework and algorithm. We derive a Markov Chain 2) Patch-level sparse regularization: An alternative to the
Monte Carlo (MCMC) sampling algorithm for stochastic opti- image-level sparsity constraint hg (x) is a patch-level, or local
mization of the dictionary variables in the objective function. regularization function hl (x), which enforces that patches
In addition, we consider including a sparse total variation (TV) (square sub-regions of the image) have a sparse representation
penalty, for which we perform efficient optimization using the according to a dictionary. One possible general form of such
alternating direction method of multipliers (ADMM). We then a regularization function is,
show the advantages of the proposed Bayesian nonparametric N
regularization framework on several CS-MRI problems in
X γ
hl (x) = kRi x − Dαi k22 + f (αi , D), (3)
Section IV. i=1
2
2) Draw a probability πk ∈ [0, 1] for each dk : K-SVD denoising (PSNR) BPFA denoising (PSNR)
σ
Match Mismatch Results Learned noise
πk ∼ Beta(cγ/K, c(1 − γ/K)), k = 1, . . . , K.
20/255 32.28 28.94 32.88 20.43/255
3) Draw precision values for noise and each weight 25/255 31.08 28.60 31.81 25.46/255
30/255 29.99 28.35 30.94 30.47/255
γε ∼ Gam(g0 , h0 ), γs ∼ Gam(e0 , f0 ).
th
4) For the i patch in x:
a) Draw the vector si ∼ N (0, γs−1 IK ).
b) Draw the binary vector zi with zik ∼ Bern(πk ).
c) Define αi = si ◦ zi by an element-wise product.
d) Sample noisy patch Ri x ∼ N (Dαi , γε−1 IP ).
5) Construct the image x as the average of all Ri x that
overlap on a given pixel.
learning proceeds by first learning the dictionary representa- and zeros elsewhere. Let Ψ = [ψ1T , . . . , ψN
T T
] be the resulting
tion of each patch, Ri x ≈ Dαi . The P denoised reconstruction 2N × N difference matrix for the entire image. The TV
of x using BPFA is then xBPFA = P1 i RiT Dαi . coefficients are β = Ψx ∈ C2N , and the isotropic TV penalty
P q 2
We show an example using 6 × 6 patches extracted from 2
P
is T V (x) = i kψ xk
i 2 = i |β| 2i−1 + |β|2i , where i
the noisy 512 × 512 image shown in Figure 1(a). In Figure ranges over the pixels in the MR image. For optimization we
1(b) we show the resulting denoised image. For this problem use the alternating direction method of multipliers (ADMM)
we truncated the dictionary size to K = 108 and set all other [31], [30]. ADMM works by performing dual ascent on the
model parameters to one. In Figures 1(c) and 1(d) we show augmented Lagrangian objective function introduced for the
some statistics from dictionary learning. For example, Figure total variation coefficients. For completeness, we give a brief
1(c) shows the values of πk sorted, where we see that fewer review of ADMM in the appendix.
than 100 elements are used by the data, many of which are very
sparsely used. Figure 1(d) shows the empirical distribution of
A. Algorithm
the number of elements used per patch. We see the ability of
the model to adapt the sparsity to the complexity of the patch. We present an algorithm for finding a local optimal solution
In Table I we show PSNR results for three noise variance to the non-convex objective function given in (5). We can write
levels. For K-SVD, we consider the case when the error this objective as
parameter matches the ground truth, and when it mismatches L(x, ϕ) = λg i kψi xk2 + i γ2ε kRi x − Dαi k22
P P
it by a magnitude of five. As expected, when K-SVD does not
+ i f (ϕi ) + λ2 kFu x − yk22 .
P
have an appropriate parameter setting the performance suffers. (6)
BPFA on the other hand adaptively infers this value, which
We seek to minimize this function with respect to x and the
helps improve the denoising.
dictionary learning variables ϕi = {D, si , zi , γε , γs , π}.
Our first step is to put the objective into a more suitable
III. CS-MRI WITH BPFA AND TV PENALTY
form. We begin by defining the TV coefficients for the ith
We next present our approach for reconstructing single- pixel as β i := [β2i−1 β2i ]T = ψi x. We introduce the vector of
channel MR images from highly undersampled k-space data. Lagrange multipliers ηi , and then split β i from ψi x by relaxing
In reference to the discussion in Section II, we consider a the equality via an augmented Lagrangian. This results in the
sparsity constraint of the form objective function
λ N
arg min λg hg (x) + hl (x) + kFu x − yk22 , (5) X ρ
x,ϕ 2 L(x, β, η, ϕ) = λg kβ i k2 + ηiT (ψi x − β i ) + kψi x − β i k22
i=1
2
N
X γε
hg (x) := T V (x), hl (x) := kRi x − Dαi k22 + f (ϕi ). N
X γε
i=1
2 + kRi x − Dαi k22 + f (ϕi )
i=1
2
For the local regularization function hl (x) we use BPFA λ
as given in Algorithm 1 in Section II-B. The parameters + kFu x − yk22 . (7)
2
to be optimized for this penalty are contained in the set
ϕi = {D, si , zi , γε , γs , π}, and are defined in Algorithm 1. From the ADMM theory [32], this objective will have (local)
We note that only si and zi vary in i, while the rest are optimal values β ∗i and x∗ with β ∗i = ψi x∗ , and so the equality
shared by all patches. The regularization term γε is a model constraints will be satisfied [31].3 Optimizing this function can
variable that corresponds to an inverse variance parameter be split into three separate sub-problems: one for TV, one for
of the multivariate Gaussian likelihood. This likelihood is BPFA and one for updating the reconstruction x. Following the
equivalently viewed as the squared error penalty term in hl (x) discussion of ADMM in the appendix, we define ui = (1/ρ)ηi
in (5). This term acts as the sparse basis for the image and and complete the square in the first line of (7). We then cycle
also aids in producing a denoised reconstruction, as discussed through the following three sub-problems,
in Sections II-B, III-B and IV-B. For the global regularization (P 1) β 0i = arg minβ λg kβk2 + ρ2 kψi x − β + ui k22 ,
function hg (x) we use the total variation of the image. This
term encourages homogeneity within contiguous regions of u0i = ui + ψi x − β 0i , i = 1, . . . , N,
the image, while still allowing for sharp jumps in pixel value ϕ0 = arg minϕ
P γε 2
(P 2) i 2 kRi x − Dαi k2 + f (ϕi ),
at edges due to the underlying `1 penalty. The regularization
P ρ 0
parameters λg , γε and λ control the trade-off between the (P 3) x0 = arg minx 0 2
i 2 kψi x − β i + ui k2
terms in this optimization. Since we sample a new value of P γε0
γε with each iteration of the algorithm discussed shortly, this + i 2 kRi x − D0 αi0 k22 + λ2 kFu x − yk22 .
trade-off is adaptively changing. Solutions for sub-problems P 1 and P 3 are globally
For the total variation penalty T V (x) we use the isotropic optimal (conditioned on the most recent values of all other
TV model. Let ψi be the 2 × N difference operator for pixel i. parameters). We cannot solve P 2 analytically since the
Each row of ψi contains a 1 centered on pixel i, and the first optimal values for the set of all BPFA variables do not
row also has a −1 on the pixel directly above pixel i, while
the second has a −1 corresponding to the pixel to the right, 3 For a fixed D, α1:N and x the solution is also globally optimal.
5
Algorithm 2 Outline of algorithm D. The addition of correlated Gaussian noise in the complex
Input: y – Undersampled k-space data plane generates the sample from the conditional posterior of
Output: x – Reconstructed MR image D. Since both the number of pixels and γε will tend to be
Initialize: x = FuH y and each ui = 0. Sample D from prior. very large, the variance of the noise is small and the mean
term dominates the update for D.
Step 1. P 1: Optimize each β i via shrinkage.
Step 2. Update Lagrange multiplier vectors ui . b) Sample sparse coding αi : Sampling αi entails sam-
Step 3. P 2: Gibbs sample BPFA variables once. pling sik and zik for each k. We sample these values us-
Step 4. P 3: Solve for x using Fourier domain. ing block sampling. We recall that to block sample two
if not converged then return to Step 1. variables from their joint conditional posterior distribution,
(s, z) ∼ p(s, z|−), one can first sample z from the marginal
distribution, z ∼ p(z|−), and then sample s|z ∼ p(s|z, −)
from the conditional distribution. The other sampling direction
have a closed form solution. Our approach for P 2 is to use
is possible as well, but for our problem sampling z → s|z is
stochastic optimization by Gibbs sampling each variable of
more efficient for finding a mode of the objective function.
BPFA conditioned on current values of all other variables.
We next present the updates for each sub-problem. We give We define ri,−k to be the residual error in approximating
an outline in Algorithm 2. the ith patch with the current values P
from BPFA minus the k th
dictionary element, ri,−k = Ri x − j6=k (sij zij )dj . We then
1) Algorithm for P1 (total variation): We can solve for β i sample zik from its conditional posterior Bernoulli distribution
exactly for each pixel i = 1, . . . , N by using a generalized zik ∼ pik δ1 + qik δ0 , where following a simplification,
shrinkage operation [31], − 21
pik ∝ πk 1 + (γε /γs )dHk dk × (10)
0 λg ψi x + u i nγ o
β i = max kψi x + ui k2 − ,0 · . (8) exp
ε
Re(dH 2 H
ρ kψi x + ui k2 k r i,−k ) /(γs /γε + d k d k ) ,
2
We recall that β i corresponds to the 2-dimensional TV qik ∝ 1 − πk . (11)
coefficients for pixel i, with differences in one direction
The symbol H denotes the conjugate transpose. The proba-
vertically and horizontally. We then update the corresponding
bilities can be obtained by dividing both of these terms by
Lagrange multiplier, u0i = ui + ψi x − β 0i .
their sum. We observe that the probability that zik = 1 takes
into account how well dictionary element dk correlates with
2) Algorithm for P2 (BPFA): We update the parameters of
the residual ri,−k . After sampling zik we sample the corre-
BPFA using Gibbs sampling. We are therefore stochastically
sponding weight sik from its conditional posterior Gaussian
optimizing (7), but only for this sub-problem. With reference
distribution,
to Algorithm 1, the P2 sub-problem entails sampling new
values for the complex dictionary D, the binary vectors zi and
!
Re(dHk r i,−k ) 1
real-valued weights si (with which we construct αi = si ◦ zi sik |zik ∼ N zik , . (12)
through the element-wise product), the precisions γε and γs , γs /γε + dH H
k d k γs + γε zik d k d k
and the probabilities π1:K , with πk giving the probability that
When zik = 1, the mean of sik is the regularized least squares
zik = 1. In principle, there is no limit to the number of samples
solution and the variance will be small if γε is large. When
that can be made, with the final sample giving the updates
zik = 0, sik can is sampled from the prior, but does not factor
used in the other sub-problems. We found that a single sample
in the model in this case.
is sufficient in practice and leads to a faster algorithm. We
c) Sample γε and γs : We next sample from the condi-
describe the sampling procedure below.
tional gamma posterior distribution of the noise precision and
a) Sample dictionary D: We define the P × N matrix
weight precision,
X = [R1 x, . . . , RN x], which is a complex matrix of all
vectorized patches extracted from the image x. We also define γε ∼ Gam g0 + 12 P N, h0 + 12 i kRi x − Dαi k22 , (13)
P
the K ×N matrix α = [α1 , . . . , αN ] containing the dictionary
γs ∼ Gam(e0 + 21 i,k zik , f0 + 21 i,k zik s2ik ).
P P
weight coefficients for the corresponding columns in X such (14)
that Dα is an approximation of X to which we add noise
The expected value of each variable is the first term of the
from a circularly-symmetric complex normal distribution. The
distribution divided by the second, which is close to the inverse
update for the dictionary D is
of the average empirical error for γε .
D = XαT (ααT + (P/γε )IP )−1 + E, (9) d) Sample πk : Sample each πk from its conditional beta
Ep,:
ind
∼ T
CN (0, (γε αα + P IP ) −1
), p = 1, . . . , P, posterior distribution,
PN PN
where Ep,: is the pth row of E. To sample this, we can first πk ∼ Beta a0 + i=1 zik , b0 + i=1 (1 − zik ) . (15)
draw Ep,: from a multivariate Gaussian distribution with this
covariance structure, followed by an i.i.d. uniform rotation of The parameters to the beta distribution include counts of how
each value in the complex plane. We note that the first term many times dictionary element dk was used by a patch.
in Equation (9) is the `2 -regularized least squares solution for
6
B. Experiments on a GE phantom
We consider a noisy synthetic example to highlight the
advantage of dictionary learning for CS-MRI. In Figure 3
we show results on a 256 × 256 GE phantom with additive
noise having standard deviation σ = 0.1. In this experiment
we use BPFA without TV to reconstruct the original image
(c) BPFA denoising (xBPFA ) (d) Total variation reconstruction
using 30% Cartesian sampling. We show the reconstruction
using zero-filling in Figure 3(a). Since λ = 10100 , we see
in Figure 3(b) that BPFA essentially helps reconstruct the
underlying noisy image for x. However, using the denoising
property of the BPFA model shown in Figure 1, we obtain the
denoised reconstruction of Figure 3(c) by focusing on xBPFA
from Equation (16). This is in contrast with the best result
(e) Dictionary (magnitude) we could obtain with TV in Figure 3(d), which places the TV
penalty on the reconstructed image. As discussed, for TV the
0.6
0.22
setting of λ relative to λg is important. We set λ = 1 and
0.2
0.5
0.18 swept through λg ∈ (0, 0.15), showing the result with highest
0.16 0.4 PSNR in Figure 3(d). Similar to Figure 1 we show statistics
0.14
0.12 0.3
from the BPFA model in Figures 3(e)-(g). We see that roughly
0.1 80 dictionary elements were used (the unused noisy elements
0.08 0.2
0.06 in Figure 3(e) are draws from the prior). We note that 2.28
0.1
0.04 elements were used on average by a patch given that at least
0.02
0 one was used, which discounts the black regions.
0 20 40 60 80 0 1 2 3 4 5 6 7 8 9
Fig. 3. GE data with noise (σ = 0.1) and 30% Cartesian sampling. BPFA
C. Experiments on real-valued (synthetic) MRI
(b) reconstructs the original noisy image, and (c) denoises the reconstruction
simultaneously. (d) TV denoises as part of the reconstruction. Also shown
For our synthetic MRI experiments, we consider two pub-
are the dictionary learning variables sorted by πk . (e) the dictionary, (f) the licly available real-valued 512 × 512 MRI9 of a shoulder and
distribution on the dictionary, πk . (g) The normalized histogram of number lumbar. We construct these problems by applying the relevant
of the dictionary elements used per patch.
sampling mask to the projection of real-valued MRI into k-
space. Though using such real-valued MRI data may not reflect
clinical reality, we include this idealized setting to provide a
A. Set-up complete set of experiments similar to other papers [3], [14],
[15]. We evaluate the performance of our algorithm using
For all images, we extract 6 × 6 patches where each pixel PSNR and compare with Sparse MRI [3], DLMRI [14] and
defines the upper left corner of a patch and wrap around the PBDW [15]. Although the original data is real-valued, we learn
image at the boundaries; we investigate different patch sizes complex dictionaries since the reconstructions are complex.
later to show that this is a reasonable size. We initialize x by We consider our algorithm with and without the total variation
zero-filling in k-space. We use a dictionary with K = 108 penalty, denoted BPFA+TV and BPFA, respectively.
initial dictionary elements, recalling that the final number of
dictionary elements will be smaller due to the sparse BPFA 8 As discussed in Section II-B, in theory K can be infinitely large.
prior. If 108 is found to be too small, K can be increased with 9 www3.americanradiology.com/pls/web1/wwimggal.vmg/wwimggal.vmg
8
TABLE II
PSNR RESULTS FOR REAL - VALUED L UMBAR MRI AS FUNCTION OF
SAMPLING PERCENTAGE AND MASK (C ARTESIAN WITH RANDOM PHASE
ENCODES , 2D RANDOM AND PSEUDO RADIAL ).
TABLE III
PSNR RESULTS FOR REAL - VALUED S HOULDER MRI AS FUNCTION OF (c) DLMRI (d) Sparse MRI
SAMPLING PERCENTAGE AND MASK (C ARTESIAN WITH RANDOM
PHASE ENCODES , 2D RANDOM AND PSEUDO RADIAL ). Fig. 4. Absolute errors for 30% Cartesian sampling of synthetic lumbar MRI.
0.1 0.1
TABLE IV
PSNR/SSIM RESULTS FOR COMPLEX - VALUED B RAIN MRI AS A FUNCTION OF SAMPLING PERCENTAGE . S AMPLING MASKS INCLUDE
C ARTESIAN SAMPLING WITH RANDOM PHASE ENCODES , 2D RANDOM SAMPLING AND PSEUDO RADIAL SAMPLING .
(a) Original (b) BPFA+TV (c) BPFA (d) PBDW (e) DLMRI
40
38
36
BPFA+TV
34 BPFA
32
30
28
0 200 400 600 800 1000
(f) PSNR vs iteration (g) BPFA+TV error (h) BPFA error (i) PBDW error (j) DLMRI error
Fig. 6. Reconstruction results for 25% pseudo radial sampling of a complex-valued MRI of the brain.
34
SSIM as performance measures. We show these values in
33
Table IV for each algorithm, sampling mask and sampling
rate. As with the synthetic MRI, we see that our algorithm 32
(a) Original (b) BPFA+TV: PSNR = 39.64 (c) BPFA: PSNR = 38.21 (d) PBDW: PSNR = 37.89 (e) DLMRI: PSNR = 35.05
40
39
38
37
36
35
BPFA+TV
34
BPFA
33
32
31
30
0 200 400 600 800 1000
(f) PSNR vs iteration (g) BPFA+TV error (h) BPFA error (i) PBDW error (j) DLMRI error
Fig. 7. Reconstruction results for 35% 2D random sampling of a complex-valued MRI of a lemon.
TABLE V
PSNR FOR 35% C ARTESIAN SAMPLING OF COMPLEX - VALUED B RAIN
MRI FOR VARIOUS NOISE STANDARD DEVIATIONS . (λ = 10100 )
TABLE VI
PSNR AS A FUNCTION OF PATCH SIZE FOR A REAL - VALUED AND
COMPLEX - VALUED B RAIN MRI WITH C ARTESIAN SAMPLING .
TABLE VII
T OTAL RUNTIME IN MINUTES ( SECONDS / ITERATION ). W E RAN 1000
ITERATIONS OF BPFA, 100 OF DLMRI AND 10 OF S PARSE MRI.
empirical probability
Sum of probabilities 4 30% sampling
0.1
30% sampling. We see that they are similar in their shape, but 3.5
0.08
the number of elements increases as the sampling percentage 3
0.06
increases since more complex information about the image is 2.5
10% sampling
2 0.04
contained in the k-space measurements. We again note that 20% sampling
1.5 30% sampling 0.02
from BPFA. We can read off the average number of elements (d) BPFA weights (cumulative) (e) Dictionary elements per patch
used per patch by looking at the right-most value. We see that
more elements are used per patch as the fraction of observed Fig. 10. Radial sampling for the Brain MRI. (a)-(c) The learned dictionary
for various sampling rates. The noisy elements towards the end of each were
k-space increases. We also see that for 10%, 20% and 30% unused and are samples from the prior. (d) The cumulative function of the
sampling, roughly 70, 80 and 95, respectively, of the 108 total sorted πk from BPFA for each sampling rate. This gives information on
dictionary elements were significantly used, as indicated by the sparsity and average usage of the dictionary. (e) The distribution on the number
of elements used per patch for each sampling rate.
leveling off of these functions. This highlights the adaptive
property of the nonparametric beta process prior. In Figure
10(e) we show the empirical distribution on the number of coding update.11 We note that inference for the BPFA model
dictionary elements used per patch for each sampling rate. We is easily parallelizable—as are the other dictionary learning
see that there are two modes, one for the empty background algorithms—which can speed up processing time.
and one for the foreground, and the second mode tends to The proposed method has several advantages, which we
increase as the sampling rate increases. The adaptability of believe leads to the improvement in performance. A significant
this value to each patch is another characteristic of the beta advantage is the adaptive learning of the dictionary size
process model. and per-patch sparsity level using a nonparametric stochastic
We also performed an experiment with varying patch sizes process that is naturally suited for this problem. Several other
and show our results in Table VI. We see that the results dictionary learning parameters such as the noise variance and
are not very sensitive to this setting and that comparisons the variances of the score weights are adjusted as well through
using 6 × 6 patches are meaningful. We also compare the a natural MCMC sampling approach. These benefits have
runtime for different algorithms in Table VII, showing both been investigated in other applications of this model [25], and
the total runtime of each algorithm and the per-iteration naturally translate here since CS-MRI with BPFA is closely
times using an Intel Xeon CPU E5-1620 at 3.60GHz, 16.0G related to image denoising as we have shown.
ram. However, we note that we arguably ran more iterations Another advantage of our model is the Markov Chain
than necessary for these algorithms; the BPFA algorithms Monte Carlo inference algorithm itself. In highly non-convex
generally produced high quality results in half the number of Bayesian models (or similar models with a Bayesian interpre-
iterations, as did DLMRI (the authors of [14] recommend 20 tation), it is often observed by the statistics community that
iterations), while Sparse MRI uses 5 iterations as default and MCMC sampling can outperform deterministic methods, and
the performance didn’t improve beyond 10 iterations. We note
that the speed-up over DLMRI arises from the lack of the OMP 11 BPFA is significantly faster than K-SVD in Matlab because it requires
algorithm, which in Matlab is much slower than our sparse fewer loops. This difference may not be as large with other coding languages.
12
[25] M. Zhou, H. Chen, J. Paisley, L. Ren, L. Li, Z. Xing, D. Dunson, G. Qin Lin is currently a graduate student in the De-
Sapiro and L. Carin, “Nonparametric Bayesian dictionary learning for partment of Communication Engineering at Xiamen
analysis of noisy and incomplete images,” IEEE Trans. Image Process., University. His research interests includes computer
vol. 21, no. 1, pp. 130-144, Jun. 2012. vision, machine learning and data mining.
[26] X. Ding, L. He and L. Carin, “Bayesian robust principal component
analysis,” IEEE Transactions on Image Processing, vol. 20, no. 12, pp.
3419-3430.
[27] N. Hjort, “Nonparametric Bayes estimators based on beta processes in
models for life history data,” Annals of Statistics, vol. 18, pp. 1259-1294,
1990.
[28] R. Thibaux and M. Jordan, “Hierarchical beta processes and the Indian
buffet process,” in International Conference on Artificial Intelligence
and Statistics, San Juan, Puerto Rico, 2007.
[29] D. Gabay and B. Mercier, “A dual algorithm for the solution of nonlinear Xinghao Ding was born in Hefei, China in 1977.
variational problems via finite-element approximations,” Computers and He received the B.S. and Ph.D degrees from the
Mathematics with Applications, vol. 2, pp. 17-40, 1976. Department of Precision Instruments at Hefei Uni-
[30] W. Yin, S. Osher, D. Goldfarb and J. Darbon, “Bregman iterative algo- versity of Technology in Hefei, China in 1998 and
rithms for L1 minimization with applications to compressed sensing,” 2003.
SIAM Journal on Imaging Sciences, vol. 1, no. 1, pp. 143-168, 2008. From September 2009 to March 2011, he was a
[31] T. Goldstein and S. Osher, “The split Bregman method for L1 regularized postdoctoral researcher in the Department of Electri-
problems,” SIAM Journal on Imaging Sciences, vol. 2, no. 2, pp. 323- cal and Computer Engineering at Duke University in
343, 2009. Durham, NC. Since 2011 he has been a Professor in
[32] S. Boyd, N. Parikh, E. Chu, B. Peleato and J. Eckstein, “Distributed the School of Information Science and Engineering
optimization and statistical learning via the alternating direction method at Xiamen University. His main research interests
of multipliers,” Foundations and Trends in Machine Learning, vol. 3, include image processing, sparse signal representation, and machine learning.
no. 1, pp. 1-122, 2010.
[33] J. Paisley, L. Carin and D. Blei, “Variational inference for stick-breaking
beta process priors,” in International Conference on Machine Learning,
Bellevue, WA, 2011. Xueyang Fu is currently a graduate student in
[34] D. Knowles and Z. Ghahramani, “Infinite sparse factor analysis and the Department of Communication Engineering at
infinite independent components analysis,” in Independent Component Xiamen University. His research interests include
Analysis and Signal Separation, Springer, pp 381-388, 2007. image processing, sparse representation and machine
[35] E. Fox, E. Sudderth, M.I. Jordan and A.S. Willsky, “Sharing features learning.
among dynamical systems with beta processes,” in Advances in Neural
Information Processing, Vancouver, B.C., 2011.
[36] T. Griffiths and Z. Ghahramani, “Infinite latent feature models and the
Indian buffet process,” in Advances in Neural Information Processing
Systems, MIT Press, pp 475-482, 2006.
[37] Z. Wang, A. C. Bovik, H. R. Sheikh and E. P. Simoncelli, “Image
quality assessment: From error visibility to structural similarity,” IEEE
Transactions on Image Processing, vol. 13, no. 4, pp. 600-612, 2004.
[38] C. Faes, J.T. Ormerod and M.P. Wand, “Variational Bayesian inference
Xiao-Ping Zhang (M’97, SM’02) received B.S. and
for parametric and nonparametric regression with missing data,” Journal
Ph.D. degrees from Tsinghua University, in 1992 and
of the American Statistical Association, vol. 106, pp 959-971, 2011.
1996, respectively, both in Electronic Engineering.
He holds an MBA in Finance, Economics and En-
trepreneurship with Honors from the University of
Chicago Booth School of Business, Chicago, IL.
Since Fall 2000, he has been with the Department
of Electrical and Computer Engineering, Ryerson
University, where he is now Professor, Director of
Yue Huang received the B.S. degree in Electrical Communication and Signal Processing Applications
Engineering from Xiamen University in 2005, and Laboratory (CASPAL). He has served as Program
the Ph.D. degree in Biomedical Engineering from Director of Graduate Studies. He is cross appointed to the Finance Department
Tsinghua University in 2010. Since 2010 she is an at the Ted Rogers School of Management at Ryerson University. Prior to
Assistant Professor of the School of Information joining Ryerson, he was a Senior DSP Engineer at SAM Technology, Inc., San
Science and Engineering at Xiamen University. Her Francisco, and a consultant at San Francisco Brain Research Institute. He held
main research interests include image processing, research and teaching positions at the Communication Research Laboratory,
machine learning, and biomedical engineering. McMaster University, and worked as a postdoctoral fellow at the Beckman
Institute, the University of Illinois at Urbana-Champaign, and the University
of Texas, San Antonio. His research interests include statistical signal pro-
cessing, multimedia content analysis, sensor networks and electronic systems,
computational intelligence, and applications in bioinformatics, finance, and
marketing. He is a frequent consultant for biotech companies and investment
firms. He is cofounder and CEO for EidoSearch, an Ontario based company
offering a content-based search and analysis engine for financial data.
Dr. Zhang is a registered Professional Engineer in Ontario, Canada, a
Senior Member of IEEE and a member of Beta Gamma Sigma Honor
John Paisley is an assistant professor in the Depart-
Society. He is the general chair for MMSP’15, publicity chair for ICME’06
ment of Electrical Engineering at Columbia Univer-
and program chair for ICIC’05 and ICIC’10. He served as guest editor
sity. Prior to that he was a postdoctoral researcher in
for Multimedia Tools and Applications, and the International Journal of
the Computer Science departments at UC Berkeley
Semantic Computing. He is a tutorial speaker in ACMMM2011, ISCAS2013,
and Princeton University. He received the B.S., M.S.
ICIP2013 and ICASSP2014. He is currently an Associate Editor for IEEE
and Ph.D. degrees in Electrical and Computer En-
Transactions on Signal Processing, IEEE Transactions on Multimedia, IEEE
gineering from Duke University in 2004, 2007 and
Signal Processing letters and for Journal of Multimedia.
2010. His research is in the area of statistical ma-
chine learning and focuses on probabilistic modeling
and inference techniques, Bayesian nonparametric
methods, and text and image processing.