Skip to main content

Local sparsity enhanced compressed sensing magnetic resonance imaging in uniform discrete curvelet domain



Compressed sensing(CS) has been well applied to speed up imaging by exploring image sparsity over predefined basis functions or learnt dictionary. Firstly, the sparse representation is generally obtained in a single transform domain by using wavelet-like methods, which cannot produce optimal sparsity considering sparsity, data adaptivity and computational complexity. Secondly, most state-of-the-art reconstruction models seldom consider composite regularization upon the various structural features of images and transform coefficients sub-bands. Therefore, these two points lead to high sampling rates for reconstructing high-quality images.


In this paper, an efficient composite sparsity structure is proposed. It learns adaptive dictionary from lowpass uniform discrete curvelet transform sub-band coefficients patches. Consistent with the sparsity structure, a novel composite regularization reconstruction model is developed to improve reconstruction results from highly undersampled k-space data. It is established via minimizing spatial image and lowpass sub-band coefficients total variation regularization, transform sub-bands coefficients l 1 sparse regularization and constraining k-space measurements fidelity. A new augmented Lagrangian method is then introduced to optimize the reconstruction model. It updates representation coefficients of lowpass sub-band coefficients over dictionary, transform sub-bands coefficients and k-space measurements upon the ideas of constrained split augmented Lagrangian shrinkage algorithm.


Experimental results on in vivo data show that the proposed method obtains high-quality reconstructed images. The reconstructed images exhibit the least aliasing artifacts and reconstruction error among current CS MRI methods.


The proposed sparsity structure can fit and provide hierarchical sparsity for magnetic resonance images simultaneously, bridging the gap between predefined sparse representation methods and explicit dictionary. The new augmented Lagrangian method provides solutions fully complying to the composite regularization reconstruction model with fast convergence speed.

Peer Review reports


Compressed sensing(CS) was first presented in the literature of Information Theory as an abstract mathematical idea [1, 2]. The fundamental idea behind CS is: rather than first sampling at a high rate and then compressing the sampled data, directly sensing the data in a compressed form (at a lower sampling rate) is preferred. CS points out that a signal can be recovered exactly from a small set of random, linear and nonadaptive measurements if it has a sparse representation. Suppose \(\textbf {x}\in \mathbb {C}^{p}\) denotes the unknown compressible (N-sparse) signal to be reconstructed. Ψ denotes a tight frame sparse transform matrix. Then x can be sparsely represented as α=Ψ x, where α0=N(Np). It is possible to measure a relatively small number of “random" linear combinations of signal (much smaller than the number of signal samples nominally defining it) allowing accurate reconstruction, which is comparable to that attainable with direct knowledge of the N most important coefficients. The measurement process is denoted as y=Φ x, where \(\boldsymbol {\Phi }\in \mathbb {C}^{m\times p} \left (m\ll p\right)\) denotes measurement matrix irrelevant to the sparse transform basis. Thus

$$ \textbf{y}=\boldsymbol{\Phi}\textbf{x}=\boldsymbol{\Phi}\boldsymbol{\Psi}^{-1}\boldsymbol{\alpha} $$

in which Φ Ψ −1 is termed as the sensing matrix. The sensing matrix should satisfy three properties including the null space property, restricted isometry property and bounded coherence [3]. Given measurements y and the sensing matrix, the reconstruction problem turns out to be an optimization problem of the form

$$ \arg\min_{\textbf{x}, \hat{\boldsymbol{\alpha}}}\left\|\hat{\boldsymbol{\alpha}}\right\|_{0} \ \ \text{s.t.} \ \boldsymbol{\Phi}\boldsymbol{\Psi}^{-1}\hat{\boldsymbol{\alpha}}=\textbf{y} $$

(2) can be solved by various nonlinear reconstruction approaches.

In magnetic resonance imaging(MRI), the sampled combinations are simply individual Fourier coefficients (k-space samples). MRI is a relatively slow imaging modality at a limited data acquisition speed. Undersampling k-space allows speeding up imaging but introduces aliasing in the reconstructed magnetic resonance(MR) images simultaneously, because it violates the Nyquist sampling theorem. Compared with that by the sinc function interpolation using sampled data restricted by Nyquist sampling theorem, CS enables MR image reconstruction with little or no visual information loss from randomly undersampled k-space measurements. Hence, it is natural to introduce CS into undersampled MRI. The emerging method to reduce MRI scanning time via CS is termed CS MRI [4, 5]. Three requirements for successful CS MRI are: the MR image can be sparsely represented (compressible); the aliasing artifacts brought by k-space undersampling are incoherent (noise like) in the transform domain; then CS solves the general reconstruction formulation using nonlinear method by constraining both sparsity and k-space measurements consistency. In CS MRI, incoherent random, radial and spiral trajectories [4, 6, 7], etc, are used to acquire measurements from k-space.

Sparsity is of vital importance for reducing artifacts in CS MRI reconstruction. The generally used sparse representation methods are spatial finite difference [4, 8, 9], discrete wavelet transform(DWT) [4, 8, 9], sharp frequency localization contourlet(SFLCT) [10, 11], discrete curvelet transform using fast algorithm(FDCT) [1214], discrete shearlet transform(DST) [15, 16], sparsity along temporal axis for dynamic cardiac imaging [1719] and the combination of some of these transforms [4, 20, 21]. Dictionary has also been introduced for sparse representation and adaptive data fitting [22, 23] and it is learnt from intermediate reconstructed or fully sampled images. Furthermore, double sparsity model has been proposed likewise. It combines adaptive dictionary learning(DL) with predefined sparse priors for signal flexible representations, stability under noise and reducing overfitting [24, 25]. Besides, nonlocal processing [26] methods have been explored as well based on the similarity of image patches [2730] and sparsity originated from this similarity [31, 32]. Established on the above sparse representation approaches, various CS MRI methods have been presented for handling the ill-posed linear inverse problem, including convex, nonsmooth sparse regularization (l 1 and total variation) based LDP [4], TVCMRI [33], iterative thresholding CS MRI based on SFLCT [11], FCSA [21], NLTV-MRI upon nonlocal total variation(TV) [9], reconstruction under wavelet structured sparsity(WaTMRI) studied in [34], reconstruction via using DL [3537] and PANO [32] by using variable splitting(VS) and quadratic penalty reconstruction technique [38] incorporated with patch-based nonlocal operator, etc. Besides, a novel individual MRI reconstruction framework of low-rank modeling for local k-space neighborhoods(LORAKS) [39] was also proposed. LORAKS generated support and phase constraints in a fundamentally different way from more direct regularized methods [4, 40]. Among these reconstruction methods, DWT based MRI reconstruction gave rise to feature loss and edge blur with numerous aliasing in reconstructed images. WaTMRI provided new thought for CS MRI by making full use of the coefficients structural relationship. PANO was recently proposed to sparsify MR images by using the similarity of image patches, achieved considerably lower reconstruction error and allowed us to establish a general formulation to constrain the sparsity of similar patches and data consistency. The availability of guide image and how the gridding process affect PANO imaging with nonCartesian sampling remain to be carefully analyzed. Besides, LORAKS provided very flexible implementation and was easily incorporated with other constraints. Furthermore, 3D dynamic parallel imaging has also been presented and was of great significance for practical MRI applications. 3D dynamic parallel imaging was generally established on TV and sparsity along temporal axis [1719] and structured low-rank matrix completion [4145].

In this paper, a novel composite sparsity structure is developed, which is inspired by double sparsity model. In this composite sparsity structure, uniform discrete curvelet transform(UDCT) [46] decomposes MR images hierarchically into one lowpass sub-band and several other highpass sub-bands. Then an adaptive dictionary is learnt from the hardly sparse lowpass sub-band coefficients patches. This adaptive DL allows a smaller amount of calculation with little (or no) decrease of efficiency compared with the double sparsity model. UDCT has quite similar properties to those of wrapping-based FDCT, such as C(logN)3 N −2 mean square error(MSE) decay rate for C 2-singularities signal with N most important transform coefficients in the curvelet expansion, tight frame property, highly directional sensitivity and anisotropy in the sense that they both employ alias free subsampling in frequency domain. Additionally, UDCT possesses some specialities making it superior over FDCT in CS MRI applications, such as a smaller redundancy of 4 and clear coefficients parent-children relationship. The goal of the proposed composite sparsity structure is to capture various directional features of images hierarchically, provide more flexible and sparse representations for lowpass sub-band adaptively, and reduce overfitting and computational complexity simultaneously. Consistent with this composite sparsity structure, one reconstruction model is provided. It involves minimizing UDCT sub-bands coefficients l 1 regularization, image and lowpass sub-band coefficients TV penalty and constraining k-space measurements fidelity. Then a new fast convergent augmented Lagrangian(AL) reconstruction method is presented to solve the reconstruction model. It is established on the constrained split augmented Lagrangian shrinkage algorithm(C-SALSA) [47], translates the original formulation into another constrained one via VS and then solves the constrained one by using the variant of ADMM [48, 49](ADMM-2 [47]). The ADMM-2 resulting from our reconstruction problem involves quadratic problem (which can be solved exactly in closed form), l 1 regularization, a shrinkage operation and an orthogonal projection on a ball.

The remainder of this paper is organized as follows. Section “Methods” depicts the prior work, the proposed CS MRI method including the composite sparsity structure and corresponding reconstruction model, and its validity in detail. In section “Results and discussions” some reconstruction results are exhibited for the proposed method and current CS MRI techniques. The ability in handling noise, convergence speed and influences of the proposed reconstruction model parameters fluctuation are analyzed, etc. Finally, conclusions and future work are explicit in section Conclusions and future work.


Compressed sensing MRI prior art

In CS MRI, \(\textbf {x}\in \mathbb {C}^{p}\) denotes the vector form of the 2D MR image to be reconstructed and y=F u x denotes the sensing procedure in k-space, where \(\textbf {F}_{u}\in \mathbb {C}^{m\times p}\left (m \ll p\right)\) means undersampled Fourier Encoding matrix and \(\textbf {y}\in \mathbb {C}^{m}\) represents k-space measurements. Assume Ψ represents an analytical sparse transform operator or the inverse of a set of signals learnt from image patches, the sparse representation is defined as α=Ψ x. Reconstructing the unknown MR image x from measurements y by using CS is to solve the constrained CS MRI optimization problem (3)

$$ \min_{\textbf{x}} \left\|\boldsymbol{\Psi} \textbf{x}\right\|_{1} \ \ \text{s.t.} \ \left\|\textbf{F}_{u} \textbf{x}-\textbf{y}\right\|^{2}_{2} \leq \boldsymbol{\varepsilon} $$

where \(\boldsymbol {\varepsilon }\in \mathbb {C}^{m}\) represents the allowed noise vector in reconstructed image, l 1 regularization enforces sparsity of Ψ x and error constraint fits the reconstruction to the sampled k-space measurements. Finite difference referred to as TV is typically added into (3) for noise reduction and spatial homogeneity enhancement. Then the formulation is

$$ \min_{\textbf{x}} \left\|\boldsymbol{\Psi} \textbf{x}\right\|_{1}+ \lambda TV\left(\textbf{x}\right) \ \ \text{s.t.} \ \left\|\textbf{F}_{u} \textbf{x}-\textbf{y}\right\|^{2}_{2} \leq \boldsymbol{\varepsilon} $$

where λ>0 denotes the weight of TV regularization. This becomes problem formulation settled by classical LDP when Ψ represents DWT. Besides, most state-of-the-art AL based reconstruction methods consider one unconstrained problem rather than (4), such as TVCMRI, FCSA, iterative thresholding CS MRI based on SFLCT and SALSA [50], etc

$$ \min_{\textbf{x}} \lambda_{1}\left\|\boldsymbol{\Psi} \textbf{x}\right\|_{1}+ \lambda_{2} TV\left(\textbf{x}\right)+\frac{1}{2}\left\|\textbf{F}_{u} \textbf{x}-\textbf{y}\right\|^{2}_{2} $$

Additionally, reconstruction based on explicit DL from image patches has been explored

$$ \begin{aligned} &\min_{\textbf{x},\textbf{D},\boldsymbol{\Gamma}} \sum\limits_{ij}\left\|\textbf{R}_{ij}\textbf{x}-\textbf{D}\boldsymbol{\alpha}_{ij}\right\|^{2}_{2}+ \lambda \left\|\textbf{F}_{u} \textbf{x}-\textbf{y}\right\|^{2}_{2} \ \text{s.t.} \ \left\|\boldsymbol{\alpha}_{ij}\right\|_{0} \\ &\quad\leq T_{0}, \forall i,j \end{aligned} $$

(6) is the formulation settled by DLMRI integrating DL with image reconstruction simultaneously (refer to [35]), where matrix \(\textbf {R}_{\textit {ij}}\in \mathbb {C}^{n \times p}\) is an operator that extracts maximum overlapped patch x ij as a column vector from x, denoted as x ij =R ij x. \(\sum _{\textit {ij}}\textbf {R}_{\textit {ij}}^{T} \textbf {R}_{\textit {ij}}\in \mathbb {C}^{p \times p}\) represents a diagonal matrix, in which the diagonal elements correspond to pixel locations of x. Γ is sparse representation set {α ij } ij of all training patches of image. T 0 denotes sparse threshold and D the explicit dictionary. Here DL problem is depicted in detail as a foundation of the proposed sparsity structure. The DL problem is

$$ \min_{\textbf{D}, \boldsymbol{\Gamma}} \sum\limits_{ij}\left\|\textbf{R}_{ij}\textbf{x}-\textbf{D}\boldsymbol{\alpha}_{ij}\right\|^{2}_{2} \ \ \text{s.t.} \ \left\|\boldsymbol{\alpha}_{ij}\right\|_{0} \leq T_{0}, \forall i,j $$

Formulation (7) is nonconvex and NP-hard [51, 52] because it comes down to sparse encoding problem for fixed D and x. K-SVD [53, 54] is a simple but efficient approach to attack (7). It simultaneously implements dictionary update step where each atom of D renews sequentially and corresponding sparse encoding for image patches that currently use it. The dictionary atom update involves computing K singular value decompositions(SVDs), once for each atom.

Proposed method

In this section, a composite regularization CS MRI method established on a novel composite sparsity structure is presented. In the sparsity structure, UDCT decomposes spatial image into one lowpass sub-band and several highpass sub-bands. Patch-based dictionary is learnt from the lowpass sub-band coefficients patches via Sparse K-SVD [25]. Then a novel composite regularization reconstruction model is thereby established and solved via VS and ADMM-2. The reconstruction model involves spatial image and transform coefficients regularization and k-space data fitting. The framework in Fig. 1 shows clearly the implementation process of the proposed method, in which the unknown MR image x is initialized with a zero-filling reconstructed image via direct inverse Fourier transform to k-space measurements, denoted as \(\textbf {x}_{0}=\textbf {F}^{H}_{u}\textbf {y}\). UDCT decomposes both the real and imaginary parts of x 0 into S levels, each level possessing 2κ s directional sub-bands. The real and imaginary parts of complex-valued MR image are handled separately because UDCT can only perfectly deal with real-valued data. Take the real part of zero-filling reconstructed image for example, the lowpass UDCT sub-band is divided into maximum overlapped patches (for dividing method, refer to [35]) as training database for DL to enhance its sparsity. The obtained dictionary D r (r=1(0) denotes result over real (imaginary) part) is the result of Sparse K-SVD to the training database. The sparse encodings set are referred to as the double sparse coefficients for all training lowpass UDCT sub-band patches over learnt D r. For imaginary part, the same procedure is implemented. Let x 0 be the initial intermediate image and (D r) the pseudo-inverse of D r. The reconstruction step starts afterwards. All nonoverlapping vector form patches (n×1 sized) are arrayed to produce a matrix from lowpass UDCT sub-band of intermediate image. Results of (D r) multiplying with the above matrix are the representation coefficients \(\boldsymbol {\Gamma }^{r}_{R}\) (differing from double sparse coefficients) of lowpass UDCT sub-band coefficients over the dictionary. They are generally not sparse but easier to handle in our reconstruction approach. The composite regularization reconstruction formulation is solved by using VS and ADMM-2 based on C-SALSA thoughts in an iterative procedure (an updated intermediate image for once iteration), which involves modifying the representation coefficients, UDCT sub-bands coefficients and k-space measurements. The proposed method is named as local sparsity enhanced CS MRI(LSECSMRI). Formulations and implementations of the proposed sparsity structure and relevant reconstruction model are described in detail in the following content.

Fig. 1
figure 1

Framework of local sparsity enhanced CS MRI reconstruction

Composite sparsity structure

To the best of our knowledge, DWT is not applicable for 2D image sparse representation [55, 56] because of its very limited directions and incapacity to capture geometric regularity along singularities of surfaces. Multi-scale geometric analysis(MGA) methods like contourlet transform [56], nonsubsampled contourlet transform(NSCT) [57], SFLCT, FDCT and DST, etc, conquer the defects of DWT. Contourlet transform lacks shift-invariance though, which causes pseudo-Gibbs phenomena around singularities. The resulting contourlets cannot ensure good frequency localization and exhibit some fuzzy artifacts inevitably with a low redundancy of 4/3. NSCT is overcompletely designed with better frequency selectivity, regularity and fully shift-invariance. But it possesses very high redundancy and large time consumption. SFLCT is a semi-redundant contourlet transform providing flexible redundancy. It only increases redundancy in the lowpass filter l 0(ω) to reduce pseudo-Gibbs phenomena, which are mainly induced by down-sample in the lowpass filter. The redundancy can be 2.33 if we don’t down sample l 0(ω) and 1.33 if we down sample it by setting (d,d)=(2,2), where d is the down-sample parameter that determines the redundancy of contourlet. But SFLCT performs negatively in capturing clear directional features. Similarly, DST provides low redundancy less than or equal to 2. The needle-shaped elements of FDCT allow very high directional sensitivity and anisotropy and are thus very efficient in representing line-like edges. Nevertheless, FDCT possesses too high redundancy, which makes it sub-optimal in sparse representation, either.

UDCT [46] is recently proposed as an innovative implementation of discrete curvelet transform disposing real-valued signals. Utilizing the ideas of FFT-based discrete curvelet transform and filter-bank based contourlet transform, UDCT is designed as a perfect multi-resolution reconstruction filter bank(FB) and executed by FFT algorithm, possessing advantages of the both. The discrete curvelet functions in UDCT are defined by a parameterized family of smooth windowed functions that satisfy two conditions: 1) 2π periodic; 2) their squares form a partition of unity and the centers of the curvelet functions at each resolution are positioned on a uniform lattice. Moreover, the lattices of lower scales are subset of those at higher scales, guaranteeing clear parent-children relationship. UDCT can provide flexible instead of fixed number of clear directions at each scale to capture various directional geometrical structures of image accurately. Besides, the forward and inverse transform form a tight and self-dual frame with an acceptable redundancy of 4, allowing the input real-valued images to be perfectly reconstructed. In terms of the implementation of UDCT, once all the curvelet windows are computed, the actual forward and inverse UDCT computations are straightforward. UDCT has asymptotic approximation properties: for image x with C 2 (C is a constant) singularities, the best N-term approximation x N (N is the number of most important coefficients allowing reconstruction) in the curvelet expansion is given by [12, 55, 58]

$$ \left\|\textbf{x}-\textbf{x}_{N}\right\|^{2}_{2} \leq CN^{-2}\left(\text{logN}\right)^{3} \ \ N\rightarrow{} \infty $$

This property is known as the optimal sparsity. Therefore, UDCT is considered as the optimal predefined sparse method and is introduced into CS MRI in this paper.

UDCT belongs to predefined sparse priors, which implies that it lacks adaptivity. Explicit dictionary representations in spatial domain gain a higher degree of freedom in the training but sacrifice efficiency of the result. Besides, this kind of explicit dictionary cannot describe hierarchical structures. Incorporating the advantages of UDCT with DL and compensating for the defects of each other, an efficient composite sparsity structure is proposed for local sparsity enhancement. This structure learns adaptive dictionary over lowpass UDCT sub-band coefficients patches. For real (imaginary) part of image x, the DL problem is deduced as

$$ \min_{\textbf{D}^{r}, \boldsymbol{\Gamma}^{r}} \sum\limits_{ij}\left\|\textbf{R}_{ij}(\boldsymbol{\Psi} \textbf{x}^{r})_{l}-\textbf{D}^{r}\boldsymbol{\alpha}_{ij}\right\|^{2}_{2} \ \ \text{s.t.} \ \left\| \boldsymbol{\alpha}_{ij}\right\|_{0} \leq T_{0}, \forall i,j $$

In (9), Ψ represents UDCT operator. x r denotes real (imaginary) part if r=1(0) and the subscript l denotes lowpass sub-band. The resulting D r is thus dictionary for real (imaginary) part. Γ r is double sparse representation set of {α ij } ij . The proposed composite sparsity structure has some advantages. Compared with predefined sparse priors, it provides adaptivity and sparser representations for lowpass sub-band coefficients according to the different structural features of lowpass and highpass UDCT sub-bands coefficients. Compared with explicit dictionary, it allows hierarchical sparsity (depending mostly on UDCT) and reduces overfitting and instability in handling noise. Contrast with double sparsity model, it reduces the amount of calculations for not training dictionaries over highpass UDCT sub-bands, which are generally sparse enough. Therefore, the proposed sparse method can fine fit and hierarchically sparsify MR images with important characteristics preserved and avoid wasting time, making a big difference for improved MR image reconstruction performance. In Fig. 2, an example of dictionary trained by using Sparse K-SVD according to the proposed sparsity structure is exhibited. The used image is complex-valued T2-weighted brain image of slice 27 (MR T2wBrain_slice_27 of 256×256 sized). It is acquired from a healthy volunteer at a 3T Siemens Trio Tim MRI scanner, using the T2-weighted turbo spin echo sequence (T R/T E=6100/99 ms, 220×220 mm field of view, 3 mm slice thickness) [37]. The training maximum overlapped patches are 8×8 sized and obtained from lowpass UDCT sub-band coefficients after S=1 level’s UDCT decomposition for real part of MR T2wBrain_slice_27. One can see that the resulting dictionary is a highly structured matrix, implying several favorable properties.

Fig. 2
figure 2

Dictionary training. (a) MR T2wBrain_slice_27, (b) real part lowpass UDCT sub-band coefficients for DL, (c) initial dictionary from 8×8 sized coefficient patches and (d) trained dictionary of 64×100. Dictionary atoms are represented using 2 coefficients each

Compressed sensing MRI reconstruction

The proposed reconstruction model and involved reconstruction procedure are demonstrated in this section. It has been proved that composite regularization performs better than either spatial image TV regularization or sparse coefficients l 1 regularization in reconstruction [5, 21, 33]. Besides, the lowpass UDCT sub-band α l is the rough approximation of spatial image to be reconstructed and contains large amounts of information, which indicates that TV regularization on lowpass UDCT sub-band coefficients may further improve edge details and suppress noise, promoting the reconstruction quality. According to the different structural features of spatial image, lowpass sub-band α l and highpass sub-bands α h , a new composite regularization reconstruction model is thereby proposed to handle various regularization efficiently in this paper. The reconstruction model can be denoted as

$$\begin{array}{@{}rcl@{}} \widetilde{\boldsymbol{\alpha}}&=&\min_{\boldsymbol{\Gamma}^{r}_{R},\boldsymbol{\alpha}} TV\left(\boldsymbol{\Psi}^{-1}\boldsymbol{\alpha}\right)+\left\|{\boldsymbol{\alpha}_{h}}\right\|_{1}+TV\left(\textbf{D}^{r}\boldsymbol{\Gamma}^{r}_{R}\right)+\left\|\textbf{D}^{r}\boldsymbol{\Gamma}^{r}_{R}\right\|_{1}\\ \textbf{x}&=&\boldsymbol{\Psi}^{-1}\widetilde{\boldsymbol{\alpha}} \ \ \text{s.t.} \ \left\|\textbf{F}_{u} \textbf{x}-\textbf{y}\right\|^{2}_{2} \leq \boldsymbol{\varepsilon} \end{array} $$

where \(\boldsymbol {\varepsilon }\in \mathbb {C}^{m}\) is proportional to the noise standard deviation and controls the allowed noise level. The subscript h means highpass UDCT sub-bands coefficients. Spatial image TV regularization T V(Ψ −1 α) and lowpass UDCT sub-band coefficients TV regularization \(TV\left (\textbf {D}^{r}\boldsymbol {\Gamma }^{r}_{R}\right)\) regularize image without smoothing the boundaries, guaranteeing clear edge details in the reconstructed result. UDCT sub-bands coefficients l 1 regularization realizes sparsity and automatic selection of the most important characteristics. Undersampled k-space data fidelity term makes the reconstruction fitting the measurements. This reconstruction model is effective for embedding efficient composite regularization according to the different structural features of spatial image, lowpass and highpass UDCT sub-bands coefficients. Since TV regularization is capable of maintaining the boundaries of the objects, the reconstructed edge details can be further strengthened by two level of TV regularization. Besides, integrating \(\boldsymbol {\Gamma }^{r}_{R}\) modification into lowpass UDCT sub-band coefficients TV and l 1 regularization can guarantee the accuracy of updated \(\boldsymbol {\Gamma }^{r}_{R}\), reduce the complexity of the reconstruction model and overfitting significantly.

Basic thought to settle this reconstruction model is based on C-SALSA. And it is slightly different from C-SALSA because the designed composite regularization contains more than one regularization terms. Since parameter ε is clearly defined and easy to set, the previous AL based methods ignore it and introduce other regularization parameters λ 1(2). They are thus inefficient [47], such as TVCMRI, FCSA and SALSA, etc. While in the proposed method, the constrained problem (10) is equivalent to an unconstrained one with a discontinuous objective based on the thoughts of C-SALSA

$$ {\begin{aligned} &\min_{\boldsymbol{\Gamma}^{r}_{R},\boldsymbol{\alpha},\textbf{x}} \lambda_{1}\left[TV\left(\boldsymbol{\Psi}^{-1}\boldsymbol{\alpha}\right)+\left\|{\boldsymbol{\alpha}}_{h}\right\|_{1}+TV\left(\textbf{D}^{r}\boldsymbol{\Gamma}^{r}_{R}\right)+\left\|\textbf{D}^{r}\boldsymbol{\Gamma}^{r}_{R}\right\|_{1}\right]\\ &\quad+\lambda_{2}\mathscr{L}_{E(\boldsymbol{\varepsilon},\textbf{I},\textbf{y})}\left(\textbf{F}_{u} \textbf{x}\right) \end{aligned}} $$

where λ 1 measures the weight of TV and l 1 regularization, λ 2 measures the weight of k-space data fidelity and E (ε,I,y) is simply a closed ε-radius Euclidean ball centered at y. As is shown, (11) can be treated as a special case of E q.30 in [47] by defining \(\Phi \left (\cdot \right)=TV\left (\boldsymbol {\Psi }^{-1}\boldsymbol {\alpha }\right)+\left \|\boldsymbol {\alpha }_{h}\right \|_{1}+TV\left (\textbf {D}^{r}\boldsymbol {\Gamma }^{r}_{R}\right)+\left \|\textbf {D}^{r}\boldsymbol {\Gamma }^{r}_{R}\right \|_{1}\). It has the form of E q.13 in [47] with J=5 number of functions. These functions in (11) are closed, proper and convex (for details, refer to [47]). Minimization problem (11) is allowed to be decoupled into 5 independent and resoluble ones by a particular mapping way. These 5 independent subproblems include spatial image TV regularization T V(Ψ −1 α), lowpass UDCT coefficients TV regularization \(TV\left (\textbf {D}^{r}\boldsymbol {\Gamma }^{r}_{R}\right)\), UDCT sub-bands coefficients l 1 regularization \(\left \|\textbf {D}^{r}\boldsymbol {\Gamma }^{r}_{R}\right \|_{1}\) and α h 1 and k-space data fidelity \(\mathscr {L}_{E(\boldsymbol {\varepsilon },\textbf {I},\textbf {y})}\left (\textbf {F}_{u} \textbf {x}\right)\). Since all the functions in (11) are closed, proper, convex and [I Ψ F u ]T has full column rank (Ψ itself is a full column rank matrix), convergence of ADMM-2 involving problem (11) is guaranteed according to Theorem1 (Eckstein-Bertsekas, [59]). The solution of (11) is enforced to approach that of (10) via slowly taking λ 1(2) to very large values by multiplying with a continuous factor ρ (a continuation process with initial values as λ 10, λ 20 and ρ>1). Introducing ADMM-2 to this particular case requires the definition of the Moreau proximal maps associated with l 1 regularization, TV regularization and \(\mathscr {L}_{E(\boldsymbol {\varepsilon }, \textbf {I}, \textbf {y})}\). Since the input of regularizer can be spatial image, UDCT sub-bands coefficients and representation coefficients, μ is introduced to represent the input of regularization uniformly. The Moreau proximal map function of regularization \(\Phi \left (\boldsymbol {\mu }\right):\mathbb {C}^{p}\rightarrow {}\mathbb {C}^{p}\) is denoted as

$$ \boldsymbol{\Theta}_{\Phi}\left(\hat{\boldsymbol{\mu}}\right)=\arg \min_{\boldsymbol{\mu}}\frac{1}{2}\left\|\boldsymbol{\mu}-\hat{\boldsymbol{\mu}}\right\|^{2}_{2}+\Phi\left(\boldsymbol{\mu}\right) $$

where \(\hat {\boldsymbol {\mu }}\) is the result of mapping to μ according to the mapping relation \(\mathbb {C}^{p}\rightarrow {}\mathbb {C}^{p}\). A simple soft threshold method \(\boldsymbol {\Theta }_{\lambda _{1}\Phi }(\cdot)=soft(\cdot, \lambda _{1})\) solves the l 1 regularization relevant to α h 1 and \(\left \|\textbf {D}^{r}\boldsymbol {\Gamma }^{r}_{R}\right \|_{1}\). The available fast Chambolle’s algorithm [60] handles TV regularization efficiently. Define ν=F u x via VS technique, then the Moreau proximal map of \(\mathscr {L}_{E(\boldsymbol {\varepsilon },\textbf {I},\textbf {y})}\) is simply the orthogonal projection of ν on the closed ε-radius ball centered at y, which can be attacked by

$$ \boldsymbol{\Theta}_{\lambda_{2}\mathscr{L}_{E(\boldsymbol{\varepsilon},\textbf{I},\textbf{y})}}\left(\boldsymbol{\nu}\right)=\left\{ \begin{array}{ll} \textbf{y}+\boldsymbol{\varepsilon}\frac{\boldsymbol{\nu}-\textbf{y}}{\left\|\boldsymbol{\nu}-\textbf{y}\right\|_{2}} &\text{if} \left\|\boldsymbol{\nu}-\textbf{y}\right\|_{2} \textgreater \boldsymbol{\varepsilon}\\ \boldsymbol{\nu} &\text{if} \left\|\boldsymbol{\nu}-\textbf{y}\right\|_{2} \leq \boldsymbol{\varepsilon} \end{array} \right. $$

In ADMM-2, error minimization with respect to ν aims at fitting measurements of k-space. Regularization terms with respect to spatial image, α h and \(\boldsymbol {\Gamma }^{r}_{R}\) avoid overfitting. The fidelity and regularization terms optimization are implemented alternatively. \(\boldsymbol {\Gamma }^{r}_{R}\) is modified from the weighted average between results of \(TV\left (\textbf {D}^{r}\boldsymbol {\Gamma }^{r}_{R}\right)\) and \(\left \|\textbf {D}^{r}\boldsymbol {\Gamma }^{r}_{R}\right \|_{1}\). Then lowpass UDCT sub-band coefficients are obtained as the weighted average between results of T V(Ψ −1 α) and \(\textbf {D}^{r}\boldsymbol {\Gamma }^{r}_{R}\). The modified highpass UDCT sub-bands coefficients are obtained as the weighted average between results of T V(Ψ −1 α) and α h 1. Then image in spatial domain is acquired as the result of inverse UDCT to UDCT coefficients. The subproblem with respect to E(ε,I,y) can be solved via (13) efficiently. The ultimate reconstructed image x i (i as counter of iterations for ADMM-2) is the result of the reweighted average between the above spatial domain image through regularization penalty and result of (13) for once iteration. Such process reduces reconstruction error and brings about a high-quality reconstructed image efficiently. The flowchart for LSECSMRI reconstruction in Fig. 3 exhibits clearly the reconstruction process based on the proposed sparsity structure.

Fig. 3
figure 3

Flowchart for LSECSMRI reconstruction

Summary of LSECSMRI

In LSECSMRI procedure, Ψ is uniform discrete curvelet decomposition operator of S levels (s=1,2,…,S) with 2κ s directions for each level, performing on real and imaginary parts of complex-valued MR data, respectively. The obtained UDCT sub-bands coefficients, representation coefficients of lowpass UDCT sub-band coefficients over D r and measurements y are sent into the composite regularization reconstruction model for MR image reconstruction. The proposed reconstruction method has some advantages as follows. Firstly, the reconstruction method handles the lowpass sub-band and highpass directional sub-bands coefficients regularization separately with different regularization methods, allowing effective reconstruction. Besides, adding the indicator function of ε into the objective makes the resulting problem equivalent to the original problem (10) based on the thoughts of C-SALSA. The resulting problem can be decoupled into several subproblems, which are easy to solve with fast convergent algorithms. Additionally, \(\boldsymbol {\Gamma }^{r}_{R}\) is modified by using the result of TV and l 1 regularization on lowpass UDCT sub-band coefficients, which reduces the computational complexity and overfitting significantly. Hence, the proposed reconstruction approach provides accurate solution along with rapid convergence speed. Superiorities of LSECSMRI are confirmed experimentally later.

Results and discussions

Experimental setup

Experiments are performed under nonuniform sampling schemes at various sampling rates (defined as ratio of \(\frac {m}{p}\in \left [0,1\right ]\)) in this section. The performance of LSECSMRI is analyzed from 4 different aspects. Images used in the experiments are complex-valued MR T2wBrain_slice_27, complex-valued water phantom acquired at 7T Varian MRI system with spin echo sequence (T R/T E=200/100 ms, 80×80 mm field of view, 2 mm slice thickness) [37] and real-valued MBA_T2_slice006 (image courtesy of Densities of MR images are normalized to a maximum amplitude of 1 for simulated experiments, via dividing each element value by the maximum module value of image pixels. The mainly applied sampling schemes (binary masks with values of 0 and 1) are Cartesian sampling with random phase encodes, 2D random sampling and pseudo radial sampling [6], etc. Setting of sampling rate depends on numerous experiments. It is appropriate if unobvious visual improvement of reconstructed quality is obtained with sampling rate going on increasing. In simulation, k-space measurements are obtained via dot multiplication between Fourier transform coefficients of raw image and sampling matrix at given sampling rate. Raw images and sampling schemes used in experiments are exhibited in Fig. 4. The proposed algorithm is compared with DLMRI, iterative thresholding CS MRI based on more redundant SFLCT(MRSFLCT based CS MRI), LORAKS (rank constraint r S =90 and neighborhood size R=4 in k-space) and PANO. All experiments are implemented in MATLAB R2011b of a 64-bit Windows 7 operating system with an Intel Xeon E5 CPU at 2.80 GHz and 8 GB memory. Matlab implementations of compared DLMRI, MRSFLCT based CS MRI, LORAKS and PANO are available from the authors’ websites,, 8 iterations of DL and reconstruction alternation procedure are adopted in DLMRI. Parameters needed in LSECSMRI are manually and empirically set for optimal reconstruction results via numerous of tests. Take UDCT decomposition level S and directional sub-bands in each level 2κ s for example, when they reach certain values, increasing their values doesn’t lead to significant improvement of reconstruction quality but increases computational complexity. Trade-off values are thus obtained by measuring the reconstruction quality and computational complexity. In practice, first commonly set S=3 and 2κ s =12, then reduce and increase them to observe the changes of the reconstruction quality. Therefore, to reconstruct images in Fig. 4(a) and (b), UDCT decomposition of S=1 and 2κ s =12 acts on intermediate images to obtain 13 coefficient sub-bands, including one lowpass sub-band and 12 highpass directional sub-bands for real and imaginary parts separately. For reconstructing image in Fig. 4(c), UDCT decomposition of S=3 and 2κ s =12 is employed. Size of dictionary is 64×100. For T2wBrain_slice_27 reconstruction, initial values of λ 1 and λ 2 are λ 10=λ 20=0.005, continuous factor is ρ=1.3. For water phantom reconstruction, λ 10=λ 20=6, continuous factor is ρ=1.4. The preset maximum number of ADMM-2 iterations I=60 is used as ultimate iteration stop criterion. Numerical metrics of quality assessment for reconstructed images are peak signal-to-noise ratio(PSNR) (in dB), transferred edge information(TEI) [61] and relative l 2 norm error(RLNE) [36].

Fig. 4
figure 4

Raw images and sampling schemes. (a) T2wBrain_slice_27, (b) MBA_T2_slice006, (c) water phantom, (d) Cartesian sampling, (e) 2D variable density random sampling and (f) pseudo radial sampling

Comparison with earlier CS MRI methods

The whole performance of LSECSMRI is tested on raw images in Fig. 4(a)-(c) in this section. The proposed method is compared with DLMRI, MRSFLCT based CS MRI, LORAKS and PANO. Computational time is recorded accordingly. For DLMRI reconstruction, the recorded computational time is that of the first iteration. And the computational time of each iteration increases with iteration number.

Using Cartesian undersampling in Fig. 4(d) at 0.35 sampling rate for raw image in Fig. 4(a), DLMRI, MRSFLCT based CS MRI, LORAKS, PANO and LSECSMRI reconstructed results are demonstrated in Fig. 5. Figure 5(a)-(c) indicate clearly that DLMRI, MRSFLCT based CS MRI and LORAKS reconstructed images show severe edge blurring, aliasing and artifacts, implying the poor performance of these methods. While PANO and LSECSMRI reconstructed images exhibit clear edge details and few artifacts. The local regions of reconstructed images in Fig. 5(d)-(f) are scaled up (by a factor of 2) for detailed observations. They clearly show that LSECSMRI performs slightly better in reconstructing clear curve-like details with 1.1dB higher in PSNR, 0.0145 higher in TEI and 0.0109 lower in RLNE compared with PANO (as areas marked in red arrow show). The corresponding difference image in Fig. 5(k) indicates that LSECSMRI reconstructed image obtains the least reconstruction error at the cost of large amounts of time consumption. The computational time is not the tough problem for it can be reduced by using the significant parallel imaging on graphics processing unit(GPU) like the ones in [18, 41, 62], which will be considered in our future work.

Fig. 5
figure 5

Performance of algorithms under Cartesian undersampling scheme at 35 % sampling rate. (a)–(e) Reconstructed images using DLMRI, MRSFLCT based CS MRI, LORAKS, PANO and LSECSMRI, respectively, (f) reconstructed image from fully sampled k-space data, (g)–(k) difference images between fully sampled MR image and images in (a)–(e) with gray scale of [0,0.25], respectively. PSNRs of them are 31.61, 32.28, 28.53, 36.33 and 3 7 . 4 0. TEIs are 0.6109, 0.6306, 0.4886, 0.7529, and 0 . 7 6 7 4. RLNEs are 0.1620, 0.1500, 0.2310, 0.0941 and 0 . 0 8 3 2 separately. And computational time is 56.5+80.4sec, 30.5sec, 664.7sec, 296.8sec and 50+492sec, respectively

Figure 6(a)-(c) presents PSNR, TEI and RLNE indices versus sampling rates of 2D variable density random sampling scheme separately for DLMRI, MRSFLCT based CS MRI, LORAKS, PANO and LSECSMRI reconstructed T2wBrain_slice_27. Figure 6(a)-(c) exhibit that PSNRs and TEIs of DLMRI, MRSFLCT based CS MRI, LORAKS and PANO reconstructed images are lower than those of LSECSMRI reconstructed image overall. When sampling rate is low (between 0.10 and 0.15), LSECSMRI reconstructed image obtain slightly higher PSNR, TEI and lower RLNE compared with that of PANO. While sampling rate gradually increases (sampling rate in the range of 0.15 to 0.45), LSECSMRI reconstructed image achieves considerable higher PSNR, TEI and lower RLNE than PANO reconstructed image. For instance, the PSNR, TEI and RLNE are 40.46dB, 0.8446 and 0.0585 separately for LSECSMRI reconstructed result at 0.25 sampling rate. To obtain comparable results, the sampling rate is approximate 0.30 for PANO, 0.45 for MRSFLCT based method, 0.53 for DLMRI and 0.60 for LORAKS. These indicate that LSECSMRI can use lower sampling rate to obtain comparable reconstruction result as the other four methods under high sampling rates. When sampling rate is high (0.50 and higher), the PSNRs of LSECSMRI reconstructed image are staying at the highest level among the compared five methods. The TEI and RLNE curves of PANO and LSECSMRI reconstructed images are comparable and tend to level off, implying that the reconstructed MR images have already obtain the most information from undersampled data and further more sampled data merely increase the data redundancy and calculations. The overall results demonstrate that LSECSMRI can obtain excellent sparsity structure and reconstruction performance from 2D variable density random undersampling scheme.

Fig. 6
figure 6

Reconstructed T2wBrain_slice_27 indices versus sampling rates of 2D variable density random sampling schemes. (a) PSNR versus sampling rates, (b) TEI versus sampling rates and (c) RLNE versus sampling rates

Figure 7(a)-(e) exhibit reconstructed results from DLMRI, MRSFLCT based CS MRI, LORAKS, PANO and LSECSMRI separately for image in Fig. 4 (b) under Cartesian undersampling scheme at 0.35 sampling rate. The difference images in Fig. 7(f)-(j) show that LSECSMRI reconstructed image possesses the least artifacts and reconstructed error among the compared methods. PSNR of LSECSMRI reconstructed image is 35.80dB, separately 4.85, 5.55, 7.47 and 0.59dB higher than that of DLMRI, MRSFLCT based CS MRI, LORAKS and PANO reconstructed images. RLNE of LSECSMRI reconstructed image is 0.0659, separately 0.0493, 0.0590, 0.0899 and 0.0046 lower than that of DLMRI, MRSFLCT based CS MRI, LORAKS and PANO reconstructed images. These indicate that LSECSMRI can obtain preeminent reconstruction performance among state-of-the-art methods.

Fig. 7
figure 7

Reconstructed MBA_T2_slice006 under Cartesian undersampling scheme at 0.35 sampling rate. (a)–(e) Reconstructed images from DLMRI, MRSFLCT based CS MRI, LORAKS, PANO and LSECSMRI, (f)–(j) difference images between fully sampled MR image and images in (a)–(e) with gray scale of [0,0.25], respectively. PSNRs of them are 30.95dB, 30.25dB, 28.33dB, 35.21dB and 3 5 . 8 0dB. TEIs are 0.5806, 0.5872, 0.5218, 0.7513, and 0 . 7 7 6 5. RLNEs are 0.1152, 0.1249, 0.1558, 0.0705 and 0 . 0 6 5 9 separately. And computational time is 62.7+90.5sec, 30.5sec, 711.3sec, 291.2sec and 52.7+514.6sec separately

Figure 8 exhibits reconstructed results for complex-valued water phantom. K-space measurements are obtained via pseudo radial undersampling scheme at 0.3020 sampling rate. Local regions in Fig. 8(g)-(l) (scaled by a factor of 2 to visualize details) exhibit that LSECSMRI shows great superiority in reconstructing clear edges and textures with the least blurring among the compared five methods. While DLMRI reconstructed image introduces severe edge blurring and MRSFLCT based CS MRI reconstructed image exhibits disordered directions.

Fig. 8
figure 8

Pseudo radial undersampling at 0.3020 sampling rate for water phantom reconstruction. (a)–(f) Reconstructed images from DLMRI, MRSFLCT based CS MRI, LORAKS, PANO, LSECSMRI and fully sampled k-space data and (g)–(l) local regions from (a)–(f). PSNRs for images in (a)–(e) are 33.14dB, 34.97dB, 31.83dB, 35.76dB and 3 5 . 9 2dB. RLNEs for images in (a)–(e) are 0.0464, 0.0375, 0.0539, 0.0343 and 0 . 0 3 3 6, respectively

Results on noisy data

In this section, the ability in handling noise is demonstrated for the proposed method. Random gaussian white noise with standard deviation 5.1 is added to the original k-space data. PSNR, TEI and RLNE are 36.19dB, 0.8208 and 0.0956 for fully sampled T2wBrain_slice_27 and 34.90dB, 0.8813 and 0.0731 for fully sampled MBA_T2_slice006, respectively. In simulation, regularization parameters are manually optimized to obtain maximum PSNRs, TEIs and minimum RLNEs for the compared five methods.

Table 1 and Table 2 show numerical values of PSNRs, TEIs and RLNEs for reconstructed T2wBrain_slice_27 and MBA_T2_slice006 under Cartesian sampling scheme at various sampling rates, respectively. Table 1 and Table 2 exhibit similar change trend. They both show that PSNR and TEI of LSECSMRI reconstructed result are always the highest compared with those of the other four methods reconstructed results when sampling rate is between 0.15 and 0.85, implying superior edge information transfer, low reconstruction error and minimum reconstruction noise of LSECSMRI reconstructed result. As is seen, the PSNR is 34.31dB for LSECSMRI reconstructed T2wBrain_slice_27 at 0.35 sampling rate. To obtain the comparable result, the sampling rate is approximate 0.40 for PANO, 0.75 for LORAKS, 0.60 for MRSFLCT based method and over 0.55 for DLMRI. These indicate that LSECSMRI can use lower sampling rate to obtain comparable reconstructed results as the other four methods at high sampling rates. LORAKS reconstructed result obtains the highest PSNR, TEI and the lowest RLNE when sampling rate reaches 0.95, which is a sign that the significance of sparsity and CS cannot be reflected when sampling rate approaches whole sampling.

Table 1 Reconstructed T2wBrain_slice_27 quality indices versus Cartesian sampling rates
Table 2 Reconstructed MBA_T2_slice006 quality indices versus Cartesian sampling rates

Convergence analysis

Convergence performance of LSECSMRI reconstruction method is analyzed in this section. It is evaluated by MSE decline curve versus successive iterations. Images in Fig. 4(a)-(b) using Cartesian undersampling scheme at 0.35 sampling rate and Fig. 4(c) under pseudo radial undersampling scheme at 0.3020 sampling rate are used for test. LSECSMRI reconstruction is compared with TVCMRI, FCSA, UDCT based C-SALSA reconstruction with Φ in E q.30 in [47] representing l 1 regularization(UDCS_l1) and UDCT based C-SALSA reconstruction with Φ in E q.30 in [47] representing TV regularization(UDCS_TV). The maximum ADMM-2 iteration number is I=70. All the parameters are manually optimized for maximum PSNRs, TEIs and minimum RLNEs in reconstruction.

Figure 9(a)-(c) exhibit MSE decline curves versus iteration number by using the compared five reconstruction methods for reconstructing images in Fig. 4(a)-(c), respectively. The graphs in the second row show the MSE decline curves in fine scale when iteration number is greater than 30 for UDCS_l1, UDCS_TV and LSECSMRI reconstruction separately. It is concluded from Fig. 9(a)-(c) that LSECSMRI can obtain quite rapid convergence speed with an apparently low MSE, indicating that LSECSMRI reconstruction model is rational and the corresponding reconstruction algorithm is efficient in reconstructing high-quality images.

Fig. 9
figure 9

Reconstructed MSE decline versus iteration. (a) Reconstructing Fig. 4(a) under Cartesian undersampling at 0.35 sampling rate, (b) reconstructing Fig. 4(b) under Cartesian undersampling at 0.35 sampling rate and (c) reconstructing Fig. 4(c) under pseudo radial undersampling at 0.3020 sampling rate

Parameters in LSECSMRI

Major parameters in LSECSMRI include UDCT decomposition level (S), number of UDCT directional sub-bands for each level, size of coefficient patches (n) to train dictionary, number of dictionary atoms (K), initial values λ 10 and λ 20 of regularization parameters and the continuous factor ρ. Influence of these parameters fluctuation on LSECSMRI reconstruction performance is evaluated by RLNE. The test image is T2wBrain_slice_27. Cartesian undersampling scheme at 0.35 sampling rate is used to undersample k-space data of the test image. Experiments indicate that LSECSMRI performs better when λ 1=λ 2 than λ 1λ 2. Set the initial parameters values as S=1, 2κ s =12, n=64, K=100, λ 10=λ 20=0.005 and ρ=1.3.

Increasing S within the allowed scope of image size increases the reconstruction error slightly, as is shown in Fig. 10(a). Figure 10(b) illustrates that appropriate number of directional sub-bands gives rise to the least reconstruction error. Reducing and increasing the number both lead to a drop in the reconstructed quality. While size of lowpass UDCT sub-band coefficients patches for DL and number of dictionary atoms seldom affect the reconstruction error, as is shown in Fig. 10(c) and Fig. 10(d). One explanation is that the update of \(\boldsymbol {\Gamma }^{r}_{R}\) depends mainly upon the result of TV and l 1 regularization on lowpass UDCT sub-band coefficients. And the TV and l 1 regularization on lowpass UDCT sub-band coefficients are solved by stably convergent algorithms, which guarantees the validity and stability to dictionary size of the reconstructed results. Since parameter λ 1(2) measures the weight between regularization and data consistency, it makes sense to analyze how the initial regularization parameter λ 10 (or λ 20) and continuous factor ρ influence the reconstruction. Figure 10(e) and Fig. 10(f) show influences of λ 10 and ρ on the reconstruction performance separately. As is exhibited in Fig. 10(e), increasing λ 10 reduces the reconstruction error when λ 10 is relatively small (λ 10≤0.005 for T2wBrain_slice_27). Whereas too large λ 10 (λ 10>0.01) will increase the reconstruction error. Similar fluctuation trend is obtained for ρ and ρ=1.3 is suggested for T2wBrain_slice_27 reconstruction.

Fig. 10
figure 10

RLNEs versus LSECSMRI parameters for T2wBrain_slice_27 under Cartesian undersampling at 0.35 sampling rate. (a) S (UDCT decomposition level), (b) number of directional sub-bands, (c) n (patch size), (d) number of dictionary atoms (K), (e) initial regularization parameter λ 10 (or λ 20) and (f) continuous factor ρ when other parameters are fixed

In short words, a series of experiments indicate that the proposed CS MRI method owns preeminent sparsity structure and reconstruction with rapidly convergent speed, and thereby outperforms current CS MRI techniques in reconstructing high-quality image with clear edge details at a low sampling rate. Experiments on noise, convergence speed and parameters demonstrate its superiority in suppressing noise, convergence and robustness to parameters fluctuation among current CS MRI methods.

Conclusions and future work

A novel local sparsity enhanced composite sparsity structure is presented, in which UDCT decomposes image to produce structural sparsity and later dictionary is learnt from the lowpass UDCT sub-band coefficients to adaptively sparsify images further. Reconstruction model is then proposed for significant MR images reconstruction established on the proposed sparse representation approach in this paper. Comparing reconstruction performance indicates that the proposed sparsity structure obtains preeminent structural sparsity. The proposed reconstruction model optimizes sparse regularization and constrains measurements fidelity to recover original signal efficiently with a rapidly and stably convergent speed. Experimental results on LSECSMRI agree well with the theoretical analysis, and exhibit superiority in reconstructing highly undersampled MR images under a variety of sampling schemes compared with current CS MRI frameworks. Since the proposed method is simply tested by three MR images in this paper, its universality remains to be investigated. Besides, handling the real and imaginary parts separately doubles the amount of calculations. Further improvements and verifications on the method are subjects of ongoing research and can be made from the following three aspects: (1) test the method on more datasets acquired in real applications; (2) introduce LSECSMRI into practical 3D MRI application by adding the sparse regularization term defined along the temporal axis into the reconstruction model; (3) minimize the modified reconstruction model by using ADMM based methods with partially parallel imaging(PPI) on GPU and faster languages such as C/C++ to speed up the imaging.

Ethics statement

Brain image of MBA_T2_slice006 was downloaded from Harvard Medical School( The rest human images were acquired from healthy subjects under the approval of the Institute Review Board of Xiamen University and written consent was obtained from the participants. The data were analyzed anonymously.


  1. Donoho DL. Compressed sensing. Inform Theory IEEE Trans. 2006; 52(4):1289–306. doi:10.1109/TIT.2006.871582.

    Article  Google Scholar 

  2. Candès EJ, Romberg J, Tao T. Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. Inform Theory IEEE Trans. 2006; 52(2):489–509.

    Article  Google Scholar 

  3. Baraniuk R. Compressive sensing. IEEE Signal Process Mag. 2007; 24(4):118–21.

    Article  Google Scholar 

  4. Lustig M, Donoho D, Pauly JM. Sparse mri: The application of compressed sensing for rapid mr imaging. Magn Reson Med. 2007; 58(6):1182–95.

    Article  PubMed  Google Scholar 

  5. Lustig M, Donoho DL, Santos JM, Pauly JM. Compressed sensing mri. Signal Process Mag IEEE. 2008; 25(2):72–82. doi:10.1109/MSP.2007.914728.

    Article  Google Scholar 

  6. Chen Y, Ye X, Huang F. A novel method and fast algorithm for mr image reconstruction with significantly under-sampled data. Inverse Probl Imaging. 2010; 4(2):223–40.

    Article  CAS  Google Scholar 

  7. Santos JM, Cunningham CH, Lustig M, Hargreaves BA, Hu BS, Nishimura DG, Pauly JM. Single breath-hold whole-heart mra using variable-density spirals at 3t. Magn Reson Med. 2006; 55(2):371–9.

    Article  PubMed  Google Scholar 

  8. Haldar JP, Hernando D, Liang ZP. Compressed-sensing mri with random encoding. Med Imaging IEEE Trans. 2011; 30(4):893–903.

    Article  Google Scholar 

  9. Huang J, Yang F. Compressed magnetic resonance imaging based on wavelet sparsity and nonlocal total variation. In: Biomedical Imaging (ISBI), 2012 9th IEEE International Symposium On. Barcelona: IEEE: 2012. p. 968–71.

    Google Scholar 

  10. Lu Y, Do MN. A new contourlet transform with sharp frequency localization. In: Image Processing, 2006 IEEE International Conference On. Atlanta, GA: IEEE: 2006. p. 1629–1632.

    Google Scholar 

  11. Qu X, Zhang W, Guo D, Cai C, Cai S, Chen Z. Iterative thresholding compressed sensing mri based on contourlet transform. Inverse Probl Sci Eng. 2010; 18(6):737–58.

    Article  Google Scholar 

  12. Candes E, Demanet L, Donoho D, Ying L. Fast discrete curvelet transforms. Multiscale Model Simul. 2006; 5(3):861–99.

    Article  Google Scholar 

  13. Ma J. Compressed sensing by inverse scale space and curvelet thresholding. Appl Math Comput. 2008; 206(2):980–8.

    Google Scholar 

  14. Plonka G, Ma J. Curvelet-wavelet regularized split bregman iteration for compressed sensing. Int J Wavelets, Multiresolution Inform Process. 2011; 9(01):79–110.

    Article  Google Scholar 

  15. Lim WQ. The discrete shearlet transform: A new directional transform and compactly supported shearlet frames. Image Process IEEE Trans. 2010; 19(5):1166–1180.

    Article  Google Scholar 

  16. Qin J, Guo W. An efficient compressive sensing mr image reconstruction scheme. In: Biomedical Imaging (ISBI), 2013 IEEE 10th International Symposium On. San Francisco, CA: IEEE: 2013. p. 306–9.

    Google Scholar 

  17. Jung H, Sung K, Nayak KS, Kim EY, Ye JC. k-t focuss: A general compressed sensing framework for high resolution dynamic mri. Magn Reson Med. 2009; 61(1):103–16.

    Article  PubMed  Google Scholar 

  18. Otazo R, Kim D, Axel L, Sodickson DK. Combination of compressed sensing and parallel imaging for highly accelerated first-pass cardiac perfusion mri. Magn Reson Med. 2010; 64(3):767–76.

    Article  PubMed  PubMed Central  Google Scholar 

  19. Bilen C, Wang Y, Selesnick IW. High-speed compressed sensing reconstruction in dynamic parallel mri using augmented lagrangian and parallel processing. Emerg Select Topics Circuits Syst IEEE J. 2012; 2(3):370–9.

    Article  Google Scholar 

  20. Qu X, Cao X, Guo D, Hu C, Chen Z. Combined sparsifying transforms for compressed sensing mri. Electron Lett. 2010; 46(2):121–3.

    Article  Google Scholar 

  21. Huang J, Zhang S, Metaxas D. Efficient mr image reconstruction for compressed mr imaging. Med Image Anal. 2011; 15(5):670–9.

    Article  PubMed  Google Scholar 

  22. Lewicki MS, Sejnowski TJ. Learning overcomplete representations. Neural Comput. 2000; 12(2):337–65.

    Article  CAS  PubMed  Google Scholar 

  23. Rauhut H, Schnass K, Vandergheynst P. Compressed sensing and redundant dictionaries. Inform Theory IEEE Trans. 2008; 54(5):2210–19. doi:10.1109/TIT.2008.920190.

    Article  Google Scholar 

  24. Mairal J, Sapiro G, Elad M. Multiscale sparse image representationwith learned dictionaries. In: Image Processing, 2007. ICIP 2007. IEEE International Conference On. San Antonio, TX: IEEE: 2007. p. 105.

    Google Scholar 

  25. Rubinstein R, Zibulevsky M, Elad M. Double sparsity: Learning sparse dictionaries for sparse signal approximation. Signal Process IEEE Trans. 2010; 58(3):1553–64.

    Article  Google Scholar 

  26. Dabov K, Foi A, Katkovnik V, Egiazarian K. Image denoising by sparse 3-d transform-domain collaborative filtering. Image Process IEEE Trans. 2007; 16(8):2080–95.

    Article  Google Scholar 

  27. Adluru G, Tasdizen T, Schabel MC, DiBella EV. Reconstruction of 3d dynamic contrast-enhanced magnetic resonance imaging using nonlocal means. J Magn Reson Imaging. 2010; 32(5):1217–27.

    Article  PubMed  Google Scholar 

  28. Fang S, Ying K, Zhao L, Cheng J. Coherence regularization for sense reconstruction with a nonlocal operator (cornol). Magn Reson Med. 2010; 64(5):1413–25.

    Article  PubMed  Google Scholar 

  29. Liang D, Wang H, Chang Y, Ying L. Sensitivity encoding reconstruction with nonlocal total variation regularization. Magn Reson Med. 2011; 65(5):1384–92.

    Article  PubMed  Google Scholar 

  30. Wong A, Mishra A, Fieguth P, Clausi DA. Sparse reconstruction of breast mri using homotopic minimization in a regional sparsified domain. Biomed Eng IEEE Trans. 2013; 60(3):743–52.

    Article  Google Scholar 

  31. Akçakaya M, Basha TA, Goddu B, Goepfert LA, Kissinger KV, Tarokh V, Manning WJ, Nezafat R. Low-dimensional-structure self-learning and thresholding: Regularization beyond compressed sensing for mri reconstruction. Magn Reson Med. 2011; 66(3):756–67.

    Article  PubMed  PubMed Central  Google Scholar 

  32. Qu X, Hou Y, Lam F, Guo D, Zhong J, Chen Z. Magnetic resonance image reconstruction from undersampled measurements using a patch-based nonlocal operator. Med Image Anal. 2014; 18(6):843–56.

    Article  PubMed  Google Scholar 

  33. Ma S, Yin W, Zhang Y, Chakraborty A. An efficient algorithm for compressed mr imaging using total variation and wavelets. In: Computer Vision and Pattern Recognition, 2008. CVPR 2008. IEEE Conference On. Anchorage, AK: IEEE: 2008. p. 1–8.

    Google Scholar 

  34. Chen C, Huang J. Compressive sensing mri with wavelet tree sparsity. In: Advances in Neural Information Processing Systems. Beijing: IEEE: 2012. p. 1115–1123.

    Google Scholar 

  35. Ravishankar S, Bresler Y. Mr image reconstruction from highly undersampled k-space data by dictionary learning. Med Imaging IEEE Trans. 2011; 30(5):1028–41. doi:10.1109/TMI.2010.2090538.

    Article  Google Scholar 

  36. Qu X, Guo D, Ning B, Hou Y, Lin Y, Cai S, Chen Z. Undersampled mri reconstruction with patch-based directional wavelets. Magn Reson Imaging. 2012; 30(7):964–77.

    Article  PubMed  Google Scholar 

  37. Ning B, Qu X, Guo D, Hu C, Chen Z. Magnetic resonance image reconstruction using trained geometric directions in 2d redundant wavelets domain and non-convex optimization. Magn Reson Imaging. 2013; 31(9):1611–22.

    Article  PubMed  Google Scholar 

  38. Yang J, Zhang Y, Yin W. A fast alternating direction method for tvl1-l2 signal reconstruction from partial fourier data. Selected Topics Signal Process IEEE J. 2010; 4(2):288–97.

    Article  Google Scholar 

  39. Haldar J. Low-rank modeling of local k-space neighborhoods (loraks) for constrained mri. Med Imaging IEEE Trans. 2014; 33(3):668–81.

    Article  Google Scholar 

  40. Zhao F, Noll DC, Nielsen JF, Fessler JA. Separate magnitude and phase regularization via compressed sensing. Med Imaging IEEE Trans. 2012; 31(9):1713–23.

    Article  Google Scholar 

  41. Shin PJ, Larson PE, Ohliger MA, Elad M, Pauly JM, Vigneron DB, Lustig M. Calibrationless parallel imaging reconstruction based on structured low-rank matrix completion. Magn Reson Med. 2013; 72(4):959–70.

    Article  PubMed  PubMed Central  Google Scholar 

  42. Lam F, Zhao B, Liu Y, Liang ZP, Weiner M, Schuff N. Accelerated fmri using low-rank model and sparsity constraints. In: Proc. Int. Soc. Magn. Reson. Med. Salt Lake City. p. 2013.

  43. Haldar J. Calibrationless partial fourier reconstruction of mr images with slowly-varying phase: A rank-deficient matrix recovery approach. In: Proc. ISMRM Workshop on Data Sampling & Image Reconstruction. Sedona. p. 2013.

  44. Jain P, Netrapalli P, Sanghavi S. Low-rank matrix completion using alternating minimization. In: Proceedings of the Forty-fifth Annual ACM Symposium on Theory of Computing. ACM. p. 2013. arXiv:1212.0467v1.

  45. Otazo R, Candès E, Sodickson DK. Low-rank plus sparse matrix decomposition for accelerated dynamic mri with separation of background and dynamic components. Magn Reson Med. 2015; 73(3):1125–36. doi:10.1002/mrm.25240.

    Article  PubMed  Google Scholar 

  46. Nguyen TT, Chauris H. Uniform discrete curvelet transform. Signal Process IEEE Trans. 2010; 58(7):3618–34. doi:10.1109/TSP.2010.2047666.

    Article  Google Scholar 

  47. Afonso MV, Bioucas-Dias JM, Figueiredo MAT. An augmented lagrangian approach to the constrained optimization formulation of imaging inverse problems. Image Process IEEE Trans. 2011; 20(3):681–95. doi:10.1109/TIP.2010.2076294.

    Article  Google Scholar 

  48. Yang J, Zhang Y. Alternating direction algorithms for l 1-problems in compressive sensing. SIAM J Sci Comput. 2011; 33(1):250–78.

    Article  CAS  Google Scholar 

  49. Esser E. Applications of lagrangian-based alternating direction methods and connections to split bregman. CAM Report. 2009; 9:31.

    Google Scholar 

  50. Afonso MV, Bioucas-Dias JM, Figueiredo MA. Fast image recovery using variable splitting and constrained optimization. Image Process IEEE Trans. 2010; 19(9):2345–56.

    Article  Google Scholar 

  51. Natarajan BK. Sparse approximate solutions to linear systems. SIAM J Comput. 1995; 24(2):227–34.

    Article  Google Scholar 

  52. Chartrand R. Exact reconstruction of sparse signals via nonconvex minimization. Signal Process Lett IEEE. 2007; 14(10):707–10. doi:10.1109/LSP.2007.898300.

    Article  Google Scholar 

  53. Aharon M, Elad M, Bruckstein A. k -svd: An algorithm for designing overcomplete dictionaries for sparse representation. Signal Process IEEE Trans. 2006; 54(11):4311–322. doi:10.1109/TSP.2006.881199.

    Article  Google Scholar 

  54. Elad M, Aharon M. Image denoising via sparse and redundant representations over learned dictionaries. Image Process IEEE Trans. 2006; 15(12):3736–45. doi:10.1109/TIP.2006.881969.

    Article  Google Scholar 

  55. Candes EJ, Donoho DL. Curvelets: A surprisingly effective nonadaptive representation for objects with edges. Technical report, DTIC Document Stanford, Calif: Stanford University; 2000.

    Google Scholar 

  56. Do MN, Vetterli M. The contourlet transform: an efficient directional multiresolution image representation. Image Process IEEE Trans. 2005; 14(12):2091–106.

    Article  Google Scholar 

  57. Da Cunha AL, Zhou J, Do MN. The nonsubsampled contourlet transform: theory, design, and applications. Image Process IEEE Trans. 2006; 15(10):3089–101.

    Article  Google Scholar 

  58. Candes EJ, Donoho DL. New tight frames of curvelets and optimal representations of objects with piecewise c2 singularities. Commun Pure Appl Math. 2004; 57(2):219–66.

    Article  Google Scholar 

  59. Eckstein J, Bertsekas DP. On the Douglas–Rachford splitting method and the proximal point algorithm for maximal monotone operators. Math Program. 1992; 55(1–3):293–318.

    Article  Google Scholar 

  60. Chambolle A. An algorithm for total variation minimization and applications. J Math Imaging Vis. 2004; 20(1–2):89–97.

    Google Scholar 

  61. Qu G, Zhang D, Yan P. Information measure for performance of image fusion. Electron Lett. 2002; 38(7):313–5.

    Article  Google Scholar 

  62. Liang D, Liu B, Wang J, Ying L. Accelerating sense using compressed sensing. Magn Reson Med. 2009; 62(6):1574–1584.

    Article  PubMed  Google Scholar 

Download references


The authors would like to thank Xiaobo Qu for providing raw MR data and useful advice. This work was partially supported by National Natural Science Foundation of China (Grant No.61175012, 61201421, 61201422), Science Foundation of Gansu Province of China (Grant No.1208RJZA265), Specialized Research Fund for the Doctoral Program of Higher Education of China (Grant No.20110211110026), the Fundamental Research Funds for the Central Universities of China (Grant No.lzujbky-2013-k06, lzujbky-2015-108).

Author information

Authors and Affiliations


Corresponding author

Correspondence to Yide Ma.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

BXY and MY were the principal investigators, participated in the framework design, carried out experiments and drafted the manuscript. YDM participated in the design of the framework, helped drafting and revising the manuscript. JWZ and KZ helped analyzing experimental results and revising the manuscript. All authors read and approved the final manuscript.

Rights and permissions

Open Access  This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made.

The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.

To view a copy of this licence, visit

The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Yang, B., Yuan, M., Ma, Y. et al. Local sparsity enhanced compressed sensing magnetic resonance imaging in uniform discrete curvelet domain. BMC Med Imaging 15, 28 (2015).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI:


  • Compressed sensing
  • Magnetic resonance imaging
  • Uniform discrete curvelet transform
  • Dictionary learning
  • Augmented Lagrangian