1 Introduction

Acquired MR signals lie in a frequency-phase space known as k-space, and MR images are obtained via the inverse Fourier transform (FT) of a finite k-space acquisition. Because of scan time and SNR limitations, the outer parts of the k-space, containing the high-frequency information of the image, are generally not recorded. Therefore, all MR images are affected by Gibbs ringing, a well-known artefact consisting of spurious oscillations in the proximity of sharp image gradients at tissue boundaries. These artifacts often occur at high-contrast interfaces since the partial k-space sampling cuts off the high-frequency components in these zones. Such artifacts degrade the quality of the images and make quantitative measurement difficult, hampering clinical diagnosis. Although Gibbs ringing is a feature of all MR images, it becomes increasingly significant for low resolution or quantitative modalities such as diffusion MRI and when it is necessary to increase the resolution of MR images to facilitate the doctors’ diagnosis. In these cases, the MR images are reconstructed on a finer grid by using zero-filling (ZF), which sets the missing high-frequency components to zero, causing severe ringing artifacts.

Moreover, in many MR imaging applications, a short scan time is required to observe dynamic processes such as the beating heart, the passage of a contrast bolus guidance of an interventional procedure, or to reduce artifacts from physiological motion [1,2,3,4]. In such applications, only a portion of the k-space is acquired, and the missing data are usually simply replaced by zeros. Despite the intensive research into acquisition and reconstruction algorithms, the ZF method is still most commonly used by clinical doctors since, although presenting truncation artifacts, it is stable and fast.

For all these reasons, a significant quantity of MR images containing truncation artifacts is archived and re-exploited for the purpose of diagnosis, treatment evaluation and disease monitoring. Therefore, the development of effective post-processing approaches to remove truncation artifacts from these acquired MR images are highly desirable. In particular, it is desirable to remove as much as possible the ringing artifacts without adding excessive blur to the images.

There is a considerable amount of literature on the post-processing of MR images, which can be summarized as follows. One approach consists in recovering the missing information by data extrapolation, see [5] and reference therein. Variational approaches have been proposed based on the Total Variation regularization [2,3,4, 6, 7], the \(L_1\) norm of the shearlet transform [8, 9] and the weighted Total Variation [10, 11]. Finally, [12] proposes a method based on space domain filters presenting a Matlab software tool which is often used in clinical practice.

In the present paper, we model the Gibbs artefact removal as a constrained variational problem balancing data discrepancy and sparsity-promoting regularization functions such as Total Variation (TV), Total Generalized Variation (TGV) and \(L_1\) norm of Wavelet transform (L1W). The data discrepancy can be represented by the absolute distance between the data and the unknown enhanced image (denoising model), or it can reproduce the truncation in the k-space as a convolution of the underlying image with a sinc-function.

Concerning the regularization functions, there are several papers in the literature about the Total Variation (TV) and second-order Total Variation (Total Generalized Variation) applied to the denoising model [4, 6]. Moreover, the Wavelet transform has proven to be effective in noise removal [13] and regularization based on wavelet transform is investigated in the compressed sensing context (see [14] and references therein).

To the best of our knowledge, in the context of Gibbs artefact removal, a thorough investigation of TV, TGV and L1W for the convolution model is not present in the literature. For this reason, we focus our analysis on the denoising and convolutive models with such regularization functions. The variational problems are solved by adapting the Alternate Directions Method of Multipliers (ADMM) to the specific objective functions.

This work aims to assess the efficacy of the proposed variational models by running an extended set of numerical experiments in which the ringing artifacts are introduced on synthetic and real data on high resolution acquired brain images.

The structure of the paper is the following: Sect. 2 discusses in detail the models and their numerical implementation; Sect. 3 presents the most significant numerical results. Finally, Sect. 4 presents some conclusions.

2 Models and methods

The Gibbs ringing artifacts arise because the Fourier series cannot represent a discontinuity within a finite number of terms. The ringing effect is especially visible for acquisitions at low resolution around sharp edges. Low pass filtering techniques reduce this artefact at the expense of introducing image blur. For this reason, edge preserving variational filtering techniques are studied. One approach [2, 3] is to compute the enhanced image \(x_e\) by solving the following denoising problem:

$$\begin{aligned} x_e = \arg \min _{x \ge 0} \left\{ \frac{\lambda }{2}\Vert x-y \Vert _2^2 + {\mathcal {R}}(x) \right\} \end{aligned}$$
(1)

where \(y\in {\mathbb {R}}^n\), \(x\in {\mathbb {R}}^n \), \(n=n_xn_y\), respectively are the data and unknown images of size \(n_x \times n_y\), and \({\mathcal {R}}(x)\) represents an edge-preserving penalty function.

A sharp cut-off or truncation in the k-space is equivalent to a convolution in the spatial domain with a sinc function. In the discrete FT reconstruction of MR images, the cut-off frequency equals the frequency of the sinc function that is convolved with the image. The oscillating lobes of the sinc function result in the ringing pattern around sharp edges. Therefore, an enhanced image \(x_c\) can be obtained by the following convolutive model:

$$\begin{aligned} x_c = \arg \min _{x \ge 0} \left\{ \frac{\lambda }{2}\Vert sinc *x-y \Vert _2^2 + {\mathcal {R}}(x) \right\} . \end{aligned}$$
(2)

Concerning the regularization functions, we focus our analysis on models (1) and (2) with the following regularization functions:

  • Total Variation function [15], defined in discrete form as:

    $$\begin{aligned} {\mathcal {R}}(x)= TV(x) \equiv \Vert \nabla x \Vert _1, \end{aligned}$$

    where \(\nabla \in {\mathbb {R}}^{2n \times n}\) represents the discrete gradient operator.

  • Second order Total Generalized Variation (TGV) function [16]:

    $$\begin{aligned} {\mathcal {R}}(x)= TGV(x) \equiv \min \limits _{w\in {\mathbb {R}}^{2n}} \displaystyle {\alpha _0 \left\| \nabla x - w\right\| _1} + {\alpha _1 \left\| Ew\right\| _1}, \quad \alpha _0,\alpha _1\in (0,\,1) \end{aligned}$$

    where \(E \in {\mathbb {R}}^{4n \times 2n} \) represents the symmetrized derivative operator.

  • \(L_1\) norm of the discrete wavelet transform W [17]:

    $$\begin{aligned} {\mathcal {R}}(x)=\Vert Wx \Vert _1. \end{aligned}$$

where \(W\in {\mathbb {R}}^{n \times n}\) represents an orthogonal wavelet transform.

We solve the optimization problems (1) and (2) by the ADMM method by reformulating them as separable convex optimization problems with two blocks. Firstly, for the reader’s convenience, let us recall the classical two-blocks ADMM method for the solution of the following problem

$$\begin{aligned} \begin{array}{ll} \min \limits _{x,z} &{} \displaystyle f_1(x) + f_2(z) \\ \hbox {s.t.}&{} A_1x + A_2z = b, \\ &{} x\in \Omega _1, \, z\in \Omega _2 \end{array} \end{aligned}$$
(3)

where \(A_i\in {\mathbb {R}}^{p\times n_i}\) (\(i=1,2\)) are given linear operators, \(b\in {\mathbb {R}}^p\) is a given vector and \(\Omega _i\subset {\mathbb {R}}^{n_1}\) (\(i=1,2\)) are closed convex sets. The functions \(f_i:{\mathbb {R}}^{n_i}\rightarrow {\mathbb {R}}\) (\(i=1,2\)) are convex functions on \(\Omega _i\), respectively. The augmented Lagrangian function of problem (3) has the form

$$\begin{aligned} {\mathcal {L}}_\rho (x,z,\mu ) = f_1(x) + f_2(z) + \mu ^T( A_1x + A_2z -b)+\frac{\rho }{2}\Vert A_1x + A_2z -b\Vert ^2 \end{aligned}$$

where \(\mu \in {\mathbb {R}}^p\) is a vector of Lagrange multipliers and \(\rho >0\) is a penalty parameter. Given a chosen initial point \((x^{(0)},z^{(0)},\mu ^{(0)})\in {\mathbb {R}}^{n_1+n_2+p}\), the ADMM method consists of the iteration

$$\begin{aligned} x^{(k+1)}= & {} \displaystyle \arg \min \limits _{x\in \Omega _1} {\mathcal {L}}_\rho (x,z^{(k)},\mu ^{(k)}),\nonumber \\ z^{(k+1)}= & {} \displaystyle \arg \min \limits _{z\in \Omega _2} {\mathcal {L}}_\rho (x^{(k+1)},z,\mu ^{(k)}),\nonumber \\ \mu ^{(k+1)}= & {} \displaystyle \mu ^{(k)} + \rho \left( A_1x^{(k+1)} + A_2z^{(k+1)} -b\right) . \end{aligned}$$
(4)

As far as TV and L1W regularization is concerned, we can reformulate problems (1) and (2) as follows

$$\begin{aligned} \begin{array}{ll} \min \limits _{x,z} &{} \displaystyle \frac{\lambda }{2} \Vert Hx-y\Vert ^2 + \Vert z\Vert _1 \\ \hbox {s.t.}&{} Lx - z = 0, \\ &{} x\ge 0 \end{array} \end{aligned}$$
(5)

where H is the identity matrix for model (1) or it derives from the discretization of the convolution product with the sinc finction for model (2), and \(L=\nabla \) or \(L=W\) for TV and L1W regularization, respectively. Clearly, problem (5) is a separable problem of the form (3) with \(f_1(x) = \frac{\lambda }{2} \Vert Hx-y\Vert ^2\), \(f_2(z)=\Vert z\Vert _1\), \(A_1=L\), \(A_2=-I\), \(b=0\), \(\Omega _1=\{x \ : \ x\ge 0 \}\); we can now apply the ADMM method to problem (5). In iterations (4), the subproblem in x is a bound constrained quadratic programming problem; we solve it by the variable projection (VP) method introduced in [18], a gradient projection method using a limited minimization rule as linesearch [19] and the Barzilai and Borwei rule [20] for steplength selection. The subproblem in z can be rewritten as

$$\begin{aligned} \begin{array}{ll} \min \limits _{z}&\displaystyle \Vert z\Vert _1 + \frac{\rho }{2}\Vert A_1 x^{(k+1)}+A_2 z -b +\frac{\mu ^{(k)}}{\rho }\Vert ^2 . \end{array} \end{aligned}$$
(6)

This subproblem is separable and can be solved exactly by computing the proximal operator of \(\Vert z\Vert _1\).

For the solution of problems (1) and (2) with TGV regularization, we use a two-blocks ADMM method originally proposed in [21] for the restoration of directional images under Poisson noise. To this end, we reformulate the problems as follows:

$$\begin{aligned} \begin{array}{ll} \min \limits _{x,w,z_1,z_2,z_3,z_4} &{} \displaystyle \frac{\lambda }{2}\Vert z_1-y\Vert ^2 + \alpha _0\,\Vert z_2\Vert _1 + \alpha _1\,\Vert z_3\Vert _1 + \chi _{{\mathbb {R}}_+^n}(z_4) \\ \hbox {s.t.}&{} z_1 = H x,\\ &{} z_2 = \nabla x - w,\\ &{} z_3 = E w,\\ &{} z_4 = x \end{array} \end{aligned}$$
(7)

where \(z_1\in {\mathbb {R}}^n\), \(z_2\in {\mathbb {R}}^{2n}\), \(z_3\in {\mathbb {R}}^{4n}\), \(z_4\in {\mathbb {R}}^{n}\), and \(\chi _{{\mathbb {R}}_+^n}(\cdot )\) is the characteristic function of the nonnegative orthant in \({\mathbb {R}}^n\). By introducing the auxiliary variables \({\tilde{x}} = ( x, w )\) and \(z = ( z_1, z_2, z_3, z_4 )\) we can reformulate problem (7) as a problem of the form (3) where we define the two separable functions as

$$\begin{aligned} f_1({\tilde{x}})= 0,\quad f_2(z) = \frac{\lambda }{2}\Vert z_1-y\Vert ^2 + \alpha _0\,\Vert z_2\Vert _1 + (1-\alpha )\,\Vert z_3\Vert _1 + \chi _{{\mathbb {R}}_+^n}(z_4), \end{aligned}$$
(8)

and the matrices \(A_1\in {\mathbb {R}}^{8n\times 3n}\) and \(A_2\in {\mathbb {R}}^{8n\times 8n}\) as

$$\begin{aligned} A_1 = \left[ \begin{array}{rr} H &{} 0 \\ \nabla &{} -I_{2n} \\ 0 &{} E \\ I_n &{} 0 \end{array} \right] , \quad A_2= \left[ \begin{array}{rrrr} {-I_n} &{} 0 &{} 0 &{} 0 \\ 0 &{} -I_{2n} &{} 0 &{} 0 \\ 0 &{} 0 &{} -I_{4n} &{} 0 \\ 0 &{} 0 &{} 0 &{} -I_n\end{array} \right] . \end{aligned}$$
(9)

The ADMM iterations can now be applied to problem (7); we underline that both the \({\tilde{x}}\) and z subproblems can be solved exactly at a low cost and we refer the reader to [21] for a deeper discussion on the solution of the ADMM subproblems.

The ADMM method for solving two-block convex minimization problems has been studied extensively in the literature and its global convergence has been proved [22, 23] when the two subproblems are solved exactly. This is the case of our version of ADMM specialized for TGV regularization. In our implementation of ADMM for TV and L1W regularization, the x subproblem is solved to a high precision, i.e., the iterations of the inner VP method have been stopped when the relative distance between two successive values of the objective function becomes less than \(10^{-6}\). We have observed practical convergence of the corresponding ADMM scheme. However, we underline that variants of inexact ADMM have been proposed which are based on different strategies for the control of the lack of exactness [24,25,26,27].

3 Numerical results

To assess the effectiveness of the proposed models and regularization functions, we developed numerical tests using synthetic and brain images acquired by MRI scanner. The numerical results have been computed using Matlab R2021a on a Intel Core i5 processor with 2.50GHZ and Windows operating system.

In our experiments, the matrix W of L1W regularization represents an orthogonal Haar wavelet transform with four levels. As regards the implementation of the ADMM method, the initial guess is chosen as the null vector, i.e., \(x^{(0)}=0\), \(z^{(0)}=0\) and \(\mu ^{(0)}=0\). Moreover, the ADMM iterations are terminated when

$$\begin{aligned} \frac{\Vert x^{(k+1)}-x^{(k) \Vert }}{\Vert x^{(k+1)} \Vert } \le 10^{-3} \end{aligned}$$

or after a maximum number of 100 iterations. The VP method for the solution of the inner x subproblems, for TV and L1W regularization, is arrested when the residual norm becomes less than \(10^{-6}\) or after 1000 iterations. The values of the TGV weights are chosen as \(\alpha _0 = \beta \) and \(\alpha _1=(1-\beta )\) with \(\beta =0.5\). Indeed we found that, for smaller values of \(\beta \), TGV is less effective in removing Gibbs artifacts while, for larger values, the restored MR images tend to be too smooth. Finally, the value of the ADMM penalty parameter is set as \(\rho = 10\). Obviously, the choice of \(\rho \) affects the ADMM convergence rate but it also influences the quality of the restored images which are too smooth, for large value of \(\rho \). The choice \(\rho =10\) gives a good tradeoff between convergence rate and solution quality.

Synthetic data

Simulated \(128\times 128\) raw MRI data have been created by using the Matlab function mriphantom.m [28]. This function analytically generates raw data from k-space coordinates along a Cartesian trajectory using the continuous Shepp and Logan head phantom function. Noisy k-space data is also used by adding Gaussian white noise of level equal to 0.025.

Figure 1 shows the exact phantom image and the image obtained via inverse FT of the simulated k-data without and with added noise. In order to better visualize the ringing artifacts in the printed figures, Fig. 1 also shows the negative images whose colormap limits are set to [-0.5,0]. Since artifacts are more evident in the phantom negative images, we will always show them in the following.

Fig. 1
figure 1

Top: Exact phantom (left) and negative (right) images. Middle: noiseless data y (left) and negative \(-y\) (right) images. Bottom: noisy data y (left) and negative \(-y\) (right) images. The negative images have scaled colorbar

To compare the models, we heuristically compute the optimal value of the regularization parameter corresponding to a particular error metric. In preliminary investigations, we considered the most commonly used error metrics [29]: Root Mean Square Error (RMSE), Improved Signal to Noise Ratio (ISNR), Structural SIMilarity index (SSIM), Feature SIMilarity index (FSIM). Repeating the same analysis for all the models and the regularization functions, we observed that the best RMSE and ISNR values are obtained for too small \(\lambda \) values, producing excessively blurred images. Conversely, both optimal SSIM and FSIM values are obtained for larger values of \(\lambda \) giving results of better visual quality; in most cases \(\lambda _\text {SSIM} < \lambda _\text {FSIM}\). As an example, we report in Fig. 2 the values of RMSE, ISNR, SSIM and FSIM computed by the TGV-sinc model for different \(\lambda \) and represent the optimal value with a red circle. We observe that RMSE, ISNR are optimal at the same \(\lambda \) while SSIM and FSIM are optimal for larger \(\lambda \) values.

Fig. 2
figure 2

Behavior of the RMSE, ISNR, SSIM and FSIM error measures vs. \(\lambda \) for the TGV-sinc model. The red circle represents the optimal values

In Fig. 3, we represent the images computed by the TGV-sinc model with the best regularization parameter for each metric (corresponding to the red circle in Fig. 2). It is evident that the RMSE and ISNR are not proper metrics to measure the quality of the computed images, while SSIM and FSIM give visually similar results. For this reason, in the remaining tests, we use the SSIM as metric to heuristically set the optimal value of the regularization parameter for each model and regularization functional.

Fig. 3
figure 3

Reconstructed images obtained by the TGV-sinc model with a value of \(\lambda \) optimal with respect to: (A) RMSE (\(\lambda =1.353\)); (B) ISNR (\(\lambda =1.353\)); (C) SSIM (\(\lambda =24.2012\)); (D) FSIM (\(\lambda =67.342\))

Table 1 Numerical results for the synthetic data

For all the considered models, in Table 1, we report the value of the regularization parameters (determined as previously explained) and the corresponding RMSE, ISNR and SSIM values; we also report the number of ADMM iterations (column iter) and computation time in seconds (column time), using both noiseless and noisy data. In all cases, the best SSIM values (highlighted in bold) are reached by TGV regularization, while the worst SSIM values are always reached by L1W regularization. Moreover, L1W regularization has the highest computational cost, as reported in the last two columns of Table 1.

For the noiseless data, Figs. 4 and 5 show the reconstructions obtained by the denoise (1) and sinc (2) models, respectively. The error images are also depicted, i.e., the absolute difference between the exact phantom image and the restored one. For better visualization of the differences, the negative values of the error images are displayed with colormap limits set to [-0.1,0]. Finally, for the noisy data, Figs. 6 and 7 display the restored and error images obtained by model (1) and (2), respectively. A visual inspection of the figures shows that TGV regularization is always more effective in removing ringing artifacts than both TV and L1W regularization. Moreover, from Figs. 4-7, we observe that the TGV-sinc model can remove ringing artifacts more effectively as compared to the TGV-denoise model, even if the SSIM values reached by the TGV-denoising model are slightly higher than those of TGV-sinc model (see Table 1 where the best values are written in bold). Hence, for synthetic data, the TGV-sinc model appears the more appropriate model for the removal of ringing artifacts.

Fig. 4
figure 4

Noiseless synthetic data: reconstructed (top) and difference (bottom) images obtained by model (1)

Fig. 5
figure 5

Noiseless synthetic data: reconstructed (top) and difference (bottom) images obtained by model (2)

Fig. 6
figure 6

Noisy synthetic data: reconstructed (top) and difference (bottom) images obtained by model (1)

Fig. 7
figure 7

Noisy synthetic data: reconstructed (top) and difference (bottom) images obtained by model (2)

Real MRI data We consider the problem of obtaining a high-resolution MR image from a low-resolution one. This process is often required for clinical purposes, posterior analysis, or postprocessing, such as registration or segmentation. Figure 8 shows the exact high resolution image of size \(256\times 256\) and the acquired low resolution image of size \(128\times 128\).

Since the denoising model (1) is generally less effective, we focus our attention on the sinc model (2) coupled with TV, L1W and TGV regularizations. Figure 9 reports the high-resolution images obtained by ZF and the sinc model with the considered regularization functionals. The red arrows highlight ringing artifacts in the ZF image; by comparing the high-resolution images, we observe that both TV and TGV regularization effectively remove the artifacts, but, as expected, TV tends to produce small flat regions in the image. On the other hand, L1W performs worse, and some artifacts are still present in the corresponding high-resolution image. Finally, Table 2 displays the numerical results confirming the superiority of TGV regularization.

To evaluate the performance of the proposed approach in the presence of noisy data, we add Gaussian white noise of level 0.1 to the low-resolution k-space data. Figure 10 shows the obtained high resolution images. It is evident that the ZF and L1W images are degraded by noise, while TV and TGV regularization can remove noise from the high-resolution image. The numerical results are reported in Table 2; they show TGV superior accuracy, as measured by the SSIM index.

Fig. 8
figure 8

Real MR image: exact high resolution (left) and low resolution (right) images of size \(256\times 256\) and \(128 \times 128\), respectively

Fig. 9
figure 9

High resolution images obtained from noiseless data by the sinc model with TV, L1W, TGV regularization and by the ZF method (from top to bottom and from left to right)

Fig. 10
figure 10

High resolution images obtained from noisy data by the sinc model with TV, L1W, TGV regularization and by the ZF method (from top to bottom and from left to right)

Table 2 Numerical results for the real MRI data

4 Conclusion

In this paper, we investigate the efficacy of the denoising and sinc convolution model in removing Gibbs artifacts using TV, TGV and L1W penalties. From observing the results, we conclude that the sinc convolution model coupled with TGV penalty is, in general, the most effective and that L1W produces the worst results in almost all cases.

Therefore, future developments will focus on the automatic setting of the TGV parameters by exploiting the Uniform Penalty principle [30, 31].