Abstract
This article focuses on exploiting superconvergence to obtain more accurate multi-resolution analysis. Specifically, we concentrate on enhancing the quality of passing of information between scales by implementing the Smoothness-Increasing Accuracy-Conserving (SIAC) filtering combined with multi-wavelets. This allows for a more accurate approximation when passing information between meshes of different resolutions. Although this article presents the details of the SIAC filter using the standard discontinuous Galerkin method, these techniques are easily extendable to other types of data.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Application-driven developments of complex fluid flows at multiple scales can benefit from deep understanding of the mathematical aspects of multi-resolution analysis (MRA) combined with existing knowledge of accuracy enhancement. This allows for more accurate approximations when passing information between meshes of different resolutions. In this article, we combine a natural technique for simulations at multiple scales, that of multi-wavelet MRA, and that of accuracy enhancing post-processing. Specifically, the techniques introduced in this article reduce the error in the approximation when moving information between coarse and fine grids (and vice versa). This is accomplished by exploiting the underlying properties of the numerical scheme such as the superconvergence paired with the multi-wavelet/MRA. This ability to reduce the errors when moving information from coarse data to finer data is demonstrated in Fig. 1. In this figure, the \(L^2\) and \(L^\infty\) errors are presented in log-log scale for six different meshes. Specifically, the initial information is given on a mesh consisting of twenty elements, and the mesh is refined five times so that the final mesh consists of 640 elements. Errors are given for a simple \(L^2\) projection onto a finer grid as well as constructing an enhanced reconstruction before mesh refinement.
The MRA is a useful technique for expressing the discontinuous Galerkin (DG) approximation. It is based on a series of papers by Alpert, Belykin, and collaborators [1, 9], and was introduced in the DG framework in [1]. It has since been used by Cheng, Gao, and collaborators for sparse representations [7, 15, 25, 26, 31], by Müller and collaborators for adaptivity [3, 6, 13], and by Vuik and Ryan for discontinuity detection [29, 30]. It is based on the idea that a DG approximation over a mesh consisting of 2N elements, denoted \(u_h^{2N}(x,t)\), and be written as a DG approximation over a mesh consisting of N elements, denoted \(u_{2h}^N(x,t)\), together with the multi-wavelet information. Mathematically, the expression is the following:
Here, \(\psi ^N_{k,j}(x)\) are called multi-wavelets and are generally taken to be those defined by Alpert [1]. They are piecewise polynomials of degree p, defined on an element. \(d_{k,j}^N\) are the multi-wavelet coefficients that can be thought of as the detail coefficients between a mesh consisting of N elements and a mesh consisting of 2N elements, whereas the approximation on a mesh consisting of N elements is the average.
We can further improve the reconstruction by using the information from neighboring cells combined with the idea of superconvergence. This produces an alternative set of multi-wavelets that creates an improved reconstruction and reduces the error when passing the approximation from a coarse grid to a finer grid. It combines accuracy extraction techniques with the MRA. This combination creates an alternative set of multi-wavelets that allows for error reduction while allowing information to be passed from a fine scale to a coarse scale and vice versa.
The property of the superconvergence for Galerkin methods means that the solution is converging faster than the “optimal” convergence rate, which is tied to more accurate propagation of waves and allows for the physically relevant eigenvalue to dominate in long time simulations [8]. For the DG, the optimal rate is \(p+1\). In [4], it was shown that a specially designed post-processor can extract the higher order information of the dispersion and dissipation error and create an approximation that converges with \(2p+1\) order accuracy. This post-processing filtering has been further developed as the Smoothness-Increasing Accuracy-Conserving (SIAC) filtering [10,11,12, 14, 18,19,20, 22, 24]. The ability to extract the higher order information will be exploited to allow for accurate transfer of information between scales.
The SIAC kernel has mostly been developed for DG methods. In [20], it was demonstrated that the approximation properties of the SIAC procedure actually lend themselves to describing a family of filtering kernels. Specifically, the smoothness increasing the property of the SIAC filtering is a direct result of using the B-spline filtering, the accuracy preserving property (i.e., superconvergence) of the SIAC filtering is attained through choosing a specific number of B-splines and imposing a polynomial reproduction constraint. The authors in [20] also studied variations of the symmetric SIAC kernel that still attain superconvergence properties while having different computational costs and smoothness properties. From the application point of view, the results can help practitioners design new kernels of interest with different design criteria.
It should be noted that combining these ideas leads to a variant of the Nyström reconstruction [21]. The Nyström reconstruction provides for more accurate reconstruction using convolution. Here, we present these ideas for DG approximations, but expect that they can be translated to other areas as well. The outline of this paper is the following. In Sect. 2, we present the DG approximation as well as how it can be written in terms of a multi-wavelet MRA. We also discuss the SIAC kernel that will be used to generate an alternative set of wavelets. In Sect. 3, we show how to combine these ideas to generate an alternative set of multi-wavelets and multi-wavelet coefficients. Finally, numerical examples are presented in Sect. 4.
2 Background
In this section, the discussion will be limited to the important components that serve as the mathematical formulation of the DG method, the MRA, and a specific filtering that both extracts accuracy and maintains accurate wave propagation properties. That is, a discussion of the SIAC filtering.
2.1 MRA
To illustrate the usefulness of the DG-MRA, an example of translating from a coarse scale to a finer scale is given in Fig. 2. In this figure, the approximation space for the coarsest mesh (level 0) consisting of one element is given in the top left. The approximation space for a mesh consisting of two elements (level 1) is given in the top right. We can relate these two approximations spaces by adding the multi-wavelet information where the multi-wavelets are given on the bottom row. In other words, the DG approximation space on level 1 can be rewritten as a direct sum of the scaling function basis of the coarse level 0 together with multi-wavelets. The coefficients multiplying the multi-wavelets contain the details needed to go from level 0 to level 1. Hence,
where \(d_{\ell j}^{0}\) are the detail coefficients.
The DG-MRA can be generalized for multiple levels. Here, \(\psi _{\ell j}^m(x)\) are multi-wavelets that allow for translation from one level to another. This means that by subtracting details, the approximation on a coarser level can be obtained. This knowledge has already been established for the DG approximation basis itself [29, 30].
To provide more details of how the MRA works in the general case, consider a DG approximation for a given equation. To construct such an approximation, it is necessary to first break the domain up into N elements. In one dimension, denote elements as \(I_j,\, j=1,\cdots ,N\) such that the domain can be written as \(\Omega = \mathop \cup \limits _{j=1}^N\, I_j\). On each element, the DG approximation is formed of piecewise polynomials of degree less than or equal to p, denoted \(\mathbb {P}^p(I_j).\) The entire DG approximation space is then given by
where h denotes the element size. However, this is directly related to a multi-resolution approximation. In Fig. 3, the DG space is given at level n with the element size \(h=\frac{|\Omega |}{N}\) and \(N=2^n\). The approximation itself is represented as
where \(\phi _{\ell j}^n(x) \in V_h^p\) are the basis functions in the approximation. The DG basis functions are also the scaling functions of the MRA on level n. The finer details are given by the multi-wavelet space, which is defined as the orthogonal complement of the scaling function space on a given level. For example, the DG approximation space on level n, denoted \(V_h^p\), can be rewritten as a direct sum of the scaling function basis on level \(n-1\) (with the element size 2h) and its orthogonal complement:
The coefficients multiplying the wavelets in the space \(W_{2h}^p\) contain the details needed to go from level \(n-1\) to level n. Hence,
This means that by subtracting details, the approximation on a coarser level can be obtained. For example, to get from level n to level 0, the coarsest level, the DG approximation at level 0 can be written as
where \(d_{\ell j}^m\) are the detail coefficients and \(\psi _{\ell j}^m(x)\) are multi-wavelets that allow for translation from one level to another. A diagram of the different multi-resolution levels and their scaling function bases is given in Fig. 3. The extra information provided by the multi-wavelets allows for translation from one level to another.
2.2 SIAC Filtering
The symmetric SIAC kernel was originally introduced in [2, 4]. It was extended to include the superconvergent derivative information in [22, 27]. The utility for post-processing near the boundaries or shocks was addressed through one-sided SIAC kernels in [23, 28]. Recent investigations into nonlinear equations show that it may be possible to improve on the filtered solution by modifying the kernel to account for the reduced regularity assumptions [12, 16, 17]. Traditionally, introducing these filters in engineering applications can be challenging since a tensor product filter grows in support size as the field dimension increases, becoming computationally expensive. Recently, this was overcome with the introduction of the line SIAC filter [5]. This one-dimensional filter for multi-dimensional data preserves the properties of traditional tensor product filtering, including smoothness and improvement in the convergence rate, at a greatly reduced computational cost.
The SIAC filter uses the continuous convolution
where \(u_h\) denotes the approximation at final time and the kernel (symmetric) is a linear combination of central B-splines,
Here, \(\gamma\) denotes the B-splines centers and \(c_\gamma\) their weights. The kernel subindex H in (8) acts as a scaling factor, i.e., \(K_H(x-y)=\frac{1}{H} K\left( \frac{x-y}{H}\right) =\frac{1}{H}K(\eta ).\) To give an idea of the filter size, for uniform meshes, the usual scaling choice is \(H=h\), where h denotes the element size. The superindexes \((r+1,\ell )\) indicate the number of B-splines employed to build the kernel \((r+1)\) and the spline order (\(\ell\)). The convolution kernel then has a support of size \((2r+\ell )H\) around the point being post-processed. For typical DG solutions, the superconvergent extracting SIAC kernel employs \(r+1=2p+1\) B-Splines of order \(\ell =p+1\) (p denotes the highest polynomial degree employed for the numerical approximation). In multiple dimensions, computational costs can be saved by applying this filter as a line integral [5].
In this article, we take \(2p+1\) B-splines of order \(\ell =1.\) Using higher order splines will lead to over-smoothing, a decrease in accuracy, and a decrease in accuracy with refinement. In this instance, the first three kernels are
These kernels will be utilized to obtain more accurate approximations when refining coarse mesh information.
3 SIAC-MRA
In this section, we introduce an enhanced reconstruction using the SIAC-MRA. This provides a basis of alternative multi-wavelets as well as alternative multi-wavelet coefficients. To do so, we introduce the following notation:
with \(\xi _j = \frac{2}{\Delta x_j}(x-x_j)\) and
This is the approximation space of the post-processed solution. There is a right counterpart to these functions defined as
However, this expression can be further simplified if the basis functions are taken to be the orthonormal Legendre polynomials, \(\phi ^{(k)}(\xi ) = \sqrt{k+\frac{1}{2}}P^{k}(\xi )\) in \([-1,1]\). In this case, we can write \(\mathcal {X}_{\rm R}^{(k)}(\eta ) = - \mathcal {X}_{\rm L}^{(k)}(\eta ) = -\frac{\sqrt{k+1/2}}{2(2k+1)}\left( P^{(k+1)}(\eta )-P^{(k-1)}(\eta )\right)\) when \(k>0\). For \(k=0\), we have \(\mathcal {X}_{\rm L}^{(0)}(\eta ) = \frac{1}{2\sqrt{2}}(\eta +1)\) and \(\mathcal {X}_{\rm R}^{(0)}(\eta ) = \frac{1}{2\sqrt{2}}(1-\eta )\) which means that \(\mathcal {X}_{\rm R}^{(0)}(\eta ) = -\mathcal {X}_{\rm L}^{(0)}(\eta )+\sqrt{\frac{1}{2}}\). Plots of these functions are given in Fig. 4 (left).
To describe the procedure of the enhanced reconstruction, we begin with the DG approximation represented by orthonormal Legendre polynomials:
on a mesh consisting of N elements. Next, using a post-processor of \(2p+1\) B-Splines of order one we obtain the post-processed solution:
where \(\bar{x} = x_j + \frac{\Delta x}{2}\zeta\). The above is valid when \(\zeta \in (-1,0)\). Here,
Similarly, if \(\zeta \in (0,1)\), the post-processed solution can be written as
where
We then need to obtain the approximation on the finer mesh, which can be expressed in terms of the following approximation space:
This gives the coefficients for the refined approximation as
The inner products used above are defined in the following way:
Hence, an alternative multi-wavelet coefficient expression of
can be obtained, where \(\mathcal {P}_{h/2}\) denotes the projection onto the piecewise polynomial approximation space (15) consisting of 2N elements, and \(\widetilde{d}_j^{(k)}\) are new multi-wavelets coefficients that may be inserted into (6).
Remark 1
(Errors) The enhanced reconstruction using the SIAC is of order \(p+2\). This allows for a reduced error constant when refining the mesh.
4 Numerical Results
In this section, we present the effectiveness of the enhanced reconstruction for a series of test problems involving both projections and time-evolution equations. We begin with an application to an analytical function and move on to a discontinuous function. Finally, we compare the time evolution of the usual MRA solution against the enhanced reconstruction.
4.1 Analytic Function
In this section, we consider
The function is first projected onto a mesh consisting of twenty elements using a piecewise orthonormal Legendre basis. This approximation is then further projected onto a series of five finer meshes, where the number of elements are doubled for every refinement. The projection using the standard orthonormal Legendre polynomials is compared with enhanced reconstructions that exploit the SIAC information. We compare using the enhanced reconstruction for every refinement versus the enhanced reconstruction only for the first refinement. The log-log plots of the errors for \(\mathbb {P}^1-\mathbb {P}^4\) are given in Fig. 5 and pointwise errors in Fig. 6. The corresponding error tables are given in Table 1. Figure 5 and Table 1 show that even if the enhanced reconstruction is used just for the first refinement, it is beneficial. The errors can be further reduced using the enhanced reconstruction for each refinement thereafter. In other words, the magnitude of the error decreases with the first refinement, and the errors continue to decrease if enhancement is done with each refinement. This is because the affect of the SIAC is the greatest on the original approximation. It should be noted that using the SIAC and then refining always produces better errors than refinement by itself.
4.2 Two-Dimensional Analytic Function
We repeat the experiment in Sect. 4.1 for two dimensions, that is, we consider the two-dimensional function
The log-log plots of the errors are given in Fig. 7 and the corresponding error table is given in Table 2. Similar to the one-dimensional case, we find that using the enhanced reconstruction is beneficial, even if it is only performed for the first refinement.
4.3 Discontinuous Function
In this section, we consider a simple box function,
Note that the projections can achieve machine zero provided the discontinuities are at domain boundaries. For illustrative purposes, we only consider \(p=1, 2\). Plots of the function using a simple projection as well as enhanced reconstructions are given in Fig. 8. The log-log plots of N vs. error are given in Fig. 9. We see that it is still beneficial to use the enhanced reconstruction for the first refinement. However, continually using the enhanced reconstruction, while still better, may not be beneficial. The results can be seen in Table 3 and Fig. 10. The errors are calculated away from the discontinuities.
4.4 Time Evolution
In this example, we consider the linear transport equation
The initial approximation is done for twenty elements and the errors are calculated after four periods in time (\(T=8\)). The same calculation is performed where the initialization uses the \(L^2\) projection onto twenty elements and then is refined either with or without the enhanced reconstruction. The log-log plots of the errors for \(p=1\), 2, 3, 4 is given in Fig.11. After the initial refinement, a better approximation is obtained using an enhanced reconstruction. This matches the corresponding \(L^2\) and \(L^\infty\) errors given in Table 4. Pointwise errors for \(p=3,\, 4\) are given in Fig. 12. The errors for the enhanced reconstruction have fewer oscillations.
5 Conclusions
In this article, we demonstrated that it is possible to construct a more accurate MRA that exploits the multi-scale structure of the underlying numerical approximation while taking advantage of the hidden accuracy of the approximation. This allows for reducing errors when passing numerical information from coarse scales to fine scales (and vice versa). We specifically utilized the SIAC filtering combined with multi-wavelets. Although this article presents the details of the SIAC filtering using the standard DG method, these techniques are easily extendable to other types of data.
References
Alpert, B.K.: A class of bases in \({L}^2\) for the sparse representation of integral operators. SIAM J. Math. Anal. 24, 246–262 (1993)
Bramble, J.H., Schatz, A.H.: Higher order local accuracy by averaging in the finite element method. Math. Comput. 31, 94–111 (1977)
Caviedes-Voulliéme, D., Gerhard, N., Sikstel, A., Müller, S.: Multiwavelet-based mesh adaptivity with discontinuous Galerkin schemes: exploring 2d shallow water problems. Adv. Water Resour. 138, 103559 (2020)
Cockburn, B., Luskin, M., Shu, C.-W., Suli, E.: Enhanced accuracy by post-processing for finite element methods for hyperbolic equations. Math. Comput. 72, 577–606 (2003)
Docampo-Sánchez, J., Ryan, J.K., Mirzargar, M., Kirby, R.M.: Multi-dimensional filtering: reducing the dimension through rotation. SIAM J. Sci. Comput. 39, A2179–A2200 (2017)
Gerhard, N., Iacono, F., May, G., Müller, R.S.S.: A high-order discontinuous Galerkin discretization with multiwavelet-based grid adaptation for compressible flows. J. Sci. Comput. 62, 25–52 (2015)
Guo, W., Cheng, Y.: A sparse grid discontinuous Galerkin method for high-dimensional transport equations and its application to kinetic simulations. SIAM J. Sci. Comput. 38, A3381–A3409 (2016)
Guo, W., Zhong, X., Qiu, J.-M.: Superconvergence of discontinuous Galerkin and local discontinuous Galerkin methods: eigen-structure analysis based on Fourier approach. J. Comput. Phys. 235, 458–485 (2013)
Hovhannisyan, N., Müller, S., Schäfer, R.: Adaptive Multiresolution Discontinuous Galerkin Schemes for Conservation Laws. Report 311, Institut für Geometrie und Praktische Mathematik, Aachen (2010). http://www.igpm.rwth-aachen.de/en/reports2010
Ji, L., van Slingerland, P., Ryan, J.K., Vuik, K.: Superconvergent error estimates for position-dependent smoothness-increasing accuracy-conserving post-processing of discontinuous Galerkin solutions. Math. Comput. 83, 2239–2262 (2014)
Ji, L., Yan, X., Ryan, J.K.: Accuracy enhancement for the linear convection-diffusion equation in multiple dimensions. Math. Comput. 81, 1929–1950 (2012)
Ji, L., Yan, X., Ryan, J.K.: Negative-order norm estimates for nonlinear hyperbolic conservation laws. J. Sci. Comput. 54, 269–310 (2013)
Kesserwani, G., Caviedes-Voulliéme, D., Gerhard, N., Müller, S.: Multiwavelet discontinuous Galerkin h-adaptive shallow water model. Comput. Methods Appl. Mech. Eng. 294, 56–71 (2015)
King, J., Mirzaee, H., Ryan, J.K., Kirby, R.M.: Smoothness-Increasing Accuracy-Conserving (SIAC) filtering for discontinuous Galerkin solutions: improved errors versus higher-order accuracy. J. Sci. Comput. 53, 129–149 (2012)
Liu, Y., Cheng, Y., Chen, S., Zhang, Y.-T.: Krylov implicit integration factor discontinuous Galerkin methods on sparse grids for high dimensional reaction-diffusion equations. J. Comput. Phys. 388, 90–102 (2019)
Meng, X., Ryan, J.K.: Discontinuous Galerkin methods for nonlinear scalar hyperbolic conservation laws: divided difference estimates and accuracy enhancement. Numer. Math. 136, 27–73 (2017)
Meng, X., Ryan, J.K.: Divided difference estimates and accuracy enhancement of discontinuous Galerkin methods for nonlinear symmetric systems of hyperbolic conservation laws. IMA J. Numer. Anal. 38, 125–155 (2018)
Mirzaee, H., Ji, L., Ryan, J.K., Kirby, R.M.: Smoothness-Increasing Accuracy-Conserving (SIAC) post-processing for discontinuous Galerkin solutions over structured triangular meshes. SIAM J. Numer. Anal. 49, 1899–1920 (2011)
Mirzaee, H., King, J., Ryan, J.K., Kirby, R.M.: Smoothness-Increasing Accuracy-Conserving (SIAC) filters for discontinuous Galerkin solutions over unstructured triangular meshes. SIAM J. Sci. Comput. 35, A212–A230 (2013)
Mirzargar, M., Ryan, J.K., Kirby, R.M.: Smoothness-Increasing Accuracy-Conserving (SIAC) filtering and quasi-interpolation: a unified view. J. Sci. Comput. 67, 237–261 (2016)
Nyström, E.J.: Über die praktische auflösung von integralgleichungen mit anwendungen auf randwertaufgaben. Acta Math. 54, 185–204 (1930)
Ryan, J.K., Cockburn, B.: Local derivative post-processing for the discontinuous Galerkin method. J. Comput. Phys. 228, 8642–8664 (2009)
Ryan, J.K., Shu, C.-W.: One-sided post-processing for the discontinuous Galerkin method. Methods Appl. Anal. 10, 295–307 (2003)
Ryan, J.K.: Exploiting superconvergence through Smoothness-Increasing Accuracy-Conserving (SIAC) filtering. In: Kirby, R., Berzins, M., Hesthaven, J. (eds.) Spectral and High Order Methods for Partial Differential Equations ICOSAHOM 2014. Lecture Notes in Computational Science and Engineering, vol. 106, pp. 87–102. Springer, Cham (2015)
Tao, Z., Chenand, A., Zhang, M., Cheng, Y.: Sparse grid central discontinuous Galerkin method for linear hyperbolic systems in high dimensions. SIAM J. Sci. Comput. 41, A1626–A1651 (2019)
Tao, Z., Guo, W., Cheng, Y.: Sparse grid discontinuous Galerkin methods for the Vlasov-Maxwell system. J. Comput. Phys. 3, 100022 (2019)
Thomée, V.: High order local approximations to derivatives in the finite element method. Math. Comput. 31, 652–660 (1977)
Van Slingerland, P., Ryan, J.K., Vuik, K.: Position-dependent Smoothness-Increasing Accuracy-Conserving (SIAC) filtering for accuracy for improving discontinuous Galerkin solutions. SIAM J. Sci. Comput. 33, 802–825 (2011)
Vuik, M.J., Ryan, J.K.: Multiwavelet troubled-cell indicator for discontinuity detection of discontinuous Galerkin schemes. J. Comput. Phys. 270, 138–160 (2014)
Vuik, M.J., Ryan, J.K.: Multiwavelets and jumps in DG approximations. In: Kirby, R., Berzins, M., Hesthaven, J. (eds.) Spectral and High Order Methods for Partial Differential Equations ICOSAHOM 2014. Lecture Notes in Computational Science and Engineering, vol. 106, pp. 503–511. Springer, Cham (2015)
Wang, Z., Tang, Q., Guo, W., Cheng, Y.: Sparse grid discontinuous Galerkin methods for high-dimensional elliptic equations. J. Comput. Phys. 314, 244–263 (2016)
Acknowledgements
This work was motivated by discussions with Dr. Venke Sankaran (Edwards Air Force Research Lab, USA) and was performed while visiting the Applied Mathematics group at Heinrich-Heine University, Düsseldorf, Germany. Research supported by the Air Force Office of Scientific Research (AFOSR) Computational Mathematics Program (Program Manager: Dr. Fariba Fahroo) under Grant numbers FA9550-18-1-0486 and FA9550-19-S-0003.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interest
On behalf of all authors, the corresponding author states that there is no conflict of interest.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ryan, J.K. Capitalizing on Superconvergence for More Accurate Multi-Resolution Discontinuous Galerkin Methods. Commun. Appl. Math. Comput. 4, 417–436 (2022). https://doi.org/10.1007/s42967-021-00121-w
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s42967-021-00121-w
Keywords
- Multi-resolution analysis
- Multi-wavelets
- Discontinuous Galerkin
- Smoothness-Increasing Accuracy-Conserving (SIAC)
- Post-processing
- Superconvergence
- Accuracy enhancement