Abstract
The paper incorporates new methods of numerical linear algebra for the approximation of the biharmonic equation with potential, namely, numerical solution of the Dirichlet problem for
High-order discrete finite difference operators are presented, constructed on the basis of discrete Hermitian derivatives, and the associated Discrete Biharmonic Operator (DBO). It is shown that the matrices associated with the discrete operator belong to a class of quasiseparable matrices of low rank matrices. The application of quasiseparable representation of rank structured matrices yields fast and stable algorithm for variable potentials c(x). Numerical examples corroborate the claim of high order accuracy of the algorithm, with optimal complexity O(N).
Similar content being viewed by others
1 Introduction
In this paper, we propose a novel algorithm for the numerical solution of the biharmonic equation in an interval, involving a potential, as follows:
Here, the function \(\phi (x)\) and the potential c(x) are given functions.
It is assumed that zero is not an eigenvalue of the operator \(\left( \frac{d}{dx}\right) ^4+c(x)\) so that (1.1) has a unique solution. That can always be achieved by a suitable constant shift of the potential.
We use a high-order finite-difference scheme, based on the Hermitian derivative and the resulting “Discrete Biharmonic Operator” (DBO) as outlined in the paper [10]. In particular, the scheme used here is not new, but the algorithm offered for the resolution of the ensuing linear system is new, leading to an optimal O(N) complexity. This is achieved by an application of the theory of quasiseparable matrices as is explained below. The implementation of the scheme leads to a system of linear algebraic equations
with \(N\times N\) tridiagonal matrices matrices A, B, C, D and a given N-dimensional column \({\mathfrak y}\). Here, N is the number of the nodes in the grid. Assuming that the matrix D is invertible, the \(2N\times 2N\) system is reduced to \(N\times N\) system
The fact that the discrete solution \(\mathfrak {u}\) converges (as \(N\rightarrow \infty \)) to the exact solution u(x) has been established in [3].
The paper [2] dealt with the case of zero potential \(c(x)\equiv 0.\) In this case, the proposed algorithm was based on the FFT transform and led to an \(O(N\log N)\) complexity algorithm. In a more general case of a nonzero potential, the matrix A is not Toeplitz and so the FFT-based algorithm does not work. This case is assumed to be treated via perturbation theory.
In the present paper, we suggest a completely different paradigm for the algebraic treatment of the system (1.3). It is based on the theory of quasiseparable representations of matrices and their inversion. It permits the incorporation of a general potential. We refer to the treatise [8] for a detailed exposition of this theory.
Here, we need to consider a rather special case of the abovementioned theory. We show that in the context of the system (1.3) with a general symmetric tridiagonal A, the matrix of the problem Z belongs to a special class of rank structured matrices. More precisely, the part below the first subdiagonal in the matrix Z (\(j\le i-2\)) is given via a product of scalar parameters that can be readily calculated.
The full matrix is given by the following formulas:
The parameters in the RHS of (1.5) are computed in a straightforward way from the given matrices \(A,\,B,\,C,\,D\,\,\) [7, Algorithm 6.1]. Indeed, this is the “heart of the algorithm” proposed here.
In fact, assuming that Z is a strongly regular matrix, the parameters in (1.5) are readily implemented in expressing Z as a classical “Left\(\cdot \) Diag\(\cdot \)Right” product
where \(\Gamma \) is an invertible diagonal matrix and L is a lower triangular matrix with zero diagonal. Based on this factorization, the solution to (1.4) follows a well-known procedure.
Overall, we obtain a linear O(N) complexity algorithm for any nonconstant potential. We present some numerical tests that corroborate the linearity and the good accuracy of our algorithm for a variety of potentials. The condition that Z is strongly regular is not necessary. In our next publications, we intend to present another inversion algorithm free of this restriction. Surely this alternative algorithm is longer and more complicated than the algorithm of the present paper.
The paper consists of thirteen sections. In Section 2, we present the formulation of the boundary value problem, describe the Hermitian finite-difference operators, and obtain the discrete analog of the basic equation. In Section 3, we recall the basic facts pertaining to the quasiseparable representation which is the foundation of the algorithm. In Section 4, we reduce the system (1.3) to matrix equation in the class of rank structured matrices. In Section 5, we present an inversion algorithm for the tridiagonal matrix D, which is the main tool in the quasiseparable frame. The proof is postponed to Appendix 1 in Section 13. Section 6 is in fact the central part of the paper. Here, we show that the part \(i>j+1\) of the matrix Z of the problem admits the representation (1.5) and derive an algorithm to compute the parameters of the representation (1.5) ( by means of quasiseparable generators) as well as the elements of the diagonal and subdiagonal of the symmetric matrix Z. These data are precisely the input for the “Left\(\cdot \) Diag\(\cdot \)Right” factorization algorithm in [7, Theorem 7.1]. The construction of (1.6) is described in Section 7, subject to the assumption that Z is a strongly regular matrix. Based on this factorization, we recall in Section 8 the classical procedure for solving (1.4) when Z is given by (1.6). In Section 9, we summarize the final complete algorithm in the form as it is used in the numerical tests. In Section 10, we consider the case of zero potential for which the Green function of the problem is explicitly available, see (10.3).
Section 11 contains the results of numerical experiments. Appendix 1 (Section 13) contains the proof of the inversion algorithm for the tridiagonal matrix D. Finally, in Appendix 2 (Section 14), we present for completeness the basic foundation of the quasiseparable representation of matrices.
NOTATION
Submatrices in this paper are indicated in MATLAB style, i.e., for a matrix A, A(m : n, t : s) selects rows m to n and columns t to s. A colon without an index range selects all the rows and columns, for instance, for an \(m\times n\) matrix B, we have \(B(:,j)=B(1:m,j)\) for any j with \(1\le j\le n\).
2 The model Dirichlet problem and the discretization
We consider the numerical solution of the biharmonic equation on an interval
subject to Dirichlet boundary conditions
Here, \(\phi (x)\) is a continuous function. By “solution of the problem” (2.1)-(2.2), we mean a continuous function u(x) on the segment [0, 1] having a fourth-order derivative \(u^{(4)}(x)\) in [0, 1].
To get the numerical solution of the problem (2.1)-(2.2), we start with the definition of the grid on the interval. In this paper, we use a uniform grid
We denote the values of the potential and the right part of (2.1) on the grid by
and form the \(N+2\)-dimensional columns \(c=\textrm{col}(c_j)_{j=0}^{N+1},\;\phi =\textrm{col}(\phi _j)_{j=0}^{N+1}\). The corresponding unknown values of the solution u(x) and its derivative \(u^{\prime }(x)\) on this grid are displayed as the columns of the size \(N+2\) as \(u^*=\textrm{col}(u(x_j))_{j=0}^{N+1}\) and \((u^*)^{\prime }=\textrm{col}(u^{\prime }(x_j))_{j=0}^{N+1}\).
In the numerical solution of the problem, we are looking for an approximation of the solution, i.e., an \(N+2\)-dimensional vector column \(\tilde{u}=\textrm{col}({\mathfrak u}_j)_{j=0}^{N+1}\), and of its derivatives, i.e., a vector \(\tilde{u}^{\prime }=\textrm{col}(({\mathfrak u}_x)_j)_{j=0}^{N+1}\). We apply here an approach suggested in the paper [4, Appendix] and the monograph [10]. In accordance with these references, the values \({\mathfrak u}_j,({\mathfrak u}_x)_j\) are determined via a system of linear algebraic equations
The boundary value problem (2.1), (2.2) is thus reduced to a problem in numerical linear algebra.
The last equalities in (2.3) surely follow from the boundary value conditions (2.2). Following (2.3), we set \(h_0=\frac{h^4}{12}\) and define the \(N\times N\) matrices
Here, \(A_0,B_0,C_0,D_0\) are tridiagonal matrices. Set
We have
Thus, we have reduced (2.3) to the computation of the N-dimensional vectors \({\mathfrak u}=\textrm{col}({\mathfrak u}_j)_{j=1}^N\) and \({\mathfrak u}_x=\textrm{col}(({\mathfrak u}_x)_j)_{j=1}^N\) as components of the solution of the system of linear algebraic equations
with the given N-dimensional vector
We set \({\mathfrak y}=\frac{h^4}{12}{\mathfrak f}\).
3 The scalar quasiseparable representations
Let \(\mathcal A\) be an \(N\times N\) matrix and let l be a nonnegative entire number. We say that the sets of numbers \(p(i)\;(i=l+1,\dots ,N-l),\;q(j)\;(j=1,\dots ,N-l),\; a(k)\;(k=2,\dots ,N-l)\) form a scalar quasiseparable representation of the part \(i-j\ge l\) of \(\mathcal {A}\) if its entries are given by the equalities
The parameters p(i), q(j), a(k) of this representation are called scalar quasiseparable generators. The part \(i-j\ge l\) admits a scalar quasiseparable representation if and only if
see [7, Section 5.2].
For instance, for a tridiagonal matrix, the strictly lower triangular part \(i-j>1\) admits a scalar quasiseparable representation with generators \(p(i)=1,\;a(k)=0,\;q(j)=\mathcal {A}(j,j+1)\). The inverse of a tridiagonal matrix is a so called Green matrix (see [7, Section 6.3]). Recall that a matrix \(\mathcal {B}\) is said to be a “lower Green of order one matrix” if the relations
hold. Hence, it follows that the lower triangular part \(\{i-j\ge 0\}\) of the matrix \(\mathcal {B}\) has a scalar quasiseparable representation. Such a representation is obtained in Section 5 for the matrix \(D^{-1}\). The scalar quasiseparable representation
obtained in Theorem 6.1 for the part \(\{i-j\ge 2\}\) of the matrix Z plays a crucial role in the design of the fast algorithm.
4 The class of matrices and equations
The system (2.8) belongs to a wider class of systems of linear algebraic equations. We consider systems of the form
with \(N\times N\) matrices A, B, C, D and the given N-dimensional column \({\mathfrak y}\).
Assume that the matrix D and the matrix \(Z=A-BD^{-1}C\) are invertible. Using the equation \(C{\mathfrak u}+D{\mathfrak u}_x=0\) and invertibility of the matrix D, we reduce the system (4.1) to the equation
Comparing with (2.8), we have
with the matrices \(A_0,B_0,C_0,D_0\) and the vector column \({\mathfrak f}\) as in (2.4), (2.5) and (2.9). Here, \(A_0,B_0,C_0,D_0\) are tridiagonal matrices. Using the formula (4.2), we get
We now assume that the matrices A, B, C, D are tridiagonal and real. Moreover, the matrices A and D are symmetric, and the matrices B and C satisfy the condition \(B=C^t\); this means that the matrix Z is real and symmetric.
Furthermore, we assume that the tridiagonal matrix D and the matrix Z are strongly regular. For a matrix \(\mathcal {A}=\{a_{ij}\}_{i,j=1}^N\) (see for instance [8, Section 1]), this means that all the leading submatrices \(\mathcal {A}_k=\{a_{ij}\}_{i,j=1}^k\) are invertible. In particular, the matrix \(\mathcal {A}\) itself is invertible. A strongly regular matrix admits the factorization
where \(\Gamma =\textrm{diag}\{\gamma _1,\gamma _2,\dots ,\gamma _N\}\) is an invertible diagonal matrix with diagonal entries
and L is a lower triangular matrix with zero diagonal. The strong regularity means that the values \(\gamma _k\) in (4.5) are well defined and nonzeros.
The condition that Z is strongly regular is not necessary. In our next publications, we intend to present another inversion algorithm free of this restriction. Surely this alternative algorithm is longer and more complicated than the algorithm of the present paper.
5 The inversion of a tridiagonal matrix
The first step of the procedure is the inversion of a tridiagonal real symmetric matrix. This problem was treated by many authors, see [6, 9, 12] and the literature cited therein. However, as far as we know, the quasiseparable representations of the inverses to tridiagonal matrices have never been considered in detail. We present here an algorithm for the computation of the quasiseparable representation of the inverse to a strongly regular tridiagonal matrix. The following lemma seems to be new. The advantages of using of the quasiseparable structure in the inversion of tridiagonal and, more general, band matrices were discussed in the recent paper [5] by Paola Boito and the second author.
Lemma 5.1
Let D be a strongly regular tridiagonal matrix with nonzero entries
Then, the values \(\gamma _k,\;k=1,\dots ,N\) are determined via recursive relations
Moreover, the matrix \(D^{-1}\) is given by the formulas
where the parameters of the representation are determined via
and
The proof is carried out in Appendix 1.
One can check easily that the matrix \(D_0\) in (2.6) strongly regular. For instance, one can prove that in this case, \(\gamma _k\ge 1/3,\;k=1,2,\dots \). Indeed, we have \(\gamma _1=2/3>1/3\), and if \(\gamma _{k-1}\ge 1/3\) then using (5.1) with \({\mathfrak d}_k=2/3,\;b_{k-1}=1/6\), we get \(\gamma _k=2/3-\frac{1}{\gamma _{k-1}}\frac{1}{30}\ge 2/3-1/12>1/3\).
6 The matrix of the problem
The material in this section is the “heart of the matter” of this paper. We derive for the matrix of the problem a representation that fits the direct application of an algorithm presented in the paper [7]. More precisely, we show that the part \(i-j\ge 2\) of the matrix Z defined in (4.2) admits a scalar quasiseparable representation and we compute the corresponding scalar quasiseparable generators, as well as diagonal and subdiagonal entries of this symmetric matrix.
Theorem 6.1
Assume that the conditions of Lemma 5.1 hold. Applying the algorithm from this lemma to the matrix D, we determine the values \(a(k)\;(k=1,\dots ,N-1)\) and \(\theta _k\;(k=1,\dots ,N-1)\). Set \(a(0)=a(N)=\theta _0=\theta _{N+1}=0\). Assume that
The matrix
is given by the formulas
where
with auxiliary parameters
Proof
Consider first the matrix product \(S=BD^{-1}C\). From the definition of the matrices B and C, it follows that the rows of the matrix \(B=C^t\) and the columns of the matrix C have the form
with taking empty matrices for zero and negative indices. Let Q be an \(N\times N\) matrix, the matrix \(S_Q=BQC\) is obtained by the formulas
and next using (6.9), we get
Hence, the main diagonal of the matrix \(S_Q\) is determined via
If the matrix Q is real symmetric, we have \(Q(i-1,i+1)=Q(i+1,i-1)\). Therefore,
For the subdiagonal entries, i.e., for \(i=j+1\), for the real and symmetric Q, we have
Now, we can give the details of the matrix Z. Assume that \(Q=D^{-1}\) and apply Lemma 5.1. Using (5.2), we get
Inserting this in (6.11) and using (6.1), we get
Notice that
and using (6.7), (6.8), we get (6.3).
Next using (5.2), we have
Inserting (6.13) in (6.12) and using (6.1), we obtain
and using (6.8), we get (6.4).
Finally, for \(1\le j<j+2\le i\le N\) using (5.2) again, we have
Inserting this in (6.10), we get
Using (6.1), (6.7), and (6.8), we obtain
with \(p(i-1),q(j)\) as in (6.5), (6.6) which completes the proof.\(\square \)
One can check that the condition \(c(x)\ge 0\), which implies \(c_k\ge 0\), is sufficient for the matrix Z to be strongly regular.
7 The LR factorization
We are now able to apply certain results and algorithms obtained in the paper [7]. Using Theorem 7.1 from [7], we have the following.
Theorem 7.1
Let Z be the matrix from Theorem 6.1. Assume that Z is strongly regular, meaning that it admits the factorization
where \(\Gamma \) is an invertible diagonal matrix and L is a lower triangular matrix with zero diagonal.
The nonzero part of the matrix L is given by the formulas
where p(i), a(k) are the same as in Theorem 6.1 and the values \(\beta _i\;(i=2,\dots ,N-1),\;g(j)\;(j=1,\dots ,N-2)\) as well as the diagonal entries of the matrix \(\Gamma =\textrm{diag}\left( w_1,\dots ,w_N\right) \) are determined via the following algorithm.
-
1.
Compute
$$ w_1=d_1,\;\beta _1=\delta _1/w_1,\;g(1)=q(1)/w_1,\; h_1=\left( \begin{array}{c}\beta _1\\ g(1)\end{array}\right) \left( \begin{array}{cc}\delta _1&q(1)\end{array}\right) $$ -
2.
For \(k=2,\dots ,N-2\) Compute the \(3\times 3\) matrix
$$ H_k=\left( \begin{array}{cc}1&{}0\\ 0&{}p(k)\\ 0&{}a(k)\end{array}\right) h_{k-1} \left( \begin{array}{ccc}1&{}0&{}0\\ 0&{}p(k)&{}a(k)\end{array}\right) , $$$$ w_k=d_k-H_k(1,1), $$$$ e_k=\left( \begin{array}{cc}\delta _k\\ q(k)\end{array}\right) -H_k(2:3,1),\; \left( \begin{array}{c}\beta _k\\ g(k)\end{array}\right) =e_k\frac{1}{w_k}, $$$$ h_k=H_k(2:3,2:3)+\left( \begin{array}{c}\beta _k\\ g(k)\end{array}\right) \cdot e_k^T. $$Recall that here \(H_k(2:3,1)\) selects the entries 2,3 from the first column of the matrix \(H_k\) and so on.
-
3.
Compute the \(2\times 2\) matrix
$$ H_{N-1}=\left( \begin{array}{cc}1&{}0\\ 0&{}p(N-1)\end{array}\right) h_{N-2} \left( \begin{array}{ccc}1&{}0\\ 0&{}p(N-1)\end{array}\right) , $$$$\begin{aligned}\begin{gathered} w_{N-1}=d_{N-1}-H_{N-1}(1,1),\\ \beta _{N-1}=(\delta _{N-1}-H_{N-1}(2,1))/w_{N-1},\\ h_{N-1}=H_{N-1}(2,2)+\beta _{N-1} (\delta _{N-1}-H_{N-1}(2,1)). \end{gathered}\end{aligned}$$ -
4.
Compute \(w_N=d_N-h_{N-1}\).
8 The solution of the system
Based on the LR factorization in Theorem 7.1, we obtain easily the solution of the equation \(Z{\mathfrak u}={\mathfrak y}\) as it is shown [7, Section 6, Algorithm 8.1].
Algorithm 8.1
1. Compute the solution v of the equation \((I+L)v={\mathfrak y}\).
1.1. Start with \(v(1)={\mathfrak y}(1)\) and compute \(v(2)={\mathfrak y}(2)-\beta _1 v(1)\).
1.2. Compute \(z_2=g(1)v(1),\;v(3)={\mathfrak y}(3)-\beta _2v(2)-p(2)z_2\).
1.3. For \(i=3,\dots ,N-1\) compute
2. Compute the solution \(\chi \) of the equation \(\Gamma \chi =v\).
3. Compute the solution \({\mathfrak u}\) of the equation \((I+L^t){\mathfrak u}=\chi \).
3.1. Start with \(u(N)=\chi (N)\) and compute \({\mathfrak u}(N-1)=\chi (N-1)-\beta _{N-1}{\mathfrak u}(N)\).
3.2. Compute \(\rho _{N-2}=p(N-1){\mathfrak u}(N),\; {\mathfrak u}(N-2)=\chi (N-2)-\beta _{N-2}{\mathfrak u}(N-1)-g(N-2)\rho _{N-2}\).
3.3. For \(i=N-3,\dots ,1\) compute
9 The complete algorithm
Combining the results of the preceding sections, we can now state the full algorithm:
Algorithm 9.1
1. Following Lemma 5.1, compute the representation of the matrix \(D^{-1}\) as follows.
Determine auxiliary variables
and set
Compute
2. Following Theorem 6.1, determine the representation of the matrix Z as follows. Compute
and then compute
3. Following Theorem 7.1, compute the factorization (7.1).
3.1. Compute
3.2. For \(k=2,\dots ,N-2\) compute
3.3. Compute
3.4. Compute \(w_N=d_N-h_{N-1}\).
4. Following Algorithm 8.1, compute the solution of the system \(Z {\mathfrak u}={\mathfrak y}\).
4.0. Compute \({\mathfrak y}(i)=\frac{h^4}{12}\phi _i,\;i=1,\quad ,N\).
4.1.1. Start with \(v(1)={\mathfrak y}(1)\) and compute \(v(2)={\mathfrak y}(2)-\beta _1 v(1)\).
4.1.2. Compute \(z_2=g(1)v(1),\;v(3)={\mathfrak y}(3)-\beta _2v(2)-p(2)z_2\).
4.1.3. For \(i=3,\dots ,N-1\) compute
4.2. Compute the solution \(\chi \) of the equation \(\Gamma \chi =v\).
4.3.1. Start with \({\mathfrak u}(N)=\chi (N)\) and compute \({\mathfrak u}(N-1)=\chi (N-1)-\beta _{N-1}{\mathfrak u}(N)\).
4.3.2. Compute \(\rho _{N-2}=p(N-1){\mathfrak u}(N),\; {\mathfrak u}(N-2)=\chi (N-2)-\beta _{N-2}{\mathfrak u}(N-1)-g(N-2)\rho _{N-2}\).
4.3.3. For \(i=N-3,\dots ,1\) compute
Here, \(z_i,\rho _i\) are auxiliary variables.
All this algorithm requires O(N), i.e., linear in N, number of arithmetic operations, in contrast to \(O(N^2)\) in a general case and also with \(O(N\log N)\) in the based on FFT algorithm in the case of constant potential.
Numerical evidence is given below in Section 11.4.
10 A fast algorithm for the zero potential
For the case of zero potential \(c(x)=0\) in [4] using the kernel method, an explicit solution of (4.3) was obtained via
with
Notice that the expression in the brackets may be done as
Setting
we get
Thus, using (10.1), we get
which implies
with
Here, the sums are obtained recursively via relations
Combining (10.4) and (10.5) together, we obtain the following procedure of linear O(N) complexity.
Algorithm 10.1
Set \(\chi _0=0_{2\times 1}\) and for \(i=1,\dots ,N\) compute \(\chi _i=\chi _{i-1}+\psi ^t(x_i)\phi _i\).
Set \(z_N=0_{2\times 1}\) and for \(i=N-1,\dots ,1\) compute \(z_i=z_{i+1}+\varphi ^t(x_i)\phi _i\).
For \(i=1,\dots ,N\) compute \(u_i=h(\varphi (x_i)\chi _i+\psi (x_i)z_i)\).
11 Numerical results
In the numerical results section, we consider problems of the form
We measure the error \(e(x_i)= u_{comp})_i-u_{exact}(x_i)\) using the norms \(\vert e\vert _2\) and \(\vert e\vert _{\infty }\) where
The rate of convergence is defined as follows. Assume that the discrete \(l_2\) norm \(\vert e\vert _2\) on a coarse mesh \(h_1\) is \(\vert e\vert _{2,h_1}\) and that the discrete \(l_2\) norm \(\vert e\vert _2\) on a finer mesh \(h_2=0.5 h_1\) is \(\vert e\vert _{2,h_2}\), then the rate of convergence is defined as
Similarly for the rate of convergence for the discrete infinity norm.
11.1 Example 1
Here, we consider the problem (11.6) with \(c(x)=1\).
The exact solution is \(u(x)=\sin ^2(\pi x)\); therefore, the right-hand side \(\phi (x)\) of the differential equation is \(\phi (x)=-8\pi ^4 \cos (2 \pi x)+x~sin^2(\pi x)\). The numerical errors are shown in Table 1.
11.2 Example 2
Here, we consider the problem (11.6) with \(c(x)=x\)
The exact solution is \(u(x)=\sin ^2(\pi x)\); therefore, the right-hand side \(\phi (x)\) of the differential equation is \(\phi (x)=-8\pi ^4 cos(2 \pi x)+sin^2(\pi x)\). The numerical errors are shown in Table 2.
In Fig. 1, we display the plots of the numerical solution (left) and the error (right).
11.3 Example 3
Here, we consider the problem (11.6) with \(c(x)=1/((x-0.5)^2+\epsilon )\).
The numerical errors are shown in Table 3.
We pick an exact solution u of the form \(u=p(x) \sin (1/q(x))\), where \(p(x)=16x^2(1-x)^2\), \(q(x)=(x-0.5)^2+\epsilon \), with \(\epsilon =0.05\). In addition, \(c(x)=1/q(x)\). The right-hand side \(\phi (x)\) of the equation is \(\phi (x)=u^{(4)}(x)+ c(x) u(x)\).
In Fig. 2, we display the plots of the numerical solution (left) and the error (right).
11.4 Numerical evidence on the complexity of the algorithm
In Table 4, we display computational times in seconds as N grows for Example 3.
All this algorithm requires O(N), i.e., linear in N, number of arithmetic operations, in contrast to \(O(N^2)\) in a general case and also with \(O(N\log N)\) in the based on FFT algorithm in the case of constant potential.
12 Summary
We have performed several numerical tests which assure the fourth order convergence of the scheme for the biharmonic equation also in case of the presence of a variable potential c(x). In the example \(c(x)=1\) (a positive constant), in the second example \(c(x)=x\) (a non negative function on the interval (0, 1)), and in the third example \(c(x)=x\) but for the case of highly oscillatory solution. In Table 4, we have displayed the computational time for the whole numerical procedure in cases where \(N=2047, 4096, 8191, 15381\) (N is doubled throughout the test), and we obtain the running time is growing as O(N), as expected from the design of the scheme.
Availability of data and materials
Not applicable.
References
Ben-Artzi, M., Croisille, J.P., Fishelov, D.: Navier-Stokes Equations in Planar Domain. World Scientific, (2013)
Ben-Artzi, M., Croisille, J.P., Fishelov, D.: A fast direct solver for the biharmonic problem in a rectangular grid. SIAM J. Sci. Comput. 31, 303–333 (2008)
Ben-Artzi, M., Croisille, J.P., Fishelov, D., Katzir, R.: Discrete fourth-order Sturm-Liouville problems. IMA J. Numer. Anal. 38, 1475–1522 (2018)
Ben-Artzi, M., Katriel, G.: Spline functions, the biharmonic operator and approximate eigenvalues. Numer. Math. 141(4), 839–880 (2019)
Boito, P., Eidelman, Y.: Computations of quasiseparable representations of Green matrices. arXiv:2308 02701v1 [math.NA] 4 Aug 2023
da Fonseca, C.M., Petronilho, J.: Explicit inverses of some tridiagonal matrices. Linear Algebra Appl. 325, 7–21 (2001)
Eidelman, Y., Gohberg, I.: Out-off-band quaasiseparable matrices. Linear Algebra Appl. 429, 266–289 (2008)
Eidelman, Y., Gohberg, I., Haimovici, I.: Separable type representations of matrices and fast algorithms. Vol. 1. Basics. Completion problems. Multiplication and inversion algorithms, Operator Theory: Advances and Applications, Birkhäuser, (2013)
El-Mikkawy, M., Karawia, A.: Inversion of general tridiagonal matrices. Appl. Math. Lett. 19(8), 712–720 (2006)
Fishelov, D., Ben-Artzi, M., Croisille, J.P.: Recent advances in the study of a fourth-order compact scheme for the one-dimensional biharmonic equations. J Sci. Comput. 53, 55–79 (2012)
Golub, G., Van Loan, C.: Matrix computations. The Johns Hopkins University Press, Baltimore (2013)
Meurant, G.: A review on the inverse of symmetric tridiagonal and block tridiagonal matrices. SIAM J. Matrix Anal. Appl. 13(3), 707–728 (1992)
Acknowledgements
Not applicable.
Funding
Open access funding provided by Hebrew University of Jerusalem. Not applicable.
Author information
Authors and Affiliations
Contributions
Authors M.B and Y.E. wrote the main manuscript. D.F. prepared Figures 1–2 and Tables 1–3. All authors reviewed the paper.
Corresponding author
Ethics declarations
Ethics approval
Not applicable.
Consent to participate
Not applicable.
Consent for publication
The authors approved it for publication.
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix 1
Proof of Lemma 5.1
Using the Gaussian eliminations algorithm, see for instance [11, p.165,177], we get
with
where the numbers \(\gamma _k,\;k=1,\dots ,N\) and \(f_k,\;k=1\dots ,N-1\) are determined via (5.1), (5.3). The matrix L is the product of elementary matrices \(F=\tilde{F}_1\cdots \tilde{F}_{N-1}\), where
Here, \(\otimes \) stands for a tensor product. Consider the inverse matrix \(A^{-1}\). We have
One can see easily that
with
and also
and
Next, we compute the matrix \(A^{-1}\) via the factorization
We set
it is clear that
Next set
Define the \(N-k+1\)-dimensional vector columns \(T_k,\;k=N,\dots ,1\) via recursive relations
with the parameters \(\theta _k,\;a(k)\) as in (5.5), (5.4). The relations (13.13) are equivalent to the formulas
Set \(G_k=A^{-1}(k:N,k:N),\;k=1,\dots ,N\). Thus, (5.1) is equivalent to the relations
We prove (13.15) by the backward induction. For \(j=N\) using we have
Assume that for some k with \(N-2\ge k\ge 1\), the relations
hold. This implies that
To get such a representation for \(G_k\), we use the formulas (13.1)-(13.12). Using the formulas (13.1), (13.2), (13.5), and (13.12), we get
Using the formula \(\Delta _{k-1}=(\tilde{L}_{k-1})^t\Delta _k\tilde{L}_{k-1}\) and (13.3), (13.4), (13.6), (13.7), we obtain
So we get that in fact
From here, using also (13.9) and (13.12), (13.10), we obtain
Moreover, using the formula \(\Delta _k=(\tilde{L}_k)^t\Delta _{k+1}\tilde{L}_k\), we have
and next using (5.4) and
From here, using the direct computations, we get
Hence, using (reftk), we get
which completes the proof.\(\square \)
Appendix 2
The scalar quasiseparable representations are a particular case of so called quasiseparable (separable type) representations of matrices studied in details in [8]. We present here the basic definitions and facts.
Let \(\mathcal A\) be an \(N\times N\) matrix. Consider the ranks of the maximal submatrices of \(\mathcal A\) entirely located in the strictly lower triangular part
The numbers \(r_k,\;k=1,\dots ,N-1\) are called lower rank numbers of the matrix \(\mathcal A\). Let \(r=\max _{1\le k\le N-1} r_k\) be the maximal lower rank number of \(\mathcal A\), we say that \(\mathcal A\) is lower quasiseparable of order r.
In accordance with the lower rank numbers, see [7, Section 5.2], the strictly lower triangular part of \(\mathcal A\) admits the following parametrization. There exist the integers \(t_k,\;k=1,\dots ,N\) and the \(t_{k-1}\)-dimensional rows \(p(k)\;(k=2,\dots ,N)\), the \(t_k\)-dimensional columns \(q(k)\;(k=1,\dots ,N-1)\), the \(t_k\times t_{k-1}\) matrices \(a(k)\;(k=2,\dots ,N-1)\) such that the equalities
hold. The representation (14.1) is called the lower quasiseparable representation of \(\mathcal A\), and the parameters \(p(k)\;(k=2,\dots ,N),\;q(k)\;(k=1,\dots ,N-1),a(k)(k=2,\dots ,N-1)\) are said to be lower quasiseparable generators of the matrix \(\mathcal A\) with orders \(t_k,\;k=1,\dots ,N-1\). The representation (14.1) is not unique, but for a matrix \(\mathcal A\) with the lower rank numbers \(r_k,\;k=1,\dots ,N-1\) for any set of lower quasiseparable generators, the orders \(t_k\) satisfy the inequalities ([7, Lemma 5.8])
Moreover ([7, Theorem 5.9, Corollary 5.0]), there exist a lower quasiseparable representation (14.1) with minimal orders, i.e., with orders equal to the corresponding rank numbers
For an \(N\times N\) matrix with the given elements, a set of lower quasiseparable generators may be determined in \(O(N^2)\) arithmetic operations ([7, Theorem 5.9]). In many cases, in particular in this paper quasiseparable generators, are given in advance.
The quasiseparable generators are not unique, and the choice depends on the concrete algorithm we use.
A similar structure is defined for the upper triangular part of a matrix. Consider the ranks of the maximal submatrices of \(\mathcal A\) entirely located in the strictly upper triangular part
The numbers \(s_k,\;k=1,\dots ,N-1\) are called upper rank numbers of the matrix \(\mathcal A\). Let \(s=\max _{1\le k\le N-1} s_k\) be the maximal upper rank number of \(\mathcal A\), we say that \(\mathcal A\) is upper quasiseparable of order s.
In accordance with the upper rank numbers, the strictly upper triangular part of \(\mathcal A\) admits the parametrization.
with the \(s_k\)-dimensional rows \(g(k)\;(k=1,\dots ,N-1)\), the \(s_{k-1}\)-dimensional columns\(h(k)\;(k=2,\dots ,N)\) and the \(s_{k-1}\times s_k\) matrices \(b(k)\;(k=2,\dots ,N-1)\). The representation (14.2) is called the upper quasiseparable representation of \(\mathcal A\), and the parameters of this representation \(g(k)\;(k=1,\dots ,N-1),\;h(k) (k=2,\dots ,N),\;b(k)\;(k=2,\dots ,N-1)\) are said to be upper quasiseparable generators of the matrix \(\mathcal A\) with orders \(s_k,\;k=1,\dots ,N-1\). A matrix which is lower quasiseparable of order r and upper quasiseparable of order s is said to be quasiseparable of order (r, s).
For the products of matrices in (14.1) and (14.2), we use the following notations. For \(i>j\), we define the operation \(a^>_{ij}\) via
and for \(i<j\) the operation \(b^<_{ij}\) via
It is clear that the relations
and
hold.
Thus, the entries of any matrix \(\mathcal A\) are represented in the form
with the lower and upper quasiseparable generators defined above and diagonal entries \(d(i)\;i=1,\dots ,N\). For a symmetric \(\mathcal A\), upper generators are not needed and we have
Every matrix is quasiseparable with some order. If we take the order (r, s) to be fixed, we obtain a special class of matrices. Since for an invertible \(\mathcal A\) the rank numbers of the inverse are the same ([8, Corollary 6.3]) as for the original matrix, such a class is invariant under inversion. For instance, the class of lower quasiseparable of order one matrices contains tridiagonal and unitary Hessenberg matrices as well as the inverses to them.
Quasiseparable representations of the form (14.4) are used also for block matrices. For a block square matrix with entries of sizes \(m_i\times n_j,\;i,j=1,\dots ,N\), where \(m_i,n_j\ge 0\) and \(\sum _{i=1}^Nm_i= \sum _{i=1}^Nn_i\), the quasiseparable representation is defined via (14.4) with matrix entries. Here, quasiseparable generators are the matrices \(p(i)\;(i=2,\dots ,N),\;q(j)\;(j=1,\dots ,N-1),\;a(k)\;(k=2,\dots ,N-1); g(i)\;(i=1,\dots ,N-1),\;h(j)\;(j=2,\dots ,N),\;b(k)\;(k=2,\dots ,N-1); d(k)\;(k=1,\dots ,N)\) of sizes \(m_i\times t_{i-1},t_j\times n_j, t_k\times t_{k-1};m_i\times s_i,s_{j-1}\times n_j,s_{k-1}\times s_k, m_k\times n_k\), respectively.
In practice, quasiseparable representations are used for matrices with quasiseparable orders r, s essentially smaller than the size N of a matrix \(\mathcal A\). In this case, the matrix with \(N^2\) entries turns out to be parametrized via O(N) parameters which are quasiseparable generators. Moreover, using quasiseparable representation, one can obtain various algorithms with lower complexity. The multiplication by vector algorithm via quasiseparable generator costs O(N) arithmetic operations instead of \(O(N^2)\) in the unstructured algorithms and the solution of a system of linear algebraic is performed via O(N) operations in contrast to \(O(N^3)\) in the general case. In the applications to numerical methods for differential equations, we obtain algorithms with linear in the number of the points in the grid complexity.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ben-Artzi, M., Eidelman, Y. & Fishelov, D. Numerical solution of the boundary value problems for the biharmonic equations via quasiseparable representations. Numer Algor (2024). https://doi.org/10.1007/s11075-024-01809-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11075-024-01809-9
Keywords
- Biharmonic equations
- Potential
- Dirichlet problem
- Numerical solution
- Quassiseparable representation of matrices
- Hermitian derivative
- High-order difference scheme