Abstract
In this paper, the trapezoidal rule for the Grünwald-Letnikov operator is derived. It is a trapezoidal rule in the sense that the formula yields the exact Grünwald-Letnikov derivative/integral of a piecewise linear function. Firstly, the formula for evenly spaced points is derived, and is used as a basis to derive the equivalent formula for arbitrary abscissae. Further, an analytic bound on the residual error is derived, which depends on a bound on the second derivative of the function. The derived trapezoidal rule can therefore be used to compute fractional integrals and derivatives to within a given error tolerance. Through numerical testing it is shown that the new formula yields results that are orders-of-magnitude more accurate than the classical formula, even for arbitrary functions. A simple adaptive algorithm is proposed for computing the result of applying the Grünwald-Letnikov operator to a function to within a desired accuracy.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
The numerical calculation of fractional order derivatives and integrals is critical to the simulation of fractional order systems. Specifically, a good approximation is essential to the numerical solution of fractional order differential equations [1, 2], as well as partial fractional order differential equations [3, 4]. One of the more important operators in Fractional Calculus is the Grünwald-Letnikov operator [5–7] of arbitrary order \(\alpha \), since it encompasses both the Riemann integral as the special case \(\alpha =-1\) as well as the definition of the derivative as the limit of a ratio of differences as the special case \(\alpha =1\). In terms of numerical algorithms, much attention has been given to specific fractional derivatives, such as the Caputo derivative [8, 9], and the Riemann-Liouville derivative [9, 10], including some higher order formulas. Strangely, relatively little attention has been given to the actual numerical computation of the Grünwald-Letnikov operator, in spite of the fact that its classical approximation appears to be the most commonly used formula for the numerical computation of fractional order derivatives. The classic approach to computing Grünwald-Letnikov derivatives and integrals is to take the definition of the operator at a point \(x>0\), i.e.,
where \({\varGamma }(z)\) is the Gamma Function, and to simply take a finite value for n, equal to the number of abscissae. This yields the coefficients of a linear combination of function values for the approximation. This particular approach is advocated in the majority of the literature, e.g., [6, 7, 10]. The formula has the advantage that it can be computed relatively simply using the recurrence formula for the Gamma function [7]. This approach is prevalent enough that there have been recent analyses as to the accuracy of this classic formula [11]. There are, however, two significant disadvantages due to the nature of the formula: Firstly, since the formula is only exact in the limit \(n\rightarrow \infty \), it only yields accurate results for a very large number of abscissae; secondly, the coefficients only make sense when the abscissae are evenly spaced, limiting the use of the approximation to evenly spaced function evaluations. In practical terms, this means that more often than not, a large amount of computation must be undertaken in order to yield acceptable results. An alternative approximation is given by Podlubny et al. [12]; their approximation for differentiation involves the matrix inversion of an integration operator, which unfortunately can lead to instability for a larger number of abscissae.
The new approach taken in this paper is to compute the necessary coefficients in order to yield the exact result of the Grünwald-Letnikov operator applied to a piecewise linear function. In the case of \(\alpha =-1\), this approach is the well known trapezoidal rule for numerical integration. In fact, the generalization of the trapezoidal rule derived in this paper yields exactly the important limit cases: The trapezoidal rule for integration, when \(\alpha =-1\); the appropriate identity operator for \(\alpha =0\); and finally, the backwards difference operator for the case \(\alpha =1\). Given that the new approximation is exact for piecewise linear functions over arbitrary spacing, it follows that the new method can be used to obtain more accurate solutions to fractional differential equations than the classic method with far less computational work. The method is also effective for large spacing, which is elusive to the classic method.
In order that the proposed trapezoidal rule can be used as a veritable theorem of numerical analysis, we derive an analytic bound on the residual error of the approximation. Specifically, if the second derivative of the function to which the trapezoidal rule is applied can be bounded, then the residual error of the computation can be bounded. The bound is valid provided that \(f^{\prime \prime }(x)\) is continuous, and that \(\alpha \le 1\). The bound therefore covers all of integration, i.e., \(\alpha \le 0\), as well as differentiation for \(\alpha \) in the range \(0 \le \alpha \le 1\). The case \(\alpha =1\) corresponds to backward differences, which are known to have low accuracy for computing derivatives, especially when spacing h is not chosen with care [13]. Further, the case \(\alpha =2\) corresponds to a second derivative, whereby, the second derivative of a piecewise linear function vanishes everywhere. It can therefore be said, that if a practitioner wishes to compute a derivative of order \(1 < \alpha < 2\), a higher order formula should be used. The error bound, which is valid for \(\alpha <1\), can be used for two immediate applications: Firstly, the bound can be used to calculate the numerical result to within a given error tolerance; an example of this is given in the present paper; secondly, the error bound can be used to justify the short-memory principle for numerical computation over very large data sets [7]. Previously this approach has been justified only by specific examples; the proposed error bound would give a precise bound for the general case.
Another aspect of the proposed method is the matrix-based approach. Since integration and differentiation are linear operators, it is not surprising that standard methods for approximating integrals and derivatives are to determine an appropriate linear combination of a finite number of function values. In particular, fractional order operators often require a linear combination of all function values. To this end, a matrix based approach to numerical Fractional Calculus greatly simplifies results, and leads to more accurate solutions (Cf. [2, 3]). Specifically, it is known that the classic Grünwald-Letnikov approximation is a Toeplitz matrix [2, 9]. Toeplitz matrices have a number of special properties which can simplify computation [14]. It is shown in this paper that the same advantage holds for trapezoidal rule for the Grünwald-Letnikov operator since for evenly spaced points it can be written as a sum of Toeplitz matrices. In general, throughout the paper we derive formulæ for the approximation of a linear operator over a partition of an interval, namely, evaluation at each of the points \(x_1,x_2,\ldots ,x_n\). It is assumed that the lower limit of the Grünwald-Letnikov operator is \(x_1\), i.e., the first node, whereby in the case of even spacing it suffices to assume it to be zero. Firstly, we derive the operator for the special case of evenly spaced abscissae. This provides us with sufficient groundwork to approach the case of arbitrary spacing. We thereby follow with a derivation of the general trapezoidal rule for the Grünwald-Letnikov operator, indicating the key steps in the derivation. Furthermore, we provide \(\text {MATLAB}^{\circledR }\) code for the implementation of the evenly spaced case for both the trapezoidal rule, and the proposed error bound matrix, in order that the formulæ can be independently verified. By means of extensive numerical testing, we demonstrate that the newly derived formulæ are indeed exact for piecewise linear functions, and can be used for accurate computation of fractional derivatives and integrals of arbitrary functions.
2 Evenly spaced abscissae
Since we are interested in computing the value of the Grünwald-Letnikov operator applied to a piecewise linear function, it suffices to consider its value at an arbitrary index \(i+1\). In order to derive the trapezoidal rule, we therefore recall the formula for the Grünwald-Letnikov operator (1) of an arbitrary function evaluated at the point \(x_{i+1}\) with \(x_1=0\), i.e.,
Since we are interested in asymptotically large values of n, we may divide each interval into n points, yielding a total of \(in+1\) points. This yields the formula,
For evenly spaced points with spacing h and lower limit \(x_1=0\), we have \(x_{i+1} = ih\); thereby, partitioning the summation over each interval, we have,
Each term in the summation over j now corresponds to a single interval, namely, the interval \([x_{i-j},x_{i-j+1}]\). It should be noted that this remains an exact expression for the Grünwald-Letnikov operator. For a piecewise linear approximation to the function f(x), the function is approximated by a straight line over each interval. This can be accomplished best for our purposes by defining the piecewise linear function,
with \(j=0,\ldots ,i-1\), with each segment defined by the Lagrange interpolating polynomial,
For the Grünwald-Letnikov operator, we simply need to index the function values over the index k. Thus, by substituting the indexed values of x into the interpolating functions, the individual linear portions of the function take the form,
Note that the indexing is “backwards” in the variable x (or time, say). That is, we have for each j,
Substituting the indexed function values of the piecewise linear representation into the exact formula for the Grünwald-Letnikov operator, we yield,
By rearranging, this formula can be written as,
where the constants \(\hat{C}_{i,j}^{(1)}\) and \(\hat{C}_{i,j}^{(2)}\) are given as,
The evaluation of these summations and passage to the limit can be evaluated using well known formulae (See, e.g., [6]). This yields the expressions,
for the limiting case of \(j=0\), as well as,
for \(j=1,\ldots ,i-1\). Finally, using the observation that,
we can write the approximation as,
where the constants are given as,
for \(j=0\), and,
for \(j=1,\ldots ,i-1\). Note that the isolation of the limit case \(j=0\) is important for positive values of \(\alpha \).
Another crucial observation is that the coefficient values themselves depend only on the index j, and not on the index i. This greatly simplifies the computation of the coefficients, as the structure of the computation defines a sum of Toeplitz matrices. Namely, if we relabel the coefficients as,
for \(p=1,2\), then we can define the matrix,
and the Toeplitz matrices,
for \(p=1,2\). The matrix discretization of the Grünwald-Letnikov operator is then given by the expression,
where the matrix \(\mathsf {J}\) is a circulant shift operator, defined as,
This Toeplitz structure thereby leads to greatly simplified MATLAB code for generating the matrix discretization of the operator, as can be seen in Sect. 6. Specifically, since there are \(2(n-1)\) coefficients that require computing, they can be computed in order \(\mathcal {O}\left( n\right) \) time.
3 Arbitrary spacing
For arbitrary spacing, we take the same approach, however, the formulae are slightly more complicated. In order to calculate the value of the Grünwald-Letnikov operator at the point, \(x_{i+1}\), we divide the interval \([x_1,x_{i+1}]\) into n points. In turn, we divide each individual interval \([x_ {i-j},x_{i-j+1}]\) into \(n_{j+1}\) points. This partitioning and indexing of intervals is shown in Fig. 1.
We further denote the partial sums of the divisions of each interval as,
such that \(N_0 = 0\), \(N_1 = n_1\), \(N_2 = n_1 + n_2\), and so on until \(N_i = n\). In this case, the indexing of the function values over the irregularly spaced intervals takes the form,
For this partitioning of the intervals, the formula for the Grünwald-Letnikov operator with lower limit \(x_1\) now reads,
In a similar manner to the evenly spaced case, this can be written as,
where the coefficients are computed by evaluating the sums over k analytically, followed by the passage to the limit. The key observation necessary in evaluating the limits is the fact that for the values of \(n_j\) and \(N_j\), we have the asymptotic relations,
In this vein, if we define the differences,
then the coefficients of the approximation are given as,
for the case \(j=0\), and,
for \(j=1,\ldots ,n-1\). In the case of arbitrary spacing, the coefficients do not have a Toeplitz structure, and the computation of the general coefficients is therefore an order \(\mathcal {O}\left( n\log {n}\right) \) operation.
4 Initial value and interpolation over the first interval
An issue that is of computational importance is the value of the Grünwald-Letnikov operator at the lower limit, \(x=x_1\). Depending on the value of \(f(x_1)\) and the order of the operator \(\alpha \), the initial value,
can take on any value in \([-\infty ,\infty ]\). Furthermore, the function resulting from the Grünwald-Letnikov operator can have strong asymptotic behavior over this initial interval. The interpolation function can be derived in a straightforward manner from the interpolation function over the first interval. In this case, the Newton form of the interpolating function is more appropriate, namely,
where \(h_1 = x_2-x_1\). By applying the monomial rule for the Grünwald-Letnikov operator, we yield the function,
This function can be used to interpolate the result over the interval \([x_1,x_2]\), as well as determine the value at \(x=x_1\) by means of a limiting process.
5 Error analysis
As with any numerical approximation, an essential component of the approximation is its error analysis. For the trapezoidal rule proposed herein, an error term can be estimated based on the remainder formula for Lagrange interpolation. This derivation is followed by the error analysis for arbitrary spacing.
5.1 Evenly spaced points
We can write an exact representation of a function, f(x) over the interval \([x_{i-j},x_{i-j+1}]\), by means of the Lagrange interpolation formula and its error term. Specifically, a continuous function f(x) can be written as,
where the function g(x) is again the piecewise linear representation of f(x) and r(x) is a remainder function. The remainder is also defined piecewise as,
The functions, \(r_{i-j}(x)\), are given by the Lagrange remainder,
where \(\xi _{i-j}(x) \in [x_{i-j},x_{i-j+1}]\) is an unknown value that depends on x. Thus, by applying the Grünwald-Letnikov operator to both sides of Eq. (45), we obtain,
The remainder term can be computed by substituting the remainder function into the formula for the Grünwald-Letnikov operator obtained in Eq. (4), namely,
By substituting the indexed values of x into the remainder function, we have,
Substituting this relation into the previous equation yields,
This relation can be written more compactly by defining the constants,
and
The expression for the Grünwald-Letnikov operator applied to the remainder term can thereby be written as,
The coefficients \(p_k\) in the summation are given by a recurrence relation [7] as,
Hence, if \(\alpha <0\), then the coefficients, \(p_k\), are zero or greater for all k. Similarly, if \(0 \le \alpha \le 1\), then clearly, \(p_0 > 0\), but \(p_k < 0\) for \(k>0\). In a similar manner, the coefficients \(q_k\) are determined by a polynomial function which is quadratic in k, and has roots at \(k=jn\) and \(k=(j+1)n\). Between these roots, the function is negative, and hence we can say that the coefficients \(q_k\le 0\) for all k. Putting this together, we can establish the fact that the product of the coefficients does not change sign when \(\alpha \le 1\), i.e.,
The product \(p_k q_k\) represents successive values of a continuous function, which does not change sign over the interval \([x_{i-j},x_{i-j+1}]\). Therefore, if the second derivative \(f^{\prime \prime }(x)\) is continuous over the interval, then by the generalized mean value theorem [15], we can write,
where \(\eta _{i-j}\) is a value in the interval \([x_{i-j},x_{i-j+1}]\). By rearranging, we have,
Finally, by evaluating the sum and the limit, we have,
where the coefficients, \(\kappa _j\), are given as,
and
for \(j=1,2,\ldots ,i-1\). Since these coefficients depend only on the index j, they again define a Toeplitz matrix. That is, for the purpose of computing error bounds, we can define the matrix,
This matrix can be used to compute upper and lower bounds on the residual error of the computed result. Provided the second derivative can be bounded, this can be done regardless if the exact solution is known or not. In order to compute the error bounds, we substitute the result in Eq. (61) into (48) to yield,
This expression describes the residual difference between the exact solution and the trapezoidal rule; the relation itself is exact, recalling however that the values \(\eta _{i-j}\) are generally unknown.
By defining the vectors,
as well as,
the relation can be written for all points, \(x_2,\ldots ,x_n\) as,
This relation can be used to bound the residual error of the computation. Specifically, if the second derivative is continuous, then by the extreme value theorem [15], it can be bounded over each interval such that,
We therefore define the vectors,
To determine a bound, the sign of the coefficients in the matrix \(\mathsf {R}\) are important. Firstly, for the range of \(\alpha \) that is of interest, i.e., \(\alpha <1\), the common factor of the matrix \(\mathsf {R}\) is always positive such that,
Further, since for \(\alpha < 0\) the coefficients \(\kappa _k\) were obtained as the sum of terms that are less than or equal to zero, the coefficients are negative, i.e.,
Similarly, for the range \(0 < \alpha <1\) the coefficients were obtained as the sum of terms that are greater or equal to zero, hence the coefficients are positive, i.e.,
We may now take advantage of the fact that in the specified cases, namely when \(\alpha <1\), the non-zero entries of the matrix \(\mathsf {R}\) all have the same sign. When the Grünwald-Letnikov operator is computed with \(0\le \alpha \le 1\), then the operator represents fractional differentiation; the non-zero entries of \(\mathsf {R}\) are all positive. The residual error is therefore bounded by the relation,
Similarly, for the case \(\alpha <0\), the Grünwald-Letnikov operator represents fractional integration; the non-zero entries of \(\mathsf {R}\) are all negative, and hence the residual error is bounded by,
With these relations it is therefore possible to bound the residual error for all cases \(\alpha < 1\), if a bound on the second derivative can be computed. For a given function, if its second derivative can be computed in a straightforward manner, then the bounds over each interval can be computed, e.g., using interval analysis [16]. The obvious benefit to this bound is that a simple algorithm can be derived to adjust the spacing such that the residual is guaranteed to be smaller than a given tolerance. For such an algorithm to be efficient, we require the error bound for arbitrary spacing, which is developed in the following.
5.2 Arbitrary spacing
The error bound for arbitrary spacing is computed in a similar manner as to that for even spacing. The remainder function, discretized for arbitrary spacing is given as,
over the interval \([x_{i-j},x_{i-j+1}]\), where again, \(\xi _{i-j}(k) \in [x_{i-j},x_{i-j+1}]\), is an unknown value. As with the evenly spaced case, the Grünwald-Letnikov operator applied to the error term takes the form,
with \(\eta _{i-j}\in [x_{i-j},x_{i-j+1}]\). The coefficients for the error bound are given as,
for \(j=0\), as well as,
for \(j=1,\ldots ,i-1\), with \(u_{i,j}\) and \(v_{i,j}\) given as in Eqs. (36) and (37). The identical argument follows in terms of the signs of the coefficients, and hence the error bounds are given in terms of Eqs. (74) and (75).
5.3 Backwards compatibility
It is straightforward to show that the formulae are backwards compatible with the known error formulae for the trapezoidal rule. For the case \(\alpha =1\), the trapezoidal rule is equivalent to the backward-difference formula. In this case, simple substitution shows that \(\kappa _0=1\), and \(\kappa _j = 0\) for \(j=1,\ldots ,n-2\). Therefore, the error bound is given in terms of the matrix,
which is the known formula for the error of backward-differences [17]. For the case of \(\alpha =-1\), we have simple integration. Again, simple substitution shows that, \(\kappa _j=-1\) for \(j=0,1,\ldots ,n-2\), and therefore the bound matrix is given as,
where \(\mathsf {L}\) is the \((n-1)\times (n-1)\) lower triangular matrix,
Again, this is the well known error term for the classical trapezoidal rule for integration [17].
6 Numerical implementation and testing
6.1 MATLAB Code
The following MATLAB codeFootnote 1 generates a matrix \(\mathsf {S}\) for the trapezoidal approximation to the Grünwald-Letnikov operator for evenly spaced points. The input to the function is the spacing,
, the number of abscissae
, and the order,
:
Note that this code has been arranged for compactness and readability; it is not necessarily optimized for speed or memory allotment. The code for generating the error bound matrix is given as:
As for the code for arbitrary spacing, the structure of the coefficients is much the same as that for evenly spaced points, except for the notion of the Toeplitz matrix. The necessary matrix can be computed by means of a double nested loop, with i ranging from 1 to \(n-1\), and j ranging from 1 to \(i-1\). The complexity of computing all coefficients for arbitrary spacing is therefore \(\mathcal {O}\left( n\log n\right) \).
6.2 Numerical testing
In this section we utilize four numerical tests to demonstrate the functionality of the newly proposed formulae:
-
1.
Testing the formulae for a linear function with even spacing, showing that the results are exact. For this test, we compare the results to the classic Grünwald-Letnikov approximation and show that it is not exact (even for linear functions) and only achieves accurate results for a very large number of abscissae.
-
2.
Testing the formulae for a linear function with arbitrary spacing, in this case with randomized abscissae. The results of the new formulae are also exact in this case. This test also demonstrates the interpolation formula for the first interval derived in Sect. 4.
-
3.
We then demonstrate the full functionality of the formulae by applying the numerical Grünwald-Letnikov operator to a sinusoidal function. We compare the classic approximation to the new trapezoidal rule with both evenly spaced points, as well as arbitrary spacing. For the case of arbitrary spacing, the nodes were chosen such that they yield an optimal piecewise-linear approximation of the sinusoid.
-
4.
Finally, we use the new error bound to develop a simple algorithm for calculating the Grünwald-Letnikov integral or derivative of a function to a given accuracy tolerance without knowing the exact solution, but knowing only the second derivative.
The results of the first test are shown in Fig. 2 for the values \(\alpha =\frac{3}{2},\frac{1}{2},-\frac{1}{2},-\frac{3}{2}\). As expected, applying the new formula to a linear function evaluated at evenly spaced points yields exact results. For this particular test we chose \(n=31\) points, such that the spacingFootnote 2 was \(h=\frac{1}{10}\). Even for this relatively small spacing the classical approximation yields a significant error in the approximation. For a much larger number of abscissae the classic approximation will become more accurate; however, due to the fact that we are always restricted to finite precision arithmetic, the classic approximation will always have a bias.
The second experiment is simply to show that the new approximation also yields exact results when applied to a linear function evaluated at arbitrarily spaced points. The results are shown in Fig. 3, whereby the function was evaluated at \(n=10\) arbitrarily spaced points, and the same values of \(\alpha \) as above. This test shows that the results are still exact for linear functions evaluated at large intervals of unequal lengths, and that the interpolation formula over the interval \([x_1,x_2]\) is also exact for this case.
To demonstrate the possibilities for applying the new formulae in practise, they have been applied with \(\alpha =-\frac{1}{2}\) to the trigonometric function, \(f(x)=\cos {x}\); the results are shown in Fig. 4. Firstly, the classic approximation was applied to the function with \(n=49\) evenly spaced points. Secondly, the trapezoidal rule for even-spacing was applied to the same points. Thirdly, we determined a spacing numerically which best represents the cosine function as a piecewise linear function, with the same number of nodes. Clearly the results show that the classic approximation to the Grünwald-Letnikov operator still has significant bias, especially near the initial values. The evenly spaced trapezoidal rule provides a much better approximation. In fact, in the residual plot it is shown to be roughly one to two orders-of-magnitude more accurate. The residual plot also shows that a better choice of spacing, in combination with the trapezoidal rule for arbitrary spacing, can be used to yield more accurate results with the same number of points. Figure 5 shows the residuals for the two new methods for this example. In addition, it shows the computed bounds on the residuals for both methods. Clearly, corroborating the theory, the residual lies between the two computed bounds. An interesting effect, is that although the arbitrary spacing has made the solution more accurate, the bound about the residual is slightly larger. This is due to the fact that in some places where the second derivative is small, the spacing, h, was chosen to be larger. In this case, the error bound increases with \(h^{\frac{5}{2}}\), and therefore increases with increased spacing values.
As a final example, the error bound developed in this paper can be used to compute a numerical approximation to the Grünwald-Letnikov operator applied to a continuous function to within a given tolerance of accuracy. As an example, we chose the function,
since its second derivative can be computed with relative ease, but its fractional derivative not. The aim was to compute the half-derivative, \(\alpha =\frac{1}{2}\), at a set of points in the interval \(x\in [0,(\frac{13\pi }{2})^{\frac{5}{9}}]\), to within a tolerance of \(t=\pm 0.025\). Choosing an initial partition of \(n=15\) evenly spaced points, we computed the bounds on the error—which only involves bounding the second derivative. For any point where the residual exceeded the tolerance, we added the mid-point in the preceding interval to the set of abscissae. This process is continued until the residual bound for all points lies within the tolerance \(t=\pm 0.025\). Finally, the trapezoidal rule is computed at the set of adapted abscissae, which is guaranteed to be accurate to within the tolerance. The results of this process are shown in Fig. 6. Here, the initial trapezoidal approximation with \(n=15\) is shown. Further, the final set of function evaluations are shown, following the adaptive computation. The result is shown for the initial trapezoidal approximation, as well as the final adaptive computation of the half-derivative. In addition, the solution is also interpolated over the first interval; since it is a half-derivative with \(f(x_1)=1\), it tends asymptotically to \(+\infty \). Figure 7 shows the bounds on the second derivative of the function at the final partitioning, as well as the residual bounds. The initial trapezoidal approximation with \(n=15\) is shown to have large bounds, which are outside of the tolerance. Following the adaptive scheme, the final residual bounds lie entirely within the proposed tolerance.
7 Outlook and conclusion
In this paper, we derived formulae for the exact application of the Grünwald-Letnikov operator to piecewise linear functions, essentially a trapezoidal rule for the Grünwald-Letnikov operator. Since the formulae were derived for arbitrary spacing, the formulae can be used to advantage by selecting abscissae such that higher accuracy is obtained with fewer points. Further, we derived an analytic bound on the residual error for the computation. The error bound can be used to compute the result to within a given tolerance. The error bound can also be used to compute a rigorous error estimate when applying the short memory principle in practical computations. Future work is to extend the result to higher order formulas, which should be possible based on the results of the present paper.
Notes
The following code was written and tested in MATLAB R2013b; no special toolboxes are required. Note that
is used instead of
to aid readability.
The choice of \(h=0.1\) is not the best since it has no exact binary representation; it was chosen to demonstrate that the formulas remain accurate also in this case.
References
Harker M, O’Leary P (2014) Numerical solution of fractional order differential equations via matrix-based methods, Fractional Calculus: theory, vol 1. Nova Science Publishers, Hauppauge, NY
Podlubny I (2000) Matrix approach to discrete fractional calculus. Fract Calc Appl Anal 3(4):359–386
Harker M, O’Leary P (2015) Sylvester equations and the numerical solution of partial fractional differential equations. J Comput Phys 293:370–384
Podlubny I, Chechkin A, Skovranek T, Chen YQ, Jara BMV (2009) Matrix approach to discrete fractional calculus II: partial fractional differential equations. J Comput Phys 228(8):3137–3153
Grünwald A (1867) Ueber ”begrenzte” derivationen und deren anwendung. Zeitschrift f Mathematik u Physik 12(6):441–480
Miller K, Ross B (1993) An introduction to the fractional calculus and fractional differential equations. Wiley, New York
Podlubny I (1999) Fractional differential equations, mathematics in science and engineering, vol 198. Academic Press, San Diego
Diethelm K, Ford N, Freed A, Luchko Y (2005) Algorithms for the fractional calculus: a selection of numerical methods. Comput Methods Appl Mech Eng 194:743–773
Harker M, O’Leary P (2014) Polynomial accurate numerical fractional order integration and differentiation. In: International conference on fractional differentiation and its applications (ICFDA), 2014, pp 1–6
Oldham K, Spanier J (1974) The fractional calculus: theory and applications of differentiation and integration to arbitrary order. Dover Publications Inc, Minneola, USA
TenreiroMachado J (2015) Numerical calculation of the left and right fractional derivatives. J Comput Phys 293:96–103
Podlubny I, Skovranek T, Jara BMV, Petras I, Verbitsky V, Chen Y (2013) Matrix approach to discrete fractional calculus III: non-equidistant grids, variable step length and distributed orders. Philos Trans R Soc Lond A Math Phys Eng Sci 371(1990)
Press W, Teukolsky S, Vetterling W, Flannery B (2007) Numerical recipes: the art of scientific computing, 3rd edn. Cambridge University Press, Cambridge
Golub G, Van Loan C (1996) Matrix computations, 3rd edn. The Johns Hopkins University Press, Baltimore
Courant R, John F (1989) Introduction to calculus and analysis, classics in mathematics, vol I & II. Springer, Berlin
Moore R, Kearfott R, Cloud M (2009) Introduction to interval analysis. SIAM, Philadelphia
Burden R, Faires J (2005) Numerical Analysis, 8th edn. Thomson Learning, Inc, Belmont
Acknowledgments
Open access funding provided by Montanuniversität Leoben.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.
About this article
Cite this article
Harker, M., O’Leary, P. Trapezoidal rule and its error analysis for the Grünwald-Letnikov operator. Int. J. Dynam. Control 5, 18–29 (2017). https://doi.org/10.1007/s40435-016-0236-z
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40435-016-0236-z