1 Introduction

Over the past decades, fractional order systems have attracted increasing attention from the control community, since many engineering plants and processes cannot be described concisely and precisely without introducing fractional order calculus[1, 2]. Due to more and more scholars devoting themselves to the fractional order field, a tremendous amount of valuable results on system identification[3,4], controllability and observability[5, 6], stability analysis[79] and controller synthesis[1012] of fractional order systems have been reported in the literature. Many fundamentals and applications of fractional order control systems can be found in [3] and the reference therein.

The reconstruction of system state from its input and output has received a great deal of attention recently. Doye et al.[13] studied the fractional order Luenberger observer and observer-based controller design. Further, it was extended to the uncertain case via linear matrix inequality approach[14]. There are some special Luenberger state observers, such as the discrete form observer[15] and proportional integral (PI) observer[16]. With the frequency distributed model[17] of a fractional order system proposed, an interesting result is obtained that the conventional state is a pseudo state of the system, not the true state. Based on the frequency distributed model, Sabatier et al.[18, 19] discussed the observability of a class of fractional order systems and proposed a class of Luenberger observers.

Despite the plentiful achievements, some critical problems still need further investigation. The system parameters should be known. For the case where no a priori knowledge of the system parameters is available, the so-called adaptive observers should be introduced. The basic idea in the approach is to use a Luenberger observer that will continuously adapt to the parameters. The related research in the integer order area has been reported in [2023], while there exist few results in the literature investigating this point in fractional order area. There are two main reasons. Firstly, it is difficult to prove the stability of the closed-loop system. Secondly, the parameter identification belongs to the category of identification of continuous systems[24], in which there are still many problems to be solved. To the best of our knowledge, looking for an adaptive observer for fractional order system still remains open.

Motivated by these observations, this article focuses on adaptive observation for fractional order system. The theoretical contributions of this work are as follows. Firstly, a novel adaptive observation scheme is derived, which continuously adapts to the parameters in the observer. Secondly, based on the identification of continuous system, a modified parameter identified algorithm of fractional order system is proposed. The former study solves the problem “viable or not”, making the impossible possible. The latter one improves the performance of the fractional order system identification.

The remainder of this article is organized as follows. Section 2 briefly provides some preliminaries for fractional order system and problem statement. A novel adaptive observation scheme is proposed in Section 3. In Section 4, two simulation examples are provided to illustrate the validity of the proposed approach. Finally, Section 5 concludes the study.

2 Preliminaries

Consider the following linear time invariant single-input single-output (LTI SISO) plant

$$G(s) = {{{b_1}{s^{(m - 1)\alpha}} + {b_2}{s^{(m - 2)\alpha}} + \cdots + {b_m}} \over {{s^{n\alpha}} + {a_1}{s^{(n - 1)\alpha}} + {a_2}{s^{(n - 2)\alpha}} + \cdots + {a_n}}}$$
(1)

where α ∈ (0, 1) is the fractional commensurate order, u(t) is assumed as a piecewise continuous bounded function of time, y(t) is the measureable output, a i (i = 1, 2, ⋯,n) and b j (j = 1, 2, ⋯, m) are constants but unknown, m, n are known positive integers and m < n, and the system is supposed to be stable.

In this study, the Caputo’s definition is adopted for fractional order derivative, i.e.,

$$_cD_t^\alpha f(t) \buildrel \Delta \over = {1 \over {\Gamma (k - \alpha)}}\int_c^t {{{{f^{(k)}}} \over {{{(t - \tau)}^{\alpha - k + 1}}}}{\rm{d}}\tau}$$
(2)

where k is a positive integer and k − 1 ≤ αk. To simplify the notation, we denote the fractional derivative of order α as Dα instead of \(_0D_t^\alpha\) in this work.

This study aims at constructing a scheme that estimates both the plant parameters

$$\left\{{\matrix{{{a_p} = {{[{a_1}\quad {a_2}\quad \cdots \quad {a_n}]}^{\rm{T}}}} \hfill \cr{{b_p} = {{[{b_1}\quad {b_2}\quad \cdots \quad {b_m}]}^{\rm{T}}}} \hfill \cr}} \right.$$
(3)

as well as the system pseudo state x(t) using only input u(t) and output measurements y(t).

To construct the above scheme, we first introduce a definition as follows.

  • Definition 1. Matrix A of the system

    $${D^\alpha}x(t) = Ax(t)$$
    (4)

    is said to be a stable matrix if and only if it satisfies

    $$\vert \arg ({\rm{spec}}(A))\vert > {{\alpha \pi} \over 2}$$
    (5)

    where arg (z) denotes the principle argument of z and spec (A) denotes the spectrum of A.

3 Fractional order adaptive observer

3.1 Observer design

A minimal realization of the system (1) can be expressed as

$$\left\{{\matrix{{{D^\alpha}x(t) = Ax(t) + Bu(t)} \hfill \cr{y(t) = Cx(t) + v(t)} \hfill \cr}} \right.$$
(6)

where the stable matrix ARn×n, B, CTRn, {A, B} is controllable and {A, C} is observable, and the measurement noise v (t) is white noise with variance \(\sigma _v^2\).

On the basis of this transformation, the problem becomes how to construct a scheme that estimates both the plant parameters, i.e., A, B, C as well as the state vector x(t). A good starting point for designing an adaptive observer is the Luenberger observer[22, 23]

$$\left\{{\matrix{{{D^\alpha}\hat x(t) = A\hat x(t) + Bu(t) + L[y(t) - \hat y(t)]} \hfill \cr{\hat y(t) = C\hat x(t)} \hfill \cr}} \right.$$
(7)

where L is chosen so that ALC is a stable matrix and guarantees that \(\hat x(t)\) converges to x (t) exponentially for any initial condition x (0) and any input u (t).

In order to construct the structure of the adaptive observer, the unknown matrices A, B and C in the Luenberger observer (7) should be replaced with their estimates \(\hat A,\hat B\) and \(\hat C\) respectively, generated by some adaptive law.

$$\left\{{\matrix{{{D^\alpha}\hat x(t) = \hat A(t)\hat x(t) + \hat B(t)u(t) + L[y(t) - \hat y(t)]} \hfill \cr{\hat y(t) = \hat C(t)\hat x(t).} \hfill \cr}} \right.$$
(8)

There are mainly 3 disadvantages in such the structure of the adaptive observer:

  1. 1)

    The estimate matrix \({\hat A}\) cannot be guaranteed as stable, which implies that the estimate state \(\hat x(t)\) may not be convergent.

  2. 2)

    Since the state space realization result of the transfer matrix G(s) is not unique, matrices A, B and C are not unique either.

  3. 3)

    The number of the parameters to be estimated is n2 + 2n, while the number of the unknown parameters in G(s) is m + n.

To eliminate these disadvantages, a special realization form called the observable canonical form is developed. Therefore, matrices A, B and C in (6) are given by

((9))

Consequently, there are m + n parameters to be estimated and we no longer need to estimate the matrix C. Then, related parameters in the adaptive observer (8) can be rewritten as

((10))

where \({{\hat a}_p}(t)\) and \({{\hat b}_p}(t)\) are the estimates of vectors a p and b p , respectively, and a o Rn is chosen such that

$${A_o} = \left[ {- {a_o}\left\vert {{{{I_{n - 1}}} \over {{0_{1 \times (n - 1)}}}}} \right.} \right]$$
(11)

is a stable matrix.

  • Remark 1. Matrix A o in (11) is not unique. Regarding the stability, any stable matrices can be adopted for A o . In addition, the convergence speed of the state observation depends on A o .

3.2 Parameter estimation

The corresponding differential equation of system (1) can be expressed as

$$\matrix{\hfill {{D^{n\alpha}}y(t) + \sum\limits_{i = 1}^n {{a_i}{D^{(n - i)\alpha}}y(t) - {D^{n\alpha}}v(t) -}} \cr\hfill {\sum\limits_{i = 1}^n {{a_i}{D^{(n - i)\alpha}}v(t) = \sum\limits_{j = 1}^m {{b_j}{D^{(m - j)\alpha}}u(t).}}} \cr}$$
(12)

To obtain a regression-like form of system (12), two methods will be introduced.

In method 1, define the differential operator \({p^\alpha} = {D^\alpha},D({p^\alpha}) = {p^{n\alpha}} + {a_1}{p^{(n - 1)\alpha}} + \cdots + {a_n}\) and \(N({p^\alpha}) = {b_1}{p^{(m - 1)\alpha}} + {b_2}{p^{(m - 2)\alpha}} + \cdots + {b_m}\). Then, system (12) can be written in an alternative time-domain differential operator form as

$$D({p^\alpha})y(t) - D({p^\alpha})v(t) = N({p^\alpha})u(t).$$
(13)

Filtering the signals on both sides of (13) with the following stable filter

$$F({p^\alpha}) = {p^{n\alpha}} + {f_1}{p^{(n - 1)\alpha}} + \cdots + {f_n}$$
(14)

we have

$$\matrix{{y(t) = {{F({p^\alpha}) - D({p^\alpha})} \over {F({p^\alpha})}}y(t) + {{N({p^\alpha})} \over {F({p^\alpha})}}u(t) +} \hfill \cr{\quad \quad \;\;{{D({p^\alpha})} \over {F({p^\alpha})}}v(t).} \hfill \cr}$$
(15)

By defining some new variabilities, (15) can be expressed in a regression-like form as

$$y(t) = {\varphi ^{\rm{T}}}(t)\theta + v(t)$$
(16)

where

$$\matrix{{\varphi (t) = \left[ {{{{p^{(n - 1)\alpha}}y(t)} \over {F({p^\alpha})}} \cdots {{y(t)} \over {F({p^\alpha})}}{{{p^{(m - 1)\alpha}}u(t)} \over {F({p^\alpha})}} \cdots {{u(t)} \over {F({p^\alpha})}}} \right.} \hfill \cr{\quad \quad \quad {{\left. {{{{p^{(n - 1)\alpha}}v(t)} \over {F({p^\alpha})}} \cdots {{v(t)} \over {F({p^\alpha})}}} \right]}^{\rm{T}}}} \hfill \cr{\theta = {{[{f^{\rm{T}}} - a_p^{\rm{T}}\quad b_p^{\rm{T}}\quad c_p^{\rm{T}}]}^{\rm{T}}}} \hfill \cr{f = {{[{f_1}\quad {f_2}\quad \cdots \quad {f_n}]}^{\rm{T}}}} \hfill \cr{c = {{[{c_1}\quad {c_2}\quad \cdots \quad {c_n}]}^{\rm{T}}}.} \hfill \cr}$$

In method 2, considering the generalized Poisson moment functional

$${1 \over {H({p^\alpha})}} = {\left({{1 \over {{p^\alpha} + \omega}}} \right)^\kappa}$$
(17)

with κ > n, system (12) can be rewritten as

$$\matrix{\hfill {{{{p^{n\alpha}}y(t)} \over {H({p^\alpha})}} + \sum\limits_{i = 1}^n {{a_i}{{{p^{(n - i)\alpha}}y(t)} \over {H({p^\alpha})}} - {{{p^{n\alpha}}v(t)} \over {h({p^\alpha})}}} -} \cr\hfill {\sum\limits_{i = 1}^n {{a_i}{{{p^{(n - i)\alpha}}v(t)} \over {H({p^\alpha})}} = \sum\limits_{j = 1}^m {{b_j}} {{{p^{(n - j)\alpha}}u(t)} \over {H({p^\alpha})}}.}} \cr}$$
(18)

By defining \(\bar y(t) = {{{p^{n\alpha}}y(t)} \over {H({p^\alpha})}}\), system (18) can be written in a regression-like form as

$$\bar y(t) = {\bar \varphi ^{\rm{T}}}(t)\bar \theta + v(t)$$
(19)

where the regressor and the parameter vectors are now defined by

$$\matrix{{\bar \varphi (t) = \left[ {{{- {p^{(n - 1)\alpha}}y(t)} \over {H({p^\alpha})}} \cdots {{- y(t)} \over {H({p^\alpha})}}{{{p^{(m - 1)\alpha}}u(t)} \over {H({p^\alpha})}} \cdots {{u(t)} \over {H({p^\alpha})}}} \right.} \hfill \cr{{{\left. {\quad \quad \quad {{{p^{(n - 1)\alpha}}v(t)} \over {H({p^\alpha})}} \cdots {{v(t)} \over {H({p^\alpha})}}} \right]}^{\rm{T}}}} \hfill \cr{\bar \theta ={{[a_p^{\rm{T}}\quad b_p^{\rm{T}}\quad c_p^{\rm{T}}]}^{\rm{T}}},{c_p} = {{[{c_1}\quad {c_2}\quad \cdots \quad {c_n}]}^{\rm{T}}}.} \hfill \cr}$$

Since it is very similar to estimate a p and b p based on the least squares method through (16) and (19), respectively, only the former one is introduced.

With regard to (16), at any time instant t = t k , k = 1, 2, ⋯, L, the following standard linear regression-like form can be obtained as

$$y({t_k}) = {\varphi ^{\rm{T}}}({t_k})\theta + v({t_k}).$$
(20)

Now, from L available samples of the input and output signals observed at discrete times t1, ⋯,t L , not necessarily uniformly spaced, the linear least squares based parameter estimates are given by

$$\hat \theta ({t_L}) = {\left[ {\sum\limits_{k = 1}^L {\varphi ({t_k}){\varphi ^{\rm{T}}}({t_k})}} \right]^{- 1}}\left[ {\sum\limits_{k = 1}^L {\varphi ({t_k}){y ^{\rm{T}}}({t_k})}} \right].$$
(21)

\(\hat \theta ({t_L})\) cannot converge to θ, because ϕ (t k ) is correlated with v (tk). To correct the deviations of the parameter estimate, we replace p (t k ) with

$$\matrix{{\hat \varphi ({t_k}) = \left[ {{{{p^{(n - 1)\alpha}}y(t)} \over {F({p^\alpha})}} \cdots {{y({t_k})} \over {F({p^\alpha})}}{{{p^{(m - 1)\alpha}}u({t_k})} \over {F({p^\alpha})}} \cdots {{u({t_k})} \over {F({p^\alpha})}}} \right.} \hfill \cr{{{\left. {\quad \quad \quad {{{p^{(n - 1)\alpha}}\hat v(t)} \over {F({p^\alpha})}} \cdots {{\hat v(t)} \over {F({p^\alpha})}}} \right]}^{\rm{T}}}} \hfill \cr}$$
(22)

where \(\hat v({t_k})\) is the estimation of v (t k ).

$$\hat v({t_k}) = y({t_{k - 1}}) - {\hat \varphi ^{\rm{T}}}({t_{k - 1}})\hat \theta ({t_{k - 1}}).$$
(23)

Consequently, the modified solution \(\hat \theta ({t_L})\) can be expressed as

$$\hat \theta ({t_L}) = {\left[ {\sum\limits_{k = 1}^L {\hat \varphi ({t_k}){{\hat \varphi}^{\rm{T}}}({t_k})}} \right]^{- 1}}\left[ {\sum\limits_{k = 1}^L {\hat \varphi ({t_k}){y^{\rm{T}}}({t_k})}} \right].$$
(24)

To estimate the parameter on line, the matrix inversion theorem is used to get the recursive least square algorithm as

$$\left\{{\matrix{{K({t_{k + 1}}) = {{P({t_k})\hat \varphi ({t_{k + 1}})} \over {1 + {{\hat \varphi}^{\rm{T}}}({t_{k + 1}})P({t_k})\hat \varphi ({t_{k + 1}})}}} \hfill \cr{\varepsilon ({t_{k + 1}}) = y({t_{k + 1}}) - {{\hat \varphi}^{\rm{T}}}({t_{k + 1}})\hat \theta ({t_k})} \hfill \cr{\hat \theta ({t_{k + 1}}) = \hat \theta ({t_k}) + K({t_{k + 1}})\theta ({t_{k + 1}})} \hfill \cr{P({t_{k + 1}}) = [{I_{m + 2n}} - K({t_{k + 1}}){{\hat \varphi}^{\rm{T}}}({t_{k + 1}})]P({t_k})} \hfill \cr{\hat v({t_{k + 1}}) = y({t_k}) - {{\hat \varphi}^{\rm{T}}}({t_k})\hat \theta ({t_k})} \hfill \cr}} \right.$$
(25)

where \(k > 0,\hat v({t_1}) = 0,\hat \theta ({t_1})\) can be chosen arbitrarily, and P (t1) is usually selected as 103∼8 Im+2n.

As a result, one can get the original system parameter estimation from

$$\hat \theta ({t_k}) = \left[ {\matrix{{f - {{\hat a}_p}({t_k})} \cr{{{\hat b}_p}({t_k})} \cr{{{\hat c}_p}({t_k})} \cr}} \right]$$
(26)
$$\hat \bar \theta ({t_k}) = \left[ {\matrix{{{{\hat a}_p}({t_k})} \cr{{{\hat b}_p}({t_k})} \cr{{{\hat c}_p}({t_k})} \cr}} \right].$$
(27)
  • Remark 2. The filter F (pα) in (14) is not unique. Regarding the stability and the maximum order of , any filters can be adopted for F (pα).

  • Remark 3. The generalized Poisson moment functional \({1 \over {H({p^\alpha})}}\) in (17) is not unique. Although any κn is feasible, one usually selects κ = n. In addition, ω can be selected to guarantee system (1) and \({1 \over {H({p^\alpha})}}\) have a similar bandwidth.

  • Remark 4. Actually, parameter c p can be calculated from parameter a p and the coefficients of H(pα) or F(pα). However, it is not affected that one can identify c p and a p independently.

  • Remark 5. The two methods realize the parameter adaptation by predicting the output y(i) and the filtered output \(\bar y(t)\), receptively. Additionally, the parameter estimation results depend on H(pα), F(pα) and other parameters. As a result, it is hard to say which one is better.

3.3 Stability analysis

For any time t k t < t k +1, we have

$$\left\{{\matrix{{{{\hat a}_p}(t) = {{\hat a}_p}({t_k})} \hfill \cr{{{\hat b}_p}(t) = {{\hat b}_p}({t_k}).} \hfill \cr}} \right.$$
(28)

Based on the previous result, we are ready to present the adaptive observer for the fractional order system.

  • Theorem 1. For plant (6), if the adaptive observer is designed as (8), (10) and (25), then all the signals in the closed-loop adaptive system are global uniformly bounded. And if u(t) can guarantee that φ (t) is persistent excitation, then the parameter estimation and state observation are achieved as

    $$\left\{{\matrix{{\mathop {\lim}\limits_{t \rightarrow \infty} [\theta - \hat \theta (t)] = 0} \hfill \cr{\mathop {\lim}\limits_{t \rightarrow \infty} [x(t) - \hat x(t)] = 0.} \hfill \cr}} \right.$$
    (29)
  • Proof. The observer equation we design can be written as

    $$\matrix{{{D^\alpha}\hat x(t) = {A_o}\hat x(t) + [\hat A(t) - {A_o}]x(t) +} \hfill \cr{\quad \quad \quad \quad {{\hat b}_p}(t)u(t) + L(t)v(t).} \hfill \cr}$$
    (30)

Since A o is a stable matrix and \({{\hat b}_p},\hat A(t),L(t),x(t),u(t),v(t)\) are bounded, it follows that \(\hat x(t)\) is bounded, which in turn implies that all signals are bounded.

Based on the convergence properties of the least squares for the linear regression form problem, one concludes that \(\hat \theta ({t_L})\) is an unbiased estimation of θ. In other words, the equation \(\mathop {lim}\limits_{t \rightarrow \infty} [\theta - \hat \theta (t)] = 0\) holds.

Define the state observation error as

$$\tilde x(t) = x(t) - \hat x(t).$$
(31)

Then, \(\tilde x(t)\) satisfies

$${D^\alpha}\tilde x(t) = {A_o}\tilde x(t) + {\tilde b_p}(t)u(t) - {\tilde a_p}(t)[y(t) + v(t)]$$
(32)

where \({{\tilde a}_p}(t) = {a_p} - {{\hat a}_p}(t)\) and \({{\tilde b}_p}(t) = {b_p} - {{\hat b}_p}(t)\) are the parameter errors.

Since \(\theta - \hat \theta (t) \rightarrow 0\) as t → ∞, it follows that \({{\tilde a}_p}(t) \rightarrow 0\) and \({{\tilde b}_p}(t) \rightarrow 0\). Since u(t),y(t) and v(t) are bounded, the error equation consists of a homogenous part which is exponentially stable, and an input which is decaying to zero. This implies that \(\tilde x(t) \rightarrow 0\) as t → ∞. □

  • Remark 6. The application range of the proposed adaptive observer should be pointed out. Such an adaptive observer applies to the linear fractional order systems where the unknown parameters appear linearly in the dynamics. To estimate the unknown parameters precisely, the inputs to the systems must satisfy the conditions of persistent excitation.

4 Numerical examples

Numerical examples illustrated in this section are implemented via the piecewise numerical approximation algorithm. To get more information about the algorithm, one can refer to [25]. For all the numerical examples, consider the following LTI SISO fractional order plant

$$G(s) = {{{s^{0.6}} + 2} \over {{s^{1.2}} + 5{s^{0.6}} + 10}}.$$

The pseudo state space realization of observable canonical form is

ThestablematrixAo is selected as

The variance of the white noise v(t) is selected as

$$\sigma _v^2 = 0.000\;1.$$

.

The related initial conditions are selected as

The multi-sine signal is selected as the system input u (t), which is shown in Fig. 1.

  • Example 1. Construct an adaptive observer with the parameter identification based on the regression-like form (16). Select the stable filter

    $$F({p^\alpha}) = {p^{1.2}} + 4{p^{0.6}} + 4.$$

    Then, one has the simulation results shown in Figs. 24.

  • Example 2. Construct an adaptive observer with the parameter identification based on the regression-like form (19). Select the generalized Poisson moment functional

    $${1 \over {H({p^\alpha})}} = {\left({{1 \over {{p^{0.6}} + 2}}} \right)^2}.$$

    Then, one has the simulation results shown in Figs. 57.

Fig. 1
figure 1

System input u(t) in Examples 1 and 2

Fig. 2
figure 2

Identification of parameter a p in Example 1

Fig. 3
figure 3

Identification of parameter b p in Example 1

Fig. 4
figure 4

Observation of system state x(t) in Example 1

Fig. 5
figure 5

Identification of parameter a p in Example 2

Fig. 6
figure 6

Identification of parameter b p in Example 2

Fig. 7
figure 7

Observation of system state x(t) in Example 2

It can be seen from the results of Examples 1 and 2 that both adaptive observers make the state estimations approach the actual states precisely, despite the presence of unknown system parameters and measurement noise. However, the transient responses of the adaptive observer with method 1 are better than those of the adaptive observer with method 2. The observation error of adaptive observer due to method 1 is less than that of the latter one.

5 Conclusions

In this article, a new adaptive observer design scheme is proposed for LTI SISO fractional order plant. Both the state and the unknown parameters can be efficiently observed. Motivated by the noise estimation, a modified least squares solution of the parameters is developed. For the scheme of the proposed adaptive observer, the stability of the closed-loop system and the parameter convergence have been discussed in detail, with the parameters being identified rapidly and accurately. Simulation results from numerical examples are provided to demonstrate the advantages and effectiveness of the approaches proposed in this article. It is believed that the approaches provide a new avenue to solve the problem. Future research subjects will include how to extend the results to multiple-input multiple-output case and how to control the plant based on the proposed adaptive observer.