1 Introduction

1.1 Introduction I: motivation and background information

One of the two main objectives of the present research is to study the algorithmic and computational aspect of a class of nonlinear automatic control dynamical systems that are modeled by multiple valued iterative dynamical systems (MVIDS) [16]. We call such a model a multiple valued iterative dynamics model (MVID-Model) [16], and the corresponding computational algorithm, a multiple valued iterative dynamics algorithm (MVID-Algorithm) [15].

The MVID-modeling is useful characterizing and analyzing the steady state set and the long term behavior of a non-linear control dynamical system, when more traditional tools (such as Lyapunov stability theory) fail [12, 16]. A typical failure takes place when a sudden large disturbance fundamentally changes the qualitative behavior of the system [15]. Such a phenomenon is commonly referred as a bifurcation. Here, “sudden” means the discontinuity with respect to time, and “large” refers to the magnitude bigger than the bifurcation point. See, for instance, [1] or [30], for more detail on the bifurcation theory in general.

It is well known that the bifurcation often generates the singularity (discontinuity and/or non-differentiability), and such bifurcations are abundant in nature. See, for instance, [6,7,8,9, 11,12,13,14,15,16,17,18, 20, 21], for this issue in various viewpoints. Due to the singularity, most of the traditional calculus-based models and techniques no longer work as effectively, in such cases. The disturbance that causes the singularity is called the singular disturbance. A disturbed control dynamical system that includes the singular disturbance is called a singularly disturbed control dynamical system (SDCDS).

The particular topic we aim to study is the MVID model of SDCDS. The mathematical detail of this model will be elaborated in Sect. 2. Here in the introduction, we will have only a general overview through a limited visualization (Figs. 123). The illustrations in this paper are confined to 2D for easier visualization. Our MVID algorithms and the justification theorems, however, are stated for arbitrary finite dimension This is a standard treatment in control and automation theory, as in [25,26,27,28,29] and many others.

Fig. 1
figure 1

Mutually interfering examples of the steady state sets of the MVID models of SDCDS’s from [15]. The multi-fractal structure is generated by the interference of imperfectly tuned multiple controllers [15]

Fig. 2
figure 2

Examples of the steady state sets of partially optimal SDCDS’s under a non-interference condition from [19]. The non-interference condition certifies the simple fractal structure [17]

Fig. 3
figure 3

Examples of the steady state sets of fully optimal SDCDS’s under a non-interference condition from [17]. These show near-ideal cases

As exemplified by Figs. 12 and 3, the steady state set of a MVID model of a SDCDS often exhibits a fractal structure. The analyses of the models and algorithms associated to these figures will be discussed in Sect. 4. See, also, [15, 17, 20] for the elaborations in different perspectives. The emergence of the fractal structure is quite contrary to the traditional models of disturbed control dynamical system, in which case, the steady state set tend to be a single convex polytope, as exemplified by [25,26,27,28,29], for instance.

The fractal structure exemplified by Figs. 12 and 3 is a natural consequence of the singular disturbance. Indeed, similar phenomena in nature have long been known and discussed in classical literature such as [24]. For MVID models of SDCDS, the fractal structure of Figs. 12 and 3 is the result of a competition between multiple attracting points (imperfectly tuned multiple controllers) that are pushed away by (a) repelling point(s) in the interior (unstable target state). The applications of the singular disturbance closely related to this phenomenon include non-linear physics [13, 14, 18], and digital signal processing [6,7,8,9]. In this paper, we consider a similar development in control and automation theory, partly as a continuation of [15, 16, 20, 21]. See, also, [17] for a similar development in the viewpoint of an optimization research.

1.2 Introduction II: main results

This paper discusses the theory (Sects. 2 and 3) and the implementation (Sect. 4) of a class of MVID-models of SDCDS. The main contribution of this paper comes from the theory part. It consists of the justification theorems of the MVID algorithms, namely, Theorems 2.32.7 and  2.8. These theorems literally justify the computational algorithms MVID models, as we will see in the implementation stage (Sect. 4).

Theorem 2.3 was completed in [16], so it will not be re-proved here. The first original contribution of this paper is Theorem 2.7. It completes the preliminary works and the unsolved questions of [16], and provides the theoretical foundation of the numerical studies of [15] and [20]. The second main result is Theorem 2.8. It requires broader understanding on the fractal theory, and shares some key ideas with [17]. Prior reading of [17] will not be necessary, however. All technical details are established here, over the foundation of Theorem 2.7.

The implementation part of this paper is relatively minor. Its main purpose is to provide stronger motivation for proposing the MVID modeling and proving the justification theorems. Therefore, only a brief survey of analyses on known visualization algorithms will be discussed, with a due emphasis on the role of the justification theorems. Other aspects of the algorithmic research, such as optimization, will be discussed in a separate paper.

2 Multiple valued iterative dynamics (MVID) modeling

2.1 MVID model

A classical model of discrete-time non-linear disturbed control dynamical system (DCDS) can be written as follows, where the map G is called the feedback control law, and the variables \(v_k\) and \(w_k\) represent the disturbance.

$$\begin{aligned} {\left\{ \begin{array}{ll} F: (x_k, u_k, w_k) \mapsto x_{k+1}, \\ G: (x_k, v_k) \mapsto u_k. \end{array}\right. } \end{aligned}$$
(2.1)

The classical model assumes the continuity for each variable, and most of the time, negative Lyapunov exponents. It cannot handle, therefore, a sudden large disturbance, or a singular disturbance that includes the discontinuity (typically for u, v or w). Also, as pointed out in [12] and [16], the classical model (2.1) is somewhat problematic in that it includes not only unknown but also unknowable variables. If all disturbance variables, say \(u_k\), \(v_k\) and \(w_k\), could be determined, they would not have been disturbances to begin with. Partly to overcome this dilemma, and in another part to accommodate the singular disturbance, the author proposed a new approach, the multiple valued iterative dynamics model (MVID model) [10,11,12, 15, 16, 21]. Let us elaborate it as follows.

Definition 2.1

(Multiple Valued Iterative Dynamics Model [16]). Let X, Y be non-empty sets, and \({\mathscr {P}}(X)\), \({\mathscr {P}}(Y)\) be their power sets. We say a set function \(f: {\mathscr {P}}(X) \rightarrow {\mathscr {P}}(Y)\) is a multiple valued map (function) from X to Y if

$$\begin{aligned} f(S) = \bigcup \{ f(x): x \in S \}, \end{aligned}$$
(2.2)

for all \(S \subset X\). Here, f(x) is the abbreviation of \(f(\{x\})\). In particular, if \(X = Y\), we call the dynamical system on X given by the iteration of f in \({\mathscr {P}}(X)\), the multiple valued iterative dynamical system (MVIDS). If an MVIDS was used to model a disturbed control dynamical system, we call such a model, a multiple valued iterative dynamics model (MVID Model).

In a sense, the MVID modeling is a generalization of the classical model (2.1) through the iterative dynamics of \(f: x_k \mapsto \{F(x_k, G(x_k, v_k), w_k)\}\), for all possible or for all statistically meaningful disturbances captured by \(v_k\) and \(w_k\) values. See [12] for more detail on this generalization, including the proof of the validity theorem.

Remark 2.2

Besides the author’s MVID modeling, alternative set-dynamics models were developed by other researchers such as [2] and [26]. The main ideas of all these models are more or less equivalent.

2.2 Advantages of MVID model

The advantages of the MVID modeling are twofold. First, it is well equipped to handle sudden large disturbances, or singular disturbances [15, 16], which most of the traditional models of disturbed control dynamics models fail to cover. See [11, 12, 15, 16] for the author’s treatment on this issue. The singular disturbance is an important topic, because the bifurcation of the qualitative behavior of the dynamics due to the singular disturbance is a common occurrence in nonlinear physics. See [13, 14, 17, 18] for more detail on this connection.

The second advantage is the computational effectiveness. Because the MVID modeling does not add more variables, say \(u_k\), \(v_k\) and \(w_k\) as in the classical DCDS model of (2.1), the algorithm can be made simpler and the computation quicker. [15] and [16] exemplify this issue via a comparison of computation-speed. See, also, [20] for a similar issue on a slightly different system.

2.3 Maximal invariance in MVID model

The controllability and the approximate control of maximal invariance are important topics in any nonlinear control and automation theory. For instance, a closed loop system [23] of an automatic control system runs in the maximal quasi-invariant set [21,22,23]. However, the concept of maximal invariance is a good deal more complicated in MVID models, compared to that of classical DCDS models.

Theorem 2.3

(Maximal Invariant Sets and Their Properties [16]). Let X be a non-empty set, whose points we will call states. Let Y be a non-empty subset of X, which represent a set of desirable states. Finally, let \(f: {\mathscr {P}}(X) \rightarrow {\mathscr {P}}(X)\) be a multiple valued self map on X, whose iteration forms an MVIDS on X. We define the strong locally maximal quasi-invariant subset, \({\mathcal {M}}_s^+(Y)\), the weak locally maximal quasi-invariant subset, \({\mathcal {M}}_w^+(Y)\), the strong locally maximal full-invariant subset, \({\mathcal {M}}_s (Y)\), and the weak locally maximal full-invariant subset, \({\mathcal {M}}_w (Y)\) of Y as follows.

$$\begin{aligned} {\mathcal {M}}_s^+ (Y)&= \bigcup \{ S \subset Y : \, \forall x \in S, \, f(x) \subset S\}, \nonumber \\ {\mathcal {M}}_w^+ (Y)&= \bigcup \{ S \subset Y : \, \forall x \in S, \, f(x) \cap S \ne \emptyset \}, \nonumber \\ {\mathcal {M}}_s (Y)&= \bigcup \{ S \subset Y : \, \forall x \in S, \, x \in f(S) \text { and } f(x) \subset S\}, \nonumber \\ {\mathcal {M}}_w (Y)&= \bigcup \{ S \subset Y : \, \forall x \in S, \, x \in f(S) \text { and } f(x) \cap S \ne \emptyset \}. \end{aligned}$$
(2.3)

Then, \({\mathcal {M}}_s^+(Y)\) turns out to be the set of all states whose forward dynamics always stays in Y, while \({\mathcal {M}}_w^+(Y)\) is the set of all states whose forward dynamics can stay in Y all the time. Also, \({\mathcal {M}}_s(Y)\) is the steady-state set of all dynamics that must always stay within Y, while \({\mathcal {M}}_w(Y)\) is the steady-state set of all possible dynamics within Y.

Proof

See [16]. \(\square\)

Remark 2.4

All maximal invariant sets mentioned in Theorem 2.3 are the optimal bounds in terms of the set inclusion. The strong maximal invariance (both quasi and full) is important in traditional robust control problem, while the weak maximal invariance is important for inverse problems. Many other cases such as adaptive control problems fall in between these two optimal cases.

Remark 2.5

Theorem 2.3 is, in fact, a superficial combination of a number of theorems in [16], re-written to fit the purpose of this paper. In [16], the concept of the “orbit-chains” was used to characterize the maximal invariant sets. Here, most of the in-depth descriptions not directly relevant to our main topic were snipped out for brevity.

2.4 The controllability theorem

Now, let us turn ourselves to the controllability / reachability theorem, the main contribution of this paper. First, we begin with the following definition, which upgrades similar concepts of [16].

Definition 2.6

(Restricted Predecessor/Successor Operators). Let X, Y and f be as in Theorem 2.3. We define the set functions \((f_s|_Y)^{-1} : {\mathscr {P}}(X) \rightarrow {\mathscr {P}}(Y)\) and \((f_w|_Y)^{-1} : {\mathscr {P}}(X) \rightarrow {\mathscr {P}}(Y)\) as,

$$\begin{aligned} {\left\{ \begin{array}{ll} (f_s|_Y)^{-1} (S) = \{y \in Y : f(y) \subset S\}, \\ (f_w|_Y)^{-1} (S) = \{y \in Y : f(y) \cap S \ne \emptyset \}. \end{array}\right. } \end{aligned}$$
(2.4)

We call the set functions \((f_s|_Y)^{-1}\) and \((f_w|_Y)^{-1}\), the strong predecessor operator of f restricted to Y and the weak predecessor operator of f restricted to Y. Similarly, we restrict the forward application of f to Y as follows to create the set function \((f|_Y) : {\mathscr {P}}(X) \rightarrow {\mathscr {P}}(Y)\) .

$$\begin{aligned} (f|_Y) (S) = Y \cap f(S) = \{y \in Y : y \in f(S) \}. \end{aligned}$$
(2.5)

We call \(f|_Y\), the successor operator of f restricted to Y.

It turns out that \((f_w|_Y)^{-1}\) and \((f|_Y)\) are multiple valued maps (Definition 2.1), but \((f_s|_Y)^{-1}\) is not. The proof is simple and left to the readers.

If \(Y = X\), then the predecessor operators are said to be unrestricted, and we abbreviate them as \(f_s^{-1}\) and \(f_w^{-1}\), respectively. Using these abbreviations, we can re-write the equality (2.4) in the following form for all non-empty subset Y of X.

$$\begin{aligned} (f_s|_Y)^{-1} (S) = Y \cap f_s^{-1}(S) \quad \text {and} \quad (f_w|_Y)^{-1} (S) = Y \cap f_w^{-1}(S). \end{aligned}$$
(2.6)

Using the iterative dynamics of the restricted predecessor and successor operators, we can rigorously elaborate the preliminary findings of [16] as follows.

Theorem 2.7

(Controllability Theorem). In addition to the conditions of Theorem 2.3, let us assume that both f(y) and \(f_w^{-1}(y) = \{x \in X: y \in f(x)\}\) are finite. Furthermore, let

$$\begin{aligned} (f_s|_Y)^{-\infty }(S)&= \bigcap _{n = 0}^{\infty } (f_s|_Y)^{-n}(S), \nonumber \\ (f_w|_Y)^{-\infty }(S)&= \bigcap _{n = 0}^{\infty } (f_w|_Y)^{-n}(S), \nonumber \\ (f|_Y)^{\infty } (S)&= \bigcap _{n=0}^{\infty } (f|_Y)^{n} (S). \end{aligned}$$
(2.7)

Then, the following descending chain inequalities hold.

$$\begin{aligned}&Y \supset (f_s|_Y)^{-1} (Y) \supset (f_s|_Y)^{-2} (Y) \supset \cdots \supset (f_s|_Y)^{-\infty }(Y)\nonumber \\&\quad = {\mathcal {M}}_s^+ (Y), \nonumber \\&Y \supset (f_w|_Y)^{-1} (Y) \supset (f_w|_Y)^{-2} (Y) \supset \cdots \supset (f_w|_Y)^{-\infty }(Y)\nonumber \\&\quad = {\mathcal {M}}_w^+ (Y), \nonumber \\&Y \supset (f|_Y)^{1} \circ (f_s|_Y)^{-1}(Y) \supset \cdots \supset (f|_Y)^{\infty } \circ (f_s|_Y)^{-\infty }(Y) \nonumber \\&\quad = {\mathcal {M}}_s (Y), \nonumber \\&Y \supset (f|_Y)^{1} (Y) \cap (f_w|_Y)^{-1} (Y) \supset \cdots \supset (f|_Y)^{\infty } (Y) \nonumber \\&\qquad \cap (f_w|_Y)^{-\infty } (Y) = {\mathcal {M}}_w (Y). \end{aligned}$$
(2.8)

That is, all maximal invariant sets are controllable (reachable) in a countably many steps.

Proof

The proof will be presented in Sect. 3.1 of Sect. 3. \(\square\)

The finite step approximate control problem is meaningful only if the target set is reachable in a countable step. Through Theorem 2.7, we just established the the reachability of various maximal invariant sets (strong versus weak, and quasi versus full) under the finiteness condition (both f(y) and \(f_w^{-1}(y)\) are finite).

2.5 The global attractor theorem

While Theorem 2.3 established the initial set and the terminal set (steady-state set) in terms of the maximal invariance, Theorem 2.7 proved that those sets are controllable under the finiteness condition mentioned in the beginning of Theorem 2.7. However, these theorems contribute little toward the question whether the iterates of the sets (or the individual orbits of states) actually converge to the terminal set (steady state set). This is an essential step toward the full justification of the finite step numerical approximation problem, the topic for the next section.

The following theorem fills in this gap for our purpose. In fact, it completes the justification of the finite step approximation algorithm of our MVID model (Sect. 4.1). A simpler version of this theorem was done in [17]. It was for the problems on self-similar fractals and their dimension computation. For the purpose of this paper, we generalize it accordingly as follows.

Theorem 2.8

(Global Attractor Theorem). Let \(v_1, \ldots , v_N\) be points in \({\mathbb {R}}^n\) such that they form the vertices of their convex hull \(K(v_1, \ldots , v_N)\). For each \(i \in \{1, \ldots , N\}\), let \(f_i: {\mathbb {R}}^n \rightarrow {\mathbb {R}}^n\) be a contraction map with the contraction ratio (Lyapunov multiplier) \(r_i \in (0, 1)\) that fixes \(v_i \in {\mathbb {R}}^n\). That is,

$$\begin{aligned} d(f_i(x), \, f_i(y)) \le r_i \cdot d(x, \, y) \quad \text {and} \quad f_i(v_i) = v_i. \end{aligned}$$
(2.9)

Suppose further that \(f_i (K(v_1, \ldots , v_N)) \subset K(v_1, \ldots , v_N)\). Finally, let \(f: {\mathscr {P}}({\mathbb {R}}^n) \rightarrow {\mathscr {P}}({\mathbb {R}}^n)\) be a set function defined by,

$$\begin{aligned} f(S) = f_1(S) \cup \cdots \cup f_N(S). \end{aligned}$$
(2.10)

Then, every compact subset Y of \({\mathbb {R}}^n\) that contains \(K(v_1, \ldots , v_N)\) satisfies,

$$\begin{aligned} {\mathcal {M}}_s (Y) = {\mathcal {M}}_w (Y) = \bigcap _{k=0}^{\infty } f^k(Y) = \bigcap _{k=0}^{\infty } f^k ( K(v_1, \ldots , v_N)). \end{aligned}$$
(2.11)

Moreover, the set given by the above equality is the global attractor in a sense that

$$\begin{aligned} d \left( f^{m} (x), \,\, \bigcap _{k=0}^{\infty } f^k ( K(v_1, \ldots , v_N) ) \right) \rightarrow \, 0, \end{aligned}$$
(2.12)

as \(m \rightarrow \infty\), for any \(x \in {\mathbb {R}}^n\). Here, d represents the standard distance function in \({\mathbb {R}}^n\).

Proof

The proof will be presented in Sect. 3.2 of Sect. 3. \(\square\)

The following corollary is related to the interference between multiple controllers, and partly justifies the pattern formation of the interference-noise, a well known phenomenon that goes back to early 1970s [24].

Corollary 2.9

(Invariant Fractal). In addition to the conditions of Theorem 2.8, let us impose the non-interference condition given as follows.

$$\begin{aligned} i \ne j \,\,\Longrightarrow \,\, int \left( f_i (K(v_1, \ldots , v_N)) \cap f_j (K(v_1, \ldots , v_N)) \right) = \emptyset . \end{aligned}$$
(2.13)

Then, the steady state set, \(\bigcap _{k=0}^{\infty } f^k ( K(v_1, \ldots , v_N))\) has a simple fractal structure. Moreover, if each \(f_i\) is an affine or a piecewise affine contraction, then the steady state set is a self-similar fractal.

Proof

This is an immediate corollary of Theorem 2.8, in terms of the construction of fractals, multi-fractals, and self-similar fractals. The equality (2.13), which corresponds to the non-interference in signal processing and control theory, is called the open set condition [5] in fractal geometry. See, for instance, [3] or [4] for detail. \(\square\)

3 The proof of the justification theorems

This is a technical section that concerns the proof of the justification theorems, Theorems 2.7 and 2.8. Theorem 2.3 is done throughly in [16], so we will not reproduce it here.

3.1 The Proof of Theorem 2.7

We divide the proof in three parts for easier reading.

The Proof of Theorem 2.7, Part I

First, we prove the first, the second and the fourth descending chain inequalities of Theorem 2.7. From the definitions of the restricted predecessor and successor operators, we can prove that these operators preserve the set inclusion as follows. Let \(S \subset T\), and \(y \in Y\). Then,

$$\begin{aligned} y&\in (f_s|_Y)^{-1}(S) \Rightarrow f(y) \subset S \Rightarrow f(y) \\&\quad \subset T \Rightarrow y \in (f_s|_Y)^{-1}(T), \\ y&\in (f_w|_Y)^{-1}(S) \Rightarrow f(y) \cap S \ne \emptyset \Rightarrow f(y) \\&\quad \cap T \ne \emptyset \Rightarrow y \in (f_w|_Y)^{-1}(T), \\ y&\in (f|_Y)(S) \Rightarrow y \in f(S) \Rightarrow y \in f(T) \Rightarrow y \in (f|_Y)(T). \end{aligned}$$

Therefore, repeatedly applying the restricted predecessor operators \((f_s|_Y)^{-1}\), \((f_w|_Y)^{-1}\), and the restricted successor operator \((f|_Y)\) to the initial set inequalities \(Y \supset (f_s|_Y)^{-1}(Y)\), \(Y \supset (f_w|_Y)^{-1}(Y)\), and \(Y \supset (f|_Y)(Y)\), we get the following descending chain inequalities.

$$\begin{aligned} {\left\{ \begin{array}{ll} Y \supset (f_s|_Y)^{-1} (Y) \supset (f_s|_Y)^{-2} (Y) \supset \cdots \supset (f_s|_Y)^{-\infty }(Y), \\ Y \supset (f_w|_Y)^{-1} (Y) \supset (f_w|_Y)^{-2} (Y) \supset \cdots \supset (f_w|_Y)^{-\infty }(Y), \\ Y \supset (f|_Y)^{1} (Y) \supset (f|_Y)^{2} (Y) \supset \cdots \supset (f|_Y)^{\infty }(Y). \end{array}\right. } \end{aligned}$$
(3.1)

The first two chains of the set inequality (3.1) are the same as those of Theorem 2.7. Intersecting the second chain and the third chain term by term, we get the fourth chain of Theorem 2.7.

The third chain of Theorem 2.7 follows from the principal diagonal chain of the following diagram. This diagram consists of the first chain of the inequality (3.1) arranged vertically, and the iteration of the restricted successor operator arranged horizontally. Here, each arrow (\(\rightarrow\)) stands for going down a descending set inclusion (\(\supset\)).

This completes the proof of the the descending chain part of Theorem 2.7. \(\square\)

The Proof of Theorem 2.7, Part II

The descending chain part of (2.8) is done. We must prove the final equalities. First, we prove the following inclusion.

$$\begin{aligned} {\left\{ \begin{array}{ll} {\mathcal {M}}_s^+ (Y) &{}\subset \,\, (f_s|_Y)^{-\infty }(Y), \\ {\mathcal {M}}_w^+ (Y) &{}\subset \,\, (f_w|_Y)^{-\infty }(Y), \\ {\mathcal {M}}_s (Y) &{}\subset \,\, (f|_Y)^{\infty } \left[ (f_s|_Y)^{-\infty }(Y) \right] , \\ {\mathcal {M}}_w (Y) &{}\subset \,\, (f|_Y)^{\infty } (Y) \cap (f_w|_Y)^{-\infty }(Y). \end{array}\right. } \end{aligned}$$
(3.2)

As described in Theorem 2.3, \({\mathcal {M}}_s^+(Y)\) consists of the states \(x_0 \in Y\), where any forward dynamics \((x_k)_{k=0}^{\infty }\), \(x_k \in f(x_{k-1})\), \(k \in \{1, 2, 3, \ldots \}\), of \(x_0\) must stay entirely in Y, that is, \(\forall x_k \in Y\), for any choice of forward orbit \((x_k)_{k=0}^{\infty }\). Therefore, \(\forall x_1 \in f(x_0)\), \(x_1 \in Y\), thus yielding \(f(x_0) \subset Y\), or \(x_0 \in (f_s|_Y)^{-1}(Y)\). Similarly, \(\forall x_2 \in f^2(x_0) = f \left( f(x_0) \right) \subset Y\), or \(x_2 \in (f_s|_Y)^{-1}\left( (f_s|_Y)^{-1}(Y) \right) = (f_s|_Y)^{-2}(Y)\). Repeating this process, we get, \(x_0 \in (f_s|_Y)^{-k}(Y)\) for all \(k \in \{1, 2, 3, \ldots \}\), and upon taking the intersection, we get \({\mathcal {M}}_s^+(Y) \subset (f_s|_Y)^{-\infty }(Y)\).

On the other hand, \({\mathcal {M}}_w^+(Y)\) consists of the states \(x_0 \in Y\), which has at least one forward dynamics \((x_k)_{k=0}^{\infty }\), \(x_k \in f(x_{k-1})\), \(k \in \{1, 2, 3, \ldots \}\), so that \(\forall x_k \in Y\). Here, \(x_1 \in f(x_0) \cap Y\), or \(x_0 \in (f_w|_Y)^{-1}(Y)\). Similarly, \(x_2 \in f(x_1) \cap Y\). This condition, coupled with \(x_1 \in f(x_0) \cap Y\), yield \(x_0 \in (f_w|_Y)^{-2}(Y)\). Repeating this process, we get \(x_0 \in (f_w|_Y)^{-k}(Y)\) for all \(k \in \{1, 2, 3, \ldots \}\), and upon taking the intersection, we get \({\mathcal {M}}_w^+(Y) \subset (f_w|_Y)^{-\infty }(Y)\).

From the definitions of the maximal full invariance in the equality (2.3), one can clearly see that \({\mathcal {M}}_s(Y) \subset {\mathcal {M}}_s^+(Y) \subset (f_s|_Y)^{-\infty }(Y)\) and \({\mathcal {M}}_w(Y) \subset {\mathcal {M}}_w^+(Y) \subset (f_w|_Y)^{-\infty }(Y)\). Thus, from the strong/weak local invariance of \({\mathcal {M}}_s(Y)\) and \({\mathcal {M}}_w(Y)\), we get \({\mathcal {M}}_s(Y) = (f|_Y)\left( {\mathcal {M}}_s(Y)\right) = (f|_Y)^2\left( {\mathcal {M}}_s(Y)\right) = \cdots = (f|_Y)^{\infty }\left( {\mathcal {M}}_s(Y)\right) \subset (f|_Y)^{\infty } \left[ (f_s|_Y)^{-\infty }(Y) \right]\), and \({\mathcal {M}}_w(Y) \subset (f|_Y)^{\infty } \left( {\mathcal {M}}_w(Y) \right) \cap (f_w|_Y)^{-\infty }(Y) \subset (f|_Y)^{\infty } (Y) \cap (f_w|_Y)^{-\infty }(Y)\). \(\square\)

The Proof of Theorem 2.7, Part III

The final step is to prove the other direction of the inclusion. That is,

$$\begin{aligned} {\left\{ \begin{array}{ll} {\mathcal {M}}_s^+ (Y) &{}\supset \,\, (f_s|_Y)^{-\infty }(Y), \\ {\mathcal {M}}_w^+ (Y) &{}\supset \,\, (f_w|_Y)^{-\infty }(Y), \\ {\mathcal {M}}_s (Y) &{}\supset \,\, (f|_Y)^{\infty } \left[ (f_s|_Y)^{-\infty }(Y) \right] , \\ {\mathcal {M}}_w (Y) &{}\supset \,\, (f|_Y)^{\infty } (Y) \cap (f_w|_Y)^{-\infty }(Y). \end{array}\right. } \end{aligned}$$
(3.3)

Recall that we assumed the finiteness condition, that is, both f(y) and \(f_w^{-1}(y)\) are finite sets. It turns out, however, the first inclusion of (3.3) holds in general, without the finiteness condition. The proof of the other inclusions depend upon the finiteness condition.

Given \(x_0 \in (f_s|_Y)^{-\infty }(Y)\), choose any forward orbit \((x_k)_{k=0}^{\infty }\), \(x_k \in f(x_{k-1})\), \(k \in \{1, 2, 3, \ldots \}\). Then, \(\forall x_k \in Y\), because \(x_0 \in (f_s|_Y)^{-\infty }(Y) \subset (f_s|_Y)^{-k}(Y)\). Hence, \(x_0 \in {\mathcal {M}}_s^+(Y)\), by Theorem 2.3. This proves the first inclusion, \((f_s|_Y)^{-\infty }(Y) \subset {\mathcal {M}}_s^+(Y)\).

To prove the second inclusion, choose any \(y_0 \in (f_w|_Y)^{-\infty }(Y)\). That is, \(y_0 \in (f_w|_Y)^{-n}(Y)\) for all \(n \in \{1, 2, 3, \ldots \}\), or equivalently, \(x_0, x_1, \ldots , x_n \in Y\) for some forward orbit \((x_k)_{k=0}^{n}\), \(x_k \in f(x_{k-1})\), \(k \in \{1, 2, \ldots , n\}\) and \(x_0 = y_0\). This holds for all n’s, which define infinitely many forward orbits \((x_k)_{k=0}^{n}\), \(n \in \{1, 2, 3, \ldots \}\). Because \(f(y_0)\) is finite, this process cannot generate infinitely many \(x_1\)’s. Instead, a certain \(x_1\) must be shared by infinitely many forward orbit chains. Let us fix this \(x_1\) and name it \(y_1 \in Y\). Then, by this construction, there are infinitely many forward orbits of \(y_0\) that go through \(y_1\). We apply the finiteness condition (this time, \(f(y_1)\) is finite) to the orbits that begin with \(y_0\) and then \(y_1\). By the same argument, we get \(y_2 \in Y\), such that there are infinitely many forward orbits of \(y_0\) that go through \(y_0\), \(y_1\), and then, \(y_2\). Repeating this process, we get the forward orbit \((y_k)_{k=0}^{\infty }\) that stays within Y entirely. Hence, \(y_0 \in {\mathcal {M}}_w^{+}(Y)\), by Theorem 2.3, consequently yielding the second inclusion, \((f_w|_Y)^{-\infty }(Y) \subset {\mathcal {M}}_w^+(Y)\).

The fourth inclusion is based on the same idea as the above, thus relying upon the finiteness condition. Chose any \(y_0 \in (f|_Y)^{\infty }(Y) \cap (f_w|_Y)^{-\infty }(Y)\). Because \(y_0 \in (f_w|_Y)^{-\infty }(Y)\), there exists a certain forward orbit \((y_k)_{k=0}^{\infty }\) of \(y_0\), \(y_k \in f(x_{k-1})\), \(k \in \{1, 2, 3, \ldots \}\) within Y. Also, from \(y_0 \in (f|_Y)^{\infty }(Y)\), for any \(n \in \{1, 2, 3, \ldots \}\), there exists a certain \(x_{-n} \in Y\) such that the orbit \((x_k)_{k=-n}^{0}\), \(x_k \in f(x_{k-1})\), \(x_0 = y_0\) within Y. Applying the same finiteness argument backward, we get a backward orbit \((y_k)_{k \rightarrow - \infty }^{0}\) in Y, and consequently, a doubly infinite orbit \((y_k)_{k \rightarrow - \infty }^{\infty }\) in Y. Setting \(S = \{y_k\}_{k \rightarrow - \infty }^{k \rightarrow \infty }\), we can see that S includes \(y_0\) and and satisfies the condition for \({\mathcal {M}}_w (Y)\) of (2.3).

The third inclusion is based upon the fact that \((f_s|_Y)^{-\infty } (Y) = {\mathcal {M}}_s^{+}(Y)\), which we proved in the second paragraph. Because \({\mathcal {M}}_s^{+}(Y)\) is strongly quasi-invariant under \(f|_Y\), we can restrict ourselves to the iterative dynamics of the self map, \(f|_Y : {\mathcal {M}}_s^{+}(Y) \rightarrow {\mathcal {M}}_s^{+}(Y)\). Therefore, we can apply the known results for global maximal invariance studied in [21]. Since f is finite-to-one in entire X, so is \(f|_Y\). Hence, the third inclusion follows as in [21]. \(\square\)

3.2 The Proof of Theorem 2.8

This time, we divide the proof in four parts.

The Proof of Theorem 2.8, Part I

First, we prove the maximal invariance of the convex hull. Let \(K = K(v_1, \ldots , v_N)\). Because K is quasi-invariant under each \(f_i\), it is quasi-invariant under f, too. Therefore, it can be regarded as the whole-space of the iterative MVID of f. Applying this to the equality (2.6), we get

$$\begin{aligned} {\left\{ \begin{array}{ll} (f_s|_K)^{-1}(K) = K \cap f_s^{-1}(K) = K, \\ (f_w|_K)^{-1}(K) = K \cap f_w^{-1}(K) = K. \end{array}\right. } \end{aligned}$$

Repeating the iteration arbitrarily many times, we get the following conclusion.

$$\begin{aligned} (f_s|_K)^{-n}(K) = (f_w|_K)^{-n}(K) = K, \end{aligned}$$

for any \(n \in \{0, 1, 2, \ldots \}\). Finally, applying Theorem 2.7, we conclude,

$$\begin{aligned} {\mathcal {M}}_s(K) = {\mathcal {M}}_w(K) = \bigcap _{k=0}^{\infty }f^k (K). \end{aligned}$$
(3.4)

\(\square\)

The Proof of Theorem 2.8, Part II

Next, we prove the last equality of (2.11). Because \(K \subset Y\), we must have

$$\begin{aligned} \bigcap _{k=0}^{\infty } f^k ( K ) \subset \bigcap _{k=0}^{\infty } f^k (Y). \end{aligned}$$
(3.5)

Let us prove the other direction. Choose any \(y_0 \in \bigcap _{k=0}^{\infty } f^k(Y)\). Then, for any \(k \in \{1, 2, 3, \ldots \}\),

$$\begin{aligned} y_0 = f_{i_1} \circ \cdots \circ f_{i_k} (y_{-k}), \end{aligned}$$
(3.6)

for some \(i_1, \ldots , i_k \in \{1, \ldots , N\}\) and \(y_{-k} \in Y\).

Observe that \(\bigcap _{k=0}^{\infty } f^k( K )\) is a nonempty compact set, because it is a nested intersection of compact sets. Therefore, we can select an element \(x_{-k} \in \bigcap _{k=0}^{\infty } f^k( K )\). Because K is strongly invariant, as we discussed in the first part of the proof,

$$\begin{aligned} x_0 = f_{i_1} \circ \cdots \circ f_{i_k} (x_{-k}) \in \bigcap _{k=0}^{\infty } f^k( K ). \end{aligned}$$
(3.7)

Here, \(i_1, \ldots , i_k\) are the same indices as in the equality (3.6).

Let \(\rho = \max \{r_1, \ldots , r_N\} \in (0, 1)\). Then, because each \(f_i\) is a contraction with the contraction ratio \(r_i \le \rho < 1\), we get the following result from the equalities (3.6) and (3.7).

$$\begin{aligned}&d\left( y_0, \,\, \bigcap _{k=0}^{\infty } f^k( K ) \right) \\&\le d(y_0, \, x_0) \\&\le r_{i_1} d(y_{-1}, \, x_{-1}) \\&\cdots \\&\le (r_{i_1} \cdots r_{i_k}) \cdot d(y_{-k}, \, x_{-k}) \\&\le \rho ^k \cdot diam(X). \end{aligned}$$

Here, k is an arbitrary positive integer and \(0< \rho < 1\). Thus, we conclude

$$\begin{aligned} d\left( y_0, \, \, \bigcap _{k=0}^{\infty } f^k( K ) \right) = 0. \end{aligned}$$
(3.8)

Now, recall that \(\bigcap _{k=0}^{\infty } f^k( K )\) is compact. Therefore, the equality (3.8) implies,

$$\begin{aligned} y_0 \in \bigcap _{k=0}^{\infty } f^k( K ). \end{aligned}$$

Because \(y_0\) was selected arbitrarily in Y, we get the following conclusion.

$$\begin{aligned} \bigcap _{k=0}^{\infty } f^k (Y) \subset \bigcap _{k=0}^{\infty } f^k ( K ). \end{aligned}$$
(3.9)

Combining the set inequalities (3.5) and (3.9), we get the desired equality,

$$\begin{aligned} \bigcap _{k=0}^{\infty } f^k (Y) = \bigcap _{k=0}^{\infty } f^k (K). \end{aligned}$$
(3.10)

\(\square\)

The Proof of Theorem 2.8, Part III

The third part is the maximality. Because \(K \subset Y\), \({\mathcal {M}_s}(K) \subset {\mathcal {M}_s}(Y)\) and \({\mathcal {M}_w}(K) \subset {\mathcal {M}_w}(Y)\). We must prove the other direction. From Theorem 2.7, \({\mathcal {M}_s}(Y) = (f|_Y)^{\infty } \circ (f_s|_Y)^{-\infty }(Y)\) and \({\mathcal {M}_w}(Y) = (f|_Y)^{\infty }(Y) \cap (f_w|_Y)^{-\infty }(Y)\). Removing the restriction from the restricted predecessor/successor operators, we get the following ascending chains.

$$\begin{aligned} {\mathcal {M}_s}(Y)&= (f|_Y)^{\infty } \circ (f_s|_Y)^{-\infty }(Y) \subset (f|_Y)^{\infty } (Y) \\&\quad \subset (f|_{\mathbb {R}^n})^{\infty }(Y),\\ {\mathcal {M}_w}(Y)&= (f|_Y)^{\infty }(Y) \cap (f_w|_Y)^{-\infty }(Y) \subset (f|_Y)^{\infty }(Y) \\&\quad \subset (f|_{\mathbb {R}^n})^{\infty }(Y). \end{aligned}$$

Applying the equalities (2.7) and (3.10) to the last quantity, we get,

$$\begin{aligned} {\mathcal {M}_s}(Y) \subset \bigcap _{k=0}^{\infty } f^k(Y)&= \bigcap _{k=0}^{\infty } f^k( K ) = {\mathcal {M}_s}(K),\\ {\mathcal {M}_w}(Y) \subset \bigcap _{k=0}^{\infty } f^k(Y)&= \bigcap _{k=0}^{\infty } f^k( K ) = {\mathcal {M}_w}(K). \end{aligned}$$

This completes the proof of the equality (2.11). \(\square\)

The Proof of Theorem 2.8, Part IV

The final piece is the global attraction. Given \(x \in {\mathbb {R}}^n\), put \(Y = K (x, v_1, \ldots , v_N)\). Then, the desired result (2.12) follows immediately from the equality (2.11). \(\square\)

4 The implementation of the MVID algorithms

The goal of this section is to provide further motivation for our theoretical research and main theorems. We exemplify and analyze the usage of Theorems 2.7 and 2.8 in the implementation stage of the MVID algorithms.

4.1 The finite step approximate control

Let us begin with the simplest examples. Figures 4 and 5 depict the implementation of the equality (2.11). Both examples show the case when the same controllers are placed symmetrically around a repeller. They illustrate, from the left to right, \(\bigcap _{k=0}^{k=1}f^k(K)\), \(\bigcap _{k=0}^{k=2}f^k(K)\), \(\bigcap _{k=0}^{k=3}f^k(K)\) and \(\bigcap _{k=0}^{k=5}f^k(K)\). Or equivalently, as in Theorem 2.7,

$$\begin{aligned} \bigcap _{k=0}^{k=1}f^k(K) \supset \bigcap _{k=0}^{k=2}f^k(K) \supset \bigcap _{k=0}^{k=3}f^k(K) \supset \cdots \supset \bigcap _{k=0}^{k=5}f^k(K). \end{aligned}$$
(4.1)
Fig. 4
figure 4

An example of a forward iteration of an MVID Algorithm for a class of Nonlinear SDCDS Modeling (\(N = 7\))

Fig. 5
figure 5

An example of a forward iteration of an MVID Algorithm for a class of Nonlinear SDCDS Modeling (\(N = 8\))

Due to the equalities (2.11) and (2.12), we can use \(\bigcap _{k=0}^{k=n}f^k(K)\) of the descending chain (4.1) as a finite-step approximation of the maximal invariant sets, \({\mathcal {M}_s}(Y)\) or \({\mathcal {M}_w}(Y)\), for any compact set Y that includes \(K = K(v_1, \ldots , v_N)\) (Theorem 2.8). The fourth pictures (from the left) in Figs. 4 and 5 depict this finite-step approximation. Also, Figs. 1 and 2 in the introduction show similar finite-step approximations of the maximal invariant sets (only the maximal invariant sets, no descending chains).

Another implication of Theorem 2.8 is the speed of the attraction. The maximal Lyapunov multiplier of \(f_1, \ldots , f_N\) is \(\bar{r} = \max \{r_1, \ldots , r_N\}\), and the condition of Theorem 2.8 includes, \(0< \bar{r} < 1\). Hence, the order of attraction is exponential. For systems with small Lyapunov exponents as in Fig. 4 (\(\bar{r} \approx 0.307979\)) and in Fig. 5 (\(\bar{r} \approx 0.292893\)), one can get a reasonably good approximation with only a small number of iteration (\(n = 5\) for both cases). The same is true for Figs. 1 and 2.

4.2 The non-interference condition

Another important feature of of Figs. 4 and 5 is the optimality. Figures 4 and 5 illustrate the MVID models with the largest possible contraction ratio (Lyapunov multiplier) for given number of controllers, under the non-interference condition (2.13) of Corollary 2.9. See [17] for detail.

Upon relaxing the optimality condition a little, we get non-symmetric examples as in Fig. 2. The lower bound of the Lyapunov multiplier is the ratio between the diameter of the smallest subpolygon and that of the original polygon. One can clearly see that the ratio is much smaller for the examples of Fig. 2 compared to that of the symmetric cases (Figs. 4 and  5). See [17] for the proof.

What if we go further and lift the non-interference condition (2.13) too? The consequence is illustrated in Fig.1 as the overlapping subpolygons. In this case, the optimization of Lyapunov multiplier is meaningless. With no concern on the overlap and/or competitions from the nearby controllers, one can take the Lyapunov multipliers as large as desired, at worst even \(\bar{r} > 1\). Clearly, this is not a realistic model of well-functioning automatic control, and its failure justifies the relevance of the non-interference condition. See [15] for detail.

4.3 The quasi-invariance condition

One imperfection of this paper is the quasi-invariance condition of the convex hull \(K(v_1, \ldots , v_N)\) in Theorem 2.8. That is,

$$\begin{aligned} f_i(K(v_1, \ldots , v_N) \subset K(v_1, \ldots , v_N), \qquad i \in \{1, \ldots , N\}. \end{aligned}$$
(4.2)

For the purpose of the MVID modeling that we had been considering so far, this restriction was appropriate. Figure 6 depict the justification. The first picture illustrates the effect of the disturbance variables \(v_k\) and \(w_k\) of the classical DCDS model (2.1), when they are applied as white noise. The second picture shows the aforementioned white noise being captured within a MVID model with slightly weaker contraction, or slightly higher Lyapunov multiplier. See [16] for detail.

Fig. 6
figure 6

Non-singular disturbance within MVID model [16]

Figure 7, on the other hand, depict a fundamentally different example. Here, the disturbance includes a significant and consistent rotation component, and thus the quasi-invariance condition (4.2) fails to apply. Note that the forward iterates of convex hull K are not bounded within K in Fig. 7. The second and the third pictures of Fig. 3 illustrate the same situation, only simpler and more optimal.

Fig. 7
figure 7

An example of a forward iteration of an MVID Algorithm for a class of Nonlinear SDCDS Modeling, when a rotation component is added

As shown in Figs. 3 and  7, this change creates an interesting twist to our problem that is worth studying as a future project. The fractal structure of the limit set is not too difficult to characterize. As for the possible maximal invariance and the controllability, I do not have complete solution as of this moment. I do conjecture affirmatively, and plan to develop the current research beyond this restriction.

5 Conclusion

Through this paper, we studied the theory and implementation of MVID modeling for a class of simple singularly disturbed nonlinear control dynamical systems. The original contribution of this paper was mainly on the theory part. We presented three justification theorems for our algorithmic research and the finite step approximate control of the steady state sets (terminal sets). The first (Theorem 2.3) was a summary of results from the author’s earlier publication [16]. The second (Theorem 2.7) and the third (Theorem 2.8) were new, original contributions of this paper.

In terms of the implementation, we focused upon two aspects. The visualization (Figs. 123457), and the justification (Sect. 4.1). We discussed, also, the role of the non-interference condition (2.13) of Corollary 2.9, and its connection to the optimization of Lyapunov multipliers.

There is still some work to be done beyond the present research. As mentioned in Sect. 4.3, one may be able to generalize Theorem 2.8 further by going beyond the quasi-invariance condition (4.2). Numerical experiments support this conjecture (Fig. 7, for instance). For now, this issue is left as a possible future research project.