Abstract
We show that an invariant measure of a Markov operator that is contracting in the Hutchison distance, acting on a space of Borel probability measures on a Polish space, depends continuously on the given operator. In addition, we establish an estimate for a distance between invariant measures. Some applications to the weak limit of iterates of random-valued functions (in particular, the so-called random affine maps, occuring, e.g., in perpetuities analysis) are also given.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Markov operators play important role for analyzing random dynamical systems (shortly RDS, see [4, Chapter 1] for precise definition) including such systems with random perturbations. Randomness of RDS allows us to consider movements of points via the evolution of probability measures describing the distribution of points on the state space X being a metric space, in general. Consequently, we can define a (linear) transformation P acting on the space \({\mathcal {M}}_1(X)\) of all Borel probability measures on X into itself. A probability measure \(\mu _*\) such that \(P\mu _*=\mu _*\) is said to be an invariant measure or stationary distribution with respect to P. In case of compactness of X, the space \({\mathcal {M}}_1(X)\) is compact with respect to the weak topology of measures (formally \(\hbox {weak}^*\) topology, since the space of measures is identified, via Riesz representation, with the dual of some space of continuous functions), which follows from the tightness of \({\mathcal {M}}_1(X)\), due to the Prokhorov theorem, and we can use the Markov–Kakutani fixed point theorem or the Krylov-Bogolyubov theorem to obtain an invariant measure for any continuous operator \(P:{\mathcal {M}}_1(X)\rightarrow {\mathcal {M}}_1(X)\). The existence of an invariant measure for RDS has many important consequences, for example, in ergodic theory. There are many conditions which imply the uniqueness as well as ergodicity of invariant measures. For a fuller context we refer the reader to [24, 31] and the references given there.
Let \({\varvec{\Phi }}\) be a family of Markov operators and assume that \(\mu ^*_P\) denotes a (unique) invariant measure for \(P\in {\varvec{\Phi }}\). By changing (in some sense) operators \(P\in {\varvec{\Phi }}\) it is of interest to know whether \(\mu ^*_P\) is changing continuously. Thus we shall consider an operator
and its continuity. Moreover, the aim of the paper is to establish some estimation of a distance between stationary distributions \(\mu ^*_P,\mu ^*_Q\) of operators P and Q, respectively. Our research problem is related to the study of strongly stable Markov chains, which were first examined by Aïssani and Kartashov [1]. (The interested reader is referred to [21] for further information; cf. also [22]). However, it should be emphasized that our viewpoint sheds some new light on the problem of the stability of solutions to linear iterative functional equations. Such a problem is investigated by Baron in [7, 8]. Generalizations of his results, based on theorems of our work, will appear in a forthcoming paper written by the second author.
Recently in [11] the continuous dependence of an invariant measure of piecewise-deterministic Markov processes has been established. Other studies of this type were conducted in [6], where it was proved that the limit in law of the sequence of iterates \(f^n\) of contracting in average random-valued functions f depends continuously in the Fortet–Mourier metric on the given function. (To be more precise see Corollary 4.1 and Remark 4.2 given below.) These iterates \(f^n\) are prototype of RDS; see Sect. 2 for details. One of the aims of the work is to obtain generalizations of the main results of [6], by introducing the Hutchinson distance in the space \(\{P\mu : P\in {{\varvec{\Phi }}},\mu \;{\text {has\;the\;first\;moment\;finite}}\}\) in case when \({{\varvec{\Phi }}}\) is a family of contractive Markov operators.
The organization of our paper goes as follows. In Sect. 2 some basic notions and facts concernig Markov operators as well as the iterates of random-valued functions are indicated. Main results in general settings are contained in Sect. 3. It will be shown that a parameterized version of the Banach fixed-point theorem allows us to get continuity of operator (1.1). However, the classical Banach fixed-point theorem will bring additional information, which is important from the applications point of view. These apllications presented in the last two sections concern random iterations and perpetuities.
2 Notions and Basic Facts
Throughout the work we assume that \((X,\varrho )\) is a Polish space, i.e. a separable and complete metric space. Let \({\mathcal {B}}(X)\) stand for the \(\sigma \)-algebra of all Borel subsets of X. By \({\mathcal {M}}_1(X)\) we denote the space of all probability measures on \({\mathcal {B}}(X)\). Let B(X) denote the space of all real-valued bounded Borel–measurable functions equipped with the supremum norm \(||\cdot ||_\infty \) and C(X) be the subspace of bounded continuous functions. As abbreviation we will write \(\int \varphi d\mu \) instead of \(\int _X\varphi d\mu \), where \(\varphi \in B(X)\) and \(\mu \in {\mathcal {M}}_1(X)\). Recall that a sequence of measures \((\mu _n)\) converges weakly to \(\mu \), if \(\int \varphi d\mu _n \xrightarrow [n\rightarrow \infty ]{}\int \varphi d\mu \) for every \(\varphi \in C(X)\). It is well known (see [13, Theorem 11.3.3]) that this convergence is metrizable by the Lévy–Prokhorov metric or by the Fortet–Mourier metric (known also as the bounded Lipschitz distance) [14]
where
Putting \(Lip_1^b(X)=Lip_1(X)\cap B(X)\) define now
Clearly, \(d_H\) is a distance function, called the Hutchinson metric [18] (also called the 1-Wasserstein distance or Kantorovich–Rubinstein distance [30]), however for some arguments it may be infinite. Moreover \(d_{FM}(\mu ,\nu )\le d_H(\mu ,\nu )\) and in the case when the space X is bounded we have \( d_H(\mu ,\nu )\le {\text {diam}}(X) \times d_{FM}(\mu ,\nu )\) for any \(\mu ,\nu \in {\mathcal {M}}_1(X)\), i.e. metrics \(d_{FM}, d_H\) are equivalent. The classical work on metrics on measures are [9, 10, 13].
Throughout this paper we shall consider a regular Markov operator \(P:{\mathcal {M}}_1(X)\rightarrow {\mathcal {M}}_1(X)\), i.e. P is a linear operator (here linearity is restricted to coefficients that are nonnegative and sum up to one only) and there exists (adjoint or dual) operator \(P^*:B(X)\rightarrow B(X)\) such that \(\int \varphi dP\mu =\int P^*\varphi d\mu \) for any \(\varphi \in B(X)\) and \(\mu \in {\mathcal {M}}_1(X)\). Moreover, if \(P^*:B(X)\rightarrow B(X)\) is a linear operator, \(P^*\textbf{1}_X=\textbf{1}_X\), \(P^*f_n\searrow 0\) for \(f_n\searrow 0\), and \(P^*\varphi \ge 0\) if \(\varphi \ge 0\), then the operator P given by \(P\mu (A)=\int P^*\textbf{1}_A (x)\mu (dx)\), \(A\in {\mathcal {B}}(X)\), is a Markov operator (with adjoint \(P^*\)). An operator P is called asymptotically stable if there exists a stationary distribution \(\mu _*\) such that
Clearly, the stationary distribution \(\mu _*\) satisfying (2.1) is unique.
Assume that \((\Omega , {\mathcal {A}}, {\mathbb {P}})\) is a probability space. A function \(f:X\times \Omega \rightarrow X\) is said to be random-valued function (shortly rv-function) if it is measurable with respect to the product \(\sigma \)-algebra \({{\mathcal {B}}(X)}\otimes {{\mathcal {A}}}\). Having an rv-function f we will examine a regular Markov operator \(P:{\mathcal {M}}_1(X)\rightarrow {\mathcal {M}}_1(X)\) defined by
One can show that P is a transition operator for a sequence of iterates of rv-functions in the sense of Baron and Kuczma [5]; cf. [12]. More precisely, \(P\pi _n(x,\cdot )=\pi _{n+1}(x,\cdot )\), where
denotes the distribution of the nth-iterate of f defined inductively as follows
for x from X and \(\omega =(\omega _1,\omega _2,\dots )\) from \(\Omega ^{\infty }\) being \(\Omega ^{{\mathbb {N}}}\). Note that \(f^n:X\times \Omega ^{\infty }\rightarrow X\) is an rv-function on the product probability space \((\Omega ^{\infty }, {{\mathcal {A}}}^{\infty }, {\mathbb {P}}^{\infty })\). More exactly, the n-th iterate \(f^n\) is \({{\mathcal {B}}(X)}\otimes {{\mathcal {A}}}_n\)-measurable, where \({{\mathcal {A}}}_n\) denotes the \(\sigma \)-algebra of all the sets of the form
with A from the product \(\sigma \)-algebra \({{\mathcal {A}}}^n\). The iterates \(f^n\) are prototype of RDS; see [4, Definition 1.1.1]. Namely, the formula \( \varphi (n,\omega ,x)=f^n(x,\omega ) \) defines a special RDS \(\varphi :{\mathbb {N}}\times \Omega ^\infty \times X\rightarrow X\) over the flow \(\theta (n):\Omega ^\infty \rightarrow \Omega ^\infty \) of iterates of the shift \( \theta (n)(\omega _1,\omega _2,\dots )=(\omega _{n+1},\omega _{n+2},\dots ),n\in {\mathbb {N}}\).
3 General Approach
As it was mentioned in Sect. 1 some kind of continuity of operator (1.1) can be obtained using a parameterized version of Banach’s fixed-point theorem for contractive maps. For the convenience of the reader we quote this claim from [16, p. 18], thus making our exposition self-contained. Recall only that a family \(\{H_\gamma :\gamma \in \Gamma \}\) of maps acting on a metric space \(({\mathcal {X}}, d)\) into itself is called \(\lambda \)-contractive, where \(0\le \lambda <1\), provided for some \(M>0\) and some \(0<\kappa \le 1\), we have
-
(i)
\(d(H_\gamma (x), H_\gamma (y))\le \lambda d(x,y)\) for all \(\gamma \in \Gamma \) and \(x,y\in {\mathcal {X}}\)
-
(ii)
\(d(H_{{\gamma }_1}(x),H_{\gamma _2}(x))\le M\big (\sigma (\gamma _1,\gamma _2)\big )^\kappa \) for all \(x\in {\mathcal {X}}\) and \(\gamma _1, \gamma _2\in \Gamma \),
where \(\sigma \) stands for a metric on \(\Gamma \). (Here and in the sequel all metric spaces are assumed implicitly to be non-empty.)
Theorem 3.1
Let \(({\mathcal {X}}, d)\) be a complete metric space, \((\Gamma ,\sigma )\) be a metric space and let \(\{H_\gamma :\gamma \in \Gamma \}\) be a family of contractive maps of \({\mathcal {X}}\) into itself, i.e. for any \(\gamma \) the map \(H_\gamma \) is Lipschitzian with contraction constant less than 1. Assume that for every \(x\in {\mathcal {X}}\) the map \( \Gamma \ni \gamma \longmapsto H_\gamma (x)\in {\mathcal {X}} \) is continuous and for every \(\gamma \in \Gamma \) let \(x_\gamma \) be the unique fixed point of \(H_\gamma \). Then the map
is continuous whenever at least one of the following two conditions holds:
-
(i)
the space \({\mathcal {X}}\) is locally compact,
-
(ii)
the family \(\{H_\gamma :\gamma \in \Gamma \}\) is \(\lambda \)-contractive.
Remark 3.1
The continuity of (3.1) in case of (i) follows from [27, Theorem 2]. To see that (ii) implies the continuity of (3.1) it is enough to observe that:
and, in consequence,
It is well known that the space \(({\mathcal {M}}_1(X),d_{FM})\) is compact if and only if X is. (Note however that the local compactness X is not inherited by the space \({\mathcal {M}}_1(X)\) in general. To see this let us consider \(X={\mathbb {R}}\) for simplicity, and fix weak neighbourhood \({\mathcal {O}}\) of the Dirac measure \(\delta _0\). Then we can find \(p,q\in (0,1)\) such that \(p+q=1\) and \(p\delta _0+q\delta _x\in {\mathcal {O}}\) for every \(x\in X\). Consequently, \({\mathcal {O}}\ni p\delta _0+q\delta _n\xrightarrow [n\rightarrow \infty ]{} p\delta _0+q\delta _{\infty }\not \in {\mathcal {M}}_1({\mathbb {R}})\), i.e. the closure of \({\mathcal {O}}\) is not compact. Cf. [26, 28] and [3, Remark 7.1.9].) According to this fact we may use part (i) of Theorem 3.1 to get directly the following result concerning some kind of continuity of operator (1.1).
Theorem 3.2
Assume that the space X is compact and T is a metric space and let \(\{P_t: t\in T\}\) be a family of Markov operators which are contractions in the Fortet–Mourier metric, i.e. for every \(t\in T\) there exists \(0\le \lambda _{P_t}<1\) such that
Assume that \(\mu ^*_{P_t}\) denotes the invariant measure for \(P_t\). If for every \(\mu \in {\mathcal {M}}_1(X)\) the operator \(T \ni t\longmapsto P_t\mu \in \mathcal ({\mathcal {M}}_1(X),d_{FM})\) is continuous, then
is also continuous.
In other theorems, we will not assume the compactness of X. Suppose now that \(\mu _0\in {\mathcal {M}}_1(X)\) is a fixed measure and define a set \({\mathcal {M}}(\mu _0)\) by
We start with a result being a slight modification of [20, Theorem 4.1], which will be needed in the proof of the main result of this section.
Proposition 3.1
Let \(P:{\mathcal {M}}_1(X)\rightarrow {\mathcal {M}}_1(X)\) be a Markov operator and assume that there exists \(\lambda \in [0,1)\) such that
where \({\mathcal {M}}(\mu _0)\) is given by (3.2). Assume moreover that
Then the operator P has a unique invariant measure \(\mu ^*\in {\mathcal {M}}(\mu _0)\). Furthermore, we have a geometric rate of convergence, i.e
If, additionally,
then the operator P is asymptotically stable.
Proof
Assume that P satisfies (3.3) with some \(\lambda \in [0,1)\). Then it is easy to check that condition (3.4) implies \(P({\mathcal {M}}(\mu _0))\subset {\mathcal {M}}(\mu _0)\). According to [20, Theorem 3.3] the metric space \(({\mathcal {M}}(\mu _0),d_H)\) is complete. The use of the classical Banach fixed point theorem shows that there is an invariant measure \(\mu ^*\in {\mathcal {M}}(\mu _0)\) and (3.5) holds.
Assume moreover (3.6). To prove asymptotic stability of P fix \(\mu \in {\mathcal {M}}_1(X)\) and \(\varepsilon >0\). Since the set \({\mathcal {M}}(\mu _0)\) is a dense subset of the space \({\mathcal {M}}_1(X)\) with the Fortet–Mourier metric (see [20, Theorem 3.3]; cf. also [25], p. 8) it follows that there is \(\nu \in {\mathcal {M}}(\mu _0)\) such that \(d_{FM}(\mu ,\nu )<\varepsilon \). By (3.5) we have \( d_H(P^n\nu ,\mu ^*)<\varepsilon \) for large enough \(n\in {\mathbb {N}}\), say \(n\ge n_0\). Then for \(n\ge n_0\) we obtain
which ends the proof. \(\square \)
Consider the family \({\varvec{\Phi }}\) of all Markov operators \(P:{\mathcal {M}}_1(X)\rightarrow {\mathcal {M}}_1(X)\) such that (3.3) holds with \(\lambda =\lambda _P\in [0,1)\), and (3.4) is fulfilled. On account of Proposition 3.1 operator \(P\in {\varvec{\Phi }}\) has an invariant measure \(\mu ^*_P\in {\mathcal {M}}(\mu _0)\), which is uniquely determined. Therefore we have an operator
and we are interested in its continuity.
Applying part (ii) of Theorem 3.1 and [20, Lemma 4.3] we obtain the continuity of operator (3.7) on suitable subsets of \({\varvec{\Phi }}\).
Theorem 3.3
Let \((T,\sigma )\) be a metric space and let \(\{P_t: t\in T\}\) be a subset of \({\varvec{\Phi }}\) for which \(\sup _{t\in T}\lambda _{P_t}<1\). Assume that for some \(M>0\) and some \(0<\kappa \le 1\) we have
for all \(\mu \in {\mathcal {M}}(\mu _0)\) and \(t_1, t_2\in T\). Then
is continuous.
In next theorems we will not assume the Hölder’s condition (3.8). Furthermore, our goal is to obtain some additional properties concerning continuity like estimation of the distance between limit distributions.
From now on, \({\mathbb {N}}_0\) denotes the set of all nonnegative integers.
Theorem 3.4
If \(P, Q\in {\varvec{\Phi }}\), then
where \(\mu ^*_P,\mu ^*_Q\) is the invariant measure for P, Q, respectively, and,
Proof
Assume that \(P, Q\in {\varvec{\Phi }}\) and fix \(\mu \in {\mathcal {M}}(\mu _0)\). Clearly, P satisfies condition (3.5), by Proposition 3.1. In particular the sequence \((P^n\mu )\) converges in metric \(d_H\) to \(\mu ^*_P\). The same concerns the sequence \((Q^n\mu )\), which tends to \(\mu ^*_Q\).
We will show that
for every \(n\in {\mathbb {N}}\). To do this it is enough to use an induction observing that if the statement (3.11) holds for some arbitrarily fixed n, then it holds for \(n+1\), namely,
Therefore by (3.11) we have
for every \(n\in {\mathbb {N}}\). By symmetry argument we obtain also
Consequently, passing to the limit we have
Taking infimum of all \(\mu \in {\mathcal {M}}(\mu _0)\) we get (3.9). \(\square \)
Remark 3.2
Note that in the proof of Proposition 3.1 as well as in the proof of Theorem 3.4 we do not use linearity of Markov operators. Theorem 3.4 can be easily reformulated to the case of a general abstract contraction mappings acting on any (complete) metric space into itself. However our goal is to analyze random iteration, thus we skip an abstract case.
Remark 3.3
Since
it follows that \(\alpha _P^Q(\mu )\) is finite for \(P, Q\in {\varvec{\Phi }}\) and \(\mu \in {\mathcal {M}}(\mu _0)\).
Relying on [6, Remark 1], we note the following.
Remark 3.4
-
(i)
Since \( d_{FM}(\mu ,\nu )\le \min \{d_{H}( \mu ,\nu ),2\}, \) we see that from (3.9) we have
$$\begin{aligned} d_{FM}(\mu ^*_P,\mu ^*_Q)\le & {} \min \left\{ \frac{1}{1-\lambda _P} \inf _{\mu \in {\mathcal {M}}(\mu _0)}{\alpha _Q^P(\mu )},\right. \nonumber \\{} & {} \left. \frac{1}{1-\lambda _Q}\inf _{\mu \in {\mathcal {M}}(\mu _0)}{\alpha _P^Q(\mu )},2\right\} , \end{aligned}$$(3.12)where \(\alpha _P^Q(\mu ), \alpha _Q^P(\mu )\), are given by (3.10).
-
(ii)
Note that the right-hand side of (3.9) and of (3.12) are optimal in the sense that if X has at least two elements, then there are \(P, Q\in {\varvec{\Phi }}\) such that a distance between \(\mu ^*_P\) and \(\mu ^*_Q\) is positive and inequality in (3.9) as well as in (3.12) must actually be an equality. Indeed, let x, y be different and define \(P,Q\in {\varvec{\Phi }}\) by \(P\mu =\delta _x, Q\mu =\delta _y\) for \(\mu \in {\mathcal {M}}_1(X)\). Then \(\lambda _P=\lambda _Q=0\), \(\mu ^*_P=\delta _x\), \(\mu ^*_Q=\delta _y\) and
$$\begin{aligned} \alpha _P^Q(\mu )=\alpha _Q^P(\mu )=d_H(\delta _x,\delta _y)=\rho (x,y),\quad d_{FM}(\mu ^*_P,\mu ^*_Q)=\min \{\rho (x,y),2\}. \end{aligned}$$
\(\square \)
Now we will show that the conditions of the above theorem can be expressed in terms of the adjoint operators. To do this let us consider the family \({\varvec{\Psi }}\) of all regular Markov operators \(P:{\mathcal {M}}_1(X)\rightarrow {\mathcal {M}}_1(X)\) with adjoint operator \(P^*:B(X)\rightarrow B(X)\) such that
with \(\lambda _P\in [0,1)\), and (3.4) is fulfilled.
Proposition 3.5
We have \({\varvec{\Psi }}\subset {{\varvec{\Phi }}}\). Moreover, if \(P\in {\varvec{\Psi }}\), then (3.6) holds and P is asymptotically stable.
Proof
Inclusion \({\varvec{\Psi }}\subset {{\varvec{\Phi }}}\) follows from [20, Lemma 4.3]. If \(P\in {\varvec{\Psi }}\) and \(\varphi \in Lip_1(X)\), \(||\varphi ||_\infty \le 1\), then \(P^*\varphi \in Lip_1(X), ||P^*\varphi ||_\infty \le 1\) and consequently we have
which shows (3.6). To finish the proof it is enough to apply Proposition 3.1. \(\square \)
Remark 3.6
An assertion concerning asymptotic stability of \(P\in {\varvec{\Psi }}\) is in fact a part of [20, Theorem 4.1] and of [25, Theorem 3.2]. Proofs of Propositions 3.1, 3.5 are based on ideas from [20].
According to Theorem 3.4 and Proposition 3.5 we have the following corollary, which ensures an estimation of the limit distributions without assumption (3.8).
Corollary 3.1
If \(P, Q\in {\varvec{\Psi }}\), then
holds, where \(\mu ^*_P,\mu ^*_Q\) is the invariant measures for P, Q, respectively, and,
4 Applications to Random-Valued Functions
Assume that \(f:X\times \Omega \rightarrow X\) is a fixed rv-function. Recall that
Note that for any fixed \(x\in X\) and \(B\in {\mathcal {B}}(X)\) the function \(\pi _1(x,\cdot )\) is a probability distribution on X and \(\pi _1(\cdot ,B)\) is a Borel-measurable function. In consequence the operator \(P^*\) given by
is adjoint to (2.2), and in addition,
for \(x\in X\) and \(B\in {\mathcal {B}}(X)\).
Following [7] we consider the set \({{\varvec{\Upsilon }}_{{\textbf{rv}}}}\) of all rv-functions \(f:X \times \Omega \rightarrow X\) such that
with a \(\lambda _f\in [0,1)\), and
for some (thus for all) \(x_0\in X\).
We say that \(f\in {{\varvec{\Upsilon }}_{{\textbf{rv}}}}\) is a kernel of P, if \(P^*\) has form (4.1). By \({{\varvec{\Upsilon }}}\) we will denote all regular Markov operators with adjoint operator \(P^*:B(X)\rightarrow B(X)\) given by (4.1) with kernel \(f\in {{\varvec{\Upsilon }}_{\textbf{rv}}}\).
In the main result of this section a role of the family \({\mathcal {M}}(\mu _0)\) given by (3.2) will be played by a family \({\mathcal {M}}_1^1(X)\) defined as
Note that the set \({\mathcal {M}}_1^1(X)\) is independent of \(x_0\). This set consists of all Borel measures on X with the first moment finite. One can show that \({\mathcal {M}}_1^1(X)={\mathcal {M}}(\delta _{x_0})\); see [20, Lemma 3.1].
Theorem 4.1
We have \({{\varvec{\Upsilon }}}\subset {\varvec{\Psi }}\). If \(P, Q \in {{\varvec{\Upsilon }}}\), then (3.14) holds, where \(\mu ^*_P,\mu ^*_Q\in {\mathcal {M}}_1^1(X)\) are the invariant measures for P, Q, respectively, and \(\alpha _P^Q(\mu )\), \(\alpha _Q^P(\mu )\) are defined by (3.15), (3.16). Moreover, if
for \(n\in {\mathbb {N}}_0\) and \(x\in X\), then for any \(\mu \in {\mathcal {M}}_1^1(X)\) we have
where \(f,g\in {{\varvec{\Upsilon }}_{\textbf{rv}}}\) is a kernel of P and Q, respectively. In particular,
where
Proof
Assume that \(f,g\in {{\varvec{\Upsilon }}_{\textbf{rv}}}\) are as in the statement and fix \(\varphi \in Lip_1^b(X)\). Observe that
i.e. (3.13) holds. Moreover,
and by (4.3) we obtain (3.4) with \(\mu =\delta _{x_0}\). Clearly, the same conclusion can be drawn for any rv-function from \({{\varvec{\Upsilon }}_{\textbf{rv}}}\), which shows that \({{\varvec{\Upsilon }}}\subset {\varvec{\Psi }}\).
Due to Corollary 3.1 we have (3.14), where \(\lambda _P=\lambda _f, \lambda _Q=\lambda _g\). In addition, an easy calculation shows that
hence
i.e. (4.4) is fulfilled.
It is obvious that for any \(x\in X\) we have
hence,
Since roles f and g are symmetrical, the proof is finished. \(\square \)
To illustrate our theorem, let us state a corollary, which is the main result of [6]. Recall only that by the convergence in distribution or in law of the sequence of iterates \((f^n(x, \cdot ))\) we mean that the sequence \((\pi _n(x,\cdot ))\) converges weakly to a probability distribution.
Corollary 4.1
If \(f,g\in {{\varvec{\Upsilon }}_{\textbf{rv}}}\), then the sequences of iterates \((f^n(x,\cdot ))\), \((g^n(x,\cdot ))\) are convergent in law to the probability distributions \(\pi ^f, \pi ^g\in {\mathcal {M}}_1^1(X)\), respectively, the limits do not depend on \(x\in X\), and
Proof
Note that under above notation operators P, Q with kernels f, g are asymptotically stable and according to (4.2) the sequences of iterates \((f^n(x,\cdot ))\), \((g^n(x,\cdot ))\) converge in law to probability distributions \(\pi ^f\), \(\pi ^g\), which are in fact invariant measures of P, Q, respectively. Moreover, according to (4.6) for any \(y\in X\) we have
In consequence, by Theorem 4.1 we have (4.7). \(\square \)
Remark 4.2
-
(i)
Originally in [6], instead of metric \(d_{H}\), there was \(d_{FM}\). This has been strengthened in the paper [7] to the Hutchinson metric in a vector case and the fact that distributions \(\pi ^f, \pi ^g\) belong to \({\mathcal {M}}_1^1(X)\).
-
(ii)
As it was shown in [7, 8] there are many applications of Corollary 4.1. In next two sections we enlarge possible applications of Theorem 4.1 (cf. Remark 5.1).
5 Applications to Random Affine Maps and Perpetuities
Assume now that \((X,||\cdot ||)\) is a separable Banach space. An important class of rv-functions is a family of so called random affine maps. Following [15] an rv-function \(f:X\times \Omega \rightarrow X\) is said to be a random affine map if it has a form
for some random variables \(\xi _f:\Omega \rightarrow {\mathbb {R}}, \eta _f:\Omega \rightarrow X\). We normally omit the argument \(\omega \) and just write \(f(x,\cdot )=\xi _f x+\eta _f\). From now on we will use the symbols \({\mathbb {E}}\) and \({\mathbb {E}}_\infty \) to denote expectations with respect to \({\mathbb {P}}\) and \({\mathbb {P}}^\infty \), respectively. For any random variable \(\zeta :\Delta \rightarrow X\) on a probability space \((\Delta , \nu )\) the symbol \({\mathcal {L}}(\zeta )\) will be reserved for the distribution of \(\zeta \), i.e. \({\mathcal {L}}(\zeta )(B)=\nu (\zeta \in B)\) for \(B\in {\mathcal {B}}(X)\).
Denote by \({{\varvec{\Upsilon }}_{{\textbf{aff}}}}\) the set of all random affine maps of the form (5.1) such that
Clearly, \({{\varvec{\Upsilon }}_{{\textbf{aff}}}}\) is included in \({{\varvec{\Upsilon }}_{{\textbf{rv}}}}\).
Remark 5.1
Suppose that X is unbounded. Let \(f(x,\cdot )=\xi _f x+\eta _f\) and \(g(x,\cdot )=\eta _g x+\eta _g\) with integrable random variables \(\xi _f, \eta _f, \xi _g, \eta _g\). Then, assuming \({\mathbb {P}}(\xi _f\not = \xi _g)>0\), we see that
Therefore
if and only if \(\xi _f=\xi _g\) a.s. \(\square \)
Above remark shows that Corollary 4.1 is useful for random affine maps f, g only in the case, where \(\xi _f=\xi _g\) with probability 1. Our goal is to omit this restriction by the use of Theorem 4.1. The main result of this section reads as follows.
Theorem 5.1
If \(f,g\in {{\varvec{\Upsilon }}_{\textbf{aff}}}\), then the sequences of iterates \((f^n(x,\cdot )), (g^n(x,\cdot ))\) are convergent in law to some probability distributions \(\pi ^f, \pi ^g\in {\mathcal {M}}_1^1(X)\), respectively, the limits do not depend on \(x\in X\), and
with
Proof
Fix \(f,g\in {{\varvec{\Upsilon }}_{\textbf{aff}}}\). On account of Corollary 4.1 the sequences \((f^n(x,\cdot ))\), \((g^n(x,\cdot ))\) are convergent in law to \(\pi ^f, \pi ^g\in {\mathcal {M}}_1^1(X)\), respectively. (Moreover these probability distributions are attractive in geometric rate on \({\mathcal {M}}_1^1(X)\); see Propositions 3.1 and 3.5.)
Let \(\xi _n,\eta _n\) be random variables on \(\Omega ^\infty \) defined as
for \(n\in {\mathbb {N}}\). Obviously, \({\mathcal {L}}(\xi _n)={\mathcal {L}}(\xi _g)\) and \({\mathcal {L}}(\eta _n)={\mathcal {L}}(\eta _g)\), hence
By easy induction we have
and, in particular
for \(n\in {\mathbb {N}}\). Then
for \(n\in {\mathbb {N}}, \omega \in \Omega ^\infty \) and \(\varpi \in \Omega \). Since \(\eta _j, \xi _{j+1},\dots , \xi _n\) are independent, from the above and (5.4) it follows that
for \(n\in {\mathbb {N}}\); it is also true for \(n=0\), since
By definition (4.6) this implies that
Of course the same inequality holds if we replace the subscripts g by f and the superscripts f by g. Combining these two inequalities and applying estimation (4.5) from Theorem 4.1 with \(\lambda _f={\mathbb {E}}|\xi _f|\) and \(\lambda _g={\mathbb {E}}|\xi _g|\), we obtain (5.3). \(\square \)
As we have seen in the preliminary section the sequence of iterates \((f^n(x,\cdot ))\) is of forward (or inner) iteration type. Let us consider an operator \(\sigma _n:\Omega ^\infty \rightarrow \Omega ^\infty \) given by
The sequence \(\big (f^n(x,\sigma _n(\cdot ))\big )\) forms backward (known also as outer) iterations and probability distribution of \(f^n(x,\sigma _n(\cdot ))\) coincides with \(\pi _n(x,\cdot )\), because operator \(\sigma _n\) preserves probability \({\mathbb {P}}^\infty \). The advantage of using backward iterations lies in the fact that they may converge almost sure to a limit. In the case of random affine map (5.1) we have
where \(\xi _n(\omega )=\xi _f(\omega _n)\), \(\eta _n(\omega )=\eta _f(\omega _n)\) for \(\omega =(\omega _1,\omega _2,\dots )\in \Omega ^\infty \). It can be shown that the sequence \((\prod _{k=1}^n \xi _k)_{n\in {\mathbb {N}}}\) converges a.s. to zero by the Kolmogorov strong law of large numbers, provided \(-\infty<{\mathbb {E}}\log |{\xi _f}|<0\) in case \({\mathbb {P}}(\xi _f=0)=0\). If, additionally \({\mathbb {E}}\log \max \{||\eta _f||,1\}<+\infty \), then the sequence \(\big (\sum _{k=1}^n\eta _k\prod _{j=1}^{k-1}\xi _j\big )_{n\in {\mathbb {N}}}\) converges absolutely a.s., by the application of [17, Theorem 2] (cf. [15, Corollary 2.14]) to the i.i.d. sequence \((\xi _n,||\eta _n||)\). The limit is the sum of the series
which is (in the case \(X={\mathbb {R}}\)) the probabilistic formulation of the actuarial notion of a perpetuity. For more details and interesting examples of perpetuities we refer the reader to [19]; papers [15, 17, 23, 29] provide complete characterizations of its a.s. convergence.
Since distribution of (5.5) coincides with \(\pi _f\) we will be able to use Theorem 5.1 to yield that perpetuities change continuously (of course, assumptions like (5.2) are needed). So let us denote by \({\varvec{\Sigma }}\) a family of all pairs of random variables \(\xi :\Omega \rightarrow {\mathbb {R}}, \eta :\Omega \rightarrow X\) such that
We henceforth identify \((\xi ,\eta )\in {\varvec{\Sigma }}\) with the series (5.5), where \((\xi _n,\eta _n), n\in {\mathbb {N}},\) is a sequence of independent random variables (on an arbitrary probability space, not necessary on the product \(\Omega ^\infty \)), identically distributed as \((\xi ,\eta )\).
Suppose that \((\xi ,\eta )\in {\varvec{\Sigma }}\). Let \({\mathbb {P}}(\xi =0)=0\) holds. Assuming integrability of \(\log |{\xi }|\) by the Jensen inequality we obtain \( {\mathbb {E}}\log |{\xi }|\le \log {\mathbb {E}}|\xi |<0. \) Moreover,
Summarizing, we have \(-\infty<{\mathbb {E}}\log |{\xi }|<0, {\mathbb {E}}\log {\max \left\{ ||\eta ||,1\right\} }<\infty \) and due to perpetuity convergence theorem series (5.5) converges a.s. However, it turns out that to get the convergence of (5.5) it is enough to assume (5.6) without any additional assumptions like \({\mathbb {P}}(\xi =0)=0\) or integrability of \(\log |{\xi }|\). Namely, we have the following lemma, which gives more information.
Lemma 5.1
Assume that \((\xi ,\eta )\in {\varvec{\Sigma }}\). Let \((\xi _n,\eta _n), n\in {\mathbb {N}},\) be a sequence of independent random variables on an arbitrary probability space, say on \((\Omega , {\mathcal {A}}, {\mathbb {P}})\), identically distributed as \((\xi ,\eta )\). Then
In particular series (5.5) converges absolutely a.s.
Proof
Obviously, \(\eta _{k+1}, \xi _{1},\dots , \xi _k\) are independent, \({\mathbb {E}}|\xi _n|={\mathbb {E}}|\xi |\) and \({\mathbb {E}}||\eta _n||={\mathbb {E}}||\eta ||\) for \(k,n\in {\mathbb {N}}\). Using the Lebesgue monotone convergence theorem we conclude that
i.e. (5.7) holds. Therefore \(\sum _{n=1}^\infty ||\eta _n||\prod _{k=1}^{n-1}|\xi _k|<\infty \) a.s. \(\square \)
It is worth pointing out that under some additional assumptions condition (5.7) is equivalent to (5.6) by [2, Theorem 1.4].
Applying Theorem 5.1 we will show how the Hutchinson distance between perpetuities changes depending on the expected value of some random variables.
Proposition 5.1
If \((\xi ,\eta ), (\zeta ,\vartheta )\in {\varvec{\Sigma }}\), then
with
Proof
Observe that if \((\xi ,\eta )\in {\varvec{\Sigma }}\), then the probability distribution of series (5.5) depends only on the distributions of \(\xi \) and \(\eta \). (Obviously, this series is a.s. convergent by Lemma 5.1.) Indeed, assume that all \(\xi _n,\eta _n\) are defined on a probability space \((\Theta , {\mathbb {B}})\) and have the same distribution as \(\xi , \eta \), respectively, i.e. \( {\mathbb {B}}\circ \xi _n^{-1}={\mathbb {P}}\circ \xi ^{-1}=\mu , \quad {\mathbb {B}}\circ \eta _n^{-1}={\mathbb {P}}\circ \eta ^{-1}=\nu \) for any \(n\in {\mathbb {N}}\). Then
where \(\Phi (x_1,\dots ,x_n)=x_1\cdots x_n, B\in {\mathcal {B}}(X)\), provided \((\xi _n,\eta _n), n\in {\mathbb {N}},\) are independent. From this it follows that the distribution of series (5.5) is just the same as the distribution of perpetuity with \(\xi _n,\eta _n\) given on the product \(\Omega ^\infty \) by \( \xi _n(\omega )=\xi (\omega _n), \eta _n(\omega )=\eta (\omega _n)\). Thus we can apply Theorem 5.1 to get desired result. \(\square \)
References
Aïssani, D., Kartashov, N.V.: Ergodicity and stability of Markov chains with respect to operator topology in the space of transition kernels. Dokl. Akad. Nauk. Ukr. SSR Ser. A 11, 3–5 (1983)
Alsmeyer, G., Iksanov, A., Rösler, U.: On distributional properties of perpetuities. J. Theor. Probab. 22, 666–682 (2009)
Ambrosio, L., Gigli, N., Savaré, G.: Gradient Flows in Metric Spaces and in the Space of Probability Measures. Lectures in Mathematics. Birkhäuser, Basel (2005)
Arnold, L.: Random Dynamical Systems. Springer, Berlin (1998)
Baron, K., Kuczma, M.: Iteration of random-valued functions on the unit interval. Colloq. Math. 37, 263–269 (1977)
Baron, K.: On the continuous dependence in a problem of convergence of iterates of random-valued functions. Grazer Math. Ber. 363, 1–6 (2015)
Baron, K.: Remarks connected with the weak limit of iterates of some random-valued functions and iterative functional equations. Ann. Math. Sil. 34, 36–44 (2020)
Baron, K.: Around the weak limit of iterates of some random-valued functions. Ann. Univ. Sci. Bp. Sect. Comput. 51, 31–37 (2020)
Billingsley, P.: Probability and Measure. Wiley, New York (2012)
Bogachev, V.I.: Measure Theory, vol. I, II. Springer, Berlin (2007)
Czapla, D., Hille, S.C., Horbacz, K., Wojewódka-Ścia̧żko, H.: Continuous dependence of an invariant measure on the jump rate of a piecewise-deterministic Markov process. Math. Biosci. Eng. 17, 1059–1073 (2020)
Diamond, Ph.: A stochastic functional equation. Aequat. Math. 15, 225–233 (1977)
Dudley, R.M.: Real Analysis and Probability. Cambridge Studies in Advanced Mathematics, vol. 74. Cambridge Univerity Press, Cambridge (2002)
Fortet, R., Mourier, E.: Convergence de la répartition empirique vers la répartition théorique. Ann. Sci. Ecole Norm. Sup. 70, 267–285 (1953)
Goldie, C.M., Maller, R.A.: Stability of perpetuities. Ann. Probab. 28, 1195–1218 (2000)
Granas, A., Dugundji, J.: Fixed Point Theory. Springer, New York (2003)
Grincevičjus, A.K.: On the continuity of the distribution of a sum of dependent variables connected with independent walks on lines. Theor. Probab. Appl. 19, 163–168 (1974)
Hutchinson, J.: Fractals and self-similarity. Indiana Univ. Math. J. 30, 713–747 (1981)
Iksanov, A.: Renewal Theory for Perturbed Random Walks and Similar Processes, Probability and Its Applications. Birkhäuser, Basel (2016)
Kapica, R.: The geometric rate of convergence of random iteration in the Hutchinson distance. Aequat. Math. 93, 149–160 (2019)
Kartashov, N.V.: Strong Stable Markov Chains. TbiMC Scientific Publishers, VSP, Utrecht (1996)
Kato, T.: Perturbation Theory for Linear Operators, 2nd corr. print. of the 2nd ed. Grundlehren der mathematischen Wissenschaften, vol. 132. Springer, Berlin (1984)
Kesten, H.: Random difference equations and renewal theory for products of random matrices. Acta Math. 131, 207–248 (1973)
Lasota, A., Mackey, M.C.: Chaos, Fractals and Noise. Stochastic Aspects of Dynamics. Applied Mathematical Sciences, vol. 97. Springer, New York (1994)
Lasota, A.: From fractals to stochastic differential equations. In: Lecture Notes in Physics, vol. 457, pp. 235–255 (1995)
Luther, N.Y.: Locally compact spaces of measures. Proc. Am. Math. Soc. 25, 541–547 (1970)
Nadler, S.B., Jr.: Sequences of contractions and fixed points. Pac. J. Math. 27, 579–585 (1968)
Topsøe, F.: Compactness and tightness in a space of measures with the topology of weak convergence. Math. Scand. 34, 187–210 (1974)
Vervaat, W.: On a stochastic difference equation and a representation of non-negative infinitely divisible random variables. Adv. Appl. Prob. 11, 750–783 (1979)
Villani, C.: Optimal Transport. Old and New. Grundlehren der Mathematischen Wissenschaften, Fundamental Principles of Mathematical Sciences, vol. 338. Springer, Berlin (2009)
Zaharopol, R.: Invariant Probabilities of Transition Functions, Probability and Its Applications. Springer, Berlin (2014)
Acknowledgements
This work was supported by the Faculty of Applied Mathematics AGH UST statutory tasks within subsidy of Ministry of Science and Higher Education. The authors are greatly indebted to the reviewers for insightful and helpful comments and remarks on the manuscript.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
No potential conflict of interest was reported by the authors.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kapica, R., Komorek, D. Continuous Dependence in a Problem of Convergence of Random Iteration. Qual. Theory Dyn. Syst. 22, 65 (2023). https://doi.org/10.1007/s12346-023-00763-6
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s12346-023-00763-6
Keywords
- Random iteration
- Markov operator
- Invariant measure
- Hutchinson distance
- Fortet–Mourier distance
- Continuous dependence on the given function
- Perpetuities