Symbolically Quantifying Response Time in Stochastic Models Using Moments and Semirings
Abstract
We study quantitative properties of the response time in stochastic models. For instance, we are interested in quantifying bounds such that a high percentage of the runs answers a query within these bounds. To study such problems, computing probabilities on a statespace blownup by a factor depending on the bound could be used, but this solution is not satisfactory when the bound is large.
In this paper, we propose a new symbolic method to quantify bounds on the response time, using the moments of the distribution of simple stochastic systems. We prove that the distribution (and hence the bounds) is uniquely defined given its moments. We provide optimal bounds for the response time over all distributions having a pair of these moments. We explain how to symbolically compute in polynomial time any moment of the distribution of response times using adequatelydefined semirings. This allows us to compute optimal bounds in parametric models and to reduce complexity for computing optimal bounds in hierarchical models.
1 Introduction
Response time has been considered lately as an important property of systems [8, 15, 21]. In this context, one does not simply want a query to be answered eventually, but to be answered in a reasonable amount of time. In the modelchecking community, problems on response time have been studied mainly qualitatively, in the context of (pure, that is non stochastic) twoplayer games [8, 21]. There, one looks for a strategy ensuring that the limsup of response time is finite. It ensures that under this strategy, there will be a bound on the response time to any query. This has been extended in [15] to a quantitative setting, where one wants to optimize the mean response time in a pure twoplayer game.
In this paper, we consider stochastic systems. In such systems, the response time is a random variable, unlikely to be bounded as even a single probabilistic loop on a reachable state will make the response time longer than T for a set of runs of small but positive probability, no matter T. Instead, we propose to quantify such response times. One way to do that is to obtain the distribution of response times. Another way is to compute, for a probability \(0<p<1\), the bound T that is satisfied (by a set of runs) with probability at least \(1p\). In this paper, we tackle both problems. For that, we use the concept of moments of the distribution of response times, as described next.
The moment of order r of a probability distribution \(\delta \) over \(\mathbb {R}\) or \(\mathbb {R}^+\) is defined as the integral of \(x^r \delta (x)\) over the support of \(\delta \), when defined (that is if \(x^r \delta (x)\) is measurable and the integral is defined). For instance, the moment of order 1 is the expected value of \(\delta \), while the moment of order 2 allows one to compute the standard deviation of \(\delta \). Inspired by the computation of entropy for automata [10] (see also [1] for the computation of entropy for (nonZeno) timedautomata), we design new semirings in which each moment corresponds to the sum of weights of runs reaching a state. This construction can be applied to probabilistic automata (that is, labeled discrete time Markov chains), as well as labeled continuous time Markov chains, where time is continuous and is drawn according to some rate. Adapting the FloydWarshall algorithm provides a symbolic way to perform the computation of the n first moments in time cubic in the number of states of the Markov Chain, and quadratic in n. For any n, we can thus compute the value of the first n moments. In some sense, we extend the approach of [12, 16] from computing probabilities to computing any moments. This allows us to evaluate the distribution of response times in two ways:
Firstly, thanks to the symbolic expression of moments, we prove that there is a unique distribution having the moments of a distribution of response times of a probabilistic automaton. We can then build a sequence of distributions matching the first n moments, for instance the maximal entropy one [11]. Here, maximal entropy means assuming the least information besides these moments. This sequence of distributions is then ensured to converge in law towards the distribution of response times.
Secondly, we study optimal symbolic bounds on the time to answer a high percentage of queries, obtained from moments. The Tchebychev inequality provides optimal symbolic bounds when considering the space of distributions having one given moment, of any order i. We obtain bounds optimal in the space of distributions having two given moments, of any orders i, j. We show how this improves Tchebychev bounds on some example. Having symbolic methods allows for instance to deal with parametric systems where the parameters represent uncertain probabilities. In this case, we can compute optimal bounds satisfying all valuations of parameters. For hierarchical systems [3], which are compact representations of large systems, our symbolic method allows to design a much more efficient algorithm (e.g. it does not consider twice the same component) to compute the moments, and thus the bounds. Missing proofs can be found in [5].
Related Work: Response times in stochastic systems have been studied for a long time by the perf.eval. community under the name “first passage times”, e.g. in [22]. Techniques used in this community to compute moments of Markov chains are mostly based on numerical methods, e.g. [13]. While [13] has the same complexity as our symbolic technique, it is very efficient on explicit models. However, these numerical methods are less adaptable than our symbolic algorithm, in particular concerning parametric or hierarchical systems.
Concerning the determinacy of the distribution given moments, it is known [20] that phasetype distributions of order n are determined by their first \(2n1\) moments. First passage distribution time in Markov chains with n states are phase type distribution of order n. However, [20] does not help characterizing bounds as it does not ensure that a nonphase type distribution cannot have the exact same moments as a phase type distribution, unlike our result.
Bounding the response time has also been studied in the perf.eval. community. Again, methods used there are mostly numerical [6, 19]. In [19] (pp. 68–69), a symbolic bound is also provided in the particular case of moments of order 1, 2 and 3. In [2], it is shown how to use the two first moments of response time across various components to compute general bounds, using techniques close to ours, but restricted to moments of order 1 and 2. In our paper, we provide optimal bounds for any order \((i,j) \in \mathbb {N}^2\). Taking into account moments of order \(i,j>3\) is important when the proportion of runs to answer is close to 1.
Last, computing moments find other applications. For instance, in [4, 7, 14], complex functions describing the evolution of molecular species are approximated using the first k moments, for some k.
2 Probabilistic Automata
We first introduce a simple class of models, namely probabilistic automata (also called labeled discrete time Markov chains), on which we can demonstrate our techniques. Later, we will extend our results to handle continuous time, considering ContinuousTime Markov Chains (CTMC), as well as parametric and hierarchical systems.
Definition 1

S is a finite set of states,

\(Pr: S \times \varSigma \times S \rightarrow [0,1]\) is a stochastic transition function such that for all \(s\in S\), Open image in new window : the weights of paths leaving s sum to 1,

\(\delta _0: S \rightarrow [0,1]\) is the initial distribution over states such that Open image in new window .
Example 1
For instance, the model depicted on Fig. 1 is a probabilistic automaton with 3 states \(\{1,2,3\}\). There is a transition between 1 and 2 labeled query with probability 1. From state 2, with probability .9 we stay in state 2 with a transition labeled wait, and with probability .1 we go to state 3 with a transition labeled response. We loop in state 3 with probability 1.
A finite sequence \(\pi = s_0,a_1,s_1, \ldots , a_n, s_n \in (S \varSigma )^n S\) is called a finite path starting from \(s_0\) and ending in \(s_n\), and a transition \(t\in \pi \) if \(t=s_i a_{i+1}s_{i+1}\) for some i. We denote \(\pi =n\) the length of the path \(\pi \). For a path \(\pi _1\) ending in \(s_n\) and a path \(\pi _2\) starting from \(s_n\), we can define the concatenated path \(\pi _1 \cdot \pi _2\) where the last node of \(\pi _1\) and the first node of \(\pi _2\) are merged. A path \(\pi _1\) is a prefix of \(\pi \) if there exists a path \(\pi _2\) such that \(\pi _1 \cdot \pi _2 = \pi \).
For a path \(\pi \) starting in a state \(s_0\), we define Open image in new window the probability that a path with prefix \(\pi \) is executed from \(s_0\). A path \(\pi \) is realizable if \(\mathbb {P}(\pi )>0\).
Let s be a state, and \(\varPi \) be a set of finite paths starting from s such that no path in \(\varPi \) is a prefix of another path in \(\varPi \). Then the probability that a path starting from s has a prefix in \(\varPi \) is Open image in new window . We say that \(\varPi \) is disjoint if no path \(\rho \) of \(\varPi \) is a prefix of another path \(\rho ' \ne \rho \) of \(\varPi \) or similarly, \(Cyl(\rho )\cap Cyl(\rho ')=\emptyset \) with \(Cyl(\rho )=\{\pi , \rho \text { prefix of }\pi \}\).
Some labels of an automaton will be of particular interest concerning response time. Let \(\varSigma _Q \subseteq \varSigma \) be a subset of labels standing for queries, and \(\varSigma _R \subseteq \varSigma \) be a subset of labels standing for responses. For simplicity, we will assume that there is a unique query type \(\varSigma _Q=\{q\}\) and a unique response type \(\varSigma _R=\{r\}\), with \(q \ne r\). We will also assume that there is no path with two (similar) queries q. To handle cases with several query/response types, it suffices for each type to consider only queries and answers of that type and disregard other types.
Problem Statement: We are interested in quantifying the time between queries and responses, called the response time, which is a random variable. A way to quantify it is to produce the distribution of response times, either for each transition labeled by a query, or averaged on these transitions, weighted by the probability to see each of these transitions. Another way is to answer modelchecking questions such as: what is the smallest delay T such that the mass of paths unanswered after T units of time is smaller than some probability p?
To compute both the distribution and the delay T, we will use the so called moments of the distribution of response times. The moment of order 1 is the mean value, and the moment of order 2 allows to compute the standard deviation.
3 Symbolically Computing Moments Using Semirings
In this section, we define moments and explain how to compute them symbolically using appropriatelydefined semirings.
Let X be the random variable of the response time. If all queries are answered, then X takes values in Open image in new window , else X takes values in Open image in new window . Let p(x) be the probability that the response is obtained x units of time after the query, that is, the probability that \(X=x\). Variable p is a distribution over response time, with \(\sum _{x } p(x)=1\).
Definition 2
For \(p: \mathbb {N} \rightarrow [0,1]\) and \(n \in \mathbb {N}\), we define the nth moment of p by \(\sum _{x \in \mathbb {N}} p(x) \cdot x^n=E(X^n)\), that is the expected value of \(X^n\).
3.1 Semirings Associated with Moments
We will compute moments of the distribution of response times by considering each query individually. We can then take e.g. the average over all queries (as we assumed that there are no two queries on the same path). Thus, we first fix a state q, target of a transition labeled by a query. State q symbolizes that a query has just been asked. We then let R be the set of target states of transitions labeled by a response. A state is in R if a response to this query has just been given. For instance, on Fig. 1, we have \(q=2\) and \(R=\{3\}\).
We introduce a set of semirings that will allow us to compute symbolically the moment of order n of the distribution of response times to the query associated with state q, for all \(n \in \mathbb {N}\). We will compute the moment inductively on a disjoint subset \(\varPi \) of paths of A from q to R. For an integer n, we denote Open image in new window . Let \(\mathbf {Path}_q^R\) be the set of paths in the automaton A between q and the first occurrence of R. Notice that \(\mathbf {Path}_q^R\) is disjoint. Thus, we have that \(\mu _n(\mathbf {Path}_q^R)\) is the moment of order n of the distribution of response times to the query associated with state q. To avoid some heavy notations, when R is reduced to one state t, let \(\mu _n(\mathbf {Path}_s^t)\) be the set of paths between s to the first occurrence of t and we denote \(\mu _n(s,t)=\mu _n(\mathbf {Path}_s^t)\).
We now give some properties of \(\mu \). Let \(\varPi _1\) be a set of paths ending in some state s and let \(\varPi _2\) be a set of paths starting from s. We denote by \(\varPi _1 \cdot \varPi _2\) the set of paths \(\rho _1 \rho _2\) with \(\rho _1 \in \varPi _1\) and \(\rho _2 \in \varPi _2\).
Proposition 1
For all n, we have Open image in new window

\((x_0,\dots ,x_n)\oplus _n (y_0,\dots ,y_n) = (x_0+y_0,\dots , x_n+y_n)\)

\((x_0,\dots ,x_n)\otimes _n (y_0,\dots ,y_n) = (z_0,\dots , z_n)\) with Open image in new window
The neutral element for \(\oplus _n\) is \(\overline{0}_n =(0,\dots ,0)\). \(\overline{0}_n\) is an annihilator for \(\otimes _n\). The neutral element for \(\otimes _n\) is \(\overline{1}_n = (1,0,\dots , 0)\). In the following, we will denote the different laws and elements by \(\oplus \), \(\otimes \), \(\overline{0}\) and \(\overline{1}\).
Proposition 2
For \(n\ge 0\), \((\mathbb R^{n+1}, \oplus , \otimes , \overline{0}, \overline{1})\) defines a commutative semiring.
Notice that if for all \(i \le n\), we have \(x_i=\mu _i(\varPi _1)\) and \(y_i=\mu _i(\varPi _2)\), denoting \((z_0,\dots , z_n) = (x_0,\dots ,x_n)\otimes _n (y_0,\dots ,y_n)\), we get \(\mu _i(\varPi _1 \cdot \varPi _2)=z_i\). Further, if both \(\varPi _1,\varPi _2\) are disjoint, and if no path of \(\varPi _1\) (resp. \(\varPi _2\)) is a prefix of a path of \(\varPi _2\) (resp. \(\varPi _1\)), then \(\mu _i(\varPi _1 \cup \varPi _2)=x_i+y_i\).
3.2 Computations in a Semiring
Following the FloydWarshall algorithm to sum weights of paths reaching a state, we will decompose inductively \(\mathbf {Path}_q^R\) using operations \(\cup \) and \(\cdot \). We will then use the semiring \((\mathbb R^{n+1}, \oplus , \otimes , \overline{0}, \overline{1})\) to perform these computations inductively. The induction will be over the number of states in S. Let G be a subset of S disjoint with R: \(G \cap R = \emptyset \). For all state \(s \in S \setminus R\), we define \(\mathbf {Path}^t_s(G)=\{ s_0 \cdots s_n \mid s_0=s, s_n=t, \forall 1 \le i \le n1, s_i \in G\}\) the set of paths from state s to state t using only states G, except for the initial state, which is s and for the last state which is t, even if \(s,t \in R\) or \(s,t \notin G\).
For a set of paths \(\varPi \), we define \(w_n(\varPi )=(\mathbb {P}(\varPi ),\mu _1(\varPi ),\dots ,\mu _n(\varPi ))\). Let \(g \in G\) be a state of G. A path \(\rho \) in \(\mathbf {Path}^t_s(G)\) has two possibilities: either it does not use g, or it uses g one or several times. We deduce the inductive formula:
Proposition 3
Proof

\(\rho _0\) is in \(\mathbf {Path}^g_s(G \setminus \{g\})\),

\(\rho _k\) is in \(\mathbf {Path}^t_{g}(G \setminus \{g\})\),

and for all \(0<j<k\), \(\rho _j \in \mathbf {Path}^g_g(G \setminus \{g\})\).
In order to use this formula, we need to compute Open image in new window , which represents what happens along a cycle from g to g. Let \((g,\varPi )\) a pair with g a state and \(\varPi \) a set of paths (cycles) using g exactly twice: the first state and the last states are g. The pair \((g,\mathbf {Path}_g^g(G \setminus \{g\}))\) satisfies this property. We define Open image in new window . The restriction on \((r,\varPi )\) ensures that Open image in new window is disjoint. We show that \(w_n^*(\varPi )\) is defined in most cases, namely when \(\mathbb {P}(\varPi )<1\).
Proposition 4
Notice that \(P(\varPi )=1\) describes cases where s cannot reach t (as \(t \notin G\), if \(\mathbb {P}(w_n(\mathbf {Path}_g^g(G))=1\), it would mean that every path reaching g stays in G forever, and in particular never meets t). Thus, we first compute the set of states \(S_1\) from which there exists a path to R. Notice that for each set \(\varPi \) of paths ending in \(g \in S_1 \setminus R\), we have \(\mathbb {P}(\varPi )<1\), because there is a positive probability to reach R from g, which is not captured by paths in \(\varPi \).
3.3 A Symbolic Algorithm
From the inductive formulae to compute set of paths from subsets of paths and to compute \(w_n^*(\varPi )[i]\) from \(w_n^*(\varPi )[j]\) for \(j<i\), we deduce Algorithm 1, following the ideas of FloydWarshall, incrementally adding non response states from \(S_1 \setminus R\), which can be used as intermediate states. Notice that states in \(S \setminus S_1\) cannot reach R anyway. This algorithm is symbolic (or algebraic) in that every constant (e.g. Pr(s, a, t)) can be replaced by a variable (see e.g. Sect. 4.2).
Theorem 1
Let \(A=(S,\delta ,\delta _0)\) be a probabilistic automaton. One can compute \(\mu _i(s,t)\) for all \(i \le n\) and \(s,t \in S\) in time \(O(n^2 \times S^3)\).
Proof
In Algorithm 1, after running the outer forloop on \(g_1,\dots ,g_j\), we have \(w_n(s,t)[n] = \mu _n(\mathbf {Path}_s^t(\{g_1,\ldots , g_j\}))\). At the end of Algorithm 1, we obtain \(w_n(s,t)[n]=\mu _n(\mathbf {Path}_s^t)= \mu _n(s,t)\).
To obtain \(\mu _i(s,t)\) for all \(i \le n\), it suffices to run Algorithm 1 inductively on moment of order \(1, \ldots , n\). Computing \(w^*_n[i](s,t)\) in the inner forloop takes time O(i) as \(w_n[j](s,t)=w_j[j](s,t)\) has already been computed inductively for all \(j<i\). This yields the complexity of \(O(\sum _{j=1}^n i \times S^3)=O(n^2 \times S^3)\). \(\square \)
Now, for each query q, we have \(\mu _i(\mathbf {Path}_q^R) = \sum _{r \in R} \mu _i(q,r)\), as \(\mathbf {Path}_q^{r_1}\) and \(\mathbf {Path}_q^{r_2}\) have no path prefix of each other for \(r_1 \ne r_2, r_1,r_2 \in R\). Now, the moment of order n of the distribution of response times of q is formally either \(\infty \) if \(\mu _0(\mathbf {Path}_q^R)<1\) (there is positive probability to never answer q, that is have infinite response time), and \(\mu _n(\mathbf {Path}_q^R)\) otherwise.
Example 2
For the example of Fig. 1, unfolding the algorithm for \(n=2\) (that is for probability, and moments of order 1 and 2) gives after initialization:
\(w(1,2)=(1,1,1)\), \(w(2,2)=(0.9,0.9,0.9)\), \(w(2,3)=(0.1,0.1,0.1)\), and \(w(1,3)=(0,0,0)\), as there is no direct transition from state 1 to state 3.

\(w(2,2) \leftarrow w(2,2)\oplus w(2,2)\otimes w(2,2)^* \otimes w(2,2) = w(2,2)\otimes w(2,2)^*\)

\(w(2,3) \leftarrow w(2,3)\oplus w(2,2)\otimes w(2,2)^* \otimes w(2,3) = w(2,3)\otimes w(2,2)^*\)

\(w(1,3) \leftarrow w(1,3)\oplus w(1,2)\otimes w(2,2)^* \otimes w(2,3)\)
We have \(w(2,2)^* = (\frac{1}{10.9}, \frac{0.9}{(10.9)^2}, \frac{0.9}{(10.9)^2} + \frac{2\times 0.9^2}{(10.9)^3}) = (10,90,1710)\)
At the end of the algorithm, we obtain \(\mu _i(2,3)=\mu _i(\mathbf {Path}_{2}^{\{2\}})=w(2,3) = (0.1,0.1,0.1)\otimes (10,90,1710) = (1,10,190)\). Hence, in this probabilistic automata, the probability of responding to the query is 1, in a mean time of 10, with a standard deviation of \(\sqrt{19010^2}=9.5\).
3.4 Extension to Continuous Time
We now extend the symbolic computation of moments to continuous time Markov Chains (CTMCs). In order to be as close as possible to the setting of probabilistic automata, we use the sojourn time representation of CTMCs. This representation is fully equivalent with the more usual representation of CTMCs with transition rates, see Chap. 7.3 of [9].
Definition 3

\((S,Pr, \delta _0)\) is a probabilistic automata, and

for all s, \(\lambda _s\) is the sojourn parameter associated with state s. That is, the PDF function of the sojourn time is \(X_s(t)=\lambda _s e^{ \lambda _s \cdot t}\) and the probability to stay in s at least t units of time is \(e^{ \lambda _s \cdot t}\).
In this continuous context, we need integrals instead of sums to define the ith moment of a variable X: \(\mu _i(X)=\int _0^\infty X(t) t^i dt=1\). For every state \(s \in S\), let \(X_{s}(t)=\lambda _s e^{ \lambda _s \cdot t}\). For all i, for all s, \(\mu _i(X_s)\) is well defined and \(\mu _i(X_s)=\frac{i!}{\lambda _s^i}\)
We can easily extend the computation of moments for CTMCs. The inductive formulas for probabilities and moments of the reaching time distribution remain unchanged. We only need to change the definition of moments for every transition, which is input at the initialization phase of the Algorithm 1: for all \(s,t\in S\), we set \(w_n(s,t)\) to be \((w^0(s,t),w^1(s,t),\dots ,w^n(s,t))\), where Open image in new window and Open image in new window for all \(i\in [1,n]\).
Theorem 2
Let \(A=(S,Pr,\delta _0,(\lambda _s)_{s \in S})\) be a CTMC. One can compute \(\mu _i(s,t)\) for all \(i \le n\) and \(s,t \in S\) in time \(O(n^2 \times S^3)\).
4 Uniqueness of Distribution, Parameters and Hierarchy
In this section, we present cases where having a symbolic algorithm allows efficient techniques, compared to numerical methods. We start with hierarchical systems which are a way to compactly describe systems. Then, we present the possibility to work on systems with parameters. Finally, thanks to the symbolic expression of moments, we prove that there is a unique distribution having the moments of a distribution of reaching times of a (continuoustime) Markov chain.
4.1 Hierarchical Probabilistic Automata
Definition 4

\(S_i\) is the finite set of states of module i,

\(s_i^0\in S_i\) is the initial state of module i, and \(s_i^f\) the finial state of module i,

\(Pr_i: S_i\setminus \{s_i^f\} \times \varSigma \times S_i \rightarrow [0,1]\) is a stochastic transition function such that for all \(s\in S_i \setminus \{s_f\}\) (resp. \(s \in S_1\) for \(i=1\)), Open image in new window ,

\(\lambda _i: S_i \rightarrow \{i+1, \ldots , n\}\) is a partial mapping associating some states of \(S_i\) from module i to deeper modules.
Intuitively, the system starts in module 1, in state \(s_1^0\). Each time a state \(s \in S_i\) associated with a module \(j>i\), that is \(\lambda _i(s)=j\), is entered by a transition \(t \rightarrow s\), the system goes to state \(s_j^0\) and stays in \(S_j\) till \(s_j^f\) is seen, in which case it comes back to state s and takes a transition \(s \rightarrow t'\) (according to the probability distribution from s). This process can be repeated from any state in a module i to any module j as long as \(j>i\).
To define the semantics of \((S_i,Pr_i,\lambda _i,s_i^0,s_i^f)_{1 \le i \le n}\) formally, we inductively replace states associated with the deepest module by their definition. Indeed, nodes from the deepest module are not associated with any module by definition. Once every module has been replaced, a (flat) probabilistic automaton is obtained with the intended semantics.
Hence, HPA have the same expressive power as probabilistic automata. Yet, they may be much more compact: we denote by A the size of the description of the hierarchical automaton and by \(\Vert A\Vert \) the size of the unfolded automaton. The interest of such a description is that it may be exponentially smaller than the size of the unfolded automaton, as depicted in Fig. 2: here, every module contains two copies of the next module, with the exception of the last one. While the number of states in the description is linear (4n), the number of states in the unfolded automaton is equal to \(3\cdot 2^n2\).
The symbolic Algorithm 1 is naturally modular, in that computations on a module used several times can be performed only once by considering states of the deepest module first. Indeed, one module can be summarized by three information items: the probability (and moments) to answer the query in this module, the probability (and moments) to leave this module without answering the query in the module and the probability to stay forever in this module without answering the query. Then the information can be used for shallower modules: every time a state s in a module i is associated with the deepest module, it can be replaced by this small set of states containing all the relevant information about the deepest module (and computed only once). Then, this process can be repeated to eliminate modules recursively. This leads to a complexity in the small size A of the compact HPA representation rather than in the large size A of the unfolded PA:
Theorem 3
Let A be an HPA with k modules of size at most m. The n first moments of the distribution associated with A can be computed in time \(O(n^2 k m^3)\).
Not only does Theorem 3 reduces the complexity for hierarchical representations with redundancy (\(O(n^2 k)\) for the example in Fig. 2 instead of \(O(n^2 2^{3k})\) when running the algorithm in [13] on the equivalent flat PA), it also gives a better complexity on structure without redundancy. Consider the example in Fig. 3, without redundancy, with an unfolded PA with \(3k+1\) states. Theorem 3 takes time \(O(n^2 k 3^3)\), while the algorithm in [13] on the equivalent flat PA would take time \(O(n^2 (3k)^3)\).
4.2 Parametric Systems
The first case is when parameters help with redundancy. Often, stochastic systems reuse the same constructions, but with different probability values. This would be naturally encoded as a module M of a hierarchical system using a set of parameters P. This module M would be used several times, with different values of parameters specified in each module using it.
In this case, one can run Algorithm 1 on M, using the parameter values literally in the equations. This yields rational functions \(f_n:[0,1]^P \rightarrow (0,1]\) of the parameters expressing the moments of order n for module M, for all n. For instance with the example of Fig. 4, the probability to reach state 4 from state 1 is equal to \(\frac{2p+4}{5p+4}\), and the mean time is equal to \(\frac{112+44p12p^2}{(5p+4)(2p+4)}\). Each time module M is used, \(f_n\) can be evaluated using the value of the parameters P for this particular usage.
Another possible use of parameters is to model uncertainty of values. In the example of Fig. 4, we may not know exactly the value of parameter p, but only know that it is above 0.8. In this case, one may be interested of synthesizing the largest (resp. smallest) moment of order n which is smaller (resp. larger) than the moment of any system realizing the parametric system, that is where p is replaced by any value above 0.8. This will be particularly interesting in the next section discussing bounds. To do so, one can use the rational function \(f_n\) to compute its minimal and maximal values (e.g. deriving it and looking for 0 with Euler’s method). In this way, we also obtain the best/worst value for p.
4.3 Uniqueness of the Distribution
Last, we use the symbolic expression of moments obtained in Sect. 3 in order to prove the uniqueness of the distribution having moments of first passage times of (continuoustime) Markov chains. Thus this distribution is the distribution of response times of the system considered.
Notice that in general, there may be several distributions that correspond to a given sequence of moments \((\mu _n)_{n \in \mathbb {N}}\). This would compromise approximating the distribution using moments, as there would not be a unique such distribution.
Example 3
Let us consider a distribution \(\delta \) on \(\mathbb {R}^+\). If \(\delta \) has the sequence of moments \(\{\mu _n=n! \mid n \in \mathbb {N}\}\), then \(\delta \) is the exponential distribution with parameter 1. Similarly, the sequence of moments \(\{\mu _n=(2n)! \mid n \in \mathbb {N}\}\) for a distribution on \(\mathbb {R}^+\) is characteristic of the square of the exponential distribution of parameter 1.
Now, consider the cube of the exponential distribution of parameter 1. Its sequence of moments is \(\{\mu _n=(3n!) \mid n \in \mathbb {N}\}\). However, there exist an infinite number of distributions with this sequence of moments [18].
We now prove answer positively to the Stieljes moment problem for the case of the distribution of response time in a (continuoustime) Markov chain, that is its sequence of moments respects the Carleman’s condition from year 1922, that guarantees the uniqueness of the distribution. The condition is that \(\sum _{n \in \mathbb {N}} \mu _n(\delta )^{\frac{1}{2n}} = \infty \).
Theorem 4
Let A be a probabilistic automaton or a CTMC. For all \(n \in \mathbb {N}\), let \(\mu _n\) be the moment of order n of the times of first passage in a set of state R of A. Then there exists a unique distribution \(\delta \) such that \(\mu _n(\delta )=\mu _n\) for all \(n \in \mathbb {N}\).
Sketch of Proof: We first consider CTMC where all states have the same sojourn time \(\lambda \). Then, a path that uses i transitions to answer a query will follow the gamma distribution with parameters (\(i,\lambda \)). We have a symbolic expression for moments of this distribution thanks to Sect. 3. This can be used to minimize \(\sum _{n=0}^\infty \mu _n(\delta )^{\frac{1}{2n}}\) by a diverging sum.
For general CTMCs, we use the fact that \(\mathbb E(\Gamma (i,\lambda _1)^n)\le \mathbb E((E(\lambda _1)+\dots +E(\lambda _i))^n)\) iff \(\lambda _1=min(\lambda _j)_{j=1}^i\). It allows us to minimize the Carleman’s sum of the CTMC considered by the Carleman’s sum of the CTMC where all sojourn times are replaced by the smallest sojourn time \(\lambda \), hence the divergence.
The case of probabilistic automaton is simpler. \(\square \)
We show how this theorem allows to approximate distribution \(\delta \) in the next subsection.
4.4 A Sequence of Distributions Converging Towards \(\delta \)
Since we have unicity of the distribution corresponding to the sequence of moments of the distribution of response time of a probabilistic automaton, we obtain the following convergence in law:
Proposition 5
([17]). Let \(\delta \) be the distribution of response times of a probabilistic automaton. Let \((\delta _i)_{i\in \mathbb {N}}\) be a sequence of distributions on \(\mathbb {R}^+\) such that for all n, Open image in new window . Then, if \(C_i\) is the cumulative distribution function of \(\delta _i\) and C the cumulative distribution function of \(\delta \), then for all x Open image in new window .
Thus, C can be approximated by taking a sequence \((\delta _n)_{n \in \mathbb {N}}\) of distribution such that for all \(i \le n\), \(\mu _i(\delta _n) = \mu _i(\delta )\). A reasonable choice for \(\delta _n\) is to consider the distribution of maximal entropy corresponding to the moments \(\mu _1, \ldots , \mu _n\), as presented in [11]. The distribution of maximal entropy can be understood as the distribution that assume the least information. It can be approximated as close as desired, for instance \(\frac{1}{n}\) close to the distribution of maximal entropy having moments \((\mu _1(\delta ), \ldots , \mu _n(\delta ))\). Applying Proposition 5, we thus obtain that the cumulative distribution function associated with \(\delta _i\) converges towards the cumulative distribution function associated with \(\delta \).
5 Bounding the Response Time
We now explain how to use moments in order to obtain optimal bounds on the response time. First, notice that as soon as there exists a loop between a query and a response (as in Fig. 1), then there will be runs with arbitrarily long response times, although there might be probability 1 to eventually answer every query (which is the case for Fig. 1). We thus turn to a more quantitative evaluation of the response time.
Let \(0<p<1\). We are interested in a bound T on the delay between a query and a response such that more than \(1p\) of the queries are answered before this bound. For a distribution Open image in new window of response times, we denote by \(B(\delta ,p)\) the lowest T such that the probability to have a response time above T is lower than p. Equivalently, we look for the highest T such that the probability of a response time above T is at least p.
We place ourselves in the general setting of continuous distributions, where Dirac delta functions are allowed for simplicity. Discrete distributions form a special case, with delta functions at integer values. One could get rid of Dirac delta functions by \(\epsilon \)approximating them without changing the moments, obtaining the same bounds as we prove here.
5.1 Tchebychev Bounds Associated with One Moment
Let \(i \in \mathbb {N}\) and \(\mu _i >0\). We let \(\varDelta _{i,\mu _i}\) be the set of distributions of response time which have \(\mu _i\) as moment of order i. We are interested in bounding \(B(\delta ,p)\) for all \(\delta \in \varDelta _{i,\mu _i}\), that is for all distributions with \(\mu _i\) as moment of order i. Such a bound is provided by Tchebychev inequality, and it is optimal:
Proposition 6
Let \(i \in \mathbb {N}\) and \(\mu _i\). Let \(\alpha _i(\mu _i,p)=\root i \of {\frac{\mu _i}{p}}.\) Then for all \(\delta \in \varDelta _{i,\mu _i}\), we have \(B(\delta ,p) \le \alpha _i(\mu _i,p)\). Further, \(\exists \delta \in \varDelta _{i,\mu _i}\) such that \(B(\delta ,p) = \alpha _i(\mu _i,p)\).
Proof
It suffices to remark that \(\mu _i > p b^i\) for b the bound we want to reach. Further, this bound is trivially optimal: it suffices to consider a distribution with a Dirac of mass \((1p)\) at 0 and a Dirac of mass p at \(\alpha _i(\mu _i,p)\). \(\square \)
Given a probabilistic automaton, let \(\delta \) be its associated distribution of response time. We can compute its associated moments \(\mu _i\) using Algorithm 1, described in the previous section. We thus know that \(\delta \in \varDelta _{i,\mu _i}\). Given different values of i, one can compute the different moments and apply for each of the Tchebychev bound and use the minimal bound obtained.
Understanding the relationship between the \(\alpha _i\) is thus important. For \(i<j\), one can use Jensen’s inequality for the convex function Open image in new window , and obtain: \((\mu _i)^j \le (\mu _j)^i\). For instance, \(\mu _1^2 < \mu _2\).
For \(p=1\), this gives \(\alpha _i(p=1)<\alpha _j(p=1)\). On the other hand, for p sufficiently close to 0, we have \(\alpha _j(p)<\alpha _i(p)\). That is, when p is very small, moments of high orders will give better bounds than moments of lower order. On the other hand, if p is not that small, moments of small order will suffice.
5.2 Optimal Bounds for a Pair of Moments
We now explain how to extend Tchebychev bounds to pairs of moments: We consider the set of distributions where two moments are fixed. Let \(i<j\) be two orders of moments and \(\mu _i,\mu _j >0\). We denote by \(\varDelta _{i,\mu _i}^{j,\mu _j}\) the set of distributions with \(\mu _i,\mu _j\) as moments of order i, j respectively. As \(\varDelta _{i,\mu _i}^{j,\mu _j}\) is strictly included into \(\varDelta _{i,\mu _i}\) and in \(\varDelta _{j,\mu _j}\), \(\min (\alpha _i(p),\alpha _j(p))\) is a bound for any \(\delta \in \varDelta _{i,\mu _i}^{j,\mu _j}\). However, it may be the case that \(\min (\alpha _i(p),\alpha _j(p))\) is not optimal. We now provide optimal bounds \(\alpha _i^j(p)\) for any pair \(i<j\) of order of moments and probability p:
Theorem 5

\(\alpha _i\) if \(\alpha _i \le \alpha _j\),
 \((\frac{\mu _jM}{p})^{\frac{1}{j}}\) otherwise, where \(0 \le M \le \mu _j\) is the smallest positive real root of:$$\mu _i=(1p)^{\frac{ji}{j}} M^{\frac{i}{j}} + p^{\frac{ji}{j}} (\mu _j  M)^{\frac{i}{j}}.$$
For all \(\delta \in \varDelta _{i,\mu _i}^{j,\mu _j}\), we have \(B(\delta ,p) \le \alpha _i^j\), and \(\exists \delta \in \varDelta _{i,\mu _i}^{j,\mu _j}\) with \(B(\delta ,p)=\alpha _i^j\)
Example 4
i  \(\mu _i\)  \(\alpha _i(0.1)\)  \(\alpha _i^{i1}(0.1)\)  \(\alpha _i(0.01)\)  \(\alpha _i^{i1}(0.01)\) 

1  10  100  100  1000  1000 
2  190  43.6  38.5  137.8  104.9 
3  5410  37.8  36.8  81.5  73.9 
4  205390  37.9  37.8  67.4  63.8 
5  9747010  39.6  37.9  64.2  61.43 
6  555066190  42.1  39.6  62.8  61.47 
For \(p=0.1\), it is not useful to consider moments of order higher than 3. For \(p=0.01\), moment of order 5 provides better bounds than moment of lower orders.
Lemma 1
\((f^n(0))_{n \in \mathbb {N}}\) is strictly increasing and converges towards M.
We show how to \(\epsilon \)approximate the optimal bound B of a parametric probabilistic automaton A with set of parameters P, that is such that for all \(val \in V^P\), the probabilistic automaton A with valuation val for parameter values has a bound \(b(val) \le B\) and there exists a \(val \in V^P\) such that \(b(val)=B\). First, we obtain the moments as symbolic functions of the parameters using Sect. 4.2. Then, we compute \(M_1=f(0)\) as a function of the parameters, using Lemma 1 and replacing \(\mu _i,\mu _j\) by their expression. One can then compute the minimum \(m_1\) of function \(M_1\) over all the parameters. We then proceed with \(M_2=f(m_1)\), and so on till obtaining a value m. This allows to obtain a lower bound m over values of M for all parameter values. Computing the largest \(\mu _j\) over all parameters allows to obtain an upper bound \(B_{up}\): \(B \le B_{up}=(\frac{\mu _j  m}{p})^{\frac{1}{j}}\). A lower bound \(B_{lw}\) is easily obtained by considering the value \(\ge m\) of M for the parameters maximizing \(\mu _j\). If the distance between \(B_{up}\) and \(B_{lw}\) is larger than \(\epsilon \), one can partition the space of parameter values in zones and proceed in the same way on each zone, forgetting zones for which \(B_{up}\) is lower than the \(B_{lw}\) of another zone, till the distance between \(\max (B_{lw})\) and \(\max (B_{up})\) over zones is smaller than \(\epsilon \).
6 Conclusion
In this paper, we have shown how to compute moments symbolically for probabilistic automata and CTMCs, using adequately defined semirings. This method has the same complexity as the efficient numerical methods already known [13]. The proof of this symbolic computation allows proving that there is a unique distribution of response time corresponding to a probabilistic automaton or a CTMC. This allows obtaining simple approximated distributions scheme converging in law towards the distribution of response time. The symbolic computation of moments also allows computing moments in compact (hierarchical) models faster, as well as finding lowest/highest value of moments in parametric systems.
We also provide optimal bounds on the delay after which very few queries stay unanswered. It is optimal when considering distribution displaying a given pair of moments, and we showed on a simple example how this improves Tchebychev bounds. This can be used efficiently to obtain bounds for compact (hierarchical) models or to compute an optimal bound which fulfills the response of almost all queries even for systems where some parameter values are not known exactly.
References
 1.Asarin, E., Basset, N., Degorre, A.: Entropy of regular timed languages. In: Information and Computation, vol. 241, pp. 142–176. Elsevier (2015)MathSciNetCrossRefGoogle Scholar
 2.Angrish, R., Chakraborty, S.: Probabilistic timing analysis of asynchronous systems with moments of delay. In: ASYNC 2002. IEEE (2002)Google Scholar
 3.Alur, R.: Formal analysis of hierarchical state machines. In: Verification: Theory and Practice, pp. 42–66 (2002)Google Scholar
 4.Backenköhler, M., Bortolussi, L., Wolf, V.: Generalized method of moments for stochastic reaction networks in equilibrium. In: Bartocci, E., Lio, P., Paoletti, N. (eds.) CMSB 2016. LNCS, vol. 9859, pp. 15–29. Springer, Cham (2016). https://doi.org/10.1007/9783319451770_2CrossRefMATHGoogle Scholar
 5.Bazille, H., Fabre, E., Genest, B.; Symbolically quantifying response time in stochastic models using moments and semirings. https://perso.crans.org/~genest/BFG18.pdf
 6.Bradley, J., Dingle, N., Harder, U., Harrison, P., Knottenbelt, W.: Response time densities and quantiles in large Markov and semiMarkov Models. In: Performance Evaluation of Parallel, Distributed and Emergent Systems, vol. 1 (2006)Google Scholar
 7.Bogomolov, S., Henzinger, T.A., Podelski, A., Ruess, J., Schilling, C.: Adaptive moment closure for parameter inference of biochemical reaction networks. In: Roux, O., Bourdon, J. (eds.) CMSB 2015. LNCS, vol. 9308, pp. 77–89. Springer, Cham (2015). https://doi.org/10.1007/9783319234014_8CrossRefGoogle Scholar
 8.Chatterjee, K., Henzinger, T.A., Horn, F.: The complexity of requestresponse games. In: Dediu, A.H., Inenaga, S., MartínVide, C. (eds.) LATA 2011. LNCS, vol. 6638, pp. 227–237. Springer, Heidelberg (2011). https://doi.org/10.1007/9783642212543_17CrossRefGoogle Scholar
 9.Cassandras, C., Lafortune, S.: Introduction to Discrete Event Systems. Springer, Boston (2007). https://doi.org/10.1007/9780387686127CrossRefMATHGoogle Scholar
 10.Cortes, C., Mohri, M., Rastogi, A., Riley, M.: On the computation of the relative entropy of probabilistic automata. Int. J. Found. Comput. Sci. (IJFCS) 19(1), 219–242 (2006)MathSciNetCrossRefGoogle Scholar
 11.Cover, T., Thomas, J.: Elements of Information Theory. Wiley, New York (2006)MATHGoogle Scholar
 12.Daws, C.: Symbolic and parametric model checking of discretetime Markov chains. In: Liu, Z., Araki, K. (eds.) ICTAC 2004. LNCS, vol. 3407, pp. 280–294. Springer, Heidelberg (2005). https://doi.org/10.1007/9783540318620_21CrossRefMATHGoogle Scholar
 13.Dayar, T., Akar, N.: Computing moments of first passage times to a subset of states in Markov chains. SIAM J. Matrix Anal. Appl. 27(2), 396–412 (2005)MathSciNetCrossRefGoogle Scholar
 14.Gonzalez, A.M., Uhlendorf, J., Schaul, J., Cinquemani, E., Batt, G., FerrariTrecate, G.: Identification of biological models from singlecell data: a comparison between mixedeffects and momentbased inference. In: ECC 2013, pp. 3652–3657. IEEE (2013)Google Scholar
 15.Horn, F., Thomas, W., Wallmeier, N., Zimmerman, M.: Optimal strategy synthesis for requestresponse games. RAIRO 49(3), 179–203 (2015)MathSciNetMATHGoogle Scholar
 16.Jansen, N., Corzilius, F., Volk, M., Wimmer, R., Ábrahám, E., Katoen, J.P., Becker, B.: Accelerating parametric probabilistic verification. In: Norman, G., Sanders, W. (eds.) QEST 2014. LNCS, vol. 8657, pp. 404–420. Springer, Cham (2014). https://doi.org/10.1007/9783319106960_31CrossRefGoogle Scholar
 17.Prohorov, Y., Rozanov, Y.: Probability Theory, Basic Concepts \(\cdot \) Limit Theorems Random Processes. Springer, Heidelberg (1969). Translated from RussianGoogle Scholar
 18.Stoyanov, J.: Determinacy of distributions by their moments. In: ICMSM 2006 (2006)Google Scholar
 19.Tari, Á: Moments based bounds in stochastic models, Ph.D. Thesis. Budapesti Műszaki és Gazdaságtudományi Egyetem (2005)Google Scholar
 20.Telek, M., Horvéth, G.: A minimal representation of Markov arrival processes and a moments matching method. Perform. Eval. 64(9–12), 1153–1168 (2007)CrossRefGoogle Scholar
 21.Wallmeier, N., Hütten, P., Thomas, W.: Symbolic synthesis of finitestate controllers for RequestResponse specifications. In: CIAA 2003 (2003)CrossRefGoogle Scholar
 22.Yao, D.: Firstpassagetime moments of Markov processes. J. Appl. Probab. 22(4), 939–945 (1985)MathSciNetCrossRefGoogle Scholar
Copyright information
Open Access This chapter is licensed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license and indicate if changes were made. The images or other third party material in this book are included in the book's Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the book's Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder.