1 Introduction

Planning is a problem with great importance in a wide range of fields. Generally, planning is the process of analyzing how a system changes its state based on the occurrence of actions, followed by choosing which actions to apply to the system to achieve a desired goal. Examples of planning include: path planning, navigation planning, manipulation planning, perception planning, and job-shop scheduling. Research has been done across these different notions of planning to make the process more efficient. According to Qiang (2012), an efficient planning algorithm is the “one able to find a good plan quickly”.

Focusing on the development of general algorithms for planning, Ghallab et al. (2004) proposed employing discrete-event system models (a model widely used in the field of computing). Discrete-event systems can be represented by discrete graphs, where nodes are states and transitions are actions. Based on this representation, the classical planning problem can be seen as a search for a path in a graph from the initial state to a goal state with the optimization of a criterion. However, planning problems are highly complex and it can be difficult to explicitly model a system’s behavior. To deal with this difficulty, the use of Supervisory Control Theory (SCT) has emerged as an interesting tool.

With SCT, the system and specifications are modeled separately and then tools are applied to generate the supervisor, which can achieve the most permissive closed-loop behavior that maintains safety and non-blockingness. One approach to solve planning problems is to use the supervisor as the search space to find a plan that optimizes a defined criterion, such as time or energy [(Hill and Lafortune 2017; Pena et al. 2016; Dulce-Galindo et al. 2019; Hagebring and Lennartson 2018; Ware and Su 2017; Su 2012; Fabre and Jezequel 2009), and (Bravo et al. 2018)].

The concept of synthesizing an optimal supervisor, when not directly related to the application of planning, can be seen as the solution of this type of problem [(Huang and Kumar 2007; Sengupta and Lafortune 1993; Asarin and Maler 1999; Brandin and Wonham 1994; Su et al. 2012) and (Hyun-Wook et al. 2010)]. In general, the synthesis of an optimal supervisor is beyond traditional supervisor synthesis, which only restricts the system to the set of legal behaviors. An optimal supervisor further limits the system’s behaviors to those that minimize or maximize some criterion. Most existing approaches to synthesizing an optimal supervisor rely on a monolithic approach to modeling the system, which can have prohibitive computational complexity.

Efforts have been directed to improve the scalability of planning using SCT. Modular and compositional algorithms have been applied when a system can be decomposed into sub-systems.

Su (2013) proposes a distributed approach to time optimal supervisory control, but it requires the synthesis of a global coordinator through the composition of all local components and requirements, which also can have prohibitive complexity.

Hagebring and Lennartson (2018) and Ware and Su (2017) represent their models as standard tick automata, which has the effect of dramatically increasing their model sizes, thereby reducing the applicability of the approaches for more complex problems. Hagebring and Lennartson (2018) propose the application of local optimization and abstraction to each sub-system, which removes all non-optimal local paths and merges sequences of local events into single events. This is followed by a compositional optimization, which synchronizes a sub-set of components generating a larger sub-system to which a second iteration of the local optimization is applied. This cycle is repeated until only one model remains, representing the optimal global solution by construction.

Ware and Su (2017) propose two methods for synthesizing the time-optimal accepting trace for timed automata with only controllable events based on abstraction and pruning. For the first method, the sub-systems are composed together and all local events are abstracted. When a single automaton is achieved, Dijkstra’s algorithm is applied, producing the time-optimal accepting trace. The second method is a modification of the first algorithm, applying pruning instead of abstraction to reduce the size of the automata. However, when a trace is returned, it is not guaranteed to be optimal since the pruning can remove all time-optimal accepting traces.

Bravo et al. (2018) generate results specifically for production systems with symmetrically reachable idle states (states with inactive behavior). Such a system is shown to be decomposable into factors that are optimized independently and then the partial solutions are concatenated to generate the minimum makespan controllable sublanguage.

van Putten et al. (2020) propose an approach for the synthesis of a throughput-optimal supervisor for a manufacturing system. In their modeling, they use the concept of activities which express some deterministic functionality, such as moving a product, in terms of actions. With that they create a high-level model composed by a plant model, which describes the activity sequences that are available, and specifications models, that describes the order in which activities should be executed. The lower-level model has for each activity an acyclic graph that defines their constituent actions and dependencies between them. The system’s behavior, including timing, is captured from the lower level by variables and guards of the automata used in the high-level models. A supervisor is synthesized for the high-level, but for that they remove all the variables related to time from the models prior to synthesis and return them once the supervisor is obtained. This procedure affects the guarantee of controllability and nonblockingness of the supervisor. To the resultant monolithic supervisor, they apply a game theoretic method, called ratio games, to find the control that optimizes the throughput.

Hill and Lafortune (2017) develop an approach for generating approximately time optimal supervisory control logic for a cooperative multi-robot system employing hierarchy and decomposition based on the results of Hill and Lafortune (2016). These works do not employ tick automata and do not rely on the identification of idle states. The work of (Hill and Lafortune 2016) proposes a hierarchical approach based on a new notion, cost equivalence, that aggregates states with common future costs, as shown in Fig. 1. This equivalence generates an abstraction of the supervisor that with additional conditions, such as an additive cost function and only transitions with zero cost can be “hidden,” guarantees that an optimal plan found on the abstraction is also optimal when applied to the underlying full supervisor.

Fig. 1
figure 1

Hierarchical planning (Hill and Lafortune 2016)

The work of this paper improves upon (Hill and Lafortune 2016) by developing conditions under which the abstraction can be generated more efficiently. Specifically, this paper is an extended version of the conference paper (Vilela and Hill 2020) where the proposed requirements are relaxed and additional details are provided, including a more developed example. In this work, we employ a new notion of equivalence introduced by Vilela and Hill (2020) based on cost equivalence and weak bisimulation that we term priced-observation equivalence. This class of equivalence is used to generate an abstraction based on the aggregation of states with futures that share the same costs and observed event labels. This new equivalence allows the abstraction to be generated compositionally, rather than first having to synthesize the full global supervisor as done by Hill and Lafortune (2016). The work of Hill and Lafortune (2016) and Vilela and Hill (2020) include as part of their process an approach for artificially reducing the cost of certain transitions to zero. In this work, we provide a relaxation of the conditions to perform this transformation, and demonstrate that it can be performed compositionally.

The remainder of this paper is divided into four sections: Section 2 provides the necessary preliminary definitions, Section 3 develops the main theoretical results, Section 4 provides a detailed example, and Section 5 presents the paper’s conclusions.

2 Preliminaries

This paper employs discrete-event models where behaviors are represented using strings of events taken from a finite alphabet Σ. Σ is the set of all finite strings of events in Σ, including the empty string ε, and the concatenation of strings s,u ∈Σ is written as su. A subset \(L \subseteq {\Sigma }^{*}\) is called a language. The possible behaviors of a system are represented by its automaton generator. Specifically, this paper employs weighted automata, \(G = (Q,{\Sigma },\rightarrow , q_{0}, Q_{m}, f)\), where Q is a finite set of states, Σ is a finite set of events, \(\rightarrow \ \subseteq Q \times {\Sigma } \times Q\) is the (partial) state transition relation, q0Q is the initial state, \(Q_{m} \subseteq Q\) is the finite set of marked states, and f is the cost function that represents a mapping from a transition, Q ×Σ× Q, to a nonnegative real number, \([0, \infty )\). The cost will typically represent the expenditure of energy or time associated with a transition. The transition relation is written in infix notation, \(x \ \overset {\sigma }{\longrightarrow } \ y\), and is extended to traces in the natural way. The cost of a trace s = σ1σ2...σn− 1 is defined as the sum of the costs of each transition, \(C(q_{1}\overset {s}{\rightarrow }q_{n}) = {\sum }_{i=1}^{n-1}f(q_{i}\overset {\sigma _{i}}{\rightarrow }q_{i+1})\). G is assumed to be deterministic, therefore, if \(x \overset {\sigma }{\longrightarrow }y_{1}\) and \(x \overset {\sigma }{\longrightarrow }y_{2}\), then y1 = y2. The generated language of an automaton G is \(L(G):= \{ s \in {\Sigma }^{*} | q_{0} \overset {s}{\longrightarrow }q, \ s.t. \ q \in Q \}\) and the marked language is \(L_{m}(G):=\{ s\in {\Sigma }^{*} | q_{0} \overset {s}{\longrightarrow }q,\ s.t.\ q \in Q_{m} \}\).

Given an automaton \(G = ({\Sigma },Q,\rightarrow , Q_{m}, q_{0}, f)\), a state qQ is termed accessible if \(q_{0} \ \overset {s}{\longrightarrow } \ q\) for some s ∈Σ. An automaton is accessible if every state is accessible and the accessible component of an automaton G is obtained by the operation Ac(G). For the purposes of this paper, all the automata are considered accessible. The synchronous operation of two automata G1 and G2 is modeled using the parallel composition operator, ∥, as defined in the following.

Definition 1

The parallel composition of two weighted automata \(G_{1}=(Q_{1}, {\Sigma }_{1}, \rightarrow _{1}, q_{0,1}, Q_{m,1}, f_{1})\) and \(G_{2}=(Q_{2}, {\Sigma }_{2}, \rightarrow _{2},q_{0,2},Q_{m,2},f_{2})\) is denoted G1G2 = Ac(Q1 × Q21 \(\cup {\Sigma }_2, \rightarrow _{1,2} ,(q_{0,1},q_{0,2}), Q_{m,1} \times Q_{m,2}, f_{1,2})\), where \(\rightarrow _{1,2}\) and f1,2 are defined as follows:

  • if σ ∈Σ1 ∩Σ2, then \((q_1,q_2) \overset {\sigma }{\longrightarrow }_{1,2}(q_{1}^{\prime },q_{2}^{\prime })\) and \(f_{1,2}((q_1,q_2) \overset {\sigma }{\rightarrow }_{1,2}(q_{1}^{\prime },q_{2}^{\prime }))=max(f_1(q_1 \overset {\sigma }{\rightarrow _{1}}q_{1}^{\prime }), f_{2}(q_{2}\overset {\sigma }{\rightarrow _{2}}q_{2}^{\prime }))\);

  • if σ ∈Σ1 and σ∉Σ2, then \((q_{1},q_{2}) \overset {\sigma }{\longrightarrow }_{1,2}(q_{1}^{\prime },q_{2})\) and \(f_{1,2}((q_{1},q_{2}) \overset {\sigma }{\longrightarrow }_{1,2}(q_{1}^{\prime },q_{2}))=f_{1}(q_{1}\overset {\sigma }{\longrightarrow _{1}}q_{1}^{\prime })\);

  • if σ∉Σ1 and σ ∈Σ2, then \((q_{1},q_{2}) \overset {\sigma }{\longrightarrow }_{1,2}(q_{1},q_{2}^{\prime })\) and \(f_{1,2}((q_{1},q_{2}) \overset {\sigma }{\longrightarrow }_{1,2}(q_{1},q_{2}^{\prime }))=f_{2}(q_{2}\overset {\sigma }{\longrightarrow _{2}}q_{2}^{\prime })\);

  • otherwise, \(\rightarrow _{1,2}\) and f1,2 are undefined.

In this definition, the function \(\max \limits \) is used to establish the cost of transitions with shared events. This decision makes sense, in particular, when costs are related to time. For example, if a task is running in parallel in two automata, the one that takes the greater amount of time determines the overall required time for both automata to complete the task.

The natural projection \(P_{i}: {\Sigma }^{*} \rightarrow {\Sigma }_{i}^{*}\) maps strings in Σ to strings in \({\Sigma }_{i}^{*}\) by erasing all events not contained in Σi. We can also define the inverse projection \(P_{i}^{-1}(t):=\{ s \in {\Sigma }^{*} : P_{i}(s)=t \}\). These definitions can be naturally extended to languages and then applied to provide an alternate definition of the parallel composition in terms of languages: \(L(G_{1}\|G_{2}) := P_{1}^{-1}(L(G_{1})) \cap P_{2}^{-1}(L(G_{2}))\) and Lm(G1G2) := \(P_{1}^{-1}(L_{m}(G_{1})) \cap P_{2}^{-1}(L_{m}(G_{2}))\). For further details on discrete-event systems and the base SCT, the reader is referred to Cassandras and Lafortune (2008).

A partition automaton, Gh, is generated from an automaton \(G=(Q, {\Sigma }, \rightarrow , q_{0}, Q_{m},\) f) by aggregating its states into distinct sets such that transitions inside a partition group cannot be observed. Defined more formally, a partition automaton Gh is a weighted automaton \(G^{h}=(Q^{h}, {\Sigma }^{h}, \rightarrow _h, {q_{0}^{h}}, {Q_{m}^{h}}, f^{h})\), where Qh is a set of partition states \(\{ {q_{1}^{h}},{q_{2}^{h}},...,{q_{n}^{h}} \}\) such that \(\dot {\cup }q^{h}=Q\), \({q_{i}^{h}} \neq \emptyset \), 1 ≤ i ≤|Qh|, and \({q_{i}^{h}}\cap {q_{j}^{h}}=\emptyset \) for ij. The initial state of the partition automaton is defined such that \(q_{0} \in {q_{0}^{h}}\), and marked states are defined such that \(q^{h}\in {Q_{m}^{h}} \Leftrightarrow \exists q \in Q_{m}\) s.t. qqh. The transition relation \(\rightarrow _{h} \subseteq Q^{h} \times {\Sigma }^{h} \times Q^{h}\) includes transitions from the original automaton that are “between” partitions. A transition \(q^{h}\overset {\sigma }{\longrightarrow _h}q^{h^{\prime }}\) exists in Gh if and only if there exists a transition \(q\overset {\sigma }{\longrightarrow }q^{\prime }\) in the original automaton such that qqh and \(q^{\prime } \in q^{h^{\prime }}\). The transition’s cost in the partition automata \(f^{h}(q^{h}\overset {\sigma }{\rightarrow _h}q^{h^{\prime }})\) will equal \(f(q\overset {\sigma }{\rightarrow }q^{\prime })\). This situation is represented in Fig. 2.

Fig. 2
figure 2

Cost of transitions in a partition automaton

For brevity, we introduce the notation, \(x \overset {s}{\Rightarrow } y\), to represent how a string in an automaton G is observed in its corresponding partition automaton Gh. Otherwise stated, \(x \overset {s}{\Rightarrow } y\) with s = σ1σ2...σn ∈Σh, denotes the existence of \(x \overset {r}{\rightarrow } y\) with exactly the order of events in s, but with an arbitrary number of “hidden” events, τ ∈Σ∖Σh, shuffled with the observable events in s, that is, r = t1σ1t2σ2t3...tnσntn+ 1, where \(t_{i} \in \{\Sigma \setminus {\Sigma }^{h}\}^{*}\). Further, we use the following notation to capture the cost of an observed string s ∈Σh, \(C(q_1 \overset {s}{\Rightarrow }q_{n+1})= {\sum }_{\tiny {1\leq i \leq n: \sigma _i\in {\Sigma }^h}}f(q_i \overset {\sigma _i}{\longrightarrow }q_{i+1})\).

A path in G whose transitions are visited in an order consistent with the order of the transitions in Gh, is termed a realization of the corresponding “high-level” path. This consistency of ordering is named trace-cost consistency for individual trajectories and is denoted: \(q_1 \rightarrow q_n \in {q_{1}^{h}} \rightarrow _h {q_{k}^{h}}\).

In Hill and Lafortune (2016), the partition automaton is generated by aggregating states that are cost equivalent. This notion of equivalence requires that states must have observed futures with the same cost and consistent marking. The partitions are generated such that only transitions with zero cost are “hidden”. Such partition systems are termed zero-cost reachable. The following result from Hill and Lafortune (2016) demonstrates that when partition systems maintain this property and employ cost equivalence, then an optimal plan chosen in the abstracted automaton has at least one realization in the underlying automaton and all such realizations are optimal. We define the optimal path between two states of G, \(q_1 \rightarrow ^{00} q_n\), as the sequence of transitions between q1 and qn with the smallest path cost.

Theorem 1 (Hill and Lafortune (2016))

Let \(G^{h}=(Q^h,{\Sigma }^h,\rightarrow _h, {q_{0}^{h}}, {Q_{m}^{h}}, f^{h})\) be a partition automaton generated by cost-equivalent abstraction of a weighted automaton \(G=(Q, {\Sigma }, \rightarrow , q_0, Q_m,f)\) such that {G,Gh} is zero-cost reachable. Let an optimal marked path between two states in Gh be \({q_{1}^{h}}\rightarrow _{h}^{00}{q_{k}^{h}} \in {Q_{m}^{h}}\). Then there exists at least one trace-cost consistent realization of this path in G, and all such realizations are optimal, \(q_{1}\rightarrow ^{00} q_n \in {q_{1}^{h}}\rightarrow _{h}^{00}{q_{k}^{h}}\).

Hill and Lafortune (2016) further introduce a technique to improve the amount of reduction that can be achieved for a given automaton. The idea is to create transitions that artificially have zero cost, while maintaining the overall cost of each marked path. For that, the authors introduce the concept of sequential dependence. A pair of events (σa, σb) is defined to be a sequentially-dependent ordering if the occurrence of the event σa in a marked path is always eventually followed by the event σb. Consider a sequentially-dependent ordering σaσb in an automaton G, where σa is the independent event with cost \(f(p\overset {\sigma _a}{\longrightarrow }p^{\prime })=c_a\), and σb is the dependent event with cost \(f(q\overset {\sigma _b}{\longrightarrow }q^{\prime })=c_b\). If the cost of the dependent transitions are consistent in G, it is possible to perform a “lumping” operation, which transfers the cost of all transitions with label σb to transitions with label σa, \(f(p\overset {\sigma _a}{\longrightarrow }p^{\prime })=c_a+c_b\) and \(f(q\overset {\sigma _b}{\longrightarrow }q^{\prime })=0\). The rest of the automaton G remains unaltered, resulting in the lumped automaton G. The property of costs being consistent for all transitions with the same event labels is referred to as time-separability.

Hill and Lafortune (2016) prove in their Proposition 4 that all marked paths in a lumped automaton G possess the same cost as in the original automaton G. This result allows an abstraction of G to be used for planning.

3 Main results

The main results of this paper improve upon (Hill and Lafortune 2016) by introducing a new notion of equivalence that allows the abstraction to be generated compositionally. This concept was introduced in the conference paper (Vilela and Hill 2020). Here we provide additional details and employ a definition of sequential dependence that is relaxed from the version employed in Vilela and Hill (2020) and Hill and Lafortune (2016).

The work of Hill and Lafortune (2016) relies on the existence of sequentially-dependent orderings of events to create transitions with “virtual” zero-cost to improve the amount of abstraction achieved in the generation of the high-level in the hierarchy. However, their definition of sequential dependence is restricted by a consistent ordering of the occurrence of dependent events. Based on this definition, the pair of events (σa, σb) in the automaton G presented in Fig. 3 is not considered sequentially-dependent, since if we choose an arbitrary marked path that includes the event σa, we can have the order σaσb or σbσa. In this paper, we relax the sequential dependence requirement such that a specific ordering between the dependent events is not necessary. We will classify the pair of events (σa, σb) in the example of Fig. 3 to be a sequentially-dependent pair without requiring them to be a sequentially-dependent ordering. Under this new definition of sequential dependence, we also define a new lump operation. We will show that the new lump operation will not affect the overall cost of marked paths through the automaton G, maintaining the hierarchical approach to planning allowed by Theorem 1. Further, we will show that the compositional approach to generating the abstraction proposed in Vilela and Hill (2020) holds under the new relaxed sequential dependence definition.

Fig. 3
figure 3

Sequential dependence of the pair of events (σa, σb)

This new notion of sequential dependence is defined for a pair of events (σa, σb), such that for an arbitrary marked path, the occurrence of one of the events in the path implies the occurrence of the other event. For example, if we have two occurrences of σa in a string, then we will have two occurrences of σb as well, without regard to their ordering. The marked path could include any permutation of these events, {σaσaσbσb, σaσbσaσb, σaσbσbσa, σbσaσbσa, σbσaσaσb, σbσbσaσa}. We will represent such sets by a permutation function. This set, for example, will be represented \(Perm({\sigma _{a}^{2}},{\sigma _{b}^{2}})\), where \(\sigma _{a}^2 = \sigma _{a}\sigma _{a}\) and \(\sigma _{b}^2 = \sigma _{b}\sigma _{b}\). The new relaxed notion of sequential dependence is formally presented in Definition 2.

Definition 2

Consider an automaton \(G=(Q,{\Sigma }, \rightarrow , q_0,Q_m,f)\) which possesses a pair of events (σa, σb) with corresponding natural projections \(P_a:{\Sigma }^{*} \rightarrow \sigma _a^{*}\) and \(P_b:{\Sigma }^{*} \rightarrow \sigma _b^{*}\). The pair of events is sequentially dependent in G if ∀s ∈Σ such that \(q_0 \overset {s}{\rightarrow }q\) exists for some qQm with \(P_a(s)={\sigma _{a}^{n}}\), then \(P_b(s)={\sigma _{b}^{n}}\) where \(n \in \mathbb {N}\).

Based on this new sequential dependence definition, we update the definition of the lump operation in Definition 3. Because there are not specific dependent and independent events, the cost can be transferred in either direction between the events. For example, from σa to σb or from σb to σa. However, the direction of the lumping must be kept consistent throughout in the entire automaton.

Definition 3

Let there be a weighted automaton \(G=(Q,{\Sigma }, \rightarrow , q_0,Q_m,f)\), with σa, σb ∈Σ, where the pair of events (σa, σb) is sequentially dependent. If the pair of events are time-separable in G, with \(f(p\overset {\sigma _{a}}{\longrightarrow }p^{\prime })=c_{a}\), and \(f(q\overset {\sigma _{b}}{\longrightarrow }q^{\prime })=c_{b}\), then a lumped automaton \(G^{*}=(Q,{\Sigma }, \rightarrow , q_{0},Q_{m},f^{*})\) can be created by setting \(f^{*}(q\overset {\sigma _{b}}{\longrightarrow }q^{\prime })=0\), and \(f^{*}(p\overset {\sigma _{a}}{\longrightarrow }p^{\prime })=c_{a}+c_{b}\). The rest of the automaton remains unaltered.

Now, we need to show that the cost of a marked sequence is kept unaltered in the lumped automaton G following this new definition of the lumping operation.

Proposition 1

Let there be an automaton \(G=(Q, {\Sigma }, \rightarrow , q_{0}, Q_{m}, f)\) with a time-sepa-rable and sequentially-dependent pair of events (σa, σb) from which the lumped automaton \(G^{*}=(Q, {\Sigma }, \rightarrow , q_{0}, Q_{m}, f^{*})\) is generated. For all s ∈Σ such that \(q_{0} \overset {s}{\rightarrow }q_{n}\) exists for some qnQm, it is true that \(C(q_{0} \overset {s}{\rightarrow }q_{n})=C^{*}(q_{0} \overset {s}{\rightarrow }q_{n})\).

Because the cost of any marked sequence s in G is the same as in G, we have that an optimal plan found in G is also an optimal plan in G, as settled before by Hill and Lafortune (2016).

Hill and Lafortune (2016) employ the original weighted automaton G to generate a modified lumped automaton G, which is then used to generate the cost-equivalent, zero-cost reachable partition automaton Gh. The reduced Gh is used to find an optimal global plan (marked path). It is proven by Hill and Lafortune (2016) that all realizations in G of an optimal plan in Gh are also optimal, and that any path realized in G has equal cost in the original automaton G. Therefore, any realization in G of an optimal plan in (G)h is optimal as well (captured by Theorem 1).

This paper follows the same process of finding an optimal global plan developed by Hill and Lafortune (2016), however, it proposes a more efficient approach for generating the abstraction Gh. Given that G, generally, is composed of a set of smaller automata, G1G2∥…∥Gn, this paper finds conditions under which Gh can be generated compositionally, without the need to compute G first as was done in Hill and Lafortune (2016). Specifically, we show that it is possible to distribute the “lumping” operation over the parallel composition operator, and if the partitioning employs a newly defined class of equivalence, then Gh can be generated compositionally, Fig. 4. The new notion of priced-observation equivalence is stronger than the cost equivalence introduced in Hill and Lafortune (2016). Therefore, all results from Hill and Lafortune (2016) related to optimality of a plan between a partition automaton and its underlying automaton are also valid under priced-observation equivalence.

Fig. 4
figure 4

Different approaches to create the high-level of the hierarchy

With priced-observation equivalence, states must have futures with the same cost and the same observed event labels, as well as consistent marking. One can consider this equivalence a combination of cost equivalence and weak bisimulation.

Definition 4

Let there be two weighted automata, \(G_{1}=(Q_{1}, {\Sigma }, \rightarrow _{1}, q_{0,1}, Q_{m,1},f_{1})\) and \(G_{2}=(Q_{2}, {\Sigma }, \rightarrow _{2}, q_{0,2}, Q_{m,2},f_{2})\), and a set of hidden events (Σ ∖Σh). An equivalence relation \(\approx _{PO} \subseteq Q_{1} \times Q_{2}\) is a priced-observation equivalence if for any states q1Q1 and q2Q2, q1POq2 implies:

  • (i) if \(q_{1} \overset {s}{\Rightarrow }q_{1}^{\prime }\) for some s ∈Σh, then \(\exists q_{2}^{\prime } \in Q_{2}\) s.t. \(q_{2} \overset {s}{\Rightarrow } q_{2}^{\prime }\) where \(C(q_{1} \overset {s}{\Rightarrow }q_{1}^{\prime })= C(q_{2} \overset {s}{\Rightarrow } q_{2}^{\prime })\)\(q_{1}^{\prime } \in Q_{m,1} \Rightarrow q_{2}^{\prime } \in Q_{m,2}\)\(q_{1}^{\prime } \approx _{PO} q_{2}^{\prime }\); and

  • (ii) if \(q_{2} \overset {s}{\Rightarrow }q_{2}^{\prime }\) for some s ∈Σh, then \(\exists q_{1}^{\prime } \in Q_{1}\) s.t. \(q_{1} \overset {s}{\Rightarrow } q_{1}^{\prime }\) where \(C(q_{2} \overset {s}{\Rightarrow }q_{2}^{\prime })= C(q_{1} \overset {s}{\Rightarrow } q_{1}^{\prime })\)\(q_{2}^{\prime } \in Q_{m,2} \Rightarrow q_{1}^{\prime } \in Q_{m,1}\)\(q_{1}^{\prime } \approx _{PO} q_{2}^{\prime }\).

G1 and G2 are said to be priced-observation equivalent if q0,1POq0,2. This follows from Definition 4 since all reachable states also will be priced-observation equivalent. As an example of priced-observation equivalence, we have the automata G and Gh in Fig. 5, where the events {β1, β2} are abstracted away to generate Gh. In the abstraction, β1 and β2 are replaced by the silent event τ. Two groups of priced-observation equivalent states are then generated, resulting in the automaton Gh in Fig. 5c, which is priced-observation equivalent to G, GPOGh.

Fig. 5
figure 5

Example of priced-observation equivalence

We will show now that the global partition automaton (abstraction) can be generated compositionally. This requires that priced-observation equivalence be a congruence with respect to parallel composition. We will show that this is true when a partition system {G,Gh} is generated by a priced-observation equivalent abstraction, GPOGh, and the shared events are kept observable.

Proposition 2

Let there be a weighted automaton \(G_{1}=(Q_{1}, {\Sigma }_{1}, \rightarrow _{1}, q_{0,1}, Q_{m,1}, f_{1})\) from which a partition automaton \({G_{1}^{h}}=({Q_{1}^{h}},{{\Sigma }_{1}^{h}}, {\rightarrow _{1}^{h}}, q_{0,1}^{h}, Q_{m,1}^{h}, {f_{1}^{h}})\) is generated by priced-observation equivalent abstraction, \(G_{1}\approx _{PO} {G_{1}^{h}}\). For any third automaton \(G_{2}=(Q_{2}, {\Sigma }_{2}, \rightarrow _{2}, q_{0,2}, Q_{m,2}, f_{2})\), if the set of observed events \({{\Sigma }_{1}^{h}} \supseteq {\Sigma }_{1} \cap {\Sigma }_{2}\), then \(G_{1}\|G_{2} \approx _{PO} {G_{1}^{h}}\|G_{2}\).

Proof

In this proof we will establish that \(G_{1}\|G_{2} \approx _{PO} {G_{1}^{h}}\|G_{2}\) by demonstrating that \((q_{0,1},q_{0,2})\approx _{PO}(q_{0,1}^{h},q_{0,2})\).

  • From Definition 1 we have the initial state (q0,1, q0,2) for G1G2 and the initial state \((q_{0,1}^{h},q_{0,2})\) for \({G_{1}^{h}}\|G_{2}\).

    1) Let there be an observation \((q_{0,1},q_{0,2}) \overset {\sigma }{\Longrightarrow }_{1,2}(q_{1,1},q_{1,2})\) in G1G2, such that \(\sigma \in {{\Sigma }_{1}^{h}} \cup {\Sigma }_{2}\) and \({{\Sigma }_{1}^{h}} \supseteq {\Sigma }_{1} \cap {\Sigma }_{2}\). We will show that \(\exists (q_{0,1}^{h},q_{0,2})\overset {\sigma }{\Longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{1,2})\) in \({G_{1}^{h}}\|G_{2}\) with \(C((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\Longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{1,2}))=C((q_{0,1},q_{0,2}) \overset {\sigma }{\Longrightarrow }_{1,2}(q_{1,1},q_{1,2}))\) and that \((q_{1,1}^{h}, q_{1,2})\) agrees in marking with (q1,1, q1,2).

    The observation \((q_{0,1},q_{0,2}) \overset {\sigma }{\Longrightarrow }_{1,2}(q_{1,1},q_{1,2})\) implies the existence of the sequence \((q_{0,1},q_{0,2})\overset {t_1}{\longrightarrow }_{1,2}(q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{1,2})\overset {t_2}{\longrightarrow }_{1,2}(q_{1,1},q_{1,2})\) where \(t_i \in (({\Sigma }_1 \setminus {{\Sigma }_{1}^{h}}) \setminus {\Sigma }_2)^{*}\), since by assumption \({{\Sigma }_{1}^{h}} \supseteq {\Sigma }_1 \cap {\Sigma }_2\). Analysing this sequence of transitions by parts we have:

    1.i) \((q_{0,1},q_{0,2})\overset {t_1}{\longrightarrow }_{1,2}(q_{0,1}^{\prime },q_{0,2})\): We know that \(t_1 \in (({\Sigma }_1 \setminus {{\Sigma }_{1}^{h}}) \setminus {\Sigma }_2)^{*}\), then invoking Definition 1, we have that \(\exists q_{0,1}\overset {t_1}{\longrightarrow }_{1}q_{0,1}^{\prime }\), but \(\nexists q_{0,2}\overset {t_1}{\longrightarrow }_{2}q_{1,2}\). By construction of \({G_{1}^{h}}\), we have that \(q_{0,1}\approx _{PO}q_{0,1}^{\prime }\) and that \(q_{0,1}\approx _{PO}q_{0,1}^{h}\). Therefore, by transitivity, \(q_{0,1}\approx _{PO}q_{0,1}^{\prime }\approx _{PO}q_{0,1}^{h}\). Furthermore, by Definition 1, \((q_{0,1}^{h},q_{0,2}) \in Q_{m,1}^{h} \times Q_{m,2} \iff (q_{0,1},q_{0,2}) \in Q_{m,1} \times Q_{m,2}\).

    1.ii) \((q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{1,2})\): Since \(\sigma \in {\Sigma }_1^{h} \cup {\Sigma }_2\), there are three possible cases:

    1.ii.a) \(\sigma \in {\Sigma }_1^{h} \setminus {\Sigma }_2\): By Definition 1 we know that \(\exists q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime }\) and \(\nexists q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}\) and \(f_{1,2}((q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{0,2}))=f_1(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime })\). By assumption \(G_1 \approx _{PO} {G_{1}^{h}}\), we have \(q_{0,1}^{\prime }\approx _{PO}q_{0,1}^{h}\). Given \(\exists q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime }\) and invoking Definition 4, we have that \(\exists q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^h}q_{1,1}^{h}\) with \({f_{1}^{h}}(q_{0,1}^{h}\overset {\sigma }{\longrightarrow _{1^h}}q_{1,1}^{h})=f_{1}(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime })\) and \(q_{1,1}^{h} \in Q_{m,1}^{h} \iff q_{1,1}^{\prime } \in Q_{m,1}\). Therefore, by Definition 1, in \({G_{1}^{h}}\|G_2\) \(\exists (q_{0,1}^{h},q_{0,2})\overset {\sigma }{\longrightarrow }_{1^h,2}(q_{1,1}^{h},q_{0,2})\) with cost \(f_{1^h,2}((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\longrightarrow }_{1^h,2}(q_{1,1}^{h},q_{0,2}))= {f_{1}^{h}}(q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^h}q_{1,1}^{h})= f_{1}(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime })=f_{1,2}((q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{0,2}))\) and \((q_{1,1}^{h},q_{0,2}) \in Q_{m,1}^{h} \times Q_{m,2} \iff (q_{1,1}^{\prime },q_{0,2}) \in Q_{m,1} \times Q_{m,2}\).

    1.ii.b) \(\sigma \in {\Sigma }_1^h \cap {\Sigma }_2\): By Definition 1 we know that \(\exists q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime }\) and \(\exists q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}\) and \(f_{1,2}((q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{1,2}))=max(f_1(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime }),f_2(q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}))\). Since \(\exists q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_1q_{1,1}^{\prime }\) and \(q_{0,1} \approx _{PO} q_{0,1}^{\prime } \approx _{PO} q_{0,1}^{h}\), invoking Definition 4 provides that \(\exists q_{0,1}^{h} \overset {\sigma }{\longrightarrow }_{1^h}q_{1,1}^{h}\) in \({G_{1}^{h}}\) with cost \({f_{1}^{h}}(q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^h}q_{1,1}^{h})=f_1(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime })\) and \(q_{1,1}^{h} \in Q_{m,1}^h \iff q_{1,1}^{\prime } \in Q_{m,1}\). Therefore, by Definition 1, we have that \(\exists (q_{0,1}^{h},q_{0,2})\overset {\sigma }{\longrightarrow }_{1^h,2}(q_{1,1}^{h},q_{1,2})\) in \({G_{1}^{h}}\|G_2\) with cost \(f_{1^h,2}((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\longrightarrow }_{1^h,2}(q_{1,1}^{h},q_{1,2}))= max({f_{1}^{h}}(q_{0,1}^{h} \overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h}),f_{2}(q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}))=max(f_{1}(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime }),f_{2}(q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}))=f_{1,2}((q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{1,2}))\) and \((q_{1,1}^{h},q_{1,2}) \in Q_{m,1}^{h} \times Q_{m,2} \iff (q_{1,1}^{\prime },q_{1,2}) \in Q_{m,1} \times Q_{m,2}\).

    1.ii.c) \(\sigma \in {\Sigma }_{2} \setminus {{\Sigma }_{1}^{h}}\): This case follows from logic similar to that employed in 1.ii.a)

    1.iii) \((q_{1,1}^{\prime },q_{1,2})\overset {t_{2}}{\longrightarrow }_{1,2}(q_{1,1},q_{1,2})\): We know that \(t_{2} \in (({\Sigma }_{1} \setminus {{\Sigma }_{1}^{h}}) \setminus {\Sigma }_{2})^{*}\) thus this case follows from logic similar to that employed in 1.i).

  • So far we have shown that given \((q_{0,1},q_{0,2}) \overset {\sigma }{\Longrightarrow }_{1,2}(q_{1,1},q_{1,2})\) in G1G2, then \(\exists (q_{0,1}^{h},q_{0,2})\overset {\sigma }{\Longrightarrow }_{1^h,2}(q_{1,1}^{h},q_{1,2})\) in \({G_{1}^{h}}\|G_2\) with \( C((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\Longrightarrow }_{1^h,2}(q_{1,1}^{h},q_{1,2}))=C((q_{0,1}, q_{0,2})\overset {\sigma }{\Longrightarrow }_{1,2}(q_{1,1},q_{1,2}))\) and \((q_{1,1}^{h},q_{1,2})\) and (q1,1, q1,2) agree in marking. Applying the preceding logic over subsequent observations, we end up with the result that all observed sequences that exist in G1G2 also exist in \({G_{1}^{h}}\|G_{2}\) with the same cost and marking.

    2) The next step is to show the opposite direction. Let there be an observation \((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\Longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{1,2})\) in \({G_{1}^{h}}\|G_{2}\), such that \(\sigma \in {{\Sigma }_{1}^{h}} \cup {\Sigma }_{2}\) and \({{\Sigma }_{1}^{h}} \supseteq {\Sigma }_{1} \cap {\Sigma }_{2}\).

    The observation \((q_{0,1}^{h},q_{0,2}) \overset {\sigma }{\Longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{1,2})\) implies the existence of the sequence \((q_{0,1}^{h},q_{0,2})\overset {t_{1}\sigma t_{2}}{\longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{1,2})\), where \(t_{i} \in (({\Sigma }_{1} \setminus {{\Sigma }_{1}^{h}}) \setminus {\Sigma }_{2})^{*}\), therefore, the observation \((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\Longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{1,2})\) implies that \((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{1,2})\) in \({G_{1}^{h}}\|G_{2}\).

    We will now show the existence of the observation in G1G2 in a similar manner to direction 1) by proving the existence of a sequence \((q_{0,1},q_{0,2})\overset {t_{1}}{\longrightarrow }_{1,2}(q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{1,2})\overset {t_{2}}{\longrightarrow }_{1,2}(q_{1,1},q_{1,2})\) with consistent cost and marking.

    2.i) \((q_{0,1},q_{0,2})\overset {t_{1}}{\longrightarrow }_{1,2}(q_{0,1}^{\prime },q_{0,2})\): By assumption \({G_{1}^{h}} \approx _{PO} G_{1}\), thus \(q_{0,1}^{h}\approx _{PO} q_{0,1}\). However, in G1 there are hidden events, given \(t_{1} \in (({\Sigma }_{1} \setminus {{\Sigma }_{1}^{h}}) \setminus {\Sigma }_{2})^{*}\), then \(\exists q_{0,1}\overset {t_{1}}{\longrightarrow }_{1}q_{0,1}^{\prime }\) with \(q_{0,1} \approx _{PO} q_{0,1}^{\prime }\), by construction. By transitivity, \(q_{0,1}^{h} \approx _{PO}q_{0,1}^{\prime }\). Since t1∉Σ2, Definition 1 then provides that \(\exists (q_{0,1},q_{0,2})\overset {t_{1}}{\longrightarrow }_{1,2}(q_{0,1}^{\prime },q_{0,2})\) and \((q_{0,1}^{h},q_{0,2}) \in Q_{m,1}^{h} \times Q_{m,2} \iff (q_{0,1},q_{0,2}) \in Q_{m,1} \times Q_{m,2}\).

    2.ii) \((q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1},q_{1,2})\): Since \(\sigma \in {{\Sigma }_{1}^{h}} \cup {\Sigma }_{2}\), there are three possible cases:

    2.ii.a) \(\sigma \in {{\Sigma }_{1}^{h}}\setminus {\Sigma }_{2}\): By Definition 1 we know that \(\exists q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h}\) and \(\nexists q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}\) and \(f_{1,2}((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{0,2}))={f_{1}^{h}}(q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h})\). By assumption \({G_{1}^{h}} \approx _{PO} G_{1}\), we have \(q_{0,1}^{h}\approx _{PO}q_{0,1}^{\prime }\). Given \(\exists q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h}\) and invoking Definition 4, we have that \(\exists q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime }\) with \(f_{1}(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow _{1}}q_{1,1}^{\prime })={f_{1}^{h}}(q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h})\) and \(q_{1,1}^{\prime } \in Q_{m,1} \iff q_{1,1}^{h} \in Q_{m,1}^{h}\). Therefore, by Definition 1, in G1G2 \(\exists (q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{0,2})\) with cost \(f_{1,2}((q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{0,2}))= f_{1}(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime })={f_{1}^{h}}(q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h})=f_{1^{h},2}((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{0,2}))\) and \((q_{1,1}^{\prime },q_{0,2}) \in Q_{m,1} \times Q_{m,2} \iff (q_{1,1}^{h},q_{0,2}) \in Q_{m,1}^{h} \times Q_{m,2}\).

    2.ii.b) \(\sigma \in {{\Sigma }_{1}^{h}}\cap {\Sigma }_{2}\): By Definition 1 we know that \(\exists q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h}\) and \(\exists q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}\) and \(f_{1,2}((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\longrightarrow }_{1^{h},2}(q_{1,1}^{h},q_{1,2}))=max({f_{1}^{h}}(q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h}),f_{2}(q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}))\). Since \(\exists q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h}\) and \(q_{0,1}^{h} \approx _{PO} q_{0,1}^{\prime }\), invoking Definition 4 provides that \(\exists q_{0,1}^{\prime } \overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime }\) in G1 with cost \(f_{1}(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime })={f_{1}^{h}}(q_{0,1}^{h}\overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h})\) and \(q_{1,1}^{\prime } \in Q_{m,1} \iff q_{1,1}^{h} \in Q_{m,1}^{h}\). Therefore, by Definition 1, we have that \(\exists (q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{1,2})\) in G1G2 with cost \(f_{1,2}((q_{0,1}^{\prime },q_{0,2})\overset {\sigma }{\longrightarrow }_{1,2}(q_{1,1}^{\prime },q_{1,2}))=max(f_{1}(q_{0,1}^{\prime }\overset {\sigma }{\longrightarrow }_{1}q_{1,1}^{\prime }),f_{2}(q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}))=max({f_{1}^{h}}(q_{0,1}^{h} \overset {\sigma }{\longrightarrow }_{1^{h}}q_{1,1}^{h}),f_{2}(q_{0,2}\overset {\sigma }{\longrightarrow }_{2}q_{1,2}))=f_{1^{h},2}((q_{0,1}^{h},q_{0,2})\overset {\sigma }{\longrightarrow }_{1^h,2}(q_{1,1}^{h},q_{1,2}))\) and \((q_{1,1}^{\prime },q_{1,2}) \in Q_{m,1} \times Q_{m,2} \iff (q_{1,1}^{h},q_{1,2}) \in Q_{m,1}^{h} \times Q_{m,2}\).

    2.ii.c) \(\sigma \in {\Sigma }_2 \setminus {{\Sigma }_{1}^{h}}\): This case follows from logic similar to that employed in 2.ii.a).

    2.iii) \((q_{1,1}^{\prime },q_{1,2})\overset {t_2}{\longrightarrow }_{1,2}(q_{1,1},q_{1,2})\): This case follows from logic similar to that employed in 2.i).

  • Applying this same logic to subsequent observations, we end up with the result that all observed sequences in \({G_{1}^{h}}\|G_2\) also exist in G1G2, with the same cost and marking.

  • Based on 1) and 2) and invoking Definition 4, we have that \((q_{0,1},q_{0,2})\approx _{PO}(q_{0,1}^{h},q_{0,2})\), thus \(G_{1}\|G_{2} \approx _{PO} {G_{1}^{h}}\|G_{2}\).

The fact that priced-observation equivalence is a congruence with respect to the parallel composition is an important result that allows the global partition automaton Gh to be generated compositionally as proven in the following theorem.

Theorem 2

Let there be two weighted automata \(G_{1}=(Q_{1}, {\Sigma }_{1}, \rightarrow _{1}, q_{0,1}, Q_{m,1}, f_{1})\) and \(G_{2}=(Q_{2},{\Sigma }_{2}, \rightarrow _{2}, q_{0,2}, Q_{m,2}, f_{2})\). If the partition automata are generated by a priced-observation equivalent abstraction and \({\Sigma }^{h} \supseteq {\Sigma }_{1} \cap {\Sigma }_{2}\), then \(G^{h}=(G_{1}\|G_{2})^{h} \approx _{PO}G_1^h\|G_2^h\).

Proof

Generating a partition automaton for G1 using priced-observation equivalence, we have \(G_1 \approx _{PO} G_1^h\). By Proposition 2, we then have \(G_1 \| G_2 \approx _{PO} {G_{1}^{h}}\|G_2\). Generating a partition automaton for G2, again using priced-observation equivalence, we have \(G_2 \approx _{PO} {G_{2}^{h}}\). Again invoking Proposition 2 with \(G_1^{h}\) as the third automaton, we have \({G_{1}^{h}}\|G_2 \approx _{PO} {G_{1}^{h}}\|{G_{2}^{h}}\) since parallel composition is commutative. By transitivity, we end up with \(G_1\|G_2 \approx _{PO} {G_{1}^{h}}\|{G_{2}^{h}}\). By construction, GPOGh, so G1G2PO(G1G2)h. Therefore, again by transitivity, we have \(G^h=(G_1\|G_2)^h \approx _{PO} {G_{1}^{h}}\|{G_{2}^{h}}\). □

Now we desire that the relaxed version of sequential dependence be preserved over parallel composition. Proposition 3 shows that this is true if the pair of events is sequentially dependent in all automata that have those events in their alphabet.

Proposition 3

Let there be two weighted automata \(G_1=(Q_1, {\Sigma }_1, \rightarrow _1, q_{0,1}, Q_{m,1},f_1)\) and \(G_2=(Q_2, {\Sigma }_2, \rightarrow _2, q_{0,2}, Q_{m,2},f_2)\), where G1 has a sequentially-dependent pair of events (σa, σb). The parallel composition G1G2 will also have the sequentially-dependent pair of events (σa, σb), if:

  • \(\sigma _a,\sigma _b \in \left ({\Sigma }_1 \cap {\Sigma }_2 \right )\) and G2 also has the sequentially-dependent pair of events (σa, σb); or,

  • \(\sigma _a,\sigma _b \in \left ({\Sigma }_1 \setminus {\Sigma }_2 \right )\).

Proof

Let Σ = Σ1 ∪Σ2, \(P_a:{\Sigma }^{*} \rightarrow \sigma _a^{*}\), \(P_b:{\Sigma }^{*} \rightarrow \sigma _b^{*}\), \(P_{ab}:{\Sigma }^{*} \rightarrow \{\sigma _a,\sigma _b\}^{*}\), \(P_1:{\Sigma }^{*} \rightarrow {\Sigma }_{1}^{*}\), and \(P_2:{\Sigma }^{*} \rightarrow {\Sigma }_{2}^{*}\). It is given that the pair of events (σa, σb) is sequentially-dependent in G1 and σa, σb ∈Σ1.

  • Following Definition 2, we first need to show that ∀sLm(G1G2) for which \(P_a(s)={\sigma _{a}^{n}}\), we have that \(P_{b}(s)={\sigma _{b}^{n}}\).

  • Let \(s \in L_m(G_1\|G_2) = P_{1}^{-1}(L_m(G_1)) \cap P_{2}^{-1}(L_m(G_2))\), and \(P_a(s)={\sigma _{a}^{n}}\). Thus, \(s\in P_{1}^{-1}(L_m(G_{1}))\) and \(s\in P_{2}^{-1}(L_{m}(G_{2}))\).

  • Since \(s \in P_{1}^{-1}(L_{m}(G_{1}))\), ∃s1Lm(G1) such that P1(s) = s1.

  • Also since \(s \in P_{2}^{-1}(L_{m}(G_{2}))\), ∃s2Lm(G2) such that P2(s) = s2.

  • Given σa ∈Σ1 and \(P_{1}(P_{a}(s))={\sigma _{a}^{n}}\), then \(P_{a}(P_{1}(s))=P_{a}(s_{1})={\sigma _{a}^{n}}.\)

  • G1 has the sequentially-dependent pair of events (σa, σb), so given s1Lm(G1) and \(P_{a}(s_{1})={\sigma _{a}^{n}}\), from Definition 2, \(P_{b}(s_{1})={\sigma _{b}^{n}}\). Also, \(P_{ab}(s_{1})=r_{1} \in Perm({\sigma _{a}^{n}},{\sigma _{b}^{n}})\).

  • Now analyzing the two possible cases:

  • i) for the case σa, σb ∈Σ1 ∩Σ2: from the assumption \(P_{a}(s)={\sigma _{a}^{n}}\), we have \(P_{a}(P_{2}(s))=P_{a}(s_{2})={\sigma _{a}^{n}}\) since σa ∈Σ2. The pair of events (σa, σb) is also sequentially dependent in G2, then from Definition 2, \(P_{b}(s_{2})={\sigma _{b}^{n}}\) and \(P_{ab}(s_{2})=r_{2}\in Perm({\sigma _{a}^{n}},{\sigma _{b}^{n}})\). Since \(P_{ab}(P_{1}^{-1}(s_{1}))=r_{1}\) and \(P_{ab}(P_{2}^{-1}(s_{2}))=r_{2}\), \(P_{ab}(P_{1}^{-1}(s_{1}))\cap P_{ab}(P_{2}^{-1}(s_{2})) = \epsilon \) if r1r2, or \(P_{ab}(P_{1}^{-1}(s_{1}))\cap P_{ab}(P_{2}^{-1}(s_{2})) = r_{1}\), if r1 = r2. Therefore, \(P_{ab}(P_{1}^{-1}(s_{1}))\cap P_{ab}(P_{2}^{-1}(s_{2}))\) equals ε or r1.

  • ii) for σa, σb ∈Σ1 ∖Σ2, Pab(s2) = ε, and \(P_{ab}(P_{2}^{-1}(s_{2}))= (\sigma _{a},\sigma _{b})^{*}\). Therefore, \(P_{ab}(P_{1}^{-1}(s_{1}))\cap P_{ab}(P_{2}^{-1}(s_{2})) = r_{1}\)

  • Considering both i) and ii), \(P_{ab}(P_{1}^{-1}(s_{1})) \cap P_{ab}(P_{2}^{-1}(s_{2})) \subset \{ \epsilon , r_{1}\}\). From a property of natural projection, \(P_{ab}(P_{1}^{-1}(s_{1}) \cap P_{2}^{-1}(s_{2})) \subseteq P_{ab}(P_{1}^{-1}(s_{1})) \cap P_{ab}(P_{2}^{-1}(s_{2})) \subset \{ \epsilon , r_{1}\}\).

  • Because \(s \in P_{1}^{-1}(s_{1}) \cap P_{2}^{-1}(s_{2}) \) we have Pab(s) ⊂{𝜖,r1}. This provides that Pab(s) = ε or Pab(s) = r1.

  • Since it is given \(P_{a}(s)={\sigma _{a}^{n}}\), this removes the possibility of Pab(s) = ε. Therefore, Pab(s) = r1.

  • Since \(r_{1} \in Perm({\sigma _{a}^{n}},{\sigma _{b}^{n}})\), we have that \(P_{b}(s)={\sigma _{b}^{n}}\). Therefore, the pair of events (σa, σb) is sequentially-dependent in G1G2.

The sequential dependence property can then be employed to generate the lumped automaton G as defined in Definition 3. Specifically, it is desired to perform the “lumping” compositionally. Conditions under which the lumping operation can be distributed over parallel composition, \(G^{*}=G_{1}^{*}\|G_{2}^{*}\), are presented in Propositions 4 and 5, which correspond to cases i) and ii) of Proposition 3, respectively.

Proposition 4

Let there be two weighted automata \(G_{1}=(Q_{1}, {\Sigma }_{1},\rightarrow _{1},q_{0,1}, Q_{m,1}, f_{1})\) and \(G_{2}=(Q_{2}, {\Sigma }_{2},\rightarrow _{2},q_{0,2}, Q_{m,2}, f_{2})\) with sequentially dependent pair of events (σa, σb) such that σa, σb ∈ (Σ1 ∩Σ2). If G1 and G2 are time-separable for σa and σb, with transition costs such that \(f_1(q_{a,1}\overset {\sigma _a}{\longrightarrow _1}q_{a,1}^{\prime }) \geq f_2(q_{a,2}\overset {\sigma _a}{\longrightarrow _2}q_{a,2}^{\prime })\) and \(f_1(q_{b,1}\overset {\sigma _b}{\longrightarrow _1}q_{b,1}^{\prime }) \geq f_2(q_{b,2}\overset {\sigma _b}{\longrightarrow _2}q_{b,2}^{\prime })\), then \(G^{*}=(G_1\|G_2)^{*}=G_1^{*}\|G_2^{*}\).

Proof

Let there be a pair of events (σa, σb) that is sequentially dependent in G1 and G2, where σa, σb ∈ (Σ1 ∩Σ2). It follows from Proposition 3 Case i), that the pair of events (σa, σb) is also sequentially dependent in G = G1||G2. We wish to show that the cost of corresponding transitions with labels σa and σb are the same in (G1G2) as in \(G_{1}^{*}\|G_{2}^{*}\). Letting f be the cost function for (G1G2) and \(f_{1,2}^{*}\) the cost function for \(G_{1}^{*}\|G_{2}^{*}\), we will show that \(f^{*}((q_{a,1},q_{a,2}) \overset {\sigma _a}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))=f_{1,2}^{*}((q_{a,1},q_{a,2}) \overset {\sigma _a}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))\) and \(f^{*}((q_{b,1},q_{b,2}) \overset {\sigma _b}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))=f_{1,2}^{*}((q_{b,1},q_{b,2}) \overset {\sigma _{b}}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))\), for all such transitions.

i) We begin with (G1G2). Consider transitions corresponding to σa \(q_{a,1}\overset {\sigma _{a}}{\longrightarrow _{1}}q_{a,1}^{\prime }\) in G1 and \(q_{a,2}\overset {\sigma _{a}}{\longrightarrow _{2}}q_{a,2^{\prime }}\) in G2. Noting that σa, σb ∈ (Σ1 ∩Σ2) and following Definition 1, we have the transition \((q_{a,1},q_{a,2}) \overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime })\) in G1G2, with cost \(f((q_{a,1},q_{a,2})\overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))=max(f_{1}(q_{a,1}\overset {\sigma _{a}}{\rightarrow _{1}}q_{a,1}^{\prime }),\)\(f_{2}(q_{a,2}\overset {\sigma _{a}}{\longrightarrow _{2}}q_{a,2}^{\prime }))\). By assumption that \(f_{1}(q_{a,1}\overset {\sigma _{a}}{\longrightarrow _{1}}q_{a,1}^{\prime }) \geq f_{2}(q_{a,2}\overset {\sigma _{a}}{\longrightarrow _{2}}q_{a,2}^{\prime }) \), we then have that \(f((q_{a,1},q_{a,2})\overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))=f_{1}(q_{a,1}\overset {\sigma _{a}}{\rightarrow _{1}}q_{a,1}^{\prime })\). Following the same logic for σb transitions \(q_{b,1}\overset {\sigma _{b}}{\longrightarrow _{1}}q_{b,1}^{\prime }\) in G1 and \(q_{b,2}\overset {\sigma _{b}}{\longrightarrow _{2}}q_{b,2^{\prime }}\) in G2, we have that \(f((q_{b,1},q_{b,2}) \overset {\sigma _{b}}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))=f_{1}(q_{b,1}\overset {\sigma _{b}}{\longrightarrow _{1}}q_{b,1}^{\prime })\).

Applying the lump operation of Definition 3 to generate (G1G2), lumping the cost from σb to σa, we have that \(f^{*}((q_{a,1},q_{a,2})\overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))=f((q_{a,1},q_{a,2})\overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))+f((q_{b,1},q_{b,2})\overset {\sigma _{b}}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))= f_{1}(q_{a,1}\overset {\sigma _{a}}{\rightarrow _{1}}q_{a,1}^{\prime }) + f_{1}(q_{b,1}\overset {\sigma _{b}}{\longrightarrow _{1}}q_{b,1}^{\prime })\), and \(f^{*}((q_{b,1},q_{b,2})\overset {\sigma _{b}}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))=0\).

ii) Now we analyze \(G_{1}^{*}\|G_{2}^{*}\). The costs associated with the transitions related to σa and σb, lumping the cost from σb to σa, for \(G_{1}^{*}\) are as follows: \(f_{1}^{*}(q_{a,1}\overset {\sigma _{a}}{\longrightarrow _{1}}q_{a,1}^{\prime })=f_{1}(q_{a,1}\overset {\sigma _{a}}{\longrightarrow _{1}}q_{a,1}^{\prime })+f_{1}(q_{b,1}\overset {\sigma _{b}}{\longrightarrow _{1}}q_{b,1}^{\prime })\) and \(f_{1}^{*}(q_{b,1}\overset {\sigma _{b}}{\longrightarrow _{1}}q_{b,1}^{\prime })=0\); similarly for \(G_{2}^{*}\): \(f_{2}^{*}(q_{a,2}\overset {\sigma _{a}}{\longrightarrow _{2}}q_{a,2}^{\prime })=f_{2}(q_{a,2}\overset {\sigma _{a}}{\longrightarrow _{2}}q_{a,2}^{\prime })+f_{2}(q_{b,2}\overset {\sigma _{b}}{\longrightarrow _{2}}q_{b,2}^{\prime })\) and \(f_{2}^{*}(q_{b,2}\overset {\sigma _{b}}{\longrightarrow _{2}}q_{b,2}^{\prime })=0\).

Noting that σa, σb ∈ (Σ1 ∩Σ2) and invoking Definition 1, the transition costs in \(G_{1}^{*} \| G_{2}^{*}\) are: \(f_{1,2}^{*}((q_{a,1},q_{a,2}))\overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))=max(f_{1}^{*}(q_{a,1}\overset {\sigma _{a}}{\longrightarrow _{1}}q_{a,1}^{\prime }),f_{2}^{*}(q_{a,2}\overset {\sigma _{a}}{\longrightarrow _{2}}q_{a,2}^{\prime }))\), by assumption \(f_{1}(q_{a,1}\overset {\sigma _{a}}{\longrightarrow _{1}}q_{a,1}^{\prime }) \geq f_{2}(q_{a,2}\overset {\sigma _{a}}{\longrightarrow _{2}}q_{a,2}^{\prime })\) and \(f_{1}(q_{b,1}\overset {\sigma _{b}}{\longrightarrow _{1}}q_{b,1}^{\prime }) \geq f_{2}(q_{b,2}\overset {\sigma _{b}}{\longrightarrow _{2}}q_{b,2}^{\prime })\), therefore, \(f_{1}(q_{a,1}\overset {\sigma _{a}}{\longrightarrow _{1}}q_{a,1}^{\prime }) + f_{1}(q_{b,1}\overset {\sigma _{b}}{\longrightarrow _{1}}q_{b,1}^{\prime }) \geq f_{2}(q_{a,2}\overset {\sigma _{a}}{\longrightarrow _{2}}q_{a,2}^{\prime })+f_{2}(q_{b,2}\overset {\sigma _b}{\longrightarrow _2}q_{b,2}^{\prime })\), and thus, \(f_{1,2}^{*}((q_{a,1},q_{a,2}))\overset {\sigma _a}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))=f_{1}^{*}(q_{a,1}\overset {\sigma _a}{\longrightarrow _1}q_{a,1}^{\prime })=\)\(f_1(q_{a,1}\overset {\sigma _{a}}{\longrightarrow _{1}}q_{a,1}^{\prime }) + f_{1}(q_{b,1}\overset {\sigma _{b}}{\longrightarrow _{1}}q_{b,1}^{\prime })\) and \(f_{1,2}^{*}((q_{b,1},q_{b,2})\overset {\sigma _{b}}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))=0\).

  • Comparing results from i) and ii), we have \(f^{*}((q_{a,1},q_{a,2})\overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))=f_{1,2}^{*}((q_{a,1},q_{a,2})\overset {\sigma _a}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))=f_1(q_{a,1}\overset {\sigma _a}{\longrightarrow _1}q_{a,1}^{\prime })+f_1(q_{b,1}\overset {\sigma _b}{\longrightarrow _1}q_{b,1}^{\prime })\) and \(f^{*}((q_{b,1},q_{b,2})\overset {\sigma _{b}}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))=f_{1,2}^{*}((q_{b,1},q_{b,2})\overset {\sigma _{b}}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))=0\).

  • The same logic is valid if we lump the cost of σa to σb.

  • Having demonstrated that the transitions have the same cost, in general, we have demonstrated that \(G^{*}=(G_{1}\|G_{2})^{*}=G_{1}^{*}\|G_{2}^{*}\).

Proposition 5

Let there be two weighted automata \(G_{1}=(Q_{1}, {\Sigma }_{1},\rightarrow _{1},q_{0,1}, Q_{m,1}, f_{1})\) and \(G_{2}=(Q_{2}, {\Sigma }_{2},\rightarrow _{2},q_{0,2}, Q_{m,2}, f_{2})\) with σa, σb ∈Σ1 ∖Σ2. If the pair of events (σa, σb) is sequentially dependent and time-separable in G1, then \(G^{*}=(G_{1}\|G_{2})^{*}=G_{1}^{*}\|G_{2}^{*}\).

Proof

Let there be the pair of events (σa, σb) that are sequentially dependent in G1, where σa, σb ∈Σ1 ∖Σ2. We wish to show that the cost of corresponding transitions with labels σa and σb are the same in (G1G2) as in \(G_{1}^{*}\|G_{2}^{*}\). Letting f be the cost function for (G1G2) and \(f_{1,2}^{*}\) the cost function for \(G_{1}^{*}\|G_{2}^{*}\), we will show that \(f^{*}((q_{a,1},q_{a,2}) \overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))=f_{1,2}^{*}((q_{a,1},q_{a,2}) \overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}^{\prime }))\) and \(f^{*}((q_{b,1},q_{b,2})\overset {\sigma _b}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))=f_{1,2}^{*}((q_{b,1},q_{b,2}) \overset {\sigma _b}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}^{\prime }))\), for all such transitions.

i) We begin with (G1G2). Consider the transitions \(q_{a,1}\overset {\sigma _a}{\longrightarrow _1}q_{a,1}^{\prime }\) and \(q_{b,1}\overset {\sigma _b}{\longrightarrow _1}q_{b,1}^{\prime }\) in G1. Noting that σa, σb ∈Σ1 ∖Σ2 and following Definition 1, we have the transitions \((q_{a,1},q_{a,2}) \overset {\sigma _a}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2})\) and \((q_{b,1},q_{b,2})\overset {\sigma _b}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2})\) in G1G2, with costs \(f((q_{a,1},q_{a,2})\overset {\sigma _a}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}))=f_1(q_{a,1}\overset {\sigma _a}{\rightarrow _1}q_{a,1}^{\prime })\) and \(f((q_{b,1},q_{b,2})\overset {\sigma _b}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}))=f_1(q_{b,1}\overset {\sigma _b}{\rightarrow _1}q_{b,1}^{\prime })\). Applying the lump operation of Definition 3, lumping the cost of σb to σa, to generate (G1G2), we have \(f^{*}((q_{a,1},q_{a,2})\overset {\sigma _a}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}))=f_1(q_{a,1}\overset {\sigma _a}{\rightarrow _1}q_{a,1}^{\prime })+f_1(q_{b,1}\overset {\sigma _b}{\rightarrow _1}q_{b,1}^{\prime })\) and \(f^{*}((q_{b,1},q_{b,2})\overset {\sigma _b}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}))=0\).

ii) Now we analyze \(G_{1}^{*}\|G_{2}^{*}\). The costs associated with the transitions of \(G_1^{*}\) maintaining the same order of lumping done before, are as follows: \(f_{1}^{*}(q_{a,1}\overset {\sigma _a}{\longrightarrow _1}q_{a,1}^{\prime })=f_1(q_{a,1}\overset {\sigma _a}{\longrightarrow _1}q_{a,1}^{\prime })+f_1(q_{b,1}\overset {\sigma _b}{\longrightarrow _1}q_{b,1}^{\prime })\) and \(f_{1}^{*}(q_{b,1}\overset {\sigma _b}{\longrightarrow _1}q_{b,1}^{\prime })=0\). Since σa, σb∉Σ2, and following Definition 1, we have \(f_{1,2}^{*}((q_{a,1},q_{a,2})\overset {\sigma _a}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}))=f_{1}^{*}(q_{a,1}\overset {\sigma _a}{\rightarrow _1}q_{a,1}^{\prime })=f_1(q_{a,1}\overset {\sigma _a}{\rightarrow _1}q_{a,1}^{\prime })+f_1(q_{b,1}\overset {\sigma _b}{\rightarrow _1}q_{b,1}^{\prime })\) and \(f_{1,2}^{*}((q_{b,1},q_{b,2})\overset {\sigma _b}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}))=f_{1}^{*}(q_{b,1}\overset {\sigma _{b}}{\rightarrow _{1}}q_{b,1}^{\prime })=0\).

  • Comparing results from i) and ii), we have \(f^{*}((q_{a,1},q_{a,2})\overset {\sigma _{a}}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}))=f_{1,2}^{*}((q_{a,1},q_{a,2})\overset {\sigma _a}{\longrightarrow }(q_{a,1}^{\prime },q_{a,2}))=f_1(q_{a,1}\overset {\sigma _a}{\rightarrow _1}q_{a,1}^{\prime })+f_1(q_{b,1}\overset {\sigma _b}{\rightarrow _1}q_{b,1}^{\prime })\) and \(f^{*}((q_{b,1},q_{b,2})\overset {\sigma _{b}}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2}))=f_{1,2}^{*}((q_{b,1},q_{b,2})\overset {\sigma _{b}}{\longrightarrow }(q_{b,1}^{\prime },q_{b,2})=0\).

  • The same logic is valid if we lump the cost of σa to σb.

  • Having demonstrated that the transitions have the same cost, in general, we have demonstrated that \(G^{*}=(G_{1}\|G_{2})^{*}=G_{1}^{*}\|G_{2}^{*}\).

The conditions of Propositions 4 and 5 might seem restrictive, however, many systems are composed of plant components with disjoint event sets where all interaction takes place through specifications. In this case, the transitions within the specifications may be assumed to have zero cost with each event (and its cost) “owned” by a single plant component. In such a situation, the required conditions would be satisfied to distribute the lump operation over parallel composition as defined in Definition 1.

Figure 6 provides an example that illustrates Proposition 4. The automaton G is composed of two component automata, such that G = G1G2. The pair of events (σa, σb) is sequentially dependent in G1 and G2. For the transitions with labels σa and σb, each instance has greater cost in G2 than in G1. As predicted, \(G_{1}^{*} \| G_{2}^{*}\) results in the same automaton as (G1||G2).

Fig. 6
figure 6

Example: Distribution of lump operation over parallel composition

Now that we have shown it is possible to distribute the lump operation over parallel composition and that the partition automaton can be generated compositionally, we can combine these results to present the main theorem of this paper. This result requires that events shared between two weighted automata, G1 and G2, cannot be hidden. Furthermore, if the shared events are sequentially dependent, then all transitions for each dependent pair of events must have greater or equal cost in one of the two automata. Under these conditions, the abstracted automaton (G)h can be generated compositionally \((G^{*})^{h} \approx _{PO} (G_{1}^{*})^{h} \|(G_{2}^{*})^{h}\), without the need to first construct the full unabstracted automaton G. This significantly improves the scalability of the results from Hill and Lafortune (2016) by making it possible to generate the abstracted model, while avoiding the explosion of the state space that can arise from the construction of the global model G.

Theorem 3

Let there be two weighted automata \(G_{1}=(Q_{1}, {\Sigma }_{1}, \rightarrow _{1}, q_{0,1}, Q_{m,1}, f_{1})\) and \(G_{2}=(Q_{2},{\Sigma }_{2}, \rightarrow _{2}, q_{0,2}, Q_{m,2}, f_{2})\). For each sequentially-dependent pair of events (σa, σb), either let σa, σb ∈Σ1 ∖Σ2 or σa, σb ∈ (Σ1 ∩Σ2), and let G1 and G2 be time-separable for σa and σb with costs: \(f_1(q_{a,1}\overset {\sigma _a}{\longrightarrow _1}q_{a,1}^{\prime }) \geq f_2(q_{a,2}\overset {\sigma _a}{\longrightarrow _2}q_{a,2}^{\prime })\) and \(f_1(q_{b,1} \overset {\sigma _{b}}{\longrightarrow _{1}}q_{b,1}^{\prime }) \geq f_{2}(q_{b,2}\overset {\sigma _{b}}{\longrightarrow _{2}}q_{b,2}^{\prime })\). Let G be the lumped automaton generated from G = G1G2. If the partition automata are generated by priced-observation equivalent abstraction and \({\Sigma }^{h} \supseteq {\Sigma }_{1} \cap {\Sigma }_{2}\), then \((G^{*})^{h} \approx _{PO} (G_{1}^{*})^{h} \|(G_{2}^{*})^{h}\).

Proof

By assumption, each arbitrary sequentially-dependent pair of events (σa, σb) with σa, σb ∈ (Σ1 ∩Σ2) has corresponding transitions with costs that satisfy \(f_1(q_{a,1}\overset {\sigma _a}{\longrightarrow _1}q_{a,1}^{\prime }) \geq f_2(q_{a,2}\overset {\sigma _a}{\longrightarrow _2}q_{a,2}^{\prime }) \) and \(f_1(q_{b,1} \overset {\sigma _b}{\longrightarrow _1}q_{b,1}^{\prime }) \geq f_2(q_{b,2}\overset {\sigma _b}{\longrightarrow _2}q_{b,2}^{\prime })\). Invoking Proposition 4, we have \(G^{*}=G_{1}^{*}\|G_{2}^{*}\). Similarly, each arbitrary sequentially-dependent pair of events (σa, σb) with σa, σb ∈Σ1 ∖Σ2, also provides that \(G^{*}=G_{1}^{*}\|G_{2}^{*}\) by Proposition 5. Thus \((G^{*})^h = (G_{1}^{*}\|G_{2}^{*})^h\). Applying Theorem 2 then provides that for a priced-observation equivalent abstraction with \({\Sigma }^h \supseteq {\Sigma }_1 \cap {\Sigma }_2\), we have \((G^{*})^h=(G_{1}^{*}\|G_{2}^{*})^h \approx _{PO} (G_{1}^{*})^h \| (G_{2}^{*})^h\). □

4 Computational examples

In order to illustrate the advantages of hierarchical planning using the compositional approach to abstraction proposed in this paper, we will apply the technique to the example of cooperative robot control introduced in Hill and Lafortune (2017).

In the approach of this paper, a planning technique is applied to the abstraction of a set of automata composed together. There are no requirements on how these automata are generated, but it is likely in many cases that these automata will capture the set of safe and nonblocking behaviors of the system, in other words, the system’s supervised behavior. More specifically, the idea is to create a model for the problem by generating a classic supervisor based on the controllable and nonblocking sublanguage which provides the possible legal behaviors of the system. In the hierarchical planning proposed by Hill and Lafortune (2016), an abstraction of the supervisor is created based on their notion of cost equivalence, and this abstraction becomes the search space for the planning technique. A limitation with their work is that you still need to create the full model (supervisor) before generating the abstraction, which can be computationally prohibitive. The proposed approach of this paper makes it possible to create the abstracted model for hierarchical planning in a compositional manner. In order to apply this approach, we will first generate a set of modular supervisors, rather than generating the monolithic supervisor. In the following, the models and monolithic and compositional approaches to abstraction are described for our motivating example.

Let’s first consider a scenario with two robots, A and B, four tasks, and four regions, as illustrated in Fig. 7.

Fig. 7
figure 7

Cooperative control of multiple robots: Scenario with 2 robots, 4 tasks, and 4 regions [(Hill and Lafortune 2017)]

Each robot is physically capable of performing only one task at a time. This behavior is captured for robot A by the automaton \(R_{A_{tasks}}\) shown in Fig. 8a, where each task consists of a start event a#s and a finish event a#f. The analogous model is built for robot B, \(R_{B_{tasks}}\), where the start events have the form b#s and the finish events have the form b#f. There is a requirement on ordering precedence between the tasks as illustrated by the automata \(Spec_{tasks_{1,2}}\) and \(Spec_{tasks_{3,4}}\), shown in Fig. 9. The automaton \(Spec_{tasks_{1,2}}\) represents that task 1 must be performed prior to task 2 and by the same robot. This could represent, for example, that each odd-numbered task involves picking up an item at some location, while the corresponding even-numbered task involves dropping the item at a different location. Each transition for starting a task (a#s,b#s) possesses the cost associated with performing that task (distance to that location), while all transitions for finishing a task (a#f,b#f) have zero cost. Additionally, all start events are controllable and all finish events are uncontrollable.

Fig. 8
figure 8

Models for Robot A

Fig. 9
figure 9

Specification for task precedence

We also model how the robots move between regions (numbered 5, 6 ,7, 8), as well as the region in which each task is located. From its current position, a robot can move to any adjacent region, but it cannot move diagonally between regions. The automaton in Fig. 8b is for robot A, where events a#e signify the entry into a region. Likewise, an automaton is built for robot B, with a change in the initial state since robot B starts in region 8. As a restriction, it is specified we cannot have more than one robot in the same region at the same time, as modeled in Fig. 10. All entry events (a#e,b#e) are controllable and could be assigned a cost based on the distance from the preceding location to the point of region entry.

Fig. 10
figure 10

Specification for region avoidance, Specregions

The plant model related to task completion is \(G_{tasks}=R_{A_{tasks}}\|R_{B_{tasks}}\), and the plant model related to robot movement is \(G_{regions}=R_{A_{geo}}\|R_{B_{geo}}\). The overall plant model is G = GtasksGregions. The full model for the specifications is E = Spec1,2Spec3,4Specregions. The desired behavior for the whole system is K = Lm(GE). The monolithic supervisor is obtained by S = SupC(K,Lm(G)).

In the supervisor S, entry events will always follow a start event. Sub-trajectories through the supervisor automaton have the form that a task is started, some regions are possibly entered, then the task is completed. Such a pattern almost implies sequential dependence, except that the robot could take different paths to a task (it could enact different region entry events). If the entry events had cost zero and became candidates to be abstracted, it would greatly increase the potential for abstraction and would improve the decoupling between the high-level planning and the low-level planning. A reasonable approximation is to treat the entry events as having zero cost and treat the start events as having cost equal to the entire straight-line distance to the task. In such a case, the entry events become candidates to be abstracted while maintaining priced-observation equivalence and zero-cost reachability with respect to the original automaton.

Based on this approximation, we can consider finish and entry events, both having cost zero, to be candidates for abstraction. For an arbitrary automaton A, we will denote an abstraction Ah more specifically as \(A^{(\sigma _{1}, \sigma _{2},...,\sigma _{n})}\) to signify that the events (σ1, σ2,...,σn) have been hidden. In the monolithic approach, the high-level abstraction is generated from the monolithic supervisor S = SupC(K,Lm(G)) by abstracting all finish and entry events at once Sh = S(a#f,b#f,a#e,b#e).

In order to apply Theorem 2 to generate our high-level abstraction compositionally, we first generate a set of modular supervisors that meet all of the provided specifications for our system of robots. We will employ the approach of Hill and Tilbury (2008) to generate the modular supervisors incrementally. Consider the following steps as an example for the scenario with two robots, four tasks and four regions:

  • Savoid = SupC(Lm(GregionsSpecregions),Lm(Gregions));

  • \(S^{h}_{avoid}=S^{(a\#e,b\#e)}_{avoid}\);

  • \(G_{12}=G_{tasks}\|S^{h}_{avoid}\);

  • \(S_{12}=SupC(L_{m}(G_{12}\|Spec_{tasks_{1,2}}),L_{m}(G_{12}))\);

  • \(G_{34}=S^{(a1f,a2f,b1f,b2f)}_{12}\);

  • \(S_{34}=SupC(L_{m}(G_{34}\|Spec_{tasks_{3,4}}),L_{m}(G_{34}))\).

The conjunction of these modular supervisors, SavoidS12S34, is guaranteed by construction to meet all the safety specifications and to be nonblocking by Hill and Tilbury (2008). Therefore, for this example, we can consider that the overall supervisor S is composed of these modular supervisors. Now, given S = SavoidS12S34 and based on Theorem 2, we have that the abstraction of the supervisor S can be performed compositionally (based on the transitivity of the equivalence). Thus, the high-level of the hierarchy can be built compositionally, resulting in \(S^h \approx _{PO} S^{(a\#e,b\#e,a\#f,b\#f)}_{avoid}\|S_{12}^{(a\#f,b\#f)}\|S_{34}^{(a\#f,b\#f)}\).

For this example scenario, the size of necessary steps to create the abstracted upper level of the hierarchy are presented in Table 1 for the monolithic approach and in Table 2 for the compositional approach. The monolithic approach requires the creation of an automaton with 672 states and 2512 transitions, while in the compositional approach the largest automaton that is generated has 48 states and 136 transitions. This is a reduction of 93% in the number of states and 95% in the number of transitions. Both approaches result in an upper-level abstraction with 16 states and 32 transitions. Since all the events abstracted (entry and finish events) have cost zero and the abstractions have priced-observation equivalence, it is possible to solve the high-level control of the multiple robots for this scenario in the upper-level abstraction as proposed in Theorem 1.

Table 1 Size of automata generated in the monolithic approach for a scenario of 2 robots, 4 tasks, and 4 regions
Table 2 Size of automata generated in the compositional approach for a scenario of 2 robots, 4 tasks, and 4 regions

Both approaches, monolithic and compositional, were applied to different scenarios of the cooperative robot control application proposed in Hill and Lafortune (2017), where the notation #R#T#R indicates the number of robots, number of tasks, and number of regions. These results are displayed in Table 3, which shows the size of the biggest automaton created during the monolithic and compositional approaches, as well the final abstracted supervisor size.

Table 3 Comparison between monolithic and compositional approaches for abstraction

5 Conclusion

The work of Hill and Lafortune (2016) developed a hierarchical approach to planning in order to reduce the complexity of planning operations, specifically, to accelerate the on-line planning that may need to occur in reaction to unpredictable events or new information. However, the step of first generating the global supervisor before the necessary abstraction can be applied restricts the size of systems that can be addressed by this approach. The present paper is an extension of the conference paper (Vilela and Hill 2020), which demonstrates that under certain conditions it is possible to generate the abstraction of a system compositionally using the new notion of priced-observation equivalence. This equivalence guarantees that an optimal plan found in the abstraction is also optimal in the global supervisor as proven by Theorem 1 of Hill and Lafortune (2016).

In an effort to improve the amount of reduction obtained by the abstraction, (Hill and Lafortune 2016) proposed an operation that creates transitions with artificial zero cost, if they are part of a sequentially-dependent ordering of events, without changing the overall cost of a marked path. However, there are cases where there is a sequential dependence relation between a pair of events, but they do not follow a specific ordering as required by Hill and Lafortune (2016). In this paper, a relaxation of the sequential dependence property was proposed to include these cases. It was shown that the overall cost of marked paths in the resulting lumped automaton based on the relaxed sequential dependence is kept unaltered, maintaining the optimal relation between upper and lower levels proposed by Hill and Lafortune (2016).

The present paper also demonstrates that the relaxation of the sequential dependence concept does not affect the results of Vilela and Hill (2020). Specifically, it is proven that under this relaxation, it is still possible to generate the abstraction of a system compositionally.

The results compiled in Table 3 for different scenarios of the cooperative robot control example of Hill and Lafortune (2017) demonstrates the improvement in scalability achieved by the compositional approach proposed by this paper as compared to the monolithic approach to hierarchy generation employed by Hill and Lafortune (2016). For a scenario with two robots, four tasks, and four regions, employing the compositional approach provides a reduction of 93% in the number of states and 95% in the number of transitions of the largest automaton built in the process of generating the abstraction as compared to using the monolithic approach. For large scenarios, like the one consisting of three robots, ten tasks, and nine regions, the synthesis of a global supervisor is prohibitively large, precluding the possibility of even applying the monolithic approach of Hill and Lafortune (2016). The proposed compositional approach, however, is able to address the scenario and the largest automaton that it needs to construct has 23,625 states and 124,650 transitions.