Skip to main content
Log in

Action dependencies in privacy-preserving multi-agent planning

  • Published:
Autonomous Agents and Multi-Agent Systems Aims and scope Submit manuscript

Abstract

Collaborative privacy-preserving planning (CPPP) is a multi-agent planning task in which agents need to achieve a common set of goals without revealing certain private information. In many CPPP algorithms, the individual agents reason about a projection of the multi-agent problem onto a single-agent classical planning problem. For example, an agent can plan as if it controls the public actions of other agents, ignoring any private preconditions and effects theses actions may have, and use the cost of this plan as a heuristic estimate of the cost of the full, multi-agent plan. Using such a projection, however, ignores some dependencies between agents’ public actions. In particular, it does not contain dependencies between public actions of other agents caused by their private facts. We propose a projection in which these private dependencies are maintained. The benefit of our dependency-preserving projection is demonstrated by using it to produce high-level plans in a new privacy-preserving planner, and as a heuristic for guiding forward search privacy-preserving algorithms. Both are able to solve more benchmark problems than any other state-of-the-art privacy-preserving planner. This more informed projection does not explicitly expose any private fact, action, or precondition. In addition, we show that even if an adversary agent knows that an agent has some private objects of a given type (e.g., trucks), it cannot infer the number of such private objects that the agent controls. This introduces a novel form of strong privacy, which we call object-cardinality privacy, that is motivated by real-world requirements.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Notes

  1. For ease of exposition, we assume that both the preconditions and effects of all actions are consistent, e.g., for a given literal l, no effect or precondition contains both l and \(\lnot l\).

  2. The actual driverLog domain is more complex, and it includes actions for loading and unloading packages, and actions for driving the truck. We omit these actions from this example to make it simpler.

  3. Some MAFS implementations broadcast a state when it is generated instead of when it is expanded.

  4. Note that this obfuscation is only sufficient to preserve the form of privacy defined in Sect. 2.1, which is also known as weak privacy-preserving [4].

  5. In some cases, the agent initiates a distributed process for computing the heuristic value, which includes interactions with and computations by the other agents [34].

  6. Except for the original computation of the projection, which is done once per problem, and not for each generated state.

  7. Note that the pseudo code in Algorithm 3 has a slight abuse of notation in lines 8 and 14, where a set of facts is considered as a conjunction of facts and used with a \(\models \) operator.

  8. By “used”, we mean that a public action is applied with that object as a parameter.

  9. In general, the number of truly private trucks can be more than zero, e.g., when there are trucks that cannot reach any public locations.

  10. http://agents.fel.cvut.cz/codmap/results/.

  11. In the examples that we use, adding a private floor for increasing depth also adds 2 private actions (boarding and leaving the elevator on the private floor). Hence, the branching factor is also slightly increased, and not just the depth.

  12. We incorporated the private information in the local view in a similar way that we did for the \(\textit{DP}^{{ FF}}\) heuristic.

References

  1. Blum, A. L., & Furst, M. L. (1997). Fast planning through planning graph analysis. Artificial Intelligence, 90(1–2), 281–300.

    Article  Google Scholar 

  2. Borrajo, D., & Fernandez, S. (2015). MAPR and CMAP. In ICAPS proceedings of the competition of distributed and multi-agent planners (CoDMAP-15).

  3. Botea, A., Enzenberger, M., Müller, M., & Schaeffer, J. (2005). Macro-ff: improving AI planning with automatically learned macro-operators. Journal of Artificial Intelligence Research, 24, 581–621.

    Article  Google Scholar 

  4. Brafman, R. I. (2015). A privacy preserving algorithm for multi-agent planning and search. In The international joint conference on artificial intelligence (IJCAI) (pp. 1530–1536).

  5. Brafman, R. I., & Domshlak, C. (2018). From one to many: Planning for loosely coupled multi-agent systems. In ICAPS (pp. 28–35).

  6. Brafman, R. I., & Domshlak, C. (2013). On the complexity of planning for agent teams and its implications for single agent planning. Artificial Intelligence, 198, 52–71.

    Article  MathSciNet  Google Scholar 

  7. Brafman, R. I., & Shani, G. (2012). A multi-path compilation approach to contingent planning. In (AAAI).

  8. Chrpa, L. (2010). Generation of macro-operators via investigation of action dependencies in plans. The Knowledge Engineering Review, 25(3), 281–297.

    Article  Google Scholar 

  9. Gerevini, A. E., Saetti, A., & Vallati, M. (2015). Exploiting macro-actions and predicting plan length in planning as satisfiability. AI Communications, 28(2), 323–344.

    MathSciNet  MATH  Google Scholar 

  10. Haslum, P., Bonet, B., & Geffner, H., et al. (2005). New admissible heuristics for domain-independent planning. In AAAI (Vol. 5, pp. 9–13).

  11. Helmert, M. (2006). The fast downward planning system. Journal of Artificial Intelligence Research (JAIR), 26, 191–246.

    Article  Google Scholar 

  12. Helmert, M., & Domshlak, C. (2009). Landmarks, critical paths and abstractions: What’s the difference anyway? In ICAPS.

  13. Hoffmann, J. (2001). FF: The fast-forward planning system. AI Magazine, 22(3), 57.

    Google Scholar 

  14. Jakubuv, J., Tozicka, J., & Komenda, A. (2015). Multiagent planning by plan set intersection and plan verification. In Proceedings ICAART (Vol. 15).

  15. Keyder, E., & Geffner, H. (2009). Soft goals can be compiled away. Journal of Artificial Intelligence Research, 36, 547–556.

    Article  Google Scholar 

  16. Korf, R. E. (1985). Macro-operators: A weak method for learning. Artificial Intelligence, 26(1), 35–77.

    Article  MathSciNet  Google Scholar 

  17. Kovacs, D. L. (2012). A multi-agent extension of PDDL3.1. In Workshop on the international planning competition (IPC) in the international conference on automated planning and scheduling (ICAPS) (pp. 19–27).

  18. Luis, N., & Borrajo, D. (2014). Plan merging by reuse for multi-agent planning. In ICAPS workshop on distributed and multi-agent planning (DMAP).

  19. Luis, N., & Borrajo, D. (2015). PMR: Plan merging by reuse. In ICAPS proceedings of the competition of distributed and multi-agent planners (CoDMAP-15).

  20. Maliah, S., Shani, G., & Brafman, R. I. (2016) Online macro generation for privacy preserving planning. In ICAPS (pp. 216–220).

  21. Maliah, S., Shani, G., & Stern, R. (2014). Privacy preserving landmark detection. In The European conference on artificial intelligence (ECAI) (pp. 597–602).

  22. Maliah, S., Shani, G., & Stern, R. (2016). Collaborative privacy preserving multi-agent planning. In Autonomous agents and multi-agent systems (pp. 1–38).

  23. Maliah, S., Shani, G., & Stern, R. (2016). Privacy preserving lama. In: ICAPS workshop on distributed and multi-agent planning (DMAP).

  24. Maliah, S., Shani, G., & Stern, R. (2016). Stronger privacy preserving projections for multi-agent planning. In The international conference on automated planning and scheduling (ICAPS) (pp. 221–229).

  25. McAllester, D. A., & Rosenblitt, D. (1991). Systematic nonlinear planning. In AAAI (pp. 634–639).

  26. McDermott, D., Ghallab, M., Howe, A., Knoblock, C., Ram, A., Veloso, M., Weld, D., & Wilkins, D. (1998). PDDL—the planning domain definition language. Tech. rep.

  27. Minton, S. (1990). Quantitative results concerning the utility of explanation-based learning. Artificial Intelligence, 42(2–3), 363–391.

    Article  Google Scholar 

  28. Newton, M. A. H., Levine, J., Fox, M., & Long, D. (2007). Learning macro-actions for arbitrary planners and domains. In ICAPS (pp. 256–263).

  29. Nissim, R., & Brafman, R. I. (2014). Distributed heuristic forward search for multi-agent planning. Journal of Artificial Intelligence Research (JAIR), 51, 293–332.

    Article  MathSciNet  Google Scholar 

  30. Palacios, H., & Geffner, H. (2009). Compiling uncertainty away in conformant planning problems with bounded width. Journal of Artificial Intelligence Research, 35, 623–675.

    Article  MathSciNet  Google Scholar 

  31. Richter, S., & Westphal, M. (2010). The LAMA planner: Guiding cost-based anytime planning with landmarks. Journal of Artificial Intelligence Research (JAIR), 39(1), 127–177.

    Article  Google Scholar 

  32. Rintanen, J. (2008). Regression for classical and nondeterministic planning. In European conference on artificial intelligence (ECAI) (pp. 568–572).

  33. Štolba, M., Fišer, D., & Komenda, A. (2015). Admissible landmark heuristic for multi-agent planning. In International conference on automated planning and scheduling (ICAPS).

  34. Štolba, M., & Komenda, A. (2014) Relaxation heuristics for multiagent planning. In International conference on automated planning and scheduling (ICAPS).

  35. Štolba, M., & Komenda, A. (2017). The madla planner: Multi-agent planning by combination of distributed and local heuristic search. Artificial Intelligence, 252, 175–210.

    Article  MathSciNet  Google Scholar 

  36. Štolba, M., Komenda, A., & Kovacs, D. L. (2015). Competition of distributed and multiagent planners (codmap). In The international planning competition (WIPC-15) (p. 24).

  37. Torreño, A., Onaindia, E., Komenda, A., & Štolba, M. (2017). Cooperative multi-agent planning: A survey. ACM Computing Surveys (CSUR), 50(6), 84.

    Article  Google Scholar 

  38. Tozicka, J., Jakubuv, J., & Komenda, A. (2014). Generating multi-agent plans by distributed intersection of finite state machines. In ECAI.

  39. Tozicka, J., Jakubuv, J., & Komenda, A. (2015) On internally dependent public actions in multiagent planning. In Distributed and multi-agent planning (DMAP-15) (p. 18).

  40. Tozicka, J., Jakubuv, J., & Komenda, A. (2018). Recursive reductions of action dependencies for coordination-based multiagent planning. Transactions on Computational Collective Intelligence, XXVIII, 66–92.

    Google Scholar 

  41. Tozicka, J., Štolba, M., & Komenda, A. (2017). The limits of strong privacy preserving multi-agent planning. In International conference on automated planning and scheduling (ICAPS).

Download references

Acknowledgements

This work was partially supported by ISF Grant 933/13, ISF Grant 210/17, and by the Helmsley Charitable Trust through the Agricultural, Biological and Cognitive Robotics Center of Ben-Gurion University of the Negev.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Roni Stern.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Parts of this paper appeared in [24].

Appendix: Deferred heuristic evaluation and globally preferred operators

Appendix: Deferred heuristic evaluation and globally preferred operators

In our implementation of MAFS, we used two heuristic techniques that are known to be helpful in single-agent planning. The first technique is deferred heuristic evaluation [31], which means that the heuristic value of a state is computed only when a state is expanded (i.e., when it is extracted from the open list), as oppose to when it is generated (i.e., when inserted into the open list). Newly generated states are inserted into the open list with the heuristic value of their parent—the state that is currently expanded. This is extremely useful in the case of heuristics that require relatively costly computations, which is often the case in single and multi-agent domain independant planning. The benefit of deferred heuristic evaluation specifically for privacy-preserving MA-STRIPS has been established in prior work [23].

Table 8 The percentage of messages sent and states expanded during planning when using the global preferred operators, compared to using the regular preferred operators mechanism
Table 9 Coverage over the CoDMAP domains when using the global preferred operators

The second helpful technique we used is preferred operators [31], which means that we prioritize some actions—referred to as the preferred operators—over others. Which actions we prioritize depend on the specific heuristic being used. For the full DP heuristic, these are the actions that appear in the solution to the DP projection that is computed by the heuristic. For the \(DP^{{ FF}}\) heuristic and the Joint FF heuristic, the preferred actions are those that achieve preconditions of actions in the relaxed plan. Following [31], we use two queues, one for states that were generated following a preferred operator, and one for states that were generated using other actions. Priority is given to the preferred operators, expanding more often states from that list.

We implemented a global form of preferred operators, in which an agent expands states from their non-preferred operators list only if all agents report that they do not have states in their preferred operator queue. We implement this by adding to the broadcast messages a flag, notifying other agents whether the sending agent has states in its preferred operator queue. Of course, due to synchronization issues, it is possible that one agent is expanding non-preferred states while another agent has already inserted new states into its preferred operators list. Empirically, however, we observed that this global approach for using preferred operators significantly reduces the number of expanded states, and hence, the number of broadcasted messages and overall coverage.

Table 8 shows the percentage of messages sent and states expanded during planning when using the global preferred operators, compared to using the regular preferred operators mechanism. We can see here that MAFS with the full DP heuristic benefits the most from the global preferred operators. By contrast, Joint FF and \(DP^{{ FF}}\) gain less, and in some domains even perform better without the global preferred operators list. This occurs because when using global preferred operators, some agents may be idle waiting for others to expand states from their preferred operators queue. While this is usually worthwhile for the more informed full DP heuristic, it is sometimes not worthwhile for the weaker heuristics.

The use of the global preferred operators in the full DP heuristic also resulted in a much better coverage results, as can be seen in Table 9. Using the global preferred operators, MAFS with the full DP heuristic has the best coverage over all the heuristics that we tried, although still slightly less than the DPP planner. The reason why using global preferred operators has a stronger impact on the full DP heuristic is mainly because planning over the DP projection is relatively costly to planning over the delete relaxation problem, as done by the FF heuristic. Hence, reducing the number of heuristic computations, through reducing the number of state expansions, is most influential for the projection planning approach. Moreover, without the global preferred operators, full DP actually has a lower coverage compared to the other heuristics.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Maliah, S., Shani, G. & Stern, R. Action dependencies in privacy-preserving multi-agent planning. Auton Agent Multi-Agent Syst 32, 779–821 (2018). https://doi.org/10.1007/s10458-018-9394-z

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10458-018-9394-z

Keywords

Navigation