Abstract
POMDPs provide a rich framework for planning and control in partially observable domains. Recent new algorithms have greatly improved the scalability of POMDPs, to the point where they can be used in robot applications. In this paper, we describe how approximate POMDP solving can be further improved by the use of a new theoretically-motivated algorithm for selecting salient information states. We present the algorithm, called PEMA, demonstrate competitive performance on a range of navigation tasks, and show how this approach is robust to mismatches between the robot’s physical environment and the model used for planning.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
D. Braziunas and C. Boutilier. Stochastic local search for POMDP controllers. In Proceedings of the Nineteenth National Conference on Artificial Intelligence (AAAI), pages 690–696, 2004.
A. Cassandra, M. L. Littman, and N. L. Zhang. Incremental pruning: A simple, fast, exact method for partially observable Markov decision processes. In Proceedings of the Thirteenth Conference on Uncertainty in Artificial Intelligence (UAI), pages 54–61, 1997.
M. L. Littman, A. R. Cassandra, and L. P. Kaelbling. Learning policies for partially obsevable environments: Scaling up. In Proceedings of Twelfth International Conference on Machine Learning, pages 362–370, 1995.
M. Montemerlo, N. Roy, and S. Thrun. Perspectives on standardization in mobile robot programming: The Carnegie Mellon navigation (CARMEN) toolkit. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), volume 3, pages pp 2436–2441, 2003.
J. Pineau, G. Gordon, and S. Thrun. Point-based value iteration: An anytime algorithm for POMDPs. In Proceedings of the 18th International Joint Conference on Artificial Intelligence (IJCAI), pages 1025–1032, 2003.
J. Pineau, M. Montermerlo, M. Pollack, N. Roy, and S. Thrun. Towards robotic assistants in nursing homes: challenges and results. Robotics and Autonomous Systems, 42(3–4):271–281, 2003.
K.-M. Poon. A fast heuristic algorithm for decision-theoretic planning. Master’s thesis, The Hong-Kong University of Science and Technology, 2001.
P. Poupart. Exploiting Structure to Efficiently Solve Large Scale Partially Observable Markov Decision Processes. PhD thesis, University of Toronto, 2005.
P. Poupart and C. Boutilier. Bounded finite state controllers. In Advances in Neural Information Processing Systems (NIPS), volume 16, 2004.
T. Smith and R. Simmons. Heuristic search value iteration for POMDPs. In Proceedings of the Twentieth Conference on Uncertainty in Artificial Intelligence (UAI), 2004.
N. Vlassis and M. T. J. Spaan. A fast point-based algorithm for POMDPs. In Proceedings of the Belgian-Dutch Conference on Machine Learning, 2004.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2007 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Pineau, J., Gordon, G.J. (2007). POMDP Planning for Robust Robot Control. In: Thrun, S., Brooks, R., Durrant-Whyte, H. (eds) Robotics Research. Springer Tracts in Advanced Robotics, vol 28. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-48113-3_7
Download citation
DOI: https://doi.org/10.1007/978-3-540-48113-3_7
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-48110-2
Online ISBN: 978-3-540-48113-3
eBook Packages: EngineeringEngineering (R0)