Abstract
This paper analyses solutions to the time-optimal planning and execution (TOPE) problem, in which the aim is to minimise the total time required for an agent to achieve its objectives. The TOPE process provides a means of adjusting system parameters in real-time to achieve this aim. Prior work by the authors showed that agent-based planning systems employing the TOPE process can yield better performance than existing techniques, provided that a key estimation step can be run sufficiently fast and accurately. This paper describes several real-time implementations of this estimation step. A Monte-Carlo analysis compares the performance of TOPE systems using these implementations against existing state-of-the-art planning techniques. It is shown that the average case performance of the TOPE systems is significantly better than the existing methods. Since the TOPE process can be added to an existing system without modifying the internal processes, these results suggest that similar performance improvement may be obtained in a multitude of robotics applications.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
The ‘highest’ level of a hierarchical planning system is the level of control furthest from that which directly commands the agent’s actuators. This is also commonly referred to as the ‘global’ level.
The selection of optimal parameters was done by running the comparison methods with every possible combination of parameter values. For dynamic state spaces in general it is impossible to know the best parameter values in advance. Nonetheless, the results showed that if the possible combinations of parameter sets were the same as those available to the TOPE process, then even the optimal static choice was inferior to the TOPE process’ ability to change the values dynamically.
This commitment process was first described in [3] and ensures that updated plans always start from a future position of the agent, meaning that backtracking is guaranteed never to occur.
Although this treatment affects the expected execution time of the planner’s output, in practice these areas are always revealed by the simulated agent’s sensors before they are ever traversed. As a result, the measured execution time is always correct, despite the planners treating unknown regions optimistically. It is also relevant to point out that the combination of this cost value for unknown regions, and the heuristic function used, tends to reduce the computational time required by the planning algorithms. This is because the heuristic cost is closer to the true cost, which limits the number of states explored by a graph search algorithm [31, 35].
It should be noted that given these constraints, it is possible for a particular baseline technique to fail entirely if the start and goal points are blocked by a closed ring of obstacle cells in the baseline technique’s fixed cell size cost map. This point itself is indicative of the benefits of TOPE techniques that allow adjustment of system parameters, over techniques that rely on a fixed parameter set and might therefore fail unnecessarily.
Memoization refers to recording the output of a function given a particular input, such that this output can be rapidly retrieved when the same input is used again, rather than recalculating the result. With reference to plans, this would imply that for a given parameter set, the entire plan could be stored and reused if the same parameters were selected again. Since the agent’s state will differ between TOPE estimation iterations, as will the planning state space, this is not possible within the TOPE process.
The learning-discount parameter is functionally equivalent to the ‘temperature’ used in SA.
References
Allen T (2011) Time-optimal active decision making. PhD thesis, The University of Sydney
Allen T (2011) The time-optimal planning and execution problem. In: IEEE international conference on robotics and automation, Shanghai, China
Allen T, Hill A, Underwood J, Scheding S (2009) Dynamic path planning with multi-agent data fusion: the parallel hierarchical replanner. In: Proceedings of the international conference on robotics and automation, Kobe, Japan, pp 3245–3250
Bekris K, Chen B, Ladd A, Plaku E, Kavraki L (2003) Multiple query probabilistic roadmap planning using single query primitives. In: IEEE/RSJ international conference on intelligent robots and systems
Belghith K, Kabanza F, Hartman L, Nkambou R (2006) Anytime dynamic path-planning with flexible probabilistic roadmaps. In: IEEE international conference on robotics and automation
Bellman R (1957) Dynamic programming. Princeton University Press, New Jersey
Canny J (1988) The complexity of robot motion planning. MIT Press, Cambridge
Cerny V (1985) A thermodynamical approach to the travelling salesman problem: an efficient simulation algorithm. J Optim Theory Appl 45:41–51
Ferguson D, Likhachev M, Stentz A (2005) A guide to heuristic-based path planning. In: Proceedings of the international workshop on planning under uncertainty for autonomous systems. International conference on automated planning and scheduling (ICAPS)
Fournier A, Fussel D, Carpenter L (1982) Computer rendering of stochastic models. Commun ACM 25:371–384
Hammersley J, Handscomb D (1975) Monte Carlo methods. Methuen, London
Hansen E, Zhou R (2007) Anytime heuristic search. J Artif Intell Res 28:267–297
Hansen E, Zilberstein S (2001) Montioring and control of anytime algorithms: a dynamic programming approach. Artif Intell 126:139–157
Hart P, Nilsson N, Raphael B (1968) A formal basis for the heuristic determination of minimum cost paths. IEEE Trans Syst Sci Cybern 2:100–107
Hart P, Nilsson N, Raphael B (1972) Correction to “A formal basis for the heuristic determination of minimum cost paths”. SIGART Newslett 37:28–29
Horvitz E (1987) Reasoning about beliefs and actions under computational resource constraints. In: Proceedings of the third workshop on uncertainty in artifical intelligence, Seattle, WA, pp 429–444
Howard T, Green C, Kelly A, Ferguson D (2008) State space sampling of feasible motions for high-performance mobile robot navigation in complex environments. J Field Robot 25: 325–345
Karumanchi S (2010) Off-road mobility analysis from proprioceptive feedback. PhD thesis, Department of Aerospace, Mechanical, and Mechatronic Engineering, The University of Sydney
Karumanchi S, Allen T, Bailey T, Scheding S (2010) Non-parametric learning to aid path planning over slopes. Int J Robot Res 29(8):997–1018
Kavraki LE, Svestka P, Latombe JC, Overmars MH (1996) Probabilistic roadmaps for path planning in high-dimensional configuration spaces. IEEE Trans Robot Autom 12(4):566–580
Kelly A, Stentz A, Amidi O, Bode M, Bradley D, Diaz-Calderon A, Happold M, Herman H, Mandelbaum R, Pilarski T, Rander P, Thayer S, Vallidis N, Warner R (2006) Toward reliable off road autonomous vehicles operating in challenging environments. Int J Robot Res 25(5–6):449–483
Kirkpatrick S, Gelatt C, Vecchi M (1983) Optimization by simulated annealing. Science New Series 220(4598):671–680
Knepper R, Kelly A (2006) High performance state lattice planning using heuristic look-up tables. In: 2006 IEEE/RSJ international conference on intelligent robots and systems, pp 3375–3380
LaValle SM (2006) Planning algorithms. Cambridge University Press, Cambridge. http://planning.cs.uiuc.edu/
Likhachev M, Ferguson D (2009) Planning long dynamically feasible maneuvers for autonomous vehicles. Int J Robot Res 28:933–945
Likhachev M, Ferguson D, Gordon G, Stentz A, Thrun S (2005) Anytime dynamic A*: an anytime, replanning algorithm. In: Proceedings of the international conference on automated planning and scheduling (ICAPS)
Likhachev M, Ferguson D, Gordon G, Stentz A, Thrun S (2005) Anytime dynamic A*: the proofs. Technical report. CMU-RI-TR-05-12, Robotics Institute, Pittsburgh, PA
Likhachev M, Gordon G, Thrun S (2003) ARA*: anytime A* with provable bounds on sub-optimality. In: Neural information processing systems
Matthews G (2008) Asynchronous decision making for decentralised autonomous systems. PhD thesis, Department of Aerospace, Mechanical, and Mechatronic Engineering, The University of Sydney
Miller G (1986) The definition and rendering of terrain maps. Comput Graph 20(4):39–48
Pearl J (1984) Heuristics. Addison-Wesley, New York
Pivtoraiko M, Kelly A (2005) Efficient constrained path planning via search state lattices. In: The 8th international symposium on artifical intelligence, robotics, and automation in space
Pivtoraiko M, Kelly A (2005) Generating near minimal spanning control sets for constrained motion planning in discrete state spaces. In: International conference on intelligent robots and systems
Pohl I (1970) Heuristic search viewed as path finding in a graph. Artif Intell 1:193–204
Russell S, Norvig P (2003) Artifical intelligence: a modern approach, 2nd edn. Prentice Hall, New York
Russell S, Wefald E, Karnaugh M, Karp R, McAllester D, Subramanian D, Wellman M (1991) Principles of metareasoning. In: Artificial intelligence. Morgan Kaufmann, pp 400–411
Saitoh T, Suzuki M, Kuroda Y (2010) Vision-based probabilistic map estimation with an inclined surface grid for rough terrain rover navigation. Adv Robot 24:421–440
Sánchez G, Latombe JC (2003) A single-query bi-directional probabilistic roadmap planner with lazy collision checking. Int J Robot Res 6:403–417
Sutton R, Barto A (1998) Reinforcement learning. MIT Press, Cambridge
Tesauro G (1995) Temporal difference learning and TD-gammon. Commun Assoc Comput Mach 38(3):58–69
Thayer J, Ruml W (2008) Faster than weighted A*: an optimistic approach to bounded suboptimal search. In: Proceedings of the eighteenth international conference on automated planning and scheduling
Underwood J (2009) Reliable and safe autonomy for ground vehicles in unstructured environments. PhD thesis, Department of Aerospace, Mechanical, and Mechatronic Engineering, The University of Sydney
Valve Corporation: Steam Hardware Survey (2010) http://store.steampowered.com/hwsurvey/cpus/
Voss R (1987) Fractals in nature: characterization, measurement, and simulation. In: SIGGRAPH
Watkins C (1989) Learning from delayed rewards. PhD thesis, Cambridge University, Cambridge
Zilberstein S (1996) Using anytime algorithms in intelligent systems. Artif Intell Mag 17:73–83
Zilberstein S, Russell S (1995) Imprecise and approximate computation. Approximate reasoning using anytime algorithms, chap 4. Kluwer, Norwell
Acknowledgments
This work is supported in part by the Australian Research Council (ARC) Centre of Excellence programme, funded by the ARC and the New South Wales State Government. The authors are grateful to the Rio Tinto Centre for Mine Automation for the use of \(\sim \)70,000 hours of CPU time required to perform this analysis.
Author information
Authors and Affiliations
Corresponding author
Appendix A: Tabulated results of the Monte-Carlo analysis
Appendix A: Tabulated results of the Monte-Carlo analysis
This appendix tabulates (Tables 1, 2, 3, 4, 5, 6, 7, 8) the results of the Monte-Carlo analysis of TOPE estimators described in Sect. 6.1. Two tables are shown for each of the four scenarios. The first shows the percentage of runs in which each TOPE estimator outperforms each baseline technique. The second shows the inverse; the percentage of runs in which each baseline technique outperforms each TOPE estimator.
Rights and permissions
About this article
Cite this article
Allen, T., Scheding, S. Analysis of solutions to the time-optimal planning and execution problem. Intel Serv Robotics 5, 245–258 (2012). https://doi.org/10.1007/s11370-012-0116-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11370-012-0116-0