Skip to main content
Log in

Analysis of solutions to the time-optimal planning and execution problem

  • Special Issue
  • Published:
Intelligent Service Robotics Aims and scope Submit manuscript

Abstract

This paper analyses solutions to the time-optimal planning and execution (TOPE) problem, in which the aim is to minimise the total time required for an agent to achieve its objectives. The TOPE process provides a means of adjusting system parameters in real-time to achieve this aim. Prior work by the authors showed that agent-based planning systems employing the TOPE process can yield better performance than existing techniques, provided that a key estimation step can be run sufficiently fast and accurately. This paper describes several real-time implementations of this estimation step. A Monte-Carlo analysis compares the performance of TOPE systems using these implementations against existing state-of-the-art planning techniques. It is shown that the average case performance of the TOPE systems is significantly better than the existing methods. Since the TOPE process can be added to an existing system without modifying the internal processes, these results suggest that similar performance improvement may be obtained in a multitude of robotics applications.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Notes

  1. A full derivation may be found in [1] and [2].

  2. The ‘highest’ level of a hierarchical planning system is the level of control furthest from that which directly commands the agent’s actuators. This is also commonly referred to as the ‘global’ level.

  3. The selection of optimal parameters was done by running the comparison methods with every possible combination of parameter values. For dynamic state spaces in general it is impossible to know the best parameter values in advance. Nonetheless, the results showed that if the possible combinations of parameter sets were the same as those available to the TOPE process, then even the optimal static choice was inferior to the TOPE process’ ability to change the values dynamically.

  4. This commitment process was first described in [3] and ensures that updated plans always start from a future position of the agent, meaning that backtracking is guaranteed never to occur.

  5. Although this treatment affects the expected execution time of the planner’s output, in practice these areas are always revealed by the simulated agent’s sensors before they are ever traversed. As a result, the measured execution time is always correct, despite the planners treating unknown regions optimistically. It is also relevant to point out that the combination of this cost value for unknown regions, and the heuristic function used, tends to reduce the computational time required by the planning algorithms. This is because the heuristic cost is closer to the true cost, which limits the number of states explored by a graph search algorithm [31, 35].

  6. It should be noted that given these constraints, it is possible for a particular baseline technique to fail entirely if the start and goal points are blocked by a closed ring of obstacle cells in the baseline technique’s fixed cell size cost map. This point itself is indicative of the benefits of TOPE techniques that allow adjustment of system parameters, over techniques that rely on a fixed parameter set and might therefore fail unnecessarily.

  7. Memoization refers to recording the output of a function given a particular input, such that this output can be rapidly retrieved when the same input is used again, rather than recalculating the result. With reference to plans, this would imply that for a given parameter set, the entire plan could be stored and reused if the same parameters were selected again. Since the agent’s state will differ between TOPE estimation iterations, as will the planning state space, this is not possible within the TOPE process.

  8. The learning-discount parameter is functionally equivalent to the ‘temperature’ used in SA.

References

  1. Allen T (2011) Time-optimal active decision making. PhD thesis, The University of Sydney

  2. Allen T (2011) The time-optimal planning and execution problem. In: IEEE international conference on robotics and automation, Shanghai, China

  3. Allen T, Hill A, Underwood J, Scheding S (2009) Dynamic path planning with multi-agent data fusion: the parallel hierarchical replanner. In: Proceedings of the international conference on robotics and automation, Kobe, Japan, pp 3245–3250

  4. Bekris K, Chen B, Ladd A, Plaku E, Kavraki L (2003) Multiple query probabilistic roadmap planning using single query primitives. In: IEEE/RSJ international conference on intelligent robots and systems

  5. Belghith K, Kabanza F, Hartman L, Nkambou R (2006) Anytime dynamic path-planning with flexible probabilistic roadmaps. In: IEEE international conference on robotics and automation

  6. Bellman R (1957) Dynamic programming. Princeton University Press, New Jersey

    MATH  Google Scholar 

  7. Canny J (1988) The complexity of robot motion planning. MIT Press, Cambridge

    Google Scholar 

  8. Cerny V (1985) A thermodynamical approach to the travelling salesman problem: an efficient simulation algorithm. J Optim Theory Appl 45:41–51

    Article  MathSciNet  MATH  Google Scholar 

  9. Ferguson D, Likhachev M, Stentz A (2005) A guide to heuristic-based path planning. In: Proceedings of the international workshop on planning under uncertainty for autonomous systems. International conference on automated planning and scheduling (ICAPS)

  10. Fournier A, Fussel D, Carpenter L (1982) Computer rendering of stochastic models. Commun ACM 25:371–384

    Article  Google Scholar 

  11. Hammersley J, Handscomb D (1975) Monte Carlo methods. Methuen, London

    Google Scholar 

  12. Hansen E, Zhou R (2007) Anytime heuristic search. J Artif Intell Res 28:267–297

    MathSciNet  MATH  Google Scholar 

  13. Hansen E, Zilberstein S (2001) Montioring and control of anytime algorithms: a dynamic programming approach. Artif Intell 126:139–157

    Article  MathSciNet  MATH  Google Scholar 

  14. Hart P, Nilsson N, Raphael B (1968) A formal basis for the heuristic determination of minimum cost paths. IEEE Trans Syst Sci Cybern 2:100–107

    Article  Google Scholar 

  15. Hart P, Nilsson N, Raphael B (1972) Correction to “A formal basis for the heuristic determination of minimum cost paths”. SIGART Newslett 37:28–29

    Article  Google Scholar 

  16. Horvitz E (1987) Reasoning about beliefs and actions under computational resource constraints. In: Proceedings of the third workshop on uncertainty in artifical intelligence, Seattle, WA, pp 429–444

  17. Howard T, Green C, Kelly A, Ferguson D (2008) State space sampling of feasible motions for high-performance mobile robot navigation in complex environments. J Field Robot 25: 325–345

    Article  Google Scholar 

  18. Karumanchi S (2010) Off-road mobility analysis from proprioceptive feedback. PhD thesis, Department of Aerospace, Mechanical, and Mechatronic Engineering, The University of Sydney

  19. Karumanchi S, Allen T, Bailey T, Scheding S (2010) Non-parametric learning to aid path planning over slopes. Int J Robot Res 29(8):997–1018

    Article  Google Scholar 

  20. Kavraki LE, Svestka P, Latombe JC, Overmars MH (1996) Probabilistic roadmaps for path planning in high-dimensional configuration spaces. IEEE Trans Robot Autom 12(4):566–580

    Article  Google Scholar 

  21. Kelly A, Stentz A, Amidi O, Bode M, Bradley D, Diaz-Calderon A, Happold M, Herman H, Mandelbaum R, Pilarski T, Rander P, Thayer S, Vallidis N, Warner R (2006) Toward reliable off road autonomous vehicles operating in challenging environments. Int J Robot Res 25(5–6):449–483

    Google Scholar 

  22. Kirkpatrick S, Gelatt C, Vecchi M (1983) Optimization by simulated annealing. Science New Series 220(4598):671–680

    Article  MathSciNet  Google Scholar 

  23. Knepper R, Kelly A (2006) High performance state lattice planning using heuristic look-up tables. In: 2006 IEEE/RSJ international conference on intelligent robots and systems, pp 3375–3380

  24. LaValle SM (2006) Planning algorithms. Cambridge University Press, Cambridge. http://planning.cs.uiuc.edu/

  25. Likhachev M, Ferguson D (2009) Planning long dynamically feasible maneuvers for autonomous vehicles. Int J Robot Res 28:933–945

    Article  Google Scholar 

  26. Likhachev M, Ferguson D, Gordon G, Stentz A, Thrun S (2005) Anytime dynamic A*: an anytime, replanning algorithm. In: Proceedings of the international conference on automated planning and scheduling (ICAPS)

  27. Likhachev M, Ferguson D, Gordon G, Stentz A, Thrun S (2005) Anytime dynamic A*: the proofs. Technical report. CMU-RI-TR-05-12, Robotics Institute, Pittsburgh, PA

  28. Likhachev M, Gordon G, Thrun S (2003) ARA*: anytime A* with provable bounds on sub-optimality. In: Neural information processing systems

  29. Matthews G (2008) Asynchronous decision making for decentralised autonomous systems. PhD thesis, Department of Aerospace, Mechanical, and Mechatronic Engineering, The University of Sydney

  30. Miller G (1986) The definition and rendering of terrain maps. Comput Graph 20(4):39–48

    Article  Google Scholar 

  31. Pearl J (1984) Heuristics. Addison-Wesley, New York

    Google Scholar 

  32. Pivtoraiko M, Kelly A (2005) Efficient constrained path planning via search state lattices. In: The 8th international symposium on artifical intelligence, robotics, and automation in space

  33. Pivtoraiko M, Kelly A (2005) Generating near minimal spanning control sets for constrained motion planning in discrete state spaces. In: International conference on intelligent robots and systems

  34. Pohl I (1970) Heuristic search viewed as path finding in a graph. Artif Intell 1:193–204

    Article  MathSciNet  MATH  Google Scholar 

  35. Russell S, Norvig P (2003) Artifical intelligence: a modern approach, 2nd edn. Prentice Hall, New York

    Google Scholar 

  36. Russell S, Wefald E, Karnaugh M, Karp R, McAllester D, Subramanian D, Wellman M (1991) Principles of metareasoning. In: Artificial intelligence. Morgan Kaufmann, pp 400–411

  37. Saitoh T, Suzuki M, Kuroda Y (2010) Vision-based probabilistic map estimation with an inclined surface grid for rough terrain rover navigation. Adv Robot 24:421–440

    Article  Google Scholar 

  38. Sánchez G, Latombe JC (2003) A single-query bi-directional probabilistic roadmap planner with lazy collision checking. Int J Robot Res 6:403–417

    Article  Google Scholar 

  39. Sutton R, Barto A (1998) Reinforcement learning. MIT Press, Cambridge

    Google Scholar 

  40. Tesauro G (1995) Temporal difference learning and TD-gammon. Commun Assoc Comput Mach 38(3):58–69

    Google Scholar 

  41. Thayer J, Ruml W (2008) Faster than weighted A*: an optimistic approach to bounded suboptimal search. In: Proceedings of the eighteenth international conference on automated planning and scheduling

  42. Underwood J (2009) Reliable and safe autonomy for ground vehicles in unstructured environments. PhD thesis, Department of Aerospace, Mechanical, and Mechatronic Engineering, The University of Sydney

  43. Valve Corporation: Steam Hardware Survey (2010) http://store.steampowered.com/hwsurvey/cpus/

  44. Voss R (1987) Fractals in nature: characterization, measurement, and simulation. In: SIGGRAPH

  45. Watkins C (1989) Learning from delayed rewards. PhD thesis, Cambridge University, Cambridge

  46. Zilberstein S (1996) Using anytime algorithms in intelligent systems. Artif Intell Mag 17:73–83

    Google Scholar 

  47. Zilberstein S, Russell S (1995) Imprecise and approximate computation. Approximate reasoning using anytime algorithms, chap 4. Kluwer, Norwell

    Google Scholar 

Download references

Acknowledgments

This work is supported in part by the Australian Research Council (ARC) Centre of Excellence programme, funded by the ARC and the New South Wales State Government. The authors are grateful to the Rio Tinto Centre for Mine Automation for the use of \(\sim \)70,000 hours of CPU time required to perform this analysis.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Thomas Allen.

Appendix A: Tabulated results of the Monte-Carlo analysis

Appendix A: Tabulated results of the Monte-Carlo analysis

This appendix tabulates (Tables 1, 2, 3, 4, 5, 6, 7, 8) the results of the Monte-Carlo analysis of TOPE estimators described in Sect. 6.1. Two tables are shown for each of the four scenarios. The first shows the percentage of runs in which each TOPE estimator outperforms each baseline technique. The second shows the inverse; the percentage of runs in which each baseline technique outperforms each TOPE estimator.

Table 1 Percentage of runs in which each comparison method beat each baseline for a Monte-Carlo simulation over real-world cost map data
Table 2 Percentage of runs in which each baseline beat each comparison method for a Monte-Carlo simulation over real-world cost map data
Table 3 Percentage of runs in which each comparison method beat each baseline for randomised fractal terrain data with a roughness parameter of \(H=0.2\)
Table 4 Percentage of runs in which each baseline beat each comparison method for randomised fractal terrain data with a roughness parameter of \(H=0.2\)
Table 5 Percentage of runs in which each comparison method beat each baseline for randomised fractal terrain data with a roughness parameter of \(H=0.5\)
Table 6 Percentage of runs in which each baseline beat each comparison method for randomised fractal terrain data with a roughness parameter of \(H=0.5\)
Table 7 Percentage of runs in which each comparison method beat each baseline for randomised fractal terrain data with a roughness parameter of \(H=0.8\)
Table 8 Percentage of runs in which each baseline beat each comparison method for randomised fractal terrain data with a roughness parameter of \(H=0.8\)

 

Rights and permissions

Reprints and permissions

About this article

Cite this article

Allen, T., Scheding, S. Analysis of solutions to the time-optimal planning and execution problem. Intel Serv Robotics 5, 245–258 (2012). https://doi.org/10.1007/s11370-012-0116-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11370-012-0116-0

Keywords

Navigation