Skip to main content
Log in

Repeatability of Real World Training Experiments: A Case Study

  • Published:
Autonomous Robots Aims and scope Submit manuscript

Abstract

We present a case study of reinforcement learning on a real robot that learns how to back up a trailer and discuss the lessons learned about the importance of proper experimental procedure and design. We identify areas of particular concern to the experimental robotics community at large. In particular, we address concerns pertinent to robotics simulation research, implementing learning algorithms on real robotic hardware, and the difficulties involved with transferring research between the two.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Asada, M., Noda, S., Tawaratsumida, S., and Hosada, K. 1996. Purposive behavior acquisition for a real robot by vision-based reinforcement learning. Machine Learning, 23(2–3):279–303.

    Google Scholar 

  • Barto, A.G., Sutton, R.S., and Anderson, C.W. 1983. Neuronlike adaptive elements that can solve difficult learning control problems. IEEE Transactions on Systems; Man; and Cybernetics, SMC-13:834–846.

    Google Scholar 

  • Benbrahim, H., Doleac, J., Franklin, J., and Selfridge, O. 1992. Real-time learning: A ball on a beam. In Proceedings of the International Joint Conference on Neural Networks, Vol. 1, pp. 98–103.

    Google Scholar 

  • Cohen, P.R. 1995. Emperical Methods for Artificial Intelligence, MIT Press: Cambridge, MA.

    Google Scholar 

  • Colombetti, M., Dorigo, M., and Borghi, G. 1996. Behavior analysis and training: A methodology for behavior engineering. IEEE Transactions on Systems; Man; and Cybernetics, B26(3):365–380.

    Google Scholar 

  • Hougen, D.F. 1998. Connectionist reinforcement learning for control of robotic systems. Ph.D. thesis, University of Minnesota, Minneapolis, MN.

    Google Scholar 

  • Hougen, D.F., Fischer, J., Gini, M., and Slagle, J. 1996. Fast connectionist learning for trailer backing using a real robot. In Proceedings of the IEEE International Conference on Robotics and Automation, pp. 1917–1922.

  • Hougen, D.F., Fischer, J., and Johnam, D. 1994. A neural network pole balancer that learns and operates on a real robot in real time. In Proceedings of the MLC-COLT Workshop on Robot Learning, pp. 73–80.

  • Hougen, D.F., Gini, M., and Slagle, J. 1997a. Partitioning input space for reinforcement learning for control. In Proceedings of the IEEE International Conference on Neural Networks, pp. 755–760.

  • Hougen, D.F., Gini, M., and Slagle, J. 1997b. Rapid, unsupervised connectionist learning for backing a robot with two trailers. In Proceedings of the IEEE International Conference on Robotics and Automation, pp. 2950–2955.

  • Jani, Y. 1993. Application of fuzzy logic-neural network based reinforcement learning to proximity and docking operations: Special approach/docking testcase results. Technical Report NASA-CR-192294, Research Institute for Computing and Information Systems, University of Houston, Clear Lake, TX.

    Google Scholar 

  • Jervis, T. and Fallside, F. 1992. Pole balancing on a real rig using a reinforcement learning controller, Technical Report CUED/FINFENG/TR 115, Cambridge University Engineering Department, Cambridge, England.

    Google Scholar 

  • Kohonen, T.K. 1989. Self-Organizing and Associative Memory, 3rd edition, Berlin: Springer-Verlag.

    Google Scholar 

  • Kong, S.-G. and Kosko, B. 1992. Adaptive fuzzy systems for backing up a truck-and-trailer. IEEE Transactions on Neural Networks, 3(2):211–223.

    Google Scholar 

  • Koza, J.R. 1992. Agenetic approach to finding a controller to back up a tractor-trailer truck. In Automatic Control Conference, pp. 2307–2311.

  • Kretchmar, R.M. and Anderson, C.W. 1997. Comparison of CMACs and radial basis functions for local function approximators in reinforcement learning. In Proceedings of the IEEE International Conference on Neural Networks, Houston, TX, pp. 834–837.

  • Lin, C.-S. and Kim, H. 1991. Use of CMAC neural networks in reinforcement self-learning control. In Proceedings of the 1991 International Conference on Artificial Neural Networks, Espoo, Finland, pp. 1285–1288.

  • Mahadevan, S. and Connell, J. 1992. Automatic programming of behavior-based robots using reinforcement learning. Artificial Intelligence, 55:311–365.

    Google Scholar 

  • Martin, F.G. 1995. The mini board technical reference, MIT Media Laboratory, Cambridge, MA.

    Google Scholar 

  • Martin, F.G. 1998. The handy board technical reference, MIT Media Laboratory, Cambridge, MA.

    Google Scholar 

  • Matarić, M.J. 1997. Reinforcement learning in the multi-robot domain. Autonomous Robots, 4(1):73–83.

    Google Scholar 

  • Michie, D. and Chambers, R. 1968. Boxes: An experiment in adaptive control. In E. Dale and D. Michie (Eds.), Machine Intelligence, Oliver and Boyd: Edinburgh.

    Google Scholar 

  • Nguyen, D. and Widrow, B. 1990. Neural networks for self-learning control systems. IEEE Control Systems Magazine, 10(3):18–23.

    Google Scholar 

  • Russell, D.W. and Rees, S.J. 1975. System control—A case study of a statistical learning automaton. In R. Trappl and F. de P. Hanika (Eds.), Progress in Cybernetics and Systems Research, John Wiley and Sons: New York, Vol. 2.

    Google Scholar 

  • Selfridge, O., Sutton, R., and Barto, A. 1985. Training and tracking in robotics. In Proceedings of the International Joint Conference on Artificial Intelligence, Los Angeles, CA, pp. 670–672.

  • Shelton, R.O. and Peterson, J.K. 1992. Controlling a truck with an adaptive critic CMAC design. Simulation, 58(5):319–326.

    Google Scholar 

  • Woodcock, N., Hallam, N.J., and Picton, P.D. 1991. Fuzzy BOXES as an alternative to neural networks for difficult control problems. In Proc. of Applications of Artificial Intelligence in Engineering VI, Oxford, UK, pp. 903–919.

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

About this article

Cite this article

Hougen, D.F., Rybski, P.E. & Gini, M. Repeatability of Real World Training Experiments: A Case Study. Autonomous Robots 6, 281–292 (1999). https://doi.org/10.1023/A:1008984312527

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1023/A:1008984312527

Navigation