Advertisement

Detecting Motion in the Environment with a Moving Quadruped Robot

  • Peggy Fidelman
  • Thayne Coffman
  • Risto Miikkulainen
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 4434)

Abstract

For a robot in a dynamic environment, the ability to detect motion is crucial. Motion often indicates areas of the robot’s surroundings that are changing, contain another agent, or are otherwise worthy of attention. Although legs are arguably the most versatile means of locomotion for a robot, and thus the best suited to an unknown or changing domain, existing methods for motion detection either require that the robot have wheels or that its walking be extremely slow and tightly constrained. This paper presents a method for detecting motion from a quadruped robot walking at its top speed. The method is based on a neural network that learns to predict optic flow caused by its walk, thus allowing environment motion to be detected as anomalies in the flow. The system is demonstrated to be capable of detecting motion in the robot’s surroundings, forming a foundation for intelligently directed behavior in complex, changing environments.

Keywords

robot vision image processing 

References

  1. 1.
    Sony: Aibo robot (2004), http://www.sony.net/Products/aibo
  2. 2.
    Rybski, P.E., Stoeter, S.A., Erickson, M.D., Gini, M., Hougen, D.F., Papanikolopoulos, N.: A team of robotic agents for surveillance. In: Proceedings of the Fourth International Conference on Autonomous Agents (2000)Google Scholar
  3. 3.
    Lewis, M.A.: Detecting surface features during locomotion using optic flow. In: Proceedings of the IEEE International Conference on Robotics and Automation, Washington, DC, IEEE Computer Society Press, Los Alamitos (2002)Google Scholar
  4. 4.
    Horn, B.K.P., Schunck, B.: Determining optical flow. Artificial Intelligence 17, 185–203 (1983)CrossRefGoogle Scholar
  5. 5.
    Bhanu, B., Das, S., Roberts, B., Duncan, D.: A system for obstacle detection during rotorcraft low altitude flight. IEEE Transactions on Aerospace and Electronic Systems 32, 875–897 (1996)CrossRefGoogle Scholar
  6. 6.
    Young, G.S., Hong, T.H., Herman, M., Yang, J.C.: Safe navigation for autonomous vehicles: a purposive and direct solution. In: Proceedings of SPIE: Intelligent Robots and Computer Vision XII: Active Vision and 3D Methods, vol. 2056, pp. 31–42 (1993)Google Scholar
  7. 7.
    Harris, C.L., Elman, J.L.: Representing variable information with simple recurrent networks. In: Proceedings of the 11th Annual Conference of the Cognitive Science Society, pp. 635–642 (1989)Google Scholar
  8. 8.
    RoboCup Technical Committee: Sony four legged robot football league rule book (2004), http://www.tzi.de/~roefer/Rules2004/Rules2004.pdf
  9. 9.
    Stone, P., Dresner, K., Fidelman, P., Jong, N.K., Kohl, N., Kuhlmann, G., Sridharan, M., Stronger, D.: The UT Austin Villa 2004 RoboCup four-legged team: Coming of age. Technical Report UT-AI-TR-04-313, The University of Texas at Austin, Department of Computer Sciences, AI Laboratory (2004)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2007

Authors and Affiliations

  • Peggy Fidelman
    • 1
  • Thayne Coffman
    • 2
  • Risto Miikkulainen
    • 1
  1. 1.Department of Computer Sciences, The University of Texas at Austin 
  2. 2.Department of Electrical and Computer Engineering, The University of Texas at Austin 

Personalised recommendations