1.

S. Schaal, “Is imitation learning the route to humanoid robots?” *Trends in Cognitive Sciences*, vol. 3, pp. 233–242, 1999.

2.

C.M. Bishop, *Neural Networks for Pattern Recognition*, Oxford University Press: New York, 1995.

3.

C.K.I. Williams and C.E. Rasmussen, “Gaussian processes for regression,” in *Advances in Neural Information Processing Systems*, vol. 8, edited by D.S. Touretzky, M.C. Mozer, and M.E. Hasselmo, MIT Press: Cambridge, MA, pp. 514–520, 1996.

4.

V.N. Vapnik, *Estimation of Dependences Based on Empirical Data*, Springer: Berlin, 1982.

5.

C. Cortes and V. Vapnik, “Support vector networks,” *Machine Learning*, vol. 20, pp. 273–297, 1995.

6.

V. Vapnik, S. Golowich, and A. Smola, “Support vector method for function approximation, regression estimation, and signal processing,” in *Advances in Neural Information Processing Systems*, vol. 9, edited by M. Mozer, M.I. Jordan, and T. Petsche, MIT Press: Cambridge, MA, pp. 281–287, 1996.

7.

S. Schaal and C.G. Atkeson, “Constructive incremental learning from only local information,” *Neural Comput*, vol. 10, pp. 2047–2084, 1998.

8.

C.G. Atkeson and S. Schaal, “Memory-based neural networks for robot learning,” *Neurocomputing*, vol. 9, pp. 1–27, 1995.

9.

W.S. Cleveland and C. Loader, *Smoothing by Local Regression: Principles and Methods*, AT&T Bell Laboratories: Murray Hill, NY, 1995.

10.

T.J. Hastie and R.J. Tibshirani, “Nonparametric regression and classification: Part I: Nonparametric regression,” in *From Statistics to Neural Networks: Theory and Pattern Recognition Applications*, edited by V. Cherkassky, J.H. Friedman, and H. Wechsler, ASI Proceedings, Subseries F, Computer and Systems Sciences, Springer: Berlin, pp. 120–143, 1994.

11.

C.G. Atkeson, A.W. Moore, and S. Schaal, “Locally weighted learning,” *Artificial Intelligence Review*, vol. 11, pp. 11–73, 1997.

12.

C.G. Atkeson, A.W. Moore, and S. Schaal, “Locally weighted learning for control,” *Artificial Intelligence Review*, vol. 11, pp. 75–113, 1997.

13.

D. Aha, “Lazy learning,” *Artificial Intelligence Review*, vol. 11, no. 1–5, pp. 325–337, 1997.

14.

L. Ljung and T. Söderström, *Theory and Practice of Recursive Identification*, MIT Press: Cambridge, MA, 1986.

15.

W.S. Cleveland, “Robust locally weighted regression and smoothing scatterplots,” *Journal of the American Statistical Association*, vol. 74, pp. 829–836, 1979.

16.

T. Hastie and C. Loader, “Local regression: Automatic kernel carpentry,” *Statistical Science*, vol. 8, pp. 120–143, 1993.

17.

C.G. Atkeson, “Using local models to control movement,” in *Advances in Neural Information Processing Systems*, vol. 1, edited by D. Touretzky, Morgan Kaufmann: San Mateo, CA, pp. 157–183, 1989.

18.

C.G. Atkeson, “Memory-based approaches to approximating continuous functions,” in *Nonlinear Modeling and Forecasting*, edited by M. Casdagli and S. Eubank, Addison Wesley: Redwood City, CA, pp. 503–521, 1992.

19.

A.W. Moore, “Efficient memory-based learning for robot control,” Computer Laboratory, University of Cambridge, October 1990.

20.

D.G. Lowe, “Similarity metric learning for a variable-kernel classifier,” *Neural Comput*, vol. 7, pp. 72–85, 1995.

21.

H. Wold, “Soft modeling by latent variables: The nonlinear iterative partial least squares approach,” in *Perspectives in Probability and Statistics*, Papers in Honour of M.S. Bartlett, edited by J. Gani, Academic Press: London, pp. 520–540, 1975.

22.

I.E. Frank and J.H. Friedman, “Astatistical viewof some chemometric regression tools,” *Technometrics*, vol. 35, pp. 109–135, 1993.

23.

S. Schaal, S.Vijayakumar, and C.G. Atkeson, “Local dimensionality reduction,” in *Advances in Neural Information Processing Systems*, vol. 10, edited by M.I. Jordan, M.J. Kearns, and S.A. Solla, MIT Press: Cambridge, MA, pp. 633–639, 1998.

24.

S. Vijayakumar and S. Schaal, “Locally weighted projection regression: An O(n) algorithm for incremental real time learning in high dimensional spaces,” in *Proceedings of the Seventeenth International Conference on Machine Learning 2000 (ICML 2000)*, Stanford, CA.

25.

W.P. Press, B.P. Flannery, S.A. Teukolsky, and W.T. Vetterling, *Numerical Recipes in C*—“The Art of Scientific Computing,” Press Syndiacate University of Cambridge: Cambridge, MA, 1989.

26.

M. Kawato, “Internal models for motor control and trajectory planning,” *Curr. Opin. Neurobiol*, vol. 9, pp. 718–727, 1999.

27.

D.M. Wolpert, R.C. Miall, and M. Kawato, “Internal models in the cerebellum,” *Trends in Cognitive Sciences*, vol. 2, pp. 338–347, 1998.

28.

S. Schaal and C.G. Atkeson, “Robot juggling: An implementation of memory-based learning,” *Control Systems Magazine*, vol. 14, pp. 57–71, 1994.

29.

P. Dyer and S.R. McReynolds, *The Computation and Theory of Optimal Control*, Academic Press: New York, 1970.

30.

J.J. Craig, *Introduction to Robotics*, Addison-Wesley: Reading, MA, 1986.

31.

C.H. An, C.G. Atkeson, and J.M. Hollerbach, *Model-Based Control of a Robot Manipulator*, MIT Press: Cambridge, MA, 1988.

32.

J. Baillieul and D.P. Martin, “Resolution of kinematic redundancy,” in *Proceedings of Symposia in Applied Mathematics, American Mathematical Society*, pp. 49–89, 1990.