This is the second special issue of Machine Learning on the subject of reinforcement learning. The first, edited by Richard Sutton in 1992, marked the development of reinforcement learning into a major component of the machine learning field. Since then, the area has expanded further, accounting for a significant proportion of the papers at the annual International Conference on Machine Learning and attracting many new researchers.
Unable to display preview. Download preview PDF.
- Dayan, Peter & Sejnowski, Terrence J., (1994). TD(λ) converges with probability 1. Machine Learning, 14(3).Google Scholar
- Dietterich, Thomas G., (1986). Learning at the knowledge level. Machine Learning, 1(3):287–315.Google Scholar
- Dietterich, Thomas G. & Flann, Nicholas S., (1995). Explanation-based learning and reinforcement learning: A unified view. In Proceedings of the Twelfth International Conference on Machine Learning, pages 176–184, Tahoe City, California. Morgan Kaufmann.Google Scholar
- Sutton, Richard S., (1988). Learning to predict by the method of temporal differences. Machine Learning, 3(1):9–44.Google Scholar
- Tesauro, Gerald, (1995). Temporal difference learning and TD-Gammon. Communications of the ACM, pages 58–67.Google Scholar
- Tsitsiklis, John N., (1994). Asynchronous stochastic approximation and Q-learning. Machine Learning, 16(3).Google Scholar
- Watkins, C. J. C. H., (1989). Learning from Delayed Rewards. PhD thesis, King’s College, Cambridge.Google Scholar
- Zhang, Wei & Dietterich, Thomas G., (1995). A reinforcement learning approach to job-shop scheduling. In Proceedings of the Fourteenth International Joint Conference on Artificial Intelligence, pages 1114–1120. Montreal, Canada. Morgan Kaufmann.Google Scholar