Skip to main content

Task Planning in “Block World” with Deep Reinforcement Learning

  • Conference paper
  • First Online:
Biologically Inspired Cognitive Architectures (BICA) for Young Scientists (BICA 2017)

Abstract

At the moment reinforcement learning have advanced significantly with discovering new techniques and instruments for training. This paper is devoted to the application convolutional and recurrent neural networks in the task of planning with reinforcement learning problem. The aim of the work is to check whether the neural networks are fit for this problem. During the experiments in a block environment the task was to move blocks to obtain the final arrangement which was the target. Significant part of the problem is connected with the determining on the reward function and how the results are depending in reward’s calculation. The current results show that without modifying the initial problem into more straightforward ones neural networks didn’t demonstrate stable learning process. In the paper a modified reward function with sub-targets and euclidian reward calculation was used for more precise reward determination. Results have shown that none of the tested architectures were not able to achieve goal.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Institutional subscriptions

References

  1. Bentivegna, D.C., Ude, A., Atkenson, C.G., Gordon, C.: Humanoid robot learning and game playing using PC-based vision, Switzerland (2002)

    Google Scholar 

  2. Mnih, V.: Playing atari with deep reinforcement learning. In: NIPS 2013 (2013)

    Google Scholar 

  3. Finn, C., Levine, S.: Deep visual foresight for planning robot motion. In: ICRA 2017 (2017)

    Google Scholar 

  4. Mnih, V., Heess, N., Graves, A., Kavukcuoglu, K.: Recurrent models of visual attention (2014)

    Google Scholar 

  5. Katyal, K.D., Staley, E.W., Johannes, M.S., Wang, I.-J., Reiter, A., Burlina, P.: In-hand robotic manipulation via deep reinforcement learning (2017)

    Google Scholar 

Download references

Acknowledgements

The reported study was supported by RFBR, research Projects No. 16-37-60055 and No. 17-07-00281.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Alekasndr I. Panov .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer International Publishing AG

About this paper

Cite this paper

Ayunts, E., Panov, A.I. (2018). Task Planning in “Block World” with Deep Reinforcement Learning. In: Samsonovich, A., Klimov, V. (eds) Biologically Inspired Cognitive Architectures (BICA) for Young Scientists. BICA 2017. Advances in Intelligent Systems and Computing, vol 636. Springer, Cham. https://doi.org/10.1007/978-3-319-63940-6_1

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-63940-6_1

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-63939-0

  • Online ISBN: 978-3-319-63940-6

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics