Abstract
The focus of this research is the creation of a deep reinforcement learning approach to tackle the challenging task of robotic gripping through tactile sensor data feedback. Leveraging deep reinforcement learning, we have sidestepped the necessity to design features manually, which simplifies the issue and allows the robot to acquire gripping strategies via trial-and-error learning. Our technique utilizes an off-policy reinforcement learning model, integrating deep deterministic policy gradient structure and twin delayed attributes to facilitate maximum precision in gripping floating items. We have formulated a comprehensive reward function to provide the agent with precise, insightful feedback to facilitate the learning of the gripping task. The training of our model was executed solely in a simulated environment using the PyBullet framework and did not require demonstrations or pre-existing knowledge of the task. We examined a gripping task with a 3-finger Robotiq gripper for a case study, where the gripper had to approach a floating object, pursue it, and eventually grip it. This training methodology in a simulated setting allowed us to experiment with various scenarios and conditions, thereby enabling the agent to develop a resilient and adaptable grip policy.
Similar content being viewed by others
Data availability
Data is not available publicly for this paper.
References
Shan M, Guo J, Gill E (2016) Review and comparison of active space debris capturing and removal methods. Prog Aerosp Sci 80:18–32. https://doi.org/10.1016/j.paerosci.2015.11.001
Zhao P, Liu J, Wu C (2020) Survey on research and development of on-orbit active debris removal methods. Sci China Technol Sci 63(11):2188–2210. https://doi.org/10.1007/s11431-020-1661-7
Macauley MK (2015) The economics of space debris: Estimating the costs and benefits of debris mitigation. Acta Astronaut 115:160–164. https://doi.org/10.1016/j.actaastro.2015.05.006
Schaub H, Jasper LEZ, Anderson PV, McKnight DS (2015) Cost and risk assessment for spacecraft operation decisions caused by the space debris environment. Acta Astronaut 113:66–79. https://doi.org/10.1016/j.actaastro.2015.03.028
Rybus T (2018) Obstacle avoidance in space robotics: Review of major challenges and proposed solutions. Prog Aerosp Sci 101:31–48. https://doi.org/10.1016/j.paerosci.2018.07.001
Ledkov A, Aslanov V (2022) Review of contact and contactless active space debris removal approaches. Prog Aerosp Sci 134:100858. https://doi.org/10.1016/j.paerosci.2022.100858
Matney M et al (2019) The NASA orbital debris engineering model 3.1: development, verification, and validation. In: International orbital debris conference (IOC)
Papadopoulos E, Aghili F, Ma O, Lampariello R (2021) Robotic manipulation and capture in space: a survey. Front Robot AI. https://doi.org/10.3389/frobt.2021.686723
Sun Y, Falco J, Roa MA, Calli B (2022) Research challenges and progress in robotic grasping and manipulation competitions. IEEE Robot Autom Lett 7(2):874–881. https://doi.org/10.1109/LRA.2021.3129134
Mnih V et al (2013) Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602
Mnih V et al (2015) Human-level control through deep reinforcement learning. Nature 518:7540. https://doi.org/10.1038/nature14236
Lample G, Chaplot DS (2017) Playing FPS games with deep reinforcement learning. In: Proceedings of the AAAI conference on artificial intelligence. https://doi.org/10.1609/aaai.v31i1.10827
Duan Y, Chen X, Houthooft R, Schulman J, Abbeel P (2016) Benchmarking deep reinforcement learning for continuous control. In: Proceedings of the 33rd international conference on machine learning, PMLR, June 2016, pp. 1329–1338. Available https://proceedings.mlr.press/v48/duan16.html. Accessed July 17, 2023
Guo D, Sun F, Fang B, Yang C, Xi N (2017) Robotic grasping using visual and tactile sensing. Inf Sci 417:274–286. https://doi.org/10.1016/j.ins.2017.07.017
Melnik A, Lach L, Plappert M, Korthals T, Haschke R, Ritter H (2019) Tactile sensing and deep reinforcement learning for in-hand manipulation tasks. In: IROS workshop on autonomous object manipulation
Merzić H, Bogdanović M, Kappler D, Righetti L, Bohg J (2019) Leveraging contact forces for learning to grasp. In: 2019 international conference on robotics and automation (ICRA), May 2019, pp 3615–3621. https://doi.org/10.1109/ICRA.2019.8793733.
Liu H et al (2022) Multi-fingered tactile serving for grasping adjustment under partial observation. In: 2022 IEEE/RSJ international conference on intelligent robots and systems (IROS), Oct. 2022, pp 7781–7788. https://doi.org/10.1109/IROS47612.2022.9981464
Koenig A, Liu Z, Janson L, Howe R (2021) Tactile grasp refinement using deep reinforcement learning and analytic grasp stability metrics. arXiv:2109.11234 [cs, eess], Sep. 2021. Available http://arxiv.org/abs/2109.11234. Accessed: Dec. 15, 2021
Tai L, Zhang J, Liu M, Boedecker J, Burgard W (2018) A survey of deep network solutions for learning control in robotics: from reinforcement to imitation. arXiv, Apr. 8, 2018. https://doi.org/10.48550/arXiv.1612.07139.
Benning M, Celledoni E, Ehrhardt MJ, Owren B, Schönlieb C-B (2019) Deep learning as optimal control problems: models and numerical methods. arXiv, Sep. 30, 2019. https://doi.org/10.48550/arXiv.1904.05657.
Lenz I, Lee H, Saxena A (2015) Deep learning for detecting robotic grasps. Int J Robot Res 34(4–5):705–724. https://doi.org/10.1177/0278364914549607
Saxena A, Driemeyer J, Ng AY (2008) Robotic grasping of novel objects using vision. Int J Robot Res 27(2):157–173. https://doi.org/10.1177/0278364907087172
Kumar V, Hermans T, Fox D, Birchfield S, Tremblay J (2019) Contextual reinforcement learning of visuo-tactile multi-fingered grasping policies. arXiv preprint, arXiv:1911.09233
Sutton RS, Barto AG (2018) reinforcement learning: an introduction. MIT Press, Cambridge
Buşoniu L, de Bruin T, Tolić D, Kober J, Palunko I (2018) Reinforcement learning for control: performance, stability, and deep approximators. Annu Rev Control 46:8–28. https://doi.org/10.1016/j.arcontrol.2018.09.005
Li Y (2017) Deep reinforcement learning: an overview. arXiv preprint, arXiv:1701.07274
Arulkumaran K, Deisenroth MP, Brundage M, Bharath AA (2017) Deep reinforcement learning: a brief survey. IEEE Signal Process Mag 34(6):26–38. https://doi.org/10.1109/MSP.2017.2743240
Kleeberger K, Bormann R, Kraus W, Huber MF (2020) A survey on learning-based robotic grasping. Curr Robot Rep 1(4):239–249. https://doi.org/10.1007/s43154-020-00021-6
Fujimoto S, Hoof H, Meger D (2022) Addressing function approximation error in actor-critic methods. In: Proceedings of the 35th international conference on machine learning, PMLR, July 2018, pp 1587–1596. Available https://proceedings.mlr.press/v80/fujimoto18a.html. Accessed Dec. 15, 2022
Lillicrap TP et al (2019) Continuous control with deep reinforcement learning. arXiv, Jul. 05, 2019. https://doi.org/10.48550/arXiv.1509.02971
Dankwa S, Zheng W (2020) Twin-delayed DDPG: a deep reinforcement learning technique to model a continuous movement of an intelligent robot agent. In: Proceedings of the 3rd international conference on vision, image and signal processing, in ICVISP 2019, May 2020. Association for Computing Machinery, New York, NY, USA, pp 1–5. https://doi.org/10.1145/3387168.3387199.
Raffin A, Hill A, Gleave A, Kanervisto A, Ernestus M, Dormann N (2021) Stable-baselines3: reliable reinforcement learning implementations. J Mach Learn Res 22:1–8
E. Coumans and Y. Bai. (2017). Pybullet, a Python Module for Physics Simulation in Robotics, Games and Machine Learning. [Online]. Available: https://pybullet.org
Brockman G et al (2016) Openai gym. arXiv preprint arXiv:1606.01540
Acknowledgements
This work is funded by the Discovery Grant (RGPIN-2018-05991) and CREATE Grant (555425-2021) of the Natural Sciences and Engineering Research Council of Canada, and FAST grant (19FAYORA14) of the Canadian Space Agency.
Funding
Natural Sciences and Engineering Research Council of Canada, Discovery Grant/RGPIN-2018-05991, Zheng Hong Zhu, CREATE Grant/555425-2021, Zheng Hong Zhu, Canadian Space Agency, FAST grant/19FAYORA14, Zheng Hong Zhu.
Author information
Authors and Affiliations
Contributions
All the authors contributed to the study conception and design. Material preparation, data collection, and analysis were performed by BB and ZHZ. The first draft of the manuscript was written by BB, and all the authors commented on previous versions of the manuscript. All the authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Beigomi, B., Zhu, Z.H. Utilizing deep reinforcement learning for tactile-based autonomous capture of non-cooperative objects in space. AS (2023). https://doi.org/10.1007/s42401-023-00254-1
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s42401-023-00254-1