Abstract
The paper discusses an approach for teleoperating a mobile robot based on qualitative spatial relations, which are instructed through speech-based and deictic commands. Given a workspace containing a robot, a user and some objects, we exploit fuzzy reasoning criteria to describe the pertinence map between the locations in the workspace and qualitative commands incrementally acquired. We discuss the modularity features of the used reasoning technique through some use cases addressing a conjunction of spatial kernels. In particular, we address the problem of finding a suitable target location from a set of qualitative spatial relations based on symbolic reasoning and Monte Carlo simulations. Our architecture is analyzed in a scenario considering simple kernels and an almost-perfect perception of the environment. Nevertheless, the presented approach is modular and scalable, and it could be also exploited to design application where multi-modal qualitative interactions are considered.
L. Buoncompagni, S. Ghosh and M. Moura—These authors contributed equally to this work.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
available at https://github.com/EmaroLab/mmodal_teleop.
References
Antunes, A., Pizzuto, G., Cangelosi, A.: Communication with speech and gestures: applications of recurrent neural networks to robot language learning. In: Proceedings of GLU 2017 International Workshop on Grounding Language Understanding, pp. 4–7 (2017)
Bloch, I.: Fuzzy relative position between objects in image processing: a morphological approach. IEEE Trans. Pattern Anal. Mach. Intell. 21(7), 657–664 (1999)
Bloch, I., Saffiotti, A.: Why robots should use fuzzy mathematical morphology. In: Proceedings of the 1st International ICSC-NAISO Congress on Neuro-Fuzzy Technologies, La Havana, Cuba (2002)
Buoncompagni, L., Mastrogiovanni, F.: An open framework to develop and validate techniques for speech analysis. In: Proceedings of the 3rd Italian Workshop on Artificial Intelligence and Robotics A workshop of the XV International Conference of the Italian Association for Artificial Intelligence (AI*IA 2016), vol. 1834, pp. 15–20. Genova, Italy, CEUR-WS (2016). http://ceur-ws.org/Vol-1834/
Cruz, F., Parisi, G.I., Twiefel, J., Wermter, S.: Multi-modal integration of dynamic audiovisual patterns for an interactive reinforcement learning scenario. In: Intelligent Robots and Systems (IROS), 2016 IEEE/RSJ International Conference on, pp. 759–766. IEEE (2016)
Delaye, A., Anquetil, E.: Learning of fuzzy spatial relations between handwritten patterns. Int. J. Data Min., Model. Manag. 6(2), 127–147 (2014)
Huang, C.M., Mutlu, B.: Learning-based modeling of multimodal behaviors for humanlike robots. In: Proceedings of the 2014 ACM/IEEE International Conference on Human-robot Interaction, pp. 57–64. ACM (2014)
Lucignano, L., Cutugno, F., Rossi, S., Finzi, A.: A dialogue system for multimodal human-robot interaction. In: Proceedings of the 15th ACM on International Conference on Multimodal Interaction, pp. 197–204. ACM (2013)
Olsen, D.R., Goodrich, M.A.: Metrics for evaluating human-robot interactions. In: Proceedings of PERMIS. vol. 2003, p. 4 (2003)
Poncela, A., Gallardo-Estrella, L.: Command-based voice teleoperation of a mobile robot via a human-robot interface. Robotica 33(1), 1–18 (2015)
Potenza, A., Kiselev, A., Loutfi, A., Saffiotti, A.: Towards sliding autonomy in mobile robotic telepresence: a position paper. In: ECCE 2017-European Conference on Cognitive Ergonomics, 20–22 September 2017, Umeå University, Sweden (2017)
Prescott, T.J., Mitchinson, B., Conran, S.: Miro: An animal-like companion robot with a biomimetic brain-based control system. In: Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction, pp. 50–51. ACM (2017)
Ross, R.J., Shi, H., Vierhuff, T., Krieg-Brückner, B., Bateman, J.: Towards dialogue based shared control of navigating robots. In: Freksa, C., Knauff, M., Krieg-Brückner, B., Nebel, B., Barkowsky, T. (eds.) Spatial Cognition 2004. LNCS (LNAI), vol. 3343, pp. 478–499. Springer, Heidelberg (2005). https://doi.org/10.1007/978-3-540-32255-9_26
Srimal, P.A.S., Muthugala, M.V.J., Jayasekara, A.B.P.: Deictic gesture enhanced fuzzy spatial relation grounding in natural language. In: 2017 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), pp. 1–8. IEEE (2017)
Tan, J., Ju, Z., Liu, H.: Grounding spatial relations in natural language by fuzzy representation for human-robot interaction. In: 2014 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), pp. 1743–1750. IEEE (2014)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Buoncompagni, L., Ghosh, S., Moura, M., Mastrogiovanni, F. (2018). A Scalable Architecture to Design Multi-modal Interactions for Qualitative Robot Navigation. In: Ghidini, C., Magnini, B., Passerini, A., Traverso, P. (eds) AI*IA 2018 – Advances in Artificial Intelligence. AI*IA 2018. Lecture Notes in Computer Science(), vol 11298. Springer, Cham. https://doi.org/10.1007/978-3-030-03840-3_8
Download citation
DOI: https://doi.org/10.1007/978-3-030-03840-3_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-03839-7
Online ISBN: 978-3-030-03840-3
eBook Packages: Computer ScienceComputer Science (R0)