Effective Path Modification Using Intuitive Gestures

  • Hongzhe Liu
  • Seoungjae Cho
  • Yulong Xi
  • Kyungeun Cho
  • Kyhyun Um
Conference paper
Part of the Lecture Notes in Electrical Engineering book series (LNEE, volume 301)

Abstract

Human–Computer Interaction (HCI) approaches have been rapidly changing in tandem with advances in computer technologies. The Natural User Interface (NUI) concept is being actively investigated at present because it enables more intuitive and easier access than existing user interfaces. Most of the existing research on path planning aims at automatically generating paths for autonomous driving. However, current automatic path generation approaches have difficulty coping with emergency events during autonomous driving. This paper proposes an interface that overcomes this difficulty by enabling rapid path modification using gestures during autonomous driving. The gesture execution methods, algorithm to recognize gestures, and gesture recognition accuracy are described.

Keywords

Human–computer interaction Natural user interface Gesture Path modification 

Notes

Acknowledgments

This research was supported by the MSIP (Ministry of Science, ICT and Future Planning), Korea, under the ITRC (Information Technology Research Center) support program (NIPA-2013-H0301-13-4007) supervised by the NIPA (National IT Industry Promotion Agency).

References

  1. 1.
    Rong L, Kong Z, Xiang X, Yueqiao L, Zan L (2011) Research of interactive 3D virtual fitting room on web environment. In: 2011 fourth international symposium on computational intelligence and design (ISCID), vol 1, pp 32–35Google Scholar
  2. 2.
    Abid MR, Melo LBS, Petriu EM Dynamic sign language and voice recognition for smart home interactive application. In: 2013 IEEE international symposium on medical measurements and applications proceedings (MeMeA), pp 139–144Google Scholar
  3. 3.
    Gelb D, Subramanian A, Tan K (2011) Augmented reality for immersive remote collaboration. To be published and presented at IEEE workshop on person oriented vision (POV) 2011, Kona HI, 7 Jan 2011Google Scholar
  4. 4.
    Nielsen M, Störring M, Moeslund T, Granum E (2004) A procedure for developing intuitive and ergonomic. In: 5th international gesture workshop, pp 409–420Google Scholar
  5. 5.
    Kim J, Gracanin D, Matkovic K, Quek F (2009) iPhone/iPod touch as input devices for navigation in immersive virtual environments. In: IEEE proceedings of virtual reality conference 2009, March 2009Google Scholar
  6. 6.
    Zhen-zhong Y, Wei-cou Z, Xin L, Jing H (2013) Real-time local path planning for mobile robot based on Kinect. Proc Comput Eng 39(4):243–247Google Scholar
  7. 7.
    Mojtahedzadeh R (2011) Robot obstacle avoidance using the Kinect. Master’s thesis in computer science (30 ECTS credits) at the systems, control and robotics Master’s Program Royal Institute of Technology year 2011Google Scholar
  8. 8.
    Weichert F, Bachmann D, Rudak B, Fisseler D (2013) Analysis of the accuracy and robustness of the leap motion controller. Sensors 6380–6393Google Scholar
  9. 9.
    Vikram S, Li L, Russell S (2013) Handwriting and gestures in the air, recognizing on the fly. In: CHI 2013 extended abstracts, 27 April–May 2, 2013, Paris, France, ACM 978-1-4503-1952-2/13/04Google Scholar

Copyright information

© Springer Science+Business Media Dordrecht 2014

Authors and Affiliations

  • Hongzhe Liu
    • 1
  • Seoungjae Cho
    • 1
  • Yulong Xi
    • 1
  • Kyungeun Cho
    • 1
  • Kyhyun Um
    • 1
  1. 1.Department of Multimedia EngineeringDongguk UniversitySeoulSouth Korea

Personalised recommendations