Data Entry on the Move: An Examination of Nomadic Speech-Based Text Entry

  • Kathleen J. Price
  • Min Lin
  • Jinjuan Feng
  • Rich Goldman
  • Andrew Sears
  • Julie A. Jacko
Part of the Lecture Notes in Computer Science book series (LNCS, volume 3196)

Abstract

Desktop interaction solutions are often inappropriate for mobile devices due to small screen size and portability needs. Speech recognition can improve interactions by providing a relatively hands-free solution that can be used in various situations. While mobile systems are designed to be transportable, few have examined the effects of motion on mobile interactions. We investigated the effect of motion on automatic speech recognition (ASR) input for mobile devices. We examined speech recognition error rates (RER) with subjects walking or seated, while performing text input tasks and the effect of ASR enrollment conditions on RER. RER were significantly lower for seated conditions. There was a significant interaction between enrollment and task conditions. When users enrolled while seated, but completed walking tasks, RER increased. In contrast, when users enrolled while walking, but completed seated tasks, RER decreased. These results suggest changes in user training of ASR systems for mobile and seated usage.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Johnson, P.: Usability and mobility; Interactions on the move. Retrieved August 20, 03, from Department of Computer Science Web Site (1998), http://www.dcs.gla.ac.uk/~johnson/papers/mobile/HCIMD1.html
  2. 2.
    Lu, Y.-C., Xiao, Y., Sears, A., Jacko, J.: An observational and interview study on personal digital assistant (PDA) uses by clinicians in different contexts. In: Harris, D., Duffy, V., Smith, M. (eds.) Human-centred computing: Cognitive, social and ergonomic aspects, pp. 93–97. Lawrence Erlbaum Associates, Mahwah (2003)Google Scholar
  3. 3.
    Price, K.J., Sears, A.: Speech-based text entry for mobile devices. In: Stephanidis, C., Jacko, J. (eds.) Human-computer interaction: Theory and practice (Part II), pp. 766–770. Lawrence Erlbaum Associates, Mahwah (2003)Google Scholar
  4. 4.
    Karat, C.-M., Halverson, C., Karat, J., Horn, D.: Patterns of entry and correction in large vocabulary continuous speech recognition systems. In: Proceedings of CHI 1999, pp. 568–575. ACM Press, New York (1999)Google Scholar
  5. 5.
    Noyes, J.M., Frankish, C.R.: Errors and error correction in automatic speech recognition systems. Ergonomics 37, 1943–1957 (1994)CrossRefGoogle Scholar
  6. 6.
    Price, K. J., & Sears, A.: Speech-based data entry for handheld devices: Speed of entry and error correction techniques (Information Systems Department Technical Report). Baltimore, MD: UMBC, pp.1-8 (2002) Google Scholar
  7. 7.
    Sears, A., Feng, J., Oseitutu, K., Karat, C.-M.: Hands-free, speech-based navigation during dictation: Difficulties, consequences, and solutions. Human-Computer Interaction 18, 229–257 (2003)CrossRefGoogle Scholar
  8. 8.
    McCormick, J.: Speech Recognition. Government Computer News 22(22), 24–28 (2003)Google Scholar
  9. 9.
    Paterno, F.: Understanding interaction with mobile devices. Interacting With Computers 15, 473–478 (2003)CrossRefGoogle Scholar
  10. 10.
    Iacucci, G., Kuutti, K., Ranta, M.: On the move with a magic thing: Role playing in concept design of mobile services and devices. In: (Ed.): Proceedings of the conference on designing interactive systems: Processes, practices, methods, and techniques, pp. 193–202. ACM Press, New York (2000)CrossRefGoogle Scholar
  11. 11.
    Dahlbom, B., Ljungberg, F.: Mobile informatics. Scandinavian Journal of Information Systems 10(1 & 2), 227–234 (1998)Google Scholar
  12. 12.
    Brodie, J., Perry, M.: Designing for mobility, collaboration and information use by bluecollar workers. SIGGROUP Bulletin 22(3), 22–27 (2001)Google Scholar
  13. 13.
    Perry, M., O’Hara, K., Sellen, A., Brown, B., Harper, R.: Dealing with mobility: Understanding access anytime, anywhere. ACM Transactions on Computer-Human Interaction 8(4), 323–347 (2001)CrossRefGoogle Scholar
  14. 14.
    Satyanarayanan, M.: Fundamental challenges in mobile computing. In: Proceedings of the fifteenth annual ACM symposium on principles of distributed computing, pp. 1–7. ACM Press, New York (1996)CrossRefGoogle Scholar
  15. 15.
    Brewster, S., Lumsden, J., Bell, M., Hall, M., Tasker, S.: Multi-modal ’eyes-free’ interaction techniques for wearable devices. Letters to CHI 5(1), 473–480 (2003)Google Scholar
  16. 16.
    Pascoe, J., Ryan, N., Morse, D.: Using while moving: HCI issues in fieldwork environments. ACM Transactions on Computer-Human Interaction 7(3), 417–437 (2000)CrossRefGoogle Scholar
  17. 17.
    Sawhney, N., Schmandt, C.: Nomadic Radio: Speech and audio interaction for contextual messaging in nomadic environments. ACM Transactions on Computer-Human Interactions 7(3), 353–383 (2000)CrossRefGoogle Scholar
  18. 18.
    Bradford, J.H.: The human factors of speech-based interfaces. SIGCHI Bulletin 27(2), 61–67 (1995)CrossRefMathSciNetGoogle Scholar
  19. 19.
    Ward, K., Novick, D.G.: Accessibility: Hands-free documentation. In: (Ed.): Proceedings of the 21st annual international conference on documentation, pp. 147–154. ACM Press, New York (2003)CrossRefGoogle Scholar
  20. 20.
    Cohen, P.R., Oviatt, S.L.: The role of voice in human-machine communication. In: Roe, D.B., Wilpon, J. (eds.) Human-Computer Interaction by Voice, pp. 1–36. National Academy of Sciences Press, Washington (1993)Google Scholar
  21. 21.
    Shneiderman, B.: Communications of the ACM 43(9), 63–65 (2000)CrossRefGoogle Scholar
  22. 22.
    Hagen, A., Connors, D.A., Pellom, B.L.: The analysis and design of architecture systems for speech recognition on modern handheld-computing devices. In: (Ed.): Proceedings of the international symposium on systems synthesis, pp. 65–70. ACM Press, New York (2003)Google Scholar
  23. 23.
    Holzman, T.G.: Speech-audio interface for medical information management in field environments. International Journal of Speech Technology 4, 209–226 (2001)MATHCrossRefGoogle Scholar
  24. 24.
    Entwistle, M.S.: The performance of automated speech recognition systems under adverse conditions of human exertion. International Journal of Human-Computer Interaction 16(2), 127–140 (2003)CrossRefGoogle Scholar
  25. 25.
    Fiscus, J. G., Fisher, W. M., Martin, A. F., Przybocki, M. A., Pallett, D. S.: NIST evaluation of conversational speech recognition over the telephone: English and Mandarin performance results. from (2000), http://www.nist.gov/speech/publications/tw00/pdf/cts10.pdf Retrieved February 28, 2004
  26. 26.
    Hart, S.G., Staveland, L.E.: Development of NASA-TLX (Task Load Index): Results of empirical and theoretical research. In: Hancock, P.A., Meshkati, N. (eds.) Human mental workload, pp. 139–183. Elsevier Science Publishers B.V. Amsterdam (1988)CrossRefGoogle Scholar
  27. 27.
    NASA Ames Research Center: NASA Human Performance Research Group Task Load Index (NASA-TLX) instruction manual [Brochure]. Moffett Field, CA (1987) Google Scholar
  28. 28.
    Emery, V.K., Moloney, K.P., Jacko, J.A., Sainfort, F.: Assessing workload in the context of human-computer interactions: Is the NASA-TLX a suitable measurement tool? In: Laboratory for Human-Computer Interaction and Health Care Informatics, Georgia Institute of Technology, Atlanta, Georgia (February 2004)Google Scholar
  29. 29.
    Chandrasekhar, A.: Respiratory rate and pattern of breathing: To evaluate one of the vital signs. from Loyola University Medical Education Network Web Site, n.d. (2003), http://www.meddean.luc.edu/lumen/meded/medicine/pulmonar/pd/step73a.htm Retrieved October 14
  30. 30.
    Doust, J.H., Patrick, J.M.: The limitation of exercise ventilation during speech. Respiration Physiology 46, 137–147 (1981)CrossRefGoogle Scholar
  31. 31.
    Meckel, Y., Rotstein, A., Inbar, O.: The effects of speech production on physiologic responses during submaximal exercise. Medicine and Science in Sports and Exercise 34(8), 1337–1343 (2002)CrossRefGoogle Scholar
  32. 32.
    Feng, J., Sears, A., & Karat, C-M.: A longitudinal evaluation of hands-free speech-based navigation during dictation (Information Systems Department Technical Report). UMBC, Information Systems Department, ISRC, Baltimore, MD (2004) Google Scholar
  33. 33.
    Feng, J., Sears, A.: Using confidence scores to improve hands-free speech-based navigation. In: Stephanidis, C., Jacko, J. (eds.) Human-Computer Interaction: Theory and Practice, vol. 2, pp. 641–645. Lawrence Erlbaum Associates, Mahwah (2003)Google Scholar
  34. 34.
    Rollins, A.M.: Speech recognition and manner of speaking in noise and in quiet. In: (Ed.): CHI 1985 Proceedings, pp. 197–199. ACM Press, New York (1985)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2004

Authors and Affiliations

  • Kathleen J. Price
    • 1
  • Min Lin
    • 1
  • Jinjuan Feng
    • 1
  • Rich Goldman
    • 1
  • Andrew Sears
    • 1
  • Julie A. Jacko
    • 2
  1. 1.UMBC, Information Systems DepartmentInteractive Systems Research CenterBaltimoreUSA
  2. 2.Georgia Institute of Technology, School of Industrial & Systems EngineeringAtlantaUSA

Personalised recommendations