Abstract
This paper described a bi-directional translating system between Japanese and Japanese Sign Language and two experiments conducted to clarify necessary conditions for displaying easy-to-read animation of a person model speaking sign language synthesized by the system. In Experiment 1 hearing-impaired and hearing subjects judged identity of a pair of animation. Resolution and frame rate had an main effect against correct answer rate. While the hear ing had tendency to watch the whole body of the model, the hearing-impaired focused their attention chiefly on handshape and movement. In Experiment 2 the hearing-impaired read signs on animation. For correct reading above 90 % sign animation needed to be displayed with more than 10 x 8 cm, 113 x 90 pixels and 8 f/sec. This condition was much relaxed compared with that in Experiment 1.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Kurokawa, T., Morichi, T., Watanabe, S.: Bidirectional Translation between Sign Language and Japanese for Communication with Deaf-Mute People. Advances in Human Factors/Ergonomics 19B, 1109–1114 (1993)
Kawano, S., Kurokawa, T.: Development of a System Translating Japanese into Japanese Sign Language. J. Japan Soc. Fuzzy Theory and Intel. Informatics 16, 485–491 (2004)
Fujishige, E., Jan, K., Kurokawa, T.: Recognition of Japanese Sign Language Words Using Finite Automata. Human-Computer Interaction 2, 167–171 (1999)
Miyashita, J., Iwata, K., Morimoto, K., Kurokawa, T.: Improvements in Motor Control Methods of a Human Model for Sign Animation and Their Effects. In: Proc. Human Interface 2003, pp. 305–308 (2003)
Kurokawa, T.: Representation of Sign Animation for Japanese-into-Japanese Sign Language Translation. J. Visualization Soc. Japan 24(1), 273–276 (2004)
Kawano, S., Takahashi, Y., Kurokawa, T.: An Analysis of Blinking, Gazing, Nodding and Their Meanings in Sign Language Interpretation of TV News. Human Interface News and Report 11, 435–440 (1996)
Demura, T., Itou, T., Fujishige, E., Kurokawa, T.: Analysis of Deictics in Japanese Sign Language by Using a Database of JSL Sentences. Rep. 3rd IEICE Technical Meeting for Sign Language Informatics, 37–40 (2000)
Kawano, S., Kurokawa, T.: The Effects of Facial Expression on Understanding Japanese Sign Language Animation. Human-Computer Interaction 1, 783–787 (1999)
Kawano, S., Kurokawa, T.: Facial and Head Movements of a Sign Interpreter and Their Application to Japanese Sign Animation. In: Miesenberger, K., Klaus, J., Zagler, W., Burger, D. (eds.) ICCHP 2004. LNCS, vol. 3118, pp. 1172–1177. Springer, Heidelberg (2004)
Johnson, B.F., Caird, J.K.: The Effect of Frame Rate and Video Information Redundancy on the Perceptual Learning of American Sign Language Gestures. In: Proc. CHI 1996, pp. 121–122 (1996)
Shionome, T., Kamata, K., Yamamoto, H.: A Consideration on Relationship between Sign Language Perception and Screen Size. Tech. Rept. of IEICE, HCS2003-61, 29–34 (2003)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Kawano, S., Izumi, C., Kurokawa, T., Morimoto, K. (2006). Japanese JSL Translation and Searching Display Conditions for Expressing Easy-to-Understand Sign Animation. In: Miesenberger, K., Klaus, J., Zagler, W.L., Karshmer, A.I. (eds) Computers Helping People with Special Needs. ICCHP 2006. Lecture Notes in Computer Science, vol 4061. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11788713_98
Download citation
DOI: https://doi.org/10.1007/11788713_98
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-36020-9
Online ISBN: 978-3-540-36021-6
eBook Packages: Computer ScienceComputer Science (R0)