Head pose estimation in solving human-computer interaction problems
A system that estimates a user’s head pose by its video image is described. The system includes three basic algorithms, that is, segmentation, landmark detection, and motion direction estimation. The dynamics of change in geometric relations between facial landmarks on a sequence of frames is used to determine the direction of head motion. The change in the angle between the lines that connect the corners of the eyes and the tip of nose is shown to be similar to the yaw angle in their dynamics of change. The described system operates in real time (7 frames per second) and ensures a high precision estimate of motion direction (p = 0.95).
- 1.S. Anishenko, D. Shaposhnikov, R. Comley, and X. Gao, “Facial Image Segmentation Based on Mixed Color Space,” in Proc. XV Int. Conf. on Neurocybernetics (Rostov-on-Don, 2009), Vol. 2, pp. 231–234.Google Scholar
- 2.P. A. Bakaut and G. S. Kolmogorov, “Images Segmentation: Domain Boundaries Detecting,” Zarubezh. Electron. 10, 25–47 (1987).Google Scholar
- 5.P. Majaranta and K.-J. Räihä, “Text Entry by Gaze: Utilizing Eye-Tracking,” in Text Entry Systems: Mobility, Accessibility, Universality, Ed. by I. S. MacKenzie and K. Tanaka-Ishii (2007), pp. 175–187.Google Scholar
- 6.M. D. Fairchild, Color Appearance Models, 2nd ed. (Addison-Wesley, Reading (MA), 1997).Google Scholar
- 8.E. L. Cascia, S. Sclaroff, and V. Athitsos, “Fast, Reliable Head Tracking under Varying Illumination: An Approach Based on Registration of Texture-Mapped 3d Models,” Pattern Anal. Mach. Intell. 22(4) (2000).Google Scholar