Face Profile View Retrieval Using Time of Flight Camera Image Analysis

Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 9124)

Abstract

Method for profile view retrieving of the human face is presented. The depth data from the 3D camera is taken as an input. The preprocessing is, besides of standard filtration, extended by the process of filling of the holes which are present in depth data. The keypoints, defined as the nose tip and the chin are detected in user’s face and tracked. The Kalman filtering is applied to smooth the coordinates of those points which can vary with each frame because of the subject’s movement in front of the camera. Knowing the locations of keypoints and having the depth data the contour of the user’s face a profile retrieval is attempted. Further filtering and modifications are introduced to the profile view in order to enhance its representation. Data processing enhancements allow emphasizing minima and maxima in the contour signals leading to discrimination of the face profiles and enable robust facial landmarks tracking.

Keywords

Depth image Signal processing Profile view Keypoints tracking 

Notes

Acknowledgments

This work was supported by the grant No. PBS3/B3/0/2014 Project ID 246459 entitled “Multimodal biometric system for bank client identity verification” co-financed by the Polish National Centre for Research and Development.

References

  1. 1.
    Kumar, K., Chen, T., Stern, R.M.: Profile view lip reading. In: ICASSP (2007)Google Scholar
  2. 2.
    Lucey, P., Potamianos, G.: Lipreading using profile versus frontal views. In: 2006 IEEE 8th Workshop on Multimedia Signal Processing, pp. 24–28 (2006)Google Scholar
  3. 3.
    Pass, A., Zhang, J., Stewart, D.: An investigation into features for multi-view lipreading. In: 17th IEEE International Conference on Image Processing (ICIP 2010), pp. 2417–2420 (2010)Google Scholar
  4. 4.
    Navarathna, R., Dean, D., Sridharan, S., Fookes, C., Lucey, P.: Visual voice activity detection using frontal versus profile views. In: 2011 International Conference on Digital Image Computing Techniques and Applications (DICTA), pp. 134–139 (2011)Google Scholar
  5. 5.
    Dalka, P., Bratoszewski, P., Czyzewski, A.: Visual lip contour detection for the purpose of speech recognition. In: ICSES, Poland (2014)Google Scholar
  6. 6.
    Kunka, B., Kupryjanow, A., Dalka, P., Szczodrak, M., Szykulski, M., Czyzewski, A.: Multimodal english corpus for automatic speech recognition. In: Signal Processing Algorithms, Architectures, Arrangements and Application, Poland (2013)Google Scholar
  7. 7.
    Czyżewski, A., Kostek, B., Ciszewski, T., Majewicz, D.: Language material for english audiovisual speech recognition system development. J. Acoust. Soc. Am. 134(5), 4069 (2013). (abstr.) plus Proceedings of Meetings on Acoustics, No. 1, vol. 20, pp. 1 − 7, San Francisco, USA, 2.12.2013 − 6.12.2013CrossRefGoogle Scholar
  8. 8.
    Zhou, X., Bhanu, B.: Human recognition based on face profiles in video. In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2005)Google Scholar
  9. 9.
    Pantic, M., Patras, I.: Dynamics of facial expression: recognition of facial actions and their temporal segments from face profile image sequences. IEEE Trans. Syst. Man Cyber. 2(2) (2006)Google Scholar
  10. 10.
    Tomasi, C., Manduchi, R.: Bilateral filtering for gray and color images. In: IEEE International Conference on Computer Vision, Bombay, India (1998)Google Scholar
  11. 11.
    Nagao, M., Matsuyama, T.: Edge preserving smoothing. Comput. Graphics Image Proc. 9, 394–407 (1979)CrossRefGoogle Scholar
  12. 12.
    Yang, N., Kim, Y., Park, R.: Depth hole filling using the depth distribution of neighboring regions of depth holes in the kinect sensor. In: ICSPCC, Honk Kong, pp. 658–661 (2012)Google Scholar
  13. 13.
    Kim, J., Piao, N., Kim, H., Park, R.: Depth hole filling for 3-d reconstruction using color and depth images. In: ISCE, South Korea (2014)Google Scholar
  14. 14.
    Barash, D.: Bilateral filtering and anisotropic diffusion: towards a unified viewpoint. In: Kerckhove, M. (ed.) Scale-Space 2001. LNCS, vol. 2106, pp. 273–280. Springer, Heidelberg (2001)CrossRefGoogle Scholar
  15. 15.
    Briechle, K., Hanebeck, U.D.: Template matching using fast normalized cross correlation. In: Proceeding of the SPIE, Optical Pattern Recognition XII, vol. 4387(95) (2001)Google Scholar
  16. 16.
    Kalman, R.E.: A new approach to linear filtering and prediction problems. ASME J Basic Eng. 82, 35–45 (1960)CrossRefGoogle Scholar
  17. 17.
    Szwoch, G., Dalka, P., Czyżewski, A.: Resolving conflicts in object tracking for automatic detection of events in video. Elektronika: konstrukcje, Technologie, zastosowania 52(1), 52–54 (2011). ISSN 0033-2089Google Scholar
  18. 18.
    Neubeck, A., Van Gool, L.: Efficient non-maximum suppression. In: 18th International Conference on Pattern Recognition (ICPR 2006), vol. 3, pp. 850–855 (2006)Google Scholar

Copyright information

© Springer International Publishing Switzerland 2015

Authors and Affiliations

  1. 1.Multimedia Systems Department, Faculty of Electronics, Telecommunications and InformaticsGdansk University of TechnologyGdanskPoland

Personalised recommendations