Abstract
We study the problem of expression analysis for a group of people. Automatic facial expression analysis has seen much research in recent times. However, little attention has been given to the estimation of the overall expression theme conveyed by an image of a group of people. Specifically, this work focuses on formulating a framework for happiness intensity estimation for groups based on social context information. The main contributions of this paper are: a) defining automatic frameworks for group expressions; b) social features, which compute weights on expression intensities; c) an automatic face occlusion intensity detection method; and d) an ‘in the wild’ labelled database containing images having multiple subjects from different scenarios. The experiments show that the global and local contexts provide useful information for theme expression analysis, with results similar to human perception results.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Zeng, Z., Pantic, M., Roisman, G., Huang, T.: A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 39–58 (2009)
Whitehill, J., Littlewort, G., Fasel, I.R., Bartlett, M.S., Movellan, J.R.: Toward Practical Smile Detection. IEEE TPAMI, 2106–2111 (2009)
Gallagher, A., Chen, T.: Understanding Images of Groups Of People. In: IEEE CVPR, pp. 256–263 (2009)
Wang, G., Gallagher, A., Luo, J., Forsyth, D.: Seeing People in Social Context: Recognizing People and Social Relationships. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part V. LNCS, vol. 6315, pp. 169–182. Springer, Heidelberg (2010)
Stone, Z., Zickler, T., Darell, T.: Autotagging facebook: Social network context improves photo annotation. In: IEEE CVPR (2008)
Hernandez, J., Hoque, E.: MIT Mood Meter (2011), http://www.moodmeter.media.mit.edu
Fiss, J., et al.: Candid portrait selection from video. ACM Trans. Graph, 128 (2011)
Eichner, M., Ferrari, V.: We Are Family: Joint Pose Estimation of Multiple Persons. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part I. LNCS, vol. 6311, pp. 228–242. Springer, Heidelberg (2010)
Lin, D., Tang, X.: Quality-Driven Face Occlusion Detection and Recovery. In: IEEE CVPR (2007)
Rosipal, R.: Nonlinear Partial Least Squares: An Overview. In: Chemoinformatics and Advanced Machine Learning Perspectives: Complex Computational Methods and Collaborative Techniques, pp. 169–189. ACM, IGI Global (2011)
Guo, G., Mu, G.: Simultaneous dimensionality reduction and human age estimation via kernel partial least squares regression. In: IEEE CVPR, pp. 657–664 (2011)
Schwartz, W.R., Kembhavi, A., Harwood, D., Davis, L.S.: Human detection using partial least squares analysis. In: IEEE ICCV, pp. 24–31 (2009)
Schwartz, W.R., Guo, H., Davis, L.S.: A Robust and Scalable Approach to Face Identification. In: Daniilidis, K., Maragos, P., Paragios, N. (eds.) ECCV 2010, Part VI. LNCS, vol. 6316, pp. 476–489. Springer, Heidelberg (2010)
Bosch, A., Zisserman, A., Munoz, X.: Representing Shape with a Spatial Pyramid Kernel. In: CIVR, pp. 401–408 (2007)
Gehrig, T., Ekenel, H.K.: Facial action unit detection using kernel partial least squares. In: ICCV Workshops, pp. 2092–2099 (2011)
Parikh, D., Grauman, K.: Relative attributes. In: ICCV, pp. 503–510 (2011)
Tsai, G., Xu, C., Liu, J., Kuipers, B.: Real-time indoor scene understanding using bayesian filtering with motion cues. In: ICCV, pp. 121–128 (2011)
Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent dirichlet allocation. In: NIPS, pp. 601–608 (2001)
Jiang, Y.G., Ngo, C.W., Yang, J.: Towards optimal bag-of-features for object categorization and semantic video retrieval. In: CIVR, pp. 494–501 (2007)
Blei, D.M., McAuliffe, J.D.: Supervised Topic Models. In: NIPS (2007)
Zhu, J., Ahmed, A., Xing, E.P.: Medlda: maximum margin supervised topic models for regression and classification. In: ICML, p. 158 (2009)
Dhall, A., Goecke, R., Lucey, S., Gedeon, T.: Collecting large, richly annotated facial-expression databases from movies. IEEE MultiMedia 19, 34–41 (2012)
Russell, B.C., Torralba, A., Murphy, K.P., Freeman, W.T.: Labelme: A database and web-based tool for image annotation. IJCV, 157–173 (2008)
Korc, F., Schneider, D.: Annotation tool. Technical Report TR-IGG-P-2007-01, University of Bonn, Department of Photogrammetry (2007)
Viola, P.A., Jones, M.J.: Rapid object detection using a boosted cascade of simple features. In: IEEE CVPR, pp. 511–518 (2001)
Everingham, M., Sivic, J., Zisserman, A.: Hello! My name is... Buffy – Automatic Naming of Characters in TV Video. In: BMVC, pp. 899–908 (2006)
Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines (2001), http://www.csie.ntu.edu.tw/~cjlin/libsvm
Dhall, A., Asthana, A., Goecke, R., Gedeon, T.: Emotion recognition using PHOG and LPQ features. In: FG, FERA Workshop, pp. 878–883 (2011)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Dhall, A., Joshi, J., Radwan, I., Goecke, R. (2013). Finding Happiest Moments in a Social Context. In: Lee, K.M., Matsushita, Y., Rehg, J.M., Hu, Z. (eds) Computer Vision – ACCV 2012. ACCV 2012. Lecture Notes in Computer Science, vol 7725. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-37444-9_48
Download citation
DOI: https://doi.org/10.1007/978-3-642-37444-9_48
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-37443-2
Online ISBN: 978-3-642-37444-9
eBook Packages: Computer ScienceComputer Science (R0)