Categorical vs. Dimensional Representations in Multimodal Affect Detection during Learning
- Cite this paper as:
- Hussain M.S., Monkaresi H., Calvo R.A. (2012) Categorical vs. Dimensional Representations in Multimodal Affect Detection during Learning. In: Cerri S.A., Clancey W.J., Papadourakis G., Panourgia K. (eds) Intelligent Tutoring Systems. ITS 2012. Lecture Notes in Computer Science, vol 7315. Springer, Berlin, Heidelberg
Learners experience a variety of emotions during learning sessions with Intelligent Tutoring Systems (ITS). The research community is building systems that are aware of these experiences, generally represented as a category or as a point in a low-dimensional space. State-of-the-art systems detect these affective states from multimodal data, in naturalistic scenarios. This paper provides evidence of how the choice of representation affects the quality of the detection system. We present a user-independent model for detecting learners’ affective states from video and physiological signals using both the categorical and dimensional representations. Machine learning techniques are used for selecting the best subset of features and classifying the various degrees of emotions for both representations. We provide evidence that dimensional representation, particularly using valence, produces higher accuracy.
KeywordsAffect multimodality machine learning learning interaction
Unable to display preview. Download preview PDF.