Abstract
The analysis of human affective behavior has attracted increasing attention from researchers in psychology, computer science, neuroscience, and related disciplines. We focus on the recognition of the affect state of a single person from video streams. We create a model that allows to estimate the state of four affective dimensions of a person which are arousal, anticipation, power and valence. This sequence model is composed of a magnitude model of motion constructed from a set of point of interest tracked using optical flow. The state of the affective dimension is then predicted using SVM. The experimentation has been performed on a standard dataset and has showed promising results.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Fontaine, J.R., Scherer, K.R., Roesch, E.B., Ellsworth, P.C.: The world of emotions is not two-dimensional. Psychological Science 18(2), 1050–1057 (2007)
Zeng, Z., Pantic, M., Roisman, G.I., Huang, T.S.: A Survey of Affect Recognition Methods: Audio, Visual, and Spontaneous Expressions. IEEE Transactions on Pattern Analysis and Machine Intelligence 31(1), 39–58 (2009)
Calvo, R.A., D’Mello, S.: Affect Detection: An Interdisciplinary Review of Models, Methods, and Their Applications. IEEE Transactions on Affective Computing 1(1), 18–37 (2010)
Nicolaou, M.A., Gunes, H., Pantic, M.: Continuous Prediction of Spontaneous Affect from Multiple Cues and Modalities in Valence-Arousal Space. IEEE Transactions on Affective Computing 2(2), 92–105 (2011)
Gunes, H., Piccardi, M.: Automatic temporal segment detection and affect recognition from face and body display. IEEE Transactions on Systems, Man, and Cybernetics Part B 39(1), 64–84 (2009)
Chen, S., Tian, Y., Liu, Q., Metaxas, D.N.: Recognizing expressions from face and body gesture by temporal normalized motion and appearance features. Image and Vision Computing 31(2), 175–185 (2013)
Benabbas, Y., Amir, S., Lablack, A., Djeraba, C.: Human Action Recognition using Direction and Magnitude Models of Motion. In: International Conference on Computer Vision Theory and Applications, pp. 277–285 (2011)
Shi, J., Tomasi, C.: Good features to track. In: Internatioal Conference on Computer Vision and Pattern Recognition, pp. 593–600 (1994)
Viola, P., Jones, M.J.: Rapid object detection using a boosted cascade of simple features. In: Internatioal Conference on Computer Vision and Pattern Recognition, pp. 511–518 (2001)
McKeown, G., Valstar, M.F., Cowie, R., Pantic, M.: The SEMAINE Corpus of emotionally coloured character interactions. In: IEEE International Conference on Multimedia and Expo, pp. 1079–1084 (2010)
Schuller, B., Valstar, M., Eyben, F., McKeown, G., Cowie, R., Pantic, M.: AVEC 2011–the first international audio/Visual emotion challenge. In: D’Mello, S., Graesser, A., Schuller, B., Martin, J.-C. (eds.) ACII 2011, Part II. LNCS, vol. 6975, pp. 415–424. Springer, Heidelberg (2011)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Hadjerci, O., Lablack, A., Bilasco, I.M., Djeraba, C. (2014). Affect Recognition Using Magnitude Models of Motion. In: Gurrin, C., Hopfgartner, F., Hurst, W., Johansen, H., Lee, H., O’Connor, N. (eds) MultiMedia Modeling. MMM 2014. Lecture Notes in Computer Science, vol 8326. Springer, Cham. https://doi.org/10.1007/978-3-319-04117-9_33
Download citation
DOI: https://doi.org/10.1007/978-3-319-04117-9_33
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-04116-2
Online ISBN: 978-3-319-04117-9
eBook Packages: Computer ScienceComputer Science (R0)