skip to main content
10.1145/2993148.2997632acmconferencesArticle/Chapter ViewAbstractPublication Pagesicmi-mlmiConference Proceedingsconference-collections
short-paper

Video-based emotion recognition using CNN-RNN and C3D hybrid networks

Published:31 October 2016Publication History

ABSTRACT

In this paper, we present a video-based emotion recognition system submitted to the EmotiW 2016 Challenge. The core module of this system is a hybrid network that combines recurrent neural network (RNN) and 3D convolutional networks (C3D) in a late-fusion fashion. RNN and C3D encode appearance and motion information in different ways. Specifically, RNN takes appearance features extracted by convolutional neural network (CNN) over individual video frames as input and encodes motion later, while C3D models appearance and motion of video simultaneously. Combined with an audio module, our system achieved a recognition accuracy of 59.02% without using any additional emotion-labeled video clips in training set, compared to 53.8% of the winner of EmotiW 2015. Extensive experiments show that combining RNN and C3D together can improve video-based emotion recognition noticeably.

References

  1. Dhall, A., Goecke, R., Lucey, S. and Gedeon, T. 2012. Collecting large, richly annotated facial-expression databases from movies. IEEE Multimedia. Google ScholarGoogle ScholarDigital LibraryDigital Library
  2. Dhall, A., Goecke, R. and Gedeon, T. 2015. Automatic Group Happiness Intensity Analysis. IEEE Transaction on Affective Computing.Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. Yao, A., Shao, J., Ma, N. and Chen,Y. 2015. Capturing AUAware Facial Features and Their Latent Relations for Emotion Recognition in the Wild. ACM ICMI. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Liu, M., Wang, R., Li, S., Shan, S., Huang Z. and Chen, X.2014. Combining Multiple Kernel Methods on Riemannian Manifold for Emotion Recognition in the Wild. ACM ICMI. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. Ebrahimi Kahou, S., Michalski, V., Konda, K., Memisevic, R., and Pal, C. 2015. Recurrent neural networks for emotion recognition in video. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction. 467- 474. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Tran, D., Bourdev, L., Fergus, R., Torresani, L., & Paluri, M. 2015.Google ScholarGoogle Scholar
  7. Learning spatiotemporal features with 3d convolutional networks. In 2015 IEEE International Conference on Computer Vision (ICCV) .4489-4497. IEEE. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. Eyben, F., Wöllmer, M., & Schuller, B. (2010, October). Opensmile: the munich versatile and fast open-source audio feature extractor. InProceedings of the 18th ACM international conference on Multimedia. 1459-1462. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Donahue, J., Anne Hendricks, L., Guadarrama, S., Rohrbach, M., Venugopalan, S., Saenko, K., & Darrell, T. 2015. Longterm recurrent convolutional networks for visual recognition and description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.2625-2634.Google ScholarGoogle Scholar
  10. Kahou, S. E., Pal, C., Bouthillier, X., Froumenty, P., Gülçehre, Ç, Memisevic, R. and Mirza, M. 2013. Combining modality specific deep neural networks for emotion recognition in video. In Proceedings of the 15th ACM on International conference on multimodal interaction. 543-550. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell. 2014. Caffe: Convolutional architecture for fast feature embedding. In ACM MM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  12. Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D. and Rabinovich, A. 2015. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.1-9.Google ScholarGoogle Scholar
  13. He, K., Zhang, X., Ren, S. and Sun, J. 2015. Deep residual learning for image recognition. arXiv preprint arXiv:1512.03385.Google ScholarGoogle Scholar
  14. Simonyan, K., & Zisserman, A. 2014. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556.Google ScholarGoogle Scholar
  15. Parkhi, O. M., Vedaldi, A., & Zisserman, A. 2015. Deep face recognition. In British Machine Vision Conference (Vol. 1, No. 3, p. 6).Google ScholarGoogle ScholarCross RefCross Ref
  16. Deng, J., Dong, W., Socher, R., Li, L. J., Li, K. and Li, F.F., L. 2009. Imagenet: A large-scale hierarchical image database. In Computer Vision and Pattern Recognition. CVPR. 248- 255. IEEE.Google ScholarGoogle Scholar
  17. Carrier, P. L., Courville, A., Goodfellow, I. J., Mirza, M. and Bengio, Y. 2013 .FER-2013 face database. Technical report, 1365, Université de Montréal.Google ScholarGoogle Scholar
  18. Graves, A., Liwicki, M., Fernández, S., Bertolami, R., Bunke, H. and Schmidhuber, J. 2009. A novel connectionist system for unconstrained handwriting recognition. IEEE transactions on pattern analysis and machine intelligence, 31(5), 855-868. Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Sak, H., Senior, A. W. and Beaufays, F. 2014. Long shortterm memory recurrent neural network architectures for large scale acoustic modeling. In INTERSPEECH. 338-342.Google ScholarGoogle Scholar
  20. Kim, B. K., Dong, S. Y., Roh, J., Kim, G. and Lee, S. Y. 2016. Fusing Aligned and Non-Aligned Face Information for Automatic Affect Recognition in the Wild: A Deep Learning Approach. In Computer Vision and Pattern Recognition. CVPR.Google ScholarGoogle Scholar
  21. Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R. and Li, F.F. 2014. Large-scale Video Classification with Convolutional Neural Networks.Google ScholarGoogle Scholar
  22. Ng, J., Hausknecht, M., Vijayanarasimhan S., Monga R., Vinyals O., Toderici G.2015. Beyond Short Snippets: Deep Networks for Video Classification. In Computer Vision and Pattern Recognition. CVPR. 4694-4702. IEEE.Google ScholarGoogle Scholar
  23. Sharma S., Kiros R., Salakhutdinov R.2016 Action Recognition using Visual Attention. Workshop track - ICLR.Google ScholarGoogle Scholar
  24. Kaya, H., Gürpinar, F., Afshar, S. and Salah, A. A. 2015. Contrasting and Combining Least Squares Based Learners for Emotion Recognition in the Wild. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction. 459-466. ACM. Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Venugopalan, S., Rohrbach, M., Donahue, J., Mooney, R., Darrell, T.m. and Saenko, K. 2015. Sequence to sequencevideo to text. In Proceedings of the IEEE International Conference on Computer Vision. 4534-4542. Google ScholarGoogle ScholarDigital LibraryDigital Library
  26. Pan, P., Xu, Z., Yang, Y., Wu, F. and Zhuang, Y. 2015.Google ScholarGoogle Scholar
  27. Hierarchical Recurrent Neural Encoder for Video Representation with Application to Captioning. arXiv preprint arXiv:1511.03476.Google ScholarGoogle Scholar
  28. Graves, A., Mohamed, A. R. and Hinton, G. 2013. Speech recognition with deep recurrent neural networks. In 2013 IEEE international conference on acoustics, speech and signal processing. 6645-6649. IEEE.Google ScholarGoogle Scholar
  29. Dhall, A., Goecke, R., Joshi, J., Hoey, J. and Gedeon, T. 2016. EmotiW 2016: Video and Group-level Emotion Recognition Challenges, ACM ICMI 2016. Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Jianguo L., Tao W., Yimin Z. 2011. ICCV: Face Detection using SURF Cascade. In Computer Vision Workshops.Google ScholarGoogle Scholar
  31. Fernández, S., Graves, A., Schmidhuber, J. 2007. An application of recurrent neural networks to discriminative keyword spotting. In International Conference on Artificial Neural Networks. 220-229. Springer Berlin Heidelberg Google ScholarGoogle ScholarDigital LibraryDigital Library

Index Terms

  1. Video-based emotion recognition using CNN-RNN and C3D hybrid networks

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Conferences
      ICMI '16: Proceedings of the 18th ACM International Conference on Multimodal Interaction
      October 2016
      605 pages
      ISBN:9781450345569
      DOI:10.1145/2993148

      Copyright © 2016 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 31 October 2016

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • short-paper

      Acceptance Rates

      Overall Acceptance Rate453of1,080submissions,42%

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader