Skip to main content

SWOT Analysis of Behavioural Recognition Through Variable Modalities

  • Conference paper
  • First Online:
Ubiquitous Intelligent Systems (ICUIS 2021)

Part of the book series: Smart Innovation, Systems and Technologies ((SIST,volume 302))

  • 415 Accesses

Abstract

In this era of technology, recognising the human emotion to improve the human–machine interaction has become very important. The applications of emotion recognition have continued to grow in the past few years, from healthcare and security to improving customer care service and enhancing the user experience. With increased innovation, the number of use cases for this area would only grow more in the future. This forms the motivation for thoroughly analysing the market for recognition of human behaviour. The paper summarises the SWOT analysis of behavioural recognition through the use of variable modalities. This paper gives an overview of the currently employed techniques which involve the extraction of features from text, audio and video. The main objective is to get an overview of the internal potential and limitations of this branch of machine learning. The analysis brings out the merits and demerits of all the techniques discussed.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 229.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 299.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 299.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. X. Zhang, M.-J. Wang, X.-D. Guo, Multi-modal emotion recognition based on deep learning in speech, video and text, in 2020 IEEE 5th International Conference on Signal and Image Processing (ICSIP) (2020), pp. 328–333. https://doi.org/10.1109/ICSIP49896.2020.9339464

  2. V. Rozgić, S. Ananthakrishnan, S. Saleem, R. Kumar, R. Prasad, Ensemble of SVM trees for multimodal emotion recognition, in Proceedings of The 2012 Asia Pacific Signal and Information Processing Association Annual Summit and Conference (2012), pp. 1–4

    Google Scholar 

  3. S. Zhalehpour, Z. Akhtar, C. Eroglu Erdem, Multimodal emotion recognition with automatic peak frame selection, in 2014 IEEE International Symposium on Innovations in Intelligent Systems and Applications (INISTA) Proceedings (2014), pp. 116–121. https://doi.org/10.1109/INISTA.2014.6873606

  4. S. Poria, I. Chaturvedi, E. Cambria, A. Hussain, Convolutional MKL based multimodal emotion recognition and sentiment analysis, in 2016 IEEE 16th International Conference on Data Mining (ICDM) (2016), pp. 439–448. https://doi.org/10.1109/ICDM.2016.0055

  5. A. Zadeh, P.P. Liang, S. Poria, Multi-attention recurrent network for human communication comprehension, in Proceedings of the 32th AAAI Conference on Artificial Intelligence (2018), New Orleans, USA, pp. 5642–5649

    Google Scholar 

  6. A. Bandhakavi, N. Wiratunga, S. Massie, P. Deepak, Lexicon generation for emotion analysis of text. IEEE Intell. Syst. 32(1), 102–108 (2017)

    Article  Google Scholar 

  7. E. Cambria, S. Poria, D. Hazarika, K. Kwok, SenticNet 5: discovering conceptual primitives for sentiment analysis by means of context embeddings, in AAAI (2018), pp. 1795–1802

    Google Scholar 

  8. S. Poria, E. Cambria, D. Hazarika, N. Majumder, A. Zadeh, L.-P. Morency, Context-dependent sentiment analysis in user-generated videos (2017), pp. 873–883. https://doi.org/10.18653/v1/P17-1081

  9. A. Zadeh, R. Zellers, E. Pincus, L.-P. Morency, Multimodal sentiment intensity analysis in videos: facial gestures and verbal messages. IEEE Intell. Syst. 31(6), 82–88 (2016)

    Article  Google Scholar 

  10. A. Ashok, R. Elmasri, G. Natarajan, Comparing different word embeddings for multiword expression identification, in Natural Language Processing and Information Systems (Springer, Cham, 2019)

    Google Scholar 

  11. C. Fadil, R. Alvarez, C. Martínez, et al., Multimodal Emotion Recognition Using Deep Networks (Congreso Latinoamericano De Ingeniería Biomédica Claib, 2014)

    Google Scholar 

  12. C.D. Alves, G. Sullivan, Facial expression of emotion: interaction processes, emotion regulation and communication, in Handbook on Facial Expression of Emotion, vol. 1 (2013)

    Google Scholar 

  13. P.K. Atrey, M.A. Hossain, A. El Saddik, M.S. Kankanhalli, Multimodal fusion for multimedia analysis: a survey. Multimedia Syst. 16, 345–379 (2010)

    Article  Google Scholar 

  14. A. Tawari, M.M. Trivedi, Face expression recognition by cross modal data association. IEEE Trans. Multimedia 15, 1543–1552 (2013)

    Article  Google Scholar 

  15. R. Gajsek, V. Struc, F. Mihelic, Multi-modal emotion recognition using canonical correlations and acoustic features, in 2010 20th International Conference on Pattern Recognition (ICPR) (2010), pp. 4133–4136

    Google Scholar 

  16. X. Zhu, D. Ramanan, Face detection, pose estimation, and landmark localization in the wild, in 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2012), pp. 2879–2886

    Google Scholar 

  17. Y. Wang, L. Guan, A.N. Venetsanopoulos, Kernel cross-modal factor analysis for information fusion with application to bimodal emotion recognition. IEEE Trans. Multimedia 14, 597–607 (2012)

    Article  Google Scholar 

  18. M. Meghjani, F. Ferrie, G. Dudek, Bimodal information analysis for emotion recognition, in 2009 Workshop on Applications of Computer Vision (WACV) (2009), pp. 1–6

    Google Scholar 

  19. M. Li, X. Qiu, S. Peng, L. Tang, Q. Li, W. Yang, Y. Ma, Multimodal emotion recognition model based on a deep neural network with multiobjective optimization. Wirel. Commun. Mob. Comput. 2021, 10 (2021). https://doi.org/10.1155/2021/6971100, Article ID 6971100

  20. A. Kołakowska, A. Landowska, M. Szwoch, W. Szwoch, M. Wróbel, Emotion recognition and its applications. Adv. Intell. Syst. Comput. 300, 51–62 (2014). https://doi.org/10.1007/978-3-319-08491-6_5

    Article  Google Scholar 

  21. M. Dhuheir, A. Albaseer, E. Baccour, A. Erbad, M. Abdallah, M. Hamdi, Emotion recognition for healthcare surveillance systems using neural networks: a survey. Int. Wirel. Commun. Mob. Comput. (IWCMC) 2021, 681–687 (2021). https://doi.org/10.1109/IWCMC51323.2021.9498861

    Article  Google Scholar 

  22. Y. Bian, L. Zhao, H. Li, G. Yang, L. Geng, X. Deng, Research on multi-modal human-machine interface for aerospace robot, in 2015 7th International Conference on Intelligent Human-Machine Systems and Cybernetics (2015), pp. 535–538. https://doi.org/10.1109/IHMSC.2015.74

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Aakash Garg .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Sharma, A., Garg, A., Thapliyal, A., Rajput, A. (2022). SWOT Analysis of Behavioural Recognition Through Variable Modalities. In: Karuppusamy, P., García Márquez, F.P., Nguyen, T.N. (eds) Ubiquitous Intelligent Systems. ICUIS 2021. Smart Innovation, Systems and Technologies, vol 302. Springer, Singapore. https://doi.org/10.1007/978-981-19-2541-2_25

Download citation

Publish with us

Policies and ethics