Abstract
Gait emotion recognition (GER) plays a crucial role in identifying human emotions. Most previous methods apply Spatial-Temporal Graph Convolutional Networks (ST-GCN) to recognize emotions. However, these methods suffer from two serious problems: (1) they ignore the fact that the similarity between emotions with the similar emotional intensity. Consequently, fine-grained information from the low-layer network, which is essential for accurate emotion recognition, is lost. (2) They ignore that the expression of emotion is a continuous process, that is, failing to model the temporal dimension effectively. To address these issues, a novel Pyramid Hybrid Graph Convolutional Network (Pyr-HGCN) is proposed for GER. Specifically, we first introduce and enhance the pyramid structure in GER to compensate for the missing fine-grained information of the ST-GCN structure. Additionally, we design a novel Spatial-Temporal Hybrid Convolution (STHC) block, which can indirectly and simultaneously capture complex spatio-temporal correlations in long-term regions. Extensive experiments and visualizations were performed on several benchmarks, with an accuracy improvement of 0.01 to 0.02 demonstrating the effectiveness of our approach against state-of-the-art competitors.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Arunnehru, J., KalaiselviĀ Geetha, M.: Automatic human emotion recognition in surveillance video. In: Intelligent techniques in signal processing for multimedia security, pp. 321ā342. Springer (2017)
Bhatia, Y., Bari, A.H., Gavrilova, M.: A lstm-based approach for gait emotion recognition. In: 2021 IEEE 20th International Conference on Cognitive Informatics & Cognitive Computing (ICCI* CC). pp. 214ā221. IEEE (2021)
Bhattacharya, U., Mittal, T., Chandra, R., Randhavane, T., Bera, A., Manocha, D.: Step: Spatial temporal graph convolutional networks for emotion perception from gaits. In: Proceedings of the AAAI Conference on Artificial Intelligence. vol.Ā 34, pp. 1342ā1350 (2020)
Bhattacharya, U., Roncal, C., Mittal, T., Chandra, R., Kapsaskis, K., Gray, K., Bera, A., Manocha, D.: Take an emotion walk: Perceiving emotions from gaits using hierarchical attention pooling and affective mapping. In: European Conference on Computer Vision. pp. 145ā163. Springer (2020)
Chen, Y., Zhang, Z., Yuan, C., Li, B., Deng, Y., Hu, W.: Channel-wise topology refinement graph convolution for skeleton-based action recognition. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. pp. 13359ā13368 (2021)
Daoudi, M., Berretti, S., Pala, P., Delevoye, Y., Bimbo, A.D.: Emotion recognition by body movement representation on the manifold of symmetric positive definite matrices. In: International Conference on Image Analysis and Processing. pp. 550ā560. Springer (2017)
Destephe, M., Henning, A., Zecca, M., Hashimoto, K., Takanishi, A.: Perception of emotion and emotional intensity in humanoid robots gait. In: 2013 IEEE International Conference on Robotics and Biomimetics (ROBIO). pp. 1276ā1281. IEEE (2013)
Gavrilova, M.L., Ahmed, F., Bari, A.H., Liu, R., Liu, T., Maret, Y., Sieu, B.K., Sudhakar, T.: Multi-modal motion-capture-based biometric systems for emergency response and patient rehabilitation. In: Research Anthology on Rehabilitation Practices and Therapy, pp. 653ā678. IGI global (2021)
Han, X., Gao, Y., Lu, Z., Zhang, Z., Niu, D.: Research on moving object detection algorithm based on improved three frame difference method and optical flow. In: 2015 Fifth International Conference on Instrumentation and Measurement, Computer, Communication and Control (IMCCC). pp. 580ā584. IEEE (2015)
Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 7132ā7141 (2018)
Jin, L., Wang, X., Nie, X., Wang, W., Guo, Y., Yan, S., Zhao, J.: Rethinking the person localization for single-stage multi-person pose estimation. IEEE Transactions on Multimedia (2023)
Li, B., Zhu, C., Li, S., Zhu, T.: Identifying emotions from non-contact gaits information based on microsoft kinects. IEEE Trans. Affect. Comput. 9(4), 585ā591 (2016)
Li, F., Zhu, A., Liu, Z., Huo, Y., Xu, Y., Hua, G.: Pyramidal graph convolutional network for skeleton-based human action recognition. IEEE Sens. J. 21(14), 16183ā16191 (2021)
Li, Q., Han, Z., Wu, X.M.: Deeper insights into graph convolutional networks for semi-supervised learning. In: Proceedings of the AAAI conference on artificial intelligence. vol.Ā 32 (2018)
Lin, T.Y., DollĆ”r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp. 2117ā2125 (2017)
Liu, Z., Zhang, H., Chen, Z., Wang, Z., Ouyang, W.: Disentangling and unifying graph convolutions for skeleton-based action recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 143ā152 (2020)
Martinez, J., Hossain, R., Romero, J., Little, J.J.: A simple yet effective baseline for 3d human pose estimation. In: Proceedings of the IEEE international conference on computer vision. pp. 2640ā2649 (2017)
Narayanan, V., Manoghar, B.M., Dorbala, V.S., Manocha, D., Bera, A.: Proxemo: Gait-based emotion learning and multi-view proxemic fusion for socially-aware robot navigation. In: 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). pp. 8200ā8207. IEEE (2020)
Randhavane, T., Bhattacharya, U., Kapsaskis, K., Gray, K., Bera, A., Manocha, D.: Learning perceived emotion using affective and deep features for mental health applications. In: 2019 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct). pp. 395ā399. IEEE (2019)
Sheng, W., Li, X.: Multi-task learning for gait-based identity recognition and emotion recognition using attention enhanced temporal graph convolutional network. Pattern Recogn. 114, 107868 (2021)
Shi, L., Zhang, Y., Cheng, J., Lu, H.: Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. pp. 12026ā12035 (2019)
Smith, F.W., Smith, M.L.: Decoding the dynamic representation of facial expressions of emotion in explicit and incidental tasks. Neuroimage 195, 261ā271 (2019)
Teepe, T., Gilg, J., Herzog, F., Hƶrmann, S., Rigoll, G.: Towards a deeper understanding of skeleton-based gait recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. pp. 1569ā1577 (2022)
Teepe, T., Khan, A., Gilg, J., Herzog, F., Hƶrmann, S., Rigoll, G.: Gaitgraph: Graph convolutional network for skeleton-based gait recognition. In: 2021 IEEE International Conference on Image Processing (ICIP). pp. 2314ā2318. IEEE (2021)
Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., Hu, Q.: Eca-net: Efficient channel attention for deep convolutional neural networks. In: Conference on Computer Vision and Pattern Recognition (CVPR). pp. 11531ā11539 (2020)
Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: Cbam: Convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV). pp. 3ā19 (2018)
Xia, R., Ding, Z.: Emotion-cause pair extraction: A new task to emotion analysis in texts. arXiv preprint arXiv:1906.01267 (2019)
Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. In: Thirty-second AAAI conference on artificial intelligence (2018)
Yin, Y., Jing, L., Huang, F., Yang, G., Wang, Z.: Msa-gcn: Multiscale adaptive graph convolution network for gait emotion recognition. arXiv preprint arXiv:2209.08988 (2022)
Zhuang, Y., Lin, L., Tong, R., Liu, J., Iwamot, Y., Chen, Y.W.: G-gcsn: Global graph convolution shrinkage network for emotion perception from gait. In: Proceedings of the Asian Conference on Computer Vision (2020)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
Ā© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Jing, L., Yang, G., Yin, Y. (2024). Pyr-HGCN: Pyramid Hybrid Graph Convolutional Network forĀ Gait Emotion Recognition. In: Liu, Q., et al. Pattern Recognition and Computer Vision. PRCV 2023. Lecture Notes in Computer Science, vol 14429. Springer, Singapore. https://doi.org/10.1007/978-981-99-8469-5_15
Download citation
DOI: https://doi.org/10.1007/978-981-99-8469-5_15
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-8468-8
Online ISBN: 978-981-99-8469-5
eBook Packages: Computer ScienceComputer Science (R0)