single-jc.php

JACIII Vol.28 No.6 pp. 1367-1379
doi: 10.20965/jaciii.2024.p1367
(2024)

Research Paper:

Skeleton-Based Human Action Recognition with Spatial and Temporal Attention-Enhanced Graph Convolution Networks

Fen Xu ORCID Icon, Pengfei Shi, and Xiaoping Zhang

North China University of Technology
No.5 Jinyuanzhuang Road, Shijingshan District, Beijing 100144, China

Received:
April 21, 2024
Accepted:
September 25, 2024
Published:
November 20, 2024
Keywords:
skeleton-based human action recognition, attention-enhanced, graph convolution network
Abstract

Skeleton-based human action recognition has great potential for human behavior analysis owing to its simplicity and robustness in varying environments. This paper presents a spatial and temporal attention-enhanced graph convolution network (STAEGCN) for human action recognition. The spatial-temporal attention module in the network uses convolution embedding for positional information and adopts multi-head self-attention mechanism to extract spatial and temporal attention separately from the input series of the skeleton. The spatial and temporal attention are then concatenated into an entire attention map according to a specific ratio. The proposed spatial and temporal attention module was integrated with an adaptive graph convolution network to form the backbone of STAEGCN. Based on STAEGCN, a two-stream skeleton-based human action recognition model was trained and evaluated. The model performed better on both NTU RGB+D and Kinetics 400 than 2s-AGCN and its variants. It was proven that the strategy of decoupling spatial and temporal attention and combining them in a flexible way helps improve the performance of graph convolution networks in skeleton-based human action recognition.

Spatial and temporal attention-enhanced graph convolution network

Spatial and temporal attention-enhanced graph convolution network

Cite this article as:
F. Xu, P. Shi, and X. Zhang, “Skeleton-Based Human Action Recognition with Spatial and Temporal Attention-Enhanced Graph Convolution Networks,” J. Adv. Comput. Intell. Intell. Inform., Vol.28 No.6, pp. 1367-1379, 2024.
Data files:
References
  1. [1] M. Liu, H. Liu, and C. Chen, “Enhanced skeleton visualization for view invariant human action recognition,” Pattern Recognition, Vol.68, pp. 346-362, 2017. https://doi.org/10.1016/j.patcog.2017.02.030
  2. [2] C. Li, Q. Zhong, D. Xie, and S. Pu, “Co-occurrence feature learning from skeleton data for action recognition and detection with hierarchical aggregation,” arXiv:1804.06055, 2018. https://doi.org/10.48550/arXiv.1804.06055
  3. [3] Q. Ke, M. Bennamoun, S. An, F. Sohel, and F. Boussaid, “A new representation of skeleton sequences for 3D action recognition,” 2017 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 4570-4579, 2017. https://doi.org/10.1109/CVPR.2017.486
  4. [4] P. Zhang et al., “View adaptive recurrent neural networks for high performance human action recognition from skeleton data,” 2017 IEEE Int. Conf. on Computer Vision, pp. 2136-2145, 2017. https://doi.org/10.1109/ICCV.2017.233
  5. [5] S. Yan, Y. Xiong, and D. Lin, “Spatial temporal graph convolutional networks for skeleton-based action recognition,” Proc. of the AAAI Conf. on Artificial Intelligence, Vol.32, No.1, pp. 7444-7452, 2018. https://doi.org/10.1609/aaai.v32i1.12328
  6. [6] M. Li et al., “Actional-structural graph convolutional networks for skeleton-based action recognition,” 2019 IEEE/CVF Conf. on Computer Vision and Pattern Recognition, pp. 3590-3598, 2019. https://doi.org/10.1109/CVPR.2019.00371
  7. [7] L. Shi, Y. Zhang, J. Cheng, and H. Lu, “Two-stream adaptive graph convolutional networks for skeleton-based action recognition,” 2019 IEEE/CVF Conf. on Computer Vision and Pattern Recognition, pp. 12018-12027, 2019. https://doi.org/10.1109/CVPR.2019.01230
  8. [8] C. Si, Y. Jing, W. Wang, L. Wang, and T. Tan, “Skeleton-based action recognition with spatial reasoning and temporal stack learning,” Proc. of the 15th European Conf. on Computer Vision, Part 1, pp. 103-118, 2018. https://doi.org/10.1007/978-3-030-01246-5_7
  9. [9] A. Vaswani et al., “Attention is all you need,” Proc. of the 31st Int. Conf. on Neural Information Processing Systems (NIPS’17), pp. 6000-6010, 2017.
  10. [10] J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” arXiv:1810.04805, 2018. https://doi.org/10.48550/arXiv.1810.04805
  11. [11] A. Radford, K. Narasimhan, T. Salimans, and I. Sutskever, “Improving language understanding by generative pre-training,” OpenAI, 2018.
  12. [12] Z. Yang et al., “XLNet: Generalized autoregressive pretraining for language understanding,” Proc. of the 33rd Int. Conf. on Neural Information Processing Systems, pp. 5753-5763, 2019.
  13. [13] C. Raffel et al., “Exploring the limits of transfer learning with a unified text-to-text transformer,” The J. of Machine Learning Research, Vol.21, No.1, pp. 5485-5551, 2020.
  14. [14] A. F. Bobick and J. W. Davis, “The recognition of human movement using temporal templates,” IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol.23, No.3, pp. 257-267, 2001. https://doi.org/10.1109/34.910878
  15. [15] H. Fujiyoshi, A. J. Lipton, and T. Kanade, “Real-time human motion analysis by image skeletonization,” IEICE Trans. on Information and Systems, Vol.E87-D, No.1, pp. 113-120, 2004.
  16. [16] P. Dollár, V. Rabaud, G. Cottrell, and S. Belongie, “Behavior recognition via sparse spatio-temporal features,” 2005 IEEE Int. Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance, pp. 65-72, 2005. https://doi.org/10.1109/VSPETS.2005.1570899
  17. [17] K. Simonyan and A. Zisserman, “Two-stream convolutional networks for action recognition in videos,” Proc. of the 27th Int. Conf. on Neural Information Processing Systems, Vol.1, pp. 568-576, 2014.
  18. [18] T. Liu et al., “Spatial-temporal interaction learning based two-stream network for action recognition,” Information Sciences, Vol.606, pp. 864-876, 2022. https://doi.org/10.1016/j.ins.2022.05.092
  19. [19] M. Yang, Y. Guo, F. Zhou, and Z. Yang, “TS-D3D: A novel two-stream model for action recognition,” 2022 Int. Conf. on Image Processing, Computer Vision and Machine Learning, pp. 179-182, 2022. https://doi.org/10.1109/ICICML57342.2022.10009839
  20. [20] Z. Wang, H. Lu, J. Jin, and K. Hu, “Human action recognition based on improved two-stream convolution network,” Applied Sciences, Vol.12, No.12, Article No.5784, 2022. https://doi.org/10.3390/app12125784
  21. [21] S. Ji, W. Xu, M. Yang, and K. Yu, “3D convolutional neural networks for human action recognition,” IEEE Trans. on Pattern Analysis and Machine Intelligence, Vol.35, No.1, pp. 221-231, 2013. https://doi.org/10.1109/TPAMI.2012.59
  22. [22] D. Tran, L. Bourdev, R. Fergus, L. Torresani, and M. Paluri, “Learning spatiotemporal features with 3D convolutional networks,” 2015 IEEE Int. Conf. on Computer Vision, pp. 4489-4497, 2015. https://doi.org/10.1109/ICCV.2015.510
  23. [23] X. Mu et al., “DC3D: A video action recognition network based on dense connection,” 2022 10th Int. Conf. on Advanced Cloud and Big Data, pp. 133-138, 2022. https://doi.org/10.1109/CBD58033.2022.00032
  24. [24] U. De Alwis and M. Alioto, “Temporal redundancy-based computation reduction for 3D convolutional neural networks,” 2022 IEEE 4th Int. Conf. on Artificial Intelligence Circuits and Systems, pp. 86-89, 2022. https://doi.org/10.1109/AICAS54282.2022.9869903
  25. [25] Z. Cao, T. Simon, S.-E. Wei, and Y. Sheikh, “Realtime multi-person 2D pose estimation using part affinity fields,” 2017 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 1302-1310, 2017. https://doi.org/10.1109/CVPR.2017.143
  26. [26] C. Li, Q. Zhong, D. Xie, and S. Pu, “Skeleton-based action recognition with convolutional neural networks,” 2017 IEEE Int. Conf. on Multimedia & Expo Workshops, pp. 597-600, 2017. https://doi.org/10.1109/ICMEW.2017.8026285
  27. [27] S. Li, W. Li, C. Cook, C. Zhu, and Y. Gao, “Independently recurrent neural network (IndRNN): Building a longer and deeper RNN,” 2018 IEEE/CVF Conf. on Computer Vision and Pattern Recognition, pp. 5457-5466, 2018. https://doi.org/10.1109/CVPR.2018.00572
  28. [28] W. Peng, X. Hong, H. Chen, and G. Zhao, “Learning graph convolutional network for skeleton-based human action recognition by neural searching,” Proc. of the AAAI Conf. on Artificial Intelligence, Vol.34, No.3, pp. 2669-2676, 2020. https://doi.org/10.1609/aaai.v34i03.5652
  29. [29] A. Dosovitskiy et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv:2010.11929, 2020. https://doi.org/10.48550/arXiv.2010.11929
  30. [30] Z. Liu et al., “Swin transformer: Hierarchical vision transformer using shifted windows,” 2021 IEEE/CVF Int. Conf. on Computer Vision, pp. 9992-10002, 2021. https://doi.org/10.1109/ICCV48922.2021.00986
  31. [31] M. Chen et al., “Generative pretraining from pixels,” Proc. of the 37th Int. Conf. on Machine Learning, pp. 1691-1703, 2020.
  32. [32] Y. Zhang, B. Wu, W. Li, L. Duan, and C. Gan, “STST: Spatial-temporal specialized transformer for skeleton-based action recognition,” Proc. of the 29th ACM Int. Conf. on Multimedia, pp. 3229-3237, 2021. https://doi.org/10.1145/3474085.3475473
  33. [33] C. Plizzari, M. Cannici, and M. Matteucci, “Spatial temporal transformer network for skeleton-based action recognition,” Pattern Recognition: ICPR Int. Workshops and Challenges, Part 3, pp. 694-701, 2021. https://doi.org/10.1007/978-3-030-68796-0_50
  34. [34] J. Yang et al., “Recurring the transformer for video action recognition,” 2022 IEEE/CVF Conf. on Computer Vision and Pattern Recognition, pp. 14043-14053, 2022. https://doi.org/10.1109/CVPR52688.2022.01367
  35. [35] J. Kong, Y. Bian, and M. Jiang, “MTT: Multi-scale temporal transformer for skeleton-based action recognition,” IEEE Signal Processing Letters, Vol.29, pp. 528-532, 2022. https://doi.org/10.1109/LSP.2022.3142675
  36. [36] H. Qiu, B. Hou, B. Ren, and X. Zhang, “Spatio-temporal tuples transformer for skeleton-based action recognition,” arXiv:2201.02849, 2022. https://doi.org/10.48550/arXiv.2201.02849
  37. [37] L. Shi, Y. Zhang, J. Cheng, and H. Lu, “Decoupled spatial-temporal attention network for skeleton-based action-gesture recognition,” Proc. of the 15th Asian Conf. on Computer Vision, Part 5, pp. 38-53, 2021. https://doi.org/10.1007/978-3-030-69541-5_3
  38. [38] C. Wu, X.-J. Wu, and J. Kittler, “Graph2Net: Perceptually-enriched graph learning for skeleton-based action recognition,” IEEE Trans. on Circuits and Systems for Video Technology, Vol.32, No.4, pp. 2120-2132, 2022. https://doi.org/10.1109/TCSVT.2021.3085959
  39. [39] P. Geng, X. Lu, W. Li, and L. Lyu, “Hierarchical aggregated graph neural network for skeleton-based action recognition,” IEEE Trans. on Multimedia, 2024. https://doi.org/10.1109/TMM.2024.3428330
  40. [40] C. Plizzari, M. Cannici, and M. Matteucci, “Skeleton-based action recognition via spatial and temporal transformer networks,” Computer Vision and Image Understanding, Vols.208-209, Article No.103219, 2021. https://doi.org/10.1016/j.cviu.2021.103219
  41. [41] C. Feichtenhofer, H. Fan, J. Malik, and K. He, “SlowFast networks for video recognition,” 2019 IEEE/CVF Int. Conf. on Computer Vision, pp. 6201-6210, 2019. https://doi.org/10.1109/ICCV.2019.00630
  42. [42] H. Wu et al., “CvT: Introducing convolutions to vision transformers,” 2021 IEEE/CVF Int. Conf. on Computer Vision, pp. 22-31, 2021. https://doi.org/10.1109/ICCV48922.2021.00009
  43. [43] W. Kay et al., “The kinetics human action video dataset,” arXiv:1705.06950, 2017. https://doi.org/10.48550/arXiv.1705.06950
  44. [44] A. Shahroudy, J. Liu, T.-T. Ng, and G. Wang, “NTU RGB+D: A large scale dataset for 3D human activity analysis,” 2016 IEEE Conf. on Computer Vision and Pattern Recognition, pp. 1010-1019, 2016. https://doi.org/10.1109/CVPR.2016.115
  45. [45] M. Rahevar, A. Ganatra, T. Saba, A. Rehman, and S. A. Bahaj, “Spatial–temporal dynamic graph attention network for skeleton-based action recognition,” IEEE Access, Vol.11, pp. 21546-21553, 2023. https://doi.org/10.1109/ACCESS.2023.3247820
  46. [46] S. Cho, M. H. Maqbool, F. Liu, and H. Foroosh, “Self-attention network for skeleton-based human action recognition,” 2020 IEEE Winter Conf. on Applications of Computer Vision, pp. 624-633, 2020. https://doi.org/10.1109/WACV45572.2020.9093639
  47. [47] Y. Li, J. Yuan, and H. Liu, “Human skeleton-based action recognition algorithm based on spatiotemporal attention graph convolutional network model,” J. of Computer Applications, Vol.41, No.7, pp. 1915-1921, 2021 (in Chinese). https://doi.org/10.11772/j.issn.1001-9081.2020091515
  48. [48] Q. Yu, Y. Dai, K. Hirota, S. Shao, and W. Dai, “Shuffle graph convolutional network for skeleton-based action recognition,” J. Adv. Comput. Intell. Intell. Inform., Vol.27, No.5, pp. 790-800, 2023. https://doi.org/10.20965/jaciii.2023.p0790

*This site is desgined based on HTML5 and CSS3 for modern browsers, e.g. Chrome, Firefox, Safari, Edge, Opera.

Last updated on Dec. 13, 2024