Skip to main content

Asymmetric Convolution View Adaptation Networks for Skeleton-Based Human Action Recognition

  • Conference paper
  • First Online:
Advances in Computational Intelligence Systems (UKCI 2021)

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 1409))

Included in the following conference series:

Abstract

Benefiting from the extensive use of depth sensors, human action recognition plays an important role in a great number of works, including human-computer interaction, security monitoring, motion-sensing games, medical care. In most recent works, depth cameras have been used to capture required action data for action identification. Especially, feature extraction based on skeleton information has made satisfactory achievements in action recognition and has been gradually extended to various algorithms. However, the research of view invariance is not deep enough. To improve the performance of skeleton recognition, the paper proposes an improved asymmetric convolution adaptive network, achieving desirable results on the public benchmark dataset of NTU RGB + D 60. The model combines an advanced adaptive module and asymmetric convolution blocks, which can effectively extract features from the original skeleton. Ablation studies and comparative experiments indicate that the performance of the model is better than many of the most state-of-the-art algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Pham, D.-T., Nguyen, T.-N., Le, T.-L.H., Vu.: Spatio-temporal representation for skeleton-based human action recognition. In: 2020 International Conference on Multimedia Analysis and Pattern Recognition (MAPR), October 2020

    Google Scholar 

  2. Shi, L., Zhang, Y., Cheng, J., Lu, H.: Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019

    Google Scholar 

  3. Shi, L., Zhang, Y., Cheng, J., Lu, H.: Skeleton-based action recognition with directed graph neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7912–7921 (2019)

    Google Scholar 

  4. Li, C., Zhong, Q., Xie, D., Pu, S.: Cooccurrence feature learning from skeleton data for action recognition and detection with hierarchical aggregation. In: Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence, July 2018

    Google Scholar 

  5. Li, W., Zhang, Z., Liu, Z.: Action recognition based on a bag of 3D points. In: CVPR, pp. 9–14 (2010)

    Google Scholar 

  6. Yang, X., Tian, Y.: Eigen joints-based action recognition using Naive-Bayes-nearest-neighbor. In: CVPR, pp. 14–19 (2012)

    Google Scholar 

  7. Oreifej, O., Liu, Z.: HON4D: histogram of oriented 4D normals for activity recognition from depth sequences. In: CVPR, pp. 716–723 (2013)

    Google Scholar 

  8. Rahmani, H., Mian, A.: 3D action recognition from novel viewpoints. In: CVPR, p. 12 (2016)

    Google Scholar 

  9. Wang, P., Li, W., Gao, Z., Zhang, J., Tang, C., Ogunbona, P.: Deep convolutional neural networks for action recognition using depth map sequences. In: CVPR, pp. 1–8 (2015)

    Google Scholar 

  10. Shotton, J., et al.: Real-time human pose recognition in parts from single depth images. In: CVPR, pp. 1297–1304 (2011)

    Google Scholar 

  11. Pham, D.-T., Nguyen, T.-N., Le, T.-L., Vu, H.: Analyzing role of joint subset selection in human action recognition. In: 2019 6th NAFOSTED Conference on Information and Computer Science (NICS), December 2019

    Google Scholar 

  12. Yu, J., Gao, H., Yang, W., Chin, W., Kubota, N., Ju, Z.: A discriminative deep model with feature fusion and temporal attention for human action recognition. IEEE Access 8, 43243–43255 (2020)

    Article  Google Scholar 

  13. Cheng, K., Zhang, Y., Cao, C., Shi, L., Cheng, J., Lu, H.: Decoupling GCN with DropGraph module for skeleton-based action recognition. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12369, pp. 536–553. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58586-0_32

    Chapter  Google Scholar 

  14. Ding, X., Guo, Y., Ding, G., Han, J., ACNet: Strengthening the kernel skeletons for powerful CNN via asymmetric convolution blocks. IEEE Int. Conf. Comput. Vis. (ICCV) (2019)

    Google Scholar 

  15. Zhang, P., Lan, C., Xing, J., Zeng, W., Xue, J., Zheng, N.: View adaptive neural networks for high performance skeleton-based human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 41(8), 1963–1978 (2019)

    Article  Google Scholar 

  16. Shahroudy, A., Liu, J., Ng, T.-T., Wang, G.: NTU RGB+ D: a large scale dataset for 3D human activity analysis. In: CVPR (2016)

    Google Scholar 

  17. Zhang, P., Lan, C., Xing, J., Zeng, W., Xue, J., Zheng, N.: View adaptive recurrent neural networks for high performance human action recognition from skeleton data. IEEE Int. Conf. Comput. Vis. 2117–2126 (2017)

    Google Scholar 

  18. Si, C., Jing, Y., Wang, W., Wang, L., Tan, T.: Skeleton-based action recognition with spatial reasoning and temporal stack learning. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11205, pp. 106–121. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01246-5_7

    Chapter  Google Scholar 

  19. Li, T., Fan, L., Zhao, M., Liu, Y, Katabi, D.: Making the invisible visible: action recognition through walls and occlusions. IEEE Int. Conf. Comput. Vis. 872–881 (2019)

    Google Scholar 

  20. Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. AAAI Conf. Artif. Intell. (2018)

    Google Scholar 

  21. Liu, H., Zhang, L., Guan, L., Liu, M.: GFNet: a lightweight group frame network for efficient human action recognition. IEEE Int. Conf. Acoust. Speech Sig. Process. 2583–2587 (2020)

    Google Scholar 

  22. Fan, Y., Weng, S., Zhang, Y., Shi, B., Zhang, Y.: Context-aware cross-attention for skeleton-based human action recognition. IEEE Trans. Image Process. Database: Compendex, 15280–15290 (2020)

    Google Scholar 

  23. Li, S., Yi, J., Farha, Y.A., Gall, J.: Pose refinement graph convolutional network for skeleton-based action recognition. IEEE Robot. Autom. Lett. 6(2), 1028–1035 (2021)

    Article  Google Scholar 

  24. Cheng, K., Zhang, Y., He, X., Chen, W., Cheng, J., Lu, H.: Skeleton-based action recognition with shift graph convolutional network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2020)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tianyu Ma .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ma, T., Yu, J., Gao, H., Ju, Z. (2022). Asymmetric Convolution View Adaptation Networks for Skeleton-Based Human Action Recognition. In: Jansen, T., Jensen, R., Mac Parthaláin, N., Lin, CM. (eds) Advances in Computational Intelligence Systems. UKCI 2021. Advances in Intelligent Systems and Computing, vol 1409. Springer, Cham. https://doi.org/10.1007/978-3-030-87094-2_17

Download citation

Publish with us

Policies and ethics