AeS-GCN: Attention-enhanced semantic-guided graph convolutional networks for skeleton-based action recognition

  • Qing Xu
  • , Feng Liu*
  • , Ziwang Fu
  • , Aimin Zhou
  • , Jiayin Qi*
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

6 Scopus citations

Abstract

Skeleton-based action recognition has been extensively studied in recent years and applied in virtual reality, detection systems and other cases with strong requirements for low cost as well as high accuracy, but most of the existing methods mainly focus on complex architecture of deep neural networks without considering computation efficiency. To balance accuracy and computation cost well, this paper proposes a simple and efficient attention-enhanced semantic-guided graph convolutional network (AeS-GCN) for skeleton-based action recognition. Firstly, we fuse semantics of joint type and frame index and dynamics together as representation of skeleton. Then, we use spatial attention block (SAB) to explore important features in spatial structure, in which adaptive GCN layer is adopted to adaptively model skeleton topology structure. Next, we use temporal attention block (TAB) to extract latent temporal information. The model proposed is a lightweight network and achieves the state-of-the-art performance on mainstream datasets with less parameters and less computational complexity.

Original languageEnglish
Article numbere2070
JournalComputer Animation and Virtual Worlds
Volume33
Issue number3-4
DOIs
StatePublished - 1 Jun 2022

Keywords

  • adaptive graph
  • attention-enhanced
  • computation efficiency
  • computational affection
  • semantics
  • skeleton-based action recognition

Fingerprint

Dive into the research topics of 'AeS-GCN: Attention-enhanced semantic-guided graph convolutional networks for skeleton-based action recognition'. Together they form a unique fingerprint.

Cite this