Residential College | false |
Status | 已發表Published |
Auto-Learning-GCN: An Ingenious Framework for Skeleton-Based Action Recognition | |
Xin, Wentian1,2,3; Liu, Yi1,2,3; Liu, Ruyi1,2,3; Miao, Qiguang1,2,3; Shi, Cheng4; Pun, Chi Man5 | |
2024 | |
Conference Name | 6th Chinese Conference on Pattern Recognition and Computer Vision (PRCV) |
Source Publication | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
Volume | 14425 LNCS |
Pages | 29-42 |
Conference Date | OCT 13-15, 2023 |
Conference Place | Xiamen Univ, Xiamen |
Country | PEOPLES R CHINA |
Publisher | SPRINGER-VERLAG BERLINHEIDELBERGER PLATZ 3, D-14197 BERLIN, GERMANY |
Abstract | The Graph Convolutional Network (GCN) has garnered substantial interest over an extended period owing to its notable efficacy in addressing topological correlations, with particular achievements observed in skeleton-based action recognition. However, it is intriguing to note that the efficacy of the adaptive module within the existing model gradually diminishes as the learning capacity of the adjacency matrix improves. Inspired by this, AL-GCN constructs a comprehensively learnable group adjacency matrix in both spatial and temporal dimensions, thus forming an elegant and efficient GCN-based model. Firstly, the prevalent adaptive module is courageously removed, and the key points of the human body are astutely leveraged as motion compensation to provide dynamic correlation support. Secondly, a similar temporal adjacency matrix group is designed in the temporal domain to capture the global interaction characteristics more effectively. Subsequently, attention modules are purposefully crafted for both the temporal and spatial domains, which provide discriminative information between classes and improve learnability and interpretability. Finally, the Bayesian weight selection algorithm is utilized to achieve efficient and accurate fusion results for multi-stream data. On the NTU-60, NTU-120, and NW-UCLA, AL-GCN outperforms the state-of-the-art method, with up to 7.0% improvement on the challenging UAV-Human, achieving superior performance in multiple settings, all while maintaining a lower computational cost. Related code will be available on Auto-Learning-GCN. |
Keyword | Attention Mechanism Graph Convolutional Network Model Optimization Skeleton-based Action Recognition |
DOI | 10.1007/978-981-99-8429-9_3 |
URL | View the original |
Indexed By | CPCI-S |
Language | 英語English |
WOS Research Area | Computer Science |
WOS Subject | Computer Science, Artificial Intelligence ; Computer Science, Information Systems ; Computer Science, Theory & Methods |
WOS ID | WOS:001155015700003 |
Scopus ID | 2-s2.0-85180742661 |
Fulltext Access | |
Citation statistics | |
Document Type | Conference paper |
Collection | Faculty of Science and Technology DEPARTMENT OF COMPUTER AND INFORMATION SCIENCE |
Corresponding Author | Miao, Qiguang |
Affiliation | 1.School of Computer Science and Technology, Xidian University, Xi’an, China 2.Xi’an Key Laboratory of Big Data and Intelligent Vision, Xidian University, Xi’an, China 3.Ministry of Education Key Laboratory of Collaborative Intelligence Systems, Xi’an, China 4.School of Computer Science and Engineering, Xi’an University of Technology, Xi’an, China 5.Department of Computer and Information Science, University of Macau, Zhuhai, China |
Recommended Citation GB/T 7714 | Xin, Wentian,Liu, Yi,Liu, Ruyi,et al. Auto-Learning-GCN: An Ingenious Framework for Skeleton-Based Action Recognition[C]:SPRINGER-VERLAG BERLINHEIDELBERGER PLATZ 3, D-14197 BERLIN, GERMANY, 2024, 29-42. |
APA | Xin, Wentian., Liu, Yi., Liu, Ruyi., Miao, Qiguang., Shi, Cheng., & Pun, Chi Man (2024). Auto-Learning-GCN: An Ingenious Framework for Skeleton-Based Action Recognition. Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 14425 LNCS, 29-42. |
Files in This Item: | There are no files associated with this item. |
Items in the repository are protected by copyright, with all rights reserved, unless otherwise indicated.
Edit Comment