AGMS-GCN: Attention-guided multi-scale graph convolutional networks for skeleton-based action recognition
| dc.authorid | 0000-0003-4092-3785 | |
| dc.contributor.author | Kilic, Ugur | |
| dc.contributor.author | Karadag, Ozge Oztimur | |
| dc.contributor.author | Ozyer, Gulsah Tumuklu | |
| dc.date.accessioned | 2026-01-24T12:31:16Z | |
| dc.date.available | 2026-01-24T12:31:16Z | |
| dc.date.issued | 2025 | |
| dc.department | Alanya Alaaddin Keykubat Üniversitesi | |
| dc.description.abstract | Graph Convolutional Networks have the capability to model non-Euclidean data with high effectiveness. Due to this capability, they perform well on standard benchmarks for skeleton-based action recognition (SBAR). Specifically, spatial-temporal graph convolutional networks (ST-GCNs) function effectively in learning spatial- temporal relationships on skeletal graph patterns. In ST-GCN models, a fixed skeletal graph pattern is used across all layers. ST-GCN models obtain spatial-temporal features by performing standard convolution on this fixed graph topology within a local neighborhood limited by the size of the convolution kernel. This convolution kernel dimension can only model dependencies between joints at short distances and shortrange temporal dependencies. However, it fails to model long-range temporal information and long-distance joint dependencies. Effectively capturing these dependencies is key to improving the performance of ST-GCN models. In this study, we propose AGMS-GCN, an attention-guided multi-scale graph convolutional network structure that dynamically determines the weights of the dependencies between joints. In the proposed AGMSGCN architecture, new adjacency matrices that represent action-specific joint relationships are generated by obtaining spatial-temporal dependencies with the attention mechanism on the feature maps extracted using spatial-temporal graph convolutions. This enables the extraction of features that take into account both the shortand long-range spatial-temporal relationship between action-specific joints. This data-driven graph construction method provides amore robust graph representation for capturing subtle differences between different actions. In addition, actions occur through the coordinated movement of multiple body joints. However, most existing SBAR approaches overlook this coordination, considering the skeletal graph from a single-scale perspective. Consequently, these methods miss high-level contextual features necessary for distinguishing actions. The AGMS-GCN architecture addresses this shortcoming with its multi-scale structure. Comprehensive experiments demonstrate that our proposed method attains state-of-the-art (SOTA) performance on the NTU RGB+D 60 and Northwestern-UCLA datasets. It also achieves SOTA competitive performance on the NTU RGB+D 120 dataset. The source code of the proposed AGMS-GCN model is available at: https: //github.com/ugrkilc/AGMS-GCN. | |
| dc.description.sponsorship | Scientific and Technological Research Council of Turkey (TUBITAK) [124E309, 123E635]; Atatrk University under the BAP project [FDK-2023-11957]; TUBITAK; Atatuerk University | |
| dc.description.sponsorship | This study was supported by Scientific and Technological Research Council of Turkey (TUBITAK) under the Grant Numbers 124E309 and 123E635, and by Atatuerk University under the BAP project with code FDK-2023-11957. The authors thank to TUBITAK and Atatuerk University for their support. | |
| dc.identifier.doi | 10.1016/j.knosys.2025.113045 | |
| dc.identifier.issn | 0950-7051 | |
| dc.identifier.issn | 1872-7409 | |
| dc.identifier.scopus | 2-s2.0-85216491021 | |
| dc.identifier.scopusquality | Q1 | |
| dc.identifier.uri | https://doi.org/10.1016/j.knosys.2025.113045 | |
| dc.identifier.uri | https://hdl.handle.net/20.500.12868/5757 | |
| dc.identifier.volume | 311 | |
| dc.identifier.wos | WOS:001434257800001 | |
| dc.identifier.wosquality | Q1 | |
| dc.indekslendigikaynak | Web of Science | |
| dc.indekslendigikaynak | Scopus | |
| dc.language.iso | en | |
| dc.publisher | Elsevier | |
| dc.relation.ispartof | Knowledge-Based Systems | |
| dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | |
| dc.rights | info:eu-repo/semantics/closedAccess | |
| dc.snmz | KA_WoS_20260121 | |
| dc.subject | Action recognition | |
| dc.subject | Skeletal data | |
| dc.subject | Graph convolutional networks | |
| dc.subject | Attention mechanism | |
| dc.subject | Multi-scale | |
| dc.title | AGMS-GCN: Attention-guided multi-scale graph convolutional networks for skeleton-based action recognition | |
| dc.type | Article |












