Discover the SciOpen Platform and Achieve Your Research Goals with Ease.
Search articles, authors, keywords, DOl and etc.
Using skeletal information to model and recognize human actions is currently a hot research subject in the realm of Human Action Recognition (HAR). Graph Convolutional Networks (GCN) have gained popularity in this discipline due to their capacity to efficiently process graph-structured data. However, it is challenging for current models to handle distant dependencies that commonly exist between human skeleton nodes, which hinders the development of algorithms in related fields. To solve these problems, the Lightweight Multiscale Spatio-Temporal Graph Convolutional Network (LMSTGCN) is proposed. Firstly, the Lightweight Multiscale Spatial Graph Convolutional Network (LMSGCN) is constructed to capture the information in various hierarchies, and multiple inner connections between skeleton joints are captured by dividing the input features into a number of subsets along the channel direction. Secondly, the dilated convolution is incorporated into the temporal convolution to construct Lightweight Multiscale Temporal Convolutional Network (LMTCN), which allows to obtain a wider receptive field while keeping the size of the convolution kernel unchanged. Thirdly, the Spatio-Temporal Location Attention (STLAtt) module is used to identify the most informative joints in the sequence of skeletal information at a specific frame, hence improving the model’s ability to extract features and recognize actions. Finally, multi-stream data fusion input structure is used to enhance the input data and expand the feature information. Experiments on three public datasets illustrate the effectiveness of the proposed network.
L. Wang, D. Q. Huynh, and P. Koniusz, A comparative review of recent kinect-based action recognition algorithms, IEEE Trans. Image Process., vol. 29, pp. 15–28, 2020.
Q. Cai, Y. B. Deng, H. S. Li, L. Yu, and S. F. Ming, Survey on human action recognition based on deep learning, (in Chinese), Comput. Sci., vol. 47, no. 4, pp. 85–93, 2020.
H. Qian, J. Yi, and Y. Fu, Review of human action recognition based on deep learning, (in Chinese), J. Front. Comput. Sci. Technol., vol. 15, no. 3, pp. 438–455, 2021.
G. V. Reddy, K. Deepika, L. Malliga, D. Hemanand, C. Senthilkumar, S. Gopalakrishnan, and Y. Farhaoui, Human action recognition using difference of Gaussian and difference of wavelet, Big Data Mining and Analytics, vol. 6, no. 3, pp. 336–346, 2023.
P. Zhang, C. Lan, J. Xing, W. Zeng, J. Xue, and N. Zheng, View adaptive neural networks for high performance skeleton-based human action recognition, IEEE Trans. Pattern Anal. Mach. Intell., vol. 41, no. 8, pp. 1963–1978, 2019.
A. M. U. Din and S. Qureshi, Limits of depth: Over-smoothing and over-squashing in GNNs, Big Data Mining and Analytics, vol. 7, no. 1, pp. 205–216, 2024.
J. Liu, G. Wang, L. Y. Duan, K. Abdiyeva, and A. C. Kot, Skeleton-based human action recognition with global context-aware attention LSTM networks, IEEE Trans. Image Process., vol. 27, no. 4, pp. 1586–1599, 2018.
H. Xia and X. Gao, Multi-scale mixed dense graph convolution network for skeleton-based action recognition, IEEE Access, vol. 9, pp. 36475–36484, 2021.
S. H. Gao, M. M. Cheng, K. Zhao, X. Y. Zhang, M. H. Yang, and P. Torr, Res2Net: A new multi-scale backbone architecture, IEEE Trans. Pattern Anal. Mach. Intell., vol. 43, no. 2, pp. 652–662, 2021.
Y. Li, K. Li, C. Chen, X. Zhou, Z. Zeng, and K. Li, Modeling temporal patterns with dilated convolutions for time-series forecasting, ACM Trans. Knowledge Discov. Data, vol. 16, no. 1, p. 14, 2022.
J. Liu, A. Shahroudy, M. Perez, G. Wang, L. Y. Duan, and A. C. Kot, NTU RGB+D 120: A large-scale benchmark for 3D human activity understanding, IEEE Trans. Pattern Anal. Mach. Intell., vol. 42, no. 10, pp. 2684–2701, 2020.
Y. F. Song, Z. Zhang, C. Shan, and L. Wang, Constructing stronger and faster baselines for skeleton-based action recognition, IEEE Trans. Pattern Anal. Mach. Intell., vol. 45, no. 2, pp. 1474–1488, 2023.
K. Cheng, Y. Zhang, X. He, J. Cheng, and H. Lu, Extremely lightweight skeleton-based action recognition with ShiftGCN++, IEEE Trans. Image Process., vol. 30, pp. 7333–7348, 2021.
Z. Tu, J. Zhang, H. Li, Y. Chen, and J. Yuan, Joint-bone fusion graph convolutional network for semi-supervised skeleton action recognition, IEEE Trans. Multimedia, vol. 25, pp. 1819–1831, 2023.
M. Li, S. Chen, X. Chen, Y. Zhang, Y. Wang, and Q. Tian, Symbiotic graph neural networks for 3D skeleton-based human action recognition and motion prediction, IEEE Trans. Pattern Anal. Mach. Intell., vol. 44, no. 6, pp. 3316–3333, 2022.
347
Views
52
Downloads
1
Crossref
1
Web of Science
1
Scopus
0
CSCD
Altmetrics
The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).