Next Article in Journal
Robust Scan Registration for Navigation in Forest Environment Using Low-Resolution LiDAR Sensors
Next Article in Special Issue
An Interactive Image Segmentation Method Based on Multi-Level Semantic Fusion
Previous Article in Journal
A New Method for Spatial Estimation of Water Quality Using an Optimal Virtual Sensor Network and In Situ Observations: A Case Study of Chemical Oxygen Demand
Previous Article in Special Issue
An Explainable Student Fatigue Monitoring Module with Joint Facial Representation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition

1
School of Sports Center, Xi’an Jiaotong University, Xi’an 710000, China
2
School of Software Engineering, Xi’an Jiaotong University, Xi’an 710000, China
3
Institute of Artificial Intelligence and Robotics, Xi’an Jiaotong University, Xi’an 710000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(10), 4741; https://doi.org/10.3390/s23104741
Submission received: 13 April 2023 / Revised: 3 May 2023 / Accepted: 10 May 2023 / Published: 14 May 2023
(This article belongs to the Special Issue Machine Learning Based 2D/3D Sensors Data Understanding and Analysis)

Abstract

Fitness yoga is now a popular form of national fitness and sportive physical therapy. At present, Microsoft Kinect, a depth sensor, and other applications are widely used to monitor and guide yoga performance, but they are inconvenient to use and still a little expensive. To solve these problems, we propose spatial–temporal self-attention enhanced graph convolutional networks (STSAE-GCNs) that can analyze RGB yoga video data captured by cameras or smartphones. In the STSAE-GCN, we build a spatial–temporal self-attention module (STSAM), which can effectively enhance the spatial–temporal expression ability of the model and improve the performance of the proposed model. The STSAM has the characteristics of plug-and-play so that it can be applied in other skeleton-based action recognition methods and improve their performance. To prove the effectiveness of the proposed model in recognizing fitness yoga actions, we collected 960 fitness yoga action video clips in 10 action classes and built the dataset Yoga10. The recognition accuracy of the model on Yoga10 achieves 93.83%, outperforming the state-of-the-art methods, which proves that this model can better recognize fitness yoga actions and help students learn fitness yoga independently.
Keywords: fitness yoga; human action recognition; self-attention mechanism fitness yoga; human action recognition; self-attention mechanism

Share and Cite

MDPI and ACS Style

Wei, G.; Zhou, H.; Zhang, L.; Wang, J. Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition. Sensors 2023, 23, 4741. https://doi.org/10.3390/s23104741

AMA Style

Wei G, Zhou H, Zhang L, Wang J. Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition. Sensors. 2023; 23(10):4741. https://doi.org/10.3390/s23104741

Chicago/Turabian Style

Wei, Guixiang, Huijian Zhou, Liping Zhang, and Jianji Wang. 2023. "Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition" Sensors 23, no. 10: 4741. https://doi.org/10.3390/s23104741

APA Style

Wei, G., Zhou, H., Zhang, L., & Wang, J. (2023). Spatial–Temporal Self-Attention Enhanced Graph Convolutional Networks for Fitness Yoga Action Recognition. Sensors, 23(10), 4741. https://doi.org/10.3390/s23104741

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop