Next Article in Journal
Three-Dimensional Multi-Agent Foraging Strategy Based on Local Interaction
Previous Article in Journal
An Enhanced Spatial Smoothing Technique of Coherent DOA Estimation with Moving Coprime Array
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Viewpoint-Agnostic Taekwondo Action Recognition Using Synthesized Two-Dimensional Skeletal Datasets

1
Division of Mechanical and Aerospace Engineering, Konkuk University, 120 Neungdong-ro, Gwangjin-gu, Seoul 05029, Republic of Korea
2
Physical Activity and Performance Institute, Konkuk University, 120 Neungdong-ro, Gwangjin-gu, Seoul 05029, Republic of Korea
3
Department of Sports Medicine and Science, Graduate School, Konkuk University, 120 Neungdong-ro, Gwangjin-gu, Seoul 05029, Republic of Korea
4
Department of Physical Education, Konkuk University, 120 Neungdong-ro, Gwangjin-gu, Seoul 05029, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(19), 8049; https://doi.org/10.3390/s23198049
Submission received: 18 August 2023 / Revised: 18 September 2023 / Accepted: 21 September 2023 / Published: 23 September 2023
(This article belongs to the Section Sensor Networks)

Abstract

:
Issues of fairness and consistency in Taekwondo poomsae evaluation have often occurred due to the lack of an objective evaluation method. This study proposes a three-dimensional (3D) convolutional neural network–based action recognition model for an objective evaluation of Taekwondo poomsae. The model exhibits robust recognition performance regardless of variations in the viewpoints by reducing the discrepancy between the training and test images. It uses 3D skeletons of poomsae unit actions collected using a full-body motion-capture suit to generate synthesized two-dimensional (2D) skeletons from desired viewpoints. The 2D skeletons obtained from diverse viewpoints form the training dataset, on which the model is trained to ensure consistent recognition performance regardless of the viewpoint. The performance of the model was evaluated against various test datasets, including projected 2D skeletons and RGB images captured from diverse viewpoints. Comparison of the performance of the proposed model with those of previously reported action recognition models demonstrated the superiority of the proposed model, underscoring its effectiveness in recognizing and classifying Taekwondo poomsae actions.

1. Introduction

Taekwondo is a traditional Korean martial art that has become one of the most popular sports worldwide. Two types of Taekwondo competitions are conducted: gyeorugi and poomsae, which involve various movements and complex techniques. Gyeorugi requires two competing players, and objective judgments are made using a quantitative and accurate electronic scoring system. In poomsae, a single player demonstrates basic attack and defense techniques in a specific order. In this case, evaluation is subjective and qualitative, based on the opinions of the judges, except for penalties (e.g., stopping or crossing the boundaries). Owing to situational constraints, judges must evaluate multiple participants simultaneously, which may give rise to concerns of unfairness and inconsistencies in evaluations not only in competitions but also in promotional tests. To address these issues, quantitative evaluation methods using vision-based action recognition techniques have been proposed [1,2].
Vision-based human action recognition (HAR) has emerged as a prominent area of interest in computer vision and artificial intelligence. Its primary objective is to detect and analyze human actions from unknown video sequences, thereby enabling a deeper understanding and interpretation of such actions. HAR has been applied in various domains, including security [3,4,5], healthcare [6,7,8], and sports [9,10,11]. Vision-based HAR systems have been employed to support quantitative evaluation and judgment in various sports [12,13,14,15,16,17,18]. However, few studies have reported the application of action recognition technology in martial arts disciplines, such as Taekwondo [1,2]. In previous studies, action recognition approaches using RGB (color) and RGB-D (color and depth) images have been proposed. These included methods that emphasized the dominant poses associated with each action in RGB-D videos as input to a convolutional neural network (CNN) [19], as well as techniques that enhanced the structural information of body parts, joints, and temporal scales by representing sequences of depth maps as structured dynamic images [20]. However, the rapid-action characteristics of martial arts pose challenges to motion capture because of insufficient sharpness and intermittent frame loss. The dynamic nature of martial arts and wide range of actions possible therein render the RGB-D methods inadequate. Furthermore, these approaches are susceptible to domain shifts caused by environmental changes and cannot accurately predict dynamic actions.
Recent research on action recognition has incorporated skeletal data into complex human action recognition [21,22,23,24,25,26,27,28,29,30]. For instance, Du et al. proposed an architecture that divided the human skeleton into five parts and fed them into separate subnetworks instead of using recurrent neural networks to process the entire skeleton as input [28]. Yan et al. introduced an action recognition method based on graph convolutional networks (GCNs) considering the spatiotemporal features of skeletons [19]. Subsequently, GCN-related studies [23,25,26,27,29], including one by Duan et al., generated heatmaps using skeletons to address the limitations of the GCN methods, such as the accuracy of skeleton coordinates and integration with other modality data [22]. In skeleton-based action recognition, the skeleton representation provides core information that is highly relevant to human behavior. Unlike the RGB-D models, it remains robust against variations in illumination, changes in clothing, and environmental factors.
Previous approaches used for action recognition primarily relied on images obtained from a single viewpoint. However, in the context of poomsae evaluation, the same movement may appear different when captured from different viewpoints, thereby posing challenges for accurate recognition. Furthermore, single-view action recognition requires training models specific to each viewpoint, thereby necessitating retraining efforts while dealing with images captured from other viewpoints. This results in potential time and resource constraints. Moreover, single-view action recognition predominantly focuses on discerning individual movements and presents difficulties in recognizing complex movements involving multiple actions.
In this study, we propose a novel action recognition model for the evaluation of poomsae that exhibits robust recognition performance regardless of the variations in viewpoints. The model uses three-dimensional (3D) skeletons collected using a full-body motion-capture suit to create two-dimensional (2D) skeletons from a desired viewpoint. Thus, the proposed approach obtains 2D skeletal data from diverse viewpoints as part of the training data and effectively addresses the effect of observational viewpoints, ensuring consistent and reliable performance in action recognition regardless of the viewpoint. The main contributions of this study are as follows:
  • A 3D skeletal dataset comprising 16 unit actions in Taekwondo poomsae was constructed using motion data collected by employing full-body motion-capture suits.
  • Methods were proposed for generating 2D skeletons by projecting 3D skeletons from diverse viewpoints, which in turn were used to generate synthetic joint and bone heatmaps. These incorporated viewpoint-dependent action characteristics into the training dataset. This ensured consistent and reliable performance, regardless of the viewpoint.
  • The optimal camera viewpoint for action recognition of Taekwondo poomsae was determined by analyzing and evaluating the recognition performance.

2. Materials and Methods

2.1. Data Collection

Primary motion data were collected from Taekwondo experts using a full-body motion-capture suit (Xsens MVN; Xsens Corp., Enschede, The Netherlands). The suit was equipped with 17 inertial measurement unit (IMU) sensors, where each sensor measured the acceleration, angular velocity, and orientation of the body segment at the attached point along three mutually perpendicular axes. Subsequently, the raw data obtained from the motion-capture suit were processed to extract the positions of 23 joints in the human skeleton [31]. To enhance the generalizability of the action recognition model, the skeleton with 23 joints was converted to one with 16 joints, as illustrated in Figure 1. Forty Taekwondo experts participated in the data collection. The data-collection procedure involved in this research was approved by the Konkuk University Institutional Review Board (IRB) under protocol number 7001355-202004-HR-372. The 3D skeleton data gathered using the motion-capture suit did not contain any personal privacy information. Furthermore, informed consent was obtained from each participant prior to the data collection, and we made a commitment to use the data exclusively for academic research purposes. Each subject was instructed to sequentially perform the 16 unit actions of Taekwondo poomsae as a predefined data-collection protocol while wearing the motion-capture suit. Each subject executed each action 12 times and repeated the process twice, thereby generating 12 sets of executions for each unit action. Consequently, a Taekwondo unit-action dataset comprising 7680 unit-action data points (16 unit actions per subject × 12 repetitions × 40 participants) was prepared; finally, a motion database, named the Taekwondo unit action dataset of 3D skeletons (TUAD-3D), was constructed, as depicted in Figure 2.

2.1.1. 3D CNN-Based Viewpoint-Agnostic Action Recognition

The viewpoint-agnostic action recognition proposed in this study adopted a previously reported posec3d framework, which utilized a sequence of 2D skeleton heatmaps as input to a 3D CNN as the primary action recognition architecture [22]. To address the performance degradation caused by viewpoint mismatch in training and test images, this study proposed a method for using diverse-viewpoint 2D skeletons generated through the projection of 3D skeletons as the training dataset. Finally, the 2D skeletons were converted into synthetic heatmap images and used to train the action recognition network. Figure 3 illustrates the action recognition architecture proposed in this study.

2.1.2. Generation of Diverse-Viewpoint 2D Skeletons from 3D Skeleton

Figure 4 illustrates the projection of the 3D skeleton onto the image planes of various camera viewpoints to generate 2D skeletons with diverse viewpoints. In this procedure, we assumed that the camera could be rotated along a fixed orbit around the center of the 3D skeleton, as depicted in Figure 4a. The position of the camera was calculated by multiplying the rotation matrix R z θ with its initial position p 0 , as follows:
p θ = R z , θ p 0 ,
where p 0 and p θ denote the initial and rotated camera positions, respectively. To incorporate various perspectives, the joint positions of the 3D skeleton were rotated in intervals of 10°, 45°, and 90°. This rotation facilitated the projection of the 3D skeleton keypoints onto a 2D image plane, thereby transforming the 3D skeleton information into corresponding 2D skeleton information. Although the process of rotation resulted in a partial loss of position and orientation information of the rotated skeleton, it effectively enabled the representation of 2D skeleton information from diverse viewpoints. The 2D skeleton at the rotated camera position p θ was projected by multiplying the projection matrix P θ with the 3D skeleton coordinates:
s i , θ 2 D = P θ s i 3 D ,
where s i 3 D denotes the ith joint position of the 3D skeleton, \ and s i , θ 2 D is the joint position of the 2D skeleton corresponding to the camera position p θ . The projection matrix P θ can be acquired using intrinsic and extrinsic camera parameters. An intrinsic parameter characterizes the optical properties of the camera, while an extrinsic parameter matrix describes its position and orientation.

2.1.3. Generation of Synthetic Heatmap Image from 2D Skeleton

The joint positions of the 2D skeletons were employed to generate synthetic 2D heatmap images. The value assigned to each pixel coordinate within the heatmap image was determined by applying a Gaussian kernel to that coordinate. The heatmaps generated were categorized into bone and joint heatmaps. To reduce the volume of the 3D heatmaps, we implemented two techniques. The first technique involved subject-centered cropping, which involved cropping all the frames based on a minimum bounding box that enclosed the subject in a 2D pose. The frames were cropped as the participant moved within a confined area. Subsequently, the cropped frames were resized to the desired target size. The second technique involved uniform sampling, which entailed selecting a subset of frames to capture the temporal dimension. This sampling approach ensured that the frames were uniformly distributed throughout the sequence, thereby effectively reducing the computational load.
The bone heatmap serves as a visual representation of skeletal connectivity and depicts the interconnections among different segments of the skeleton, thereby facilitating the comprehension of its structural arrangement and tracking of joint movements. Conversely, the joint heatmap focuses on representing the central point of each skeletal segment. This enables the precise localization of joint positions and provides a more detailed understanding of the skeletal shape, which is utilized for motion analysis. The training and validation procedures were conducted separately for the two types of heatmaps to ensure their individual accuracies. The pixel value of the joint heatmap J i , j was calculated as follows:
J i , j = k = 1 N b J o i n t exp D i , j , u k 2 σ 2 ,
where σ is the variance of the Gaussian map and D i ,   j , u k represents the Euclidean distance between pixels i , j and the kth joint position of the 2D skeleton ( u k ). The pixel value of the bone heatmap B i ,   j was calculated as follows:
B i , j = k = 1 N b B o n e exp D i , j , b k 2 σ 2 ,
where D i , j , b k denotes the shortest distance between pixel i , j and the kth bone segment b k , which is defined by the two joint positions of the 2D skeleton. After the above-mentioned process, 2D joint and bone heatmaps were generated for each 2D skeleton. This process resulted in a 3D heatmap with dimensions of T × H × W for each action sequence, where T is the number of frames in each action and H and W represent the height and width of the heatmap image, respectively.

2.1.4. 3D CNN Architecture

The SlowFast architecture was employed to construct a 3D CNN action-classification model [32]. It comprises two distinct pathways, namely slow and fast, as illustrated in Figure 5. The slow pathway is designed to effectively retain spatial information, whereas the fast pathway preserves temporal information. By combining these two pathways, the SlowFast architecture possesses an enhanced capability in capturing both spatial and temporal features, resulting in improved accuracy for action-classification tasks.

2.1.5. Training Procedure

Three-dimensional skeletal databases of the 16 unit actions of poomsae, each performed 12 times and collected from 40 Taekwondo experts, were used to train the proposed action-classification model. Specifically, the action data of 30 experts were allocated for model training, while the remaining 10 experts were reserved for model testing. As illustrated in Figure 6, the 3D skeletal data were subjected to processing to derive a 2D skeleton representation, resulting in an expanded dataset size contingent on the number of projected viewpoints. The number of 2D skeletons used in the model training was 30 × 16 × 12 × m × n , where m and n denote the number of frames in one action and the number of viewpoint projections, respectively. During the training phase, we assessed the generalization performance via a 5-fold cross-validation. The model was trained using the stochastic gradient descent optimizer with a maximum of 240 epochs, and cross-entropy loss was employed as the chosen loss function.

2.2. Evaluation Metrics

The evaluation metrics used in the experiment were F1-score, precision, recall, and accuracy, given as follows:
p r e c i s i o n = T P T P + F P ,
r e c a l l = T P T P + F N ,
a c c u r a c y = T P + T N T P + F P + T N + F N ,
where TP (true positive) represents samples that are predicted as positive, and the ground truth also labels them as positive; FP (false positive) represents samples that are predicted as positive, but the ground truth labels them as negative; TN (true negative) represents samples that are predicted as negative, and the ground truth also labels them as negative; and FN (false negative) represents samples that are predicted as negative, but the ground truth labels them as positive. The F1-score is a metric that balances precision and recall and measures the performance of the model. A higher F1-score indicates better performance.
F 1 S c o r e = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l .

3. Results

The action recognition model was trained on four 2D skeletal datasets. Table 1 lists the configurations of the training datasets. Each training dataset comprised 2D skeletons generated by projecting a 3D skeleton at several viewing angles. The models trained using these four datasets were denoted as Models A–D, as listed in Table 1. The performances of these models were compared with one another to deduce the optimal configuration of the projection viewpoints of the 2D skeletons for the highest recognition performance.

3.1. Performance Evaluation Using Synthetic 2D Skeleton Datasets

The performance of the model was assessed using the synthesized 2D skeletal datasets mentioned earlier. The test samples for the 2D skeletal data were generated by projecting the 3D skeletons of ten individuals selected from TUHA-3D at 10° intervals across the viewpoints. Table 2 and Table 3 present the evaluation results of the joint and bone heatmaps, respectively. While training and testing with the joint heatmap, the highest performance was observed for Model D, with an accuracy of 0.9802. Similarly, while training and testing the bone heatmap, the highest performance was observed for Model D, with an accuracy of 0.9783. The performance comparison results show that the recognition accuracy increased as more 2D skeletons, projected at distinct viewing angles, were included in the training dataset.

3.2. Performance Evaluation Using 2D Skeletons Extracted from Front- and Side-View RGB Images

The performance of the proposed model was evaluated using 2D skeleton data extracted from RGB images. The test samples of the 2D skeleton data were extracted from the poomsae unit-action images captured using an RGB-D camera (Realsense 435d; Intel Corporation, Silicon Valley, CA, USA). In the data-collection procedure, two additional RGB-D cameras were installed, one in the front and one on the left-hand side of the participants, to collect test sample images. Figure 7 depicts the RGB images captured by the frontal and lateral cameras. Overall, 5527 test samples of 2D skeletons were generated from the RGB images using the HRnet pose-estimation algorithm. Table 4 and Table 5 present the evaluation results using the joint and bone heatmaps, respectively. While training and testing with the joint heatmap, the highest performance was observed for Model D, with an accuracy of 0.8705. Similarly, while training and testing the bone heatmap, the highest performance was observed for Model C, with an accuracy of 0.8761. The performance comparison results demonstrate the effectiveness of the action recognition model trained with synthetic 2D skeletons for the RGB test samples.

3.3. Performance Evaluation Using 2D Skeletons Extracted from Random-View RGB Images

Next, the performance of the model was evaluated using a 2D skeletal dataset extracted from the RGB images captured from random viewpoints. The RGB image dataset, obtained using four smartphone cameras from four distinct viewpoints, was used for the assessment. The dataset comprised 639 samples of poomsae unit actions. Figure 8 illustrates the RGB image samples used for the performance evaluation. Table 6 and Table 7 present the evaluation results using the joint and bone heatmaps, respectively. While training and testing with the joint heatmap, the highest performance was observed for Model C, with an accuracy of 0.9381. Similarly, when training and testing the bone heatmap, the highest performance was observed for Model D, with an accuracy of 0.8670. A performance comparison shows that the action recognition model trained with synthetic 2D skeletons could work on the test samples obtained from random-view RGB images.

3.4. Performance Comparison with Previously Published Models

The performance of the proposed model was compared with those of previously reported action recognition models, including posec3d [22], stgcn [29], stgcn++ [33], ctrgcn [26], and aagcn [34]. To this end, the proposed model was trained using the 2D skeletal databases of Models C and D. The previous models were trained using the 2D skeletal databases extracted from the RGB images of the poomsae unit action captured from the frontal and lateral viewpoints presented in Section 3.2. The synthetic two-dimensional skeletal dataset and random-viewpoint RGB image datasets were used as test datasets for the proposed and previously reported models. The detailed outcomes of this evaluation for the synthetic two-dimensional skeletal test dataset can be found in Table 8. It is noteworthy that both Model C and Model D achieved accuracies surpassing 0.97 within the table. The specific results of this evaluation for the random-view RGB image test dataset are enumerated in Table 9. Notably, the proposed model trained with the 2D skeletal databases of Model D exhibited superior performance, achieving an accuracy of 0.8670.

4. Discussion

This study examined the efficacy of action recognition of Taekwondo poomsae. This was achieved by examining various training and testing datasets. The performance of four models, namely Models A, B, C, and D, trained on 2D skeletal representations obtained by projecting 3D skeletons from diverse camera viewpoints, was evaluated and contrasted across distinct testing datasets. The evaluation outcomes of the 2D skeletal data obtained by projecting the 3D skeleton at 10° intervals across the viewpoints revealed that Model A achieved an accuracy of only 0.7997. In contrast, Models B, C, and D achieved accuracies of more than 0.96. This observation underscores the insufficiency of relying solely on frontal and lateral viewpoint data to recognize actions from other perspectives.
Next, the performance of the proposed model was evaluated using RGB images captured from the frontal and lateral viewpoints. Among the four models assessed, Model A again exhibited the lowest accuracies of 0.5795 and 0.6549 for the joint and bone heatmap models, respectively. In contrast, Model D demonstrated the highest accuracy of 0.8705 in the joint heatmap model, and Model C achieved the highest accuracy of 0.8761 in the bone heatmap model. This highlights the potential enhancement of the recognition performance achieved via the incorporation of projection data from different viewpoints. The assessment encompassed image data captured from random viewpoints, and 2D skeletal representations were extracted from images obtained using HRNet from those viewpoints. Model C achieved the highest joint heatmap precision of 0.9381, while Model D attained a peak skeletal heatmap precision of 0.8670. The observed decline in the accuracy of the image data evaluation can be attributed to the disparities between the 2D skeletal representations obtained via the projection of 3D skeletons and those obtained using the estimation algorithm. Despite efforts to align them by discarding significantly different keypoints, inherent misalignments still affected the accuracy. The alignment challenge can be addressed by utilizing a generative adversarial network (GAN) model trained with a dataset comprising 2D skeletons obtained through the projection of 3D skeletons and corresponding 2D skeletons extracted from the image.
To understand the limitations of Model A, we conducted qualitative analysis, which revealed that Model A struggles to accurately recognize actions under varying camera viewpoints. This limitation becomes particularly pronounced when assessing Taekwondo forms from non-frontal and non-profile angles. Model A’s shortcomings are evident as it frequently misinterprets complex movements, leading to lower accuracy scores. In contrast, Models C and D exhibit outstanding accuracy in action recognition. Qualitative analysis indicates that their robustness stems from their ability to generalize well across different camera perspectives. Even when faced with challenging viewpoints, they consistently recognize key actions. This qualitative insight underscores the potential of these models in real-world applications, where camera angles may not always be fixed. To qualitatively assess the viewpoint sensitivity of the proposed model, we analyzed their performance across a range of camera perspectives. While quantitative results indicate some sensitivity to different viewpoints, qualitative analysis reveals that the models may struggle to achieve complete accuracy when trained on limited data.
To compare the proposed approach and the existing models, the latter were trained on sample images acquired from two RGB-D cameras positioned frontally and laterally. The images were employed to extract 2D skeletal representations that were used for training. A comparative assessment was conducted by juxtaposing the models on a test dataset of random-view RGB images, where Model C exhibited better performance than the existing models by more than 10%. This observation underscores the high recognition performance of the proposed method for Taekwondo poomsae from arbitrary viewpoints. Thus, the motion-recognition methodology demonstrates high precision across distinct viewpoints and datasets. Additionally, the recognition performances of Models C and D were superior to those of other variants. Furthermore, compared with the previously reported models, the proposed model exhibited superiority in action recognition, which highlights the efficacy of the proposed approach in real-world scenarios.
This study employed various training and testing datasets, yet the variability in the data may not fully encapsulate the complexity and diversity of real-world scenarios. However, we believe that the proposed model holds potential feasibility in other domains as well. It is important to note that validating the proposed model in these alternative domains requires a significant allocation of resources. The evaluation of this study focused on Taekwondo poomsae movements, and the generalizability of the proposed approach to other action recognition tasks remains to be explored. The action dataset TUAD-3D, which was curated by Taekwondo experts, was constructed for the purpose of developing and validating the proposed methodology. However, the application of the proposed method to the assessment of Taekwondo poomsae performed by general individuals requires a diverse set of action samples. This is because individuals’ executions of the movements may exhibit significant variations. Collecting data from such individuals is a task that will be addressed in our future work.
Future research can benefit from integrating a wider array of data sources to capture diverse lighting conditions, backgrounds, and environmental factors. Given that the proposed action recognition algorithm relies on skeleton data, real-world environmental factors, such as lighting conditions and background noise, do not directly affect the recognition performance. Those environmental factors are removed in the process of skeleton extraction. However, it is important to note that these real-world environmental factors can still have an impact on the accuracy of 2D skeletons extracted through pose-estimation algorithms from images.
The computational cost of the proposed model was measured in FLOPs (floating-point operations per second). The model comprises two million parameters, reflecting a balanced model complexity. Both model training and inference were performed on a high-performance computing system equipped with an AMD® Ryzen Threadripper 3960X 24-core processor, 48 CPUs, two GeForce RTX 3090 GPUs, and 96.0 GB of memory. The model’s average inference speed was measured at 0.125 s, demonstrating efficient real-time performance.

5. Conclusions

This study constructed a TUAD-3D dataset by employing full-body motion-capture suits to collect accurate 3D skeletal data. This dataset contained 7680 samples and included 16 fundamental techniques performed by 40 Taekwondo experts. The model effectively synthesized 2D skeletal representations from the collected 3D skeletal data and integrated multiple viewpoints during the training process. This approach ensured consistent and reliable model performance, regardless of the observer’s angles and positions. Through a comprehensive evaluation of various action recognition networks, we observed that two of the model variants, Models C and D, which were trained using 3D skeleton projection, exhibited higher accuracy. The assessment results demonstrated the superiority of the proposed model over those reported previously, highlighting its effectiveness in Taekwondo poomsae action recognition and classification. Furthermore, analysis across different viewpoints and datasets revealed the significance of optimal camera viewpoint selection for training, which influenced the model performance. However, a decline in accuracy was observed during the model evaluation of image data, owing to inherent disparities between the 2D skeletal representations obtained via 3D skeleton projections and those estimated by the algorithm. Despite efforts to align these skeletal representations, their accuracies remained affected. In conclusion, this study contributes to the advancement of action recognition technology in the context of Taekwondo poomsae. The proposed model demonstrated robust performance across various viewpoints and datasets, highlighting its potential for real-world applications. Considering factors such as viewpoint selection and data alignment, further investigation and refinement of action recognition models can enhance their accuracy and performance in the field of human motion analysis.

Author Contributions

Conceptualization, H.J. and K.L.; methodology, H.J. and C.L; software, C.L.; validation, C.L., H.-Y.P. and S.-W.K.; formal analysis, H.-Y.P. and S.-W.K.; investigation, C.L.; resources, H.-Y.P. and S.-W.K.; data curation, H.-Y.P. and S.-W.K.; writing—original draft preparation, C.L.; writing—review and editing, H.J.; visualization, C.L.; supervision, H.J.; project administration, K.L.; funding acquisition, K.L. and H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by Konkuk University in 2019.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Konkuk University (7001355-202004-HR-372).

Informed Consent Statement

Informed consent was obtained from all the subjects involved in the study.

Data Availability Statement

Data cannot be provided, owing to data security reasons.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Choi, C.-H.; Joo, H.-J. Motion recognition technology based remote Taekwondo Poomsae evaluation system. Multimed. Tools Appl. 2016, 75, 13135–13148. [Google Scholar] [CrossRef]
  2. Lee, J.; Jung, H. TUHAD: Taekwondo Unit Technique Human Action Dataset with Key Frame-Based CNN Action Recognition. Sensors 2020, 20, 4871. [Google Scholar] [CrossRef] [PubMed]
  3. Andó, B.; Baglio, S.; Lombardo, C.O.; Marletta, V. An Event Polarized Paradigm for ADL Detection in AAL Context. IEEE Trans. Instrum. Meas. 2015, 64, 1814–1825. [Google Scholar] [CrossRef]
  4. Hsieh, J.; Chuang, C.; Alghyaline, S.; Chiang, H.; Chiang, C. Abnormal Scene Change Detection from a Moving Camera Using Bags of Patches and Spider-Web Map. IEEE Sens. J. 2015, 15, 2866–2881. [Google Scholar] [CrossRef]
  5. Cosar, S.; Donatiello, G.; Bogorny, V.; Garate, C.; Alvares, L.O.; Brémond, F. Toward Abnormal Trajectory and Event Detection in Video Surveillance. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 683–695. [Google Scholar] [CrossRef]
  6. Ismail, S.J.; Rahman, M.A.A.; Mazlan, S.A.; Zamzuri, H. Human gesture recognition using a low cost stereo vision in rehab activities. In Proceedings of the 2015 IEEE International Symposium on Robotics and Intelligent Sensors (IRIS), Langkawi, Malaysia, 18–20 October 2015; pp. 220–225. [Google Scholar] [CrossRef]
  7. Rafferty, J.; Nugent, C.D.; Liu, J.; Chen, L. From Activity Recognition to Intention Recognition for Assisted Living Within Smart Homes. IEEE Trans. Hum. Mach. Syst. 2017, 47, 368–379. [Google Scholar] [CrossRef]
  8. Zolfaghari, S.; Keyvanpour, M.R. SARF: Smart activity recognition framework in Ambient Assisted Living. In Proceedings of the 2016 Federated Conference on IEEE Computer Science and Information Systems (FedCSIS), Gdansk, Poland, 11–14 September 2016; pp. 1435–1443. [Google Scholar]
  9. Zhang, L.; Hsieh, J.-C.; Ting, T.-T.; Huang, Y.-C.; Ho, Y.-C.; Ku, L.-K. A Kinect based golf swing score and grade system using GMM and SVM. In Proceedings of the 5th International Congress on Image and Signal Processing (CISP 2012), Chongqing, China, 16–18 October 2012; pp. 711–715. [Google Scholar]
  10. Zhu, G.; Xu, C.; Huang, Q.; Gao, W.; Xing, L. Player action recognition in broadcast tennis video with applications to semantic analysis of sports game. In Proceedings of the 14th ACM International Conference on Multimedia, Santa Barbara, CA, USA, 23–27 October 2006; pp. 431–440. [Google Scholar]
  11. Martin, P.-E.; Benois-Pineau, J.; Péteri, R.; Morlier, J. Sport Action Recognition with Siamese Spatio-Temporal Cnns: Application to Table Tennis. In Proceedings of the 2018 International Conference on Content-Based Multimedia Indexing (CBMI), La Rochelle, France, 4–6 September 2018. [Google Scholar]
  12. Wang, S. A Deep Learning Algorithm for Special Action Recognition of Football. Mob. Inf. Syst. 2022, 2022, 6315648. [Google Scholar] [CrossRef]
  13. Leo, M.; D’Orazio, T.; Spagnolo, P.; Mazzeo, P.L.; Distante, A. Multi-view Player Action Recognition in Soccer Games. In Computer Vision/Computer Graphics CollaborationTechniques; MIRAGE 2009; Lecture Notes in Computer Science; Gagalowicz, A., Philips, W., Eds.; Springer: Berlin/Heidelberg, Germany, 2009; Volume 5496. [Google Scholar] [CrossRef]
  14. Lin, C.-H.; Tsai, M.-Y.; Chou, P.-Y. A Lightweight Fine-Grained Action Recognition Network for Basketball Foul Detection. In Proceedings of the 2021 IEEE International Conference on Consumer Electronics-Taiwan (ICCE-TW), Taipei, Taiwan, 6–8 July 2021; pp. 1–2. [Google Scholar]
  15. Ji, R. Research on Basketball Shooting Action Based on Image Feature Extraction and Machine Learning. IEEE Access 2020, 8, 138743–138751. [Google Scholar] [CrossRef]
  16. Mora, S.V.; Knottenbelt, W.J. Deep Learning for Domain-Specific Action Recognition in Tennis. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 170–178. [Google Scholar]
  17. Rahmad, N.; As’ari, M. The new Convolutional Neural Network (CNN) local feature extractor for automated badminton action recognition on vision based data. J. Phys. Conf. Ser. 2020, 1529, 022021. [Google Scholar] [CrossRef]
  18. Rahmad, N.; As’ari, M.; Soeed, K.; Zulkapri, I. Automated badminton smash recognition using convolutional neural network on the vision based data. In Proceedings of the IOP Conference Series: Materials Science and Engineering; IOP Publishing: Putrajaya, Malaysia, 2020; Volume 884, p. 012009. [Google Scholar]
  19. Ijjina, E.P.; Chalavadi, K.M. Human action recognition in RGB-D videos using motion sequence information and deep learning. Pattern Recognit. 2017, 72, 504–516. [Google Scholar] [CrossRef]
  20. Wang, P.; Wang, S.; Gao, Z.; Hou, Y.; Li, W. Structured Images for RGB-D Action Recognition. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV 2017), Venice, Italy, 22–29 October 2017; pp. 1005–1014. [Google Scholar]
  21. Trivedi, N.; Kiran, R.S. PSUMNet: Unified Modality Part Streams are All You Need for Efficient Pose-based Action Recognition. arXiv 2022, arXiv:2208.05775. [Google Scholar]
  22. Duan, H.; Zhao, Y.; Chen, K.; Lin, D.; Dai, B. Revisiting skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2969–2978. [Google Scholar]
  23. Xia, H.; Gao, X. Multi-Scale Mixed Dense Graph Convolution Network for Skeleton-Based Action Recognition. IEEE Access 2021, 9, 36475–36484. [Google Scholar] [CrossRef]
  24. Gupta, P.; Thatipelli, A.; Aggarwal, A.; Maheshwari, S.; Trivedi, N.; Das, S.; Sarvadevabhatla, R.K. Quo vadis, skeleton action recognition? Int. J. Comput. Vis. 2021, 129, 2097–2112. [Google Scholar] [CrossRef]
  25. Song, Y.F.; Zhang, Z.; Shan, C.; Wang, L. Constructing stronger and faster baselines for skeleton-based action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1474–1488. [Google Scholar] [CrossRef] [PubMed]
  26. Chen, Y.; Zhang, Z.; Yuan, C.; Li, B.; Deng, Y.; Hu, W. Channel-wise Topology Refinement Graph Convolution for Skeleton-Based Action Recognition. arXiv 2021, arXiv:2107.12213. [Google Scholar]
  27. Wang, M.; Ni, B.; Yang, X. Learning Multi-View Interactional Skeleton Graph for Action Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 45, 6940–6954. [Google Scholar] [CrossRef] [PubMed]
  28. Du, Y.; Wang, W.; Wang, L. Hierarchical recurrent neural network for skeleton based action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 8–10 June 2015; pp. 1110–1118. [Google Scholar]
  29. Yan, S.; Xiong, Y.; Lin, D. Spatial Temporal Graph Convolutional Networks for Skeleton-Based Action Recognition. arXiv 2018, arXiv:1801.07455. [Google Scholar] [CrossRef]
  30. Song, S.; Lan, C.; Xing, J.; Zeng, W.; Liu, J. An end-to-end spatio-temporal attention model for human action recognition from skeleton data. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  31. Roetenberg, D.; Luinge, H.; Slycke, P. Xsens MVN: Full 6DOF human motion tracking using miniature inertial sensors. Xsens Motion Technol. BV Tech. Rep. 2009, 1, 1–7. [Google Scholar]
  32. Feichtenhofer, C.; Fan, H.; Malik, J.; He, K. Slowfast networks for video recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6202–6211. [Google Scholar]
  33. Duan, H.; Wang, J.; Chen, K.; Lin, D. Pyskl: Towards good practices for skeleton action recognition. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 7351–7354. [Google Scholar]
  34. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Skeleton-based action recognition with multi-stream adaptive graph convolutional networks. IEEE Trans. Image Process. 2020, 29, 9532–9545. [Google Scholar] [CrossRef] [PubMed]
Figure 1. 3D skeleton data-collection procedure.
Figure 1. 3D skeleton data-collection procedure.
Sensors 23 08049 g001
Figure 2. Structure of Taekwondo unit-action dataset of 3D skeletons (TUAD-3D).
Figure 2. Structure of Taekwondo unit-action dataset of 3D skeletons (TUAD-3D).
Sensors 23 08049 g002
Figure 3. Overall architecture of the 3D CNN-based viewpoint-agnostic action recognition.
Figure 3. Overall architecture of the 3D CNN-based viewpoint-agnostic action recognition.
Sensors 23 08049 g003
Figure 4. Generation of 2D skeletons by projecting a 3D skeleton onto various image planes: (a) determination of camera viewpoints by rotating the initial viewpoint; (b) projection of the 3D skeleton onto the desired image planes.
Figure 4. Generation of 2D skeletons by projecting a 3D skeleton onto various image planes: (a) determination of camera viewpoints by rotating the initial viewpoint; (b) projection of the 3D skeleton onto the desired image planes.
Sensors 23 08049 g004
Figure 5. 3D CNN SlowFast architecture.
Figure 5. 3D CNN SlowFast architecture.
Sensors 23 08049 g005
Figure 6. The number of skeletons used in the model training.
Figure 6. The number of skeletons used in the model training.
Sensors 23 08049 g006
Figure 7. RGB image samples used for performance evaluation: (a) frontal and (b) lateral.
Figure 7. RGB image samples used for performance evaluation: (a) frontal and (b) lateral.
Sensors 23 08049 g007
Figure 8. RGB image samples captured from random viewpoints for the performance evaluation of the model.
Figure 8. RGB image samples captured from random viewpoints for the performance evaluation of the model.
Sensors 23 08049 g008
Table 1. Viewpoint configuration of 2D skeletons of four training datasets.
Table 1. Viewpoint configuration of 2D skeletons of four training datasets.
IDViewpoint ConfigurationNumber of ViewpointsNumber of Training Data
Model A 0 ° ,   90 ° 212,360
Model B 0 ° ,   90 ° , 180 ° , 270 ° 424,720
Model C 0 ° ,   45 ° ,   90 ° ,   135 ° , ,   315 ° 849,440
Model D 0 ° ,   10 ° ,   20 ° ,   30 ° ,   ,   350 ° 36222,480
Table 2. Performance evaluation results of the joint heatmap model tested using a random-projection 2D skeletal dataset.
Table 2. Performance evaluation results of the joint heatmap model tested using a random-projection 2D skeletal dataset.
ModelPrecisionRecallF1-ScoreAccuracy
Model A0.86110.83980.83730.7997
Model B0.96800.96690.96700.9669
Model C0.97690.97640.97650.9764
Model D0.98030.98020.98020.9802
Table 3. Performance evaluation results of the bone heatmap model tested using a random-projection 2D skeletal dataset.
Table 3. Performance evaluation results of the bone heatmap model tested using a random-projection 2D skeletal dataset.
ModelPrecisionRecallF1-ScoreAccuracy
Model A0.86110.83980.83730.7997
Model B0.96860.96820.96820.9682
Model C0.97690.97640.97650.9764
Model D0.97860.97830.97840.9783
Table 4. Performance evaluation results of the joint heatmap model using the 2D skeletal dataset extracted from RGB images.
Table 4. Performance evaluation results of the joint heatmap model using the 2D skeletal dataset extracted from RGB images.
ModelPrecisionRecallF1-ScoreAccuracy
Model A0.76380.76380.58540.5795
Model B0.87610.85330.85000.8516
Model C0.87050.76470.77630.7626
Model D0.89980.87170.87060.8705
Table 5. Performance evaluation results of bone heatmap model using the 2D skeletal dataset extracted from RGB images.
Table 5. Performance evaluation results of bone heatmap model using the 2D skeletal dataset extracted from RGB images.
ModelPrecisionRecallF1-ScoreAccuracy
Model A0.74530.65590.65490.6549
Model B0.87500.83030.83000.8294
Model C0.89030.87660.87520.8761
Model D0.88910.87430.87170.8732
Table 6. Performance evaluation results of the joint heatmap model using the 2D skeletal dataset extracted from the RGB images captured from random viewpoints.
Table 6. Performance evaluation results of the joint heatmap model using the 2D skeletal dataset extracted from the RGB images captured from random viewpoints.
ModelPrecisionRecallF1-ScoreAccuracy
Model A0.82980.81850.79440.8398
Model B0.92170.90090.90370.9010
Model C0.94320.93810.93840.9381
Model D0.89770.87020.86820.8623
Table 7. Performance evaluation results of the bone heatmap model using the 2D skeletal dataset extracted from the RGB images captured from random viewpoints.
Table 7. Performance evaluation results of the bone heatmap model using the 2D skeletal dataset extracted from the RGB images captured from random viewpoints.
ModelPrecisionRecallF1-ScoreAccuracy
Model A0.71420.62220.60950.6041
Model B0.86650.78620.79300.7715
Model C0.88480.84710.84820.8419
Model D0.90400.87360.87640.8670
Table 8. Performance comparison among the proposed and previous models using synthetic two-dimensional skeletal test dataset.
Table 8. Performance comparison among the proposed and previous models using synthetic two-dimensional skeletal test dataset.
ModelPrecisionRecallF1-ScoreAccuracy
stgcn [29]0.69260.59110.59570.5911
stgcn++ [33]0.67000.53270.55160.5327
ctrgcn [26]0.71290.58050.60600.5805
aagcn [34]0.74170.64710.65710.6471
posec3d [22]0.74530.58180.61180.5818
Proposed (Model C)0.97690.97640.97650.9764
Proposed (Model D)0.97860.97830.97840.9783
Table 9. Performance comparison among the proposed and previous models using random-viewpoint RGB image test dataset.
Table 9. Performance comparison among the proposed and previous models using random-viewpoint RGB image test dataset.
ModelPrecisionRecallF1-ScoreAccuracy
stgcn [29]0.72570.69150.67150.6667
stgcn++ [33]0.81440.72100.74170.7058
ctrgcn [26]0.7944065090.65870.6275
aagcn [34]0.80400.74420.75410.7261
posec3d [22]0.88250.75200.78650.7340
Proposed (Model C)0.88480.84710.84820.8419
Proposed (Model D)0.90400.87360.87640.8670
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Luo, C.; Kim, S.-W.; Park, H.-Y.; Lim, K.; Jung, H. Viewpoint-Agnostic Taekwondo Action Recognition Using Synthesized Two-Dimensional Skeletal Datasets. Sensors 2023, 23, 8049. https://doi.org/10.3390/s23198049

AMA Style

Luo C, Kim S-W, Park H-Y, Lim K, Jung H. Viewpoint-Agnostic Taekwondo Action Recognition Using Synthesized Two-Dimensional Skeletal Datasets. Sensors. 2023; 23(19):8049. https://doi.org/10.3390/s23198049

Chicago/Turabian Style

Luo, Chenglong, Sung-Woo Kim, Hun-Young Park, Kiwon Lim, and Hoeryong Jung. 2023. "Viewpoint-Agnostic Taekwondo Action Recognition Using Synthesized Two-Dimensional Skeletal Datasets" Sensors 23, no. 19: 8049. https://doi.org/10.3390/s23198049

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop