Next Article in Journal
Design, Implementation, and Evaluation of a Head and Neck MRI RF Array Integrated with a 511 keV Transmission Source for Attenuation Correction in PET/MR
Previous Article in Journal
An Efficient Electrochemical Sensor Driven by Hierarchical Hetero-Nanostructures Consisting of RuO2 Nanorods on WO3 Nanofibers for Detecting Biologically Relevant Molecules
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Context-Aware Winter Sports Based on Multivariate Sequence Learning

1
Telerobotics and Control Laboratory, Korea Advanced Institute of Science Technology, Daejeon 34141, Korea
2
Institute for Cognitive Science, Seoul National University, Seoul 08826, Korea
3
Intelligent Robotics Laboratory, Hallym University, Chuncheon 24252, Korea
*
Authors to whom correspondence should be addressed.
Sensors 2019, 19(15), 3296; https://doi.org/10.3390/s19153296
Submission received: 15 June 2019 / Revised: 22 July 2019 / Accepted: 23 July 2019 / Published: 26 July 2019
(This article belongs to the Section Intelligent Sensors)

Abstract

:
In this paper, we present an intelligent system that is capable of estimating the status of a player engaging in winter activities based on the sequence analysis of multivariate time-series sensor signals. Among the winter activities, this paper mainly focuses on downhill winter sports such as alpine skiing and snowboarding. Assuming that the mechanical vibrations generated by physical interaction between the ground surface and ski/snowboard in motion can describe the ground conditions and playing contexts, we utilize inertial and vibration signals to categorize the motion context. For example, the proposed system estimates whether the player is sitting on a ski lift or standing on the escalator, or skiing on wet or snowy ground, etc. To measure the movement of a player during a game or on the move, we develop a custom embedded system comprising a motion sensor and piezo transducer. The captured multivariate sequence signals are then trained in a supervised fashion. We adopt artificial neural network approaches (e.g., 1D convolutional neural network, and gated recurrent neural networks, such as long short-term memory and gated recurrent units). The experimental results validate the feasibility of the proposed approach.

1. Introduction

Snow deposited on a slope is typically in a state of continuous deformation [1]. Ground conditions play a significant role in a variety of winter sports and recreational activities. For example, the preparation of equipment (e.g., the type and quantity of wax applied to the base of a set of skis or a snowboard) can vary depending on the snow quality (e.g., slush or ice). Also, surfaces with different snow conditions likely lead to skiers bumping or moving in unexpected ways, forcing them to adjust their riding styles on slopes. For example, when snow is melting (e.g., slushy snow in which the snow particles are completely immersed in water), snowboarders generally tend to move their center of mass backwards, such that the nose of their board does not collide with and stick to the wet ground. Although understanding snow quality is important, the identification of such physical characteristics through remote surveillance (e.g., computer vision techniques) is challenging or too costly to practically apply over large snowfields.
Besides estimating the quality of snow deposited on the ground, we focused on estimating a participant’s motion context, defined as the context that describes the participant’s situation or state in a snow field (e.g., skiing downhill and sitting on a ski lift). In fact, detecting and reporting conditions at ski resorts are important in that low temperatures may cause significant hurt to participants under abnormal snow conditions. In addition, quantification and classification of a player’s motion are important for coaching and prediction purposes from the perspective of sports analytics [2].
To estimate both ground conditions and motion contexts, we utilized inertial and audio signals, generated from friction and motion. These signals can describe dynamic information. To this end, we first collected time-series signals with a custom-built embedded system, which incorporates a number of sensors on board. Then, we trained the captured multivariate time-series signals in a supervised learning setting (i.e., a many-to-one mapping problem) using different types of machine learning techniques. This paper primarily contributes to the literature in two ways:
  • We proposed an intelligent context-aware system that classifies both ground conditions and motion contexts based on the incoming inertial and vibrational sensor measurements.
  • We conducted an experiment to investigate whether the proposed contexts can be classified based on the collected time-series signals.
The next section discusses related research.

2. Related Work

2.1. Sequence Classification in Human-Computer Interaction

Predicting a category of a given input sequence is called sequence classification, which forms the core of a variety of human–computer interaction applications [3,4,5,6,7]. Although a number of conventional approaches, such as k-nearest neighbor classifiers, Naïve Bayes, random forest [8], and support vector machine [9], are also widely used for sequence classification, the high feature dimensionality and sequential nature of signals whose feature attributes are ordered in time [10], complicate their understanding signals properly even when using sophisticated feature engineering techniques [11].
On the other hand, feedforward neural networks, which accept input data and learn complex features internally [12,13] for approximating some functions [14], are receiving growing attention for such tasks. Notably, recurrent neural networks (RNNs), which have recurrent hidden states that encode sequential information into a fixed length of vectors [15,16], have achieved state-of-the-art performance for a variety of sequence classification tasks [17,18,19]. Socher et al. proposed an intelligent system that classifies 3D objects in RGB-D (RGB and Depth) video sequences [17] by employing convolutional neural network (CNN) and RNN-based sequence learning architecture. RNN-based architecture was widely employed in the area of gesture recognition tasks to process RGB videos [20] and inertial sensor sequences from wearable devices [21,22]. Recently, Kim and Han revealed that a multidimensional temporal sequence can be encoded as latent space vectors using gated RNNs (e.g., long short-term memory (LSTM) [23] and gated recurrent units (GRUs) [15]) [24].
In a recent pioneering work [25], Yoon reported interesting results for a classification task of time-series signals (e.g., a natural language processing task) using CNN, which was originally introduced for a computer vision task approximately 30 years ago [26]. Because CNN is capable of learning both local and global features even from sequential data while allowing parallel computation [25,27] and requiring fewer pre-processing tasks, it is widely employed in many recent time-series applications [27,28,29,30,31].

2.2. Sensor-Based Sports Analytics

Internet of Things technologies are widely adopted for various sports analytics [2,32,33,34,35]. In a recent study, Yu et al. proposed an intelligent system that could analyze the inertial signals captured from sensors attached to multiple body parts of a professional skier [34]. In another recent study, Umek et al. designed a sensor-laden sports equipment for capturing the motion of a golf club [33]. They conducted an experiment with a set of sensors, such as a strain gauge, an accelerometer, and a gyroscope, to quantify the player’s motion. Wearable devices are also widely utilized for seamless sports analytics [32]. In a commercial market, a variety of sports activities, (e.g., basketball, tennis, soccer, running, etc.) are now being assisted by sensor and machine learning technologies [35]. For example, a motion and vibration sensor attached to the grip end of tennis racket can track a player’s shot type, such as serve, smash, and volley forehand [36].

3. Proposed System

This section describes the developed prototype hardware and context understanding system that classifies a player’s motion state (e.g., sitting on the ski lift or standing on the escalator, or skiing on the snow) and ground conditions based on the measurements. As described in Section 1, we assumed that the measured signals will vary depending on bumpiness of the surface.

3.1. Hardware

A multivariate time-series dataset was collected using a custom-built embedded system based on an ATmega328 microprocessor (Arduino Nano) with multiple sensors, namely a motion sensor (MPU9250, InvenSense, San Jose, CA, USA) and a vibration sensor (LDT0-028K, Measurement Specialties, Inc., Hampton, VA, USA), composed of piezoelectric polyvinylidene difluoride polymer film for capturing the ground vibrations generated mainly due to friction. To record the sensor signals and context information simultaneously, we also added a mode selection button on the outer cover of the sensor box (see Figure 1). A MicroSD card breakout board (Adafruit Industries, New York, NY, USA) and a real-time clock (RTC) with an integrated temperature compensated crystal oscillator (DS3231, Maxim Integrated, San Jose, CA, USA) were utilized for recording the data into the memory card. Then, we analyzed the recorded sequential data in a supervised manner. Figure 1 shows the developed prototype board.

3.2. Experiment-Sequence Classification

This section describes the process of the proposed sequence learning task. Using the sensor board described in the previous section, we collected data at 250 Hz at the Elysian Ski resort, Gangwon, South Korea, from a professional snowboard player, during February 2018. Table 1 shows the relevant weather information. The snowboarder who participated in this study tried to snowboard on the same route, preferably with the same strategy. We collected 2400 samples in total, each of which was comprised of 100 time steps. We balanced the data distribution by adjusting the number of data per class. The datasets were split such that 70% was used for training and the remainder for test purposes.
Considering that temperature and humidity determines the quality of the deposited snow on ground, we set five classes as described in Table 2.
To obtain a sense of how the measurements are distributed and to verify that the training and the test data are in the same distribution, we conducted an analysis using kernel density estimation (KDE). Figure 2 shows training and test dataset distributions.
In Figure 3, different colors represent the different values of classes defined in Table 2.
The density plots of variables in Figure 3 provides information regarding their distribution, such as the spread of the variables and central tendency. Especially, linear accelerations and angular velocities in Y-axis have apparently different probability density distributions.

3.2.1. Baseline Classifier

As the baseline, we adopted a random forest classifier, which uses a large number of decision trees in the ensemble on various sub-samples of the dataset [8]. It often shows better performance (i.e., predictive accuracy) than a single decision tree in an aggregation of many decision trees, each of which may be prone to noise, because it often reduces the effect of noises [37,38]. Also, it generally demonstrates robust results against overfitting [39]. We adopted this classifier as a baseline because it does not consider the sequential nature of a given signal.
Based on the measured signals, we also generated three additional sequences, used for further feature engineering processes. Table 3 summarizes the name of the sequence used for this study and describes them.
Each sequence among the ten concurrent sequences (seven measured and three derived signals) was then used for calculating the following features: mean, median, min, max, max/min, std, skew, abs_max, min of abs, mean of abs, and standard deviation of abs, as summarized in Table 4. Therefore, 110 (11 sequences × 10 features) features were derived in total.

3.2.2. One-Dimensional Convolutional Neural Network Model

We first adopted the 1D-CNN approach, in which a set of kernels is convolved with the inputs along a single temporal dimension as in [27]. We employed a sequential model implemented in Keras, which uses a linear stack of layers, for the training process. Figure 4 shows the model structure used for this study.
The convolution parts (e.g., Conv1D followed by MaxPooling1D) in the model depicted in Figure 4 learn how to extract features from the time-series sequences. Further, a fully connected layer (e.g., Dense) learns how to associate the learnt internal features to the types of motion context and overall surface quality listed in Table 2. As discussed in Section 2.1, the CNN-based approach has advantages in terms of parallel computation, in principle.

3.2.3. Gated RNNs—LSTM and GRU

The basic RNN algorithm (i.e., RNN with a traditional tanh unit), especially one with a deep stacked architecture, poses the problem of vanishing and exploding gradients which complicates the learning of long-term dependencies. LSTM was proposed to resolve this problem as it can learn long term dependencies by utilizing the memory cell and gate units. Thus, LSTM has been adopted by many applications [40]. The memory cell stores information taken from the input and previous cells over a period of time. This information is controlled by the gate units, which are composed of the update gate, forget gate, and output gate. GRU has also been shown to outperform vanilla RNN in many applications, such as language modelling, torque generation [41], and others [24]. Compared to LSTM, GRU is composed of a smaller number of gates, and therefore, it often allows faster optimization progress than LSTM. Also, GRU performs better in smaller datasets [16]. Based on the previously described properties of gated RNNs, we constructed a classification network as seen in Figure 5.
We trained the captured data based on the network with the stochastic gradient descent algorithm using Tensorflow.

3.3. Experimental Results and Discussion

Table 5 shows the test accuracy obtained from the experiments. Here, RF (random forest) was used as the baseline classifier.
Overall, Table 5 shows that the sequence data from five different classes were reliably classified. Figure 6 shows a set of confusion matrices across all five classes when all measurements (i.e., dimension of 7) are used for inference.
Remarkably, the proposed model successfully identified not only motion contexts but also three different ground conditions affected by weather condition (e.g., powder, melt-free crust, and slush). The experimental results using fewer input signals (e.g., Acc + Gyro Y and Acc X + Gyro) also exhibited reasonable experimental results. This result can be used for reducing multiply-accumulates for high-speed/low-power inference operation during the runtime period.

4. Limitations and Future Work

4.1. Generalization of Our System

Thus far, we have proposed an intelligent sports system that understands contexts that are likely to occur during winter sports. We then validated the proposed system using various machine learning architectures. The results of the experiment validated the proposed approach.
However, owing to the nature of winter sports, experiments were not conducted under various conditions. Further, the trained model is overfitted to a single professional player; this is a disadvantage. Although our work validated the feasibility of the proposed approach, as future work, we plan to conduct additional experiments during the winter season to generalize our model.

4.2. Application to an Embedded System

The experimental results presented in this paper are derived from the signals stored using the Arduino-based embedded system, which later learned in the PC. In the future, we intend to implement this system in the embedded system for real-time inferences. To study its feasibility, we duplicated the trained model to other types of embedded machines such as Raspberry Pi 3 Model B+ (armv7l) and NVIDIA Jetson TX2 (aarch64) that supports an open source machine learning library (e.g., TensorFlow and Keras). Figure 7 shows the average inference time for a given sequence classification task with respect to the type of device.
Raspberry Pi 3 Model B+, which incorporates a 1.4 GHz 64-bit quad-core processor, costs approximately USD 50. Thus, the proposed approach can be implemented at the embedded device level.

4.3. User Interface for Labeling

Our prototype device assigns labels before motion takes place (i.e., by pressing a button outside of the box, as shown in Figure 1) due to the design characteristics of the system (i.e., it is supposed to be mounted on the board). Thus we faced difficulties in instant tagging, which is essential for capturing unexpected situations, such as a fall while skiing. As future work, we plan to develop a more user-friendly interface, capable of capturing contexts instantaneously.

5. Conclusions

In this paper, we proposed an intelligent system for winter sports that estimates status of a player engaging in winter activities based on the sequence analysis of multivariate time-series sensor signals. More specifically, the system classifies both ground conditions (e.g., powder, melt-free crust, and slush) and motion contexts (e.g., sitting on a ski lift or standing on the escalator) based on inertial and vibrational signals, assuming that surfaces under different snow conditions likely lead to skiers bumping into objects or skiing in different directions systematically.
We designed a set of sequence classifiers, such as random forest, convolutional neural network (1D-CNN), and gated recurrent neural networks (e.g., LSTM and GRU). The random forest classifier was utilized as the baseline. The experimental results validated the feasibility of the proposed approach. We expect that the proposed system will guide the design of smart systems for winter activities. In future studies, we plan to optimize the network structure to reduce the inference time so that the system can be used for real-time embedded applications.

Author Contributions

S.-C.K. and B.-K.H. conceived and implemented the proposed system. S.-C.K. and J.-K.R. conducted the experiment and analyzed the experimental data. All of the authors wrote the paper.

Funding

This research project was supported by the Sports Promotion Fund of Seoul Olympic Sports Promotion Foundation from Ministry of Culture, Sports and Tourism, and Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2016R1D1A1B03936326).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Fierz, C.; Armstrong, R.L.; Durand, Y.; Etchevers, P.; Greene, E.; McClung, D.M.; Nishimura, K.; Satyawali, P.K.; Sokratov, S.A. International Classification for Seasonal Snow on the Ground; UNESCO/IHP: Paris, France, 2009. [Google Scholar]
  2. Gowda, M.; Dhekne, A.; Shen, S.; Choudhury, R.R.; Yang, L.; Golwalkar, S.; Essanian, A. Bringing IoT to Sports Analytics; NSDI: Berkeley, CA, USA, 2017; pp. 499–513. [Google Scholar]
  3. Harrison, C.; Hudson, S.E. Scratch Input: Creating Large, Inexpensive, Unpowered and Mobile Finger Input Surfaces. In Proceedings of the 21st Annual ACM Symposium on User Interface Software and Technology, Monterey, CA, USA, 19–22 October 2008; ACM: New York, NY, USA, 2008; pp. 205–208. [Google Scholar]
  4. Kim, S.C.; Lim, S.C.; Shin, J.; Choi, J. Biometrics for electronic eyes: System authentication with embedded CMOS image sensor. IEEE Trans. Consum. Electron. 2016, 62, 209–215. [Google Scholar] [CrossRef]
  5. Lim, S.C.; Shin, J.; Kim, S.C.; Park, J. Expansion of Smartwatch Touch Interface from Touchscreen to Around Device Interface Using Infrared Line Image Sensors. Sensors 2015, 15, 16642–16653. [Google Scholar] [CrossRef] [PubMed]
  6. Schwarz, J.; Harrison, C.; Hudson, S.; Mankoff, J. Cord input: An intuitive, high-accuracy, multi-degree-of-freedom input method for mobile devices. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Atlanta, GA, USA, 10–15 April 2010; pp. 1657–1660. [Google Scholar]
  7. Sato, M.; Poupyrev, I.; Harrison, C. Touché: Enhancing touch interaction on humans, screens, liquids, and everyday objects. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, 5–10 May 2012; ACM: Austin, YX, USA, 2012; pp. 483–492. [Google Scholar]
  8. Gislason, P.O.; Benediktsson, J.A.; Sveinsson, J.R. Random forests for land cover classification. J. Pattern Recognit. Lett. 2006, 27, 294–300. [Google Scholar] [CrossRef]
  9. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  10. Bagnall, A.; Lines, J.; Bostrom, A.; Large, J.; Keogh, E. The great time series classification bake off: A review and experimental evaluation of recent algorithmic advances. Data Min. Knowl. Discov. 2017, 31, 606–660. [Google Scholar] [CrossRef] [PubMed]
  11. Xing, Z.; Pei, J.; Keogh, E. A brief survey on sequence classification. ACM SIGKDD Explor. Newsl. 2010, 12, 40–48. [Google Scholar] [CrossRef]
  12. Notley, S.; Magdon-Ismail, M. Examining the use of neural networks for feature extraction: A comparative analysis using deep learning, support vector machines, and k-nearest neighbor classifiers. arXiv 2018, arXiv:1805.02294. [Google Scholar]
  13. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning Internal Representations by Error Propagation; California Univ San Diego La Jolla Inst for Cognitive Science: San Diego, CA, USA, 1985. [Google Scholar]
  14. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  15. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; Association for Computational Linguistics: Stroudsburg, PA, USA, 2014; pp. 1724–1734. [Google Scholar]
  16. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. In Proceedings of the NIPS 2014 Workshop on Deep Learning, Montreal, Canada, 12 December 2014. [Google Scholar]
  17. Socher, R.; Huval, B.; Bath, B.; Manning, C.D.; Ng, A.Y. Convolutional-Recursive Deep Learning for 3d Object Classification, Advances in Neural Information Processing Systems; Mit Press: Cambridge, MA, USA, 2012; pp. 656–664. [Google Scholar]
  18. Fortunato, M.; Blundell, C.; Vinyals, O.J. Bayesian recurrent neural networks. arXiv 2017, arXiv:1805.02294. [Google Scholar]
  19. Wu, Y.; Schuster, M.; Chen, Z.; Le, Q.V.; Norouzi, M.; Macherey, W.; Krikun, M.; Cao, Y.; Gao, Q.; Macherey, K. Google’s neural machine translation system: Bridging the gap between human and machine translation. arXiv 2016, arXiv:1609.08144. [Google Scholar]
  20. Molchanov, P.; Yang, X.; Gupta, S.; Kim, K.; Tyree, S.; Kautz, J. Online Detection and Classification of Dynamic Hand Gestures with Recurrent 3d Convolutional Neural Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 5–21 June 2016; pp. 4207–4215. [Google Scholar]
  21. Zhu, P.; Zhou, H.; Cao, S.; Yang, P.; Xue, S. Control with Gestures: A Hand Gesture Recognition System Using Off-the-Shelf Smartwatch. In Proceedings of the 2018 4th International Conference on Big Data Computing and Communications (BIGCOM), Chicago, IL, USA, 7–9 August 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 72–77. [Google Scholar]
  22. Shin, S.; Sung, W. Dynamic hand gesture recognition for wearable devices with low complexity recurrent neural networks. In Proceedings of the Circuits and Systems (ISCAS), 2016 IEEE International Symposium on Information Theory, Barcelona, Spain, 10–15 July 2016; pp. 2274–2277. [Google Scholar]
  23. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  24. Kim, S.C.; Han, B.K. Emulating Touch Signals from Multivariate Sensor Data using Gated RNNs. In Proceedings of the 2019 ACM/IEEE International Conference on Human-Robot Interaction (HRI’19), Daegu, Korea, 11–14 March 2019. [Google Scholar]
  25. Kim, Y. Convolutional neural networks for sentence classification. arXiv 2014, arXiv:1408.5882. [Google Scholar]
  26. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  27. LeCun, Y.; Bengio, Y. Convolutional networks for images, speech, and time series. In The Handbook of Brain Theory and Neural Networks; The MIT Press: Cambridge, MA, USA; Volume 3361, p. 1995.
  28. Perol, T.; Gharbi, M.; Denolle, M. Convolutional neural network for earthquake detection and location. Sci. Adv. 2018, 4, e1700578. [Google Scholar] [CrossRef] [PubMed]
  29. Lee, S.M.; Yoon, S.M.; Cho, H. Human Activity Recognition from Accelerometer Data Using Convolutional Neural Network. In Proceedings of the 2017 IEEE International Conference on Big Data and Smart Computing (BigComp), Jeju, Korea, 13–16 February 2017; pp. 131–134. [Google Scholar]
  30. Ronao, C.A.; Cho, S.B. Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst. Appl. 2016, 59, 235–244. [Google Scholar] [CrossRef]
  31. Collobert, R.; Weston, J. A Unified Architecture for Natural Language Processing: Deep Neural Networks with Multitask Learning. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 160–167. [Google Scholar]
  32. Ebling, M.R. IoT: From sports to fashion and everything in-between. J. IEEE Pervasive Comput. 2016, 15, 2–4. [Google Scholar] [CrossRef]
  33. Umek, A.; Zhang, Y.; Tomažič, S.; Kos, A.J.S. Suitability of strain gage sensors for integration into smart sport equipment: A golf club example. Sensors 2017, 17, 916. [Google Scholar] [CrossRef] [PubMed]
  34. Yu, G.; Jang, Y.J.; Kim, J.; Kim, J.H.; Kim, H.Y.; Kim, K.; Panday, S.B.J.S. Potential of IMU sensors in performance analysis of professional alpine skiers. Sensors 2016, 16, 463. [Google Scholar] [CrossRef] [PubMed]
  35. Lightman, K. Silicon gets sporty. IEEE Spectr. 2016, 53, 48–53. [Google Scholar] [CrossRef]
  36. Smart Tennis Sensor, Sony Corporation. Available online: https://www.sony.com.au/microsite/tennis/ (accessed on 25 July 2019).
  37. Levantesi, S.; Pizzorusso, V. Application of machine learning to mortality modeling and forecasting. Risks 2019, 7, 26. [Google Scholar] [CrossRef]
  38. Donges, N. The random forest algorithm. Towards Data Science. Available online: https://towardsdatascience.com/random-forest-3a55c3aca46d (accessed on 25 July 2019).
  39. Liaw, A.; Wiener, M. Classification and regression by randomForest. R news 2002, 2, 18–22. [Google Scholar]
  40. Hwang, W.; Lim, S.C. Inferring Interaction Force from Visual Information without Using Physical Force. Sensors 2017, 17, 2455. [Google Scholar] [CrossRef] [PubMed]
  41. Han, B.K.; Kim, S.C.; Kwon, D.S. DeepSnake: Sequence Learning of Joint Torques Using a Gated Recurrent Neural Network. IEEE Access 2018, 6, 76263–76270. [Google Scholar] [CrossRef]
Figure 1. (a) Block diagram of the embedded system for data acquisition. (b) Developed prototype embedded system. (c) Developed prototype sensor box attached to the top of the snowboard deck. Three buttons (including one mode selection button) and a commercial seven-segment display were fitted to the device’s housing. A waterproof plastic case and push buttons were used. Photo (c) was taken after conducting a test run.
Figure 1. (a) Block diagram of the embedded system for data acquisition. (b) Developed prototype embedded system. (c) Developed prototype sensor box attached to the top of the snowboard deck. Three buttons (including one mode selection button) and a commercial seven-segment display were fitted to the device’s housing. A waterproof plastic case and push buttons were used. Photo (c) was taken after conducting a test run.
Sensors 19 03296 g001
Figure 2. Visualization of the distributions of the training and test datasets based on kernel density estimation (KDE) to check if the split is stratified. The x-axis is the range of values in the dataset, and y-axis represents its probability density, which captures the underlying distribution of the sensor measurements. As shown, the distribution of the datasets for training and testing is similar.
Figure 2. Visualization of the distributions of the training and test datasets based on kernel density estimation (KDE) to check if the split is stratified. The x-axis is the range of values in the dataset, and y-axis represents its probability density, which captures the underlying distribution of the sensor measurements. As shown, the distribution of the datasets for training and testing is similar.
Sensors 19 03296 g002
Figure 3. Density plots of variables in the training and test datasets with respect to classes. Differences in distributions among variables can be observed. The class index listed in the legend means the class index defined in Table 2.
Figure 3. Density plots of variables in the training and test datasets with respect to classes. Differences in distributions among variables can be observed. The class index listed in the legend means the class index defined in Table 2.
Sensors 19 03296 g003
Figure 4. Example of the 1-D CNN model used for this study. Depending on the required performance and inference time, type/combination of input data and kernel size and numbers used at each layer may change.
Figure 4. Example of the 1-D CNN model used for this study. Depending on the required performance and inference time, type/combination of input data and kernel size and numbers used at each layer may change.
Sensors 19 03296 g004
Figure 5. Recurrent neural networks (RNN)-based architecture used for the proposed motion context classification. An RNN cell is stacked on top of another one to learn more complex representations. In the experiment, long short-term memory (LSTM) and gated recurrent unit (GRU) cells were employed.
Figure 5. Recurrent neural networks (RNN)-based architecture used for the proposed motion context classification. An RNN cell is stacked on top of another one to learn more complex representations. In the experiment, long short-term memory (LSTM) and gated recurrent unit (GRU) cells were employed.
Sensors 19 03296 g005
Figure 6. Confusion matrices of the classification results. Many misclassifications are between c2 (powder) and c3 (melt-freeze crust).
Figure 6. Confusion matrices of the classification results. Many misclassifications are between c2 (powder) and c3 (melt-freeze crust).
Sensors 19 03296 g006
Figure 7. Inference time required for each signal (7 × 100 samples). From the left, the graph shows the time spent on PC (×86_64), NVIDIA Jetson TX2 (aarch64), and Raspberry Pi 3 Model B+ (armv7l).
Figure 7. Inference time required for each signal (7 × 100 samples). From the left, the graph shows the time spent on PC (×86_64), NVIDIA Jetson TX2 (aarch64), and Raspberry Pi 3 Model B+ (armv7l).
Sensors 19 03296 g007
Table 1. Weather information of the ski resort.
Table 1. Weather information of the ski resort.
DateAverage TemperatureRelative HumidityDaily PrecipitationOverall Quality
Day #17 February 2018−10.6 °C (12.9 °F)25.6%-Powder
Day #223 February 20181.9 °C (32.4 °F)81.1%5.2 mmMelt-freeze crust
Day #326 February 20183.1 °C (37.6 °F)43.1%-Slush
Table 2. Class used in the experiment.
Table 2. Class used in the experiment.
Class IndexMotion Context (and Overall Surface Quality)
Class #0Sitting on the lift
Class #1Standing on the escalator
Class #2Downhill on Day #1 (powder)
Class #3Downhill on Day #2 (melt-freeze crust)
Class #4Downhill on Day #3 (slush)
Table 3. Signals used in this study.
Table 3. Signals used in this study.
Sequence NameDescriptions
measuredacceleration_Xlinear acceleration in X axis in m/s2
acceleration_Ylinear acceleration in Y axis in m/s2
acceleration_Zlinear acceleration in Z axis in m/s2
angular_velocity_Xangular velocity in X axis in rad/s
angular_velocity_Yangular velocity in Y axis in rad/s
angular_velocity_Zangular velocity in Z axis in rad/s
vibrationvibration intensity values
generatedtotal_linear_accelerationroot mean square of the linear acceleration in m/s2
total_angular_velocityroot mean square of the angular velocity in rad/s
acc_vs_velratio of total_linear_acceleration and total_angular_velocity in m/(rad·s)
Table 4. Features explored in this study.
Table 4. Features explored in this study.
Feature NameDescriptions
meanarithmetic mean (average)
medianmedian
min (max)minimum (maximum)
max/minratio of max and min
stdstandard deviation
skewsample skewness
abs_minminimum of absolute value
abs_maxmaximum of absolute value
abs_meanarithmetic mean of absolute value
abs_stdstandard deviation of absolute value
Table 5. Experimental results—Test accuracy (%).
Table 5. Experimental results—Test accuracy (%).
Input SignalsDim.RF1D-CNNGRULSTM
Acc XYZ377.2577.0380.6077.69
Gyro XYZ383.0192.4192.3883.66
Acc + Gyro691.9196.1196.2495.63
Acc + Vib487.2587.6790.3385.55
Acc + Gyro Y491.1395.2295.4494.78
Acc X + Gyro491.1294.4496.3394.11
Acc + Gyro Y + Vib593.5394.8995.0094.22
All793.4496.2295.6797.00

Share and Cite

MDPI and ACS Style

Han, B.-K.; Ryu, J.-K.; Kim, S.-C. Context-Aware Winter Sports Based on Multivariate Sequence Learning. Sensors 2019, 19, 3296. https://doi.org/10.3390/s19153296

AMA Style

Han B-K, Ryu J-K, Kim S-C. Context-Aware Winter Sports Based on Multivariate Sequence Learning. Sensors. 2019; 19(15):3296. https://doi.org/10.3390/s19153296

Chicago/Turabian Style

Han, Byung-Kil, Je-Kwang Ryu, and Seung-Chan Kim. 2019. "Context-Aware Winter Sports Based on Multivariate Sequence Learning" Sensors 19, no. 15: 3296. https://doi.org/10.3390/s19153296

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop