Next Article in Journal
Thermosonication Process Design for Recovering Bioactive Compounds from Fennel: A Comparative Study with Conventional Extraction Techniques
Next Article in Special Issue
Comparison between Helical Axis and SARA Approaches for the Estimation of Functional Joint Axes on Multi-Body Modeling Data
Previous Article in Journal
Heterogeneous Electro-Fenton-like Designs for the Disposal of 2-Phenylphenol from Water
Previous Article in Special Issue
Hamstring Torque, Velocity and Power Elastic Band Measurements during Hip Extension and Knee Flexion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Upper Body Posture Recognition Using Inertial Sensors and Recurrent Neural Networks

1
Department of Electrical and Computer Engineering, University of California, La Jolla, San Diego, CA 92092, USA
2
School of Biomedical Engineering, College of Biomedical Engineering, Taipei Medical University, Taipei 11031, Taiwan
3
Department of Orthopedics, Shuang Ho Hospital, Taipei Medical University, Taipei 23561, Taiwan
4
Department of Orthopedics Surgery, School of Medicine, College of Medicine, Taipei Medical University, Taipei 23561, Taiwan
5
Center for Biomedical Engineering, Graduate Institute of Biomedical Engineering, College of Engineering, Chang Gung University, Taoyuan 33302, Taiwan
6
Bone and Joint Research Center, Department of Orthopedic Surgery, Chang Gung Memorial Hospital Linkou Branch, Taoyuan 33305, Taiwan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(24), 12101; https://doi.org/10.3390/app112412101
Submission received: 30 September 2021 / Revised: 11 December 2021 / Accepted: 14 December 2021 / Published: 19 December 2021
(This article belongs to the Special Issue Biomechanics and Human Motion Analysis)

Abstract

:

Featured Application

In this study, a wearable system that can recognize human posture was developed. By using long short-term memory-based recurrent neural network (LSTM-RNN) architecture, this system was able to classify posture with data measured by using an inertial measurement unit (IMU). Our results can serve as a reference for future developments of wearable systems in order to correct human posture and mitigate risks of spinal deformity.

Abstract

Inadequate sitting posture can cause imbalanced loading on the spine and result in abnormal spinal pressure, which serves as the main risk factor contributing to irreversible and chronic spinal deformity. Therefore, sitting posture recognition is important for understanding people’s sitting behaviors and for correcting inadequate postures. Recently, wearable devices embedded with microelectromechanical systems (MEMs) sensors, such as inertial measurement units (IMUs), have received increased attention in human activity recognition. In this study, a wearable device embedded with IMUs and a machine learning algorithm were developed to classify seven static sitting postures: upright, slump, lean, right and left bending, and right and left twisting. Four 9-axis IMUs were uniformly distributed between thoracic and lumbar regions (T1-L5) and aligned on a sagittal plane to acquire kinematic information about subjects’ backs during static-dynamic alternating motions. Time-domain features served as inputs to a signal-based classification model that was developed using long short-term memory-based recurrent neural network (LSTM-RNN) architecture, and the model’s classification performance was used to evaluate the relevance between sensor signals and sitting postures. Overall results from performance evaluation tests indicate that this IMU-based measurement and LSTM-RNN structural scheme was appropriate for sitting posture recognition.

1. Introduction

The spine is an important bony structure that provides support to the human body while also providing other main functions, such as protecting the spinal cord and nerve roots, resisting external forces, and enabling the performance of all human body movements. Among five spine segments, the thoracic spine consists of 12 vertebrae (T1 to T12) and is the largest segmentof the spine. Its kyphotic curve allows the spine to bear loads anteriorly and to resist tension posteriorly, protecting the spinal cord while moving or bending the body. Since the thoracic spine provides most of the stability and support for the entire trunk [1], spine deformity resulting from poor posture often occurs in this segment.
Spinal deformities are abnormal alignments or curves of the vertebral column resulting from uneven loading on the contralateral side, with many contributing factors. For example, poor posture can cause acute or chronic forms of spinal deformities. Human body posture is primarily maintained by the musculoskeletal system, which includes the spine as an important component. In order to maintain body balance and stability, the relative positions of different spine segments are rearranged, resulting in changes in the spinal curve, and this type of structural support provided by the spine can result in some degree of deformity [2]. Spinal deformities can be divided into kyphosis, lordosis, and scoliosis, and the degree of deformation is often determined by X-ray radiography and Cobb angle measurements in clinical practice [3]. Poor posture refers to postures that increase joint stress [4] and is often the main cause of spinal deformities. According to Griegel-Morris et al. [5], poor posture is prevalent among healthy people. People with more severe postural abnormalities increase their risk of experiencing muscle pain. In addition, people suffering from chronic pain exhibit poor control over maintaining an upright posture while sitting and tend to unconsciously change their sitting postures to poor postures over time [6], resulting in a vicious cycle of pain and poor posture.
Several devices, including optoelectronic motion analysis systems [7,8], radiography [9,10], and pressure sensors [11,12], are used for monitoring and measuring posture in research and clinical practice. However, the aforementioned methods require the following: (1) specific sites, such as laboratories; (2) technical personnel for data collection; (3) lengthy periods for data analysis; and (4) high system setup costs [13]. Thus, the use of miniaturized monitoring devices [14,15,16,17] has grown rapidly in recent years due to their convenience. In addition, these devices can quickly analyze posture changes and correct poor postures, achieving long-term posture correction. Wong et al. [17] used three tri-axial accelerometers to monitor sitting postural changes and compared the results with those obtained from an optoelectronic motion analysis system. For postural measurements, the RMS error was approximately 2° in quasi-static conditions, while the error was relatively large in dynamic conditions. Therefore, a gyroscope was added to the system in order to reduce errors in collecting dynamic information. In addition, Petropoulos et al. [14] used two inertial measurement units (IMUs) to perform binary classifications of normal and abnormal postures. Their results showed that the mean square error was approximately 0.15 for the pitch angle, and the heading angle was within ±10° range of motion, indicating that collected IMU signals can be used to calculate back curvature. However, this system is only suitable for monitoring sitting posture under static conditions, since it has only been confirmed for small-angle motions. Even though the systems mentioned above demonstrated high accuracy for posture recognition in static conditions, recognition of postures under dynamic conditions still needs to be improved.
Human activity recognition (HAR) aims to identify specific movements based on different sensing signals. This technique has been developed for multiple applications, such as fall detection systems, disease prevention, the Internet of Medical Things (IoMT), etc. [18]. In addition, smart wearable devices present an application of HAR for collecting kinematic parameters of the human body by using sensors. Sensor data are used to identify human activity with the aid of deep learning and other machine learning algorithms. According to Demrozi et al. [19], although many researchers prefer classic machine learning (CML) models because of their small datasets and lower dimensionality of input data, deep learning (DL) models exhibit greater accuracy in large activity datasets because DL networks are capable of feature learning or automatic feature extraction. In addition, DL networks have shown the ability to recognize human activities, such as walking, climbing ladders, and falling [20]. Therefore, HAR research that uses DL models, such as the Convolutional Neural Network (CNN) [21,22,23], the Long Short-Term Memory Network (LSTM) [24,25,26], and the Recurrent Neural Network (RNN) [27,28], has recently increased in popularity.
In this study, we aim to recognize human unsupported sitting postures by using a deep learning algorithm. IMUs were uniformly distributed along subjects’ spines in order to collect kinematic parameters of subjects’ backs in static and dynamic conditions. For the deep learning algorithm, an LSTM-based RNN (LSTM-RNN) was used to classify seven common unsupported sitting postures by using raw data from IMUs as inputs. With the aim of preventing chronic spinal problems, this system may be useful for long-term tracking of sedentary people as well as for providing alerts regarding poor posture.

2. Materials and Methods

2.1. Description of Sitting Posture

Unsupported sitting is often performed in human activity research in clinical studies because of musculoskeletal and neural system involvement in human activity [29]. The human body relies on interactions between the musculoskeletal and neural systems in the trunk to maintain balance under static and dynamic conditions. In this study, seven common sitting postures, which were defined as single-axis rotations related to spinal deformities including upright, slump, lean, right and left bending, and right and left twisting postures, were selected for subjects to perform (Figure 1).

2.2. Sensing Device

A Next Generation IMU (NGIMU, x-io Technologies, Bristol, UK) was used to collect kinematic data. This NGIMU is a 9-axis MEMs sensor that consists of a 3-axis accelerometer, 3-axis gyroscope, and 3-axis magnetometer and features an onboard attitude and heading reference system (AHRS) sensor fusion algorithm, which provides real-time spatial data on outputs, including quaternions and Euler angles. Sensor outputs included quaternions, rotation matrix, Euler angles (0~2000º/s), linear acceleration (0~16 g), and compass heading information (0~1300 μT). For the sitting posture monitoring device, 4 IMUs were uniformly distributed between thoracic and lumbar segments (T1-L5) and tightly fixed on cloth [14,15,17,30] along the sagittal plane (Figure 2). The IMU’s sampling rate was 50 Hz, which is sufficient for HAR tasks [31].
In addition, the relevance between the number of IMU sensors used in this device and the machine learning classifier’s performance was also investigated in this study. The number of IMUs used in the device was reduced to three sensors (1-3 and 2-4 as a group), two sensors (1-2, 2-3, and 3-4 as a group), and one single sensor, respectively, during human trials. All collected data were used to train the same machine learning model, and performance was evaluated by comparing F1 scores (Equation (1)) between different sensor combinations, where the F1 score is the harmonic mean of precision (TP/(TP + FP) and recall (TP/(TP + FN), TP is the number of true positives, FP is the number of false positives, and FN is the number of false negatives [32]. F1 scores are useful for evaluating machine learning systems, especially in multi-class classification. Accuracy was defined as the ratio of true positives to all samples (Equation (2)).
F 1 score = 2 × precision × recall precision + recall
Accuracy = TP + FN TP + TN + FP + FN

2.3. Procedure of Human Trials

This study involved 6 healthy adults (4 males and 2 females) with ages ranging from 20 to 65. Subjects were required to fulfill the following criteria: no critical illness within the previous year, no neuro-musculoskeletal injury within the previous year, and no recent spinal treatment. The study was conducted with the approval of the Office of Human Research designated by Taipei Medical University, Taiwan (TMU-JIRB N201802061).
Subjects were asked to sit on a chair with their backs straightened. They were then asked to follow instructions that were displayed on a screen in order to perform seven unsupported sitting postures. MATLAB (Mathworks, Natick, MA, USA) was used to guide subjects through a series of images of sitting postures, which changed every 5 s. Subjects were instructed to maintain the displayed posture until the next posture appeared on the screen. In addition, subjects were allowed to move without constraints with respect to speed, amplitude, and path of movement during posture changes. This lack of constraint in trial procedures is essential for the functionality of portable devices; for example, differences in velocity, travel path, and moved distance between trials add randomness to training datasets, thus underscoring the practicality of portable devices in daily activities, which usually exhibit few movement constraints. In total, 28 postures were performed in a single trial, with each posture displayed four times. Postures were randomly displayed in order to improve generalization in machine learning. Nine-axis IMU signals were collected by using a sampling rate of 50Hz and transmitted via Wi-Fi.

2.4. Data Labeling

In previous studies of human activity recognition, researchers manually labeled data according to video recordings, but this method was time consuming and manpower consuming. Hence, in order to simplify the data labeling process, accelerometer data for supervised learning were labeled by using a designated trial protocol. As described in the previous section, each trial contained 28 alternative postures and a 5-s period for each posture. The presentation order of each posture was also recorded. Within the 5-s period, movements were divided into two states: static, which was defined as holding a posture, and dynamic, which was defined as changing postures. Time rates of acceleration signals were compared to a threshold in order to determine these two motion states. Recognizing such regular transitions, the labeling task divided an entire trial into serial sub-problems with 5-s durations. Results and data labeling in our dataset build-up were automatically generated. When compared to manual labeling with video recordings, this protocol is more efficient, but the obscure “labeling midpoint” in motion transitions, e.g., changing movements from slump postures to lean postures, was difficult to define. Upright posture served as midpoints for changing phases as subjects could not avoid being upright when transitioning from slump postures to lean postures or from left twisting postures to right twisting postures. This caused the determination of upright postures to be more difficult than the determination of other postures. For simplification, in motion-changing phases, a dynamic period was separated into two classes by the ratio of previous and next static phases, and respective labels were assigned to each part. This provided more randomness to machine learning classifiers in order to improve the generalization of its real-world application.

2.5. Machine Learning Algorithm

For the machine learning algorithm, a recurrent neural network (RNN) was used as the architecture for the machine learning model. RNNs contain cyclic connections that make them more powerful in modeling sequence data when compared to feedforward neural networks. RNNs have been widely used in language modeling, objective detection, and speech recognition and have received increased attention in biomechanics. RNNs are often used to construct sequence-to-sequence models and have been applied to HAR tasks because of RNNs’ competence in temporal domain processing [24,33,34,35,36]. RNNs can transition between hidden states of previous time steps to subsequent time steps, processing and passing down time-dependent sequences in chronological order. Thus, sequences obtained from time-dependent signals are suitable for RNN inputs. However, as a result of long-term dependency [37], RNN performance decreases as input sequences increase. In order to solve this problem, a long short-term memory (LSTM) cell was designed. An LSTM contains four units in its structure, including an input gate, an output gate, a memory cell, and a forget gate, and can decide whether input data can be stored in memory cells, effectively solving long-term dependencies that occur in RNNs [38,39].
A long short-term memory (LSTM)-based RNN was used to improve information extraction in the temporal domain (Figure 3). Two LSTM-based layers were present in the model structure [25,32]. Each contained 128 LSTM units, with a rectified linear unit (ReLu) applied as an activation function. The experimental setup for the training procedure is listed in Table 1. Data were trained on GPU for 600 epochs by using an Adam optimizer and 10-fold cross validation to prevent overfitting. One epoch occurs when an entire dataset is passed forward and backward through the neural network only once. An Adam optimizer is an algorithm for first-order gradient-based optimization of stochastic objective functions. Cross validation is a technique for evaluating predictive models by partitioning original samples into a training set to train models and a test set to evaluate them. Raw data from IMU sensors were used as model inputs. The model output was one of seven classes representing different sitting postures.

3. Results

Six healthy adults (four males and two females), aged 20~65, were involved in this study to simulate seven sitting postures in daily life: upright, slump, lean, right and left bending, and right and left twisting postures. Raw signals were collected by using IMUs during one single trial in each corresponding axis (Figure 4). Data labeling showed serial transitions during trials (Figure 5).
The confusion matrix of classification results was shown for the LSTM-RNN machine learning classifier (Figure 6). For seven sitting postures, the model’s average accuracy increased to 99.0 ± 0.3% in 600 epochs, and the average F1 score was approximately 0.966 ± 0.012 (Table 2). In order to evaluate the performance of the trained model, a testing sequence was input with labels. According to the leave-one-subject-out (LOSO) method, test sets consisted of persons not included in the training. Training and validation sets were created by using a 95% to 5% splitting [32]. The unused dataset including the sixth subject was tested (Figure 7), resulting in a model accuracy of 81.2%. Results showed the practicality and potential of IMUs-based applications using LSTM-RNNs in sitting posture detection.
In order to determine the optimal number of sensors, we evaluated the relevance between the numbers of sensors and the classifier’s performance. After excluding specific sensors, the classifier’s performance was compared with the performance of the four-sensor IMU, which was designated as the control group in this test. According to the comparison (Figure 8), model accuracy decreased as the number of sensors was reduced. The combination of three sensors covering the thoracic area (B1~B3) exhibited an accuracy of 98.1%.

4. Discussion

This study demonstrated a process of data labeling and investigated unconstrained movements to improve HAR research. Sitting postures investigated in this study are sustained for a relatively long time in daily life. Hence, this protocol not only simplifies time-consuming procedures but is also appropriate for sitting postures detection. In Table 2, the highest and lowest accuracies were 99.3 and 98.3% for lean and upright postures, respectively. These results show the importance of postural features in this application. Upright postures, as described in the data labeling section, served as the “labeling mid-point” for many posture-changing phases, and our model showed the poorest accuracy on this class. In contrast, holding lean or bending postures results in large tilt/bending angles in sagittal and coronal planes, respectively, and these changes allow IMUs to collect detectable changes when compared to other postures.
Accelerometers have been widely used in human motion analysis [16,17,19]. In this study, we used acceleration data from IMUs to conduct data labeling. Prior to processing classification, 28 intermittent phases were extracted by using a static–dynamic alternating sequence with approximately 5 s each. During posture transition, the signal was divided into two parts and assigned to the previous posture and the next posture, respectively, in order to enhance the model’s general application. However, this form of labeling for such transition phases is imprecise and causes misclassifications for obscure definitions of posture, e.g., upright postures during changes from right bending to lean postures, as the third green rectangle shows in Figure 8. For real-world sitting posture correction devices, upright postures are more important to classify than compared to other postures. Misidentification of upright postures (such as when transitioning from one posture to another) may be an issue with our labeling method and can be fixed by adding more explanations about upright postures before trials.
In this study, we used an LSTM-RNN to conduct sequence-to-sequence classification. High classifier accuracy indicates the effectiveness of this IMUs-RNN scheme for sitting posture detection. In addition, the model exhibited a higher classification accuracy for flexion/extension (lean postures) than for axial rotation (twisting postures) due to the larger motion range of sensors during flexion/extension. This is related to the sensors’ axis; twisting supplies rotation to all sensors along our axis, especially the lowest sensor. However, signals for other axes did not change significantly when compared to other postures, and this small difference caused relatively poorer performance in these classes.
However, this model showed slight overfitting. As shown in Figure 7, when subjects were in a static state, the model recognized postures more accurately when compared to dynamic transition states as mentioned by Rivera et al. [40]. Some techniques of feature extraction were suggested for solving this problem; for example, an averaging filter can condense signals in dynamic periods with parts of signals in static periods to improve the accuracy of a shorter period. Further investigations of feature selection or preprocessing are not discussed in this article. In addition, even though seven postures were performed by one single subject, differences of velocity, travel path, and moved distance between trials added randomness to the training dataset. Due to the fact that movements were unconstrained and, therefore, natural and despite a sample of five research subjects, we believe that the resulting data represent a significant portion of the range of normal spinal postures, which is, of course, limited to a certain extent.
In addition, we observed the importance of the “order” of postures in trials; more specifically, postures around (changed from or into) slump and upright postures caused relatively greater classification error, as shown in green rectangles in Figure 7. This type of error was caused by the order of labeling procedures. It is difficult to define the changing point from right bending to slump postures (around the 1000th sample in Figure 7) because upright postures are similar to the combination of these two postures. Consequently, our classifier assigned upright postures to this changing period rather than slump postures. A more appropriate method of defining ground truth is using fuzzy logic, since most human body motions are dynamic or “quasi-static” rather than totally static, i.e., no change in signals. Changing postures occur frequently in daily activity, and it is impossible for changes from one posture to another to occur discontinuously, as our ground truth shows in Figure 7. In addition, as a result of anatomical differences between people and unconstrained trials, upright postures become more difficult to differentiate from all other postures, and misclassification error is higher than in other classes as illustrated by accuracy and F1 scores in Table 2.
For the number of sensors, the average classification accuracy was reduced with a decreasing number of sensors and dimensionality of training features. Accuracy and F1 scores were significantly reduced by using fewer than three sensors, demonstrating that a minimum number of three sensors were required to cope with posture recognition, especially in dynamic conditions. Among devices using three sensors, the three-sensor series (B1~B3) positioned at the upper trunk displayed greater accuracy than the three-sensor series positioned at the lower trunk (B2~B4). This is probably because the B4 sensor was positioned at L5, which remained almost static between coplanar rotations, such as right and left bending [41]. Conversely, the upper arrangement of sensors B1~B3 exhibited a higher accuracy of 98.1%, which was comparable to the accuracy of the device using four sensors, indicating an optimal arrangement in the sensing system.
This study has some limitations. Firstly, we evaluated only six subjects. With more subjects, it may be possible to match user-independent subjects with a subset of overall subjects based on weight, height, and various habits of activities. This would likely allow for improved user-independent classification performance. Secondly, the average accuracy when training and testing by using 10-fold cross validation was 99.0%. When testing with the excluded subject, average accuracy decreased to 81.2%. Similar variations were also reported in other papers related to human motion classification [42,43]; similarly, our model seems to perform adequately when recognizing similar patterns but performs imperfectly when used on all segments of a population. Further studies will be conducted on more postures and more subjects in order to develop a neural network that can achieve comparable generalization performances.

5. Conclusions

In this study, a wearable device was developed for sitting posture monitoring by combining IMU technology with a machine learning algorithm. An LSTM-based recurrent neural network was used to classify seven common sitting postures by inputting raw signals collected by using IMUs during human trials. The results showed that this system can effectively distinguish different sitting postures. In addition, a strategy for reducing the number of sensors is suggested, since fewer sensors would result in improved power management in such devices. The recommended number of sensors is three, which still provides good performance without compromising model robustness. With further research, this procedure may be improved by using faster automatic methods for data processing as well as by including more data for training and augmenting the database.

Author Contributions

Conceptualization, H.-H.C., H.-Y.T., S.-H.T., C.-J.C. and T.-Y.S.; methodology, H.-H.C. and H.-Y.T.; software, H.-Y.T. and S.-H.T.; validation, H.-Y.T. and S.-H.T.; formal analysis, H.-Y.T.; investigation, H.-H.C., H.-Y.T., S.-H.T. and T.-Y.S.; resources, C.-J.C. and H.-H.C.; data curation, H.-Y.T. and S.-H.T.; writing—original draft preparation, H.-Y.T. and S.-H.T.; writing—review and editing, H.-Y.T., S.-H.T., H.-H.C. and T.-Y.S.; visualization, H.-Y.T. and S.-H.T.; supervision, H.-Y.T. and H.-H.C.; project administration, C.-J.C. and H.-H.C.; funding acquisition, H.-H.C. and C.-J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology of Taiwan (MOST 107-2221-E-038-015 and MOST 109-2221-E-038-007).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Taipei Medical University.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data of this research are available to the public in the Open Science Framework at https://osf.io/2fzvm/files/.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Frost, B.A.; Camarero-Espinosa, S.; Foster, E.J. Materials for the spine: Anatomy, problems, and solutions. Materials 2019, 12, 253. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Morrissy, R.; Goldsmith, G.; Hall, E.; Kehl, D.; Cowie, G. Measurement of the Cobb angle on radiographs of patients who have. J Bone Jt. Surg. Am. 1990, 72, 320–327. [Google Scholar] [CrossRef] [Green Version]
  3. Tanure, M.C.; Pinheiro, A.P.; Oliveira, A.S. Reliability assessment of Cobb angle measurements using manual and digital methods. Spine J. 2010, 10, 769–774. [Google Scholar] [CrossRef] [PubMed]
  4. Kendall, H.O.; Kendall, F.P.; Wadsworth, G.E. Muscles, testing and function. Am. J. Phys. Med. Rehabil. 1973, 52, 43. [Google Scholar]
  5. Griegel-Morris, P.; Larson, K.; Mueller-Klaus, K.; Oatis, C.A. Incidence of common postural abnormalities in the cervical, shoulder, and thoracic regions and their association with pain in two age groups of healthy subjects. Phys. Ther. 1992, 72, 425–431. [Google Scholar] [CrossRef]
  6. Falla, D.; Jull, G.; Russell, T.; Vicenzino, B.; Hodges, P. Effect of neck exercise on sitting posture in patients with chronic neck pain. Phys. Ther. 2007, 87, 408–417. [Google Scholar] [CrossRef]
  7. Abbott, J.H. A novel approach to managing graduated return to spinal loading in patients with low back pain using the Spineangel® device: A case series report. N. Z. J. Physiother. 2008, 36, 22. [Google Scholar]
  8. Takayanagi, K.; Takahashi, K.; Yamagata, M.; Moriya, H.; Kitahara, H.; Tamaki, T. Using cineradiography for continuous dynamic-motion analysis of the lumbar spine. Spine 2001, 26, 1858–1865. [Google Scholar] [CrossRef]
  9. Foubert, N.; McKee, A.M.; Goubran, R.A.; Knoefel, F. Lying and sitting posture recognition and transition detection using a pressure sensor array. In Proceedings of the 2012 IEEE International Symposium on Medical Measurements and Applications Proceedings, Budapest, Hungary, 18–19 May 2012; pp. 1–6. [Google Scholar]
  10. De Carvalho, D.E.; Soave, D.; Ross, K.; Callaghan, J.P. Lumbar spine and pelvic posture between standing and sitting: A radiologic investigation including reliability and repeatability of the lumbar lordosis measure. J. Manip. Physiol. Ther. 2010, 33, 48–55. [Google Scholar] [CrossRef]
  11. O’Sullivan, K.; O’Sullivan, L.; Campbell, A.; O’Sullivan, P.; Dankaerts, W. Towards monitoring lumbo-pelvic posture in real-life situations: Concurrent validity of a novel posture monitor and a traditional laboratory-based motion analysis system. Man. Ther. 2012, 17, 77–83. [Google Scholar] [CrossRef]
  12. Xu, W.; Li, Z.; Huang, M.-C.; Amini, N.; Sarrafzadeh, M. ecushion: An etextile device for sitting posture monitoring. In Proceedings of the 2011 International Conference on Body Sensor Networks, Dallas, TX, USA, 23–25 May 2011; pp. 194–199. [Google Scholar]
  13. Straker, L.M.; O’Sullivan, P.B.; Smith, A.J.; Perry, M.C. Relationships between prolonged neck/shoulder pain and sitting spinal posture in male and female adolescents. Man. Ther. 2009, 14, 321–329. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Petropoulos, A.; Sikeridis, D.; Antonakopoulos, T. Wearable smart health advisors: An IMU-enabled posture monitor. IEEE Consum. Electron. Mag. 2020, 9, 20–27. [Google Scholar] [CrossRef]
  15. Petropoulos, A.; Sikeridis, D.; Antonakopoulos, T. SPoMo: IMU-based real-time sitting posture monitoring. In Proceedings of the 2017 IEEE 7th International Conference on Consumer Electronics-Berlin (ICCE-Berlin), Berlin, Germany, 3–6 September 2017; pp. 5–9. [Google Scholar]
  16. Ma, S.; Cho, W.-H.; Quan, C.-H.; Lee, S. A sitting posture recognition system based on 3 axis accelerometer. In Proceedings of the 2016 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Chiang Mai, Thailand, 5–7 October 2016; pp. 1–3. [Google Scholar]
  17. Wong, W.Y.; Wong, M.S. Detecting spinal posture change in sitting positions with tri-axial accelerometers. Gait Posture 2008, 27, 168–171. [Google Scholar] [CrossRef] [PubMed]
  18. Sousa Lima, W.; Souto, E.; El-Khatib, K.; Jalali, R.; Gama, J. Human activity recognition using inertial sensors in a smartphone: An overview. Sensors 2019, 19, 3213. [Google Scholar] [CrossRef] [Green Version]
  19. Demrozi, F.; Pravadelli, G.; Bihorac, A.; Rashidi, P. Human activity recognition using inertial, physiological and environmental sensors: A comprehensive survey. IEEE Access 2020, 8, 210816–210836. [Google Scholar] [CrossRef]
  20. Ronao, C.A.; Cho, S.-B. Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst. Appl. 2016, 59, 235–244. [Google Scholar] [CrossRef]
  21. Wang, K.; He, J.; Zhang, L. Attention-based convolutional neural network for weakly labeled human activities’ recognition with wearable sensors. IEEE Sens. J. 2019, 19, 7598–7604. [Google Scholar] [CrossRef] [Green Version]
  22. Santos, G.L.; Endo, P.T.; Monteiro, K.H.d.C.; Rocha, E.d.S.; Silva, I.; Lynn, T. Accelerometer-based human fall detection using convolutional neural networks. Sensors 2019, 19, 1644. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Sun, J.; Fu, Y.; Li, S.; He, J.; Xu, C.; Tan, L. Sequential human activity recognition based on deep convolutional network and extreme learning machine using wearable sensors. J. Sens. 2018, 2018, 8580959. [Google Scholar] [CrossRef]
  24. Pienaar, S.W.; Malekian, R. Human activity recognition using LSTM-RNN deep neural network architecture. In Proceedings of the 2019 IEEE 2nd wireless africa conference (WAC), Pretoria, South Africa, 18–20 August 2019; pp. 1–5. [Google Scholar]
  25. Steven Eyobu, O.; Han, D.S. Feature representation and data augmentation for human activity classification based on wearable IMU sensor data using a deep LSTM neural network. Sensors 2018, 18, 2892. [Google Scholar] [CrossRef] [Green Version]
  26. Ordóñez, F.J.; Roggen, D. Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Inoue, M.; Inoue, S.; Nishida, T. Deep recurrent neural network for mobile human activity recognition with high throughput. Artif. Life Robot. 2018, 23, 173–185. [Google Scholar] [CrossRef] [Green Version]
  28. Murad, A.; Pyun, J.-Y. Deep recurrent neural networks for human activity recognition. Sensors 2017, 17, 2556. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Harvey, L.A.; Ristev, D.; Hossain, M.S.; Hossain, M.A.; Bowden, J.L.; Boswell-Ruys, C.L.; Hossain, M.M.; Ben, M. Training unsupported sitting does not improve ability to sit in people with recently acquired paraplegia: A randomised trial. J. Physiother. 2011, 57, 83–90. [Google Scholar] [CrossRef] [Green Version]
  30. Severin, I.-C. Head Posture monitor based on 3 IMU sensors: Consideration toward Healthcare application. In Proceedings of the 2020 International Conference on e-Health and Bioengineering (EHB), Iasi, Romania, 29–30 October 2020; pp. 1–4. [Google Scholar]
  31. Guo, M.; Wang, Z. Segmentation and recognition of human motion sequences using wearable inertial sensors. Multimed. Tools Appl. 2018, 77, 21201–21220. [Google Scholar] [CrossRef]
  32. Zimmermann, T.; Taetz, B.; Bleser, G. IMU-to-segment assignment and orientation alignment for the lower body using deep learning. Sensors 2018, 18, 302. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Brugarolas, R.; Roberts, D.; Sherman, B.; Bozkurt, A. Posture estimation for a canine machine interface based training system. In Proceedings of the 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Diego, CA, USA, 28 August–1 September 2012; pp. 4489–4492. [Google Scholar]
  34. Senyurek, V.Y.; Imtiaz, M.H.; Belsare, P.; Tiffany, S.; Sazonov, E. A CNN-LSTM neural network for recognition of puffing in smoking episodes using wearable sensors. Biomed. Eng. Lett. 2020, 10, 195–203. [Google Scholar] [CrossRef]
  35. Hu, Y.; Zhang, X.-Q.; Xu, L.; He, F.X.; Tian, Z.; She, W.; Liu, W. Harmonic loss function for sensor-based human activity recognition based on LSTM recurrent neural networks. IEEE Access 2020, 8, 135617–135627. [Google Scholar] [CrossRef]
  36. Albaba, M.; Qassab, A.; YILMAZ, A. Human activity recognition and classification using of convolutional neural networks and recurrent neural networks. Int. J. Appl. Math. Electron. Comput. 2020, 8, 185–189. [Google Scholar] [CrossRef]
  37. Schaefer, A.M.; Udluft, S.; Zimmermann, H.-G. Learning long-term dependencies with recurrent neural networks. Neurocomputing 2008, 71, 2481–2488. [Google Scholar] [CrossRef]
  38. Zheng, J.; Xu, C.; Zhang, Z.; Li, X. Electric load forecasting in smart grids using long-short-term-memory based recurrent neural network. In Proceedings of the 2017 51st Annual Conference on Information Sciences and Systems (CISS), Baltimore, MD, USA, 22–24 March 2017; pp. 1–6. [Google Scholar]
  39. Zhao, R.; Wang, J.; Yan, R.; Mao, K. Machine health monitoring with LSTM networks. In Proceedings of the 2016 10th international conference on sensing technology (ICST), Nanjing, China, 11–13 November 2016; pp. 1–6. [Google Scholar]
  40. Rivera, P.; Valarezo, E.; Choi, M.-T.; Kim, T.-S. Recognition of human hand activities based on a single wrist imu using recurrent neural networks. Int. J. Pharma Med. Biol. Sci 2017, 6, 114–118. [Google Scholar] [CrossRef]
  41. Rousseau, M.-A.; Bradford, D.S.; Hadi, T.M.; Pedersen, K.L.; Lotz, J.C. The instant axis of rotation influences facet forces at L5/S1 during flexion/extension and lateral bending. Eur. Spine J. 2006, 15, 299–307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Beil, J.; Ehrenberger, I.; Scherer, C.; Mandery, C.; Asfour, T. Human Motion Classification Based on Multi-Modal Sensor Data for Lower Limb Exoskeletons. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 5431–5436. [Google Scholar]
  43. Rizwan, A.; Ali, N.A.; Zoha, A.; Ozturk, M.; Alomaniy, A.; Imran, M.A.; Abbasi, Q.H. “Non-invasive hydration level estimation in human body using galvanic skin response. IEEE Sens. J. 2020, 20, 4891–4900. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Seven kinds of unsupported sitting postures performed in the experiment: (a) slump, (b) upright, (c) lean, (d) right bending, (e) left bending, (f) left twisting, and (g) right twisting.
Figure 1. Seven kinds of unsupported sitting postures performed in the experiment: (a) slump, (b) upright, (c) lean, (d) right bending, (e) left bending, (f) left twisting, and (g) right twisting.
Applsci 11 12101 g001
Figure 2. (a) Nine-axis inertial measurement units (NGIMU); (b) distribution of IMUs where blue dots represent positions of IMUs relative to subjects’ spines: B1 was positioned at T1, B4 was positioned at L5, and B2 and B3 were evenly distributed between S1 and S4; and (c) prototype of the sitting posture recognition device.
Figure 2. (a) Nine-axis inertial measurement units (NGIMU); (b) distribution of IMUs where blue dots represent positions of IMUs relative to subjects’ spines: B1 was positioned at T1, B4 was positioned at L5, and B2 and B3 were evenly distributed between S1 and S4; and (c) prototype of the sitting posture recognition device.
Applsci 11 12101 g002
Figure 3. Schematic diagram of the LSTM-RNN machine learning classifier.
Figure 3. Schematic diagram of the LSTM-RNN machine learning classifier.
Applsci 11 12101 g003
Figure 4. Original time-dependent signals collected from IMU sensors in one single human trial.
Figure 4. Original time-dependent signals collected from IMU sensors in one single human trial.
Applsci 11 12101 g004
Figure 5. Results of data labeling. The blue line represents the raw signal collected from the IMU, and the orange line shows that the raw signal had been extracted to 28 intermittent phases by first-order differentiation of acceleration.
Figure 5. Results of data labeling. The blue line represents the raw signal collected from the IMU, and the orange line shows that the raw signal had been extracted to 28 intermittent phases by first-order differentiation of acceleration.
Applsci 11 12101 g005
Figure 6. Confusion matrix of classification results for the LSTM-RNN model after training collected from 4 IMU sensors. Model precision and recall were calculated and indicated as green figures in the last line and the last row, respectively. Precision: proportion of correctly classified predicted positives; recall: proportion of correctly classified actual positives.
Figure 6. Confusion matrix of classification results for the LSTM-RNN model after training collected from 4 IMU sensors. Model precision and recall were calculated and indicated as green figures in the last line and the last row, respectively. Precision: proportion of correctly classified predicted positives; recall: proportion of correctly classified actual positives.
Applsci 11 12101 g006
Figure 7. Classification performance of the trained LSTM-RNN model by using a testing sequence collected in the same protocol but not in the training dataset. The red line represents postures performed by subjects during human trials. The blue line represents postures predicted by the model with a sampling rate of 50 Hz. Classification errors are shown in green rectangles.
Figure 7. Classification performance of the trained LSTM-RNN model by using a testing sequence collected in the same protocol but not in the training dataset. The red line represents postures performed by subjects during human trials. The blue line represents postures predicted by the model with a sampling rate of 50 Hz. Classification errors are shown in green rectangles.
Applsci 11 12101 g007
Figure 8. F1 scores of the neural network using different sensor combinations.
Figure 8. F1 scores of the neural network using different sensor combinations.
Applsci 11 12101 g008
Table 1. Experimental setup in the training procedure.
Table 1. Experimental setup in the training procedure.
DescriptionValue
EnvironmentIntel Core i7-7700HQ CPU, 12 GB RAM
GPU NVIDIA GeForce GTX1050
Mini-batch size32
Epoch600
Learning rate0.001
Optimization (Backpropagation)Adam optimizer
Validation10-fold CV
Table 2. Accuracy and F1 scores of the LSTM-RNN model after training with five datasets collected from 4 IMU sensors.
Table 2. Accuracy and F1 scores of the LSTM-RNN model after training with five datasets collected from 4 IMU sensors.
Accuracy (%)F1 Score
Upright98.30.942
Slump99.20.974
Lean99.30.975
Right bending99.10.967
Left bending99.40.976
Right twisting98.90.960
Left twisting99.10.966
Average99.0 ± 0.30.966 ± 0.012
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tang, H.-Y.; Tan, S.-H.; Su, T.-Y.; Chiang, C.-J.; Chen, H.-H. Upper Body Posture Recognition Using Inertial Sensors and Recurrent Neural Networks. Appl. Sci. 2021, 11, 12101. https://doi.org/10.3390/app112412101

AMA Style

Tang H-Y, Tan S-H, Su T-Y, Chiang C-J, Chen H-H. Upper Body Posture Recognition Using Inertial Sensors and Recurrent Neural Networks. Applied Sciences. 2021; 11(24):12101. https://doi.org/10.3390/app112412101

Chicago/Turabian Style

Tang, Hao-Yuan, Shih-Hua Tan, Ting-Yu Su, Chang-Jung Chiang, and Hsiang-Ho Chen. 2021. "Upper Body Posture Recognition Using Inertial Sensors and Recurrent Neural Networks" Applied Sciences 11, no. 24: 12101. https://doi.org/10.3390/app112412101

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop