Next Article in Journal
An Efficient Attentional Image Dehazing Deep Network Using Two Color Space (ADMC2-net)
Previous Article in Journal
Stretch Sensor: Development of Biodegradable Film
Previous Article in Special Issue
Detection of SO2F2 Using a Photoacoustic Two-Chamber Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Low-Cost Inertial Measurement Unit Motion Capture System for Operation Posture Collection and Recognition

1
School of Mechatronics Engineering, Harbin Institute of Technology, Harbin 150001, China
2
School of Astronautics, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(2), 686; https://doi.org/10.3390/s24020686
Submission received: 11 December 2023 / Revised: 12 January 2024 / Accepted: 19 January 2024 / Published: 21 January 2024
(This article belongs to the Special Issue Advanced Sensors for Real-Time Monitoring Applications ‖)

Abstract

:
In factories, human posture recognition facilitates human–machine collaboration, human risk management, and workflow improvement. Compared to optical sensors, inertial sensors have the advantages of portability and resistance to obstruction, making them suitable for factories. However, existing product-level inertial sensing solutions are generally expensive. This paper proposes a low-cost human motion capture system based on BMI 160, a type of six-axis inertial measurement unit (IMU). Based on WIFI communication, the collected data are processed to obtain the displacement of human joints’ rotation angles around XYZ directions and the displacement in XYZ directions, then the human skeleton hierarchical relationship was combined to calculate the real-time human posture. Furthermore, the digital human model was been established on Unity3D to synchronously visualize and present human movements. We simulated assembly operations in a virtual reality environment for human posture data collection and posture recognition experiments. Six inertial sensors were placed on the chest, waist, knee joints, and ankle joints of both legs. There were 16,067 labeled samples obtained for posture recognition model training, and the accumulated displacement and the rotation angle of six joints in the three directions were used as input features. The bi-directional long short-term memory (BiLSTM) model was used to identify seven common operation postures: standing, slightly bending, deep bending, half-squatting, squatting, sitting, and supine, with an average accuracy of 98.24%. According to the experiment result, the proposed method could be used to develop a low-cost and effective solution to human posture recognition for factory operation.

1. Introduction

In the era of Industry 4.0, motion capture systems will find broader applications in engineering for digital human modeling [1]. In the factory, the recognition of human body movement contributes to human–machine collaboration [2] and human factor analysis [3]. In contrast to optical cameras, inertial sensors are more flexible and resistant to obstruction, making them suitable for scenarios such as automotive assembly [4,5]. Researchers have established methods for capturing full-body motion by sparse inertial sensors. Susperregi et al. [6] proposed the fusion of multiple low-cost sensors and cameras to capture human behavior, addressing data bias through data fusion. Caputo et al. [4] utilized a motion capture system to estimate the basic segment positions of the human body. He et al. [7] introduced a wavelet tensor fuzzy clustering scheme for analyzing multisensor signals to capture human behavior, achieving higher recognition accuracy compared to the fuzzy mean clustering method. Liu et al. [8] developed a segmentation procedure based on a moving average window algorithm, and introduced a double-threshold technique for automatic recognition and segmentation of calibration postures. Yi et al. [9] tracked human motion using only six inertial sensors, combining a neural kinematic estimator and a physical perception motion optimizer. Previous work has provided good guidance for achieving low-cost inertial sensor dynamic capture, giving IMUs potential application prospects in the engineering field.
Due to the inevitable presence of a large number of metal objects in the factory environment, the negative impact on magnetometers needs to be considered. Therefore, we have chosen the cost-effective six-axis sensor chip BMI160, along with the ESP8266-NodeMCU chip, IP5306 BMS charging board, and li-ion battery, to collectively form our tracker; the total cost is USD 3.60. A comparison of prices and performance of other IMU solutions is presented in Table 1, covering aspects such as price, sampling rate, accelerometer rate noise spectral density, gyroscope rate noise spectral density, interface mode, and battery life. The IMU solutions commonly used in human motion recognition research, such as Xsens MTw Awinda [10], MetaMotionR [10], Next-Generation IMU [11,12], MetaMotionC [13], Shimmer3 [14], and InvenSense MPU-9250 [15], were chosen. Through comparison, our solution demonstrates advantages in terms of pricing. Moreover, for the collection of human body movements, the sampling rate and accuracy within our solution fall within acceptable ranges, and the interface and the battery life employed in our solution are also sufficient for human motion capture.
Machine learning is commonly used in human motion recognition research, for example, the support vector machines classification model [16], Markov model [17], and random forest (RF) [18]. In the past few years, deep learning algorithms have found extensive applications in the realm of human motion recognition [19], demonstrating superior recognition performance compared to traditional algorithms [20,21]. Akkaladevi et al. [22] proposed a multilabel human action recognition framework using a spatiotemporal graph convolutional network (ST-GCN) to capture spatial and temporal relationships between joint sequences. Tang et al. [23] introduced a novel dual-branch interactive network (DIN) that incorporates the strengths of both CNNs and transformers for managing multichannel time series. Wang et al. [24] explored adaptive networks that can dynamically adjust their structure based on available computing resources, allowing for a trade-off between accuracy and speed. Dey et al. [25] utilized a three-layer stacked temporal convolutional network to predict foot angular positions. Oh et al. [26] employed a pattern recognition method based on an artificial neural network algorithm to detect different gait states. Seenath et al. [27] proposed the conformer-based human activity recognition model, which leverages attention mechanisms to better capture the temporal dynamics of human motion and improve recognition accuracy. Considering that IMU motion capture data contain both temporal and spatial information, Chen et al. [28] used a deep convolutional neural network with a bidirectional long short-term memory network (DCNN-BiLSTM) to recognize and estimate four swimming styles. Based on deep learning algorithms, the accuracy of human motion recognition can reach around 90%. Based on existing research, we will carry out IMU-based human operation posture recognition.
Existing product-level inertial motion capture devices generally require high purchasing costs. This paper aims to explore a low-cost operation motion capture system and an operation posture recognition solution based on IMU. The cost-effective core components are used to build the human motion capture system, and experimental tests are conducted in virtual factory environments. A deep neural network model is used to recognize multiple basic operation postures offline based on experimental datasets. This paper is organized as follows: Section 1 offers an overview of the research status and significance of IMU-based human motion capture and operation posture recognition. Section 2 introduces the low-cost assembly operation motion capture scheme based on IMU from the aspects of hardware configuration, motion signal processing, and human motion reproduction. Section 3 describes the operation motion capture experiment and the operation posture recognition method based on the BiLSTM model. Section 4 discusses the proposed research methods and suggests future research directions; Section 5 summarizes the proposed research work.

2. Design of a Low-Cost Motion Capture System Based on IMU

2.1. Overall Solution

This paper proposes a low-cost human motion capture system based on IMU. As shown in Figure 1, the system consists of four main components: a firmware module, a hardware module, a signal processing module, and a synchronized visualization module. In the hardware part, the core modules include the inertial measurement, communication, and charging modules. The inertial measurement module utilizes BMI160. The communication module uses the ESP8266 chip for wireless communication via WiFi. The charging module consists of a charging integration board, a battery, and a switch. The BMI160 is driven by the CH341SER. The firmware code is compiled and run in PlatformIO IDE (VSCode). The tracker signals are transmitted to the host computer via WiFi, where the collected data are processed to obtain the pose information of the sensors. The trackers are assigned to the corresponding joint positions of the digital human body based on their actual wearing positions. Combined with the hierarchical relationship of the human skeleton, the real-time calculation of human posture is performed. Finally, using the Open Sound Control (OSC) network transmission protocol, the system synchronously visualizes human motion through a 3D digital human model in Unity3D.

2.2. Hardware

The main functional components of the action tracker are the BMI160 IMU module, the ESP8266-NodeMCU module, and the IP5306 BMS charging module. Considering usability and price, the BMI160 was chosen to implement the inertial measurement function. The BMI160 chip module includes a three-axis accelerometer and a three-axis gyroscope. The chip features three 16-bit analog-to-digital converters (ADCs) for digitizing the accelerometer outputs and three 16-bit ADCs for digitizing the gyroscope outputs with standard IIC (up to 1 MHz)/SPI communication protocol. The chip can monitor an acceleration range of ±4 g and an angular velocity range of ±250°/s. The sampling rate is 100 Hz. In coordination, the ESP8266 Node MCU module, which is a version containing the ESP-12F WIFI unit with a peak power consumption of approximately 1.5 W, was selected for communication, supporting WIFI connections in the 2.4 G frequency band. Additionally, the charging module was designed using the TP4056 Type-C charging chip, an input voltage of 5 V, and a maximum charging current of 1000 mA. The 3.7 V, 1500 mAh lithium battery was chosen. Finally, two-position toggle switches were selected to control the tracker’s on and off functions. The circuit diagram and physical diagram of the tracker are shown in Figure 2a. The wires were soldered in a tightly arranged manner to minimize the size of the tracker. The tracker’s housing was 3D printed, with a total length of 54 mm, a total width of 39 mm, and a total height of 29 mm. The strap width is 25 mm, as shown in Figure 2b. In this paper, six motion trackers are used, strapped respectively to the chest, and waist, above the left knee joint, above the left ankle joint, above the right knee joint, and above the right ankle joint of the human. From top to bottom, these trackers represent the movements of the chest, waist, knee end of the femur bone, and the ankle end of the tibia bone. The wearing positions and directions of IMU are shown in Figure 2c. When wearing, the direction of the BMI160 inside each tracker is consistent, with the Y-axis pointing towards the ground and the Z-axis pointing towards the front of the body. The length values of each segment of the experimenter’s body have been pre-inputted into the terminal, and the movement status of the trunk and lower limbs can be obtained by providing joint displacement and angle. Before the motion capture experiment, the experimenter needs to make two designated postures: upright posture and skiing posture, to calibrate the initial direction of each tracker.

2.3. Signal Processing

The processing of motion signals involves two main parts: filtering and drift compensation of IMU signals. Kalman filtering algorithm is used for filtering. Human motion is irregular but within a certain activity space. The Kalman filtering algorithm is a classic method for processing IMU signals, which consists of predicting the position of the next time step, and correcting the position of the current state. The specific implementation principle is as follows.
X ^ k ¯ = A k ¯ X ^ k 1 + B k u k
P k ¯ = A k ¯ P k 1 A k T + Q
Above are prediction equations. In Equation (1), X ^ k ¯ represents the prior state estimation at time k, and X ^ k 1 represents the posterior state at time k−1, respectively. A k ¯ is a transformation matrix that represents the proportion of the previous state’s correction to the current state result. B k represents the control variable matrix, and u k is the state control vector. In Equation (2), P k ¯ represents the prior estimate covariance at time k, and P k 1 represents the posterior estimate covariance at time k 1 , Q is the covariance of the system process noise.
K k = P k ¯ H k T H k P k ¯ H k T + R
X ^ k = X ^ k ¯ + K k ( Z k H k X ^ k ¯ )
P k = ( I H k K k ) P k ¯
Equation (3) calculates the Kalman gain ( K k ), in which H k represents the prediction matrix and R is the covariance matrix of the measurement noise. Equation (4) uses two predicted values and a ratio to calculate the output X ^ k , the posterior state estimation at time k. Z k is a measurement vector. Equation (5) prepares the posteriori estimation covariance at time k( P k ) for the prediction of the next time step.
The drift compensation part mainly involves applying inverse rotation to compensate drift of the IMU. In this study, signal processing and fusion are based on the Slime VR open-source software, a recently matured open-source motion capture solution based on IMUs. Based on our experimental environment and equipment, after multiple tuning and testing sessions primarily focusing on the accuracy and stability of reproducing human motion, we finally set the filtering strength to 50% and drift compensation strength to 20%. The original signals collected by IMUs consist of XYZ tri-axis acceleration signals and XYZ tri-axis gyroscope signals. The displacement information can be obtained by integrating the acceleration signal, while the rotation angle information can be obtained by integrating the gyroscope signal.
The calculation method for obtaining the current pose from two frames of IMU data is as follows. For the acceleration data, calculate the average acceleration between the current time t and the next time t + 1 . This average acceleration over the time interval is used to approximate the velocity and displacement at t + 1 , given the initial velocity and displacement at t. Since the IMU acceleration data is represented in the body coordinate system, it needs to be transformed to the world coordinate system using the corresponding pose. Before the transformation, the bias needs to be subtracted, and after the transformation, the gravitational acceleration needs to be subtracted. For the gyroscope data, the average angular velocity over the time interval is calculated between t and the next time t + 1 . With this average angular velocity and the current pose, the pose at t + 1 can be approximated. Equations (6)–(12) show the entire integration process.
a t , w = Q t ( a t , b B a ) g
where a t , w is the acceleration of IMU at time t in the world coordinate system, Q t is the quaternion of IMU at time t, a t , b is the acceleration at time t in the body coordinate system, B a is the deviation of the body coordinate system, and g is the gravitational acceleration.
ω ¯ t = 1 2 ( ω t + ω t + 1 ) B g
Q t + 1 = Q t ( ω t Δ t )
In Equations (7) and (8), ω ¯ t is the average angular velocity, ω t is the angular velocity at time t, ω t + 1 is the angular velocity at time t + 1 , and B g is the gyroscope bias, Q t + 1 is IMU Quaternion at time t + 1 .
a t + 1 , w = Q t + 1 ( a t + 1 , b B a ) g
a ¯ t , w = 1 2 ( a t , w + a t + 1 , w )
In Equations (9) and (10), a t + 1 , w is the acceleration in the world coordinate system at time t + 1 , a t + 1 , b is the acceleration in the body coordinate system at time t + 1 , and a ¯ t , w is the average acceleration.
V t + 1 = V t + a ¯ t , w Δ t
D t + 1 = D t + V t Δ t + 1 2 a ¯ t , w Δ t 2
In Equations (11) and (12), V t is the velocity at time t, V t + 1 is the velocity at time t, D t is the displacement of the IMU at time t, and D t + 1 is the displacement of the IMU at time t + 1 .

2.4. Online Synchronized Display of Human Body Motion

The online synchronized display of human body movements is achieved based on the tracker’s pose information and the hierarchical relationship of the human body skeleton. This study uses a simplified digital human model to focus on the operational movements of the human torso and lower limbs. A 3D digital human model was built on the Unity platform. The joint composition of the digital human includes thoracic joints, lumbar joints, left and right hip joints, left and right knee joints, and left and right ankle joints. In constructing the digital human model, the thoracic, lumbar, and hip joints comprise three independent subjoints capable of generating rotation, pitch, and yaw movements. The ankle joint is generally considered a ball joint with two independent axes of rotation, while the knee joint has only one axis of rotation. The head and upper limb segments are set to default states. Figure 3 shows the skeletal and digitized human models with skinning. The lengths of the body segments are set according to the experimenter’s height (1580 mm) and standard body proportions.
Unity and the IMU host can communicate through the OSC protocol to achieve an online synchronized display of human body movements. Figure 4 shows the real-time human body movement at a certain moment and the corresponding movements of the digital human model at the same moment.

3. Human Posture Recognition with Assembly Operations

3.1. Basic Operation Postures

By observing the assembly and maintenance operation processes of large-scale equipment, several common assembly basic postures that facilitate exerting force could be summarized: standing posture, slightly bending posture, deep bending posture, half squatting posture, squatting posture, sitting posture, and supine posture. Operators could perform upper limb actions based on these basic postures, such as pushing (pulling), tightening (loosening), gripping, tapping, etc. The labels, names, and reference images of the basic working postures are shown in Table 2. The definition of postures mainly considered the range of bending angles of the torso, the range of bending angles of the hip joint, and the range of bending angles of the knee joint. Labels have been defined for these basic postures.

3.2. Operation Posture Collection Experiment

As shown in Figure 5, an immersive assembly scene was set up to facilitate participants making corresponding assembly movements based on prompts using Tecnomatix software and HTC VIVE devices. The participant wearing the tracker completed the operation tasks under instructions. The router was not connected to other devices to obtain sufficient bandwidth during the experiment, and the entire experiment process was recorded. At the same time, we tried to avoid other 2.4 G signals to prevent excessive data transmission delay caused by frequency congestion in the experimental environment. The experiment was conducted within a radius of 5 m from the router to ensure low data transmission delay. The average latency during the actual testing process was approximately 3 ms. With time, the BMI160 may experience drift, causing body parts to face the wrong direction after some time. Therefore, a calibration of the wearable device was required every 10 min.
The participants sequentially completed seven different types of work tasks under voice prompts. Each work task corresponds to a category of basic working postures. A rest period was scheduled between the fourth and fifth tasks for device reset. Table 3 displays the duration of each operation task. The experiment involved a participant with a mechanical engineering background familiar with assembly processes. The participant’s height is 1580 mm, and weight is 55 kg.

3.3. Operation Posture Recognition Method

After signal processing, the experiment data were organized as the cumulative displacement of six joints and the joint angles of six joints over time. In preparation for posture recognition, removing the preparation and rest periods and labeling the remaining periods with corresponding posture labels is necessary. As shown in Figure 6, Taking the curve of chest joint angle over time as an example, the gray area in the graph represents the excluded periods. In the remaining periods, each color represents a category of working posture.
BiLSTM is a deep learning model suitable for sequential data, and particularly effective for data with a temporal structure, such as time series. BiLSTM effectively captures contextual relationships and long-term dependencies in sequential data by combining forward and backward information. In recent years, the BiLSTM model has been commonly applied in research on IMU-based human posture recognition, demonstrating excellent recognition performance. Based on the experiment data, the BiLSTM model was used to recognize the seven basic operation postures: standing posture, slightly bending posture, deep bending posture, half-squatting posture, squatting posture, sitting posture, and supine posture. The operation posture recognition network structure is shown in Figure 7. Labeled experimental data were transformed into the dataset using a sliding window technique. The window length is 50 and the sliding size is 5; a total of 16,067 labeled samples were obtained for training. The input features included the displacement and rotation angles of six joints (chest, waist, left hip, right hip, left ankle, right ankle) in the XYZ direction, resulting in 36 features. The input layer of the network module is a 16,067 × 36 matrix. The input sequence is processed by two separate LSTM layers, each observing the sequence in both the forward and backward directions. The number of hidden neurons in each LSTM layer is 64. The input time-series data first pass through the forward layer. For each time step, the forward LSTM unit updates its internal state and produces an output. Similarly, the input sequence data also go through the backward layer. For each time step, the backward LSTM unit updates its internal state and produces an output. The outputs from both the forward and backward directions are merged. The merged representation is then passed to a fully connected layer. Finally, it is fed into an output layer for classification, using the softmax activation function to generate a probability distribution over the classes. The outputs of multiple neurons are mapped to the range of 0–1 to obtain the predicted probability distribution, which represents the probability of belonging to each category and enables posture prediction. This model was compiled using the cross-entropy loss function and adaptive moment estimation (Adam) optimizer. We divided the dataset into training and testing sets in a 4:1 ratio, the random state was set to 42. The epochs and batch size were set to 10 and 32. The initial learning rate was set as 0.001. L2 regularization with a dropout of 0.5 was selected to prevent overfitting of the model.

3.4. Operation Posture Recognition Result

The offline test was conducted on a workstation with an Intel Core i7-1165G7 CPU and NVIDIA GeForce MX 450 GPU. To reduce the random effects of the training tests, the sample order was randomly shuffled and the training test was repeated five times. The average training time was 75.08 s. Figure 8 shows the training and validation loss as the number of iterations increases. It can be observed that the loss curves of the training set and validation set tend to flatten after the 8th iteration. After the 10th generation, the test set loss remained stable below 0.05.
After the test, the average accuracy of posture prediction was 98.24%. The posture prediction transition time, including data preprocessing time and inference time, is 31 ms. Table 4 shows the accuracy, recall, and F-score of each posture prediction result. Each calculation result in the table is the average of 5 tests.
The results are summarized in Table 4, which shows that (1) the precision of the seven postures are all above 96%, with the highest precision for the bending posture at 99.74% and the lowest for the squatting posture at 96.80%; (2) the recall of the seven postures are all above 96%, with the highest recall for the bending posture at 99.45% and the lowest for the half-squatting posture at 96.57%; (3) the F-score of the seven postures are all above 97%, with the highest F-score for the bending posture at 99.56% and the lowest for the half-squatting posture at 97.42%. Overall, the recognition performance is best for the deep bending posture, while the recognition performance for the half-squatting and squatting postures is relatively poor. Figure 9 shows the distribution of the test set confusion matrix from five tests. From the confusion matrix, it can be visually observed that the model performs well in classifying most postures. In comparison, the standing posture and half-squatting posture are more prone to be misclassified as a slightly bending posture.
The comparison of the results with existing research is shown in Table 5. In aspects of accuracy and time cost, we compared our work with other IMU-based human posture recognition works. The number of recognition classes and the number of IMUs are also shown in the table. In terms of accuracy, our work achieved a 98% accuracy for the classification of seven postures using six trackers, placing it in a relatively high position compared to similar studies. Regarding the time cost, we took into account the posture prediction transition time (including data preprocessing time and inference time), as well as the IMU sampling rate. While our method does not match the performance of the approach described in reference [28], we were able to identify a greater number of postures with more considerable accuracy.

4. Discussion

This paper focuses on operation posture collection and recognition based on low-cost IMU. The proposed method is available, and the accuracy of basic posture classification recognition is satisfactory. Integrating more features and employing more complex machine learning models may result in higher recognition accuracy, but it also comes with relatively higher time costs. Based on the data in this experiment, we compared the recognition accuracy of the LSTM and BiLSTM models. We conducted five training and testing times and calculated the average accuracy of classification on the test set. The average accuracy of LSTM is 95.81%, while the average accuracy of BiLSTM is 98.24%. Compared to LSTM, BiLSTM has a greater advantage in basic posture classification for assembly tasks. However, due to the complexity of the model, BiLSTM requires longer training time. In our test, LSTM took 34.82 s and BiLSTM took 75.08 s. However, the difference in prediction time between the two models is not significant: LSTM took 15ms and BiLSTM took 31 ms.
Using wireless network transmission of data can enhance convenience, but it may lead to sudden posture distortion when the network signal is unstable, as shown in Figure 10. The occurrence of abnormal postures is related to the network signal quality. In the experimental environment, the occurrence of anomalies is rare (1–2 times/10 min) and quickly recovers to normal. However, in environments with poorer signal quality, it can be foreseen that sudden abnormal postures will affect the observation of operation movements and the accuracy of posture recognition to some extent. How to identify and ignore exceptional signals is a research question that needs to be further studied. Because abnormal postures often manifest as sudden drifts in joint positions, a possible solution is to set a threshold for joint position changes and identify abnormal postures accordingly. Alternatively, based on gathering a sufficient number of data samples, machine learning models can be employed to differentiate between normal and abnormal states. We will attempt to address this issue in future work.

5. Conclusions

For the demand for operation posture recognition in the Industry 4.0 era, this paper explores a low-cost method for collecting assembly actions and recognizing assembly postures based on IMU. The study includes the following aspects:
A low-cost human motion collection system based on IMU has been proposed. The BMI160 inertial measurement module is combined with the ESP8266 communication module to create the motion collection tracker. Motion signals are transmitted via WiFi to the computer to obtain sensor pose information. The tracker is assigned to the corresponding joint positions of the digital human body based on the actual wearing position. Real-time calculation of human posture is performed by combining the hierarchical relationship of the human body skeleton. The Unity development platform receives human motion information and presents synchronized online visualization through a 3D digital human model;
We experimentally validate the feasibility of the action collection scheme. We have simulated various assembly tasks in a virtual reality environment and collected motion information for six joints of the subjects: chest, waist, left knee, left ankle, right knee, and right ankle, which included the rotation angles around XYZ directions and the displacement in XYZ directions. The BiLSTM model was used to identify seven common assembly postures: standing posture, slightly bending posture, deep bending posture, half-squatting posture, squatting posture, sitting posture, and supine posture. The model performs well in classifying these operation postures.
Based on the experiment results, the system could serve as a low-cost solution for the basic operation posture recognition of operation tasks. Subsequent research will focus on enhancing the operation posture recognition system and testing it in real factory environments.

Author Contributions

Conceptualization, M.Y. and J.L.; methodology, M.Y. and J.L.; validation, M.Y. and T.W.; writing—original draft preparation, M.Y.; writing—review and editing, M.Y., J.L. and T.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The original experiment data are available.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Feldmann, F.; Seitz, R.; Kretschmer, V.; Bednorz, N.; Hompel, M.T. Ergonomic Evaluation of Body Postures in Order Picking Systems Using Motion Capturing. In Proceedings of the 2019 Prognostics and System Health Management Conference (PHM-Paris), Paris, France, 2–5 May 2019; pp. 204–209. [Google Scholar] [CrossRef]
  2. Roda-Sanchez, L.; Garrido-Hidalgo, C.; García, A.S.; Olivares, T.; Fernández-Caballero, A. Comparison of RGB-D and IMU-based gesture recognition for human-robot interaction in remanufacturing. Int. J. Adv. Manuf. Technol. 2018, 8, 297–308. [Google Scholar] [CrossRef]
  3. Nguyen Ngoc, H.; Lasa, G.; Iriarte, I. Human-centred design in industry 4.0: Case study review and opportunities for future research. J. Intell. Manuf. 2022, 33, 35–76. [Google Scholar] [CrossRef] [PubMed]
  4. Caputo, F.; Greco, A.; D’Amato, E.; Notaro, I.; Spada, S. On the use of Virtual Reality for a human-centered workplace design. Procedia Struct. Integr. 2023, 124, 3099–3111. [Google Scholar] [CrossRef]
  5. Fang, W.; Zheng, L.; Xu, J. Self-contained optical-inertial motion capturing for assembly planning in digital factory. Int. J. Adv. Manuf. Technol. 2017, 93, 1243–1256. [Google Scholar] [CrossRef]
  6. Susperregi, L.; Arruti, A.; Jauregi, E.; Sierra, B.; Martínez-Otzeta, J.M.; Lazkano, E.; Ansuategui, A. Fusing multiple image transformations and a thermal sensor with kinect to improve person detection ability. Eng. Appl. Artif. Intell. 2013, 26, 1980–1991. [Google Scholar] [CrossRef]
  7. He, H.; Tan, Y.; Zhang, W. A wavelet tensor fuzzy clustering scheme for multi-sensor human activity recognition. Eng. Appl. Artif. Intell. 2018, 70, 109–122. [Google Scholar] [CrossRef]
  8. Liu, Y.T.; Zhang, Y.A.; Zeng, M. Sensor to segment calibration for magnetic and inertial sensor based motion capture systems. Measurement 2019, 142, 1–9. [Google Scholar] [CrossRef]
  9. Yi, X.; Zhou, Y.; Habermann, M.; Shimada, S.; Golyanik, V.; Theobalt, C.; Xu, F. Physical inertial poser (pip): Physics-aware real-time human motion tracking from sparse inertial sensors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13167–13178. [Google Scholar]
  10. Guidolin, M.; Budau Petrea, R.A.; Roberto, O.; Reggiani, M.; Menegatti, E.; Tagliapietra, L. On the Accuracy of IMUs for Human Motion Tracking: A Comparative Evaluation. In Proceedings of the 2021 IEEE International Conference on Mechatronics (ICM), Kashiwa, Japan, 7–9 March 2021; pp. 1–6. [Google Scholar] [CrossRef]
  11. Tang, H.Y.; Tan, S.H.; Su, T.Y.; Chiang, C.J.; Chen, H.H. Upper Body Posture Recognition Using Inertial Sensors and Recurrent Neural Networks. Appl. Sci. 2021, 11, 12101. [Google Scholar] [CrossRef]
  12. Kuo, C.T.; Lin, J.J.; Jen, K.K.; Hsu, W.L.; Wang, F.C.; Tsao, T.C.; Yen, J.Y. Human Posture Transition-Time Detection Based upon Inertial Measurement Unit and Long Short-Term Memory Neural Networks. Biomimetics 2023, 8, 471. [Google Scholar] [CrossRef]
  13. Zhao, J.; Obonyo, E.; G. Bilén, S. Wearable Inertial Measurement Unit Sensing System for Musculoskeletal Disorders Prevention in Construction. Sensors 2021, 21, 1324. [Google Scholar] [CrossRef]
  14. Baghdadi, A.; Cavuoto, L.A.; Crassidis, J.L. Hip and Trunk Kinematics Estimation in Gait Through Kalman Filter Using IMU Data at the Ankle. IEEE Sens. J. 2018, 18, 4253–4260. [Google Scholar] [CrossRef]
  15. Yan, Y.; Ou, Y. Accurate fall detection by nine-axis IMU sensor. In Proceedings of the 2017 IEEE International Conference on Robotics and Biomimetics (ROBIO), Macau, China, 5–8 December 2017; pp. 854–859. [Google Scholar] [CrossRef]
  16. Pelliccia, L.; Klimant, F.; De Santis, A.; Di Gironimo, G.; Lanzotti, A.; Tarallo, A.; Putz, M.; Klimant, P. Task-based motion control of digital humans for industrial applications. Procedia CIRP 2017, 62, 535–540. [Google Scholar] [CrossRef]
  17. Bordel, B.; Alcarria, R.; Robles, T. Recognizing human activities in Industry 4.0 scenarios through an analysis-modeling-recognition algorithm and context labels. Integr. Comput. Aided Eng. 2022, 29, 83–103. [Google Scholar] [CrossRef]
  18. Ragni, F.; Archetti, L.; Roby-Brami, A.; Amici, C.; Saint-Bauzel, L. Intention Prediction and Human Health Condition Detection in Reaching Tasks with Machine Learning Techniques. Sensors 2021, 21, 5253. [Google Scholar] [CrossRef]
  19. Chen, K.; Zhang, D.; Yao, L.; Guo, B.; Yu, Z.; Liu, Y. Deep Learning for Sensor-Based Human Activity Recognition: Overview, Challenges, and Opportunities. ACM Comput. Surv. (CSUR) 2021, 54, 1–40. [Google Scholar] [CrossRef]
  20. Banjarey, K.; Sahu, S.P.; Dewangan, D.K. A survey on human activity recognition using sensors and deep learning methods. In Proceedings of the 2021 5th International Conference on Computing Methodologies and Communication (ICCMC), Erode, India, 8–10 April 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1610–1617. [Google Scholar]
  21. Bozkurt, F. A comparative study on classifying human activities using classical machine and deep learning methods. Arab. J. Sci. Eng. 2022, 47, 1507–1521. [Google Scholar] [CrossRef]
  22. Akkaladevi, S.C.; Heindl, C. Action recognition for human robot interaction in industrial applications. In Proceedings of the 2015 IEEE International Conference on Computer Graphics, Vision and Information Security (CGVIS), Bhubaneswar, India, 2–3 November 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 94–99. [Google Scholar]
  23. Tang, Y.; Zhang, L.; Wu, H.; He, J.; Song, A. Dual-Branch Interactive Networks on Multichannel Time Series for Human Activity Recognition. IEEE J. Biomed. Health Inform. 2022, 26, 5223–5234. [Google Scholar] [CrossRef]
  24. Wang, X.; Zhang, L.; Huang, W.; Wang, S.; Wu, H.; He, J.; Song, A. Deep Convolutional Networks With Tunable Speed–Accuracy Tradeoff for Human Activity Recognition Using Wearables. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  25. Dey, S.; Schilling, A.F. A function approximator model for robust online foot angle trajectory prediction using a single IMU sensor: Implication for controlling active prosthetic feet. IEEE Trans. Ind. Inform. 2022, 19, 1467–1475. [Google Scholar] [CrossRef]
  26. Oh, H.W.; Hong, Y.D. Divergent Component of Motion-Based Gait Intention Detection Method Using Motion Information From Single Leg. J. Intell. Robot. Syst. 2023, 107, 51. [Google Scholar] [CrossRef]
  27. Seenath, S.; Dharmaraj, M. Conformer-based human activity recognition using inertial measurement units. Sensors 2023, 23, 7357. [Google Scholar] [CrossRef] [PubMed]
  28. Chen, L.; Hu, D. A Deep Learning Control Strategy of IMU-based Joint Angle Estimation for Hip Power-Assisted Swimming Exoskeleton. IEEE Sens. J. 2023, 23, 15058–15070. [Google Scholar] [CrossRef]
  29. Nguyen, V.N.; Yu, H. Novel automatic posture detection for in-patient care using IMU sensors. In Proceedings of the 2013 6th IEEE Conference on Robotics, Automation and Mechatronics (RAM), Manila and Pico de Loro, Philippines, 12–15 November 2013; pp. 31–36. [Google Scholar] [CrossRef]
  30. Mascret, Q.; Bielmann, M.; Fall, C.L.; Bouyer, L.J.; Gosselin, B. Real-Time Human Physical Activity Recognition with Low Latency Prediction Feedback Using Raw IMU Data. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; pp. 239–242. [Google Scholar] [CrossRef]
Figure 1. The structure of the low-cost motion capture system based on IMU.
Figure 1. The structure of the low-cost motion capture system based on IMU.
Sensors 24 00686 g001
Figure 2. The circuit diagram and physical diagram of the tracker. (a) The status of completed welding. (b) The tracker with casing and straps attached. (c) The tracker placement.
Figure 2. The circuit diagram and physical diagram of the tracker. (a) The status of completed welding. (b) The tracker with casing and straps attached. (c) The tracker placement.
Sensors 24 00686 g002
Figure 3. The digital human model. (a) Skeletal model. (b) Digital human model with skinning.
Figure 3. The digital human model. (a) Skeletal model. (b) Digital human model with skinning.
Sensors 24 00686 g003
Figure 4. Real-time human body and digital human model.
Figure 4. Real-time human body and digital human model.
Sensors 24 00686 g004
Figure 5. Operation posture collection experiment scene.
Figure 5. Operation posture collection experiment scene.
Sensors 24 00686 g005
Figure 6. Chest joint angle over time, each color represents a category of working posture, the gray area is the excluded periods.
Figure 6. Chest joint angle over time, each color represents a category of working posture, the gray area is the excluded periods.
Sensors 24 00686 g006
Figure 7. The operation posture recognition network structure.
Figure 7. The operation posture recognition network structure.
Sensors 24 00686 g007
Figure 8. Training and validation loss.
Figure 8. Training and validation loss.
Sensors 24 00686 g008
Figure 9. The confusion matrix from five tests.
Figure 9. The confusion matrix from five tests.
Sensors 24 00686 g009
Figure 10. The sudden posture distortion when the network signal is unstable. (a) The normal posture. (b) The abnormal posture.
Figure 10. The sudden posture distortion when the network signal is unstable. (a) The normal posture. (b) The abnormal posture.
Sensors 24 00686 g010
Table 1. Comparison of prices and performance of common IMUs.
Table 1. Comparison of prices and performance of common IMUs.
Ref.IMU ModelsPrice (1 IMU)Sampling RateAcc. Noise [µg/ Hz ]Gyro. Noise [°/s/ Hz ]InterfaceBattery Life
[10]Xsens MTw Awinda$437.2120 Hz2000.01Wireless 2.4 GHz6 h
[10]MetaMotionR$98.4100 Hz3000.007Bluetooth LTE 2.4 GHz1 to 14 days
[11,12]NG IMU$273.350 Hz/100 HzNANAPreconfigured Wi-Fi router4 to 12 h
[13]MetaMotionC$75NA3000.007Bluetooth 4.01 to 14 days
[14]Shimmer3$945.551.2 Hz72.50.007BluetoothNA
[15]InvenSense MPU-9250$10.9120 Hz3000.01I 2 C S P I NA
This work$3.60100 Hz3000.007Wireless 2.4 GHz6 h
Table 2. The basic postures.
Table 2. The basic postures.
LableGesture NameGesture PictureTorso Bending Angle (°)Hip Joint Bending Angle (°)Knee Joint Bending Angle (°)
1standing postureSensors 24 00686 i001−15–15−17–5−3–5
2slightly bending postureSensors 24 00686 i00215–405–505–60
3deep bending postureSensors 24 00686 i00340–8470–1165–70
4half squatting postureSensors 24 00686 i00415–4050–11660–100
5squatting postureSensors 24 00686 i00515–4050–116100–150
6sitting postureSensors 24 00686 i006−15–1530–70−3–5
7supine postureSensors 24 00686 i007−5–4050–1005–60
Table 3. The duration of each operation task.
Table 3. The duration of each operation task.
Task NameDuration (ms)
Preparation60,000
Installing bolts in standing posture121,800
Installing bolts in slightly bending posture123,250
Installing bolts in deep bending posture116,550
Installing bolts in half squatting posture140,600
Break time60,000
Installing bolts in squatting posture118,350
Installing bolts in sitting posture86,500
Installing bolts in supine posture96,300
Table 4. Classification accuracy of the test set.
Table 4. Classification accuracy of the test set.
Posture CategoryPrecisionRecallF-Score
1—standing0.99400.98590.9905
2—slightly bending0.97640.99090.9800
3—folding0.99740.99450.9956
4—half-squatting0.98320.96570.9742
5—squatting0.96800.98710.9773
6—sitting posture0.99360.97210.9823
7—supine posture0.97520.98500.9859
Table 5. Comparison of the human posture recognition results.
Table 5. Comparison of the human posture recognition results.
Ref.ModelAccuracyPrediction TimeSampling RateRecognition ClassesNumber of IMUs
[29]ANN82.74%NA50 Hz82
[13]CLN model90.00%NANA75
[12]LSTM94.44%170 ms100 Hz34
[28]DCNN-BiLSTM96.00%8.47 ms120 Hz43
[30]RBF-SVM97.35%120 ms62 Hz83
[11]LSTM-RNN99.00%277 ms–488 ms50 Hz74
This workBiLSTM98.24%31 ms100 Hz76
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yin, M.; Li, J.; Wang, T. A Low-Cost Inertial Measurement Unit Motion Capture System for Operation Posture Collection and Recognition. Sensors 2024, 24, 686. https://doi.org/10.3390/s24020686

AMA Style

Yin M, Li J, Wang T. A Low-Cost Inertial Measurement Unit Motion Capture System for Operation Posture Collection and Recognition. Sensors. 2024; 24(2):686. https://doi.org/10.3390/s24020686

Chicago/Turabian Style

Yin, Mingyue, Jianguang Li, and Tiancong Wang. 2024. "A Low-Cost Inertial Measurement Unit Motion Capture System for Operation Posture Collection and Recognition" Sensors 24, no. 2: 686. https://doi.org/10.3390/s24020686

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop