Next Article in Journal
Study on the Equivalence Transformation between Blasting Vibration Velocity and Acceleration
Next Article in Special Issue
Correlations of Aerobic Capacity with External and Internal Load of Young Football Players during Small-Sided Games
Previous Article in Journal
Crack Detection and Analysis of Concrete Structures Based on Neural Network and Clustering
Previous Article in Special Issue
Optimization of Torque-Control Model for Quasi-Direct-Drive Knee Exoskeleton Robots Based on Regression Forecasting
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Estimating Rotational Acceleration in Shoulder and Elbow Joints Using a Transformer Algorithm and a Fusion of Biosignals

1
School of Mechanical Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
2
Zhiyuan Research Institute, Hangzhou 310000, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(6), 1726; https://doi.org/10.3390/s24061726
Submission received: 11 January 2024 / Revised: 5 March 2024 / Accepted: 6 March 2024 / Published: 7 March 2024
(This article belongs to the Special Issue Human Movement Monitoring Using Wearable Sensor Technology)

Abstract

:
In the present study, we used a transformer model and a fusion of biosignals to estimate rotational acceleration in elbow and shoulder joints. To achieve our study objectives, we proposed a mechanomyography (MMG) signal isolation technique based on a variational mode decomposition (VMD) algorithm. Our results show that the VMD algorithm delivered excellent performance in MMG signal extraction compared to the commonly used technique of empirical mode decomposition (EMD). In addition, we found that transformer models delivered estimates of joint acceleration that were more precise than those produced by mainstream time series forecasting models. The average R2 values of transformer are 0.967, 0.968, and 0.935, respectively. Finally, we found that using a fusion of signals resulted in more precise estimation performance compared to using MMG signals alone. The differences between the average R2 values are 0.041, 0.053, and 0.043, respectively. Taken together, the VMD isolation method, the transformer algorithm and the signal fusion technique described in this paper can be seen as supplying a robust framework for estimating rotational acceleration in upper-limb joints. Further study is warranted to examine the effectiveness of this framework in other musculoskeletal contexts.

1. Introduction

In recent years, multimodal signal fusion methods and deep learning techniques for human activity recognition (HAR) have been extensively studied by researchers. Several such works may be considered for comparison with the study described in this paper. Dirgová Luptáková I et al. (2022) classified human activities using transformer models and achieved an accuracy rate of 99.2%. The device used for measurements in this research was a smartphone (accelerometer and gyroscope) placed in a bag on the waist of each person [1]. Wensel J et al. developed recurrent and vision transformer models to upgrade scalability and HAR speed; they obtained a level of accuracy that matched that of ResNet-LSTM but was achieved at twice the speed [2]. Shavit Y et al. (2021) evaluated a transformer model on diverse inertial datasets obtained over 27 h from 91 users, i.e., representing varying degrees of difficulty [3]. The scholarly literature also reveals the sheer number of biosignals that can be used for the recognition of human motion intention. Other recent studies have considered methods such as surface electromyography (sEMG), mechanomyography (MMG), and electroencephalography (EEG), as well as human–machine interactive forces [4,5].
A review of the literature reveals that one vital consideration when apply biosignals to estimate human motion is to ensure that wearing sensors will not affect human movement; this was previously emphasized by Zhang L et al. (2019) [6]. A specific trend in recent research has been to utilize MMG and sEMG for the estimation of joint acceleration, building on previous work on decoding movement via these signals. In comparison with more invasive or motion-constraining sensors, MMG and sEMG sensors are able to capture salient neuromuscular dynamics without imposing restrictions on users. Furthermore, the fusion of sEMG, which measures electrical activity driving muscle contraction, with MMG, which quantifies mechanical muscular vibrations, may result in the generation of complementary information. Further studies are warranted to determine optimal sensor arrangements and signal processing methodologies for the robust estimation of joint acceleration across diverse motion contexts.
A review of the literature also reveals the emerging use of transformer algorithms in the classification of human movement. Transformer models comprise a class of deep learning architectures that use attention mechanisms for processing sequential data, including pictures or natural language. Compared to recurrent neural networks (RNNs) or convolutional neural networks (CNNs), transformers offer several advantages when used for sequence modeling tasks. These include the following: (1) the ability to capture long-range dependencies between inputs, (2) efficient parallel calculation, and (3) state-of-the-art performance with respect to various natural language processing (NLP) benchmarks [7]. In this paper, we propose a new method for estimating rotational accelerations in elbow and shoulder joints using a fusion of biosignals and transformer algorithms. By using this method, we sought to harness the representation learning and sequence modeling capabilities of transformers to obtain the accurate decoding of limb dynamics from noninvasive neuromuscular signals. The usage of MMG and EMG enhanced the wearability of sensors without impeding movement. Using a simplified transformer model without a decoder part for motion estimation not only applied the attention mechanism of the transformer model but also simplified the structure of the motion estimation model and increased the operating efficiency and estimation accuracy.
In addition, we used the variational mode decomposition (VMD) algorithm for extracting MMG signals from raw acceleration measurements. A number of preparation techniques were then used to preprocess the acquired MMG and sEMG data. Ultimately, we established a transformer-based model for estimating joint acceleration using prepared MMG signals, sEMG signals, VMD-extracted pseudo-accelerations, and 3D camera-based movement tracker accelerations as model inputs.

2. Methods

This section begins with an overview of the experimental protocols and the sensor prototype used in the present study. The method proposed for MMG signal extraction is then described, along with the processing approaches applied to the acquired MMG and sEMG datasets. Finally, we describe the transformer model that was used for estimating joint acceleration in the present work. Transformers were originally developed for machine translation, but the more recent use of attention mechanisms by transformers has seen them emerge as an alternative to recurrent neural networks for sequence modeling, with the advantages of parallelization and faster training times.

2.1. Experimental Process and Sensor Prototype

The aim of the current experiment was to predict rotational accelerations in elbow and shoulder joints throughout three physical exercises: bicep curls, arm lateral raises, and arm frontal raises. Participants were instructed to execute movements at a frequency of their own choosing rather than at a prescribed rate. This was to promote naturalistic motion patterns that approximated daily activities. Informed consent was provided by 30 healthy adult participants (aged 24–27 years, including two females) after they were given the full details of the experimental protocols. The cohort size was adequate for a preliminary study; however, the number of participants should be increased in future studies so that statistical findings more closely and surely represent general populations.
The self-determined movements of participants enabled us to assess the ability of the models to generalize across movement speeds. To this end, for each physical exercise in the present study, 20 sets of 10 repetitions were completed by the participants, with 2–3 min of rest between sets to prevent fatigue. MMG and sEMG signals were acquired from the biceps brachii, anterior deltoid muscle, middle deltoid muscle, and posterior deltoid muscle using 4 special sensors. The arrangement of the sensors is shown in Figure 1a. Simultaneously, shoulder and elbow 3D rotational accelerations were recorded utilizing an optical gesture capture system. The prescribed movements were executed by participants upon instruction, and the data were collected. This level of physical exercise was sufficient for an initial validation. Greater robustness in the results might be obtained with more demanding exercise regimens; nevertheless, we believed that our preliminary research could offer indicative evidence of the proposed methodology’s utility and contribute to the search for a more rigorous experimental design.
The special sensor (Figure 1b) consisted of a 6-axis inertial measurement unit (MPU6050, TDK InvenSense, Sunnyvale, CA, USA), 2 surface electromyography electrodes with conductive gel, a STM32F103C8T6 microcontroller (ST, STMicroelectronics, Geneva, Switzerland), a nRF24L01 wireless module (NORDIC, Trondheim, Norway), and a lithium battery. The sensor was 5 cm long and 3 cm wide and weighed 98 g (including battery). It generated slight heat under long-term use but did not affect the progress of the experiment. Sensors were affixed over the target muscles using the conductive gel. The wireless design enabled untethered data streaming to a computer at 1000 Hz. Program for signal acquisition, processing, and motion estimation was implemented in Python on a PC (AMD Ryzen 5 5600X, 3.70 GHz). Sensor positioning and orientation were standardized to reduce placement-induced variability. Such a self-contained, wireless design offers a practical framework for unobtrusive monitoring outside laboratory settings. Further miniaturization and encapsulation could further enhance user comfort and robustness, enabling more prolonged periods of usage.

2.2. MMG Extraction and Biosignal Processing Method

2.2.1. MMG Extraction Method

Mechanomyography (MMG) signals consist of low-frequency (10–50 Hz) muscular vibrations that are elicited throughout contractions. However, raw MMG recordings obtained using accelerometers contain high-frequency noise and motion artifacts. For the present study, then, signal decomposition was required to isolate the underlying MMG information. To this end, variational mode decomposition (VMD) was used. This involved an algorithm that separated a signal recursively into separate frequency–domain components by solving a constrained variational problem to minimize the mode bandwidths [8,9]. Compared to empirical mode decomposition (EMD), another method widely used for nonstationary signal analysis [10,11], VMD is characterized by greater sampling consistency and better noise robustness.
For the VMD-based MMG extraction method used in the present study, a fundamental preprocessing step was to isolate the 10–50 Hz mechanical vibrations from the raw sensor recordings. Subsequent analyses could then confirm the efficacy of VMD compared to alternatives such as EMD and also serve to characterize the effects of the tuning parameters on the preservation of balance and elimination of noise in relevant MMG signals.
The VMD algorithm had two fundamental parameters that required optimization: the penalty coefficient α and the mode number K. The number of intrinsic mode functions (IMFs) needed to decompose the signal into was determined by K. A too-low value of K raised the risk of insufficient separation and modal mixing; a too-high value of K raised the risk of identifying artificial components [12]. The bandwidths of the acquired IMFs were controlled by the penalty coefficient α, with lower values of α resulting in wider bandwidths. In previous studies, wider bandwidths have been associated with inter-mode leakage, while over-constrained bandwidths have been associated with loss of the informative signal content [13].
In light of the above, we applied differential evolution (DE), a population-based search algorithm, to adaptively tune the values of α and K. The effective use of DE begins with a population of candidate solutions. The solutions are subsequently improved using genetic operators like mutation, crossover, and selection [14]. By means of DE-based metaheuristic parameter optimization, we obtained a data-driven method to determine the penalty coefficient and VMD mode number. Use of this method removed any need for the exhaustive parameters required for tuning on a case-by-case basis.
Our application of differential evolution (DE) involved defining a fitness function to optimize the variational mode decomposition (VMD) parameters. A sparsity metric based on the uniformity of a signal’s probability distribution was provided by envelope entropy [15,16]. Inherent mode functions (IMFs) with higher levels of noise and fewer distinct components exhibited lower envelope entropy and greater sparsity. In contrast, IMFs with greater sparsity exhibited lower entropy. Therefore, envelope entropy was applied as the fitness criterion for VMD DE optimization, with the target of minimizing entropy so that maximally sparse signal decomposition was obtained. The envelope entropy E was defined as follows:
E e = j = 1 N   e j l g e j e j = a ( j ) / j = 1 N   a ( j )
where E_e is the envelope entropy, e_j is the normalized form of a(j), a(j) is the envelope signal of the signal after Hilbert transformation, and N is the number of zero mean signals. The aim of this data-driven entropy minimization approach was to tune VMD to reject artifacts and isolate the MMG signal.

2.2.2. MMG and sEMG Processing Method

MMG and sEMG signals are time series data; consequently, the preprocessing of both was required prior to the estimation of joint movement. The processing pipeline was as follows:
(1)
MMG extraction from raw accelerations using variational mode decomposition optimized by differential evolution (DE-VMD);
(2)
DC offset elimination for MMG and sEMG;
(3)
The 10–450 Hz bandpass filtering of both signals;
(4)
The full-wave rectification of both signals;
(5)
Linear envelope extraction for both signals;
(6)
The normalization of the processed signals.
The aim of this sequential pipeline was to isolate the characteristic MMG vibrations, remove noise and interference, smooth the data, and standardize the scales prior to input into the transformer architecture [17,18]. The entire preprocessing workflow is depicted in Figure 2.
Previous studies on wearable robotics have shown that system response delays in excess of 300 ms are perceivable and can impede normal human movement [19]. In the present study, we estimated offline movement; for any following real-time implementations, the preparation of continuous streaming data within this 300 ms was required. Specifically, it was necessary that signal acquisition, extraction, preprocessing, and model execution should all be completed prior to the next read cycle.
To this end, an input window duration of 1000 ms at increments of 200 ms was configured for the estimation model. This made sure that new estimates were available within the perceptual threshold; at the same time, continuity was ensured by the 800 ms overlap. Furthermore, the quantifications of data transmission and algorithmic contributions to the total delays enabled bottlenecks to be identified and addressed. The above methods were suitable for initial prototyping; however, the development of fully wearable hardware with optimizations for real-time usage is still required if fully practical human–robotic integration is to be achieved.

2.3. Transformer Model

The transformer architecture consisted of decoder and encoder components, as depicted in Figure 3. Both types of components used multi-head self-attention layers, pointwise feedforward layers, and fully connected layers. First, the input sequence was mapped by the encoder into a high-dimensional representation that encoded contextual interdependencies. The decoder then condensed this representation into the target output sequence, using masked self-attention to avoid information leakage. Modeling of long-range temporal relationships was enabled by attention in the data; nonlinear feature transformation was executed by the feedforward layers [20].
  • Encoder
As illustrated in Figure 3, the encoder consisted of stacked identical encoder layers. Each encoder layer was composed of two sublayers: a multi-head self-attention mechanism, followed by a position-wise feedforward neural network. The attention sublayer output was fed into the feedforward network, which used the same transformation at each sequence position. The two sublayers were connected with a residual connection and were then subjected to layer normalization.
2.
Decoder
The decoder consisted of a stack of identical decoder layers, similar to the encoder just described. A self-attention mechanism followed by a feedforward network was also contained in each decoder layer. Additionally, encoder–decoder attention layers were inserted between pairs of sublayers.
3.
Self-attention layer
a. Definition of attention
The definition of attention proposed by Google is as follows:
A t t e n t i o n ( Q , K , V ) = s o f t m a x Q K d k V
where Q ∈ Rn×dk, K ∈ Rm×dk, and V ∈ Rm×dv. The attention layer encodes the sequence Q into a new sequence of n × dv.
b. Multi-head attention
An enhancement of the standard attention mechanism can be represented by multi-head attention in transformers. Formally, it projects the query (Q), key (K), and value (V) inputs through linear transformations to obtain h distinct representations of each.
head i = A t t e n t i o n Q W i Q , K W i K , V W i V
Here, W i Q R d k × d ˜ k , W i K R d k × d ˜ k , a n d   W i V R d v × d ˜ v , and then
M u l t i H e a d ( Q , K , V ) = C o n c a t ( h e a d 1 , , h e a d h )
Finally, we obtain a sequence of n × h d ˜ v . The term ‘multi-head’ refers to doing the same thing repeatedly and then concatenating the results, so that
Y = M u l t i H e a d   ( X ,   X ,   X )
4.
Position embedding
Position embeddings were applied to inject sequential-order information into the transformer model. All positions in the input and output sequences were assigned corresponding vectors via a lookup table. The formula for constructing position embedding may be expressed as follows:
P E 2 i ( p ) = s i n p / 10000 2 i / d p o s P E 2 i + 1 ( p ) = c o s p / 10000 2 i / d p o s
By such means, the position p is mapped to a dpos dimensional position vector, and the value of the ith element of this vector may be expressed as PEi(p).

2.4. Application of a Transformer Model for the Estimation of Joint Acceleration

This model for estimating joint acceleration used the encoder part of the transformer for modeling only. A fully connected layer was used to project the tensor into the shape of [batch size, output length], and the output sequence was the estimating outcome of the input sequence [21]. The construction of the estimation model is illustrated in Figure 4. Application of the transformer-based model for the estimation of joint acceleration involved three key steps, as follows:
  • Padding masks were utilized to pad all input sequences to a uniform length. Masking the added padding tokens to negative infinity enabled the softmax-normalized multi-head attention to ignore these positions effectively. This permitted variable-length inputs to be operated.
  • Sinusoidal position encodings were injected into the input sequences. By assigning to each timestep a unique encoding based on sine/cosine functions at different frequencies, the model could generate sequential order information.
  • The encoded input batch was passed through the encoder, consisting of a projection layer and encoder layers. Hierarchical features were extracted by the encoder layers via feedforward processing and self-attention. The last layer condensed the representation into the target output shape [batch size, output length]. The hyperparameters included batch size, attention heads, input sequence length, output length, number of encoder layers, and learning rate. All of these could be tuned by means of a grid search. After hyperparameter optimization, the best hyperparameter combination that was consistent with these experimental data was as follows: the batch size was 64, the number of input window/the number of output window was 20, the number of attention heads was 2, and the number of encoder layers was 1.

3. Results

In this section, we presented study results that demonstrated the efficacy of the variational mode decomposition (VMD) algorithm for MMG signal isolation and biosignal preprocessing pipelines. We then presented acceleration estimation results and compared these with results obtained using mainstream time series forecasting models. Finally, we reported the performance of the transformer model in different hyperparameters.

3.1. MMG Extraction and Biosignal Processing Results

DE-VMD was utilized to separate the pseudo-accelerations of joints and mechanomyography (MMG) signals from raw accelerometer data, as described above. Case decomposition results are shown in Figure 5 and Figure 6. The results for the original acceleration, pseudo-acceleration, and two IMFs, which are components of MMG, are depicted in Figure 5a. IMF1 was pseudo-acceleration, and IMF2 and IMF3 were components of MMG. The Marginal Hilbert Spectrum of corresponding components is shown in Figure 6a, which reflect the frequency–domain energy distribution. An examination of the figure revealed the effective isolation of the MMG signal (10–50 Hz) from higher-frequency noise and lower-frequency motion artifacts. The Marginal Hilbert Spectrum is a two-dimensional representation derived from the Hilbert spectrum; it shows the total energy contribution of each frequency value. The Marginal Hilbert Spectrum helped us understand the energy distribution across different frequencies in a signal, making it a valuable tool for various applications, including analyzing blood flow, climatic features, and water waves [22,23,24]. The efficacy of DE-VMD was compared to that of alternative methods such as empirical mode decomposition (EMD) with respect to information preservation and artifact rejection.
As previously mentioned, EMD is a method widely utilized for nonstationary signal decomposition. Unlike wavelet and Fourier transforms, which make use of predefined basis functions, EMD is data-driven and intrinsically adapts to signal characteristics. This delivers all the advantages associated with the nonlinear processing of nonstationary data with high noise levels. In the present study, for comparative purposes, EMD was applied to extract MMG signals, and the results are depicted in Figure 5 and Figure 6. The results for raw acceleration, extracted MMG, and low-frequency residue signal are displayed in Figure 5b. IMF1 and IMF2 were components of MMG, and IMF3 was a low-frequency residue signal. The Marginal Hilbert Spectrum of the corresponding components is displayed in Figure 6b. An examination of the figure showed that the EMD-derived MMG exhibits greater randomness and more dispersed spectral content in comparison with the DE-VMD extraction.
Figure 7 and Figure 8 illustrate the MMG and sEMG signals before and after the multistage preprocessing pipeline. It could be seen that raw nonstationary biosignals with high variability were transformed into smoother and more stationary signals after processing. This visual trend suggested the processing methods, including VMD extraction, filtering, rectification, and normalization, adequately prepared the sEMG data and MMG for application as inputs to the transformer-based model for estimating joint acceleration.

3.2. Estimation of Shoulder and Elbow Joint Acceleration Using a Fusion of Signals

To assess the efficacy of the proposed transformer-based method for the estimation of joint acceleration, comparative experiments were conducted using recurrent neural network (RNN) and long short-term memory (LSTM) architectures. LSTMs and RNNs might be seen as representative of mainstream sequence modeling techniques that are widely used today extensively for time series forecasting tasks.
Recurrent neural networks (RNNs) are a class of recursive architectures well-suited for sequential data, as previously stated. RNNs contain cyclical connections that enable the network state at each timestep to depend on prior context [25]. This provides a type of memory in contrast to feedforward networks, which lack an intrinsic temporal mechanism. Nevertheless, standard RNNs struggle to model longer-range dependencies because of vanishing/exploding gradients in the course of backpropagation over time. This limitation might be addressed using long short-term memory (LSTM) networks with gated-cell states, which allow for the persistence of information across extended sequences [25]. While RNNs are not able to link temporally distant context, longer-range patterns can be learnt by LSTMs. By comparing the proposed transformer model with LSTMs and RNNs, we sought to determine whether it would capture complex multimodal biosignal interdependencies more effectively and achieved more precise estimates of joint acceleration compared to the other two methods.
A comparison between the results for the recognition of joint acceleration obtained using the algorithm proposed here and those obtained using traditional algorithms is presented in Figure 9 and Table 1. In the figure, the blue curves are real joint acceleration signals of 3D movement experimentally captured, the yellow curves are estimation results of the transformer model, the gray curves are estimation results of the RNN model, and the brown curves are estimation results of the LSTM model. It can be seen that the transformer model produced the best estimates of joint acceleration, with the lowest rate of overall error among the algorithms, as shown by an estimation curve, which was smooth and without mutation.

3.3. Estimation of Shoulder and Elbow Joint Acceleration Using MMG Signals

A comparison between the results for joint acceleration recognition obtained using the algorithm newly proposed here and those obtained using traditional algorithms is presented in Figure 10 and Table 2. In the figure, the blue curves are real joint acceleration signals of 3D movement experimentally captured, the yellow curves are the estimation results of the transformer model, the gray curves are estimation results of the RNN model, and the brown curves are estimation results of the LSTM model. It can be seen that the transformer model produced the best estimates of joint acceleration, with the lowest rate of overall error among the algorithms.

3.4. Estimation of Shoulder and Elbow Joint Acceleration in Different Hyperparameters

Estimations in different hyperparameters are presented in Table 3. Because the number of features is eight, and the number of attention heads should be divisible by the number of features, so the number of attention heads are arranged as eight, four, two, and one, respectively. After finding the optimal number of attention heads, the optimal values of the other hyperparameters were also found through grid search. The estimation results in different numbers of attention heads are shown in Table 3(a). In addition, the estimation results in different numbers of encoder layers and different numbers of input windows/output windows are shown in Table 3(b,c). Among all kinds of hyperparameters, the number of attention heads, the number of encoder layers, and the number of input windows/output windows are most important, and the number of batch size does not have much impact on the estimation accuracy of the model.

4. Discussion

This study highlighted the following two advantages of the proposed methodology:
(1)
The use of transformer modeling capabilities.
(2)
The fusion of complementary multimodal biosignals.
Firstly, the results shown in Figure 9 and Figure 10 and Table 1 and Table 2 indicated that the precision of the proposed method was better than that of LSTM and RNN. The transformer’s capability to model complex spatiotemporal interdependencies within multimodal sequences was found to exceed the capabilities of LSTMs and RNNs, enabling a more precise inference of movement and further indicating that attention-based modeling captured the intricate relations between neural drives, muscle contractions, and limb accelerations more effectively than LSTM or RNN.
Secondly, the results shown in Table 4 indicated that precision was improved when joint movement was estimated using a fusion of EMG, MMG, and kinematic data compared to using MMG alone. This could be attributed to the heterogenous information conveyed by the different methods in the fusion. Because MMG measured mechanical vibrations and electrical potentials were detected by sEMG, the distinct insights provided by the two methods were combined into recruitment patterns and neuromuscular activation. Moreover, the signals were differentially influenced by factors such as fatigue and muscle length. Therefore, the aim of the fusion was to give a more complete representation of the motor unit behavior that drove movement. However, each constituent method had its own unique limitations, and these must be addressed as part of an effective fusion approach.
In addition, this study found that, due to the particularity of the experimental data, the number of enhanced layers and attention heads could not be too large; otherwise, it would lead to overfitting, which would lead to a decrease in accuracy. When the number of input windows/output windows increased, the accuracy of the model would increase. This was because, when the number of input windows/output windows increased, the model could use more historical data for predictions. However, the degree of increase in the accuracy would reduce as the number of input windows/output windows increased.

5. Conclusions

In this paper, we proposed a framework that combines differential evolution variational mode decomposition (DE-VMD) and transformer modeling to estimate acceleration in human shoulder and elbow joints. To be effective, any such estimation must address longstanding challenges posed by the inherent complexity of human movement, the ambiguity of sensing modalities, and noisy environments. The methodology proposed in the present study aimed to address factors related to both measurement complexities and body dynamics.
Comparative results indicated that the proposed method offered greater stability and superior precision in decoding upper-extremity movement than LSTM and RNN. The efficacy of the VMD method was validated by a systematic evaluation of MMG extraction. A fusion of MMG, sEMG, and kinematic data resulted in an enhanced context compared to a single signal. Ultimately, the capabilities of LSTM and RNN for representation learning were exceeded by the transformer when modeling the intricate spatiotemporal relationships that govern neuromuscular coordination.
The present study must be seen as preliminary in nature; nevertheless, the results reported here might serve to substantiate the utility of combining multimodal signal decomposition, fusion, and deep sequence modeling so that human movement intention might be inferred from wearable sensors. Further validation, involving more complex tasks and other limb joints, is warranted to determine the generalizability. However, the method described in this paper might be seen as a promising paradigm for further enhancing the robustness and precision that were essential for the recognition of human movement intention.
The present study might be seen as promising in its own right; however, several further avenues of research might also be motivated by the work reported in this paper. These include the following:
1. Enhancement of interpretability and model robustness. The current training data encompassed a limited, homogenous cohort performing simple tasks. Enlarging the diversity of subjects, biomechanics, and movements would enrich the dataset to improve generalizability.
2. Optimization of sensor configurations. The use of extra modalities (e.g., EEG), positions and sensor numbers, and on-body integration would help to determine an optimal balance of information gain and user load.
3. Investigation of complementary and alternative sequence models. A comparison of the transformer fusion method reported here with graphical methods and state-of-the-art convolutional techniques might reveal distinct advantages among different architectures.
4. Translation to online decoding on hardware. The introduction of self-contained and wearable platforms might be made possible by system optimization and miniaturization so that the recognition of movement intention can be achieved in real time.
In summary, the preliminary research described in this paper demonstrated the potential utility of a fusion sequence that modeled human movement intention. Further works involving different datasets, sensors, algorithms, and systems will be critical if these biosignals are to be used practically for the robust recognition of body movement.

Author Contributions

Conceptualization, Y.B.; Methodology, Y.B., Z.W., Z.L. and M.Z.; Software, Y.B.; Validation, Y.B.; Data curation, Y.B.; Writing—original draft, Y.B.; Writing—review & editing, Y.B.; Supervision, X.G. and L.H.; Project administration, X.G. and L.H.; Funding acquisition, X.G. and L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Postgraduate Research Practice Innovation Program of Jiangsu Province.

Institutional Review Board Statement

This work involved human subjects or animals in its research. Approval of all ethical and experimental procedures and protocols was granted by the Medical Ethics Committee of Nanjing Medical University under approval number 2021-SR109.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dirgová Luptáková, I.; Kubovčík, M.; Pospíchal, J. Wearable sensor-based human activity recognition with transformer model. Sensors 2022, 22, 1911. [Google Scholar] [CrossRef] [PubMed]
  2. Wensel, J.; Ullah, H.; Munir, A. Vit-ret: Vision and recurrent transformer neural networks for human activity recognition in videos. IEEE Access 2023, 11, 72227–72249. [Google Scholar] [CrossRef]
  3. Shavit, Y.; Klein, I. Boosting inertial-based human activity recognition with transformers. IEEE Access 2021, 9, 53540–53547. [Google Scholar] [CrossRef]
  4. Gardner, M.; Mancero Castillo, C.S. A multimodal intention detection sensor suite for shared autonomy of upper-limb robotic prostheses. Sensors 2020, 20, 6097. [Google Scholar] [CrossRef] [PubMed]
  5. Caulcrick, C.; Huo, W.; Hoult, W.; Vaidyanathan, R. Human joint torque modelling with MMG and EMG during lower limb human-exoskeleton interaction. IEEE Robot. Autom. Lett. 2021, 6, 7185–7192. [Google Scholar] [CrossRef]
  6. Zhang, L.; Liu, G. sEMG based human motion intention recognition. J. Robot. 2019, 2019, 3679174. [Google Scholar] [CrossRef]
  7. Vaswani, A.; Shazeer, N. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  8. Ur Rehman, N.; Aftab, H. Multivariate variational mode decomposition. IEEE Trans. Signal Process. 2019, 67, 6039–6052. [Google Scholar] [CrossRef]
  9. Dragomiretskiy, K.; Zosso, D. Variational mode decomposition. IEEE Trans. Signal Process. 2013, 62, 531–544. [Google Scholar] [CrossRef]
  10. Rehman, N.; Mandic, D.P. Multivariate empirical mode decomposition. Proc. R. Soc. A Math. Phys. Eng. Sci. 2010, 466, 1291–1302. [Google Scholar] [CrossRef]
  11. Huang, N.E.; Shen, Z. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. London Ser. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  12. Xie, C.; Wang, D. Mechanomyography signals processing method using multivariate variational mode decomposition. In Proceedings of the 2021 14th International Symposium on Computational Intelligence and Design (ISCID), Hangzhou, China, 11–12 December 2021; pp. 278–281. [Google Scholar]
  13. Nagineni, S.; Taran, S. Features based on variational mode decomposition for identification of neuromuscular disorder using EMG signals. Health Inf. Sci. Syst. 2018, 6, 13. [Google Scholar] [CrossRef] [PubMed]
  14. Pant, M.; Zaheer, H. Differential Evolution: A review of more than two decades of research. Eng. Appl. Artif. Intell. 2020, 90, 103479. [Google Scholar]
  15. Ding, J.; Huang, L. GMPSO-VMD algorithm and its application to rolling bearing fault feature extraction. Sensors 2020, 20, 1946. [Google Scholar] [CrossRef]
  16. Zhang, F.; Sun, W. Fault diagnosis of a wind turbine gearbox based on improved variational mode algorithm and information entropy. Entropy 2021, 23, 794. [Google Scholar] [CrossRef] [PubMed]
  17. Shi, Y.; Dong, W. Human Joint Torque Estimation Based on Mechanomyography for Upper Extremity Exosuit. Electronics 2022, 11, 1335. [Google Scholar] [CrossRef]
  18. Chen, J.; Zhang, X. Surface EMG based continuous estimation of human lower limb joint angles by using deep belief networks. Biomed. Signal Process. Control 2018, 40, 335–342. [Google Scholar] [CrossRef]
  19. Englehart, K.; Hudgins, B. A robust, real-time control scheme for multifunction myoelectric control. IEEE Trans. Biomed. Eng. 2003, 50, 848–854. [Google Scholar] [CrossRef]
  20. Ahmed, S.; Nielsen, I.E. Transformers in time-series analysis: A tutorial. Circuits Syst. Signal Process. 2023, 42, 7433–7466. [Google Scholar] [CrossRef]
  21. Sahoo, B.B.; Jha, R. Long short-term memory (LSTM) recurrent neural network for low-flow hydrological time series forecasting. Acta Geophys. 2019, 67, 1471–1481. [Google Scholar] [CrossRef]
  22. Huang, W.; Shen, Z.; Huang, N.E. Engineering analysis of biological variables: An example of blood pressure over 1 day. Proc. Natl. Acad. Sci. USA 1998, 95, 4816–4821. [Google Scholar] [CrossRef]
  23. Bengulescu, M.; Blanc, P.; Wald, L. Characterizing temporal variability in measurements of surface solar radiation and its dependence on climate. Energy Procedia 2016, 97, 164–171. [Google Scholar] [CrossRef]
  24. Schlurmann, T. Spectral analysis of nonlinear water waves based on the Hilbert-Huang transformation. J. Offshore Mech. Arct. Eng. 2002, 124, 22–27. [Google Scholar] [CrossRef]
  25. Sherratt, F.; Plummer, A.; Iravani, P. Understanding LSTM network behaviour of IMU-based locomotion mode recognition for applications in prostheses and wearables. Sensors 2021, 21, 1264. [Google Scholar] [CrossRef]
Figure 1. (a) The arrangement of sensors. (b) The composition of the special sensor.
Figure 1. (a) The arrangement of sensors. (b) The composition of the special sensor.
Sensors 24 01726 g001
Figure 2. Multisource signal processing flowchart.
Figure 2. Multisource signal processing flowchart.
Sensors 24 01726 g002
Figure 3. Transformer model.
Figure 3. Transformer model.
Sensors 24 01726 g003
Figure 4. Transformer model for the estimation of joint acceleration.
Figure 4. Transformer model for the estimation of joint acceleration.
Sensors 24 01726 g004
Figure 5. (a) Results of VMD in the time domain. (b) Results of EMD in the time domain.
Figure 5. (a) Results of VMD in the time domain. (b) Results of EMD in the time domain.
Sensors 24 01726 g005aSensors 24 01726 g005b
Figure 6. (a) Marginal Hilbert Spectrum of VMD. (b) Marginal Hilbert Spectrum of EMD.
Figure 6. (a) Marginal Hilbert Spectrum of VMD. (b) Marginal Hilbert Spectrum of EMD.
Sensors 24 01726 g006aSensors 24 01726 g006b
Figure 7. (a) Raw sEMG signal. (b) Raw MMG signal.
Figure 7. (a) Raw sEMG signal. (b) Raw MMG signal.
Sensors 24 01726 g007aSensors 24 01726 g007b
Figure 8. (a) Prepared sEMG signal. (b) Prepared MMG signal.
Figure 8. (a) Prepared sEMG signal. (b) Prepared MMG signal.
Sensors 24 01726 g008aSensors 24 01726 g008b
Figure 9. (a) Estimation results for arm side raises using a fusion of signals. (b) Estimation results for arm front raises using a fusion of signals. (c) Estimation results for elbow curls using a fusion of signals.
Figure 9. (a) Estimation results for arm side raises using a fusion of signals. (b) Estimation results for arm front raises using a fusion of signals. (c) Estimation results for elbow curls using a fusion of signals.
Sensors 24 01726 g009aSensors 24 01726 g009b
Figure 10. (a) Estimation results for arm side raises using MMG signals. (b) Estimation results for arm front raises using MMG signals. (c) Estimation results for elbow curls using MMG signals.
Figure 10. (a) Estimation results for arm side raises using MMG signals. (b) Estimation results for arm front raises using MMG signals. (c) Estimation results for elbow curls using MMG signals.
Sensors 24 01726 g010aSensors 24 01726 g010b
Table 1. (a) Estimation results for arm side raises using a fusion of signals. (b) Estimation results for arm front raises using a fusion of signals. (c) Estimation results for elbow curls using a fusion of signals.
Table 1. (a) Estimation results for arm side raises using a fusion of signals. (b) Estimation results for arm front raises using a fusion of signals. (c) Estimation results for elbow curls using a fusion of signals.
(a)
AlgorithmHighest Value of R2Lowest Value of R2Mean Value of R2
Transformer0.9760.9420.967
RNN0.9560.9260.941
LSTM0.9530.9170.938
(b)
AlgorithmHighest Value of R2Lowest Value of R2Mean Value of R2
Transformer0.9790.9450.968
RNN0.9660.9310.945
LSTM0.9610.9220.939
(c)
AlgorithmHighest Value of R2Lowest Value of R2Mean Value of R2
Transformer0.9570.9010.935
RNN0.9150.8520.889
LSTM0.9310.8810.899
Table 2. (a) Estimation results for arm side raises using MMG signals. (b) Estimation results for arm front raises using MMG signals. (c) Estimation results for elbow curls using MMG signals.
Table 2. (a) Estimation results for arm side raises using MMG signals. (b) Estimation results for arm front raises using MMG signals. (c) Estimation results for elbow curls using MMG signals.
(a)
AlgorithmHighest Value of R2Lowest Value of R2Mean Value of R2
Transformer0.9520.9020.926
RNN0.9230.8610.897
LSTM0.9250.8770.902
(b)
AlgorithmHighest Value of R2Lowest Value of R2Mean Value of R2
Transformer0.9550.8870.915
RNN0.9290.8720.903
LSTM0.9260.8740.888
(c)
AlgorithmHighest Value of R2Lowest Value of R2Mean Value of R2
Transformer0.9310.8520.892
RNN0.8670.8260.841
LSTM0.9020.8360.857
Table 3. (a) Estimation results for arm side raises in different numbers of attention heads. (b) Estimation results for arm side raises considering the number of encoder layers. (c) Estimation results for arm side raises in different numbers of input windows/output windows.
Table 3. (a) Estimation results for arm side raises in different numbers of attention heads. (b) Estimation results for arm side raises considering the number of encoder layers. (c) Estimation results for arm side raises in different numbers of input windows/output windows.
(a)
the Number of Attention HeadsHighest Value of R2Lowest Value of R2Mean Value of R2
80.9190.8560.879
40.9450.8810.911
20.9790.9260.955
10.9520.8960.918
(b)
the Number of Encoder LayersHighest Value of R2Lowest Value of R2Mean Value of R2
200.6110.5520.585
150.7610.7260.745
100.8750.8510.863
50.9410.8760.915
20.970.9310.953
10.9790.9260.955
(c)
the Number of Input windows/Output WindowsHighest Value of R2Lowest Value of R2Mean Value of R2
50.7410.6180.667
100.8020.7040.752
150.8950.8290.854
200.9790.9260.955
400.9780.9270.955
Table 4. (a) Estimation results for arm side raises using the transformer model. (b) Estimation results for arm front raises using the transformer model. (c) Estimation results for elbow curls using the transformer model.
Table 4. (a) Estimation results for arm side raises using the transformer model. (b) Estimation results for arm front raises using the transformer model. (c) Estimation results for elbow curls using the transformer model.
(a)
Signal TypeHighest Value of R2Lowest Value of R2Mean Value of R2
Fusion signal0.9760.9420.967
MMG signal0.9520.9020.926
(b)
Signal TypeHighest Value of R2Lowest Value of R2Mean Value of R2
Fusion signal0.9790.9450.968
MMG signal0.9550.8870.915
(c)
Signal TypeHighest Value of R2Lowest Value of R2Mean Value of R2
Fusion signal0.9570.9010.935
MMG signal0.9310.8520.892
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bai, Y.; Guan, X.; He, L.; Wang, Z.; Li, Z.; Zhu, M. Estimating Rotational Acceleration in Shoulder and Elbow Joints Using a Transformer Algorithm and a Fusion of Biosignals. Sensors 2024, 24, 1726. https://doi.org/10.3390/s24061726

AMA Style

Bai Y, Guan X, He L, Wang Z, Li Z, Zhu M. Estimating Rotational Acceleration in Shoulder and Elbow Joints Using a Transformer Algorithm and a Fusion of Biosignals. Sensors. 2024; 24(6):1726. https://doi.org/10.3390/s24061726

Chicago/Turabian Style

Bai, Yu, Xiaorong Guan, Long He, Zheng Wang, Zhong Li, and Meng Zhu. 2024. "Estimating Rotational Acceleration in Shoulder and Elbow Joints Using a Transformer Algorithm and a Fusion of Biosignals" Sensors 24, no. 6: 1726. https://doi.org/10.3390/s24061726

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop