Skip to Content
Applied SciencesApplied Sciences
  • Article
  • Open Access

25 July 2021

Two-Step Biometrics Using Electromyogram Signal Based on Convolutional Neural Network-Long Short-Term Memory Networks

,
and
1
IT Research Institute, Chosun University, 309 Pilmun-daero, Dong-gu, Gwangju 61452, Korea
2
Interdisciplinary Program in IT-Bio Convergence System, Chosun University, 309 Pilmun-daero, Dong-gu, Gwangju 61452, Korea
*
Author to whom correspondence should be addressed.

Abstract

Electromyogram (EMG) signals cannot be forged and have the advantage of being able to change the registered data as they are characterized by the waveform, which varies depending on the gesture. In this paper, a two-step biometrics method was proposed using EMG signals based on a convolutional neural network–long short-term memory (CNN-LSTM) network. After preprocessing of the EMG signals, the time domain features and LSTM network were used to examine whether the gesture matched, and single biometrics was performed if the gesture matched. In single biometrics, EMG signals were converted into a two-dimensional spectrogram, and training and classification were performed through the CNN-LSTM network. Data fusion of the gesture recognition and single biometrics was performed in the form of an AND. The experiment used Ninapro EMG signal data as the proposed two-step biometrics method, and the results showed 83.91% gesture recognition performance and 99.17% single biometrics performance. In addition, the false acceptance rate (FAR) was observed to have been reduced by 64.7% through data fusion.

1. Introduction

Biometrics refers to the verification of identity using an individual’s behavioral and physical features. Currently, widely used biometrics include face, fingerprint, and iris recognition as physical features. However, faces, fingerprints, and irises are bodily information without liveness that can be forged using 3D printers, lenses, and silicon, raising security concerns [1]. To compensate for this issue, biometrics studies using biosignals, which are a behavioral feature, are being conducted. Biosignals are signals that measure the flow of microcurrents generated through biological activities, which cannot be replicated outside the body [2]. In addition, biosignals are signals with liveness, thereby having strong security advantages as they cannot be forged and falsified [3]. Examples of biosignals include electromyogram (EMG), electrocardiogram (ECG), and electroencephalogram (EEG) signals. Among them, the EMG signal represents a signal obtained in the form of a voltage value by measuring a microcurrent generated whenever there is a muscle contraction. There are two methods of measuring EMG signals: a needle electrode method and a surface electrode method. The needle electrode method measures an action potential occurring at a point in the muscle by inserting a needle electrode into a muscle. The surface electrode method measures an action potential by placing electrodes to the skin surface. Measurement of the EMG signal using the surface electrode method wherein an electrode is placed to the skin surface is shown in Figure 1 [1]. In terms of the electrode placement position, the signal is measured by placing the electrode at a muscle which change during experimentation or a muscle position that is frequently used in existing studies.
Figure 1. Electromyogram (EMG) signal measurement using the surface electrode method.
Biometrics studies using EMG signals are lacking compared to studies using ECG and EEG signals. However, the EEG signal gets distorted as it passes through the skull, which is a disadvantage while measuring the signal, and the ECG signal is a periodic signal, which is a drawback by itself in that the waveform of the signal cannot be changed if the registered data is hacked. By comparison, the EMG signal can be measured more easily than the EEG signal, and its registration data can be changed distinctly, unlike the ECG signal, using a feature representing a different waveform of a signal for each gesture. EMG signals in existing works have been used as gesture recognition or single biometrics.
In this paper, we proposed a method for using EMG signal-based gesture recognition and single biometrics together. The proposed method is a two-step biometrics method that uses CNN-LSTM network-based EMG signals. The time domain characteristics of the EMG signal and the LSTM network were used to determine whether the user’s gesture matches. Single biometrics was performed using the spectrogram and CNN-LSTM network if it matched the registered gesture. Data fusion of gesture recognition and single biometrics was performed in the form of AND in which the false acceptance rate (FAR) was lowered. As a result of the experiment, it was confirmed that the FAR decreased by 64.7%. The remainder of this paper is structured as follows. Existing gesture recognition and biometrics studies using bioinformation are mentioned in Section 2. A two-step biometrics method using EMG signals based on the proposed CNN-LSTM network is described in Section 3. The results of the two-step biometrics experiment using our proposed method are explained in Section 4 and Section 5 is the conclusion.

3. Materials and Methods

The two-step biometrics method using EMG signals based on the CNN-LSTM network proposed is shown in Figure 2. Noise in the signal was removed, and the EMG signal was then aligned using RMS to perform two-step biometrics using the EMG signal. After preprocessing the EMG signal, gesture recognition was performed using time domain features and the LSTM network to examine whether it matched the registered gesture, and single biometrics was performed if the gesture matched. For single biometrics, a one-dimensional EMG signal was converted into a spectrogram in the form of a two-dimensional image, and features were extracted and trained using the designed CNN-LSTM network. When the EMG signal for the two-step biometrics test was input, the above-mentioned preprocessing, gesture recognition, and spectrogram conversion were performed. Subsequently, the individual was classified using the information trained on the designed network.
Figure 2. Convolutional neural network-long short-term memory (CNN-LSTM) network-based two-step biometrics flowchart.

3.1. EMG Signal Preprocessing

The EMG signal is in the form of a voltage value obtained by measuring microcurrent generated when a muscle contracts, and it contains considerable gesture information in a low-frequency band (500 Hz or below). The measured EMG signal contains various types of noise. The types of noise included in the EMG signal are baseline wander, power line noise, and external environmental noise [23,24]. In this study, NF was used in the 60 Hz band and BPF was used in the 5–500 Hz band to remove noise. Figure 3 represents changes in EMG signals in the time-frequency domains before and after preprocessing, with blue signals representing EMG signals before preprocessing and red signals representing EMG signals after preprocessing. It can be seen in Figure 3 (in red box) that NF application removed the 60 Hz band and BPF application passed the 5–500 Hz band. The EMG signal from which noise had been removed was divided according to purpose after aligning the EMG signal using RMS prior to performing gesture recognition.
Figure 3. Comparison before and after EMG signal preprocessing in time-frequency domain of a notch filter (NF) and a band pass filter (BPF).

3.2. Gesture Recognition Method Using EMG Signals

For the two-step biometrics methods proposed herein, gesture recognition was performed as shown in Figure 4. Training and classification were performed with an LSTM network after extracting five time domain features from the preprocessed EMG signal, and the time domain features used in the experiment were mean absolute value (MAV), variance (VAR), integrated EMG (IEMG), MAV slope (MAVSLP), and waveform length (WL). MAV can detect muscle contraction, which is a feature widely used in control applications using EMG signals. VAR uses the power of the EMG signal as a feature, representing the power information of the EMG signal. IEMG is a feature used as an index to detect muscle activity, representing the force applied to a muscle. MAVSLP is an improved feature of MAV that can be calculated using the difference between adjacent MAVs, and WL is a feature representing the difference in the length of the waveform of the EMG signal. The five time domain features can be calculated using the formula shown in Figure 4, where x i denotes the EMG signal and N denotes the signal length [25,26].
Figure 4. Flowchart of gesture recognition using time domain features and the LSTM network.
Training and classification were performed using an LSTM network after the features were extracted from the time domain of the EMG signal. The LSTM network is characterized by maintaining sequence information in the time domain, thereby an important role in improving classification accuracy [27]. In addition, the LSTM network is a structure that compensates for the disadvantage of the vanishing gradient problem of RNN. Four gates (forget, input, output, updating) can be used to selectively transfer or delete information from sequential inputs to maintain dependencies in the sequence, as shown in Figure 5.
Figure 5. Basic LSTM network structure.
The LSTM network had two transfer parameters, c ( t ) and h ( t ) , calculated by using four gates. The forget gate was calculated using Equation (1), and the input gate was calculated using Equation (2). Parameters c ( t ) and h ( t ) to be transferred to the next LSTM network were calculated using Equations (3) and (4), respectively. In the equation, σ denotes the sigmoid function, denotes the convolution, and W denotes the weight. In this paper, we performed gesture recognition using an LSTM network consisting of four LSTM layers and two fully connected layers. The layer placement and parameters of the network were selected through experiments, and the designed network deployed one dropout (0.3 probability) for two LSTM networks. An LSTM network uses a many-to-one structure that receives the time domain characteristics of a multi EMG channel as input and outputs the gesture class. Network learning is done by setting adam optimizer, hidden unit 512 and batch size 128, and epoch with 100.
f t = σ ( W x f x t + W h f h t 1 + b f )
i t = σ ( W x i x t + W h i h t 1 + b i )
c t = f t c t 1 + i t tan h ( W x c x t + W h c h t 1 + b c )
h t = σ ( W x h x t + W h h h t 1 + b h )

3.3. Single Biometrics Method Using EMG Signals

The single biometrics method among the two-step biometrics methods proposed herein used a spectrogram and the CNN-LSTM network. The time domain of the EMG signal represents information on muscle activity, and the frequency domain has the advantage of being immune to noise [1]. Therefore, the one-dimensional EMG signal was converted into a two-dimensional spectrogram to simultaneously analyze time–frequency domain information. The spectrogram was generated using a window function of a fixed length within the time domain of a given signal. In the generated spectrogram, the horizontal axis represented time, and the vertical axis represented frequency information.
In the case of biometrics using the existing EMG signal, a handcrafted method is performed wherein a subject extracts features by using a specific feature extraction algorithm. However, this method lacks versatility and has a problem in that it may not be able to extract features optimized for a signal. To solve this problem, deep learning methods such as CNN and LSTM networks, which have been actively studied recently, can automatically extract optimal features from data without extracting features via the handcrafted method [28].
A CNN-LSTM network was designed herein to perform single biometrics using EMG signals as shown in Figure 6. In this paper, a single biometrics was performed using a network of three convolutional layers, three pooling layers, two LSTM layers, and two fully connected layers. Batch normalization was performed between each convolutional layer and the pooling layer, and a dropout (0.3 probability) was placed after three convolutional layers. The layer placement and parameters of the network were selected through experiments, and a rectified linear unit (ReLU) was used as the activation function. Network learning was done by setting sgdm optimizer, batch size 50, and epoch with 100. The spectrogram input to the network was normalized to 256 × 256, and the output data size, padding, stride, and number of filters used in the experiment are shown in Table 2. The stride and padding were set to 1 and 2, respectively, to ensure that the data size remained unchanged in the convolutional layer, and the stride and padding were set to 2 and 0, respectively, to reduce the data size in the pooling layer.
Figure 6. CNN-LSTM network structure for single biometrics.
Table 2. Designed CNN-LSTM network parameters.

4. Results and Discussion

The publicly available EMG signal dataset Ninapro DB2 was used to verify the performance of the two-step biometrics method using EMG signals based on the CNN-LSTM network proposed herein. The Ninapro DB2 comprises hand gestures performed by a subject, and hand gestures can be divided into three groups (groups 1–3). The data used in the experiment were EMG signals of seventeen gestures measured within group 1, and the experiment was conducted with twenty subjects. EMG signal measurements were collected at twelve channels using Delsys Trigno Wireless EMG equipment and measured at a 2000 Hz sampling rate. When measuring the EMG data, each movement lasted for five seconds, and a three seconds rest period was performed between movements. The subjects were asked to move the right hand repeatedly while measuring the EMG data and each movement was repeated six times [29].
Experimental performance was evaluated using accuracy, recall, and precision, and the above evaluation indicators can be calculated as true positive ( T P ), true negative ( T N ), false positive ( F P ), and false negative ( F N ). Accuracy refers to the performance of when the model’s true label predicts true and false label predicts false, which is an evaluation index representing the most intuitive performance of the model. The method to obtain accuracy is shown in Equation (5). Recall refers to the ratio of the model’s true label predicts true or false, which can be calculated through T P and F N as shown in Equation (6). Precision refers to the actual ratio of true classifications among all classifications by the trained model, which can be calculated through T P and F P as shown in Equation (7).
accuracy = ( T P + T N ) ( T P + T N + F P + F N )
recall = T P ( T P + F N )
precision = T P ( T P + F P )
The data used in the two-step biometrics experiment were recorded from each gesture repeated six times by twenty subjects. However, there is a problem that the amount of training and test data were insufficient when using an LSTM network for gesture recognition and a CNN-LSTM network for single biometrics. Therefore, three-fold cross-validation (each subject used data from four training sets and two test sets) was performed for gesture recognition (1:N comparison) and single biometrics (1:N comparison).
The results of the gesture recognition experiment using time domain features and LSTM network are as shown in Table 3. As a result of the experiment, the basic gestures of the seventeen hands showed an average of 83.91% accuracy, 91.26% recall, and 91.15% precision performance. In terms of the recognition performance by gesture, the 16th gesture was the best classified at 99.62%, and the 11th gesture was the worst classified. The 11th gesture was most often misclassified as the 9th gesture, indicating similar hand gestures.
Table 3. Gesture recognition experiment results using the time domain features and LSTM.
The results of comparative experiments with conventional gesture recognition studies using EMG signals are shown in Table 4. The EMG signal has a characteristic that shows different waveforms for each gesture when the same subject performs different gestures. Therefore, the performance of gesture recognition using the EMG degraded as the number of classification gesture classes increased. In this paper, we conducted more hand gesture classification experiments than other related works by utilizing seventeen gestures. In addition, the data size used in the experiment was 60ea, which is smaller than in related works (except [4]).
Table 4. Gesture recognition comparison experiment.
The results of the single biometrics experiment using spectrogram and CNN-LSTM network are as shown in Table 5. The experiments were conducted correctly using all seventeen gestures from Ninapro DB group 1. As a result of the experiment, twenty subjects showed an average of 99.17% accuracy, 99.76% recall, and 99.41% precision performance. Nine people (4th, 5th, 7th, 8th, 11th, 12th, 15th, 19th, and 20th) were classified 100%, and the 14th subject was classified 97.06%, showing the worst performance. The 14th subject was most often misclassified as the 11th subject, and the 9th, 10th, and 11th gestures were most often misclassified, due to the similarity of the hand gestures.
Table 5. Results of single biometrics experiments using the CNN-LSTM network.
The results of a comparative experiment with the conventional single biometrics study using EMG signals are shown in Table 6. The EMG signal has a characteristic that shows different waveforms for each gesture when the same subject performs different gestures. Therefore, biometrics experiments with a large number of gestures degrade performance. The single biometrics method using EMG signals based on the CNN-LSTM network performed herein was tested with seventeen gestures in Ninapro DB group 1. By comparison, most of the related studies conducted biometrics experiments using one gesture, and no study used more than six gestures. In addition, we showed 99.17% performance on twenty subjects, which was more subjects than in related works (except [12,16]).
Table 6. Single biometrics comparison experiment.
Figure 7 shows the 9th, 10th, and 11th gestures as performed by the 11th and 14th subjects, which showed the highest misclassification rates in a single biometrics test. The Ninapro DB 9th, 10th, and 11th gestures are EMG signals for performing ‘wrist supination’ operations, measured by varying axis and rotational direction. Therefore, we find that classification performance is low due to the use of EMG signals for similar hand gestures.
Figure 7. Ninapro DB (database) 11th and 14th subjects EMG signal.
The two-step biometrics method proposed herein performs single biometrics after gesture recognition, wherein data fusion of the two experimental results is performed in an AND form. This data fusion has an advantage in that the FAR can be lowered when it is performed in an AND form. Results of the comparison of FAR between two-step biometrics performing data fusion in an AND form and single biometrics is shown in Figure 8. As a result of the experiment, it was confirmed that the FAR of the two-step biometrics that was data fused in AND form was 0.0155%, which was 64.7% lower than that of single biometrics. As the FAR decreased, the probability of misclassifying an unregistered subject as a registered subject decreased [22]. Therefore, it was confirmed that the EMG based two-step biometrics method proposed in this paper was less likely to misclassify unregistered subjects as registered subjects. In addition, the existing biometrics problem in which registration data cannot be changed was solved by adding a gesture recognition process.
Figure 8. Comparison of false acceptance rate (FAR) between single biometrics and two-step biometrics.

5. Conclusions

A two-step biometrics method using EMG signals based on the CNN-LSTM network was proposed. Noise in the EMG signal was removed using NF and BPF, and preprocessing was performed with RMS. After preprocessing the EMG signal, the time domain features and LSTM network were used to examine whether the gesture matched, and single biometrics was performed if the gesture matched. In single biometrics, EMG signals were converted into a two-dimensional spectrogram, and training and classification were performed through the CNN-LSTM network. Data fusion of the gesture registration and single biometrics was performed in the form of an AND, which lowered FAR, thus proving that the proposed method was less likely to misclassify unregistered subjects as registered subjects. A publicly available EMG signal dataset (Ninapro DB) was used to evaluate the performance of the proposed two-step biometrics method using EMG signals. As a result of the experiment, seventeen hand gestures were classified with an accuracy of 83.91%, and twenty people were classified with an accuracy of 99.17%. In addition, we found that the fusion of gesture recognition and single biometrics reduced the FAR, which is important in biometrics. A two-step biometrics experiment will be conducted with a greater number of subjects and hand gestures in the future, and the hand gesture recognition performance using EMG signals will be improved. In addition, check the performance according to the spectrogram image size, and the intra- and inter-individual differences depending on each gesture of the EMG signal will be examined.

Author Contributions

Conceptualization, J.-S.K. and S.-B.P.; methodology, J.-S.K. and S.-B.P.; software, J.-S.K.; validation, J.-S.K., M.-G.K. and S.-B.P.; formal analysis, J.-S.K.; investigation, J.-S.K.; writing—original draft preparation, J.-S.K.; writing—review and editing, J.-S.K., M.-G.K. and S.-B.P.; supervision, S.-B.P.; project administration, S.-B.P.; funding acquisition, S.-B.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (No. 2017R1A6A1A03015496), the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. NRF-2021R1A2C1014033), and the HPC Support Project, supported by the Ministry of Science and ICT and NIPA.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data supporting the findings of the article are available in the Ninapro DB2 at [29].

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

EMGElectromyogram
CNN-LSTMconvolutional neural network-long short-term memory
FARfalse acceptance rate
ECGelectrocardiogram
EEGelectroencephalogram
RMSroot mean square
ZCzero crossing
KNNK-nearest neighbor
CWTcontinuous wavelet transform
PCAprincipal component analysis
GRNNgeneral regression neural network
BPFband pass filter
DBdatabase
EDeuclidean distance
SVMsupport vector machine
UDPunsupervised discriminant projection
CFAclass-dependent feature analysis
HPFhigh pass filter
LPFlow pass filter
NFnotch filter
ANNsartificial neural networks
LDAlinear discriminant analysis
DWTdiscrete wavelet transform
OCSVMone-class support vector machine
LOFlocal outlier factor
EMDempirical mode decomposition
ACautocorrelation
FFTfourier transform
OPFoptimum patch forest
DCTdiscrete cosine transform
BPbackpropagation
MFmean frequency
MAVmean absolute value
VARvariance
IEMGintegrated EMG
MAVSLPMAV slope
WLwaveform length
ReLUrectified linear unit
TPtrue positive
TNtrue negative
FPfalse positive
FNfalse negative

References

  1. Kim, J.S.; Pan, S.B. A study on EMG-based biometrics. J. Internet Serv. Inf. Secur. 2017, 7, 19–31. [Google Scholar]
  2. Kim, J.S.; Kim, S.H.; Pan, S.B. Personal recognition method using coupling image of ECG signal. Smart Media J. 2019, 8, 62–69. [Google Scholar] [CrossRef]
  3. Belgacem, N.; Fournier, R.; Ali, A.N.; Reguig, F.B. A novel biometric authentication approach using ECG and EMG signals. J. Med. Eng. Technol. 2015, 39, 226–238. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, X.; Yang, Z.; Chen, T.; Chen, D.; Huang, M.C. Cooperative sensing and wearable computing for sequential hand gesture recognition. IEEE Sens. J. 2019, 19, 5575–5583. [Google Scholar] [CrossRef]
  5. Jaramillo, A.G.; Benalcazar, M.E. Real-time hand gesture recognition with EMG using machine learning. In Proceedings of the IEEE Ecuador Technical Chapters Meeting, Salinas, Ecuador, 16–20 October 2017. [Google Scholar]
  6. Oh, D.C.; Jo, Y.U. EMG-based hand gesture classification by scale average wavelet transform and CNN. In Proceedings of the International Conference on Control, Automation and Systems, Jeju, Korea, 15–18 October 2019. [Google Scholar]
  7. Qi, J.; Jiang, G.; Li, G. Surface EMG hand gesture recognition system based on PCA and GRNN. Neural Comput. Appl. 2020, 32, 6343–6351. [Google Scholar] [CrossRef]
  8. Chen, L.; Fu, J.; Wu, Y.; Li, H.; Zheng, B. Hand gesture recognition using compact CNN via surface electromyography signals. Sensors 2020, 20, 672. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Asif, A.R.; Waris, A.; Gilani, S.O.; Jamil, M.; Ashraf, H.; Shafique, M.; Niazi, K. Performance evaluation of convolutional neural network for hand gesture recognition using EMG. Sensors 2020, 20, 1642. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Shreyas, V.; Felix, J.X.; Benjamin, C.; Marios, S. Electromyograph and keystroke dynamics for spoof-resistant biometric authentication. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, UAS, 8–10 June 2015. [Google Scholar]
  11. Shin, S.H.; Jung, J.H.; Kim, Y.T. A study of an EMG-based authentication algorithm using an artificial neural network. In Proceedings of the IEEE Sensors, Glasgow, UK, 29 October–1 November 2017. [Google Scholar]
  12. Lu, L.; Mao, J.; Wang, W.; Ding, G.; Zhang, Z. An EMG-based personal identification method using continuous wavelet transform and convolutional neural networks. In Proceedings of the IEEE Biomedical Circuits and Systems Conference, Nara, Japan, 17–19 October 2019. [Google Scholar]
  13. Shioji, R.; Ito, S.; Ito, M.; Fukumi, M. Personal authentication and hand motion recognition based on wrist EMG analysis by a convolutional neural network. In Proceedings of the IEEE International Conference on Internet of Things and Intelligence System, Bali, Indonesia, 1–3 November 2018. [Google Scholar]
  14. Lee, M.; Ryu, J.; Youn, I. Biometric personal identification based on gait analysis using surface EMG signals. In Proceedings of the IEEE International Conference on Computational Intelligence and Applications, Beijing, China, 8–11 September 2017. [Google Scholar]
  15. Morikawa, S.; Ito, S.; Ito, M.; Fukumi, M. Personal authentication by lips EMG using dry electrode and CNN. In Proceedings of the IEEE International Conference on Internet of Things and Intelligence System, Bali, Indonesia, 1–3 November 2018. [Google Scholar]
  16. Lu, L.; Mao, J.; Wang, W.; Ding, G.; Zhang, Z. A study of personal recognition method based on EMG signal. IEEE Trans. Biomed. Circuits Syst. 2020, 14, 681–691. [Google Scholar] [CrossRef] [PubMed]
  17. Li, Q.; Dong, P.; Zheng, J. Enhancing the security of pattern unlock with surface EMG-based biometrics. Appl. Sci. 2020, 10, 541. [Google Scholar] [CrossRef] [Green Version]
  18. Khan, M.U.; Choudry, Z.A.; Aziz, S.; Naqvi, S.Z.H.; Aymin, A.; Imtiaz, M.A. Biometric authentication based on EMG signals of speech. In Proceedings of the International Conference on Electrical, Communication, and Computer Engineering, Istanbul, Turkey, 12–13 June 2020. [Google Scholar]
  19. Liu, L.; Gu, X.F.; Li, J.P.; Lin, J.; Shi, J.X.; Huang, Y.Y. Research on data fusion of multiple biometric features. In Proceedings of the International Conference on Apperceiving Computing and Intelligence Analysis, Chengdu, China, 23–25 October 2009. [Google Scholar]
  20. Kumar, A.; Zhang, D. Improving biometric authentication performance from the user quality. IEEE Trans. Instrum. Meas. 2010, 59, 730–735. [Google Scholar] [CrossRef]
  21. Rahman, M.W.; Gavrilova, M. Emerging EEG and kinect face fusion for biometric identification. In Proceedings of the IEEE Symposium Series on Computational Intelligence, Honolulu, HI, USA, 27 November–1 December 2017. [Google Scholar]
  22. Faundez-Zanuy, M. Data fusion in biometrics. IEEE Aerosp. Electron. Syst. Mag. 2005, 20, 34–38. [Google Scholar] [CrossRef]
  23. Choi, Y.J.; Yu, H.J. Human motion tracking based on sEMG signal processing. J. Inst. Control Robot. Syst. 2007, 13, 769–776. [Google Scholar]
  24. Lee, C.H.; Kang, S.I.; Bae, S.H.; Kwon, J.W.; Lee, D.H. A study of a module of wrist direction recognition using EMG signals. J. RWEAT 2013, 7, 51–58. [Google Scholar]
  25. Phinyomark, A.; Limsakul, C.; Phukpattaranont, P. A novel feature extraction for robust EMG pattern recognition. J. Comput. 2009, 1, 71–80. [Google Scholar]
  26. Phinyomark, A.; Hirunviriya, S.; Limsakul, C.; Phukpattaranont, P. Evaluation of EMG feature extraction for hand movement recognition based on Euclidean distance and standard deviation. In Proceedings of the International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology, Chiang Mai, Thailand, 19–21 May 2010. [Google Scholar]
  27. Choi, G.H.; Bak, E.S.; Pan, S.B. User identification system using 2D resized spectrogram features of ECG. IEEE Access 2019, 7, 34862–34873. [Google Scholar] [CrossRef]
  28. Taherisadr, M.; Asnani, P.; Salster, S.; Dehzangi, O. ECG-based driver inattention identification during naturalistic driving using mel-frequency spectrum 2-D transform and convolutional neural networks. Smart Health 2018, 9, 50–61. [Google Scholar] [CrossRef]
  29. Atzori, M.; Gijsberts, A.; Castellini, C.; Caputo, B.; Hager, A.G.M.; Elsig, S.; Giatsidis, G.; Hassetto, F.; Muller, H. Electromyography data for non-invasive naturally-controlled robotic hand prostheses. Sci. Data 2014, 1, 1–13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.