Next Article in Journal
Motor Synergies Measurement Reveals the Relevant Role of Variability in Reward-Based Learning
Next Article in Special Issue
Feature Selection and Validation of a Machine Learning-Based Lower Limb Risk Assessment Tool: A Feasibility Study
Previous Article in Journal
Conceptual Design and Image Motion Compensation Rate Analysis of Two-Axis Fast Steering Mirror for Dynamic Scan and Stare Imaging System
Previous Article in Special Issue
Influence of the Antenna Orientation on WiFi-Based Fall Detection Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Comparative Study of Time Frequency Representation Techniques for Freeze of Gait Detection and Prediction

by
Tahjid Ashfaque Mostafa
1,*,
Sara Soltaninejad
1,
Tara L. McIsaac
2,3 and
Irene Cheng
1,*
1
Multimedia Research Center, Department of Computing Science, University of Alberta, Edmonton, AB T6G 2E8, Canada
2
Arizona School of Health Sciences, A.T. Still University, 5850 E. Still Circle, Mesa, AZ 85206, USA
3
School of Pharmacy and Health Professions, Creighton University Health Sciences, 3100 N. Central Ave., Phoenix, AZ 85013, USA
*
Authors to whom correspondence should be addressed.
Sensors 2021, 21(19), 6446; https://doi.org/10.3390/s21196446
Submission received: 13 August 2021 / Revised: 13 September 2021 / Accepted: 24 September 2021 / Published: 27 September 2021
(This article belongs to the Special Issue Perception and Intelligence Driven Sensing to Monitor Personal Health)

Abstract

:
Freezing of Gait (FOG) is an impairment that affects the majority of patients in the advanced stages of Parkinson’s Disease (PD). FOG can lead to sudden falls and injuries, negatively impacting the quality of life for the patients and their families. Rhythmic Auditory Stimulation (RAS) can be used to help patients recover from FOG and resume normal gait. RAS might be ineffective due to the latency between the start of a FOG event, its detection and initialization of RAS. We propose a system capable of both FOG prediction and detection using signals from tri-axial accelerometer sensors that will be useful in initializing RAS with minimal latency. We compared the performance of several time frequency analysis techniques, including moving windows extracted from the signals, handcrafted features, Recurrence Plots (RP), Short Time Fourier Transform (STFT), Discreet Wavelet Transform (DWT) and Pseudo Wigner Ville Distribution (PWVD) with Deep Learning (DL) based Long Short Term Memory (LSTM) and Convolutional Neural Networks (CNN). We also propose three Ensemble Network Architectures that combine all the time frequency representations and DL architectures. Experimental results show that our ensemble architectures significantly improve the performance compared with existing techniques. We also present the results of applying our method trained on a publicly available dataset to data collected from patients using wearable sensors in collaboration with A.T. Still University.

1. Introduction

Parkinson’s Disease (PD) had affected about 6.2 million people globally in 2015 [1]. Since then the number is estimated to have risen to around 10 million [2], making it one of the most widely occurring neuro-degenerative movement disorders. PD is usually much more prevalent in aging people, adults 60 and over [3]. It is characterized by the loss of nerve cells or neurons in the substantia nigra area of the basal ganglia in the human brain. Loss of these cells results in reduction of the neurotransmitter dopamine and ultimately in decreased control of body movements [3,4,5,6].
PD is characterized by a number of neurological and motor symptoms like speech impediments, olfactory dysfunctions, autonomic dysfunctions, sleep disorders, fatigue, resting tremors, stiffness of trunk and limbs, slowness of movement (bradykinesia), reduced movement (akinesia), dyskinesia, irregular stride length and gait speed, Freezing of Gait (FOG), falls and postural disorders [7]. It is difficult to diagnose PD without the manifestation of motor symptoms, which are often unlikely to appear before 50% to 70% of the neurons have been damaged [8], making it difficult to administer any kind of preventive measures. The motor symptoms usually cause functional impairments in a subject, creating difficulties in sitting and standing up. The person with PD also suffers from losing the normal pendulum motion of the arms and displaying very small steps [4,5].
One of the most common symptoms of PD is FOG, with around 50% of all PD patients being affected [9,10]. Episodes of FOG cause the person with PD to intermittently experience a sudden inability to move, which often occurs while initiating gait, making turns while walking or when experiencing stress. The subjects report a feeling of their feet being glued to the ground during these events [11]. Based on the signals received from sensors worn around the ankles, it was found that while normal walking steps occur at a frequency of 0.5 Hz to 3 Hz, FOG exhibits a frequency of 6 Hz to 8 Hz [12,13].
Typically, FOG is very difficult to estimate and predict, but it can cause the risk of falls and pose health risk for the affected elderly [14,15,16,17]. An accurate prediction and detection of FOG can reduce accidents and thus improve the quality of life of the patients and their loved ones.
Current methods for detecting FOG [18,19,20] mostly consist of tests and detailed questionnaires used to assess the frequency and severity of FOG episodes. Although somewhat accurate, these methods suffer from shortcomings because of their clinical setup not reflecting real-world scenarios. FOG events usually tend to occur at home or while the patients are performing Activities of Daily Living (ADL) [7,11,21], which are different from the clinical test setup.
With the advancement in technology, using wearable sensors to monitor movements, body temperatures, heart rates and other physical parameters has become increasingly commonplace [22]. These sensors are lightweight, comfortable and usually do not hamper a person’s daily activities while monitoring ADL. The data recorded from wearable sensors for activity detection has brought promising performance in various applications, especially when combining with modern Machine Learning (ML) and Deep Learning (DL) based techniques [23,24,25,26,27]. There are wearable sensors that use auditory stimulation to treat FOG, which help shorten the duration of FOG events [28]. However, these sensors cannot effectively stop FOG episodes because of the latency of detection, which can still be hundreds of milliseconds in the best case scenarios [29].
There have been many applications using wearable technologies along with ML and DL based techniques to monitor motor functions of PD patients, aiming to achieve more effective treatment and reduce healthcare expenses [30,31,32]. These approaches can provide an unobtrusive and comfortable experience to the patient, while collecting personalized long term relevant medical history and improving the quality of treatment. Maetzler et al. [33] state that an automatic FOG analysis and detection system could play a vital role in monitoring the occurrence and evolution of FOG events over time. Although a permanent and guaranteed cure for PD or FOG itself has not been available at this time, a sufficiently accurate automatic monitoring system might prove to be helpful in minimizing the frequency and duration of FOG events. An established FOG treatment is to use Rhythmic Auditory Stimulation (RAS) [34], which produces a rhythmic ticking sound as auditory cues to help the patient resume normal gait when a FOG is detected. RAS has shown to improve walking by maintaining the speed and amplitude of movements [35,36,37].
With the popularity of the wearable sensors, the amount of available data collected from them is increasing at a rapid pace, which facilitates the use of DL based techniques. DL is under the scope of artificial intelligence that has the capabilities to automatically extract features from data without manual feature engineering. DL based end-to-end classifiers have shown promising performance, outperforming ML based classifiers in general, if sufficient amount of training data is available. Recently, DL based approaches have been adopted to perform tasks related to Human Activity Recognition (HAR) using data from various sensors [23,24,25,26].
Deep Convolutional Neural Networks (CNNs) are common Deep Learning architectures. Lecun et al. [38] mention in their book “The Handbook of Brain Theory and Neural Networks” that CNNs can be applied to temporal signals and images to automatically extract abstract distinct features by combining several convolutional operators. Although CNNs are proficient in extracting invariant local features from data, this architecture often falls short when the data has global time dependency, which is often the case with data obtained from wearable sensors. Recurrent Neural Networks (RNNs) are able to solve this issue because the connections between the nodes of this architecture exhibit a discrete-time dynamical system [39,40]. Long Short Term Memory (LSTM) is one of the most widely used RNNs, able to model time dependency in sequential time series data using various logic gates to control a memory space [41].
Neural Network Ensembling is the learning paradigm of training a collection of neural networks to collaborate on the same task [42]. The idea of ensembling was introduced by Hansen et al. [43], who proposed that the generalization ability of a Neural Network based system can be significantly improved by training a number of neural networks and by combining their solutions to solve the same problem. A typical ensemble architecture consists of two steps, i.e., training multiple components or constituent neural networks and then creating an architecture that combines their outputs. In recent years, Ensemble Learning techniques have been applied to PD detection tasks and they have achieved significant success [44,45].
The purpose of this work is to combine some most widely used time frequency analysis techniques, with CNN and LSTM based architectures for the detection and prediction of FOG events, using data captured from a tri-axial accelerometer sensor. In order to solve the issue of detection latency, we predict the changes in gait immediately before a FOG event. If the onset of FOG events can be accurately predicted, RAS can be applied even before it starts. We use a BiDirectional LSTM architecture with raw signals and handcrafted features and it was observed that handcrafted features did not improve the performance. We explored a CNN architecture with multiple visual representation methods including Recurrence Plot (RP), Short Time Fourier Transform (STFT), Discrete Wavelet Transform (DWT) and Pseudo Wigner Ville Distribution (PWVD). Three ensemble neural network architectures were also proposed and it was observed that they significantly improve the prediction and detection performance compared to individual models. We evaluated our models with multiple metrics to ensure that our findings are unbiased.
Our research group, in collaboration with Arizona School of Health Sciences, A.T. Still University, acquired gait data from 14 PD patients recorded using APDM wearable sensors. We applied our trained neural networks and proposed a system capable of monitoring FOG from these data after we verified the performance of our models on a publicly available dataset.

2. Literature Review

Smart sensors have been commonly used as a tool for assessing motor symptoms such as FOG in PD and other movement disorders. This is possible because of the improvements in computational power of small devices [46]. Existing FOG assessment methods using these sensors can be categorized into different groups depending on the sensor types, sensor locations, extracted features, and the analytics methods. FOG detection can be conducted real-time or offline [47]. However, FOG detection and prediction are challenging tasks because of the variability of event duration and frequency. We observe that previous studies mainly captured FOG episodes that are not consistent to the patients’ normal daily activities because their data were simulated in laboratory settings. In this section, we review related work on FOG detection.
An early FOG detection method was proposed by Han et al. [48] using U-AMS (Activity Monitoring System). Wavelet power features were used for discrimination of abnormal movements in PD patients. Moore et al. [49] then proposed a threshold based method for FOG detection by defining the Freeze Index (FI), which is the ratio between the power of the signal in “freeze” band (3–8 Hz) divided by the power of the signal in the “locomotion” band ( 0.5 –3 Hz). The proposed method marks FOG episodes when FI exceeds a certain threshold. The subject dependent experimental results show 78 % correct detection of FOG (true positive rate) and 20 % false positive rate. Bachlin et al. [28] presented a real time FOG detection method by introducing a new term to Moore et al. [49] method, called Power Index (PI), which is the addition of walking band (WB) and Freezing Band (FB) that indicates the amount of movement. In [28], FOG episodes are determined using two thresholds (Freezing Threshold (FTH) and Power Threshold (PTH)) given F I > F T H and P I > P T H . In this method, once the FOG episodes are detected, the patient will get the auditory signals until his normal walking ability is resumed. They reported 73.1 % and 81.6 % for sensitivity and specificity respectively. The author also created the Daphnet dataset [28] for FOG assessment methods evaluation.
The first proposed FOG detection method based on ML was by Mazilu et al. [50]. The features for this classification were from the work of Bachlin et al. [28] with some additional features including mean, standard deviation, entropy, energy, FI and power of the acceleration signals. Random Forest (RF), Naive Bayes and K-nearest neighbour (KNN) were the ML algorithms used for classification. Motion data capture was done by a smartphone and a wrist acceleration sensor. The best obtained results were 66.25 and 95.83 for sensitivity and specificity respectively with RF using 10-fold cross-validation. In the following year, they presented another automatic FOG detection system using wearable sensor. In this work, they did multi-class analysis as the PreFOG motion was considered a new class (FOG vs. PreFOG vs. normal locomotion). Learning was conducted by studying the time domain and statistical features from the motion data. In this new work, they could improve F1 score by 8.1 % . The new automatic FOG detection method introduced auditory cueing to warn the patient about FOG episodes. In the same year (2013), a system for automatic FOG detection was proposed by Tripoliti et al. [51]. The system was based on four steps: data imputation (interpolation), band-pass filtering, entropy calculation and automatic classification (Naïve Bayes, RF, Decision Trees and Random Tree). Data was obtained from 5 healthy subjects, 5 PD patients with FOG symptoms and 5 PD patients without FOG symptoms. The results show 81.94 % sensitivity, 98.74 % specificity, 96.11 % accuracy and 98.6 % Area Under Curve (AUC) using RF. Another proposed FOG detection work in 2013 was by Moore et al. [52], which assesses seven sensors placed in different locations for gait analysis. Their analysis found that the shank and back were the most convenient places for the sensors. However, they found that using all the seven sensors could get higher and more robust performance with sensitivity 84.3 % and specificity 78.4 % .
In 2015, Zack et al. [53] presented a threshold based FOG detection technique following the approach of Moore [52] using a single tri-axial accelerometer placed at the waist. Receiver operating characteristic (ROC) curves were drawn to determine a global FI threshold to distinguish between FOG and non-FOG episodes for different tasks. In addition to the global FI threshold, they calculated the sensitivity and specificity of the FI threshold for each subject. Combining all task results, a sensitivity of 75 % and specificity of 76 % were achieved [47].
Rodríguez et al. [54] presented a novel approach for FOG detection using machine learning techniques and daily activities of the PD patients in real environments. They extracted 55 FOG related features from 21 PD patients using just a single waist-worn tri-axial accelerometer. Support Vector Machine (SVM) with leave-one-out cross-validation was used for classification in two scenarios: user independent and user dependent. Experimental results show a sensitivity of 88.09 % and specificity of 80.09 % with R-10-fold cross-validation and a sensitivity of 79.03 % and specificity of 74.67 % for leave-one-subject-out (LOSO) evaluation. After that, Sama et al. [55] decreased the number of features to 28 for the same dataset. The extracted features were sent to 8 different classifiers with greedy subset selection process, 10-fold cross-validation and different window sizes. The results of FOG detection at patients’ homes were 91.7 % and 87.4 % for sensitivity and specificity respectively, which are better than the results of Rodrigues’s method.
Orphanidou et al. [56] evaluated machine learning algorithms to identify the FOG prior to its onset. An accelerometer time series dataset containing 237 individual Freezing of Gait events from 8 patients was considered, from which features were extracted and presented to 7 machine learning classifiers. SVM achieved the highest performance in comparison with the benchmark techniques. The classification algorithm was applied to 5 s windows using 18 features, obtaining balanced accuracies (the mean value of sensitivity and specificity) of 91 % , 90 % and 82 % over the Walk, FOG and Transition classes, respectively. However, the need for systematic analysis of the problem was identified. Therefore, in their next study [56], they specifically focused on the early detection of a FOG event, through classification of the transition class using varying size time windows and time/frequency contrary to the majority of previous studies that recognized FOG only when it had occurred. In their paper, the Daphnet dataset was used with accelerometer signals obtained from sensors mounted on the ankle, thigh and trunk of the PD patients. Data augmentation was performed on the dataset to include another class label called ‘transition’ that showed the episodes before FOG occurrence. Daphnet features were sent out to a group of 5 classifiers, including Gradient Boosting (GB), Extreme Gradient Boosting, SVM, RF and Neural Networks. Experimental results show that SVM with Radial Basis (RBF) kernels has the best performance with sensitivity of 72.34 % , 91.49 % , 75.00 % and specificity values of 87.36 % , 88.51 % and 93.62 % , for FOG, transition and normal activity classes, respectively.
Deep Learning (DL) techniques have also been used for automatic FOG determination. DL can handle multi-modal data, missing information and high dimensional feature spaces. The first proposed FOG detection method using DL was by Camps et al. [57]. The proposed 1D Convolutions Neural Network (CNN) has 8 layers, which is trained using a novel spectral data representation strategy that considers information from both the previous and current signal windows. The data was collected from 21 subjects, consisting 9-channel signals recorded from a waist-worn Inertial Measurement Unit (IMU) with three tri-axial sensors: accelerometer, gyroscope and magnetometer. The experimental results show a performance of 90.6 % for the Geometric Mean (GM), an AUC of 0.88 , a sensitivity of 91.9 % and a sensibility of 89.5 % .
In 2019, San-Segundo et al. [58] presented a study to evaluate the robustness of different feature sets and ML algorithms for FOG detection using body-worn accelerometers. They used four feature sets: (Mazilu et al. [50] features, Human Activity Recognition (HAR) features, Mel Frequency Cepstral Coefficients (MFCCs) features, and Speech Quality Assessment (SQA) features). They also used four classifications (RF, multi-layer perceptron, Hidden Markov models (HMM) and deep neural networks). Evaluation was performed using a LOSO cross-validation. The best results were obtained when using the current window and three previous windows, with the feature set composed of Mazilu features [50] and MFCCs [59]. They found that the best classifier was a deep convolutions neural network achieving an AUC of 0.93 and an Equal Error Rate (EER) of 12.5 % .
In 2020, Sigcha et al. [60] evaluated some ML and DL classification and detection techniques with accelerometer signals acquired from a body worn IMU to enhance the FOG detection performance in real-world home environments. Three data representations proposed in the literature were reproduced (including Mazilu features [50], Mel Frequency Cepstral Coefficients (MFCCs) [59] and Fast Fourier Transform (FFT)) to establish a baseline using RF classifier with 10-fold cross-validation (R10fold) and LOSO. This analysis was also conducted to find the best data representation to test DL approaches including: a denoiser autoencoder, a deep neural network with CNN and a combination of CNN and LSTM layers. For comparison purposes, shallow algorithms such as one-class SVM (OC-SVM), SVM, AdaBoost and RF were tested. This study was evaluated on the data collected by Rodríguez-Martín et al. [54], which includes recordings from 21 PD patients, who manifested FOG episodes when performing ADL at their homes. The best performance for AUC was 0.93 . Their results illustrate that modeling spectral information of adjacent windows through an LSTM model can improve the performance of FOG detection without increasing the length of the analysis window.
In order to give a clear comparison of the different methods, we summarize their characteristics in Table 1. The novelty of the proposed work lies in the usage of multiple time frequency techniques as well as the usage of ensembling methodologies. Although there have been many experiments with ML and DL based technologies and feature extraction methods for FOG detection, to the best of our knowledge, our work is the first to apply time frequency representation techniques like Recurrence Plots (RP), Short Time Fourier Transform (STFT), Discreet Wavelet Transform (DWT) and Pseudo Wigner Ville Distribution (PWVD) alongside original signal and extracted features for FOG detection and prediction. We combine these techniques to design ensemble architectures. We designed a Bidirectional LSTM based architecture with four layers and trained separate instances of the architecture on Raw signal windows as well as 27 features extracted from the signal window.
We also designed a CNN architecture with four Convolutional layers, separate instances of which were trained on our visual features (RP, STFT, DWT, PWVD). Combining these models trained six different data modalities, we proposed our three ensemble architectures and evaluated their performance. The proposed models were able to both detect and predict FOG events while achieving the high scores in multiple evaluation criteria. The scores of the proposed ensemble methods either outperform or are very close to the scores of existing methods. Note that when solving a multiclass classification problem, i.e., in both identifying and predicting FOG events, our architectures achieve superior performance. We also tested the trained models on monitoring progression with real-world data, i.e., predicting FOG from accelerometer data captured using APDM wearable sensors. The result demonstrated that our approach is able to detect preFOG episodes in real world scenarios as presented in Section 5.
We believed there was room for improving the performance of existing models. None of the existing works analyzed by us had used visual representation of time series data for this task. Combining various visual representation techniques with handcrafted feature engineering and the capacity of DL based models to automatically extract relevant features from raw data was a promising avenue of research. Furthermore, although some authors including Orphanidou et al. [56] had introduced the concept of a PreFOG/Transition state before entering FOG state, their average performance scores for all classes were comparatively low and in their experiment SVM outperformed Neural Network based architectures. In other cases where DL based architectures achieved superior performance, they were mostly detecting FOG itself, not the PreFOG state. The goal of this experiment was to develop a unified solution that would be able to both predict FOG by detecting the PreFOG state and detect FOG itself with high accuracy. Additionally, we also wanted to compare the performance of various different time frequency representation techniques for this task.

3. Materials and Methods

In this work, we developed DL based techniques and used time frequency representation data as a feature set to classify as well as predict the FOG events. Experimental results show that our approach gives higher accuracy compared with existing state-of-the-art models based on tri-axial accelerometer sensor signals. The performance of each DL model was evaluated with different feature sets and multiple metrics in order to determine the optimal combination of models without bias. Finally, we are able to propose three ensemble architectures, each of which is composed of a selected set of models and features. The ensemble architectures significantly improve the performance of individual models.

3.1. Data

The publicly available DAPHNet [28] dataset was used for our experiments. The dataset contained data collected from ten PD patients, with seven male and three female experiencing regular FOG in their day to day activities. The average age of the participants was 66.4 ± 4.8 years, with an average disease duration of 13.7 ± 9.67 years. The average Hoehn and Yahr score was 2.6 ± 0.65 , indicating that the subjects had mild symptoms with mild balance impairment to moderate balance impairment [61]. Two tri-axial (3D) accelerometer sensors were attached to one of the patient’s legs: One was located at the shank just above the ankle and another was attached to the thigh slightly above the knee. A third sensor was placed at the lower back of the patient. The locations of the sensors are given in [28] and illustrated in Figure 1.
The dataset contained 237 FOG events. Synchronized video recordings were used by physiotherapists to identify the FOG events. The signal point, where the left-right steps stop alternating, is defined as the start of a FOG event. The point, where normal pattern resumed is defined as the end of the FOG event. Eight out of the ten subjects experienced FOG during the study, with the duration of FOG events ranging from 0.5 s to 40.5 s. The mean duration was 7.3 ± 6.7 s, with 50% of the FOG episodes being shorter than 5.4 s and 93.2% being shorter than 20 s. The signals were annotated in three categories:
  • 0—Not part of the experiment; user performed activities are unrelated to the experimental protocol while the sensors were installed.
  • 1—Experiment; no FOG.
  • 2—FOG.

3.2. Proposed Data Preparation Method

An overview of the preprocessing, data augmentation and feature extraction algorithms used in this study is presented in Figure 2. The major components are explained in the subsections below.

3.3. Preprocessing

In the preprocessing component, data that is irrelevant to the study is removed and signals from the three axes of each sensor is combined so that there is only one signal stream for each sensor.

3.3.1. Removing Unrelated Data

As previously described in Section 3.1, data with an annotation of 0 was not a part of the experiment; the subjects performed activities which were not in accordance with the protocol of the data collection process. In this step, the unrelated data was excluded from the dataset.
The Daphnet [28] dataset was divided per subject and we excluded subjects who did not experience FOG at all during the data collection period.

3.3.2. Calculating Magnitude of Acceleration for all three axis

Each of the accelerometers was assigned to a single channel, with the data being recorded for three channels: Ankle (A), Leg (L) and Torso (T) representing the locations of the sensor placement. Each channel contained three separate signals, with each of the signals corresponding to a single axis of the accelerometer. The axes were horizontal forward (X), vertical (Y) and horizontal lateral (Z). Thus, a set τ of nine signals was recorded for each of the patients with a sampling frequency ( f s ) of 64, as illustrated in Equation (1).
τ = { { A X , A Y , A Z } , { L X , L Y , L Z } , { T X , T Y , T Z } }
Magnitude of acceleration is the relative value of the overall acceleration at any given time instance, calculated as shown in Equation (2), combining the signals for each of three axes ( x , y , z ) into a signal.
α C = α X 2 + α Y 2 + α Z 2 , w h e r e α X , α Y , α Z τ , α C τ C
The three signals ( x , y , z ) originating from each channel ( A , L , T ) were combined to calculate the magnitude of acceleration denoted by τ C in Equation (2), resulting in three signal streams with one from each of the sensors as shown in Equation (3).
τ C = { A C , L C , T C }

3.4. Data Augmentation

We applied small non-overlapping windows to extract data from the original continuous signal. The window data immediately before the start of a FOG event was labelled with a new class PreFOG, which is essential for predicting FOG events before they occur. The number of Non-FOG samples vastly outnumber the PreFOG and FOG samples, making Non-FOG our majority class. In order to solve the issue of class imbalance, the minority classes, PreFOG and FOG, were over sampled to match the number of samples from the Non-FOG class.

3.4.1. Signal Segmentation

Non-overlapping 1-dimensional windows of length f s × w time-steps were used to extract signal α C τ C , where w is the length of the signal window in seconds.
α C = [ α 1 α 2 α w × f s ] w × f s
Since the windows were non-overlapping, shorter window lengths provided a larger dataset. Signals were segmented into window lengths ranging from 1 to 4 s.

3.4.2. Labeling PreFOG class

Mazilu et al. [62] proposed that gait cannot enter into FOG state directly from normal walking without first going through a state of deterioration. They define this state as PreFOG, which is a transition period with variable duration. Identifying this transition state would be valuable for both FOG detection and prediction. Since the duration of PreFOG might not be the same from patient to patient, for our experiment the immediate window ( w × f s time steps) before the onset of a FOG event was labeled as PreFOG. The final dataset thus had three annotations,
  • 0—Non FOG
  • 1—FOG
  • 2—Pre FOG
Figure 3 presents an illustration of combined accelerometer signal from Ankle that contains normal gait, PreFOG and FOG. PreFOG is highlighted in yellow and the FOG episode is highlighted in red.

3.4.3. SMOTE Oversampling

At this stage, the dataset was hugely imbalanced, with the majority of the data being from the Non-FOG class. Such imbalanced data would lead to most architectures ignoring the minority classes and over-classifying the majority class, although the performance on the minority classes is much more significant in this case. There are multiple ways to address this issue. One approach is to under-sample the majority class to match the number of samples in the minority classes. However, in our case, the minority samples are sparse, and under-sampling the majority class would lead to a drastic decrease in the total number of training samples. Neural Network architectures require a large number of training samples in order to perform satisfactorily, and therefore under-sampling would lead to poor performance. An alternative method is to over-sample the minority class. It involves duplicating the samples of the minority class to match the number of samples in the majority class. Although this method balances the class distribution, it does not provide the networks with any new information to learn. We decided to choose the approach proposed by Chawla et al. [63] to synthesize new samples from existing samples. This Synthetic Minority Over-sampling Technique (SMOTE) creates new synthetic plausible samples that are in the same feature space as other minority class samples.

3.5. Feature Extraction

After data augmentation, our final feature set consisted of 5 different modalities extracted from the same source signal in addition to original signal itself, α i α C as shown in Equation (5).
F e a t u r e s i = { α i , F i , R P i , S T F T i , D W T i , P W V D i } , α i α C
where:
  • α i = Moving window extracted from signal α C .
  • F i = Manually extracted feature set from α i .
  • R P i = Recurrence Plot representation of α i .
  • S T F T i = Short Time Fourier Transform representation of α i .
  • D W T i = Discrete Wavelet Transform representation of α i .
  • P W V D i = Pseudo Wigner Ville Distribution representation of α i .

3.5.1. Time and Frequency Domain Features

For each moving window α i α C , 27 features relating to the time and frequency domain were extracted, as explained in Table 2. Some of the features were selected based on previous works done by Mazilu et al. [62] who added additional features to the work of Bachlin et al. [28] including Min, Max, Range, Median, Mode, Trimmed Mean, Standard Deviation, Variance, Root Mean Square, Skewness, Kurtisis, Normalized Signal Magnitude Area, Mean Crossing Rate and Signal Vector Magnitude. Additional time domain features such as Mean Absolute Value, Median Absolute Deviation, 25th and 75th percentile values, Interquantile Range and Peak of Fourier Transform were added. Furthermore, sensor and frequency based features such as Energy, Entropy, Peak Frequency, Freeze and Locomotion Band Power, Freeze Index and Band Power were added in accordance to the work done by Gokul et al. [64].

3.5.2. Recurrence Plots

Recurrence Plots (RP) are used to represent temporal correlations of univariate series data defined in a square matrix [65]. For time series data, the matrix elements represent the times at which the amplitude of the signal recurs. If i and j are two time instances, and x ( i ) and x ( j ) are values in the time series at two recurrence time instances, the formula to compute the recurrence plot [66] is given in Equation (6)
R ( i , j ) = 1 , i f | | x ( i ) x ( j ) ϵ | | 0 , otherwise , ( ϵ i s a c u s t o m s i m i l a r i t y t h r e s h o l d )
Recurrence plots are often robust against outliers and noisy data for periodic signals. Some examples of recurrence plots for our signals can be seen in Figure 4. The plots were generated with a window length (w) of 2. It was observed that for w = 2 , Non-FOG events had no distinct pattern when represented as a recurrence plot, PreFOG events show clear distinct patterns and FOG events had patterns that were more defined than Non-FOG but less defined than PreFOG. Both x and y axes represent time for RP.

3.5.3. Short Time Fourier Transform

A Short Time Fourier Transform (STFT) is a Fourier transform that quantifies the phase content and the sinusoidal frequency of signal segments changing over time [65]. STFT is useful in capturing the time and frequency characteristics in the signals. Rajoub et al. [67] mentioned that STFT does not perform well in capturing sharp signals and patterns with varying duration. Figure 5 shows some example spectograms generated using STFT, describing magnitude over time for each of our signal types over a 2 s time window. x axis represents time and y axis represents frequency for STFT.
For w = 2 , STFT captured the difference between Non-FOG and other classes, with the spectograms for PreFOG and FOG classes being almost clear compared to that of Non-FOG. However, it was difficult to visually differentiate PreFOG and FOG from STFT alone.

3.5.4. Discrete Wavelet Transform

Discrete Wavelet Transform (DWT) is a process of decomposing a signal sequence into subsets, with each subset being a time series consisting coefficients that represent the time evolution in the corresponding frequency band [68]. A main advantage of DWT is the ability to capture both frequency and location characteristics in a time series. Haar Transform is the simplest of wavelet transforms. We used Haar sequence proposed by Haar et al. [69], which is the first known wavelet basis. The Haar wavelet can be used to analyze signals with sudden transitions, because of its non-differentiable property. Figure 6 shows sample plots of the approximation and detail coefficients of transforms for a 2 s time window, with 2 subsets. For DWT, x axis represents time and y axis frequency.
DWT representation plot for w = 2 is useful for visually identifying the Non-FOG class compared to PreFOG and FOG classes. For Non-FOG events, the approximation and detail coefficient plots are almost flat, without any large fluctuation in value, which is distinctly identifiable. The representations for PreFOG and FOG events are harder to differentiate as both representations show sudden rise and drop in their values.

3.5.5. Pseudo Wigner Ville Distribution

Pseudo Wigner Ville Distribution (PWVD) is a method to represent transient phenomena in three dimensions, i.e., time, frequency and amplitude [70]. PWVD has been proven to be effective in generating accurate time frequency representation, since its frequency and time resolutions are determined by the resolution of the signals and not by the duration [70]. Figure 7 shows some examples of PWVD computed on signals with 2 s time window from our data. For PWVD, x and y axes represent time and frequency respectively.
The Non-FOG and FOG gaits can be clearly distinguished from PWVD representations for w = 2 , as Non-FOG gaits have a clear central section compared to FOG events. Both PreFOG and FOG classes have patterns appearing in the central section, which makes it difficult to differentiate them visually.
Finally, we can conclude that since RP shows a clear distinction between Normal and PreFOG states, and STFT, DWT and PWVD show a clear distinction between Normal and FOG states, if we combine RP with any one of STFT, DWT or PWVD, it can be possible to visually identify Normal, FOG or PreFOG states. This can be utilized to identify the resumption of Normal Gait after an FOG event.

3.6. Model Structure

Based on the findings obtained from the four feature visual representations, R P , S T F T , D W T , P W V D discussed above, we introduced a Convolutional Neural Network (CNN) based model architecture. A Long Short Term Memory (LSTM) based architecture was proposed for the original signal α and the corresponding feature set F. For each data modality F e a t u r e s , an instance of the corresponding model was trained and its performance was recorded. Then, the trained model instances were combined in three ensemble network architectures, M 7 , M 8 and M 9 , as explained below. Our objective is to demonstrate that properly designed ensemble models can provide better performance than individual constituent models.

3.6.1. Basic Convolutional Neural Network

Convolutional Neural Networks (CNNs) are known for their ability to identify complex non-linear relationships between data points without hand crafted feature engineering. To complement our techniques to present time series data visually, a CNN architecture was designed, which is presented in Figure 8. The input is passed through four 2D C o n v o l u t i o n a l layers with filter sizes 64, 32, 16 and 8 respectively, a kernel size of ( 4 , 4 ) and L e a k y R e L u activation function with a negative slope coefficient and a l p h a value of 0.3 . Each of the C o n v o l u t i o n a l layers was followed by a 2D M a x P o o l i n g layer with a pool size of ( 2 , 2 ) and a D r o p o u t layer having a dropout rate of 0.25 . The data was then flattened and passed through two D e n s e layers with 100 and 50 units respectively. Each of the D e n s e layers had L e a k y R e L u activation function with alpha value of 0.3 and was followed by 2 D r o p o u t layers having a dropout rate of 0.2 . Finally, a D e n s e layer with S o f t m a x activation function of 3 units for our three output classes was added. The model was compiled with a R M S P r o p optimizer with an initial Learning rate of 0.0001 . For our four visual feature types, RP, STFT, DWT and WV, a separate instance model was trained and validated, which are labeled M 1 , M 2 , M 3 and M 4 respectively.

3.6.2. Basic Bidirectional LSTM

Long Short Term Memory (LSTM) network is a type of recurrent neural network architecture, which is suitable for learning and remembering a long sequence of input data, automatically extracting features from the raw sequence and providing comparable performance to using handcrafted features. Bidirectional LSTMs add a duplication of the first recurrent layer. The first layer is trained on the original input sequence and the duplicated layer is trained on a reversed copy of the input sequence. For our data, the use of Bidirectional LSTM is justified because the context of the whole signal sequence, instead of a linear interpretation, is relevant for FOG identification and prediction. Our Bidirectional LSTM architecture is illustrated in Figure 9. The input is passed through four Bidirectional LSTM layers stacked on top of each other with t a n h activation function and n l a y e r s hidden layers. The value of n l a y e r s is computed by Equation (7) where l i n p u t is the length of the input sequence and σ is the multiplication coefficient. The value for σ was set to 3 based on trial and error as it generated the best result. The output of L S T M was passed through a D e n s e layer with S o f t m a x activation function. The final D e n s e layer had 3 units to classify between the three output classes. An A d a m optimizer with an initial learning rate of 0.0001 was used to compile the model. One instance of this model, M 5 , was trained on the original signal α C and another instance, M 6 , was trained on the handcrafted feature set F C corresponding to the signal α C .
n l a y e r s = l i n p u t × σ

3.6.3. Ensemble Architectures

Ensemble Learning is a neural network training approach, where the predictions from multiple trained networks are combined to solve a problem [71]. In this work, three ensemble network architectures are examined (Figure 10). The constituent model set is defined as,
M c o n s t i t u e n t = { M 1 , M 2 , M 3 , M 4 , M 5 , M 6 }

Stacked Ensemble Model—M7

This model architecture is designed by combining the output predictions of all M i M c o n s t i t u e n t . The models have already been trained on their respective data, and all layers of constituent models are set as non-trainable before adding them to the ensemble model. The outputs of the models are passed through a C o n c a t e n a t i o n layer and then two D e n s e layers with 10 and 3 units respectively. The first D e n s e layer has a R e L u activation function and the final D e n s e layer has a S o f t m a x activation function. An A d a m optimizer is used with a learning rate of 0.0001 .

Average Ensemble Model—M8

This model architecture takes the average of the predicted outputs of all M i M c o n s t i t u e n t . The constituent models pre-trained on their respective data are set as non trainable, and the outputs are passed through an A v e r a g e layer. M 8 is compiled with an A d a m optimizer having a learning rate of 0.0001 .

Majority Voting—M9

For majority voting, the output is based on the majority vote of the constituent models M i M c o n s t i t u e n t . The hard voting approach is used to calculate the final outcome, where every constituent model votes for an output class and the majority vote is selected as the final prediction. In statistical terms, this is equivalent to calculating the Mode of the predictions from all constituent models.

4. Results

The focus of this work is to analyze how well our proposed solution can detect and predict the characteristics embedded in the given data. We use collected data, which already has ground truth defined, i.e., True Positive (TP), and therefore do not need to obtain user opinions to validate the outcome. Instead, we compare the model generated results with the ground truth. We follow the assessment metrics based on TP, False Positive (FP), True Negative (TN) and False Negative (FN), used in related work in order to make fair comparisons.

4.1. Evaluation Criteria

In cases, where a majority class dominates the dataset, it might be possible that the detection accuracy is very high despite the model failing to identify the minority classes. To ensure that the performance of our model is properly evaluated and it is not over classifying the majority class, a number of evaluation metrics were utilized in our study.

4.1.1. Detection Accuracy

Detection accuracy is the most widely used evaluation metric. It is defined as the fraction of predictions by a model that are accurate. Detection accuracy can be computed as in Equation (9). The output of this metric ranges between ( 0 , 1 ) , with 0 being completely inaccurate and 1 representing perfect prediction.
A c c u r a c y = N u m b e r o f c o r r e c t p r e d i c t i o n s T o t a l n u m b e r o f r e c o r d s

4.1.2. Precision, Recall/Sensitivity, Specificity, F β Score

Precision, Recall/Sensitivity, Specificity and F β Score are very important in understanding the performance of a model. Since we are evaluating a multi-class classification model, each of these metrics computes an individual class and then their weighted average is calculated.
Precision for a class is the measure of the classifier’s ability to not classify a negative sample as positive, as defined in Equation (10).
P r e c i s i o n ( A k , B k ) = | A k B k | | A k |
Recall/Sensitivity of a class measures how well the classifier can identify positive samples of a class, as defined in Equation (11).
R e c a l l ( A k , B k ) = | A k B k | | B k |
Specificity for a class is defined as the ability of a classifier to reject samples that are not a member of that class.
S p e c i f i c i t y ( C k , D k ) = | C k D k | | D k |
The F β score is calculated as the weighted harmonic mean of Precision and Recall, ranging between (0,1), with 1 being the best possible value, as presented in Equation (13).
F β ( A k , B k ) = ( 1 + β 2 ) P r e c i s i o n ( A k , B k ) × R e c a l l ( A k , B k ) β 2 P r e c i s i o n ( A k , B k ) + R e c a l l ( A k , B k )
where,
  • A k is the predictions for class k.
  • B k is the occurrences for class k.
  • C k is the predictions for samples not in class k.
  • D k is the occurrences for samples not in class k.
  • k represents a class in range 1 : K , in our case K = 3 .

4.1.3. Matthews Correlation Coefficient

Matthews Correlation Coefficient (MCC), also known as P h i Coefficient, was proposed by Matthews et al. [72] in 1975. MCC offers a balanced measure of quality for both binary and multi-class classifications, which can be used even if the classes are imbalanced. The value of this metric ranges from ( 1 , + 1 ) . A MCC value of + 1 indicates perfect prediction, 0 indicates random prediction and 1 indicates inverse predictions. Gorodkin et al. [73] generalized MCC for multiple classes as the R K statistic, defined with respect to confusion matrix C for K classes following Equation (14) presented below [74].
M C C = c × s k K p k × t k ( s 2 k K p k 2 ) × ( s 2 k K t k 2 )
where,
  • t k = i K C i k , the number of occurrences of class k.
  • p k = i K C K i , the number of predictions for class k.
  • c = k K c k k , total correct predictions.
  • s = i K j K C i j , total number of samples.

4.2. K-Fold Cross Validation

In order to get an accurate estimate of the model performances, Stratified K-fold cross validation technique was utilized to separate the data into training and testing sets, with 80 % of the data being used for training and the rest for testing, preserving the ratio of samples of each class. Since neural network models take a long time to train and evaluate, it is difficult to use high values for K. For this experiment, K was set to 5. The dataset was first shuffled, and then it was split into K unique (train, test) combinations. For each fold, a new instance of each of our models was trained using the training set and its performance on the testing set was evaluated and recorded. The evaluation performances were retained while the instances of the models were discarded. Finally, the average performance of the models across all K folds was recorded.

4.3. Normalization

The visual feature representations in R P i , S T F T i , D W T i , P W V D i are normalized using Equation (15). Since the range for RGB values in images is (0–255), each channel is normalized to the range of 0–1. Then the values are centered through division with the mean.
p n o r m a l i z e d = p o 255.0 m e a n ( p o ) , p o R P i , S T F T i , D W T i , P W V D i

4.4. Experimental Setup

For each of the training sets, it was further divided into (train, validation) sets with 80% being used for training and the rest for validation. Validation using unseen data was crucial to evaluate whether the model was learning over time by comparing its performances. Each of the models was trained for 500 epochs. The training was stopped if the validation accuracy did not improve over 50 epochs. All the experiments were done on a Machine with Ubuntu OS, 4 core Intel Xeon Processor, 62 Gigabytes of RAM and Nvidia Tesla GPU with 16 Gigabytes memory. The algorithm was tested with data from all three wearable sensors α C τ C , but we achieved the best performance with data from the ankle mounted sensor A C . All results presented here are based on A C . In order to train Convolutional Neural Networks, all image features were adjusted to the shape of (3, 128, 128) and then normalized. The runtimes presented include both training and testing of the model but do not include the preprocessing and feature extraction time. Runtimes presented here for ensemble models do not include the time for training the constituent models. All scores presented in Section 4.5 are average scores. Five instances of each model were trained and evaluated on five Folds of (train, test) sets, and their scores and runtimes were averaged. Table 3 shows the reported performance of some existing models on the same dataset, in a (0–100) range.

4.5. Metric Scores and Discussion

Since we used non-overlapping time windows, smaller window sizes yielded significantly larger amount of data, which led to better performance in neural network based architectures. We experimented with widow sizes of 2, 3, 4 s. A window size of w seconds means that our model is able to predict the start of a FOG event w seconds before it happens. We believe that a window size of 1 would lead to much better detection performance since it means more training examples for the model. However, we did not use a smaller window size of 1 because it would also decrease the time window by which we can predict the FOG event, leading to a higher resource consumption during training.
We also observed that the size to which the features in R P i , S T F T i , D W T i , P W V D i are reshaped also plays a vital role in model performance, with larger sizes producing better results. Due to resource constraints, we set this size to be ( 3 , 128 , 128 ) .
Section 4.5 presents the performance of for each model M M 1 , M 9 with signal S τ C (for each of A n k l e ( A C ) , L e g ( L C ) and T r u n k ( T C ) ) with Window Size w 2 , 3 , 4 (seconds). The tables report multiple evaluation metric scores including Accuracy, Precision, Recall/Sensitivity, Specificity, F β score, MCC score and the Runtime taken for the model to train in minutes. All scores are reported in the range of (0, 1), except MCC score, which is in the range of (−1, 1). The scores are reported in M e a n ± S t a n d a r d D e v i a t i o n format. The best scores for each model using the same modality of data but with different window sizes were reported in bold font.
From the metric scores presented in Table 4, it can be seen that Basic CNN M1 trained on RP generated from signals performs reasonably well across all metric scores. In most cases the smallest window size of 2 s yielded the best scores, but there was not a drastic decrease in performance when we increased the window sizes. Comparing sensor locations, data collected from Trunk sensor ( T C ) performed the best, followed closely by data collected from Ankle ( A C ) and Leg ( L C ).
Table 5 presents the scores for Basic CNN M2 trained on STFT plots generated from the signals. For STFT. window size of 3 seemed to provide comparatively better results, although the scores were poor when compared to the scores from RP. The data collected from Trunk sensor ( T C ) provided best results when using STFT, followed closely by data collected from Leg ( L C ) and Ankle ( A C ).
The metric scores of Basic CNN M3 using DWT are reported in Table 6. M3 achieved the highest accuracy among our models using visual features (RP, STFT, DWT, PWVD). The scores for varying window sizes were very similar, with a window size of 2 s providing the best scores for Ankle ( A C ) and Trunk ( T C ) sensor data. For data collected from the Leg ( L C ), a window size of 3 generated the best scores. Comparing the scores of the three sensors’ locations, it was noted that Trunk ( T C ) provided the best scores, followed very closely by Leg ( L C ) and Ankle ( A C ).
Table 7 notes the metric scores of Basic CNN M4 using PWVD. A window size of 3 s provided the best scores for Ankle ( A C ) and Leg ( L C ) sensor data. For data collected from the Trunk ( T C ), a window size of 4 generated the best scores. Comparing the best scores for each sensor location, the scores for all three locations were pretty similar.
Table 8 contains the scores of Bidirectional LSTM with extracted raw signal windows. The overall performance is not as good as using visual features. The performance does not experience a drastic change when the window size increases from 2 to 3 s, but we see a significant drop in performance as the window size changes from 3 to 4 s. A window size of 3 s provided comparatively better scores for Trunk ( T C ) and Leg ( L C ) sensor data. For data collected from the Ankle ( A C ), a window size of 2 generated the best scores. Data from Leg ( L C ) sensor provided the best overall scores when using bidirectional LSTM and raw signals.
The scores of Bidirectional LSTM with extracted features are presented in Table 9. The performance is slightly better than using raw signals. The performance does not experience drastic changes with changes in the window size. A window size of 2 s provided comparatively better scores for Ankle ( A C ), Trunk ( T C ) and Leg ( L C ) sensor data. Data from Ankle ( A C ) sensor provided the best overall score. The results when using Trunk ( T C ) sensor data slightly outperformed the scores when using Leg ( L C ) sensor data.
The scores from LSTMs (M5 and M6) were moderate, but the issue was the very long runtime. The time for training LSTMs on raw signals was almost 5 times and features was almost 3 times of that for training the CNNs on visual features. All three of our ensemble architectures M 7 , M 8 and M 9 , improved the scores of individual models. The majority voting model M9 had the best performance across all evaluation criteria without any extra training or parameter tuning. The scores were high for all evaluation criteria. The reported runtimes for ensemble models do not include the training time needed to prepare the constituent models.
Table 10 contains the scores of ensemble architecture M7 with all features. The overall performance is vastly superior to using individual features. The performance does not change significantly when the window size changes. A window size of 3 s generated the best scores for Ankle ( A C ) and Leg ( L C ) sensor data and a window size of 2 generated the best scores for data collected from the Trunk ( T C ). Data from Leg ( L C ) sensor provided the best overall scores. We can see that for M7, the runtimes were very large with all window sizes, which is a disadvantage considering this does not include the training time for individual models that make up the ensemble architecture. Adding the runtimes for constituent models, a significant amount of time was needed to train this model architecture.
Table 11 reports the scores of ensemble architecture M8 with all features. This architecture is similar to M7, except it adds an Average layer and calculates the average of the prediction of all constituent models, whereas M7 concatenates the predictions and uses two Dense layers to reshape the output. The overall performance is superior to using individual features and comparable to the performance of M7. The performance does not change significantly when the window size changes. For this model, window size of 3 s generated the best scores using Ankle ( A C ) and Leg ( L C ) sensor data and a window size of 2 generated the best scores for data collected from the Trunk ( T C ). Data from Trunk ( T C ) sensor provided the best overall scores. The runtimes for M8 were not that large when compared to M7, but the performances were comparable. Thus, adding the runtimes for constituent models, M8 was able to produce similar results while needing a lot less time for training.
Table 12 presents the scores of Majority voting architecture M9 with all features. This architecture is different from M7 and M8; there is no training for this method. The overall performance is similar to the performance of M7 and M8. The performance is not much affected when the window size changes. For this model, window size of 3 s generated the best scores using Ankle ( A C ) sensor data and a window size of 2 generated the best scores for data collected from the Trunk ( T C ) and Leg ( L C ). Data from Ankle ( A C ) sensor provided the best overall scores, very closely followed by Trunk ( T C ) and Leg ( L C ). However, the main strength of this model lies with its runtime, as it only outputs the majority result of its constituent models. Since there is no training time, it can generate the output in milliseconds. It can produce scores similar to M7 and M8 while not needing any extra training time.
The results were also compared with the performance of some state-of-the-art models on the same dataset, as shown in Table 3. Mazilu et al. [62] compared feature learning approaches based on time and statistical domain with unsupervised learning approaches using principle component analysis for both FOG detection and prediction. Their average sensitivity, specificity and F β score are presented for only the FOG class with both supervised and unsupervised approaches. Our proposed approach in this work outperforms their results for the FOG class. Moreover, the result they presented is only for the FOG class; their method had lower scores when identifying the PreFOG class.
Gokul et al. [64] presented a number of ML based techniques to detect FOG events and evaluate their performances with sensitivity and specificity, which are also presented in Table 3. They also experimented with multiple window sizes and achieved the best results with a window length of 4 s. Although their performance is higher than our proposed model, they solved a binary classification problem of only identifying the FOG event. Their work does not have a prediction component. They achieved the best results with Random Forest classifier, with a sensitivity score of 98.91 and a specificity score of 99.44. Our ensemble architecture results were very close to their scores, while being able to also predict the onset of a FOG event. However, compared to Gokul et al. [64], one shortcoming of our method would be the large size of the trained models, which might pose a problem in deploying the models to wearable sensors.

5. Application of Trained Model on Data Collected from APDM Sensors

We were able to obtain an additional dataset in collaboration with the Arizona School of Health Sciences, A.T. Still University. There were 14 PD patients in their experiments where gait was monitored and sensor data was recorded using APDM wearable sensors. Eight of the subjects were male and six were female. The subjects were aged between 55 and 77 years and the mean age of the subjects was 64.7. The average Hoenn & Yahr score for the subjects was 2.2 which indicated mild symptoms. The data made available to us was anonymous. Using the existing models trained on Daphnet [28] data and the Majority Voting Ensemble Architecture, our goal is to detect the onset of FOG in this real-world dataset.

5.1. Data Collection and Processing

The data was provided by the Arizona School of Health Sciences, A.T. Still University. The study was supervised, with the prior knowledge that 7 out of the 14 patients were Freezers, identified as a score > 0 on the New Freezing of Gait Questionnaire (NFOGQ) and the rest being non-Freezers, with a score of zero on the NFOGQ [20]. There were two different configurations of the Six-Minute Walk Test (6MWT), a common clinical assessment of walking endurance [75]: 50 and 100 feet. For both configurations, the duration of the study was fixed at 6 min. Each patient had to walk for 6 min continuously with a 180 degree turn after either a 50 or 100 feet walk respectively. The number of turns was less for the 100 feet configuration. For each of the 14 patients, two datasets were generated with one for each configuration. The data was cleaned and missing values were filled with zeros. Twenty-eight data files were generated with 14 containing Freezer data and 14 containing non-Freezer data.

Sensor Types and Locations

The sensors were placed in six locations on a patient’s body as illustrated in Figure 11.
The sensors were placed in Left Foot (Ankle), Right Foot (Ankle), Left Wrist, Right Wrist, Sternum and Lumber (Trunk).
For each location, the following five types of sensor were used to record data simultaneously:
  • Accelerometer.
  • Magnetometer.
  • Gyroscope.
  • Barometer.
  • Temperature.
The recorded data were processed using Moveo Explorer and Mobility Lab.

5.2. Workflow and Challenges

Figure 12 depicts our workflow for this part of the experiment. In order to make the data consistent across both Moveo Explorer and Mobility Lab, MinMax normalization was used. Since our purpose was only to monitor FOG, we discarded all the data where the patient did not experience FOG. As the result, data of 7 out of the 14 patients was discarded.
Since there were multiple sensors in multiple locations, finding the appropriate sensors suitable for our experiment was an initial step. There are multiple sensors producing a large amount of data and processing them all can be both time and resource intensive. Furthermore, it could lead to biased predictions or impacting the performance negatively. Identifying the optimal combination of sensors which was the most useful in detecting FOG posed a challenge. We decided to consider only sensor locations that overlap with Daphnet [28] sensor locations, i.e., Left and Right foot (Ankle) and Lumber (Trunk). Although data from multiple sensors as well as derived kinematic information like velocities and displacements was available, we only considered accelerations. This was because our training dataset, Daphnet [28], only provided acceleration data, and thus it was not possible for our currently pretrained architecture to use data from additional sensors like Magnetometer, Gyroscope or Barometer. Our training dataset also did not consist of data from sensors placed on the wrist or sternum area of the patient. Therefore, signals from those areas would not be helpful in this case. We combined the signals from all three axis ( X , Y , Z ) using Equation (2).
The sampling frequency for the data was 128 Hz, which is double the sampling frequency of Daphnet [28]. Since our models were trained on a sampling frequency of 64 Hz, the data needed to be down-sampled from 128 Hz to 64 Hz in order for the trained models to be effective.
Although the data was labeled, the labels were for the whole time series. That means we used one continuous 6 min time series, which had only one label indicating whether FOG occurred in that series or not. However, the exact occurrence and number of FOG were unmarked; we did not know when exactly the FOG event happened or how many FOG events are there. FOG events usually last shorter than 1 min, so there might be multiple occurrences FOG within one 6 min signal. Thus even if we did apply our models trained on Daphnet [28] to this data, it would be difficult to verify the accuracy of the results. No significant difference can be visually defined between the two types of signal for this dataset, because the very short FOG episodes occur in between 6 min of regular walking. So even the majority of Freezer data is basically similar to the non-Freezer data, with some FOG episodes in between.
In order to detect the occurrence of even a small FOG, a window size of 2 was selected. After selecting the appropriate sensors and down-sampling the data, non-overlapping moving windows of 2 s were extracted from the source signal. Then the relevant features were generated from the windows. We only chose to use visual features R P , S T F T , D W T , P W V D , since they have demonstrated good performance and were less time consuming as explained in the previous part of our experiment using the Daphnet [28] dataset.
After generating the features, the models M1, M2, M3 and M4 trained on Daphnet [28] data were used to analyze the data. There were three instances of trained models. For data from Left and Right foot, the instances of models trained on data from Ankle sensor of Daphnet was used. For data from Lumbar, models trained on data from Trunk sensor of Daphnet was used. Our training dataset Daphnet does not specify whether the sensors were located at the left or right side of the body. So Left and Right foot data was used with models trained on Ankle sensor. Finally, the results for each of the models were passed through a modified majority voting model M9, which generated the result.
For comparison, Figure 13 shows a small four window chunk of the signals from Accelerometer sensor of the right Ankle with a window size of 2 s. Since our data was down-sampled, this signal sequence has a sampling rate of 64 Hz instead of 128 Hz. The first Figure 13a shows a part of the sequence that does not contain any FOG or preFOG according to our model. Figure 13b shows a 4 s where our model identified PreFOG and FOG occurrences marked Yellow and Red respectively. When we compare the two figures, we can visually identify some differences. For the preFOG region, all the peak heights are smaller than that of non-FOG signals. The FOG region has higher peaks than that of preFOG, but the average peak height still appears to be lower than that of non-FOG signals. These patterns appear to be consistent with our observations from the Daphnet [28] data presented in Figure 3.

6. Conclusions

In this work, the performance of multiple time frequency representation techniques was compared in detecting and predicting FOG using tri-axial accelerometer sensor data from the publicly available Daphnet [28] dataset. We presented three ensemble neural network architectures comprised of multiple modalities of data and analyzed their performances. We were able to validate that ensemble network architectures significantly improve the performance over individual models.
The novelty of this work lies in the usage of multiple time frequency techniques and the usage of ensembling methodologies. Although there have been many experiments with ML and DL based technologies and feature extraction methods for FOG detection, as far as our results indicate and the literature review shows, this work is the first to apply time frequency representation techniques, i.e., Recurrence Plots, Short Time Fourier Transform, Discreet Wavelet Transform and Pseudo Wigner Ville Distribution, alongside captured signals and extracted features for FOG detection and prediction. Based on these innovative ideas, we have introduced effective ensemble architectures. Furthermore, the proposed solution is able to both detect and predict FOG events while achieving high performance scores based on multiple evaluation criteria. Our proposed ensemble methods outperform or are competitive to existing methods as demonstrated in Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12. Mazilu et al. [50] reported a sensitivity and specificity score of 66.25% and 95.23% respectively. Tripoliti et al. [51] achieved 81.94% sensitivity, 98.74% specificity and 96.11% detection accuracy. Sama et al. [55] used eight different classifiers achieving 91.7% and 87.4% for sensitivity and specificity respectively. Gokul et al. [64] report the best scores with Random Forest (89.91% Sensitivity and 99.44% Specificity), AdaBoost (97.99% specificity and 99.56% sensitivity), SVM (97.54% Sensitivity and 98.64% Specificity) and ProtoNN (95.25% Sensitivity and 99.66% Specificity) using a window size of 4 to solve a binary problem of FOG detection only. Comparatively, using a window size of 3, our Majority Voting Architecture M9 achieved accuracy, sensitivity and specificity of 98.5%, 98.5% and 97.9% solving a multiclass classification problem of both FOG detection and prediction. When we used a larger window size of 4, the sensitivity and specificity were around 96.9% and 96.7%. We also evaluated our models using MCC scores (i.e., 1 is the best), and the scores of our ensemble architectures were always above 0.9 which indicates that our approach had excellent performance in all confusion matrix categories for all classes. While solving a multiclass classification problem, i.e., in both identifying and predicting FOG events, our architectures achieve superior performance across all evaluation criteria.
We also applied the trained models to monitor the progression of real-world data, i.e., FOG from accelerometer data captured using APDM wearable sensors, which demonstrate that our approach is able to detect preFOG episodes in real world scenarios. In future works, we will integrate more sensors and data modalities, investigate more model combinations for creating ensemble architectures, reduce size and complexity of the models and finally apply the resultant models to more real-world data. More importantly, we will extend our models in identifying the preFOG class, i.e., predicting FOG events, to use real-world data from more practical wearable sensors, in order to test the potential of preventing falls by initiating RAS even before the start of the event. One of the drawbacks of the proposed system is that the performance was not verified when applied on some collected data due to the lack of proper annotations. Thus, our future goals also include properly testing the performance of our method with more precisely annotated data using wearable sensors. The ultimate objective is to deliver a system that is capable of administering RAS for the prediction of FOG and the prevention of falls.

Author Contributions

Conceptualization, T.A.M., S.S., I.C.; methodology, T.A.M.; software, T.A.M.; S.S.; validation, T.A.M.; S.S.; formal analysis, T.A.M.; investigation, T.A.M.; S.S.; resources, T.A.M., S.S., T.L.M.; data curation, T.A.M.; S.S., T.L.M.; I.C.; writing—original draft preparation, T.A.M., S.S.; writing—review and editing, T.A.M., S.S., T.L.M., I.C.; visualization, T.A.M., S.S.; supervision, I.C.; project administration, I.C.; funding acquisition, I.C. All authors have read and agreed to the published version of the manuscript.

Funding

Financial support from the Natural Sciences and Engineering Research Council of Canada (NSERC) is gratefully acknowledged.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Patient consent was waived due to the subjects of the study not being identifiable.

Data Availability Statement

The Daphnet dataset used for training the model is publicly available. Supplementary code will be made available on request to the correspondent author’s email with appropriate justification.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analysis, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
FOGFreezing of Gait
PDParkinson’s Disease
RASRhythmic Auditory Stimulation
RPRecurrence Plot
STFTShort Time Fourier Transform
DWTDiscreet Wavelet Transform
PWVDPseudo Wigner Ville Distribution
DLDeep Learning
MLMachine Learning
LSTMLong Short Term Memory
CNNConcolutional Neural Network
RNNRecurrent Neural Network
ADLActivities of Daily Living
HARHuman Activity Recognition
WBWalking Band
FBFreezing Band
FTHFreezing Threshold
PTHPower Threshold
FIFreeze Index
PIPower Index
RFRandom Forest
KNNK Nearest Neighbours
AUCArea Under Curve
ROCReceiver Operating Characteristic
SVMSupport Vector Machine
LOSOLeave One Subject Out
GBGradient Boosting
RBFRadial Basis Function
IMUInertial Measurement Unit
SQASpeech Quality Assesment
MFCCMel Frequency Cepstral Coefficients
EEREqual Error Rate
FFTFast Fourier Transform
MCCMatthews Correlation Coefficient
TPTrue Positive
FPFalse Positive
TNTrue Negative
FNFalse Negative
2AFCTwo-Alternative-Forced-Choice
DSCQSDouble Stimulus Continuous Quality Scale
U-AMSActivity Monitoring System
GMGeometric Mean
SQASpeech Quality Assessment
HMMHidden Markov Model

References

  1. Vos, T.; Allen, C.; Arora, M.; Barber, R.M.; Bhutta, Z.A.; Brown, A.; Carter, A.; Casey, D.C.; Charlson, F.J.; Chen, A.Z.; et al. Global, regional, and national incidence, prevalence, and years lived with disability for 310 diseases and injuries, 1990–2015: A systematic analysis for the Global Burden of Disease Study 2015. Lancet 2016, 388, 1545–1602. [Google Scholar] [CrossRef] [Green Version]
  2. Statistics. (n.d.). Retrieved 26 September 2021. Available online: https://www.parkinson.org/Understanding-Parkinsons/Statistics (accessed on 13 August 2021).
  3. Reeve, A.; Simcox, E.; Turnbull, D. Ageing and Parkinson’s disease: Why is advancing age the biggest risk factor? Ageing Res. Rev. 2014, 14, 19–30. [Google Scholar] [CrossRef] [PubMed]
  4. Goetz, C.G. The History of Parkinson’s Disease: Early Clinical Descriptions and Neurological Therapies. Cold Spring Harb. Perspect. Med. 2011, 1, a008862. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Wirdefeldt, K.; Adami, H.O.; Cole, P.; Trichopoulos, D.; Mandel, J. Epidemiology and etiology of Parkinson’s disease: A review of the evidence. Eur. J. Epidemiol. 2011, 26, 1–58. [Google Scholar] [CrossRef] [PubMed]
  6. Michel, P.P.; Hirsch, E.C.; Hunot, S. Understanding Dopaminergic Cell Death Pathways in Parkinson Disease. Neuron 2016, 90, 675–691. [Google Scholar] [CrossRef] [Green Version]
  7. Sveinbjornsdottir, S. The clinical symptoms of Parkinson’s disease. J. Neurochem. 2016, 139, 318–324. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Cheng, H.C.; Ulane, C.M.; Burke, R.E. Clinical progression in Parkinson disease and the neurobiology of axons. Ann. Neurol. 2010, 67, 715–725. [Google Scholar] [CrossRef]
  9. Giladi, N.; Treves, T.A.; Simon, E.S.; Shabtai, H.; Orlov, Y.; Kandinov, B.; Paleacu, D.; Korczyn, A.D. Freezing of gait in patients with advanced Parkinson’s disease. J. Neural Transm. 2001, 108, 53–61. [Google Scholar] [CrossRef]
  10. Okuma, Y.; Yanagisawa, N. The clinical spectrum of freezing of gait in Parkinson’s disease. Mov. Disord. 2008, 23, S426–S430. [Google Scholar] [CrossRef]
  11. Schaafsma, J.D.; Balash, Y.; Gurevich, T.; Bartels, A.L.; Hausdorff, J.M.; Giladi, N. Characterization of freezing of gait subtypes and the response of each to levodopa in Parkinson’s disease. Eur. J. Neurol. 2003, 10, 391–398. [Google Scholar] [CrossRef] [Green Version]
  12. Ahlrichs, C.; Samá, A.; Lawo, M.; Cabestany, J.; Rodríguez-Martín, D.; Pérez-López, C.; Sweeney, D.; Quinlan, L.R.; Laighin, G.O.; Counihan, T.; et al. Detecting freezing of gait with a tri-axial accelerometer in Parkinson’s disease patients. Med Biol. Eng. Comput. 2015, 54, 223–233. [Google Scholar] [CrossRef] [Green Version]
  13. Kim, H.; Lee, H.J.; Lee, W.; Kwon, S.; Kim, S.K.; Jeon, H.S.; Park, H.; Shin, C.W.; Yi, W.J.; Jeon, B.S.; et al. Unconstrained detection of freezing of Gait in Parkinson’s disease patients using smartphone. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015. [Google Scholar] [CrossRef]
  14. Latt, M.D.; Lord, S.R.; Morris, J.G.; Fung, V.S. Clinical and physiological assessments for elucidating falls risk in Parkinson’s disease. Mov. Disord. 2009, 24, 1280–1289. [Google Scholar] [CrossRef]
  15. Kerr, G.K.; Worringham, C.J.; Cole, M.H.; Lacherez, P.F.; Wood, J.M.; Silburn, P.A. Predictors of future falls in Parkinson disease. Neurology 2010, 75, 116–124. [Google Scholar] [CrossRef]
  16. Moore, O.; Peretz, C.; Giladi, N. Freezing of gait affects quality of life of peoples with Parkinson’s disease beyond its relationships with mobility and gait. Mov. Disord. 2007, 22, 2192–2195. [Google Scholar] [CrossRef]
  17. Okuma, Y. Freezing of Gait and Falls in Parkinson’s Disease. J. Park. Dis. 2014, 4, 255–260. [Google Scholar] [CrossRef]
  18. Giladi, N.; Shabtai, H.; Simon, E.; Biran, S.; Tal, J.; Korczyn, A. Construction of freezing of gait questionnaire for patients with Parkinsonism. Park. Relat. Disord. 2000, 6, 165–170. [Google Scholar] [CrossRef]
  19. Giladi, N.; Tal, J.; Azulay, T.; Rascol, O.; Brooks, D.J.; Melamed, E.; Oertel, W.; Poewe, W.H.; Stocchi, F.; Tolosa, E.; et al. Validation of the freezing of gait questionnaire in patients with Parkinson’s disease. Mov. Disord. 2009, 24, 655–661. [Google Scholar] [CrossRef]
  20. Nieuwboer, A.; Rochester, L.; Herman, T.; Vandenberghe, W.; Emil, G.E.; Thomaes, T.; Giladi, N. Reliability of the new freezing of gait questionnaire: Agreement between patients with Parkinson’s disease and their carers. Gait Posture 2009, 30, 459–463. [Google Scholar] [CrossRef] [PubMed]
  21. Nieuwboer, A.; Weerdt, W.D.; Dom, R.; Lesaffre, E. A frequency and correlation analysis of motor deficits in Parkinson patients. Disabil. Rehabil. 1998, 20, 142–150. [Google Scholar] [CrossRef] [PubMed]
  22. Sazonov, E. Wearable Sensors: Fundamentals, Implementation and Applications; Academic Press: Cambridge, MA, USA; Elsevier: Amsterdam, The Netherlands, 2021. [Google Scholar]
  23. Hassan, M.M.; Huda, S.; Uddin, M.Z.; Almogren, A.; Alrubaian, M. Human Activity Recognition from Body Sensor Data using Deep Learning. J. Med. Syst. 2018, 42, 99. [Google Scholar] [CrossRef]
  24. Nweke, H.F.; Teh, Y.W.; Al-garadi, M.A.; Alo, U.R. Deep learning algorithms for human activity recognition using mobile and wearable sensor networks: State of the art and research challenges. Expert Syst. Appl. 2018, 105, 233–261. [Google Scholar] [CrossRef]
  25. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep learning for sensor-based activity recognition: A survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef] [Green Version]
  26. Ordóñez, F.; Roggen, D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [Green Version]
  27. Lara, O.D.; Labrador, M.A. A Survey on Human Activity Recognition using Wearable Sensors. IEEE Commun. Surv. Tutor. 2013, 15, 1192–1209. [Google Scholar] [CrossRef]
  28. Bachlin, M.; Plotnik, M.; Roggen, D.; Maidan, I.; Hausdorff, J.M.; Giladi, N.; Troster, G. Wearable Assistant for Parkinson’s Disease Patients With the Freezing of Gait Symptom. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 436–446. [Google Scholar] [CrossRef]
  29. Jovanov, E.; Wang, E.; Verhagen, L.; Fredrickson, M.; Fratangelo, R. A real time system for detection and unfreezing of gait of Parkinson’s patients. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 3–6 September 2009. [Google Scholar] [CrossRef] [Green Version]
  30. Kubota, K.J.; Chen, J.A.; Little, M.A. Machine learning for large-scale wearable sensor data in Parkinson’s disease: Concepts, promises, pitfalls, and futures. Mov. Disord. 2016, 31, 1314–1326. [Google Scholar] [CrossRef]
  31. Matias, R.; Paixão, V.; Bouça, R.; Ferreira, J.J. A Perspective on Wearable Sensor Measurements and Data Science for Parkinson’s Disease. Front. Neurol. 2017, 8, 677. [Google Scholar] [CrossRef] [Green Version]
  32. Sánchez-Ferro, A.; Elshehabi, M.; Godinho, C.; Salkovic, D.; Hobert, M.A.; Domingos, J.; van Uem, M.; Ferreira, J.J.; Maetzler, W. New methods for the assessment of Parkinson’s disease (2005 to 2015): A systematic review. Mov. Disord. 2016, 31, 1283–1292. [Google Scholar] [CrossRef]
  33. Maetzler, W.; Domingos, J.; Srulijes, K.; Ferreira, J.J.; Bloem, B.R. Quantitative wearable sensors for objective assessment of Parkinson’s disease. Mov. Disord. 2013, 28, 1628–1637. [Google Scholar] [CrossRef]
  34. Hashimoto, T. Speculation on the responsible sites and pathophysiology of freezing of gait. Park. Relat. Disord. 2006, 12, S55–S62. [Google Scholar] [CrossRef]
  35. Arias, P.; Cudeiro, J. Effect of Rhythmic Auditory Stimulation on Gait in Parkinsonian Patients with and without Freezing of Gait. PLoS ONE 2010, 5, e9675. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Rocha, P.A.; Porfírio, G.M.; Ferraz, H.B.; Trevisani, V.F. Effects of external cues on gait parameters of Parkinson’s disease patients: A systematic review. Clin. Neurol. Neurosurg. 2014, 124, 127–134. [Google Scholar] [CrossRef]
  37. Lim, I.; van Wegen, E.; de Goede, C.; Deutekom, M.; Nieuwboer, A.; Willems, A.; Jones, D.; Rochester, L.; Kwakkel, G. Effects of external rhythmical cueing on gait in patients with Parkinson’s disease: A systematic review. Clin. Rehabil. 2005, 19, 695–713. [Google Scholar] [CrossRef]
  38. LeCun, Y.; Bengio, Y. Convolutional Networks for Images, Speech, and Time Series. In The Handbook of Brain Theory and Neural Networks; MIT Press: Cambridge, MA, USA, 1998; pp. 255–258. [Google Scholar]
  39. Pascanu, R.; Gulcehre, C.; Cho, K.; Bengio, Y. How to Construct Deep Recurrent Neural Networks. arXiv 2014, arXiv:1312.6026. [Google Scholar]
  40. Karim, F.; Majumdar, S.; Darabi, H.; Chen, S. LSTM Fully Convolutional Networks for Time Series Classification. IEEE Access 2018, 6, 1662–1669. [Google Scholar] [CrossRef]
  41. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  42. Sollich, P.; Krogh, A. Learning with ensembles: How overfitting can be useful. In Proceedings of the 1995 Conference, Denver, CO, USA, 27–30 November 1995; Volume 8, pp. 190–196. [Google Scholar]
  43. Hansen, L.; Salamon, P. Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 993–1001. [Google Scholar] [CrossRef] [Green Version]
  44. Mostafa, T.A.; Cheng, I. Parkinson’s Disease Detection with Ensemble Architectures based on ILSVRC Models. arXiv 2020, arXiv:2007.12496. [Google Scholar]
  45. Ashfaque Mostafa, T.; Cheng, I. Parkinson’s Disease Detection Using Ensemble Architecture from MR Images*. In Proceedings of the 2020 IEEE 20th International Conference on Bioinformatics and Bioengineering (BIBE), Cincinnati, OH, USA, 26–28 October 2020. [Google Scholar] [CrossRef]
  46. Camps, J.; Samà, A.; Martín, M.; Rodríguez-Martín, D.; Pérez-López, C.; Alcaine, S.; Mestre, B.; Prats, A.; Crespo, M.C.; Cabestany, J.; et al. Advances in Computational Intelligence; Rojas, I., Joya, G., Catala, A., Eds.; Springer International Publishing: Cham, Switzerland, 2017; pp. 344–355. [Google Scholar]
  47. Soltaninejad, S.; Cheng, I.; Basu, A. Kin-FOG: Automatic Simulated Freezing of Gait (FOG) Assessment System for Parkinson’s Disease. Sensors 2019, 19, 2416. [Google Scholar] [CrossRef] [Green Version]
  48. Han, J.H.; Lee, W.J.; Ahn, T.B.; Jeon, B.S.; Park, K.S. Gait analysis for freezing detection in patients with movement disorder using three dimensional acceleration system. In Proceedings of the 25th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (IEEE Cat. No.03CH37439), Cancun, Mexico, 17–21 September 2003; IEEE: Piscataway, NJ, USA, 2003. [Google Scholar] [CrossRef]
  49. Moore, S.T.; MacDougall, H.G.; Ondo, W.G. Ambulatory monitoring of freezing of gait in Parkinson’s disease. J. Neurosci. Methods 2008, 167, 340–348. [Google Scholar] [CrossRef] [PubMed]
  50. Mazilu, S.; Hardegger, M.; Zhu, Z.; Roggen, D.; Troester, G.; Plotnik, M.; Hausdorff, J. Online Detection of Freezing of Gait with Smartphones and Machine Learning Techniques. In Proceedings of the 6th International Conference on Pervasive Computing Technologies for Healthcare, San Diego, CA, USA, 21–24 May 2012; IEEE: Piscataway, NJ, USA, 2012. [Google Scholar] [CrossRef]
  51. Tripoliti, E.E.; Tzallas, A.T.; Tsipouras, M.G.; Rigas, G.; Bougia, P.; Leontiou, M.; Konitsiotis, S.; Chondrogiorgi, M.; Tsouli, S.; Fotiadis, D.I. Automatic detection of freezing of gait events in patients with Parkinson’s disease. Comput. Methods Programs Biomed. 2013, 110, 12–26. [Google Scholar] [CrossRef]
  52. Moore, S.T.; Yungher, D.A.; Morris, T.R.; Dilda, V.; MacDougall, H.G.; Shine, J.M.; Naismith, S.L.; Lewis, S.J. Autonomous identification of freezing of gait in Parkinson’s disease from lower-body segmental accelerometry. J. NeuroEng. Rehabil. 2013, 10, 19. [Google Scholar] [CrossRef] [Green Version]
  53. Zach, H.; Janssen, A.M.; Snijders, A.H.; Delval, A.; Ferraye, M.U.; Auff, E.; Weerdesteyn, V.; Bloem, B.R.; Nonnekes, J. Identifying freezing of gait in Parkinson’s disease during freezing provoking tasks using waist-mounted accelerometry. Park. Relat. Disord. 2015, 21, 1362–1366. [Google Scholar] [CrossRef]
  54. Rodríguez-Martín, D.; Samà, A.; Pérez-López, C.; Català, A.; Arostegui, J.M.M.; Cabestany, J.; Bayés, À.; Alcaine, S.; Mestre, B.; Prats, A.; et al. Home detection of freezing of gait using support vector machines through a single waist-worn triaxial accelerometer. PLoS ONE 2017, 12, e0171764. [Google Scholar] [CrossRef]
  55. Samà, A.; Rodríguez-Martín, D.; Pérez-López, C.; Català, A.; Alcaine, S.; Mestre, B.; Prats, A.; Crespo, M.C.; Bayés, À. Determining the optimal features in freezing of gait detection through a single waist accelerometer in home environments. Pattern Recognit. Lett. 2018, 105, 135–143. [Google Scholar] [CrossRef]
  56. Orphanidou, N.K.; Hussain, A.; Keight, R.; Lishoa, P.; Hind, J.; Al-Askar, H. Predicting Freezing of Gait in Parkinsons Disease Patients Using Machine Learning. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC), Rio de Janeiro, Brazil, 8–13 July 2018; IEEE: Piscataway, NJ, USA, 2018. [Google Scholar] [CrossRef]
  57. Camps, J.; Samà, A.; Martín, M.; Rodríguez-Martín, D.; Pérez-López, C.; Arostegui, J.M.M.; Cabestany, J.; Català, A.; Alcaine, S.; Mestre, B.; et al. Deep learning for freezing of gait detection in Parkinson’s disease patients in their homes using a waist-worn inertial measurement unit. Knowl.-Based Syst. 2018, 139, 119–131. [Google Scholar] [CrossRef]
  58. San-Segundo, R.; Navarro-Hellín, H.; Torres-Sánchez, R.; Hodgins, J.; De la Torre, F. Increasing Robustness in the Detection of Freezing of Gait in Parkinson’s Disease. Electronics 2019, 8, 119. [Google Scholar] [CrossRef] [Green Version]
  59. San-Segundo, R.; Montero, J.M.; Barra-Chicote, R.; Fernández, F.; Pardo, J.M. Feature extraction from smartphone inertial signals for human activity segmentation. Signal Process. 2016, 120, 359–372. [Google Scholar] [CrossRef]
  60. Sigcha, L.; Costa, N.; Pavón, I.; Costa, S.; Arezes, P.; López, J.M.; Arcas, G.D. Deep Learning Approaches for Detecting Freezing of Gait in Parkinson’s Disease Patients through On-Body Acceleration Sensors. Sensors 2020, 20, 1895. [Google Scholar] [CrossRef] [Green Version]
  61. Hoehn, M.; Yahr, M. Parkinsonism: Onset, progression, and mortality. Neurology 1967, 77, 874. [Google Scholar] [CrossRef] [Green Version]
  62. Mazilu, S.; Calatroni, A.; Gazit, E.; Roggen, D.; Hausdorff, J.M.; Tröster, G. Feature Learning for Detection and Prediction of Freezing of Gait in Parkinson’s Disease. In Machine Learning and Data Mining in Pattern Recognition; Springer: Berlin/Heidelberg, Germany, 2013; pp. 144–158. [Google Scholar] [CrossRef]
  63. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  64. Gokul, H.; Suresh, P.; Hari Vignesh, B.; Pravin Kumaar, R.; Vijayaraghavan, V. Gait Recovery System for Parkinson’s Disease using Machine Learning on Embedded Platforms. In Proceedings of the 2020 IEEE International Systems Conference (SysCon), Montreal, QC, Canada, 24 August–20 September 2020. [Google Scholar] [CrossRef]
  65. Eckmann, J.P.; Kamphorst, S.O.; Ruelle, D. Recurrence Plots of Dynamical Systems. Europhys. Lett. (EPL) 1987, 4, 973–977. [Google Scholar] [CrossRef] [Green Version]
  66. Recurrence Plot. 12 August 2021. Available online: https://en.wikipedia.org/wiki/Recurrence_plot (accessed on 13 August 2021).
  67. Rajoub, B. Characterization of biomedical signals: Feature engineering and extraction. In Biomedical Signal Processing and Artificial Intelligence in Healthcare; Academic Press: Cambridge, MA, USA, 2020; pp. 29–50. [Google Scholar] [CrossRef]
  68. Hosseinzadeh, M. Robust control applications in biomedical engineering: Control of depth of hypnosis. In Control Applications for Biomedical Engineering Systems; Academic Press: Cambridge, MA, USA, 2020; pp. 89–125. [Google Scholar] [CrossRef]
  69. Haar, A. Zur Theorie der orthogonalen Funktionensysteme. Math. Ann. 1911, 71, 38–53. [Google Scholar] [CrossRef]
  70. Shin, Y.S.; Jeon, J.J. Pseudo Wigner–Ville Time-Frequency Distribution and Its Application to Machinery Condition Monitoring. Shock Vib. 1993, 1, 65–76. [Google Scholar] [CrossRef]
  71. Zhou, Z.H.; Wu, J.; Tang, W. Corrigendum to “Ensembling neural networks: Many could be better than all” [Artificial Intelligence 137 (1–2) (2002) 239–263]. Artif. Intell. 2010, 174, 1570. [Google Scholar] [CrossRef] [Green Version]
  72. Matthews, B. Comparison of the predicted and observed secondary structure of T4 phage lysozyme. Biochim. Biophys. Acta (BBA)-Protein Struct. 1975, 405, 442–451. [Google Scholar] [CrossRef]
  73. Gorodkin, J. Comparing two K-category assignments by a K-category correlation coefficient. Comput. Biol. Chem. 2004, 28, 367–374. [Google Scholar] [CrossRef]
  74. Metrics and Scoring: Quantifying the Quality of Predictions. (n.d.). Available online: https://scikit-learn.org/stable/modules/model_evaluation.html (accessed on 13 August 2021).
  75. Steffen, T.; Seney, M. Test-Retest reliability and minimal detectable change on balance AND Ambulation tests, the 36-ITEM Short-Form Health survey, and the unified Parkinson Disease rating scale in people with parkinsonism. Phys. Ther. 2008, 88, 733–746. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Sensor placement for data collection.
Figure 1. Sensor placement for data collection.
Sensors 21 06446 g001
Figure 2. Proposed preprocessing, data augmentation and feature extraction workflow.
Figure 2. Proposed preprocessing, data augmentation and feature extraction workflow.
Sensors 21 06446 g002
Figure 3. Example of combined Accelerometer signal from Ankle, capturing the motor variations in the gait of a PD patient, containing Normal gait, followed by a window of PreFOG period (Yellow), and then a FOG event (Red).
Figure 3. Example of combined Accelerometer signal from Ankle, capturing the motor variations in the gait of a PD patient, containing Normal gait, followed by a window of PreFOG period (Yellow), and then a FOG event (Red).
Sensors 21 06446 g003
Figure 4. Examples of recurrence plot for (a) signals representing Normal walking or Non-FOG, (b) signals representing PreFOG and (c) signals representing FOG.
Figure 4. Examples of recurrence plot for (a) signals representing Normal walking or Non-FOG, (b) signals representing PreFOG and (c) signals representing FOG.
Sensors 21 06446 g004
Figure 5. Examples of Short Time Fourier Transform plot for (a) signals representing Normal walking or Non-FOG, (b) signals representing PreFOG and (c) signals representing FOG.
Figure 5. Examples of Short Time Fourier Transform plot for (a) signals representing Normal walking or Non-FOG, (b) signals representing PreFOG and (c) signals representing FOG.
Sensors 21 06446 g005
Figure 6. Discrete Wavelet Transformation plot for (a) signals representing Normal walking, (b) signals representing PreFOG and (c) signals representing FOG.
Figure 6. Discrete Wavelet Transformation plot for (a) signals representing Normal walking, (b) signals representing PreFOG and (c) signals representing FOG.
Sensors 21 06446 g006
Figure 7. Examples of Pseudo Wigner Ville Distribution plot for (a) signals representing Normal walking, (b) signals representing PreFOG and (c) signals representing FOG.
Figure 7. Examples of Pseudo Wigner Ville Distribution plot for (a) signals representing Normal walking, (b) signals representing PreFOG and (c) signals representing FOG.
Sensors 21 06446 g007
Figure 8. Proposed basic CNN Architecture with 4 recurring 2D Convolution blocks, followed by 3 Dense layers.
Figure 8. Proposed basic CNN Architecture with 4 recurring 2D Convolution blocks, followed by 3 Dense layers.
Sensors 21 06446 g008
Figure 9. Proposed basic Bidirectional LSTM Architecture with 4 recurring Bidirectional LSTM blocks, followed by a Dense Layer.
Figure 9. Proposed basic Bidirectional LSTM Architecture with 4 recurring Bidirectional LSTM blocks, followed by a Dense Layer.
Sensors 21 06446 g009
Figure 10. Proposed Ensemble Architectures, (1) M7 concatenates the output all constituent models, followed by a Dense Layer, (2) M8 Averages the outputs of all constituent models and (3) M9 calculates the majority prediction of all models using mode.
Figure 10. Proposed Ensemble Architectures, (1) M7 concatenates the output all constituent models, followed by a Dense Layer, (2) M8 Averages the outputs of all constituent models and (3) M9 calculates the majority prediction of all models using mode.
Sensors 21 06446 g010
Figure 11. Sensor placement for data collection.
Figure 11. Sensor placement for data collection.
Sensors 21 06446 g011
Figure 12. Workflow of monitoring FOG using models trained on Daphnet data.
Figure 12. Workflow of monitoring FOG using models trained on Daphnet data.
Sensors 21 06446 g012
Figure 13. Comparison of Freezer vs. non-Freezer Accelerometer signals with (a) non-Freezer and (b) Detected preFOG and FOG from the Right Leg (Ankle).
Figure 13. Comparison of Freezer vs. non-Freezer Accelerometer signals with (a) non-Freezer and (b) Detected preFOG and FOG from the Right Leg (Ankle).
Sensors 21 06446 g013
Table 1. Comparison between technologies of various approaches.
Table 1. Comparison between technologies of various approaches.
AuthorsTechnology Used
Han et al. [48]U-AMS (Activity Monitoring System) using wavelet power features for discrimination of abnormal movement.
Moore et al. [49]Threshold based method for FOG detection by defining the Freeze Index (FI), achieving 78 % True Positive Rate
Bachlin et al. [28]Added Power Index (PI) to the method proposed by Moore et al. [49]
Bachlin et al. [28]Two thresholds (FTH and PTH) given F I > F T H and P I > P T H , achieving 73.1 % sensitivity and 81.6 % specificity
Mazilu et al. [50]Additional Features along with Bachlin et al. [28] features used with ML models RF, Naive Bayes and KNN, achieving 66.25 % sensitivity and 95.23 % specificity
Tripoliti et al. [51]Data imputation (interpolation), band-pass filtering, entropy calculation, and automatic classification (Naïve Bayes, RF, Decision Trees and Random Tree), resulting in 81.94 % sensitivity, 98.74 % specificity, 96.11 % accuracy and 98.6 % AUC
Moore et al. [52]Assessed seven sensors placed in different locations and concluded that the shank and back were the most convenient places for the sensors. Using all the seven sensors achieved best performance with sensitivity 84.3 % and specificity 78.4 %
Zack et al. [53]presented a threshold based FOG detection technique following Moore [52] using ROC to determine a global FI threshold to distinguish between FOG and non-FOG episodes for different tasks, achieving sensitivity of 75 % and specificity of 76 %
Rodríguez et al. [54]used ML techniques and 55 FOG related features from 21 PD patients with SVM. reporting a sensitivity of 88.09 % and specificity of 80.09 % for user independent evaluation, and a sensitivity of 79.03 % and specificity of 74.67 % for LOSO evaluation
Sama et al. [55]Decreased the number of features to 28 for the Rodríguez et al. [54] dataset and used 8 different classifiers with greedy subset selection process, 10-fold cross-validation and different window sizes, achieving 91.7 % and 87.4 % for sensitivity and specificity respectively
Orphanidou et al. [56]Extracted features and presented to 7 machine learning classifiers, with SVM achieving the highest performance. The classification algorithm was applied to 5 s windows using 18 features, obtaining balanced accuracy (the mean value of sensitivity and specificity) of 91 % , 90 % , and 82 % over the Walk, FOG and Transition classes, respectively
Orphanidou et al. [56]Focused on the early detection, through classification of the transition class using varying size time windows and time/frequency contrary to the majority of previous studies that recognized FOG only when it had occurred by including another class label called ‘transition’ that showed the episodes before FOG occurrence. 5 classifiers were used, including Gradient Boosting (GB), Extreme Gradient Boosting, SVM, RF and Neural Networks. SVM with RBF kernels has the best performance with sensitivity of 72.34 % , 91.49 % , 75.00 % and specificity values of 87.36 % , 88.51 % and 93.62 % , for FOG, transition and normal activity classes, respectively
Camps et al. [57]Proposed 1D Convolutions Neural Network (CNN) with 8 layers, trained using a novel spectral data representation strategy that considers information from both the previous and current signal windows showing a performance of 90.6 % for the GM, an AUC of 0.88 , a sensitivity of 91.9 % and a sensibility of 89.5 %
San-Segundo et al. [58]Evaluated the robustness of different feature sets and ML algorithms for FOG detection using body-worn accelerometers with 4 feature sets: (Mazilu et al. [50], HAR, MFCCs, and SQA) using 4 classification methods (RF, multi-layer perceptron, HMM and deep neural networks).The current window and three previous windows, with the feature set composed of Mazilu features [50] and MFCCs [59] achieved best scores. Deep convolutions neural network achieved an AUC of 0.93 and an Equal Error Rate (EER) of 12.5 % .
Sigcha et al. [60]Reproduced Mazilu features [50], MFCCs [59], and FFT) to establish a baseline using RF classifier with 10-fold cross-validation (R10fold) and LOSO. Further, tested multiple DL approaches include: A denoiser autoencoder, a deep neural network with CNN and a combination of CNN and LSTM layers and compared with shallow algorithms such as one-class SVM (OC-SVM), SVM, AdaBoost and RF, achieving 0.93 AUC
Our Proposed MethodUsing all three sensors from Daphnet [28] dataset, used moving windows extracted from the original signal, handcrafted feature set and time frequency visualization techniques including RP, STFT, DWT and PWVD alongside a CNN and LSTM architecture. Evaluated the performance and used trained models to create 3 ensemble architectures which further improve the performance. The proposed method is capable of both FOG detection and prediction and the evaluation scores show better performance when compared to existing approaches.
Table 2. F i Features extracted for each α i α C .
Table 2. F i Features extracted for each α i α C .
Time Domain FeaturesDescription
Min, MaxMinimum and Maximum value of the signal
RangeDifference between the minimum and maximum value of the signal
MeanAverage value of signal
MedianMedian value of the signal
ModeModal value of the signal
Trimmed MeanTrimmed/Truncated mean of the signal
Standard DeviationDeviation of a signal compared to its mean
VarianceSquare root of the standard deviation of the signal
Root mean squareSquare root of the mean of the squared signal
Mean absolute valueMean of absolute value of the signal
Median absolute deviationMedian over the absolute deviations from the median
25th Percentile25th percentile value of the signal
75th Percentile75th percentile value of the signal
Interquantile rangeDifference between the 75th and 25th percentile of the signal
Normalized Signal Magnitude AreaSum of standardized acceleration magnitude normalized by window length
SkewnessThe degree of asymmetry in the signal
KurtosisThe degree of peakedness in the signal, signals with high kurtosis have more outliers
Mean Crossing RateThe number of times the signals goes from above average value to below average value normalized by the window length
Signal Vector MagnitudeSum of euclidean norm over the window normalized by window length
Peak of Fourier TransformMaximum magnitude of Discrete Fourier Transform of the signal normalized by the window length
Frequency Domain FeaturesDescription
EntropyMeasure of random distribution of frequency
EnergySum of squared magnitude of FFT of the signal divided by window length
Peak FrequencyMaximum frequency value in the power spectrum
Freeze Band PowerThe sum of power in Freeze band of frequencies divided by sampling frequency
Locomotion Band PowerThe sum of power in Locomotion band of frequencies divided by sampling frequency
Freeze IndexPower of signal in freeze band (3-8Hz) divided by its Power in locomotion band(0.5-3Hz)
Band PowerSum of the power in freeze band and in locomotion band
Table 3. The Metric scores of some related works on similar data.
Table 3. The Metric scores of some related works on similar data.
ModelWindow (s)SensitivitySpecificity F β Score
Mazilu et al. [62] (Unsupervised—20 Features)376.8686.2181.56
Mazilu et al. [62] (Supervised—20 Features)366.6588.7478.27
Mazilu et al. [62] (Unsupervised—25 Features)376.8685.5280.82
Mazilu et al. [62] (Supervised—25 Features)367.5888.5278.65
Decision Tree [64]496.7098.92-
Random Forest [64]498.9199.44-
AdaBoost [64]497.9999.56-
KNN [64]494.6197.38-
SVM [64]497.5498.64-
ProtoNN [64]495.2599.66-
Bonsai [64]492.998.36-
Table 4. Results of Basic CNN M1 with RP.
Table 4. Results of Basic CNN M1 with RP.
Data TypeWindow (s)AccuracyPrecisionSensitivitySpecificity F β ScoreMCCRuntime (min)
A C
(3)
20.894 ± 0.0210.929 ± 0.0060.894 ± 0.0210.933 ± 0.0070.904 ± 0.0160.687 ± 0.03952.05
30.832 ± 0.0230.905 ± 0.0110.832 ± 0.0230.926 ± 0.0100.849 ± 0.0210.648 ± 0.03517:41
40.864 ± 0.0300.901 ± 0.0200.864 ± 0.0300.925 ± 0.0200.873 ± 0.0270.695 ± 0.05811:52
L C
(3)
20.891 ± 0.0190.929 ± 0.0080.891 ± 0.0190.937 ± 0.0080.902 ± 0.0160.684 ± 0.03757:31
30.873 ± 0.0050.915 ± 0.0060.873 ± 0.0050.937 ± 0.0060.883 ± 0.0050.702 ± 0.01422:41
40.840 ± 0.0220.895 ± 0.0130.840 ± 0.0220.930 ± 0.0140.853 ± 0.0200.664 ± 0.0379:25
T C
(3)
20.926 ± 0.0130.942 ± 0.0070.926 ± 0.0130.944 ± 0.0090.930 ± 0.0110.755 ± 0.03364:58
30.876 ± 0.0210.921 ± 0.0080.876 ± 0.0210.924 ± 0.0080.889 ± 0.0170.649 ± 0.03815:24
40.849 ± 0.0000.911 ± 0.0000.849 ± 0.0000.934 ± 0.0000.863 ± 0.0000.667 ± 0.0009:59
Table 5. Results of Basic CNN M2 with STFT.
Table 5. Results of Basic CNN M2 with STFT.
Data TypeWindow (s)AccuracyPrecisionSensitivitySpecificity F β ScoreMCCRuntime (min)
A C
(3)
20.678 ± 0.0160.893 ± 0.0080.678 ± 0.01760.860 ± 0.0070.742 ± 0.0110.407 ± 0.01468.11
30.799 ± 0.0300.883 ± 0.0100.799 ± 0.0300.897 ± 0.0120.819 ± 0.0260.585 ± 0.03818:41
40.781 ± 0.0300.877 ± 0.0050.781 ± 0.0300.905 ± 0.0100.803 ± 0.0250.588 ± 0.02611:22
L C
(3)
20.719 ± 0.0230.889 ± 0.0090.719 ± 0.0230.864 ± 0.0170.762 ± 0.0190.452 ± 0.035127:59
30.815 ± 0.0250.902 ± 0.0060.815 ± 0.0250.921 ± 0.0080.834 ± 0.0210.633 ± 0.03021:02
40.746 ± 0.0360.865 ± 0.0100.746 ± 0.0360.894 ± 0.0130.770 ± 0.0310.550 ± 0.0328:55
T C
(3)
20.781 ± 0.0100.899 ± 0.0020.781 ± 0.0100.894 ± 0.0030.813 ± 0.0080.516 ± 0.00857:19
30.831 ± 0.0370.911 ± 0.0130.831 ± 0.0370.914 ± 0.0190.854 ± 0.0300.586 ± 0.06014:36
40.816 ± 0.0050.905 ± 0.0040.816 ± 0.0050.927 ± 0.0030.836 ± 0.0040.629 ± 0.0048:19
Table 6. Results of Basic CNN M3 with DWT.
Table 6. Results of Basic CNN M3 with DWT.
Data TypeWindow (s)AccuracyPrecisionSensitivitySpecificity F β ScoreMCCRuntime (min)
A C
(3)
20.939 ± 0.0020.948 ± 0.0040.939 ± 0.0020.947 ± 0.0090.942 ± 0.0020.785 ± 0.01367.44
30.922 ± 0.0310.940 ± 0.0190.922 ± 0.0310.949 ± 0.0170.926 ± 0.0280.796 ± 0.06822:39
40.906 ± 0.0220.923 ± 0.0180.906 ± 0.0220.941 ± 0.0170.910 ± 0.0210.766 ± 0.05013:48
L C
(3)
20.923 ± 0.0050.941 ± 0.0010.923 ± 0.0050.944 ± 0.0060.928 ± 0.0040.748 ± 0.003127:32
30.940 ± 0.0100.951 ± 0.0060.940 ± 0.0100.963 ± 0.0030.943 ± 0.0090.834 ± 0.02026:43
40.930 ± 0.0200.940 ± 0.0150.930 ± 0.0200.958 ± 0.0140.932 ± 0.0190.817 ± 0.04416:47
T C
(3)
20.946 ± 0.0150.954 ± 0.0100.946 ± 0.0150.952 ± 0.0100.949 ± 0.0130.811 ± 0.04255:17
30.938 ± 0.0190.949 ± 0.0120.938 ± 0.0190.952 ± 0.0080.941 ± 0.0170.784 ± 0.04822:12
40.943 ± 0.0080.953 ± 0.0050.943 ± 0.0080.971 ± 0.0030.945 ± 0.0080.839 ± 0.01912:39
Table 7. Results of Basic CNN M4 with PWVD.
Table 7. Results of Basic CNN M4 with PWVD.
Data TypeWindow (s)AccuracyPrecisionSensitivitySpecificity F β ScoreMCCRuntime (min)
A C
(3)
20.831 ± 0.0230.906 ± 0.0060.831 ± 0.0230.902 ± 0.0110.852 ± 0.0180.571 ± 0.03574.00
30.865 ± 0.0150.907 ± 0.0090.865 ± 0.0150.930 ± 0.0100.876 ± 0.0140.681 ± 0.03132:34
40.825 ± 0.0240.888 ± 0.0100.825 ± 0.0240.919 ± 0.0100.839 ± 0.0210.641 ± 0.03517:46
L C
(3)
20.811 ± 0.0110.901 ± 0.0050.811 ± 0.0110.897 ± 0.0090.836 ± 0.0090.545 ± 0.021131:00
30.871 ± 0.0200.912 ± 0.0050.871 ± 0.0200.930 ± 0.0050.881 ± 0.0160.695 ± 0.02628:58
40.831 ± 0.0250.895 ± 0.0100.831 ± 0.0250.923 ± 0.0130.846 ± 0.0210.657 ± 0.03216:27
T C
(3)
20.805 ± 0.0110.900 ± 0.0100.805 ± 0.0110.887 ± 0.0170.832 ± 0.0100.531 ± 0.03668:28
30.842 ± 0.0330.908 ± 0.0090.842 ± 0.0330.917 ± 0.0130.861 ± 0.0260.590 ± 0.05229:24
40.864 ± 0.0250.916 ± 0.0100.864 ± 0.0250.946 ± 0.0120.877 ± 0.0210.687 ± 0.04015:50
Table 8. Results of Bidirectional LSTM M5 with Raw Signals.
Table 8. Results of Bidirectional LSTM M5 with Raw Signals.
Data TypeWindow (s)AccuracyPrecisionSensitivitySpecificity F β ScoreMCCRuntime (min)
A C
(3)
20.797 ± 0.1060.896 ± 0.0270.797 ± 0.1060.917 ± 0.0480.827 ± 0.0820.527 ± 0.149302.21
30.784 ± 0.1240.903 ± 0.0180.784 ± 0.1240.939 ± 0.0530.817 ± 0.0920.597 ± 0.152149:31
40.695 ± 0.1780.832 ± 0.0530.695 ± 0.1780.774 ± 0.0740.715 ± 0.1550.441 ± 0.16873:22
L C
(3)
20.705 ± 0.1060.877 ± 0.0120.705 ± 0.1060.836 ± 0.0340.747 ± 0.0880.418 ± 0.091121:56
30.894 ± 0.0210.929 ± 0.0060.894 ± 0.0210.933 ± 0.0070.904 ± 0.0160.687 ± 0.03996.09
40.360 ± 0.3800.315 ± 0.4320.360 ± 0.3800.806 ± 0.1340.312 ± 0.4190.246 ± 0.35152:05
T C
(3)
20.715 ± 0.2700.880 ± 0.0710.715 ± 0.2700.800 ± 0.1200.744 ± 0.2390.478 ± 0.314199:10
30.778 ± 0.1420.903 ± 0.0320.778 ± 0.1420.930 ± 0.0600.814 ± 0.1110.531 ± 0.181105:22
40.441 ± 0.3120.517 ± 0.3720.441 ± 0.3120.728 ± 0.0590.411 ± 0.3050.111 ± 0.15774:20
Table 9. Results of Bidirectional LSTM M6 with Extracted features.
Table 9. Results of Bidirectional LSTM M6 with Extracted features.
Data TypeWindow (s)AccuracyPrecisionSensitivitySpecificity F β ScoreMCCRuntime (min)
A C
(3)
20.846 ± 0.0340.912 ± 0.0150.846 ± 0.0340.902 ± 0.0230.865 ± 0.0280.600 ± 0.067179.36
30.815 ± 0.0200.896 ± 0.0070.815 ± 0.0200.923 ± 0.0090.834 ± 0.0170.620 ± 0.02691:14
40.803 ± 0.0330.875 ± 0.0100.803 ± 0.0330.893 ± 0.0090.821 ± 0.0280.597 ± 0.04049:37
L C
(3)
20.822 ± 0.0130.907 ± 0.0080.822 ± 0.0130.904 ± 0.0120.845 ± 0.0100.563 ± 0.025147:41
30.812 ± 0.0110.897 ± 0.0010.812 ± 0.0110.917 ± 0.0020.832 ± 0.0080.620 ± 0.01189:48
40.783 ± 0.0700.859 ± 0.0420.783 ± 0.0700.855 ± 0.0430.801 ± 0.0620.557 ± 0.12247:42
T C
(3)
20.840 ± 0.0070.907 ± 0.0140.840 ± 0.0070.898 ± 0.0250.859 ± 0.0060.579 ± 0.039190:06
30.773 ± 0.0220.900 ± 0.0010.773 ± 0.0220.897 ± 0.0040.807 ± 0.0170.519 ± 0.01961:44
40.770 ± 0.0190.888 ± 0.0130.770 ± 0.0190.909 ± 0.0190.797 ± 0.0170.560 ± 0.03540:11
Table 10. Results of Stacked Ensemble M7 with Extracted features.
Table 10. Results of Stacked Ensemble M7 with Extracted features.
Data TypeWindow (s)AccuracyPrecisionSensitivitySpecificity F β ScoreMCCRuntime (min)
A C
(3)
20.971 ± 0.0070.972 ± 0.0070.971 ± 0.0070.956 ± 0.0120.971 ± 0.0070.885 ± 0.027200.32
30.979 ± 0.0020.979 ± 0.0020.979 ± 0.0020.977 ± 0.0050.979 ± 0.0020.934 ± 0.006157:30
40.967 ± 0.0050.967 ± 0.0070.967 ± 0.0050.967 ± 0.0120.967 ± 0.0060.905 ± 0.018108:52
L C
(3)
20.967 ± 0.0080.968 ± 0.0080.967 ± 0.0080.954 ± 0.0130.967 ± 0.0080.870 ± 0.032200:29
30.980 ± 0.0020.980 ± 0.0020.980 ± 0.0020.977 ± 0.0050.980 ± 0.0020.938 ± 0.006132:45
40.965 ± 0.0110.965 ± 0.0110.965 ± 0.0110.968 ± 0.0140.965 ± 0.0110.899 ± 0.032118:05
T C
(3)
20.972 ± 0.0090.973 ± 0.0090.972 ± 0.0090.956 ± 0.0130.972 ± 0.0090.889 ± 0.036324:54
30.971 ± 0.0060.971 ± 0.0060.971 ± 0.0060.960 ± 0.0080.971 ± 0.0060.882 ± 0.024196:53
40.967 ± 0.0090.971 ± 0.0070.967 ± 0.0090.979 ± 0.0030.968 ± 0.0090.900 ± 0.026178:51
Table 11. Results of Average Ensemble M8 with Extracted features.
Table 11. Results of Average Ensemble M8 with Extracted features.
Data TypeWindow (s)AccuracyPrecisionSensitivitySpecificity F β ScoreMCCRuntime (min)
A C
(3)
20.979 ± 0.0080.978 ± 0.0090.979 ± 0.0080.958 ± 0.0130.978 ± 0.0090.913 ± 0.034103.09
30.980 ± 0.0050.980 ± 0.0050.980 ± 0.0050.976 ± 0.0070.980 ± 0.0050.938 ± 0.01543:36
40.967 ± 0.0050.967 ± 0.0060.967 ± 0.0050.967 ± 0.0120.967 ± 0.0060.905 ± 0.01832:09
L C
(3)
20.973 ± 0.0080.973 ± 0.0080.973 ± 0.0080.956 ± 0.0130.973 ± 0.0080.893 ± 0.032107:14
30.978 ± 0.0030.978 ± 0.0020.978 ± 0.0030.976 ± 0.0030.978 ± 0.0030.931 ± 0.00851:31
40.969 ± 0.0080.970 ± 0.0080.969 ± 0.0080.969 ± 0.0130.969 ± 0.0080.911 ± 0.02424:04
T C
(3)
20.983 ± 0.0060.983 ± 0.0060.983 ± 0.0060.960 ± 0.0120.983 ± 0.0060.932 ± 0.02664:56
30.975 ± 0.0050.975 ± 0.0050.975 ± 0.0050.962 ± 0.0090.975 ± 0.0050.900 ± 0.01932:55
40.976 ± 0.0080.978 ± 0.0070.976 ± 0.0080.983 ± 0.0030.976 ± 0.0080.925 ± 0.02438:40
Table 12. Results of Majority Voting M9 with Extracted features.
Table 12. Results of Majority Voting M9 with Extracted features.
Data TypeWindow (s)AccuracyPrecisionSensitivitySpecificity F β ScoreMCCRuntime (min)
A C
(3)
20.981 ± 0.0070.980 ± 0.0070.981 ± 0.0070.951 ± 0.0150.980 ± 0.0070.921 ± 0.029< 1
30.985 ± 0.0030.985 ± 0.0030.985 ± 0.0030.979 ± 0.0060.985 ± 0.0030.953 ± 0.010< 1
40.969 ± 0.0060.969 ± 0.0070.969 ± 0.0060.967 ± 0.0120.969 ± 0.0070.911 ± 0.019< 1
L C
(3)
20.977 ± 0.0080.977 ± 0.0080.977 ± 0.0080.958 ± 0.0120.977 ± 0.0080.907 ± 0.032< 1
30.973 ± 0.0080.975 ± 0.0060.973 ± 0.0080.974 ± 0.0020.973 ± 0.0070.917 ± 0.020< 1
40.971 ± 0.0080.972 ± 0.0080.971 ± 0.0080.967 ± 0.0080.971 ± 0.0080.917 ± 0.023< 1
T C
(3)
20.983 ± 0.0070.983 ± 0.0070.983 ± 0.0070.960 ± 0.0120.983 ± 0.0070.932 ± 0.030< 1
30.977 ± 0.0030.977 ± 0.0040.977 ± 0.0030.962 ± 0.0080.976 ± 0.0040.905 ± 0.015< 1
40.976 ± 0.0110.978 ± 0.0090.976 ± 0.0110.979 ± 0.0040.976 ± 0.0100.925 ± 0.032< 1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ashfaque Mostafa, T.; Soltaninejad, S.; McIsaac, T.L.; Cheng, I. A Comparative Study of Time Frequency Representation Techniques for Freeze of Gait Detection and Prediction. Sensors 2021, 21, 6446. https://doi.org/10.3390/s21196446

AMA Style

Ashfaque Mostafa T, Soltaninejad S, McIsaac TL, Cheng I. A Comparative Study of Time Frequency Representation Techniques for Freeze of Gait Detection and Prediction. Sensors. 2021; 21(19):6446. https://doi.org/10.3390/s21196446

Chicago/Turabian Style

Ashfaque Mostafa, Tahjid, Sara Soltaninejad, Tara L. McIsaac, and Irene Cheng. 2021. "A Comparative Study of Time Frequency Representation Techniques for Freeze of Gait Detection and Prediction" Sensors 21, no. 19: 6446. https://doi.org/10.3390/s21196446

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop