Next Article in Journal
Possible Mechanisms of Hypnosis from an Interactional Perspective
Next Article in Special Issue
Change in Blood Flow Velocity Pulse Waveform during Plateau Waves of Intracranial Pressure
Previous Article in Journal
Optimal Woven EndoBridge (WEB) Device Size Selection Using Automated Volumetric Software
Previous Article in Special Issue
Classification of Prefrontal Cortex Activity Based on Functional Near-Infrared Spectroscopy Data upon Olfactory Stimulation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identification of Neurodegenerative Diseases Based on Vertical Ground Reaction Force Classification Using Time–Frequency Spectrogram and Deep Learning Neural Network Features

1
Department of Biomedical Engineering, College of Engineering, National Cheng Kung University, Tainan 701, Taiwan
2
Medical Device Innovation Center, National Cheng Kung University, Tainan 701, Taiwan
*
Author to whom correspondence should be addressed.
Brain Sci. 2021, 11(7), 902; https://doi.org/10.3390/brainsci11070902
Submission received: 16 June 2021 / Revised: 1 July 2021 / Accepted: 5 July 2021 / Published: 8 July 2021
(This article belongs to the Special Issue Neuroinformatics and Signal Processing)

Abstract

:
A novel identification algorithm using a deep learning approach was developed in this study to classify neurodegenerative diseases (NDDs) based on the vertical ground reaction force (vGRF) signal. The irregularity of NDD vGRF signals caused by gait abnormalities can indicate different force pattern variations compared to a healthy control (HC). The main purpose of this research is to help physicians in the early detection of NDDs, efficient treatment planning, and monitoring of disease progression. The detection algorithm comprises a preprocessing process, a feature transformation process, and a classification process. In the preprocessing process, the five-minute vertical ground reaction force signal was divided into 10, 30, and 60 s successive time windows. In the feature transformation process, the time–domain vGRF signal was modified into a time–frequency spectrogram using a continuous wavelet transform (CWT). Then, feature enhancement with principal component analysis (PCA) was utilized. Finally, a convolutional neural network, as a deep learning classifier, was employed in the classification process of the proposed detection algorithm and evaluated using leave-one-out cross-validation (LOOCV) and k-fold cross-validation (k-fold CV, k = 5). The proposed detection algorithm can effectively differentiate gait patterns based on a time–frequency spectrogram of a vGRF signal between HC subjects and patients with neurodegenerative diseases.

1. Introduction

Amyotrophic lateral sclerosis (ALS), Huntington’s disease (HD), and Parkinson’s disease (PD), as NDDs, are defined as diseases caused by the progressive death of neurons in different regions of the nervous system, through the loss of structure and function of neurons [1]. For example, PD is the second most prevalent NDD, with a prevalence of 0.3% in the general population, ~1% in the elderly over 60 years old, and ~3% in those aged 80 years old or more [2]. PD incidence rate ranges between 8 and 18 people out of 100,000 per year [2]. The median age at onset is 60 years, and the average time it takes for the disease to progress, from the diagnosis to death, is approximately 15 years [2]. Men show a 1.5–2 times greater prevalence of this disease and incidence compared to women [2]. In terms of the medication required for treatment, PD costs USD 2500 each year, and therapeutic surgery costs up to USD 100,000 per patient [3]. ALS is the third most prevalent NDD and the most common motor neuron disease, with an estimated annual incidence of 1.9 people out of 100,000 per year [4,5]. In the United States, 30,000 people have ALS, 30,000 have HD, and 1 million have PD [6]. As NDDs mainly affect people in their middle to late years of life, the incidence is expected to increase with the an increasingly aging population. In 2030, 1 out of every 5 Americans will be over the age of 65, and 30 years from now, more than 12 million Americans will be affected by NDDs [7]. The development of early detection, treatments, and cures for NDDs is an ultimate goal of increasing urgency. NDDs can affect a variety of bodily functions, including heart rate, respiration, speech, mental function, balance, and movement. As the central nervous system, particularly the basal ganglia, controls the general motion (flexion and extension) of lower limbs, the gait of a patient with an NDD will become abnormal (different gait pattern than a healthy subject) due to motor neuron decline [8]. ALS, also known as motor neuron disease (MND), is characterized by stiff muscles, muscle twitching, and steadily deteriorating weakness as muscles decrease in size [9,10,11]. HD is a hereditary disorder that causes the death of brain cells, resulting in a lack of coordination, a shaky gait, and uncoordinated and jerky body movements as the disease progresses [12,13,14]. PD is a long-term degenerative disorder of the central nervous system that primarily affects the motor system. Early symptoms include trembling, rigidity, slowness of movement, and difficulty walking [15,16,17]. It is reasonable to assume that these NDDs have an impact on foot force as a result of this phenomenon. For movement analyses in HC subjects and other subjects with various diseases, gait data have been developed. This type of approach is very useful for understanding the movement disorder in NDDs and has a large amount of potential in terms of presenting non-invasive automatic NDD classification methods.
Gait analysis research has been developed in the last decade, particularly using time series of stride, stance or swing intervals, ground reaction force (GRF), and foot force. Previous research has shown that feature extraction methods and machine learning can be used to classify gait features. Xia et al. suggested a method for classifying gait rhythm signals in patients with NDDs and healthy people [18]. They tested various classification models and statistical characteristics, such as a support vector machine (SVM), random forest (RF), multi-layer perceptron (MLP), and k-nearest neighbor (KNN). To perform feature extraction of PD subjects based on sensor signals, Ertugrul et al. developed shifted one-dimensional local binary patterns and used Bayes network (BayesNT), naive Bayes (NB), logistic regression (LR), partial C4.5 decision tree (PART), a rule learner approach (Jrip), functional tree (FT), and other classifiers. [19]. Using entropy parameters, Wu et al. measured signal fluctuations in gait rhythm time series of PD patients [20]. Their research aimed to calculate the approximate entropy (ApEn), normalized symbolic entropy (NSE), and signal turns count (STC) parameters for measuring stride fluctuations in PD. Nonlinear gait pattern classifications were performed using generalized linear regression analysis (GLRA) and support vector machines (SVM). Bilgin investigated the effect of feature extraction on ALS patient classification in NDD and HC subjects [21]. The input signal, compound foot (CF) force, was decoded for feature extraction using a 6-level discrete wavelet transform (DWT) with several wavelet methods. In linear discriminant analysis (LDA) and naive Bayesian classifier (NBC), the derived features were validated using 20 trials for 5-fold cross-validation.
Deep learning has demonstrated excellent performance in gait classification problems in recent years. Zeng and Wang, for example, introduced a technique based on gait dynamics to classify (diagnose) NDDs using deterministic learning theory and a recurrent neural network (RNN) [22]. Zhao et al. used dual-channel long short-term memory (LSTM)-based multi-feature extraction on gait for NDD diagnosis [23]. They developed a dual-channel LSTM model to combine gait time series and force series recorded from NDD patients in order to understand the whole gait. According to several electrocardiogram (ECG) classification studies [24,25,26,27], combining a time–frequency representation (spectrogram) with a deep neural network can improve performance in extracting the distribution of important features more easily and automatically learn complex representation features directly from data. The goal of feature enrichment in a spectrogram of a time series signal is to enrich the simplified representation by restoring topological information, neighborhood information, and association information with details [28]. A deep neural network, on the other hand, has excellent feature extraction capabilities and can automatically extract “deep features” [29], greatly improving classification accuracy.
The specific aim of this study was to observe the effectiveness of the utilization of several feature transformations from a 1D vGRF signal into a 2D time–frequency spectrogram and the combination of principal component analysis (PCA) with a deep learning network for extracting features for the classification of NDD patients. The technological impact of this paper is that it is the first to apply spectrogram- and deep learning-based networks to the gait classification problem and yield high classification accuracy. The emphasis of this paper is on gaining insight into the effectiveness of the left foot (LF), right foot (RF), and compound the foot (CF) force signals in the classification of NDDs. It warrants investigation into whether the three types of degenerative nerve diseases (ALS, HD, and PD) interfere with a patient’s ability to handle two-foot propulsion and if the major difference in vGRF is related to the type of disease the patient has.
Raw vGRF signal data from NDD and HC subjects were obtained as the system’s input using force-sensitive resistors, with the output approximately proportional to the force under the foot [30]. Continuous wavelet transform (CWT), short-time Fourier transform (STFT), and wavelet synchrosqueezed transform (WSST) feature transformations were applied to the input in order to create new features (time–frequency spectrogram) from existing ones. Then, to increase classification performance, principal component analysis (PCA) was applied to the time–frequency spectrogram by selecting the features’ principal components (PCs). Training and testing sets were created for the PCs of HC and NDD subjects. Several classification parameters were created by training the estimators on the training sets and comparing them to a test set of the HC or NDD to be categorized. In this study, a convolutional neural network (CNN) was successfully used to classify the HC and NDD in the classification stage (training and testing phase). The proposed method can effectively distinguish between HC and NDD gait patterns.

2. Materials and Methods

By transforming one-dimensional signals into two-dimensional pattern objects (images) using the feature transformation technique from a continuous wavelet transform (CWT), the proposed NDD detection algorithm attempted to extract pattern characteristics and visualization from vGRF signals in ALS, HD, PD, and HC subjects. The proposed NDD detection algorithm consists of four main steps, as shown in Figure 1: (1) signal preprocessing of NDD and the HC vGRF signal; (2) feature extraction by generating the spectrogram of the vGRF signal using CWT and PCA; (3) construction of the classifier model by feature training using Pretrained AlexNet CNN; and (4) the use of cross-validation techniques to test and analyze the effectiveness of the detection algorithm based on the classifier model.

2.1. Neuro-Degenerative Diseases Gait Dynamics Database

Hausdorff et al. presented the vGRF database used in this study (called the Gait Dynamics in Neuro-Degenerative Disease Database) online in the PhysioNet database [31]. This database’s raw signal data were obtained by using force-sensitive resistors with an output proportional to the force under the foot. When loaded, the transducer was a conductive polymer layer sensor with a changed resistance. The sensor was chosen due to its 0.05-inch thickness, temperature insensitivity, rapid dynamic response, ability to restrain an overload, and electronically simple interface. Two 1.5-in2 force-sensitive resistors were used, and the sensors were taped to an insole that was used to position them inside the shoe. The insole was made by tracing an outline of the foot onto the manila folder and then cutting out the tracing. One sensor was placed near the toes and metatarsals in the anterior part of the insole, and the other was placed near the heel on the opposite end. The two footswitches were connected in parallel and functioned as a single large sensor (the output from these two footswitches were added up). To increase the signal saturation, a 390 resistor, R1, was placed in series with this parallel connection as a voltage divider. A 5-V battery-operated circuit powered the sensors. The divider’s output voltage was fed into a voltage follower, of which the voltage output nonlinearly increased as the force was increased. The switch’s output voltage ranges from 0 V with no load to 3.5 V with a full load (closed). The analog signal was then converted into digital format and analyzed with software [30].
There are 64 recordings of information from 13 patients with ALS, 20 patients with HD, 15 patients with PD, and 16 healthy controls in the database. This database contains two types of data: raw force series data and derived time series from the raw data. The force series comprises LF force and RF force signal. Left stride interval (s), right stride interval (s), left swing interval (s), right swing interval (s), left swing interval (% of stride), right swing interval (% of stride), left stance interval (s), right stance interval (s), left stance interval (% of stride), right stance interval (% of stride), double support interval (s), and double support interval (% of stride) are contained within the time-series data.

2.2. Signal Preprocessing

During the data collection, a 5-min vGRF signal was obtained. The proposed technique took three types of vGRF signals as input: LF, RF, and CF (CF = LF + RF). Due to the length of the foot force signal, it was difficult to interpret the data even after using a CWT to transform the features. The window function, a mathematical term that is zero-valued outside of a specified interval, was used to visualize the foot force signal clearly. The time windows used in this study were 10, 30, and 60 s. The time windowing determination was helpful in obtaining more data to feed into the deep learning model and simulating more precise and fast disease predictions [32].

2.3. Feature Transformation

2.3.1. Continuous Wavelet Transform (CWT)

A continuous wavelet transform (CWT) is a signal processing technique for observing nonstationary signals’ time-varying frequency spectrum characteristics [33]. The CWT result is a time–frequency spectrogram (time–scale representation), which provides useful information on the relationship between time and frequency.
The CWT of a time series function x ( t ) L 2 ( ) with a scaling factor s +   ( s > 0 ) that controls the wavelet’s width and a translation parameter τ controls the wavelet’s location can be expressed by the following equation:
X w ( s , τ ) = 1 s x ( t ) ψ * ( t τ s )   d t
where ψ(t) is a mother wavelet, also called a window function. A Morlet or Gabor wavelet was used as the mother wavelet function in this study. This wavelet function is made up of a complex sinusoid with a Gaussian window (a complex exponential multiplied by a Gaussian window) that is specified by the following term:
ψ ω 0 ( t ) = ( e i f t e 1 2 f 2 ) e 1 2 t 2
Parameter t refers to the time and f represents the reference frequency.
The vGRF signal is represented as a time–frequency spectrogram image by the time–frequency transformation applied to the system. The image clearly shows distinct vGRF patterns for HC and NDD subjects that are not visible in the signal’s time and frequency domains. Variations in the foot pressure signal caused by temporal characteristics can also be studied using the time–frequency spectrogram. The measurement of step length, stance width, the length of the step rhythm, and step velocity are all examples of temporal characteristics, which are also known as spatial characteristics or linear gait variabilities. The CWT feature transformation results for the NDD and HC groups are shown in Figure 2 and Figure 3.

2.3.2. Short Time Fourier Transform

The short-time Fourier transform (STFT) is a series of Fourier-related transforms applied to a windowed signal to determine the sinusoidal frequency and phase content of local parts as the signal transforms over time [34]. STFT is calculated by dividing a longer time signal into shorter segments of equal lengths and then computing the Fourier transform on each shorter segment independently.
The STFT pair is given as follows:
{ X S T F T [ m , n ] = k = 0 L 1 x [ k ] g [ k m ] e j 2 π n k / L x [ k ] = m n X S T F T [ m , n ] g [ k m ] e j 2 π n k / L
where x[k] represents a signal and g[k] represents an L-point window function. The STFT of x[k] can be construed as the Fourier transform of the product x[k]g[km].

2.3.3. Wavelet Synchrosqueezed Transform (WSST)

The wavelet synchrosqueezed transform is a time–frequency analysis technique for studying multi-component signals with oscillating modes (speech waveforms, machine vibrations, and physiologic signals), with the goal of sharpening a time–frequency analysis by reallocating the signal energy in frequency [35]. The synchrosqueezing algorithm uses CWT of the input signal to generate the instantaneous frequency information. The instantaneous frequencies from the CWT output, W f , are extracted using a phase transform, ω f . This phase transform is proportionate to the first derivative of the CWT with respect to the translation, u .
ω f ( s , u ) = t W f ( s , u ) 2 π i W f ( s , u )
s are the scales, defined as s = f x f , where f x is the peak frequency and f is the frequency. Finally, “squeeze” the CWT over regions where the phase transformation is constant. The resulting instantaneous frequency value is redefined to a single value at the centroid of the CWT time–frequency region.

2.4. Principal Component Analysis (PCA) for Feature Enhancement

The main idea behind a principal component analysis (PCA) is to reduce the dimension of a dataset with a large number of interrelated variables while minimizing the amount of variance in the dataset [36,37,38]. Specifically, PCA is able to minimize input data redundancy, remove potential association, and extract the most important feature vectors for data changing directions. This is accomplished by converting the dataset into a new set of variables known as principal components (PCs), which contain decorrelated and ordered variables. The PCA technique was mathematically characterized in this study using the steps below (as shown in Figure 4).
The aim of using PCA as a feature enhancement in this study was to improve between-class separability while reducing within-class separability [32]. Its goal was to increase the performance of deep learning in extracting features and artificial intelligence in classifying data points into the correct groups. In a deep learning network, such as CNN, the gradient diffusion problem occurs [39,40], and many of the filters in the layer are highly correlated thus making it possible to detect the same feature [41], and making insignificant contributions to the classification accuracy performance. To alleviate these problems by initializing the weights of convolution kernels, PCA is employed to the unsupervised extraction of input image eigenvectors [39,41,42]. PCA can also improve the classification performance (accuracy, sensitivity, specificity, and AUC value; see Section 3 Experimental Results).

2.5. Pre-Trained Convolutional Neural Network (CNN) as Feature Extractor

As in a simple multilayer neural network (deep learning), a convolutional neural network (CNN) is made up of one or more convolutional layers (often with subsampling and pooling layers) followed by one or more fully connected layers [43]. The architecture of a CNN is designed to take advantage of the input’s 2D structure (image or signal). This is achieved using local connections and weights, which are then followed by any pooling function that produces translation-invariant features. A CNN also has the advantage of being easier to train and having fewer parameters than other fully connected networks with the same number of hidden layers. The use of a CNN in the proposed method is primarily to differentiate between the time–frequency spectrogram representation of vGRF from HC and NDD (ALS, HD, and PD) subjects.
The proposed method used a pre-trained AlexNet CNN from the MATLAB R2018a Deep Learning Toolbox™ (The MathWorks, Inc., Natick, MA, USA). There are 25 layers in the architecture, including an input layer, five convolution 2D layers, seven ReLU (activation function) layers, two cross-channel normalization layers, three max pooling 2D layers, three fully connected layers, two dropout layers (for regularization), a softmax layer (normalized exponential function), and an output layer. The time–frequency spectrogram figure of the vGRF signal yielded by the CWT is fed into the AlexNet CNN in the proposed procedure. By using the layer activations as features, the pre-trained AlexNet CNN was proposed as a feature extractor [32,44,45,46]. This is a simple, time-efficient strategy to use pre-trained networks that avoids the effort needed to train a full network. By employing this simple and fast methodology, the possibility of wearable device integration with the algorithm becomes more promising. The suggested technique used a support vector machine (SVM) for classification and used the second fully connected layer as the feature extractor [47,48] (the CNN architecture utilized in this study is described in Table 1). AlexNet CNN has been trained with numerous common images, such as cars, boats, planes, dogs, and cats, but it is also possible for the CNN to utilize the distinct properties of non-image data (1D signal)−computationally efficient and locally focused—by converting non-image data into an image, such as a binary image [49], spectrogram [29,50], recurrence plot [32], or Gramian Angular Summation Field (GASF) image [51].

2.6. Support Vector Machine (SVM) as Classifier

In this study, the NDD patients and HC subjects were automatically distinguished using a support vector machine (SVM) after being processed based on feature transformation and extraction. The aim of the SVM is to construct a hyperplane or set of hyperplanes in a high- or infinite-dimensional space, which can be used for classification, regression, or other tasks, such as outlier detection [55]. Specifically, the purpose of using SVM is to discover an optimal decision surface that splits the dataset into correct classes and has a maximum distance or margin among the classes.

2.7. Cross-Validation

Cross-validation is a statistical method for evaluating and comparing learning algorithms that divide data into two groups: one for learning or training a model (training set) and another for validating the model (testing or validation set) [56,57,58]. In order for each data point to be confirmed, the training and testing sets must cross over in consecutive rounds. There are two primary reasons to use cross-validation: first, one algorithm can be used to investigate the performance of the learned model from available data. Specifically, it is used to assess an algorithm’s generalizability. The second goal is to compare the performance of two or more different algorithms and determine which is most appropriate for the data or to compare the performance of two or more variants of the parameterized model. Leave-one-out cross-validation (LOOCV) and k-fold cross-validation (k-fold CV, k = 5) were the two cross-validation methods used in this study.

3. Experimental Results

The experiments were run on an NVIDIA GeForce GTX 1060 6 GB computer with an Intel® Core™ i5-8400 CPU @ 2.80 GHz, 2808 MHz, and 24 GB RAM, using MATLAB software (R2018a, The MathWorks, Inc., MA, USA). The number of time–frequency spectrogram images input (related to the time windowing process, where smaller time windowing results in more images and computation time becomes longer) and the number of neurons in the CNN corresponded to the calculation of the computation time (see Table 2). The proposed method’s accuracy, sensitivity, specificity, and ROC area under the curve (AUC) value were included as evaluation parameters, as specified in [59]. The learning curve contains the training loss function of the machine learning classifier, an SVM, as the result of feature extracted from fully connected layer of the pre-trained AlexNet CNN (see Figure 5).
When deciding between two or more diagnostic tests, Youden’s index is commonly used to assess the overall diagnostic test’s efficacy [60]. Youden’s index is a function of sensitivity and specificity that ranges from 0 to 1, with a value close to 1 indicating high diagnostic test effectiveness and that the test is perfect, and a value close to 0 indicating limited diagnostic test effectiveness and that the test is useless. The Youden’s index (J) is described as the sum of the two fractions representing the measurements properly diagnosed for the diseased (sensitivity) and HC (specificity) groups, overall cut-points c , < c < :
J = max c { sensitivity ( c ) + specificity ( c ) 1 }

3.1. Classification of the NDD and HC Group

In this classification scenario, there were three types of classification tasks: ALS versus HC, HD versus HC, and PD versus HC. In all classification scenarios, 13 ALS patients, 20 HD patients, 15 PD patients, and 16 HC subjects were used and observed, but the feedback signal for the proposed procedure was dependent on the time window in the time-windowing process and the frequency selection. There were 480 HC, 390 ALS, 600 HD, and 450 PD input signal numbers for the 10-s time window. The HC, ALS, HD, and PD input signal numbers were 160, 130, 200, and150, respectively, in the 30-s time window. There were 80, 65, 100, and 75 HC, ALS, HD, and PD input signal numbers in the 60-s time window, respectively. The detailed classification results are given in Table 3 and Table 4.

3.2. Classification among the NDD

In this study, a classification concepts were developed among the NDD, such as ALS vs. HD, PD vs. ALS, and HD vs. PD. The primary goal of this classification was to determine whether ALS, HD, and PD could be easily separated (the NDD group: ALS, HD, and PD). The conclusion was that the ALS group could easily be distinguished from the HD and PD groups, but that HD and PD were difficult to differentiate. In contrast to ALS vs. HD and PD vs. ALS, the HD vs. PD classification results were lower. This occurred due to the fact that both HD and PD are caused by basal ganglia degeneration, and the gait patterns of HD and PD patients are nearly identical [61]. The complete classification results are shown in Table 5 and Table 6.

3.3. Classification of All NDD in One Group with HC Group

The vGRF datasets of ALS, HD, and PD patients were merged into one group for NDD vs. HC classification, with the total number of NDD datasets varying depending on the time window. The experimental results for this classification situation are shown in Table 5 and Table 6.

3.4. Multi-Class Classification

As the physician may not know whether the patient is suffering from ALS, HD, or PD, the multi-class classification is closer to the clinical application. The entire vGRF dataset was divided into four categories based on disease patients (ALS, HD, and PD) and healthy subjects. For assessment and validation purposes, LOOCV and k-fold CV (k = 5) were also used in the multi-class classification. The detailed classification results are given in Table 7, Table 8 and Table 9.

4. Discussion

This section discusses the gait analysis of each NDD using the time and frequency analysis of the time–frequency spectrogram. Certain key features of a signal are difficult to notice with the naked eye, but time–frequency spectrogram analysis may aid in the discovery of significant time and frequency characteristics. The time–domain signal was transformed into the time–frequency domain using CWT in this research. Pattern visualization and recognition of the time–frequency spectrogram could easily be used to understand the NDD and HC gait phenomena.
This observation was limited to the CF vGRF signal. As this type of input signal is the extra force between the LF and RF force signals, it defines the correlations between the LF and RF features rather than each individual feature. As the input signal was shorter and the gait phenomenon could be studied in greater detail, a time window of 10 s was chosen. Based on the normal frequency of leg movements [62] and in order to obtain a high level of visualization, the frequency ranges of 0.1–5 Hz and 5–50 Hz were selected to observe the CWT time–frequency spectrogram in detail.

4.1. Healthy Control

The normal gait phenomenon was interpreted by observing the time–frequency spectrogram of the HC subject shown in Figure 2 (left). At the 0.1–5 Hz spectrogram, the strongest walking force magnitude (yellow) of the normal gait occurred at 1.6–2.1 Hz and was stable from the initial time until the end. This means that the foot force distribution and walking velocity of normal subjects are the same when they are walking. It was also shown that at 3 Hz and around 4.5–5 Hz, small areas, signifying the lowest force magnitude (dark blue), appeared alternately with a significant force magnitude (light blue) forming a regular pattern. This phenomenon appeared in the spectrogram caused by the CF force signal at the lowest magnitudes. There are three lowest magnitudes that can be observed in one cycle of the CF force time–domain signal (see Figure 2 (left) vGRF signal); each of these lowest magnitudes has an almost equal magnitude in every cycle of the signal. The lowest magnitudes (global minimum) that occurred at the beginning and end of the half gait cycle (only LF or RF gait cycle), close to the 0 force unit, show the toe-off and initial contact and the lowest magnitude (local minimum) that occurred in the half gait cycle exhibited when only one foot was in contact with the ground.
At the 5–50 Hz frequency range, there was also a steady, strong force level (yellow) of around 5 Hz, the same magnitude as that occurring during walking, from the initial to the end, and a significant force magnitude (light blue) still occurred up to 50 Hz and was also constant every time. Both time–frequency spectrograms indicate that the time and frequency components in the spectrogram comprise a regular pattern. This interpretation became a benchmark for the investigation into the NDD gait phenomenon. It was compared to analyze and discover the gait characteristics of NDDs based on the spectrogram.

4.2. Amyotrophic Lateral Sclerosis

For the ALS syndrome, as shown in Figure 2 (right), the most intense walking force attenuation of these patients in the 0.1–5 Hz spectrogram occurred at approximately 0.6–0.9 Hz and 1.1–1.5 Hz, which was lower than the frequency of the HC. This means that ALS patients walk in a more delayed fashion than the HC. The CF force time–domain signal shows that the lowest force magnitudes were not equal in every cycle of the CF force signal (clearly depicted in Figure 2 (right) vGRF signal); even at a specific time, the global minimum magnitudes were almost the same as the local minimum magnitudes and were not near the 0 force unit. In addition to these tendencies, there were more local minimum magnitudes along with the ALS time–domain signal. This phenomenon affects the regularity of the lowest force power pattern that typically occurs at 3 and 5 Hz. There were three frequency bands that showed the lowest force magnitude (dark blue), which appeared alternately with the significant force magnitude (light blue) forming an irregular pattern: at approximately 2–2.5 Hz, 3.5–4 Hz, and 5 Hz.
ALS patients had an unstable force magnitude (yellow) at 5 Hz, and at the 5–50 Hz frequency range, the instability only occurs at a specific time, at 6 and 7 s, and did not occur during the entire walking time. The significant force magnitude (light blue) was different every time and only reached 45 Hz.

4.3. Huntington’s Disease

Among the symptoms of HD are uncoordinated, jerky body movements that cause the patients to have severe gait abnormalities, especially in terms of their walking velocity. At specific times, it is faster than the HC, and at other times, it is slower. As shown in Figure 3 (left), on the 0.1–5 Hz spectrogram, the walking velocity of the HD patient arbitrarily changed corresponding with the time; for example, it can be seen that from the initial time to 2 s, the strongest force level (yellow) was at 1.5–2 Hz for 2 s until 4 s. The strongest force magnitude frequency decreased to 1 Hz, and there were two strong force magnitudes at 4 to 7 s, (1 Hz and 2–2.5 Hz). The CF force time–domain signal could not be distinguished between the global and local minimum as nearly all of the lowest force magnitudes were not close to the 0 force unit, which means at a specific time, both feet appeared to be in contact with the ground.
The 5–50 Hz spectrogram showed the strongest force power (yellow), where a significant force (light blue) only occurred at a specific period of time and had a different magnitude every time. Based on this observation, it can be concluded that the walking velocity of the HD subject fluctuated.

4.4. Parkinson’s Disease

As presented in Figure 3 (right), the time–frequency spectrogram of the PD subject is similar to that of the HC. The strongest force power was at 1.6–2 Hz and 1 Hz for the 0.1–5 Hz spectrogram and for the 5–50 Hz spectrogram, the strongest force magnitude (yellow) was approximately 5 Hz and the significant force power (light blue) occurred up to 50 Hz every time. However, the force magnitude was not distributed equally during the entire walking period. It was also obvious that the pattern of the lowest force magnitude was irregular at 2.5–5 Hz. This indicates that the global and local minimum magnitudes are not the same in every gait cycle. PD patients can exhibit walking velocity similar to that of a normal person, but their force distribution is typically not distributed equally due to the possibility of having a tremor.

4.5. Comparison Results with the Existing Literature

The comparison was made with the study by Zeng et al. [22]. The authors presented the gait dynamics method to classify NDD via the deterministic learning theory. They used LOOCV as the evaluation method only for ALS vs. HC, HD vs. HC, and PD vs. HC. They also employed an all-training-all-testing evaluation method for all their classification experiments, but in the current study, we did not use this method. A comparison was also made with the study by Zhao et al. [23]. They implemented dual-channel long short-term memory (LSTM)-based multi-feature extraction on gait for the diagnosis of NDD. Here, only accuracy results for ALS vs. HC, HD vs. HC, PD vs. HC, and NDD vs. HC were compared using LOOCV as the evaluation method.
We also compared our results with two studies from Tuan D. Pham [63], who proposed a novel method for gait analysis by transforming a time-series data sequence into images from which texture analysis methods and texture features of a gait can be extracted and presented the sensitivity, specificity, AUC value, and accuracy of HC vs. HD, HC vs. PD, and HC vs. ALS classifications using LOOCV as the evaluation method, and from Ren et al., who applied empirical mode decomposition in gait rhythm fluctuation analysis in neurodegenerative diseases subjects and used 10-fold cross-validation in order to overcome overfitting and obtained the AUC values of HD vs. HC, PD vs. HC, and ALS vs. HC [64]. A comparison of these studies with the results obtained using the proposed method is shown in Table 10.
In conclusion, the proposed method outperformed the classification results from Zeng et al., Zhao et al., and Ren et al. The NDD detection algorithm, proposed by Pham, obtained better results than the proposed method in the PD vs. HC classification. However, in the ALS vs. HC and HD vs. HC classifications, the proposed method achieved the same performance as the journal in terms of all evaluation parameters. However, the authors also used another method for classifying patients with PD (PD vs. HC) using linear discriminant analysis (LDA) and LOOCV was also performed as an evaluation technique with poor classification results. The accuracy only reached 77.42%.

5. Conclusions

This study used a time–frequency spectrogram based on a vGRF signal to implement a novel AI-based NDD detection algorithm. The ability to distinguish between the gait phenomena of NDD patients and a HC was achieved through pattern visualization and the recognition of the time–frequency spectrogram. By transforming the signal from the time–domain to the time–frequency domain, CWT was used to visualize the spectrogram of a gait foot force signal. By transforming the signal from the time–domain to the time–frequency domain, CWT was used to visualize the spectrogram of a gait foot force signal. To achieve good feature visualization, three-time window (10, 30, and 60 s) and three types of gait foot force signals were chosen as inputs (LF, RF, and CF force signals). Following the transformation of the original signal, feature enhancement using PCA was used to improve between-class separability while reducing within-class separability. Finally, CNN was used to classify the spectrogram images. Two types of cross-validation methods, LOOCV and k-fold CV (k = 5), were used to assess the CNN classification process, and four parameters were generated, including accuracy, sensitivity, specificity, and the AUC value. As a result, the proposed method outperformed state-of-the-art NDD detection methods in the literature for more than 95.32% of the parameters evaluated.
Despite the fact that the proposed method received importance-performance evidence, there are several significant areas in which it could be improved. First, since the proposed method was used to improve the performance of an existing database, clinical data should be obtained for verification and to address the database’s constraints (the limited number of NDD patients). Our own manufactured smart insole with an embedded 0.5” force-sensing resistor will be used to gather clinical data. Instead of walking down a long pathway, the NDD patient would be required to perform some basic daily tasks such as turning around and sitting. Second, long-term data collection for NDD progression is important for NDD patient therapy as the gait pattern of NDD patients should change over time as the disease progresses. Third, the NDD gait phenomenon based on a time–frequency spectrogram should be discussed with doctors to ensure the clinical meaning. Fourth, in order to validate and compare the efficiency of pattern visualization and recognition based on the use of a time–frequency spectrogram in NDD detection applications, other input data (such as kinetic data, temporal data, step length, and cadence) and classifiers should be used.
Based on pattern visualization and recognition using a deep learning classifier, the time–frequency spectrogram was successfully used to differentiate the gait phenomenon between NDD patients and a HC in this study. A fuzzy recurrence plot can also be used to implement and observe pattern visualization and recognition of the NDD gait phenomenon. A deep learning gait classification algorithmic based on fuzzy recurrence plot images could be used to improve NDD gait classification in the future.

Author Contributions

Conceptualization, F.S. and C.-W.L.; methodology, F.S. and C.-W.L.; software, F.S.; validation, F.S.; investigation, F.S. and C.-W.L.; resources, C.-W.L.; writing—original draft preparation, F.S.; writing—review and editing, C.-W.L.; supervision, C.-W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Ministry of Science and Technology (Taiwan), grant number 108-2628-E-006-003-MY3.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. JPND Research. What is Neurodegenerative Disease? JPND Research. 7 February 2015. Available online: https://bit.ly/2Hkzs9w (accessed on 12 July 2019).
  2. Lee, A.; Gilbert, R. Epidemiology of Parkinson Disease. Neurol. Clin. 2016, 34, 955–965. [Google Scholar] [CrossRef]
  3. Parkinson’s Disease Foundation. Statistics on Parkinson’s. Parkinson’s Disease Foundation. 2018. EIN: 13-1866796. Available online: https://bit.ly/2RCeh9H (accessed on 12 July 2019).
  4. Chiò, A.; Logroscino, G.; Traynor, B.; Collins, J.; Simeone, J.; Goldstein, L.; White, L. Global Epidemiology of Amyotrophic Lateral Sclerosis: A Systematic Review of the Published Literature. Neuroepidemiology 2013, 41, 118–130. [Google Scholar] [CrossRef] [Green Version]
  5. Renton, A.E.; Chio, A.; Traynor, B.J. State of play in amyotrophic lateral sclerosis genetics. Nat. Neurosci. 2014, 17, 17–23. [Google Scholar] [CrossRef]
  6. Agrawal, M.; Biswas, A. Molecular diagnostics of neurodegenerative disorders. Front. Mol. Biosci. 2015, 2, 54. [Google Scholar] [CrossRef] [Green Version]
  7. Harvard NeuroDiscovery Center. The Challenge of Neurodegenerative Diseases. Available online: https://bit.ly/2soDGmD (accessed on 12 July 2019).
  8. Hausdorff, J.M.; Cudkowicz, M.E.; Firtion, R.; Wei, J.Y.; Goldberger, A.L. Gait variability and basal ganglia disorders: Stride-to-stride variations of gait cycle timing in Parkinson’s disease and Huntington’s disease. Mov. Disord. 1998, 13, 428–437. [Google Scholar] [CrossRef]
  9. Brown, R.H., Jr.; Al-Chalabi, A. Amyotrophic Lateral Sclerosis. N. Engl. J. Med. 2017, 377, 1602. [Google Scholar] [CrossRef] [Green Version]
  10. Hausdorff, J.M.; Lertratanakul, A.; Cudkowicz, M.E.; Peterson, A.L.; Kaliton, D.; Goldberger, A.L. Dynamic markers of altered gait rhythm in amyotrophic lateral sclerosis. J. Appl. Physiol. 2000, 88, 2045–2053. [Google Scholar] [CrossRef] [PubMed]
  11. Zarei, S.; Carr, K.; Reiley, L.; Diaz, K.; Guerra, O.; Altamirano, P.F.; Pagani, W.; Lodin, D.; Orozco, G.; Chinea, A. A comprehensive review of amyotrophic lateral sclerosis. Surg. Neurol. Int. 2015, 6, 171. [Google Scholar] [CrossRef] [PubMed]
  12. Banaie, M.; Sarbaz, Y.; Gharibzadeh, S.; Towhidkhah, F. Huntington’s disease: Modeling the gait disorder and proposing novel treatments. J. Theor. Biol. 2008, 254, 361–367. [Google Scholar] [CrossRef]
  13. Dayalu, P.; Albin, R.L. Huntington disease: Pathogenesis and treatment. Neurol. Clin. 2015, 33, 101–114. [Google Scholar] [CrossRef]
  14. Pyo, S.J.; Kim, H.; Kim, I.S.; Park, Y.-M.; Kim, M.-J.; Lee, H.M.; Koh, S.-B. Quantitative gait analysis in patients with huntington’s disease. J. Mov. Disord. 2017, 10, 140–144. [Google Scholar] [CrossRef]
  15. National Institute of Neurological Disorders and Stroke. Parkinson’s Disease Information Page. 2016. Available online: https://bit.ly/2xTA6rL (accessed on 12 July 2019).
  16. Hoff, J.; v/d Plas, A.; Wagemans, E.; van Hilten, J. Accelerometric assessment of levodopa-induced dyskinesias in Parkinson’s disease. Mov. Disord. Off. J. Mov. Disord. Soc. 2001, 16, 58–61. [Google Scholar] [CrossRef]
  17. Pistacchi, M. Gait analysis and clinical correlations in early Parkinson’s disease. Funct. Neurol. 2017, 32, 28–34. [Google Scholar] [CrossRef] [PubMed]
  18. Xia, Y.; Gao, Q.; Ye, Q. Classification of gait rhythm signals between patients with neuro-degenerative diseases and normal subjects: Experiments with statistical features and different classification models. Biomed. Signal Process. Control. 2015, 18, 254–262. [Google Scholar] [CrossRef]
  19. Ertuğrul, Ö.F.; Kaya, Y.; Tekin, R.; Almalı, M.N. Detection of Parkinson’s disease by shifted one dimensional local binary patterns from gait. Expert Syst. Appl. 2016, 56, 156–163. [Google Scholar] [CrossRef]
  20. Wu, Y.; Chen, P.; Luo, X.; Wu, M.; Liao, L.; Yang, S.; Rangayyan, R.M. Measuring signal fluctuations in gait rhythm time series of patients with Parkinson’s disease using entropy parameters. Biomed. Signal Process. Control 2017, 31, 265–271. [Google Scholar] [CrossRef]
  21. Bilgin, S. The impact of feature extraction for the classification of amyotrophic lateral sclerosis among neurodegenerative diseases and healthy subjects. Biomed. Signal Process. Control. 2017, 31, 288–294. [Google Scholar] [CrossRef]
  22. Zeng, W.; Wang, C. Classification of neurodegenerative diseases using gait dynamics via deterministic learning. Inf. Sci. 2015, 317, 246–258. [Google Scholar] [CrossRef]
  23. Zhao, A.; Qi, L.; Dong, J.; Yu, H. Dual channel LSTM based multi-feature extraction in gait for diagnosis of Neurodegenerative diseases. Knowl. Based Syst. 2018, 145, 91–97. [Google Scholar] [CrossRef] [Green Version]
  24. Huang, J.; Chen, B.; Yao, B.; He, W. ECG arrhythmia classification using STFT-based spectrogram and convolutional neural network. IEEE Access 2019, 7, 92871–92880. [Google Scholar] [CrossRef]
  25. Xie, Q.; Tu, S.; Wang, G.; Lian, Y.; Xu, L. Feature enrichment based convolutional neural network for heartbeat classification from electrocardiogram. IEEE Access 2019, 7, 153751–153760. [Google Scholar] [CrossRef]
  26. He, W.; Wang, G.; Hu, J.; Li, C.; Guo, B.; Li, F. Simultaneous human health monitoring and time-frequency sparse representation using EEG and ECG signals. IEEE Access 2019, 7, 85985–85994. [Google Scholar] [CrossRef]
  27. Tadesse, G.A.; Javed, H.; Thanh, N.L.N.; Thai, H.D.H.; Van Tan, L.; Thwaites, L.; Clifton, D.A.; Zhu, T. Multi-modal diagnosis of infectious diseases in the developing world. IEEE J. Biomed. Health Inform. 2020, 24, 2131–2141. [Google Scholar] [CrossRef] [PubMed]
  28. Xu, L. Deep bidirectional intelligence: AlphaZero, deep IA-search, deep IA-infer, and TPC causal learning. In Applied Informatics; Springer: Berlin/Heidelberg, Germany, 2018; Volume 5, p. 5. [Google Scholar]
  29. Salem, M.; Taheri, S.; Yuan, J.-S. ECG Arrhythmia Classification Using Transfer Learning from 2-Dimensional Deep CNN Features. In Proceedings of the 2018 IEEE Biomedical Circuits and Systems Conference (BioCAS), Cleveland, OH, USA, 17–19 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–4. [Google Scholar]
  30. Hausdorff, J.M.; Ladin, Z.; Wei, J.Y. Footswitch system for measurement of the temporal parameters of gait. J. Biomech. 1995, 28, 347–351. [Google Scholar] [CrossRef]
  31. Hausdorff, J.M.; Lertratanakul, A.; Cudkowicz, M.E.; Peterson, A.L.; Kaliton, D.; Goldberger, A.L. Gait Dynamics in Neuro-Degenerative Disease Database. 21 Dec 2000; PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 2019, 101, e215–e220. Available online: https://www.physionet.org/content/gaitndd/1.0.0/ (accessed on 12 July 2019).
  32. Lin, C.-W.; Wen, T.-C.; Setiawan, F. Evaluation of vertical ground reaction forces pattern visualization in neurodegenerative diseases identification using deep learning and recurrence plot image feature extraction. Sensors 2020, 20, 3857. [Google Scholar] [CrossRef] [PubMed]
  33. Sadowsky, J. The continuous wavelet transform: A tool for signal investigation and understanding. Johns Hopkins APL Tech. Dig. 1994, 15, 306. [Google Scholar]
  34. Allen, J. Short term spectral analysis, synthesis, and modification by discrete Fourier transform. IEEE Trans. Acoust. Speech Signal Process. 1977, 25, 235–238. [Google Scholar] [CrossRef]
  35. Daubechies, I.; Lu, J.; Wu, H.-T. Synchrosqueezed wavelet transforms: An empirical mode decomposition-like tool. Appl. Comput. Harmon. Anal. 2011, 30, 243–261. [Google Scholar] [CrossRef] [Green Version]
  36. Lever, J.; Krzywinski, M.; Altman, N. Points of Significance: Principal Component Analysis; Nature Publishing Group: Berlin, Germany, 2017; Chapter 1; p. 1. [Google Scholar]
  37. Jolliffe, I.T. Introduction. In Principal Component Analysis, 2nd ed.; Springer: New York, NY, USA, 2002. [Google Scholar]
  38. Roccetti, M.; Delnevo, G.; Casini, L.; Mirri, S. An alternative approach to dimension reduction for pareto distributed data: A case study. J. Big Data 2021, 8, 1–23. [Google Scholar] [CrossRef]
  39. Ren, X.-D.; Guo, H.-N.; He, G.-C.; Xu, X.; Di, C.; Li, S.-H. Convolutional neural network based on principal component analysis initialization for image classification. In Proceedings of the 2016 IEEE First International Conference on Data Science in Cyberspace (DSC), Changsha, Chine, 13–16 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 329–334. [Google Scholar]
  40. Ng, A.; Ngiam, J.; Foo, C.Y.; Mai, Y.; Suen, C. UFLDL Tutorial. 2012. Available online: http://deeplearning.stanford.edu/wiki/index.php/UFLDL_Tutorial (accessed on 19 March 2021).
  41. Garg, I.; Panda, P.; Roy, K. A low effort approach to structured CNN design using PCA. IEEE Access 2019, 8, 1347–1360. [Google Scholar] [CrossRef]
  42. Seuret, M.; Alberti, M.; Liwicki, M.; Ingold, R. PCA-initialized deep neural networks applied to document image analysis. In Proceedings of the 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR), Kyoto, Japan, 9–15 November 2017; IEEE: Piscataway, NJ, USA, 2017; Volume 1, pp. 877–882. [Google Scholar]
  43. O’Shea, K.; Nash, R. An introduction to convolutional neural networks. arXiv 2015, arXiv:1511.08458. [Google Scholar]
  44. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef] [Green Version]
  45. Liu, Y.H. Feature extraction and image recognition with convolutional neural networks. J. Phys. Conf. Ser. 2018, 1087, 062032. [Google Scholar] [CrossRef]
  46. Garcia-Gasulla, D.; Parés, F.; Vilalta, A.; Moreno, J.; Ayguadé, E.; Labarta, J.; Cortés, U.; Suzumura, T. On the behavior of convolutional nets for feature extraction. J. Artif. Intell. Res. 2018, 61, 563–592. [Google Scholar] [CrossRef] [Green Version]
  47. Rajaraman, S.; Antani, S.K.; Poostchi, M.; Silamut, K.; Hossain, M.A.; Maude, R.J.; Jaeger, S.; Thoma, G.R. Pre-trained convolutional neural networks as feature extractors toward improved malaria parasite detection in thin blood smear images. PeerJ 2018, 6, e4568. [Google Scholar] [CrossRef]
  48. Hegde, R.B.; Prasad, K.; Hebbar, H.; Singh, B.M.K. Feature extraction using traditional image processing and convolutional neural network methods to classify white blood cells: A study. Australas. Phys. Eng. Sci. Med. 2019, 42, 627–638. [Google Scholar] [CrossRef]
  49. Li, J.; Si, Y.; Xu, T.; Jiang, S. Deep convolutional neural network based ECG classification system using information fusion and one-hot encoding techniques. Math. Probl. Eng. 2018, 2018, 7354081. [Google Scholar] [CrossRef]
  50. Liu, Q.; Cai, J.; Fan, S.-Z.; Abbod, M.F.; Shieh, J.-S.; Kung, Y.; Lin, L. Spectrum analysis of EEG signals using CNN to model patient’s consciousness level based on anesthesiologists’ experience. IEEE Access 2019, 7, 53731–53742. [Google Scholar] [CrossRef]
  51. Thanaraj, K.P.; Parvathavarthini, B.; Tanik, U.J.; Rajinikanth, V.; Kadry, S.; Kamalanand, K. Implementation of deep neural networks to classify EEG signals using gramian angular summation field for epilepsy diagnosis. arXiv 2020, arXiv:2003.04534. [Google Scholar]
  52. Lin, M.; Chen, Q.; Yan, S. Network in network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  53. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  54. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  55. Cortes, C.; Vapnik, V. Support-Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  56. Refaeilzadeh, P.; Tang, L.; Liu, H.; Liu, L.; Özsu, M.T. Cross-Validation. Springer Ref. 2011, 532–538. [Google Scholar]
  57. Wong, T.-T.; Yeh, P.-Y. Reliable Accuracy Estimates from k-Fold Cross Validation. IEEE Trans. Knowl. Data Eng. 2020, 32, 1586–1594. [Google Scholar] [CrossRef]
  58. Roccetti, M.; Delnevo, G.; Casini, L.; Salomoni, P. A Cautionary Tale for Machine Learning Design: Why we Still Need Human-Assisted Big Data Analysis. Mob. Netw. Appl. 2020, 25, 1075–1083. [Google Scholar] [CrossRef]
  59. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  60. Youden, W.J. Index for rating diagnostic tests. Cancer 1950, 3, 32–35. [Google Scholar] [CrossRef]
  61. Yang, M.; Zheng, H.; Wang, H.; McClean, S. Feature selection and construction for the discrimination of neurodegenerative diseases based on gait analysis. In Proceedings of the 2009 3rd International Conference on Pervasive Computing Technologies for Healthcare, London, UK, 1–3 April 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 1–7. [Google Scholar]
  62. Nilsson, J.; Thorstensson, A. Adaptability in frequency and amplitude of leg movements during human locomotion at different speeds. Acta Physiol. Scand. 1987, 129, 107–114. [Google Scholar] [CrossRef]
  63. Pham, T.D. Texture classification and visualization of time series of gait dynamics in patients with neuro-degenerative diseases. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 26, 188–196. [Google Scholar] [CrossRef] [PubMed]
  64. Ren, P.; Tang, S.; Fang, F.; Luo, L.; Xu, L.; Bringas-Vega, M.L.; Yao, D.; Kendrick, K.M.; Valdes-Sosa, P.A. Gait rhythm fluctuation analysis for neurodegenerative diseases by empirical mode decomposition. IEEE Trans. Biomed. Eng. 2016, 64, 52–60. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flowchart of the proposed NDD detection algorithm using CWT as the feature transformation.
Figure 1. Flowchart of the proposed NDD detection algorithm using CWT as the feature transformation.
Brainsci 11 00902 g001
Figure 2. Time–frequency spectrogram using the CWT of the CF vGRF signal of HC (left) and ALS (right) subjects in 10-s time windows.
Figure 2. Time–frequency spectrogram using the CWT of the CF vGRF signal of HC (left) and ALS (right) subjects in 10-s time windows.
Brainsci 11 00902 g002
Figure 3. Time–frequency spectrogram using the CWT of the CF vGRF signal of HD (left) and PD (right) subjects in 10-s time windows.
Figure 3. Time–frequency spectrogram using the CWT of the CF vGRF signal of HD (left) and PD (right) subjects in 10-s time windows.
Brainsci 11 00902 g003
Figure 4. Flowchart of new feature extracted reconstruction using principal component analysis (PCA) as feature enhancement purpose.
Figure 4. Flowchart of new feature extracted reconstruction using principal component analysis (PCA) as feature enhancement purpose.
Brainsci 11 00902 g004
Figure 5. The learning curve of the SVM classifier with the feature extracted from fully connected layer of pre-trained AlexNet CNN as the input.
Figure 5. The learning curve of the SVM classifier with the feature extracted from fully connected layer of pre-trained AlexNet CNN as the input.
Brainsci 11 00902 g005
Table 1. CNN architecture of the proposed method.
Table 1. CNN architecture of the proposed method.
LayerSizeHyperparameter
WeightBias
InputImage 227 × 227 × 3 --
1Convolution 55 × 55 × 96 11 × 11 × 3 × 96 1 × 1 × 96
2ReLU 1 55 × 55 × 96 --
3Cross Channel Normalization 55 × 55 × 96 --
4Max Pooling 2 27 × 27 × 96 --
5Grouped Convolution 27 × 27 × 256 5 × 5 × 48 × 128 × 2 1 × 1 × 128 × 2
6ReLU 1 27 × 27 × 256 --
7Cross Channel Normalization 27 × 27 × 256 --
8Max Pooling 2 13 × 13 × 256 --
9Convolution 13 × 13 × 384 3 × 3 × 256 × 3 1 × 1 × 384
10ReLU 1 13 × 13 × 384 --
11Grouped Convolution 13 × 13 × 384 3 × 3 × 192 × 192 × 2 1 × 1 × 192 × 2
12ReLU 1 13 × 13 × 384 --
13Grouped Convolution 13 × 13 × 256 3 × 3 × 192 × 128 × 2 1 × 1 × 128 × 2
14ReLU 1 13 × 13 × 256 --
15Max Pooling 2 6 × 6 × 256 --
16Fully Connected 1 × 1 × 4096 4096 × 9216 4096 × 1
17ReLU 1 1 × 1 × 4096 --
18Dropout (50%) 2 1 × 1 × 4096 --
19Fully Connected 1 × 1 × 4096 4096 × 4096 4096 × 1
20SVM Classification Model 3---
OutputClassification Output---
1 Rectified linear unit as activation layer. 2 Max pooling and dropout layer can prevent the overfitting [52,53,54]. 3 The optimization hyperparameter of SVM is based on Bayesian optimization with 30 iterations.
Table 2. Average computation time of the proposed method.
Table 2. Average computation time of the proposed method.
Time WindowTotal Number of vGRF SpectrogramElapsed Time (s)
LOOCVk-Fold CV (k = 5)
10 s19207933.68433.013
30 s640873.82912.708
60 s320235.9867.473
Table 3. Summary results of two-class classification states (NDD and HC groups) using PCA.
Table 3. Summary results of two-class classification states (NDD and HC groups) using PCA.
Classification TaskEvaluation ParameterProposed Method
CWT + PCA (10 s/30 s/60 s)STFT + PCA (10 s/30 s/60 s)WSST + PCA (10 s/30 s/60 s)
LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)
ALS vs. HCSen (%)100/100/100100/100/100100/100/100100/100/100100/100/100100/100/100
Spec (%)99.79/99.38/98.7799.79/99.39/98.8299.79/99.38/98.7799.79/99.39/98.82100/98.77/98.7799.79/98.79/97.65
Acc (%)99.89/99.66/99.3199.89/99.66/99.3199.89/99.66/99.3199.89/99.66/99.31100/99.31/99.3199.89/99.31/98.62
AUC0.9990/0.9969/0.99381/1/10.9990/0.9969/0.99381/1/11/0.9938/0.99380.9987/0.9962/0.9846
HD vs. HCSen (%)100/100/100100/100/100100/100/100100/100/10099.83/99.49/10099.83/100/100
Spec (%)100/100/100100/100/10099.79/100/10099.79/100/10099.79/96.36/89.89100/99.39/91.67
Acc (%)100/100/100100/100/10099.91/100/10099.91/100/10099.81/98.06/9599.91/99.72/95.56
AUC1/1/11/1/10.9990/1/11/1/10.9981/0.9793/0.94941/1/1
PD vs. HCSen (%)98.38/94.27/94.8199.13/97.47/10093.59/92.36/88.6192.86/91.09/89.7089.68/86.54/79.7194.46/89.96/60.69
Spec (%)95.17/98.69/97.4494.53/96.42/96.4791.68/89.76/93.4291.39/91.16/94.6288.06/90.26/76.7492.33/95.07/76.99
Acc (%)96.67/96.45/96.1396.45/96.77/98.0692.58/90.97/90.9791.29/90.97/90.3288.82/88.39/78.0692.47/91.29/61.94
AUC0.9678/0.9648/0.96120.9992/0.9969/0.99670.9264/0.9106/0.91010.9794/0.9659/0.96790.8887/0.8840/0.78230.9957/0.9795/0.8763
Note: bold and underlined classification results were selected as the best based on Youden’s index.
Table 4. Comparison results of all two-class classification states between PCA and non-PCA (NDD and HC group) using k-fold CV (k = 5).
Table 4. Comparison results of all two-class classification states between PCA and non-PCA (NDD and HC group) using k-fold CV (k = 5).
Classification TaskEvaluation ParameterProposed Method
CWT + PCA (10 s/30 s/60 s)STFT + PCA (10 s/30 s/60 s)WSST + PCA (10 s/30 s/60 s)
PCANon-PCAPCANon-PCAPCANon-PCA
ALS vs. HCSen (%)100/100/10094.97/96.26/89.02100/100/10087.88/88.49/93.14100/100/10075.69/76.30/86.11
Spec (%)99.79/99.39/98.8292.73/95.23/94.2199.79/99.39/98.8293.74/93.65/91.7899.79/98.79/97.6565.49/77.41/62.59
Acc (%)99.89/99.66/99.3193.56/95.52/91.7299.89/99.66/99.3190.69/91.03/91.0399.89/99.31/98.6266.90/69.31/62.07
AUC1/1/10.9809/0.9871/0.96761/1/10.9676/0.9712/0.95960.9987/0.9962/0.98460.7753/0.7698/0.6779
HD vs. HCSen (%)100/100/10093.28/91.37/93.42100/100/10085.89/89.16/87.4599.83/100/10078.84/76.27/89.42
Spec (%)100/100/10081.14/85.05/89.2799.79/100/10081.34/77.68/77.24100/99.39/91.6767.82/61.72/62.52
Acc (%)100/100/10087.04/88.06/91.1199.91/100/10082.31/81.39/77.7899.91/99.72/95.5661.76/67.78/61.11
AUC1/1/10.9431/0.9366/0.96881/1/10.8904/0.9034/0.90311/1/10.7186/0.7898/0.7969
PD vs. HCSen (%)99.13/97.47/10088.48/89.34/91.9892.86/91.09/89.7083.56/71.14/91.6794.46/89.96/60.6969.61/90.40/78.97
Spec (%)94.53/96.42/96.4784.74/89.09/86.3291.39/91.16/94.6280.85/83.56/67.1092.33/95.07/76.9962.60/55.48/54.31
Acc (%)96.45/96.77/98.0686.24/88.06/88.3991.29/90.97/90.3281.08/72.26/67.7492.47/91.29/61.9460.11/53.87/55.48
AUC0.9992/0.9969/0.99670.9222/0.9481/0.94500.9794/0.9659/0.96790.8907/0.8294/0.89540.9957/0.9795/0.87630.6908/0.6507/0.7150
Note: bold and underlined classification results were selected as the best based on Youden’s index.
Table 5. Summary results of all two-class classification states (among the NDD and all NDD in one group with HC group) using PCA.
Table 5. Summary results of all two-class classification states (among the NDD and all NDD in one group with HC group) using PCA.
Classification TaskEvaluation ParameterProposed Method
CWT + PCA (10 s/30 s/60 s)STFT + PCA (10 s/30 s/60 s)WSST + PCA (10 s/30 s/60 s)
LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)
ALS vs. HDSen (%)97.95/96.83/10097.99/98.46/10096.93/93.43/92.1999.01/96.18/95.7996.70/87.60/90.3297.88/82.21/89
Spec (%)98.67/96.08/93.4698.67/96.61/95.3298.16/98.96/94.0697.10/97.54/93.0698.49/91.54/91.2696.30/95.38/88.49
Acc (%)98.38/96.36/95.7698.38/97.27/96.9797.68/96.67/93.3397.78/96.97/93.3397.78/90/90.9196.87/86.67/85.45
AUC0.9831/0.9645/0.96730.9953/0.9958/0.98000.9755/0.9620/0.93120.9876/0.9851/0.97920.9760/0.8957/0.90790.9934/0.9687/0.9419
PD vs. ALSSen (%)99.78/99.34/98.6899.78/99.35/98.7599.78/99.34/98.6899.78/99.35/98.7599.78/98.04/94.9499.78/98.75/97.50
Spec (%)100/100/100100/100/100100/100/100100/100/100100/100/100100/100/100
Acc (%)99.88/99.64/99.2999.88/99.64/99.2999.88/99.64/99.2999.88/99.64/99.2999.88/99.64/97.1499.88/99.29/98.57
AUC0.9989/0.9967/0.99341/1/0.99230.9989/0.9967/0.99341/1/0.9923/0.99230.9989/0.9902/0.97470.9987/0.9962/0.9923
HD vs. PDSen (%)98.82/96.14/94.1798.52/95.28/96.2394.93/93.30/96.0894.61/94.78/94.8594.21/89.16/86.6089.79/93.43/96.80
Spec (%)96.72/99.30/95.8397.82/99.31/97.4295.44/96.45/97.2697.01/95.75/96.4093.05/87.07/79.4999.51/82.95/84.09
Acc (%)97.90/97.43/94.8698.19/96.86/96.5795.14/94.57/96.5795.52/94.57/94.8693.71/88.29/83.4393.14/86/89.14
AUC0.9777/0.9772/0.95000.9982/0.9953/0.99400.9519/0.9488/0.96670.9912/0.9926/0.99530.9363/0.8812/0.83040.9979/0.9540/0.9767
NDD vs. HCSen (%)99.58/98.96/97.94100/99.38/97.5896.77/95.91/96.6998.06/96.60/97.1593.85/98.49/93.6495.30/93.91/100
Spec (%)95.76/96.88/97.4095.29/96.47/97.3993.52/92.72/92.3192.09/93.73/94.2991.51/86.44/77.3899.29/88.54/72.83
Acc (%)98.59/98.44/97.8198.75/98.59/97.5095.99/95.16//95.6396.41/95.63/95.9493.33/95.16/89.3896.09/90.78/89.38
AUC0.9767/0.9792/0.97670.9991/0.9959/0.96890.9515/0.9431/0.94500.9844/0.9847/0.97420.9268/0.9246/0.85510.9978/0.9938/0.9943
Note: bold and underlined classification results were selected as the best based on Youden’s index.
Table 6. Comparison results of all two-class classification states between PCA and non-PCA (among the NDD and all NDD in one group with HC group) using k-fold CV (k = 5).
Table 6. Comparison results of all two-class classification states between PCA and non-PCA (among the NDD and all NDD in one group with HC group) using k-fold CV (k = 5).
Classification TaskEvaluation ParameterProposed Method
CWT + PCA (10 s/30 s/60 s)STFT + PCA (10 s/30 s/60 s)WSST + PCA (10 s/30 s/60 s)
LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)
ALS vs. HDSen (%)97.99/98.46/10081.75/88.74/92.6299.01/96.18/95.7986.74/83.68/82.8497.88/82.21/8948.35/55.47/55.68
Spec (%)98.67/96.61/95.3295.82/93.63/96.0997.10/97.54/93.0687.75/91.06/92.0596.30/95.38/88.4992.85/85.99/83.86
Acc (%)98.38/97.27/96.9788.89/90.61/94.5597.78/96.97/93.3386.97/86.67/83.6496.87/86.67/85.4556.56/52.12/61.82
AUC0.9953/0.9958/0.98000.9584/0.9551/0.97880.9876/0.9851/0.97920.9410/0.9510/0.95190.9934/0.9687/0.94190.6849/0.6187/0.7612
PD vs. ALSSen (%)99.78/99.35/98.7591.99/80.17/86.0899.78/99.35/98.7586.12/87.07/85.7999.78/98.75/97.5074.55/63.51/84.04
Spec (%)100/100/10088.84/85.75/79.51100/100/10083.91/74.38/76.07100/100/10063.60/86.06/69.66
Acc (%)99.88/99.64/99.2990.36/81.43/8099.88/99.64/99.2984.17/77.50/77.1499.88/99.29/98.5764.40/64.64/57.14
AUC1/1/0.99230.9513/0.8835/0.81951/1/0.9923/0.99230.9315/0.8892/0.89380.9987/0.9962/0.99230.7588/0.6604/0.6641
HD vs. PDSen (%)98.52/95.28/96.2384.34/79.74/86.8694.61/94.78/94.8572.76/76.98/75.9589.79/93.43/96.8072.56/63.81/79.10
Spec (%)97.82/99.31/97.4277.75/78.58/75.1697.01/95.75/96.4071.33/70.22/66.9899.51/82.95/84.0954.34/59.58/63.71
Acc (%)98.19/96.86/96.5780.29/77.14/78.8695.52/94.57/94.8670.48/72.57/69.7193.14/86/89.1456.48/59.71/61.71
AUC0.9982/0.9953/0.99400.8869/0.8692/0.84570.9912/0.9926/0.99530.7580/0.7922/0.78170.9979/0.9540/0.97670.6515/0.5985/0.6963
NDD vs. HCSen (%)100/99.38/97.5890.57/94.41/88.9498.06/96.60/97.1589.69/89.21/90.0195.30/93.91/10083.12/89.43/83.38
Spec (%)95.29/96.47/97.3977.63/74.75/83.4792.09/93.73/94.2965.08/58.69/66.9699.29/88.54/72.8346.99/38.03/70.67
Acc (%)98.75/98.59/97.5087.19/87.97/84.3896.41/95.63/95.9482.71/78.28/74.3896.09/90.78/89.3865.99/48.91/69.06
AUC0.9991/0.9959/0.96890.8938/0.9308/0.86340.9844/0.9847/0.97420.8478/0.8280/0.89090.9978/0.9938/0.99430.6424/0.7777/0.7500
Note: bold and underlined classification results were selected as the best based on Youden’s index.
Table 7. Summary results of all multi-class classification states using PCA for 10-s time window.
Table 7. Summary results of all multi-class classification states using PCA for 10-s time window.
Classification ClassEvaluation ParameterProposed Method
CWT + PCASTFT + PCAWSST + PCA
LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)
HCSen (%)98.7597.7192.5094.799087.71
Spec (%)99.0398.8997.9297.6496.1897.71
Acc (%)98.9698.5996.5696.9394.6495.21
AUC0.98890.98300.95210.96220.93090.9271
ALSSen (%)98.2197.9597.4497.6997.1895.90
Spec (%)99.3599.8099.3598.7698.9599.02
Acc (%)99.1199.4398.9698.5498.5998.39
AUC0.98780.98880.98390.98230.98070.9746
HDSen (%)97979593.8389.8389.33
Spec (%)98.8699.1796.5296.7496.8998.11
Acc (%)98.2898.4996.0495.8394.6995.36
AUC0.97930.98080.95760.95290.93360.9372
PDSen (%)95.3396.2285.3386.2286.2283.78
Spec (%)98.6497.2196.1996.7391.9092.52
Acc (%)97.8696.9893.6594.2790.5790.47
AUC0.96990.98080.90760.91480.89060.8815
Note: bold and underlined classification results were selected as the best based on Youden’s index.
Table 8. Summary results of all multi-class classification states using PCA for 30-s time window.
Table 8. Summary results of all multi-class classification states using PCA for 30-s time window.
Classification ClassEvaluation ParameterProposed Method
CWT + PCASTFT + PCAWSST + PCA
LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)
HCSen (%)96.2597.5089.3892.5086.2591.88
Spec (%)99.1798.5497.9297.7196.8896.46
Acc (%)98.4498.2895.7896.4194.2295.31
AUC0.97710.98020.93650.95100.91560.9417
ALSSen (%)96.9295.3896.9296.9293.0898.46
Spec (%)98.6398.8298.6398.2495.8891.76
Acc (%)98.2898.1398.2897.9795.3193.13
AUC0.97780.97100.97780.97580.94480.9511
HDSen (%)93.5093.509494.5076.5084.50
Spec (%)97.2797.7397.509595.6895
Acc (%)96.0996.4196.4194.8489.6991.72
AUC0.95390.95610.95750.94750.86090.8975
PDSen (%)92.67948884.6783.3376.67
Spec (%)98.1694.6993.8896.5390.2093.06
Acc (%)96.8894.5392.5093.7588.5989.22
AUC0.95410.94350.90940.90600.86770.8486
Note: bold and underlined classification results were selected as the best based on Youden’s index.
Table 9. Summary results of all multi-class classification states using PCA for 60-s time window.
Table 9. Summary results of all multi-class classification states using PCA for 60-s time window.
Classification ClassEvaluation ParameterProposed Method
CWT + PCASTFT + PCAWSST + PCA
LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)LOOCVk-Fold CV (k = 5)
HCSen (%)9598.7592.5097.5078.7576.25
Spec (%)99.1797.9297.5098.0492.9288.75
Acc (%)98.1398.1396.2594.3889.3885.63
AUC0.97080.98330.95000.95420.85830.8250
ALSSen (%)92.3193.8590.7786.1578.4675.38
Spec (%)98.0496.8697.2598.8298.8290.20
Acc (%)96.8896.2595.9496.2594.6987.19
AUC0.95170.95350.94010.92490.88640.8279
HDSen (%)989595947592
Spec (%)93.6495.4594.099591.3687.27
Acc (%)9595.3194.3894.6986.2588.75
AUC0.95820.95230.94550.94500.83180.8964
PDSen (%)89.3389.3385.3382.676857.33
Spec (%)97.5597.9693.8895.5183.2788.98
Acc (%)95.6395.9491.8892.5079.6981.56
AUC0.93440.93650.89610.89090.75630.7316
Note: bold and underlined classification results were selected as the best based on Youden’s index.
Table 10. A comparison of the performance of the proposed method using LOOCV.
Table 10. A comparison of the performance of the proposed method using LOOCV.
StudiesEvaluation ParameterClassification Task
ALS vs. HCHD vs. HCPD vs. HCNDD vs. HC
Zeng et al. [22]Sen (%)92.318587.50-
Spec (%)87.5081.2586.67-
Acc (%)89.6687.1087.10-
Zhao et al. [23]Acc (%)97.4394.9697.3396.42
Ren et al. [64]AUC0.89800.88100.9010-
Pham, T.D. [63]Sen (%)100100100-
Spec (%)100100100-
Acc (%)100100100-
AUC111-
The Proposed MethodSen (%)10010099.0898.96
Spec (%)10010095.9796.88
Acc (%)10010097.4298.44
AUC110.97520.9792
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Setiawan, F.; Lin, C.-W. Identification of Neurodegenerative Diseases Based on Vertical Ground Reaction Force Classification Using Time–Frequency Spectrogram and Deep Learning Neural Network Features. Brain Sci. 2021, 11, 902. https://doi.org/10.3390/brainsci11070902

AMA Style

Setiawan F, Lin C-W. Identification of Neurodegenerative Diseases Based on Vertical Ground Reaction Force Classification Using Time–Frequency Spectrogram and Deep Learning Neural Network Features. Brain Sciences. 2021; 11(7):902. https://doi.org/10.3390/brainsci11070902

Chicago/Turabian Style

Setiawan, Febryan, and Che-Wei Lin. 2021. "Identification of Neurodegenerative Diseases Based on Vertical Ground Reaction Force Classification Using Time–Frequency Spectrogram and Deep Learning Neural Network Features" Brain Sciences 11, no. 7: 902. https://doi.org/10.3390/brainsci11070902

APA Style

Setiawan, F., & Lin, C. -W. (2021). Identification of Neurodegenerative Diseases Based on Vertical Ground Reaction Force Classification Using Time–Frequency Spectrogram and Deep Learning Neural Network Features. Brain Sciences, 11(7), 902. https://doi.org/10.3390/brainsci11070902

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop