Next Article in Journal
Tracking of Mental Workload with a Mobile EEG Sensor
Next Article in Special Issue
Non-Contact Video-Based Assessment of the Respiratory Function Using a RGB-D Camera
Previous Article in Journal
Computational–Experimental Design Framework for Laser Path Length Controller
Previous Article in Special Issue
AIoT-Enabled Rehabilitation Recognition System—Exemplified by Hybrid Lower-Limb Exercises
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Implementation of a Deep Learning Algorithm Based on Vertical Ground Reaction Force Time–Frequency Features for the Detection and Severity Classification of Parkinson’s Disease

1
Department of Biomedical Engineering, College of Engineering, National Cheng Kung University, Tainan City 701, Taiwan
2
Medical Device Innovation Center, National Cheng Kung University, Tainan City 701, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(15), 5207; https://doi.org/10.3390/s21155207
Submission received: 9 July 2021 / Revised: 29 July 2021 / Accepted: 30 July 2021 / Published: 31 July 2021
(This article belongs to the Special Issue Sensors for Physiological Parameters Measurement)

Abstract

:
Conventional approaches to diagnosing Parkinson’s disease (PD) and rating its severity level are based on medical specialists’ clinical assessment of symptoms, which are subjective and can be inaccurate. These techniques are not very reliable, particularly in the early stages of the disease. A novel detection and severity classification algorithm using deep learning approaches was developed in this research to classify the PD severity level based on vertical ground reaction force (vGRF) signals. Different variations in force patterns generated by the irregularity in vGRF signals due to the gait abnormalities of PD patients can indicate their severity. The main purpose of this research is to aid physicians in detecting early stages of PD, planning efficient treatment, and monitoring disease progression. The detection algorithm comprises preprocessing, feature transformation, and classification processes. In preprocessing, the vGRF signal is divided into 10, 15, and 30 s successive time windows. In the feature transformation process, the time domain vGRF signal in windows with varying time lengths is modified into a time–frequency spectrogram using a continuous wavelet transform (CWT). Then, principal component analysis (PCA) is used for feature enhancement. Finally, different types of convolutional neural networks (CNNs) are employed as deep learning classifiers for classification. The algorithm performance was evaluated using k-fold cross-validation (kfoldCV). The best average accuracy of the proposed detection algorithm in classifying the PD severity stage classification was 96.52% using ResNet-50 with vGRF data from the PhysioNet database. The proposed detection algorithm can effectively differentiate gait patterns based on time–frequency spectrograms of vGRF signals associated with different PD severity levels.

1. Introduction

Parkinson’s disease (PD) is a neurodegenerative disease that belongs to a group of motor system disorders caused by the loss of dopamine-producing brain cells. PD is the second most common neurodegenerative disease [1]; its prevalence is approximately 0.3% in the general population, approximately 1% in individuals older than 60, and approximately 3% in people aged 80 and over [1]. The incidence of PD is 8–18 per 100,000 people. The median age at onset is 60 years, and the mean duration of the progression of the disease from diagnosis to death is approximately 15 years [1]. There is a 1.5–2-fold greater prevalence and incidence of this disease in men [1]. PD treatments cost approximately USD 2500 each year, and therapeutic surgery costs up to USD 100,000 per patient [2]. The primary PD symptoms are tremors in the hands, arms, legs, jaw, and face; rigidity (inflexibility of the limbs and trunk); bradykinesia (slowness in movement); and postural instability (balance and coordination disturbance) [3,4,5]. As these symptoms become more severe, patients may experience difficulties walking, talking, or accomplishing simple tasks. Currently, there are no blood or laboratory tests that assist in diagnosing PD. Symptoms of the disease include certain characteristic walking difficulties, such as a shortened stride length, decreased gait speed, increased stride-to-stride variation, a shambling gait, and frozen gait.
Gait analysis is used to assess and treat individuals with conditions affecting their ability to walk, such as poor health, advanced age, size, weight, and speed. A standard assessment is needed to clinically identify and evaluate gait characteristics and other phenomena in PD patients, such as gait count, walking speed, and step length. Pistacchi et al. analyzed temporal parameters (see Figure 1) in patients with early PD using 3D gait analysis-related cadence (PD patients: 102.46 ± 13.17 steps/min and healthy subjects: 113.84 ± 4.30 steps/min), stride duration (PD patients: 1.19 ± 0.18 s right limb and 1.19 ± 0.19 s left limb; healthy subjects: 0.426 ± 0.16 s right limb and 0.429 ± 0.23 s left limb), stance duration (PD patients: 0.74 ± 0.14 s right limb and 0.74 ± 0.16 s left limb; healthy subjects: 1.34 ± 1.1 s right limb and 0.83 ± 0.6 s left limb), and velocity (PD patients: 0.082 ± 0.29 m/s; healthy subjects: 1.33 ± 0.06 m/s) [5]. Sofuwa et al. concluded that individuals with PD showed a significant reduction in step length and walking speed compared with the non-PD control group [6]. These observations suggest that foot force is affected by PD. Lescano et al. aimed to analyze gait parameters, stance, swing phase duration, and the magnitude of the vertical component of the ground reaction force for the purpose of assessing whether there are statistically significant differences between PD patients in stages 2 and 2.5 (modified Hoehn and Yahr (HY) scale, see description in Table 1) [7]. Gait information has been developed for movement analysis in healthy control (CO, term defined by PhysioNet [8]) subjects and other subjects with different types of diseases. This approach is useful for understanding movement disorders arising from PD, and may be valuable in developing non-invasive automatic detection and severity classification approaches for PD.
Classification is the process of identifying the class of a new observation using a set of categories based on a training process involving observations for which the classes are known. In PD classification, various machine learning algorithms have been implemented as classifiers and combined with sophisticated feature extraction methods for dimensionality reduction. Recently, deep learning approaches, instead of conventional machine learning algorithms, have been applied to improve PD classification performance. For example, Jane et al. presented a Q-backpropagated time delay neural network (Q-BTDNN) in a clinical decision-making system (CDMS) to diagnose patients with PD (PD vs. CO) [11]. The Q-BTDNN was trained using a Q-learning induced backpropagation (Q-BP) training algorithm by generating a reinforced error signal, and the weights of the network were corrected through the backpropagation of the generated error signal. Correa et al. implemented a method to model PD patients’ difficulties in starting and ending movements by examining information from speech, handwriting, and gait [12]. These researchers trained a convolutional neural network (CNN) to classify PD patients and CO subjects. The PD population in the database was divided into three groups based upon the stage of PD: low, intermediate, or severe. Lee and Lim classified idiopathic PD patients and COs based on their gait force characteristics using a continuous wavelet transform (CWT) to generate approximate coefficients and detail coefficients [13]. Forty features were extracted from those coefficients using statistical approaches, including frequency distributions and their variabilities. The features of idiopathic PD patients and COs were classified using a neural network with weighted fuzzy membership functions (NEWFM). Zhao et al. developed a two-channel model that combined Long Short-Term Memory (LSTM) and CNNs to learn spatio-temporal patterns in gait data recorded by foot sensors [14]. The model was trained and tested on three public vGRF datasets. The model could perform multi-category classification on features such as the severity level of PD, while previous machine learning-based approaches could only perform binary classification.
As previously mentioned, only a few studies have used the deep learning approach for the detection and severity classification of PD, and some of them have used statistical features combined with machine learning methods. The drawbacks of using machine learning are the dependence of its performance on data size and the understanding of features [15,16]. Machine learning only performs well on small to medium datasets and needs a better understanding of features to represent the data. The objective of this work was to develop a deep learning classifier to help physicians screen and classify the severity of PD in patients, using vGRF spectrograms. The effectiveness of time–frequency spectrogram (feature transformation) of vGRF signals from left (LF), right (RF), and compound foot (CF = LF + RF) movements in classifying features of PD severity was investigated. Specifically, the aim was to determine whether a significant difference in vGRF is related to the specifics of disease severity, as passive (weight acceptance) and active (push off) peaks of vGRF are important gait parameters [17] and exhibit significant relevance in the detection of gait abnormalities, especially in the PD gait assessment [13,14,18,19,20]. Different deep learning algorithms (including AlexNet, ResNet-50, ResNet-101, and GoogLeNet) were also utilized with the proposed method to compare the effectiveness among classifiers.

2. Materials and Methods

The proposed PD severity classification algorithm attempts to extract pattern features and visualizations from vGRF signals in PD patients with severity stages of 0, 2, 2.5, and 3 on the HY rating scale by transforming one-dimensional time domain signals into two-dimensional patterns (images) using the feature transformation method from a CWT. The proposed PD severity classification algorithm consists of four main steps, as shown in Figure 2: (1) signal preprocessing of PD patients’ vGRF signals, (2) feature extraction from a spectrogram of the vGRF signal generated using CWT and PCA, (3) construction and training of a CNN classifier, and (4) cross-validation to evaluate the performance of the classification algorithm.

2.1. Gait in Parkinson’s Disease Database

The vGRF database used in this research, the Gait in Parkinson’s Disease Database (gaitpdb), is available online from PhysioNet [8]. The database comprises three datasets, which were contributed by Yogev et al. (Ga) [21], Hausdorff et al. (Ju) [22], and Frenkel-Toledo et al. (Si) [23,24].
The database contains information recorded from 93 idiopathic PD patients (average age: 66.3 years; 63% men and 37% women) and 73 CO subjects (average age: 66.3 years; 55% men and 45% women). Every subject was instructed to walk at their usual pace for about two minutes while wearing a pair of shoes with eight force sensors located under each insole. The raw vGRF signal data in this database were obtained using force-sensitive sensors (Ultraflex Computer Dyno Graphy, Intronic Inc., NL-7650 AB Tubbergen, The Netherlands) with the output proportional to the force under the foot in Newtons, collected at 100 samples per second (frequency of readings during movement was 100 Hz). The recordings also included two signals that reflect the sum of the eight sensor outputs from the left and right foot.
The database also contains information about each participant, including gender, age, height, weight, walking velocity, and severity level of PD. The PD severity level was assigned according to two rating scales, HY [10] and the Unified Parkinson’s Disease Rating Scale (UPDRS) [25]. The HY rating scale, widely used to represent the way in which symptoms of PD progress, defines five stages of PD, with two additional intermediate stages, 1.5 and 2.5 (Table 1) [10]. The number of participants diagnosed using the HY rating scale is shown in Table 2.

2.2. Signal Preprocessing

A two-minute foot force signal was acquired during data collection from subjects. The LF, RF, and CF vGRF signals of the CO and PD subjects were used as inputs to the proposed algorithm. It was difficult to interpret the foot force data directly, despite using a CWT to transform the features, due to the length of the foot force signal. To observe the foot force signal more accurately, a window function was employed. A window function is a mathematical construct that is zero-valued outside of selected intervals. In this research, 10, 15, and 30 s window sizes were used. The aim of the time-windowing process was to obtain shorter signal data. In the clinical application, this data collection is more convenient for the PD patient and, furthermore, reduces the fall risk. The possibility of patient injury rises if the data collection time is longer. Normalization and zero-mean processing were also used, to reduce the redundancy and dependency of data.
In 1987, Nilsson and Thorstensson observed the adaptability in the frequency and amplitude of leg movements during human locomotion at different speeds [26]. They reported that the overall range of stride frequency for normal leg movements is 0.83–1.95 Hz. The stride cycle period is defined as the time from the heel contact of one foot with the ground to the next heel contact of the same foot with the ground. The stride cycle period can be derived from the vGRF signal, and the stride frequency is the inverted value of the stride cycle duration. In conclusion, we selected two frequency ranges, 0.83–1.95 Hz and 1.95–50 Hz, for detailed observations of vGRF spectrograms among CO and PD subjects.

2.3. The Continuous Wavelet Transform

The continuous wavelet transform (CWT) is a signal processing tool used to observe the time–frequency spectrum characteristics of non-stationary signals [27]. As in the case of the Gabor transform [28], a CWT can be used to filter a signal using a dilated version of the mother wavelet, but the frequency translation is affected by dilation (scaling) and contraction. A CWT is a time–frequency transformation method, because it changes the signal time domain to the time–frequency domain. The output of a CWT is a time–frequency spectrogram (time–scale representation), which provides valuable information about the relationship between time and frequency.
A CWT consists of a time series function x t L 2 ( R ) , with a scaling or dilation factor s R +   ( s > 0 ) that controls the width of the wavelet, and a translation parameter, τ , controlling the location of the wavelet as expressed in the following equation:
X w s , τ = 1 s x t ψ * t τ s d t
where ψ t is a mother wavelet, also called a window function. The mother wavelet function used in this research was a Morlet or Gabor wavelet. This wavelet function consists of a Gaussian-windowed complex sinusoid (a complex exponential multiplied by a Gaussian window) as follows:
ψ ω 0 t = e i f t e 1 2 f 2 e 1 2 t 2
The parameter t refers to the time, and f represents the reference frequency.
The aim of the time–frequency transformation is to represent the vGRF signal (Figure 3a) as a time–frequency spectrogram image, as shown in Figure 3b,c, Figure 4 and Figure 5. The images clearly show different patterns of vGRF between CO and PD subjects that cannot be found in the time and frequency domains of the signal. Using the time–frequency spectrogram, variations in the foot pressure signal caused by temporal characteristics can also be analyzed. Temporal characteristics, also known as spatial characteristics or linear gait variabilities, consist of the measurements of step length, the stance width, the length of the step rhythm, and the step velocity.

2.4. Principal Component Analysis

The main goal of principal component analysis (PCA) is to perform dimension reduction for a dataset containing a large number of interrelated variables while, to the greatest extent possible, retaining the variations present in the dataset [29]. This reduction is achieved by transforming the dataset into a new set of variables, the principal components (PCs), which are ordered, de-correlated variables.
The PCA method is defined mathematically using the following steps. Consider a matrix X = P 1 ; P 2 ; P 3 ; ; P 2 T constructed using the spectrogram images of PDs and COs, where P is a row vector consisting of the pixels of a spectrogram image of PDs or COs, and i is the number of spectrogram images of PDs and COs. The PC is built using the equation X T X , a covariance matrix of the matrix X, to determine its eigenvalues and eigenvectors. The W matrix, an m × m matrix of weights whose columns are the eigenvectors of X T X , is obtained. Finally, the matrix for extracted feature F can be described as the full PCs’ decomposition of X , as shown in the following equation: F = X W .
The purpose of using PCA for feature enhancement was to extract fewer patterns while identifying the most important texture and pattern features. This processing was conducted in order to improve the performance of machine learning and artificial intelligence algorithms used for classifying the data points. The full PCs of each spectrogram image sample were selected to preserve the important texture and pattern features for visualization.

2.5. Convolutional Neural Network

A convolutional neural network (CNN) is composed of one or more convolutional layers, often with subsampling and pooling layers, followed by one or more fully-connected layers, as in a basic multi-layer neural network [30]. CNN was utilized to distinguish the time–frequency spectrogram representation of vGRF between PD severity stages. The convolutional layer plays the most important role in CNN performance. This layer is composed of a set of kernels (learnable filters) as parameters, which contain a small receptive field but are expanded through the full depth of the input. When the data pass through this layer, each kernel is convolved across the spatial dimensionality of the input (width and height of the input volume), resulting in the calculation of the dot product and production of a 2D activation map. The filters in the convolutional layers are edge detectors and color filters. An activation layer utilizes a non-saturating activation function f x = max 0 , x , such as a sigmoid function, in which σ x = 1 + e x 1 , to generate the output from the input produced by the previous layer. Another important concept in CNNs is pooling, also known as non-linear down-sampling. The aim of the pooling layer is to reduce the dimensionality and minimize the number of parameters and the complexity of model computation. This layer, known as the max-pooling layer, takes the input of each activation map and scales the input dimension using the MAX function. Finally, the fully connected layers attempt to generate scores from the previous activations to use for classification, as in traditional artificial neural networks (ANNs). Neurons in this layer have connections to all of the outputs of the previous layer. The performance of AlexNet, ResNet-50, ResNet-101, and GoogLeNet was examined in this study.

2.5.1. AlexNet CNN

The AlexNet architecture [31] comprises 25 layers, including an input layer, 5 convolution 2D layers, 7 rectified linear unit (ReLU) layers, 2 cross-channel normalization layers, 3 max-pooling 2D layers, 3 fully connected layers, 2 dropout layers for regularization, a softmax layer using a normalized exponential function, and an output layer. The input to the AlexNet CNN in the proposed method is a time–frequency spectrogram of the vGRF signals produced by the CWT. There are two methods for fine-tuning a pretrained AlexNet CNN: transfer learning and feature extraction. We chose the feature extraction method because it is easy to apply to pretrained networks without expending a lot of time, as it is faster than the transfer learning method and requires less training. This method applies two previous fully connected layers and uses a support vector machine (SVM) for classification.

2.5.2. ResNet-50 and ResNet-101 CNN

The main idea behind a residual network (ResNet) [32] is the presentation of a so-called “identity shortcut connection” that skips one or more layers. A shortcut (or skip) connection is used to solve the problem of vanishing or exploding gradients by using blocks that re-route the input and add to the concept learned in the previous layer. During learning, a layer learns the concepts of the previous layer and merges with inputs from that previous layer. ResNet-X refers to a residual deep neural network with X number of layers; for example, ResNet-50 indicates a ResNet developed using 50 layers. The architectures of ResNet-50 and ResNet-101 are described in Table 3.

2.5.3. GoogLeNet CNN

GoogLeNet [33] is a pretrained CNN that has 22 layers with 9 inception layers. An inception layer determines the optimal local sparse structure in a convolutional vision network, which can be approximated and covered by readily available dense components. In general, the inception layer is a network consisting of parallel convolutions of different sizes and types ( 1 × 1 , 3 × 3 , and 5 × 5 ) for the same input, which stacks all of the outputs. The exact structure of GoogLeNet is as follows:
  • An average pooling layer with a 5 × 5 filter size and a stride of 3.
  • A 1 × 1 convolution with 128 filters for dimension reduction and rectified linear activation.
  • A fully connected layer with 1024 units and rectified linear activation.
  • A dropout layer with 70% rate of dropped outputs.
  • A linear layer with softmax loss as the classifier.
Although AlexNet, ResNet-50, ResNet-101, and GoogLeNet achieved significant performance in the PD severity detection (overall accuracy ~97%), their architecture characteristics exhibited different influences on performance based on the benefits and drawbacks of the networks. The advantages and disadvantages of AlexNet, ResNet-50, ResNet-101, and GoogLeNet applied in the proposed method are summarized in Table 4.

2.6. Cross-Validation

Cross-validation is a statistical method used to assess and compare learning algorithms by dividing data into two groups: a training set used to train a model and a testing set used to test the model [36]. The training and testing sets are varied in consecutive rounds so that each data point is tested using a classifier in whose training it did not participate. There are two main purposes of using cross-validation. Cross-validation is used to quantify the generalizability of an algorithm, by testing the classifier on unseen data. The second purpose is to evaluate the performance of different algorithms and identify the best algorithm with which to classify the available data or, alternatively, to compare the performance of two or more variants of a parameterized model. In order to compare the results with the existing literature, k-fold cross-validation was utilized. Consequently, k iterations of training and testing were carried out in such a way that within each iteration, a different fold of the dataset was used for testing, while the remaining (k-1) folds were used for training. In this research, 10-fold cross-validation was applied.

3. Results

The experiments were carried out using MATLAB R2018a software on an NVIDIA GeForce GTX 1060 6 GB computer with 24 GB RAM. The computation time is affected by the number of input time–frequency spectrogram images (related to the time-windowing process, where a smaller time window will result in more images and longer computation time) and the number of neurons in the CNN. We employed multi-class classification for the COs and PD Stages 2, 2.5, and 3. This approach is representative of real-life applications, because doctors and neurologists do not have preliminary information about whether a patient is healthy or suffers from PD and, if the latter, what the severity is.
The sensitivity, specificity, accuracy, and AUC value of the proposed method were included as parameters for evaluation. The detailed definition of each evaluation parameter is provided in [37]. When selecting between diagnostic tests, Youden’s index is often applied to evaluate the effectiveness of the test [38]. Youden’s index is a function of sensitivity and specificity, and its value ranges between 0 and 1. A value close to 1 indicates that the diagnostic test’s effectiveness is relatively high and the test is close to perfect, and a value close to 0 indicates poor effectiveness, where the test is useless. Youden’s index (J) is the sum of the two fractions and indicates whether the measurements correctly diagnosed the diseased group (sensitivity) and healthy controls (specificity) over all cut-points c , < c < :
J = m a x c sensitivity c + specificity c 1

3.1. PD Severity Classification of Separated Ga, Ju, and Si Datasets

The gaitpdb database [8] contains three different vGRF datasets based on different studies: the Ga dataset describes dual tasking in PD patients, the Ju dataset indicates rhythmic auditory stimulation (RAS) in PD patients, and the Si dataset represents treadmill walking in PD patients. There are 29 PD patients and 18 CO subjects in the Ga dataset, 29 PD patients and 26 CO subjects in the Ju dataset, and 35 PD patients and 29 CO subjects in the Si dataset. The input signal for the proposed algorithm was dependent on the window size during the time-windowing process. For the 10 s window, the input signal numbers for CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 in the Ga, Ju, and Si datasets were 447, 492, 240, and 168; 199, 352, 460, and 109; and 348, 336, and 84, respectively. In the 15 s time window, the input signal numbers for CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 in the Ga, Ju, and Si datasets were 297, 328, 160, and 112; 129, 229, 300, and 72; and 232, 224, and 56, respectively. In the 30 s time window, the input signal numbers for CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 in the Ga, Ju, and Si datasets were 147, 164, 80, and 56; 58, 103, 135, and 35; and 116, 112, and 28, respectively.
The proposed method covered two kinds of classifications, multi-class (CO vs. PD Stage 2 vs. PD Stage 2.5 vs. PD Stage 3) classification and two-class (CO vs. PD) classification. In the two-class classification, PD Stage 2, 2.5, and 3 datasets were combined into one PD dataset. The best classification performance was obtained using AlexNet CNN for multi-class classification and ResNet CNN for two-class classification. The best classification result of the Ga dataset has 98.15% sensitivity, 98.16% specificity, 98.16% accuracy, and an AUC value of 0.9816 on average for multi-class classification and 99.77% sensitivity, 98.80% specificity, 99.11% accuracy, and an AUC value of 0.9995 for two-class classification. The best classification result of the Ju dataset has 98.06% sensitivity, 98.38% specificity, 98.24% accuracy, and an AUC value of 0.9822 on average for multi-class classification and 98.94% sensitivity, 99.04% specificity, 99.01% accuracy, and an AUC value of 0.9993 for two-class classification. The best classification result for the Si dataset has 97.73% sensitivity, 98.76% specificity, 98.27% accuracy, and an AUC value of 0.9825 on average for multi-class classification and 98.85% sensitivity, 98.41% specificity, 98.56% accuracy, and an AUC value of 0.9964 for two-class classification. Based on these classification results, the performance of the proposed method was not influenced by different datasets in the database, even though the data collection processes varied among these datasets.

3.2. PD Severity Classification of All Datasets (Merged)

For this classification, the three vGRF datasets in gaitpdb were merged and used as inputs to the proposed PD severity classification algorithm. For the 10 s, 15 s, and 30 s time windows, the input signal numbers for CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 were 994, 1180, 784, and 277; 658, 781, 516, and 184; and 321, 379, 243, and 91, respectively. The best result for this classification type was obtained using the ResNet CNN, with 92.08% sensitivity, 95.60% specificity, 94.58% accuracy, and an AUC value of 0.9384 on average for multi-class classification and 94.46% sensitivity, 97.69% specificity, 96.63% accuracy, and an AUC value of 0.9949 for two-class classification. The complete classification results are shown in Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12, Table 13, Table 14, Table 15 and Table 16 for multi-class and Table 17, Table 18 and Table 19 for two-class, and Table 20, Table 21, Table 22 and Table 23 summarizes the classification results.

4. Discussion

In this section, we discuss the gait analysis for each severity stage of PD based on the time and frequency analyses of the time–frequency spectrograms. Some key features of a signal are difficult to observe with the naked eye, but time–frequency spectrogram analysis can help to decipher important information regarding time and frequency characteristics. A CWT was used in this study to transform the signal from the time domain into the time–frequency domain. The gait phenomena could be identified using pattern visualization and recognition based on time–frequency spectrograms for CO subjects and PD patients with severity stages of 2, 2.5, and 3.
This observation was only performed for the CF vGRF signal. Since this type of input signal is the additive force between the left and right foot force signals, it describes the correlations between the features of the left and right feet instead of a single feature of the left or right foot. In order to further investigate the gait phenomena, a 10 s time window spectrogram was selected because the image feature was derived from a shorter input signal, and more detail can be perceived from the texture and pattern visualization of gait phenomena. For a 15 and 30 s time window spectrogram, the texture and pattern information is more compressed, and thus, the gait phenomena are blurred and not easily observed (see Figure 4 and Figure 5). The 0.1–5 Hz and 5–50 Hz frequency ranges were only applied to the detailed observations of the CWT time–frequency spectrogram and were not used for the classification.

4.1. Healthy Controls

Normal gait phenomena were interpreted by observing the time–frequency spectrogram of CO subjects, as shown in the first column of Figure 3. In the 0.1–5 Hz frequency range (Figure 3, first column, second row), the strongest walking force magnitude, represented in yellow, of the normal gait occurs at 1.6–2.1 Hz and is stable from the initial time to the end of the experiment. The foot force distributions and walking velocities for normal subjects were therefore the same when they were walking. At 2.5–3 Hz and approximately 4.5–5 Hz, small areas signifying the lowest force magnitude, shown in dark blue, alternate with a significant force magnitude, indicated by light blue, forming a regular pattern. This phenomenon appears in the spectrogram and is caused by the CF force signal at the lowest magnitudes. The three lowest magnitudes can be observed in one cycle of the CF force time domain signal (top left of Figure 3); the lowest magnitudes are almost equal in every cycle of the signal. The lowest magnitudes that occur at the beginning and end of the half gait cycle (that is, only the left or right foot gait cycle), close to the 0 force unit, show that the toe-off and initial contact and the lowest magnitude that occur between the half gait cycle are demonstrated only when one foot is in contact with the ground.
In the 5–50 Hz frequency range (Figure 3, first column, third row), a steady, strong force level, represented in yellow, occurs at approximately 5 Hz, with the same magnitude as that which occurs during walking, from the beginning to the end of the recording, and a significant force magnitude, shown in light blue, occurs up to 50 Hz in all records. Both time–frequency spectrograms indicate that the time and frequency components in the spectrogram have a regular pattern. This interpretation became a benchmark for investigating PD gait phenomena. These data were compared to analyze the gait characteristics of PD patients based on spectrogram analyses.

4.2. Parkinson’s Disease Stage 2

The time–frequency spectrograms for PD patients were similar to those of the CO spectrograms. For PD Stage 2 patients, as presented in the second column of Figure 3, the strongest force is at 1.6–2.1 Hz in the 0.1–5 Hz (Figure 3, second column, second row) frequency range, and there is a significant, strong magnitude, shown in light yellow, at 1 Hz, which is weaker than the force magnitude at 1.6–2.1 Hz. The significant force magnitude at 2.5–3 Hz and approximately 4.5–5 Hz becomes more yellow instead of light blue as in the CO spectrogram. It is also apparent that the pattern of the lowest force magnitude at 2.5–5 Hz is regular at some times and irregular at other times. This observation indicates that the magnitudes of the global and local minima are not the same in every gait cycle (Figure 3, second column, first row). In the time domain, the CF vGRF signal has fluctuating force magnitudes that cause an irregularity in the signal.
In the 5–50 Hz frequency range (Figure 3, second column, third row), the strongest force magnitude, shown in yellow, is about 5 Hz, and significant force, represented by light blue, occurs up to 50 Hz every time. However, the force magnitude is not distributed equally over the entire walking period.

4.3. Parkinson’s Disease Stage 2.5

As shown in the third column of Figure 3, the spectrogram for PD Stage 2.5 patients is not very different from the PD Stage 2 spectrogram in either frequency range. The only difference is that, in the 0.1–5 Hz frequency range, a significant, strong magnitude at 1 Hz becomes stronger, and yellow areas of force magnitude appear in the image. PD patients in the early stages—2 and 2.5—of the disease can have a walking velocity similar to that of COs, but their force distribution is typically not equally distributed, due to the presence of tremors.

4.4. Parkinson’s Disease Stage 3

Of the patients studied in this research, those with PD Stage 3 had the most severe level of disease. The spectrograms of this group exhibit the most irregular patterns of all severity levels. In the fourth column, first row of Figure 3, the CF vGRF signal has the most fluctuation and irregular force magnitudes because of the jerky movements and tremors of the patients.
In the 0.1–5 Hz frequency range (third column, second row of Figure 3), the strongest walking force magnitude, shown in yellow, occurs at a lower frequency than in stages 2 and 2.5 at 1–1.5 Hz. A significant strong force magnitude, depicted in light yellow, also appears at approximately 0.75 Hz, although the force level is not the same in every gait cycle. At 2–3 Hz and 3.5–4 Hz, significant force magnitude regions occur, as shown by colors that are more yellow.
In the 5–50 Hz frequency range (third column, third row of Figure 3), the strongest force magnitude only appears in certain gait cycles and is not equally distributed. A significant force magnitude, shown in light blue, only occurs up to 20 Hz, and forms an irregular pattern in every gait cycle.

4.5. Comparison of Results with the Existing Literature

A comparison between the proposed methodology and a study by Zhao et al. [14] is presented in Table 24. The authors carried out multi-class classification of vGRF signals for CO vs. PD Stage 2 vs. PD Stage 2.5 vs. PD Stage 3 using the same information found in the same database used for the proposed method, gaitpdb. These authors separated the classification types based on the three datasets—Ga, Ju, and Si—and used 10-fold cross-validation as the evaluation method. The two-class classification results were also compared with those of studies conducted by Maachi et al. [39], Wu et al. [40], Ertugrul et al. [41], Zeng et al. [42], Daliri [43], and Khoury et al. [44,45]. These comparison results are shown in Table 25 and Table 26. In Khoury et al.’s study, the classification types were divided based on the three datasets—Ga, Ju, and Si.
In summary, the proposed method produced almost the same classification results as those published in the existing literature, but the proposed algorithm generated better visualizations via time–frequency spectrograms associated with the progression of PD severity. The irregularity in patterns in the spectrograms is proportional to the severity level of PD. The more severe the disease, therefore, the more irregular the spectrogram’s pattern. This phenomenon could be helpful for medical specialists or neurologists in monitoring PD progression, allowing them to provide more effective and accurate medications and therapies to patients.

5. Conclusions

In this study, a deep learning algorithm was implemented based on vGRF time–frequency features for the detection and severity classification of Parkinson’s disease. Pattern visualization and recognition of the time–frequency spectrogram made it possible to successfully differentiate PD severity stages and COs. A CWT was used to generate spectrograms to visualize gait foot force signals by transforming signals from the time domain into the time–frequency domain. Three time-window sizes (10, 15, and 30 s), two frequency ranges (0.83–1.95 and 1.95–50 Hz), and three types of gait foot force signals (LF, RF, and CF force signals) were selected as inputs to obtain good feature visualization. After the original signal was transformed, PCA was applied for feature enhancement, to increase between-class separability and to reduce within-class separability. Finally, CNNs were used to perform classification. To evaluate the CNNs’ classification process, 10-fold cross-validation was performed, and the accuracy, sensitivity, specificity, and the AUC value were evaluated. The proposed method was able to achieve the highest performance for more than 97.42% of the parameters being evaluated and achieved superior performance in comparison with the detection and PD severity classification performance of state-of-the-art methods found in the literature.
Although the evidence indicates that the proposed method achieved good performance, there are several major drawbacks that could be improved. First, an existing database was used with the proposed method, and clinical data with a greater number of severity levels should be used to verify the performance and to resolve the limitation of the relatively small number of PD patients at certain severity levels in the current database. Clinical data collection will be carried out using a smart insole with an embedded 0.5” force sensing resistor of our own design. The precision and accuracy of force sensing resistor readings are also considered in order to obtain the correct representation of the vGRF signal. PD patients will be asked to perform some simple daily activities, such as turning around and sitting, instead of only walking down a long pathway. Second, long-term data collection to monitor PD progression is important for treatment decisions, since the gait patterns of PD patients appear to change with the long-term progression of the disease. Third, to further investigate the clinical meaning of the results, PD gait phenomena based on time–frequency spectrograms should be discussed with physicians. Fourth, other input data, such as kinetic data, temporal data, step length, and cadence, and other classifiers should be used to confirm and compare the effectiveness of pattern visualization and recognition based on the use of time–frequency spectrograms in PD detection.

Author Contributions

Conceptualization, F.S. and C.-W.L.; methodology, F.S. and C.-W.L.; software, F.S.; validation, F.S.; investigation, F.S. and C.-W.L.; resources, C.-W.L.; writing—original draft preparation, F.S.; writing—review and editing, C.-W.L.; supervision, C.-W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology (Taiwan), grant number 108-2628-E-006-003-MY3.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, A.; Gilbert, R. Epidemiology of Parkinson Disease. Neurol. Clin. 2016, 34, 955–965. [Google Scholar] [CrossRef]
  2. Parkinson’s Disease Foundation. Statistics on Parkinson’s EIN: 13-1866796. 2018. Available online: https://bit.ly/2RCeh9H (accessed on 12 July 2019).
  3. National Institute of Neurological Disorders and Stroke. Parkinson’s Disease Information Page. 2016. Available online: https://bit.ly/2xTA6rL (accessed on 12 July 2019).
  4. Hoff, J.I.; Plas, A.A.; Wagemans, E.A.H.; Van Hilten, J.J. Accelerometric assessment of levodopa-induced dyskinesias in Parkinson’s disease. Mov. Disord. 2001, 16, 58–61. [Google Scholar] [CrossRef]
  5. Pistacchi, M.; Gioulis, M.; Sanson, F.; De Giovannini, E.; Filippi, G.; Rossetto, F.; Marsala, S.Z. Gait analysis and clinical correlations in early Parkinson’s disease. Funct Neurol. 2017, 32, 28. [Google Scholar] [CrossRef] [PubMed]
  6. Sofuwa, O.; Nieuwboer, A.; Desloovere, K.; Willems, A.-M.; Chavret, F.; Jonkers, I. Quantitative Gait Analysis in Parkinson’s Disease: Comparison with a Healthy Control Group. Arch. Phys. Med. Rehabil. 2005, 86, 1007–1013. [Google Scholar] [CrossRef] [Green Version]
  7. Lescano, C.N.; Rodrigo, S.E.; Christian, D.A. A possible parameter for gait clinimetric evaluation in Parkinson’s disease patients. J. Phys. Conf. Ser. 2016, 705, 12019. [Google Scholar] [CrossRef]
  8. Goldberger, A.L.; Amaral, L.A.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a New Research Resource for Complex Physiologic Signals. Circulation 2003, 101, e215–e220. [Google Scholar] [CrossRef] [Green Version]
  9. Arafsha, F.; Hanna, C.; Aboualmagd, A.; Fraser, S.; El Saddik, A. Instrumented Wireless SmartInsole System for Mobile Gait Analysis: A Validation Pilot Study with Tekscan Strideway. J. Sens. Actuator Netw. 2018, 7, 36. [Google Scholar] [CrossRef] [Green Version]
  10. Hoehn, M.M.; Yahr, M.D. Parkinsonism: Onset, progression, and mortality. Neurology 1967, 17, 427. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Jane, Y.N.; Nehemiah, H.K.; Arputharaj, K. A Q-backpropagated time delay neural network for diagnosing severity of gait disturbances in Parkinson’s disease. J. Biomed. Inform. 2016, 60, 169–176. [Google Scholar] [CrossRef]
  12. Vasquez-Correa, J.C.; Arias-Vergara, T.; Orozco-Arroyave, J.R.; Eskofier, B.M.; Klucken, J.; Noth, E. Multimodal Assessment of Parkinson’s Disease: A Deep Learning Approach. IEEE J. Biomed. Health Inform. 2018, 23, 1618–1630. [Google Scholar] [CrossRef]
  13. Lee, S.-H.; Lim, J.S. Parkinson’s disease classification using gait characteristics and wavelet-based feature extraction. Expert Syst. Appl. 2012, 39, 7338–7344. [Google Scholar] [CrossRef]
  14. Zhao, A.; Qi, L.; Li, J.; Dong, J.; Yu, H. A hybrid spatio-temporal model for detection and severity rating of Parkinson’s disease from gait data. Neurocomputing 2018, 315, 1–8. [Google Scholar] [CrossRef] [Green Version]
  15. Hoffmann, A.G. General Limitations on Machine Learning. In Proceedings of the 9th European Conference on Artificial Intelligence, Stockholm, Sweden, 1 January 1990; pp. 345–347. [Google Scholar]
  16. Mastorakis, G. Human-like machine learning: Limitations and suggestions. arXiv 2018, arXiv:1811.06052. [Google Scholar]
  17. Jiang, X.; Napier, C.; Hannigan, B.; Eng, J.J.; Menon, C. Estimating Vertical Ground Reaction Force during Walking Using a Single Inertial Sensor. Sensors 2020, 20, 4345. [Google Scholar] [CrossRef]
  18. Lin, C.-W.; Wen, T.-C.; Setiawan, F. Evaluation of Vertical Ground Reaction Forces Pattern Visualization in Neurodegenerative Diseases Identification Using Deep Learning and Recurrence Plot Image Feature Extraction. Sensors 2020, 20, 3857. [Google Scholar] [CrossRef]
  19. Setiawan, F.; Lin, C.-W. Identification of Neurodegenerative Diseases Based on Vertical Ground Reaction Force Classification Using Time–Frequency Spectrogram and Deep Learning Neural Network Features. Brain Sci. 2021, 11, 902. [Google Scholar] [CrossRef]
  20. Alam, N.; Garg, A.; Munia, T.T.K.; Fazel-Rezai, R.; Tavakolian, K. Vertical ground reaction force marker for Parkinson’s disease. PLoS ONE 2017, 12, e0175951. [Google Scholar] [CrossRef]
  21. Yogev, G.; Giladi, N.; Peretz, C.; Springer, S.; Simon, E.S.; Hausdorff, J.M. Dual tasking, gait rhythmicity, and Parkinson’s disease: Which aspects of gait are attention demanding? Eur. J. Neurosci. 2005, 22, 1248–1256. [Google Scholar] [CrossRef]
  22. Hausdorff, J.M.; Lowenthal, J.; Herman, T.; Gruendlinger, L.; Peretz, C.; Giladi, N. Rhythmic auditory stimulation modulates gait variability in Parkinson’s disease. Eur. J. Neurosci. 2007, 26, 2369–2375. [Google Scholar] [CrossRef]
  23. Frenkel-Toledo, S.; Giladi, N.; Peretz, C.; Herman, T.; Gruendlinger, L.; Hausdorff, J.M. Treadmill walking as an external pacemaker to improve gait rhythm and stability in Parkinson’s disease. Mov. Disord. 2005, 20, 1109–1114. [Google Scholar] [CrossRef]
  24. Frenkel-Toledo, S.; Giladi, N.; Peretz, C.; Herman, T.; Gruendlinger, L.; Hausdorff, J.M. Effect of gait speed on gait rhythmicity in Parkinson’s disease: Variability of stride time and swing time respond differently. J. Neuroeng. Rehabil. 2005, 2, 23. [Google Scholar] [CrossRef] [Green Version]
  25. Fahn, S.; Marsden, C.D.; Calne, D.B.; Goldstein, M. Recent Developments in Parkinson’s Disease. Florham Park N. J. Macmillan Health Care Inf. 1987, 2, 153–164. [Google Scholar]
  26. Nilsson, J.; Thorstensson, A. Adaptability in frequency and amplitude of leg movements during human locomotion at different speeds. Acta Physiol. Scand. 1987, 129, 107–114. [Google Scholar] [CrossRef]
  27. Sadowsky, J. The continuous wavelet transform: A tool for signal investigation and understanding. Johns Hopkins APL Tech. Dig. 1994, 15, 306. [Google Scholar]
  28. Bernardino, A.; Santos-Victor, J. A real-time gabor primal sketch for visual attention. In Iberian Conference on Pattern Recognition and Image Analysis; Springer: Berlin/Heidelberg, Germany, 2005; pp. 335–342. [Google Scholar]
  29. Jolliffe, I.T. Introduction. In Principal Component Analysis, 2nd ed.; Springer: New York, NY, USA, 2002. [Google Scholar]
  30. O’Shea, K.; Nash, R. An introduction to convolutional neural networks. arXiv 2015, arXiv:1511.08458. [Google Scholar]
  31. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  32. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  33. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef] [Green Version]
  34. Khan, A.; Sohail, A.; Zahoora, U.; Qureshi, A.S. A survey of the recent architectures of deep convolutional neural networks. Artif. Intell. Rev. 2020, 53, 5455–5516. [Google Scholar] [CrossRef] [Green Version]
  35. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 1–74. [Google Scholar] [CrossRef]
  36. Refaeilzadeh, P.; Tang, L.; Liu, H. Cross-Validation. Encycl. Database Syst. 2009, 532–538. [Google Scholar] [CrossRef]
  37. Fawcett, T. An introduction to ROC analysis. Pattern Recognit. Lett. 2006, 27, 861–874. [Google Scholar] [CrossRef]
  38. Youden, W.J. Index for rating diagnostic tests. Cancer 1950, 3, 32–35. [Google Scholar] [CrossRef]
  39. El Maachi, I.; Bilodeau, G.-A.; Bouachir, W. Deep 1D-Convnet for accurate Parkinson disease detection and severity prediction from gait. Expert Syst. Appl. 2020, 143, 113075. [Google Scholar] [CrossRef]
  40. Wu, Y.; Chen, P.; Luo, X.; Wu, M.; Liao, L.; Yang, S.; Rangayyan, R. Measuring signal fluctuations in gait rhythm time series of patients with Parkinson’s disease using entropy parameters. Biomed. Signal Process. Control. 2017, 31, 265–271. [Google Scholar] [CrossRef]
  41. Ertuğrul, Ö.F.; Kaya, Y.; Tekin, R.; Almali, M.N. Detection of Parkinson’s disease by Shifted One Dimensional Local Binary Patterns from gait. Expert Syst. Appl. 2016, 56, 156–163. [Google Scholar] [CrossRef]
  42. Zeng, W.; Liu, F.; Wang, Q.; Wang, Y.; Ma, L.; Zhang, Y. Parkinson’s disease classification using gait analysis via deterministic learning. Neurosci. Lett. 2016, 633, 268–278. [Google Scholar] [CrossRef] [PubMed]
  43. Daliri, M.R. Chi-square distance kernel of the gaits for the diagnosis of Parkinson’s disease. Biomed. Signal Process. Control. 2013, 8, 66–70. [Google Scholar] [CrossRef]
  44. Khoury, N.; Attal, F.; Amirat, Y.; Oukhellou, L.; Mohammed, S. Data-Driven Based Approach to Aid Parkinson’s Disease Diagnosis. Sensors 2019, 19, 242. [Google Scholar] [CrossRef] [Green Version]
  45. Khoury, N.; Attal, F.; Amirat, Y.; Chibani, A.; Mohammed, S. CDTW-Based Classification for Parkinson’s Disease Diagnosis. In Proceedings of the 26th European Symposium on Artificial Neural Networks, Bruges, Belgium, 25–27 April 2018. [Google Scholar]
Figure 1. Description of beginning and ending of different gait phases in a normal gait cycle (Arafsha et al. [9]).
Figure 1. Description of beginning and ending of different gait phases in a normal gait cycle (Arafsha et al. [9]).
Sensors 21 05207 g001
Figure 2. Flowchart of the proposed PD detection and severity classification algorithm using the continuous wavelet transform as the feature transformation.
Figure 2. Flowchart of the proposed PD detection and severity classification algorithm using the continuous wavelet transform as the feature transformation.
Sensors 21 05207 g002
Figure 3. Time-frequency spectrograms using CWT of the first 10 s CF vGRF signals of CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 subjects in a 10 s time window size from Ga [21] dataset: (a) original vGRF signal, (b) 0.1–5 Hz time-frequency spectrogram, and (c) 5–50 Hz time-frequency spectrogram.
Figure 3. Time-frequency spectrograms using CWT of the first 10 s CF vGRF signals of CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 subjects in a 10 s time window size from Ga [21] dataset: (a) original vGRF signal, (b) 0.1–5 Hz time-frequency spectrogram, and (c) 5–50 Hz time-frequency spectrogram.
Sensors 21 05207 g003
Figure 4. Time-frequency spectrograms using CWT of the CF vGRF signals of CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 subjects in the 0.1–5 Hz frequency range for (a) a 10 s time window size, (b) a 15 s time window size, and (c) a 30 s time window size from Ga [21] dataset.
Figure 4. Time-frequency spectrograms using CWT of the CF vGRF signals of CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 subjects in the 0.1–5 Hz frequency range for (a) a 10 s time window size, (b) a 15 s time window size, and (c) a 30 s time window size from Ga [21] dataset.
Sensors 21 05207 g004
Figure 5. Time-frequency spectrograms using CWT of the CF vGRF signals of CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 subjects in the 5–50 Hz frequency range for (a) a 10 s time window size, (b) a 15 s time window size, and (c) a 30 s time window size from Ga [21] dataset.
Figure 5. Time-frequency spectrograms using CWT of the CF vGRF signals of CO, PD Stage 2, PD Stage 2.5, and PD Stage 3 subjects in the 5–50 Hz frequency range for (a) a 10 s time window size, (b) a 15 s time window size, and (c) a 30 s time window size from Ga [21] dataset.
Sensors 21 05207 g005
Table 1. Hoehn and Yahr (HY) scale [10] for PD severity stage.
Table 1. Hoehn and Yahr (HY) scale [10] for PD severity stage.
StagesDescription
0No signs of disease
1Symptoms are very mild; unilateral involvement only
1.5Unilateral and axial involvement
2Bilateral involvement without impairment of balance
2.5Mild bilateral disease with recovery on pull test
3Mild to moderate bilateral disease; some postural instability; physical independence
4Severe disability; still able to walk or stand unassisted
5Wheelchair bound or bedridden unless aided
Table 2. Number of subjects in three sub-datasets of Parkinson’s Disease database (gaitpdb) [8] based on the HY rating scale of severity.
Table 2. Number of subjects in three sub-datasets of Parkinson’s Disease database (gaitpdb) [8] based on the HY rating scale of severity.
AuthorStage 0Stage 2Stage 2.5Stage 3
Ga [21]181586
Ju [22]2612134
Si [23]292960
Table 3. Architectures for ResNet-50 and ResNet-101.
Table 3. Architectures for ResNet-50 and ResNet-101.
Layer NameOutput Size50 Layer101 Layer
conv1 112 × 112 7 × 7 , 64 , stride 2
conv2_x 56 × 56 3 × 3 max pool, stride 2
1 × 1 ,   64 3 × 3 ,   64 1 × 1 ,   256 × 3 1 × 1 ,   64 3 × 3 ,   64 1 × 1 ,   256 × 3
conv3_x 28 × 28 1 × 1 ,   128 3 × 3 ,   128 1 × 1 ,   512 × 4 1 × 1 ,   128 3 × 3 ,   128 1 × 1 ,   512 × 4
conv4_x 14 × 14 1 × 1 ,   256 3 × 3 ,   256 1 × 1 ,   1024 × 6 1 × 1 ,   256 3 × 3 ,   256 1 × 1 ,   1024 × 23
conv5_x 7 × 7 1 × 1 ,   512 3 × 3 ,   512 1 × 1 ,   2048 × 3 1 × 1 ,   512 3 × 3 ,   512 1 × 1 ,   2048 × 3
1 × 1 average pool, 1000-d fc, softmax
FLOPs 3.8 × 10 9 7.6 × 10 9
Table 4. The advantages and disadvantages of the AlexNet, ResNet, and GoogLeNet CNN architecture [34,35].
Table 4. The advantages and disadvantages of the AlexNet, ResNet, and GoogLeNet CNN architecture [34,35].
ArchitectureAdvantageDisadvantage
AlexNet
layer depth: 8
parameters: 60 million
  • There is low feature loss, as the ReLU activation function does not limit the output.
  • Uses data enhancement, dropout, and normalization layers to prevent the network from overfitting and improve the model generalization.
  • This model has much less depth; hence, it struggles to learn features from image sets.
  • Takes more time to achieve higher accuracy (highest accuracy achieved: 99.11%).
ResNet-50
layer depth: 50
parameters: 25.6 million ResNet-101
layer depth: 101
parameters: 44.5 million
  • Decreased the error rate for deeper networks by introducing the idea of residual learning.
  • Instead of widening the network, the increased depth of the network results in fewer additional parameters. This greatly reduces the training time and improves accuracy (highest accuracy ResNet-50: 99.20%; highest accuracy ResNet-101: 99.01%).
  • Mitigates the effect of vanishing gradient.
  • A complex architecture
  • Many layers may provide very little or no information.
  • Redundant feature-maps may happen to be relearned.
GoogLeNet
layer depth: 22
parameters: 7 million
  • Computational and memory efficiency.
  • Reduced number of parameters by using bottleneck and global average pooling layer.
  • Use of auxiliary classifiers to improve the convergence rate.
  • Lower accuracy (highest accuracy: 98.77%).
  • Its heterogeneous topology necessitates adaptation from module to module.
  • Substantially reduces the feature space because of the representation bottleneck and thus sometimes may lead to loss of useful information.
Table 5. Multi-class classification of LF from Ga dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 5. Multi-class classification of LF from Ga dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 098.2198.1198.140.981697.5499.4498.810.984997.5499.5698.890.985597.5497.8997.770.9771
Class 293.2997.7896.140.955495.5397.6696.880.965997.1596.4996.730.968293.0997.8996.140.9549
Class 2.582.5096.1293.690.893191.2595.9395.100.935988.7597.2995.770.930294.1794.6794.580.9442
Class 383.9397.3795.690.906583.3398.9897.030.911688.1099.3297.920.937173.8199.4196.210.8661
1.95–50 HzClass 097.7698.5698.290.981695.3098.8997.700.971094.6398.4497.180.965497.0997.7897.550.9743
Class 295.93 *98.25 *97.40 *0.9709 *96.5496.2696.360.964096.3494.7495.320.955493.0997.6695.990.9538
Class 2.593.75 *98.83 *97.92 *0.9629 *91.6798.5597.330.951187.5098.0196.140.927690.4297.6596.360.9403
Class 396.4399.1598.810.977994.6499.2498.660.969489.2999.24980.942692.8698.6497.920.9575
15 s0.83–1.95 HzClass 097.9898.3398.220.981693.9498.5096.990.962295.2999.3397.990.973197.98 *99.17 *98.77 *0.9857 *
Class 294.2196.6695.760.954393.9095.0894.650.944997.2695.4396.100.963496.0496.8496.540.9644
Class 2.58597.2995.090.911488.7596.7495.320.92759098.5196.990.942588.7597.6996.100.9322
Class 390.1898.6097.550.943988.3998.9897.660.936995.5499.6299.110.975891.9699.2498.330.9560
1.95–50 HzClass 096.3098.3397.660.973193.9498.5096.990.962295.9696.3396.210.961595.969897.320.9698
Class 293.2997.5495.990.954293.9095.0894.650.944991.7795.7894.310.937894.5195.7895.320.9515
Class 2.588.1397.5695.880.928488.7596.7495.320.927589.3896.4795.210.929284.3896.7494.540.9056
Class 394.6497.9697.550.963088.3998.9897.660.936981.2599.3697.100.903184.8298.4796.770.9165
30 s0.83–1.95 HzClass 095.2498.3397.320.967995.9298.3397.540.971391.8499.6797.090.957594.5696.3395.750.9545
Class 293.9097.1795.970.955492.6895.4194.410.940495.7391.8793.290.938090.8594.7093.290.9278
Class 2.59597.5597.090.96279095.6494.630.92828594.8293.060.89918597.8295.530.9141
Class 394.6499.4998.880.970783.9399.7497.760.918476.7999.7496.870.882692.8698.7297.990.9579
1.95–50 HzClass 095.929897.320.969693.209997.090.961093.2097.3395.970.952795.2497.3396.640.9629
Class 293.2995.7694.850.945393.9094.3594.180.941293.3092.5893.060.932488.4194.3592.170.9138
Class 2.58098.3795.080.891886.2594.8293.290.905482.5095.9193.510.892176.2594.2891.050.8526
Class 398.21 *97.44 *97.54 *0.9783 *80.3698.9896.640.896776.7999.2396.420.880183.9397.7095.970.9081
Note: * selected by Youden’s index criteria as the best classification result.
Table 6. Multi-class classification of LF from Ju dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 6. Multi-class classification of LF from Ju dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 094.4799.0298.210.967596.98 *99.67 *99.20 *0.9833 *93.9799.0298.130.965097.4998.4898.300.9798
Class 293.7598.0596.700.959096.59 *98.44 *97.86 *0.9751 *95.1794.6694.820.949293.7595.9695.270.9486
Class 2.594.1394.8594.550.944998.2693.6495.540.959592.6195.6194.380.941193.2694.0993.750.9368
Class 380.7397.7396.070.892370.6499.9097.050.852780.7399.5197.680.901271.5699.5196.790.8553
1.95–50 HzClass 096.4898.7098.300.975994.4799.4698.570.969694.9798.9198.210.969492.4698.7097.590.9558
Class 292.9098.7096.880.958093.1897.0195.800.950995.4594.79950.951293.4796.3595.450.9491
Class 2.597.39 *96.52 *96.88 *0.9695 *96.5295.4595.890.959992.8396.9795.270.949093.7097.1295.710.9541
Class 390.8399.1198.300.949789.9199.6098.660.947687.1699.7098.480.934393.5898.5298.040.9605
15 s0.83–1.95 HzClass 093.8099.1798.220.964896.1298.8498.360.974893.8098.6797.810.962393.0299.5098.360.9626
Class 295.6397.6096.990.966291.279895.890.946492.1497.4195.750.947796.5195.8196.030.9616
Class 2.594.3396.9895.890.956696.6795.3595,890.960196.6794.8895.620.95789297.2195.070.9460
Class 388.8998.0297.120.934690.2899.2498.360.947683.3399.2497.670.912990.2898.1897.400.9423
1.95–50 HzClass 090.7098.5097.120.946087.6099.1797.120.933893.8098.1797.400.959889.929896.580.9396
Class 293.0197.0195.750.950193.8996.0195.340.949593.4594.6194.250.940388.2196.4193.840.9231
Class 2.594.6797.4496.300.960594.3396.0595.340.951991.6795.5893.970.936294.6793.9594.250.9431
Class 394.4498.0297.670.962388.8998.0297.120.934679.1799.0997.120.891386.1198.4897.260.9230
30 s0.83–1.95 HzClass 086.2198.5296.370.923791.3810098.490.956989.6610098.190.948384.4898.1795.770.9133
Class 292.2393.8693.350.930596.1296.4996.370.963096.1295.6195.770.958785.4494.7491.840.9009
Class 2.591.8596.9494.860.944096.3093.8894.860.950996.3089.2992.150.927996.3092.3593.960.9432
Class 397.14 *98.65 *98.49 *0.9790 *77.1499.3296.980.882351.4399.6694.560.755585.7199.3297.890.9252
1.95–50 HzClass 089.6697.0795.770.933679.3110096.370.896681.0398.5395.470.897882.7694.8792.750.8882
Class 283.5096.0592.150.897785.4493.4290.940.894390.2990.7990.630.905479.6192.9888.820.8630
Class 2.593.3393.3793.350.933595.5687.2490.630.914091.1187.7689.120.894389.6393.3791.840.9150
Class 391.4398.3197.580.948771.4398.8996.070.852148.5799.3293.960.739582.8697.3095.770.9008
Note: * selected by Youden’s index criteria as the best classification result.
Table 7. Multi-class classification of LF from Si dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 7. Multi-class classification of LF from Si dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 097.9998.1098.050.980497.7099.5298.700.986197.9997.3897.660.976897.1396.6796.880.9690
Class 294.6491.2092.710.929297.9295.6096.610.967695.8394.9195.310.953794.3592.8293.490.9358
Class 2.561.9098.3994.400.801586.9099.2797.920.930982.1499.5697.660.908573.8199.1296.350.8647
1.95–50 HzClass 098.56 *99.05 *98.83 *0.9881 *96.8499.0598.050.979496.2696.4396.350.963595.9898.5797.400.9727
Class 295.83 *98.84 *97.53 *0.9734 *97.3296.0696.610.966995.5493.5294.400.945394.6493.9894.270.9431
Class 2.598.81 *98.39 *98.44 *0.9860 *91.6799.1298.310.953980.9599.8597.790.904084.5298.1096.610.9131
15 s0.83–1.95 HzClass 097.4197.5097.460.974696.1299.2997.850.977096.5597.8697.270.972099.1497.1498.050.9814
Class 292.4194.7993.750.936096.8894.1095.310.954995.9895.1495.510.955693.3097.5795.700.9544
Class 2.582.1497.5995.900.898785.7198.9097.460.923187.5099.1297.850.933191.0798.4697.660.9477
1.95–50 HzClass 094.8398.5796.880.967093.5398.5796.290.960596.1297.5096.880.968196.1296.7996.480.9645
Class 293.3095.4994.530.943996.4392.3694.140.943995.0993.4094.140.942595.5494.4494.920.9499
Class 2.596.4397.3797.270.969085.7198.9097.460.923180.3698.9096.880.896385.7199.5698.050.9264
30 s0.83–1.95 HzClass 093.9796.4395.310.952093.9797.8696.090.959194.8396.4395.700.956395.6997.8696.8896.77
Class 292.8693.7593.360.933094.6493.7594.140.942095.5488.1991.410.918792.8693.7593.3693.30
Class 2.589.2998.2597.270.937789.2998.2597.270.937757.1499.5694.920.783582.1497.3795.7089.76
1.95–50 HzClass 097.4194.2995.700.958593.1098.5796.090.958494.839594.920.949193.1098.5796.090.9584
Class 288.3993.7591.410.910793.7589.5891.410.916790.1889.5889.8489.8891.0793.0692.190.9206
Class 2.57597.3794.920.861871.4397.3794.530.844064.2997.8194.1481.0592.8696.4996.090.9467
Note: * selected by Youden’s index criteria as the best classification result.
Table 8. Multi-class classification of RF from Ga dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 8. Multi-class classification of RF from Ga dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 098.8898.1198.370.985097.7699.1198.660.984498.43 *99 *98.81 *0.9872 *96.2098.89980.9754
Class 295.3395.5695.470.954494.9297.5496.590.962395.5397.1996.590.963693.5096.6195.470.9505
Class 2.576.2596.2192.650.862391.6797.0296.070.943489.5895.9394.800.927689.1794.9493.910.9205
Class 378.5797.9695.550.882790.4899.2498.140.948680.9599.3297.030.901479.7698.8196.440.8929
1.95–50 HzClass 097.7699.6799.030.987194.8598.8997.550.968796.2097.7897.250.969996.4297.5697.180.9699
Class 297.97 *97.78 *97.85 *0.9787 *96.1496.0296.070.960894.3196.2695.550.952890.0496.9694.430.9350
Class 2.589.1798.2896.660.937391.6798.2897.100.94989097.2995.990.936486.6794.8593.390.9076
Class 393.4598.64980.960594.05 *99.24 *98.59 *0.9664 *86.9099.2497.700.930782.1498.1396.140.9014
15 s0.83–1.95 HzClass 097.319998.440.981596.6399.5098.550.980795.969997.990.974898.9898.1798.100.9807
Class 293.9098.2496.660.960796.9597.0196.990.969896.0496.1396.100.960990.8596.6694.540.9376
Class 2.596.88 *96.88 *96.88 *0.9688 *92.5097.8396.880.951692.5097.1596.320.948387.5095.7994.310.9165
Class 391.9699.8298.660.957992.8699.4998.660.961788.3999.8798.440.941389.2999.1197.880.9420
1.95–50 HzClass 098.3298.3398.330.983295.2998.5097.440.968993.9498.6797.100.963095.9697.3396.880.9665
Class 294.2198.7797.100.964992.3897.0195.320.947095.4393.6794.310.945591.1694.7393.420.9294
Class 2.587.5097.9696.100.927390.6396.6195.540.936283.7595.9393.760.898478.1397.5694.090.8784
Class 394.6497.7197.320.961791.9698.4797.660.952279.4698.9896.540.892295.5497.8397.550.9669
30 s0.83–1.95 HzClass 096.6098.6797.990.976395.2499.6798.210.974594.5699.3397.760.969593.2098.3396.640.9577
Class 296.3496.4796.420.964096.3497.1796.870.967696.3492.5893.960.944695.7392.2393.510.9398
Class 2.588.7598.3796.640.935696.259796.870.966381.2594.2891.950.877677.5097.2893.740.8739
Class 392.8698.9898.210.959289.2999.4998.210.943969.6499.4995.750.845785.7198.4796.870.9209
1.95–50 HzClass 093.2097.3395.970.952791.8497.6795.750.947593.8897.6796.420.957793.8896.3395.530.9511
Class 292.0794.3593.510.932188.4194.3592.170.913891.4692.2391.950.918485.9893.2990.600.8963
Class 2.58097.5594.410.88779094.0193.290.920081.2592.6490.600.869573.7594.0190.380.8388
Class 392.8697.4496.870.951582.1498.9896.870.905664.2999.4995.080.818982.1497.1995.300.8966
Note: * selected by Youden’s index criteria as the best classification result.
Table 9. Multi-class classification of RF from Ju dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 9. Multi-class classification of RF from Ju dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 096.9899.3598.930.981795.9899.4698.840.977295.4899.5798.840.975294.9799.1398.390.9705
Class 294.8998.3197.230.966097.1697.7997.590.974797.73 *97.40 *97.50 *0.9756 *96.5995.7095.980.9615
Class 2.595.4394.70950.950796.3095.4595.800.958895.87 *96.36 *96.16 *0.9612 *94.1392.1292.950.9313
Class 378.9098.6296.700.887677.0699.2197.050.881481.6599.2197.500.904356.8899.7095.540.7829
1.95–50 HzClass 097.99 *99.24 *99.02 *0.9861 *92.9699.2498.130.961093.9798.8197.950.963990.9598.9197.500.9493
Class 293.4798.4496.880.959591.4897.1495.360.943195.7494.5394.910.951393.7595.0594.640.9440
Class 2.594.1396.8295.710.954795.6594.3994.910.950289.5797.8894.460.937291.7496.9794.820.9435
Class 395.4198.1297.860.967787.1698.8197.680.929890.8398.2297.500.945292.6698.2297.680.9544
15 s0.83–1.95 HzClass 096.1299.1798.630.976596.9099.3398.900.981293.0299.5098.360.962689.1599.1797.400.9416
Class 295.6397.4196.850.965295.6397.6096.990.966295.6396.4196.160960295.2096.2195.890.9570
Class 2.59396.9895.340.94999497.9196.300.959595.6796.5196.160.960994.6796.2895.620.9547
Class 388.8998.0297.120.934695.83 *98.48 *98.22 *0.9716 *90.2899.5498.630.949186.1198.3397.120.9222
1.95–50 HzClass 093.0299.1798.080.961088.379997.120.936990.7099.6798.080.951884.5097.5095.210.9100
Class 294.3295.4195.070.948793.0194.8194.250.939196.9494.6195.340.957884.7294.2191.230.8946
Class 2.591.6796.7494.660.942194.3394.8894.660.946193.3393.7293.560.935394.3392.7993.420.9356
Class 391.6798.3397.670.950083.3399.0997.530.912169.4499.2496.300.843484.7298.7897.400.9175
30 s0.83–1.95 HzClass 086.2199.6397.280.929286.2199.2796.980.927487.9310097.890.939793.1097.0796.370.9509
Class 296.1293.4294.260.947795.1593.8694.260.945089.3295.6193.660.924787.3893.8691.840.9062
Class 2.590.3797.4594.560.939192.5996.4394.860.945197.0486.7390.940.918990.3795.4193.350.9289
Class 394.2997.9797.580.961388.5798.6597.580.936157.1499.6695.170.784088.5798.9997.890.9378
1.95–50 HzClass 084.4898.1795.770.913382.7698.9096.070.908386.2198.1796.070.921982.7698.1795.470.9046
Class 291.2693.8693.050.925686.4192.9890.940.897091.2692.5492.150.919088.3592.9891.540.9067
Class 2.594.8195.4195.170.951193.3391.8492.450.925990.3789.8090.030.900887.4193.8891.240.9064
Class 388.5799.6698.490.941285.7198.9997.580.923554.2998.6593.960.764785.7196.2895.170.9100
Note: * selected by Youden’s index criteria as the best classification result.
Table 10. Multi-class classification of RF from Si dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 10. Multi-class classification of RF from Si dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 097.9997.8697.920.979297.9297.4198.330.978799.43 *97.86 *98.57 *0.9864 *96.2697.1496.740.9670
Class 295.2494.2194.660.947395.9695.5496.300.959295.83 *97.22 *96.61 *0.9653 *95.2489.5892.060.9241
Class 2.577.3898.8396.480.881197.7990.4898.680.945888.1099.2798.050.936859.5299.1294.790.7932
1.95–50 HzClass 098.2897.3897.790.978394.2599.2997.010.967797.9996.1997.010.970996.8496.9096.880.9687
Class 292.2697.4595.180.948698.2193.7595.700.959895.2495.8395.570.955488.1096.0692.580.9208
Class 2.592.8697.6697.140.952690.4899.4298.440.949585.7199.8598.310.927891.6795.9195.440.9379
15 s0.83–1.95 HzClass 097.4197.8697.660.976498.2897.8698.050.980795.2698.5797.070.969298.2897.1497.660.9771
Class 292.4192.7192.580.925695.0996.5395.900.958196.8893.4094.920.951492.8695.4994.340.9417
Class 2.571.4397.3794.530.844089.2998.9097.850.940985.7199.3497.850.925382.1498.0396.290.9008
1.95–50 HzClass 096.9896.0796.480.965393.5397.8695.900.957096.5594.6495.510.956094.4095.3694.920.9488
Class 293.3096.5395.120.949293.7594.1093.950.939290.6394.4492.770.925390.1894.4492.580.9231
Class 2.592.8698.9098.240.958894.64 *98.03 *97.66 *0.9633 *83.9398.4696.880.912092.8697.8197.270.9533
30 s0.83–1.95 HzClass 089.6699.2994.920.944793.1097.8695.700.954893.109594.140.940591.3892.8692.190.9212
Class 297.3288.1992.190.927696.4390.9793.360.937091.0790.2890.630.906786.6192.3689.840.8948
Class 2.578.5798.6896.480.886378.5799.1296.880.88857598.2595.700.866292.8697.3796.880.9511
1.95–50 HzClass 096.5595.7196.090.961390.5297.1494.140.938395.6994.2994.920.949992.2497.1494.920.9469
Class 288.3993.7591.410.910791.9689.5890.630.907791.9687.5089.450.897388.3993.7591.410.9107
Class 2.578.5796.4994.530.875382.1497.3795.700.89765099.1293.750.745696.4395.6195.700.9602
Note: * selected by Youden’s index criteria as the best classification result.
Table 11. Multi-class classification of CF Ga Dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 11. Multi-class classification of CF Ga Dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 096.4298.5697.850.974997.3299.4498.740.983896.8799.4498.590.981696.2099.3398.290.9777
Class 294.5195.9195.400.952194.1197.5496.290.958297.1595.6796.210.964194.5195.5695.170.9503
Class 2.590.4293.9593.320.921892.9296.2195.620.945687.9296.4894.950.922081.6794.9492.580.8830
Class 368.4599.6695.770.840689.2999.3298.070.943082.1499.4997.330.908280.3698.2295.9908929
1.95–50 HzClass 097.9999.4498.960.987295.5398.7897.700.971596.8797.7897.480.973296.8799.2298.440.9805
Class 296.5498.4897.770.975194.9296.0295.620.954794.5195.5695.170.950394.5196.8495.990.9568
Class 2.592.5098.2897.250.953990.4297.5696.290.93998598.4696.070.917392.0896.7595.920.9442
Class 395.8398.8198.440.973292.2699.2498.370.957594.0598.9898.370.965190.4899.4998.370.9498
15 s0.83–1.95 HzClass 095.9697.6797.100.968197.6499.1798.660.984096.6399.1798.330.979092.5998.8396.770.9571
Class 292.9996.3195.090.946592.9998.4296.430.957093.6097.0195.760.953092.6894.9094.090.9379
Class 2.587.5096.3494.760.91929096.0794.980.930388.7594.4493.420.915983.1394.3092.310.8871
Class 383.9398.6096.770.912691.0798.3497.440.947177.6898.6095.990.881477.6897.4594.980.8757
1.95–50 HzClass 098.65 *99 *98.89 *0.9883 *97.3199.1798.550.982496.3097.5097.100.969096.6399.1798.330.9790
Class 296.65 *98.42 *97.77 *0.9753 *96.0497.5496.990.967994.8294.3894.540.946095.4397.7296.880.9657
Class 2.591.88 *99.19 *97.88 *0.9553 *92.5097.6996.770.951083.1397.2994.760.90219096.2095.090.9310
Class 399.11 *98.98 *99 *0.9904 *91.9699.2498.330.956083.9390.2497.320.915883.9398.3496.540.9114
30 s0.83–1.95 HzClass 091.1696.6794.850.939193.2097.6796.200.954391.8499.3396.870.955991.8497.3395.530.9459
Class 287.2090.8189.490.890090.8593.9992.840.924294.5192.5893.290.935589.0293.6491.950.9133
Class 2.57095.1090.600.825587.5095.1093.740.913082.5095.1092.840.88808595.1093.290.9005
Class 387.5097.1995.970.923482.1499.2397.090.906980.3698.7296.420.895489.2998.9897.760.9413
1.95–50 HzClass 095.2497.6796.870.964592.5298.3396.420.954393.8897.3396.200.956195.9294.6795.080.9529
Class 291.4696.1194.410.937993.2994.7094.180.940094.5193.6493.960.940884.7696.4792.170.9061
Class 2.58596.7394.630.90878595.9193.960.904687.5094.8293.510.911691.2595.1094.410.9317
Class 391.0797.9597.090.945183.9397.9596.200.909467.8699.7495.750.838083.9399.2397.320.9158
Note: * selected by Youden’s index criteria as the best classification result.
Table 12. Multi-class classification of CF from Ju dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 12. Multi-class classification of CF from Ju dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 093.9799.0298.130.965096.4899.5799.020.980295.4899.8999.110.976894.4799.5798.660.9702
Class 292.0597.7995.980.949296.02 *98.18 *97.50 *0.9710 *94.6098.3197.140.964594.0397.1496.160.9558
Class 2.596.0993.1894.380.946397.3996.2196.700.968098.2693.3395.360.958092.6193.1892.950.9290
Class 377.0698.8196.700.879486.2499.5198.210.928777.0699.7097.500.883874.3197.7395.450.8602
1.95–50 HzClass 098.49 *98.91 *98.84 *0.9870 *93.4799.0298.040.962596.9898.2698.040.976291.9698.7097.500.9533
Class 293.1898.8397.050.960094.3296.2295.630.952793.4795.70950.945892.3396.4895.180.9441
Class 2.596.52 *97.88 *97.32 *0.9720 *94.3597.1295.980.957393.4897.1295.630.953096.0996.5296.340.9630
Class 3100 *99.01 *99.11 *0.9951 *92.6699.0198.390.958488.9999.7098.660.943593.5899.4198.840.9649
15 s0.83–1.95 HzClass 096.1298.6798.220.974095.3599.3396.830.973492.2599.5098.220.958792.2598.8497.670.9554
Class 293.8996.2195.480.950594.3297.4196.440.958694.3297.2196.300.957693.0197.2195.890.9511
Class 2.59296.2894.520.941493.6794.6594.250.941696.6792.7994.380.947393.3394.4293.970.9388
Class 387.5098.6397.530.930780.5698.1896.440.893770.8399.0996.300.849679.1797.5795.750.8837
1.95–50 HzClass 095.3598.3497.810.968488.3799.5097.530.939493.0298.5097.530.957686.829795.210.9191
Class 292.589896.300.952992.5895.0194.250.937992.5894.2193.700.933986.9094.6192.190.9076
Class 2.59497.4496.030.957294.3395.3594.930.948492.3393.0292.740.926893.3395.8194.790.9457
Class 394.4497.8797.530.961691.6798.9498.220.953069.4499.5496.580.844991.6798.4897.810.9507
30 s0.83–1.95 HzClass 089.6698.5396.980.940981.0310096.680.905270.6910094.860.853493.1099.2798.190.9619
Class 289.3293.4292.150.913796.1292.1193.350.941192.2391.6791.840.919592.2395.1894.260.9370
Class 2.586.6794.3991.240.905390.3795.9293.660.931496.3089.2992.150.927986.6794.9091.540.9078
Class 388.5796.9696.070.927791.4398.3197.580.948768.5799.6696.370.841288.5796.2895.470.9243
1.95–50 HzClass 091.3898.1796.980.947784.4810097.280.922494.8398.9098.190.968693.1099.2798.190.9619
Class 293.2096.0595.170.946393.2095.1894.560.941995.1591.6792.750.934195.1594.7494.860.9494
Class 2.592.5996.4394.860.945193.3393.8893.660.936185.9392.3589.730.891489.6394.3992.450.9201
Class 385.7197.9796.680.918485.7197.6496.370.916762.8698.9995.170.80928098.3196.370.8916
Note: * selected by Youden’s index criteria as the best classification result.
Table 13. Multi-class classification of CF from Si dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 13. Multi-class classification of CF from Si dataset for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 097.4195.7196.480.965696.8497.6297.270.972397.1398.5797.920.978593.9799.2996.880.9663
Class 292.5694.6893.750.936294.9495.3795.180.951697.0293.5295.050.952797.0290.0593.100.9354
Class 2.582.1498.8397.010.904989.2998.9897.920.941378.5799.4297.140.889972.6298.8395.960.8572
1.95–50 HzClass 097.41 *99.52 *98.57 *0.9847 *95.9897.6296.880.968099.1495.2497.010.971997.1397.1497.140.9713
Class 294.94 *97.69 *96.48 *0.9631 *95.2496.0695.700.956593.4597.4595.700.954593.1596.0694.790.9461
Class 2.597.6297.6697.660.976495.2498.9898.570.971189.2999.5698.440.944290.4898.2597.400.9436
15 s0.83–1.95 HzClass 094.8395.7195.310.952797.4197.5097.460.974694.8397.5096.290.961693.9798.5796.480.9627
Class 293.3093.7593.550.935392.8696.5394.920.946991.9694.1093.160.930394.6491.6792.970.9315
Class 2.589.2999.3498.240.943192.8698.0397.460.954491.0797.5996.880.943380.3698.0396.090.8919
1.95–50 HzClass 099.1495.7197.270.974394.8396.4395.700.956396.129595.510.955694.8399.2997.270.9706
Class 289.7399.3195.120.945291.9695.1493.750.935593.3092.7192.970.930196.8892.3694.340.9462
Class 2.598.21 *97.37 *97.46 *0.9779 *94.6498.0397.660.963376.7999.5697.070.881780.3698.6896.680.8952
30 s0.83–1.95 HzClass 090.5295.7193.360.931296.5597.8697.270.972093.9710097.270.969890.5293.5792.190.9204
Class 292.8689.5891.020.912293.7592.3692.970.930697.3291.6794.140.944989.2990.2889.840.8978
Class 2.582.1498.6896.880.90417598.2595.700.866282.1498.6896.880.904185.7198.2596.880.9198
1.95–50 HzClass 093.1097.8695.700.954893.9797.8696.090.959195.6996.4396.090.960691.3893.5792.580.9248
Class 293.7593.7593.750.937592.8691.6792.190.922695.5490.2892.580.929190.1890.2890.230.9023
Class 2.592.8697.8197.270.953378.5797.3795.310.879764.2999.5695.700.819282.1498.6896.880.9041
Note: * selected by Youden’s index criteria as the best classification result.
Table 14. Multi-class classification of LF from all datasets for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 14. Multi-class classification of LF from all datasets for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 092.1594.5693.820.933593.8696.8895.950.953794.27 *97.68 *96.63 *0.9597 *94.3795.5895.210.9497
Class 281.6189.4986.620.855584.9295.0991.380.900087.63 *94.06 *91.72 *0.9085 *76.2790.5185.320.8339
Class 2.577.8192.3788.840.850991.45 *93.35 *92.89 *0.9240 *89.1693.9692.800.915679.2188.8286.490.8402
Class 366.7998.7896.040.827880.1499.0997.470.896280.5199.3297.710.899166.4399.1996.380.8281
1.95–50 HzClass 088.6393.6292.090.911390.1494.2092.9592.1780.6896.1291.380.884085.6193.40910.8950
Class 275.2586.8682.630.810675.5987.4583.1281.5279.4982.2981.270.808974.2482.7779.660.7851
Class 2.575.6492.5388.440.840975.3892.1788.1083.7776.1591.6487.880.838966.2092.8686.400.7953
Class 387.7398.8597.900.932987.7398.8897.9393.3182.3199.5398.050.909293.14 *98.17 *97.74 *0.9566 *
15 s0.83–1.95 HzClass 086.0294.4691.870.902492.4095.3494.440.938791.1996.4294.810.938086.0294.6091.960.9031
Class 278.8788.8185.180.838484.2592.7889.670.885287.9691.4690.180.897176.708883.870.8235
Class 2.583.5390.3388.690.869386.0595.1392.940.905985.4794.4592.290.899676.1691.4487.750.8380
Class 360.3398.7795.470.795586.4198.7797.710.925973.3799.1396.910.862581.5297.4996.120.8951
1.95–50 HzClass 086.0293.1190.930.895784.5095.8192.330.901689.2194.8093.080.920188.9190.3489.900.8963
Class 264.4086.6078.490.755077.9883.5881.530.807879.5181.9681.070.807468.7682.7077.810.7573
Class 2.576.7487.3184.760.820370.9391.1986.300.810664.9291.3784.990.781562.9891.6284.710.7730
Class 384.2498.5297.290.913882.0798.5297.100.902969.0299.6497.010.843380.9898.5297.010.8975
30 s0.83–1.95 HzClass 083.8096.9192.840.903689.4196.6394.390.930286.9296.4993.520.917082.8794.8191.100.8884
Class 280.4783.9782.690.822284.7086.7285.980.857186.2885.6585.880.859680.4783.0582.110.8176
Class 2.564.2091.4085.010.778076.5492.5488.780.845476.1391.9188.200.840267.4991.9186.170.7970
Class 383.5296.5095.360.900172.5399.5897.200.860559.3499.3695.840.793572.5397.7795.550.8515
1.95–50 HzClass 075.7094.3988.590.850583.1894.2590.810.887179.4494.9590.140.872083.1888.5086.850.8584
Class 268.6078.3274.760.734676.5280.1578.820.783370.1880.1576.500.751758.3180.6172.440.6946
Class 2.562.5587.7481.820.751461.3289.2582.690.752967.4986.2281.820.768560.0886.8580.560.7347
Class 382.4297.3596.030.898857.1497.8894.290.775163.7498.3095.260.810271.4397.6795.360.8455
Note: * selected by Youden’s index criteria as the best classification result.
Table 15. Multi-class classification of RF from all datasets for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 15. Multi-class classification of RF from all datasets for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 092.0594.4293.690.932494.97 *96.34 *95.92 *0.9566 *93.6697.5996.380.956390.6495.3193.880.9298
Class 284.5890.8088.530.876983.9895.04910.895187.88 *93.72 *91.59 *0.9080 *71.6989.8883.250.8079
Class 2.582.4094.5791.620.884989.6793.0692.240.913788.6593.6492.430.911480.7487.0385.500.8388
Class 372.5699.0596.790.858177.2699.0297.160.881475.8199.2297.220.875268.2398.8296.200.8352
1.95–50 HzClass 089.6492.5991.680.911289.5495.0993.380.923188.7393.9892.360.913582.9093.8490.4888.37
Class 267.1290.2781.820.786980.3486.4784.230.834173.9085.1181.020.795070.3482.8278.2776.58
Class 2.577.4290.3787.230.839074.4993.5188.900.840071.1791.1986.340.811873.2188.9885.1681.10
Class 394.95 *97.30 *97.09 *0.9612 *86.6499.1298.050.928884.4898.9297.680.917079.4299.1297.4389.27
15 s0.83–1.95 HzClass 091.7994.7393.830.932692.4096.4995.230.944590.1297.8495.470.939890.5895.4193.920.9299
Class 280.5491.0287.190.857885.5394.1891.020.898691.0490.4390.650.907378.3690.6586.160.8450
Class 2.582.9592.3690.090.876589.53 *94.82 *93.55 *0.9218 *84.8894.8992.470.898879.6591.9989.010.8582
Class 373.3799.0396.820.862087.5098.7297.760.931175.5499.3497.290.874483.1597.8596.590.9050
1.95–50 HzClass 083.2893.6590.460.884785.2696.6993.170.909786.6394.8792.330.907582.3792.5189.390.8744
Class 270.8184.6879.620.777579.3984.4682.610.819275.4285.3581.720.803869.2780.7176.530.7499
Class 2.575.1989.8386.300.825170.9391.0786.210.81007590.5186.770.827664.5389.7783.680.7715
Class 383.1598.4797.150.908182.0798.1696.770.901179.3599.0897.380.892179.8998.1196.540.8900
30 s0.83–1.95 HzClass 090.0393.9792.750.920090.9796.2194.580.935989.1095.5193.520.923078.5094.9589.850.8673
Class 277.3189.6285.110.834684.1791.4588.780.878181.7990.0887.040.859477.3183.0580.950.8018
Class 2.576.5493.1789.260.848686.0193.3091.590.896586.0190.7789.650.883977.7890.3987.430.8408
Class 387.9197.7796.910.928478.0299.2697.390.886460.4499.5896.130.800171.4398.7396.320.8508
1.95–50 HzClass 083.1890.6088.300.868986.9292.5790.810.897481.6293.8390.040.877280.6992.5788.880.8663
Class 253.5684.4373.110.689967.5582.6077.080.750764.9182.6076.110.737565.4480.3174.850.7287
Class 2.568.3183.0679.590.756959.6787.1080.660.733974.0783.4481.240.787660.9186.0980.170.7350
Class 372.5396.9294.780.847364.8497.2494.390.810453.8599.1595.160.765065.9397.1494.390.8154
Note: * selected by Youden’s index criteria as the best classification result.
Table 16. Multi-class classification of CF from all datasets for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 16. Multi-class classification of CF from all datasets for CO (Class 0), PD Stage 2 (Class 2), PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDisease Severity (Class)AlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzClass 088.4395.8193.540.921291.65 *96.21 *94.81 *0.9393 *89.3497.4194.930.933788.7395.2793.260.9200
Class 278.8190.4686.210.846487.37 *91.19 *89.80 *0.8928 *88.8189.2989.120.890578.3989.5985.500.8399
Class 2.585.9791.1989.920.885885.20 *95.55 *93.04 *0.9038 *83.6794.5391.900.891081.3891.1588.780.8626
Class 378.3499.0297.250.886883.7599.3998.050.915778.3499.5397.710.889375.4598.5196.540.8698
1.95–50 HzClass 087.8395.3193.010.915788.7396.5694.160.926590.4496.3994.560.934190.4495.7294.100.9308
Class 283.9885.2684.790.846282.2987.3585.500.848280.3488.4785.500.844073.7388.6183.180.8117
Class 2.571.5695.6889.830.836278.0693.1589.490.856079.0891.8488.750.854676.2890.0986.740.8318
Class 392.06 *98.61 *98.05 *0.9534 *8799.1998.150.931082.3199.2697.810.907889.5398.4197.650.9397
15 s0.83–1.95 HzClass 088.1595.3493.130.917491.0395.6194.200.933286.1797.1693.780.916786.3294.4691.960.9039
Class 282.5987.9285.970.852582.7191.6888.410.872085.9288.5187.560.873278.8786.6083.780.8274
Class 2.581.5993.3590.510.874785.6693.9691.960.898183.1493.9691.350.885576.1692.2488.360.8420
Class 377.7299.5497.660.886384.7898.9897.760.918883.7098.8797.570.912976.6398.4196.540.8752
1.95–50 HzClass 089.3693.7992.430.915785.1195.7592.470.904386.4796.4993.410.914886.9394.0691.870.9049
Class 278.7583.4381.720.810978.7586.6083.730.826778.7586.3883.590.825670.0487.0480.830.7854
Class 2.564.1594.1586.910.791578.6892.0588.830.853778.8890.2087.470.845477.5288.9186.160.8321
Class 389.1398.5297.710.938285.8798.6797.570.922776.6399.4497.480.880381.5298.6797.190.9010
30 s0.83–1.95 HzClass 082.2494.3990.620.883287.5496.91940.922381.9397.3492.550.896385.3691.8789.850.8861
Class 275.4682.7580.080.791084.7088.0986.850.863981.7986.8785.010.843374.4182.7579.690.7858
Class 2.567.9091.2885.780.795979.4292.4189.360.859282.7290.9088.970.868163.7993.6886.650.7873
Class 382.4297.6796.320.900473.6398.7396.520.861875.8298.5296.520.871785.7197.4596.420.9158
1.95–50 HzClass 083.1892.0189.260.875983.1993.9790.620.885784.4292.2989.850.883580.0591.3087.810.8568
Class 267.8179.5475.240.736871.2482.6078.430.769259.1087.0276.790.730657.5281.2272.530.6937
Class 2.556.7989.6381.910.732167.9089.3884.330.786478.6083.6982.500.811568.3184.0780.370.7619
Class 382.4297.4596.130.899479.1297.9996.320.885570.3398.4195.940.843772.5398.3096.030.8542
Note: * selected by Youden’s index criteria as the best classification result.
Table 17. Two-class classification of LF for CO and PD (Class 2, Class 2.5, and Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 17. Two-class classification of LF for CO and PD (Class 2, Class 2.5, and Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDatasetAlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzGa96.0299.1297.920.996498.8798.4698.590.997198.9099.0298.960.999695.9598.6397.620.9938
Ju96.6699.5799.020.994497.9499.2599.020.999298.49 *98.95 *98.84 *0.9968 *93.9399.5798.480.9936
Si96.5897.71970.990898.85 *98.41 *98.56 *0.9964 *97.5897.5497.390.997195.0197.8796.360.9926
All86.9797.0993.600.975793.5097.0895.920.992795.28 *96.52 *96.07 *0.9915 *90.7996.0294.250.9731
1.95–50 HzGa96.2798.5997.690.993997.7797.1097.250.996096.3498.4797.700.997598.2297.7297.850.9892
Ju96.0999.1498.570.991796.1198.5098.040.996094.6098.9398.030.999095.1998.7198.030.9860
Si97.0497.0396.870.997798.2596.7897.400.994694.0997.8395.970.995096.8896.7396.750.9893
All84.8695.7092.060.964589.2195.8493.690.982789.3594.4392.830.972883.849591.280.9643
15 s0.83–1.95 HzGa97.9998.0397.990.994499.01 *99.17 *99.11 *0.9996 *97.4298.6998.220.997197.3798.5498.110.9959
Ju94.7898.3697.670.994094.8799.1798.350.997592.3898.8797.530.998495.1898.8598.080.9902
Si95.4897.9996.490.994698.8097.5798.050.996197.0596.6096.690.998397.5297.9597.660.9913
All87.6994.9492.620.965390.4096.41 94.480.988490.7496.3994.530.987589.5393.8792.520.9577
1.95–50 HzGa95.4998.8497.660.997497.6497.5897.550.993395.5497.9096.990.997196.3798.2097.550.9936
Ju96.8597.9097.670.985293.1696.8096.170.991193.6997.9396.990.984892.0696.6395.620.9616
Si98.3096.2397.060.994394.6796.6195.320.992397.1196.8796.870.995096.279695.900.9835
All89.7492.1791.170.941686.4893.9991.450.966089.7393.1491.730.968581.4294.3189.950.9499
30 s0.83–1.95 HzGa93.5698.0996.210.994198.5297.0997.530.997996.2896.1795.980.998093.3997.7695.990.9889
Ju91.9098.2396.670.988397.5098.5898.200.998010098.2798.490.998093.8197.1896.370.9794
Si96.8697.8597.310.99329495.1494.520.990694.3295.5294.120.993295.6394.2594.540.9830
All91.4694.2693.330.971491.5294.9693.810.989292.3894.94940.985688.1491.9590.720.9310
1.95–50 HzGa93.1797.4995.760.992895.0898.6897.320.995593.8897.7896.180.994491.9397.0195.090.9862
Ju86.2198.5496.090.975498.5796.5896.650.986895.4896.5996.080.998777.1195.6891.560.9370
Si97.6394.4095.320.978894.2494.7894.180.988493.1097.3194.890.990993.0793.8193.290.9767
All84.4589.83880.915386.3293.1290.720.963591.9087.8888.690.942181.509287.710.9272
Note: * selected by Youden’s index criteria as the best classification result.
Table 18. Two-class classification of RF for CO and PD (Class 2, Class 2.5, and Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 18. Two-class classification of RF for CO and PD (Class 2, Class 2.5, and Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDatasetAlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzGa96.0399.4498.220.998398.4798.4898.440.998297.8298.3798.140.999597.1898.47980.9915
Ju95.7899.4698.750.997097.5298.7298.480.996695.6699.3598.660.999197.97 *98.72 *98.57 *0.9895 *
Si96.8498.8697.780.997298.62 *98.63 *98.57 *0.9994 *97.8499.0798.440.999097.7497.0697.270.9972
All91.3396.9495.120.974992.2997.6695.920.992994.46 *97.69 *96.63 *0.9949 *90.0597.0694.780.9763
1.95–50 HzGa99.56 *98.18 *98.59 *0.9941 *97.8397.6297.620.997493.8898.4596.810.994097.1597.9397.630.9922
Ju96.3399.5798.930.996993.9998.8297.860.996396.5498.3197.950.999394.5897.8797.230.9729
Si97.1898.6297.780.997698.2896.5797.270.996795.3797.2896.220.995897.4997.0597.130.9909
All82.2096.4291.410.966189.9295.0793.450.978089.0394.7992.920.974785.9993.9391.280.9590
15 s0.83–1.95 HzGa98.0398.5698.320.99649998.2198.440.997696.9098.0497.550.998895.8198.8397.770.9939
Ju94.5998.6897.940.994697.1899.0398.630.999296.2699.1898.640.994895.4597.9297.400.9831
Si97.6598.2797.850.995097.9398.2698.040.999197.6397.6197.460.998998.0198.6498.240.9938
All90.9796.1694.490.973992.2997.2595.610.991493.9196.4295.600.988391.2295.1793.920.9584
1.95–50 HzGa97.5399.0298.440.998098.2296.2296.770.995894.0497.5296.200.994294.6797.5896.430.9784
Ju94.6698.6897.950.991195.0397.4096.850.991995.2998.4097.530.996892.0997.2696.310.9574
Si96.3397.8797.070.995496.5695.5095.910.991494.2095.3094.330.990093.4197.5695.510.9857
All85.6193.4890.840.947188.8495.2093.080.977988.5894.2192.330.971682.6791.9088.780.9363
30 s0.83–1.95 HzGa97.3797.8197.540.991794.6598.4196.870.998296.7597.7797.330.998295.1398.0696.850.9848
Ju97.5098.2597.890.991788.6998.6196.370.99369597.2996.370.98489097.5095.800.9970
Si93.7695.5594.180.988497.7496.9096.850.993092.7793.6792.940.983792.2593.5992.540.9740
All87.4895.1092.470.968890.5796.1394.190.983891.0694.5493.330.986285.2490.0988.490.9137
1.95–50 HzGa95.6097.4796.640.987394.3198.1196.430.992593.8298.3596.650.992992.9697.1295.520.9828
Ju10096.5696.990.976792.8696.8695.810.992788.2197.5395.480.988388.8396.1394.900.9736
Si95.4395.4294.920.991396.2594.1994.540.989590.4597.7193.720.988194.2294.5094.150.9784
All79.6593.8488.690.946784.6394.5490.720.968482.2593.6589.650.958079.2492.54880.9250
Note: * selected by Youden’s index criteria as the best classification result.
Table 19. Two-class classification of CF for CO and PD (Class 2, Class 2.5, and Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Table 19. Two-class classification of CF for CO and PD (Class 2, Class 2.5, and Class 3) using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation.
Time WindowFrequency RangeDatasetAlexNetResNet-50ResNet-101GoogLeNet
Sen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUCSen (%)Spec (%)Acc (%)AUC
10 s0.83–1.95 HzGa98.0398.7998.520.998299.77 *98.80 *99.11 *0.9995 *99.1098.8098.890.999499.1297.7598.140.9875
Ju95.1699.3598.570.991996.7099.2498.750.999198.94 *99.04 *99.01 *0.9993 *94.3199.2498.310.9864
Si9797.6797.270.99699897.2497.530.998898.88 *97.35 *97.92 *0.9984 *97.8196.3596.870.9899
All89.2294.7192.980.963290.9496.4394.680.988992.71 *95.11 *94.37 *0.9856 *90.4494.3693.170.9533
1.95–50 HzGa98.2699.3498.960.997697.2699.0298.360.996896.1198.8997.920.997897.5397.8197.700.9900
Ju97.1998.9498.570.991995.0998.7198.040.992795.3098.8398.130.996195.5998.4197.860.9758
Si98.6597.4997.920.996997.2297.4597.260.996895.6399.0597.400.997496.0998.5897.400.9934
All88.3896.1893.660.975790.8595.8894.190.984392.0795.4194.310.984490.2996.0594.190.9755
15 s0.83–1.95 HzGa97.0997.4597.210.989998.3799.35990.999898.6798.8498.770.999196.9096.9296.880.9846
Ju95.3598.7098.080.986094.9999.0198.220.999096.4398.6998.220.998894.0799.0298.080.9978
Si93.8397.1395.510.994597.5797.6597.450.995098.2394.9496.280.995396.3396.5296.290.9856
All89.0594.3692.710.973088.9096.1393.740.987091.8393.7692.990.984288.1595.4293.030.9580
1.95–50 HzGa97.2799.5098.670.997795.1898.3597.210.996496.4798.5297.770.996996.7298.1997.660.9941
Ju93.4298.6997.530.991095.5296.9796.580.988892.7898.3497.260.996787.3097.0795.210.9610
Si94.5498.9396.690.994497.0996.9996.870.995991.8798.5495.120.995195.7697.1996.310.9938
All87.2095.0192.470.966090.8194.6593.320.976690.1494.5493.130.977385.7294.5791.630.9650
30 s0.83–1.95 HzGa90.9695.6193.510.986496.7997.4197.090.995997.2195.9296.190.993894.0895.5294.840.9863
Ju89.2997.8596.080.978196.9096.8596.680.988191.2496.4695.490.985392.7498.2496.980.9790
Si92.1495.1593.370.987697.339897.230.997498.3296.6797.290.994394.7296.2894.880.9731
All85.8493.2590.810.967188.5996.2393.610.986589.8693.2292.170.980183.3992.5589.270.9222
1.95–50 HzGa94.9898.0996.870.992496.0296.1995.960.991893.1998.7296.650.992493.3198.0596.200.9881
Ju96.9098.9398.480.982794.9098.5897.870.998197.1499.3098.800.989295.7198.9298.210.9881
Si96.8393.2394.140.981298.4096.0896.860.992394.1498.0895.720.988893.7493.3592.970.9841
All81.3892.9188.690.940486.9794.9692.060.969285.9194.4291.400.967085.3389.9188.300.9126
Note: * selected by Youden’s index criteria as the best classification result.
Table 20. Multi-class (CO (Class 0) vs. PD Stage 2 (Class 2) vs. PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3)) and two-class classification (CO vs. PD (Class 2, Class 2.5, and Class 3)) summary using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation for Ga dataset.
Table 20. Multi-class (CO (Class 0) vs. PD Stage 2 (Class 2) vs. PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3)) and two-class classification (CO vs. PD (Class 2, Class 2.5, and Class 3)) summary using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation for Ga dataset.
DatasetvGRF SignalCNN ClassifierTime WindowFrequency RangeClassification TaskEvaluation Parameters
Sen (%)Spec (%)Acc (%)AUC
GaLFGoogLeNet15 s0.83–1.95 HzMulti-ClassClass 097.9899.1798.770.9857
AlexNet10 s1.95–50 HzClass 295.9398.2597.400.9709
AlexNet10 s1.95–50 HzClass 2.593.7598.8397.920.9629
AlexNet30 s1.95–50 HzClass 398.2197.4497.540.9783
ResNet-5015 s0.83–1.95 HzTwo-Class99.0199.1799.110.9996
RFResNet-10110 s0.83–1.95 HzMulti-ClassClass 098.439998.810.9872
AlexNet10 s1.95–50 HzClass 297.97 *97.78 *97.85 *0.9787 *
AlexNet15 s0.83–1.95 HzClass 2.596.88 *96.88 *96.88 *0.9688 *
ResNet-5010 s1.95–50 HzClass 394.0599.2498.590.9664
AlexNet10 s1.95–50 HzTwo-Class99.5698.1898.590.9941
CFAlexNet15 s1.95–50 HzMulti-ClassClass 098.65 *99 *98.89 *0.9883 *
AlexNet15 s1.95–50 HzClass 296.6598.4297.770.9753
AlexNet15 s1.95–50 HzClass 2.591.8899.1997.880.9553
AlexNet15 s1.95–50 HzClass 399.11 *98.98 *99 *0.9904 *
ResNet5010 s0.83–1.95 HzTwo-Class99.77 *98.80 *99.11 *0.9995 *
Note: * denotes the best classification result and was selected using Youden’s index criteria.
Table 21. Multi-class (CO (Class 0) vs. PD Stage 2 (Class 2) vs. PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3)) and two-class classification (CO vs. PD (Class 2, Class 2.5, and Class 3)) summary using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation for Ju dataset.
Table 21. Multi-class (CO (Class 0) vs. PD Stage 2 (Class 2) vs. PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3)) and two-class classification (CO vs. PD (Class 2, Class 2.5, and Class 3)) summary using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation for Ju dataset.
DatasetvGRF SignalCNN ClassifierTime WindowFrequency RangeClassification TaskEvaluation Parameters
Sen (%)Spec (%)Acc (%)AUC
JuLFResNet-5010 s0.83–1.95 HzMulti-ClassClass 096.9899.6799.200.9833
ResNet-5010 s0.83–1.95 HzClass 296.5998.4497.860.9751
AlexNet10 s1.95–50 HzClass 2.597.3996.5296.880.9695
AlexNet30 s0.83–1.95 HzClass 397.1498.6598.490.9790
ResNet-10110 s0.83–1.95 HzTwo-Class98.4998.9598.840.9968
RFAlexNet10 s1.95–50 HzMulti-ClassClass 097.99 *99.24 *99.02 *0.9861 *
ResNet-10110 s0.83–1.95 HzClass 297.73 *97.40 *97.50 *0.9756 *
ResNet-10110 s0.83–1.95 HzClass 2.595.8796.3696.160.9612
ResNet-5015 s0.83–1.95 HzClass 395.8398.4898.220.9716
GoogLeNet10 s0.83–1.95 HzTwo-Class97.9798.7298.570.9895
CFAlexNet10 s1.95–50 HzMulti-ClassClass 098.4998.9198.840.9870
ResNet-5010 s0.83–1.95 HzClass 296.0298.1897.500.9710
AlexNet10 s1.95–50 HzClass 2.596.52 *97.88 *97.32 *0.9720 *
AlexNet10 s1.95–50 HzClass 3100 *99.01 *99.11 *0.9951 *
ResNet-10110 s0.83–1.95 HzTwo-Class98.94 *99.04 *99.01 *0.9993 *
Note: * denotes the best classification result and was selected using Youden’s index criteria.
Table 22. Multi-class (CO (Class 0) vs. PD Stage 2 (Class 2) vs. PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3)) and two-class classification (CO vs. PD (Class 2, Class 2.5, and Class 3)) summary using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation for Si dataset.
Table 22. Multi-class (CO (Class 0) vs. PD Stage 2 (Class 2) vs. PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3)) and two-class classification (CO vs. PD (Class 2, Class 2.5, and Class 3)) summary using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation for Si dataset.
DatasetvGRF SignalCNN
Classifier
Time WindowFrequency RangeClassification TaskEvaluation Parameters
Sen (%)Spec (%)Acc (%)AUC
SiLFAlexNet10 s1.95–50 HzMulti-ClassClass 098.56 *99.05 *98.83 *0.9881 *
AlexNet10 s1.95–50 HzClass 295.83 *98.84 *97.53 *0.9734 *
AlexNet10 s1.95–50 HzClass 2.598.81 *98.39 *98.44 *0.9860 *
ResNet-5010 s0.83–1.95 HzTwo-Class98.85 *98.41 *98.56 *0.9964 *
RFResNet-10110 s0.83–1.95 HzMulti-ClassClass 099.4397.8698.570.9864
ResNet-10110 s0.83–1.95 HzClass 295.8397.2296.610.9653
ResNet-5015 s1.95–50 HzClass 2.594.6498.0397.660.9633
ResNet-5010 s0.83–1.95 HzTwo-Class98.6298.6398.570.9994
CFAlexNet10 s1.95–50 HzMulti-ClassClass 097.4199.5298.570.9847
AlexNet10 s1.95–50 HzClass 294.9497.6996.480.9631
AlexNet15 s1.95–50 HzClass 2.598.2197.3797.460.9779
ResNet-10110 s0.83–1.95 HzTwo-Class98.8897.3597.920.9984
Note: * denotes the best classification result and was selected using Youden’s index criteria.
Table 23. Multi-class (CO (Class 0) vs. PD Stage 2 (Class 2) vs. PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3)) and two-class classification (CO vs. PD (Class 2, Class 2.5, and Class 3)) Summary using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation for all datasets.
Table 23. Multi-class (CO (Class 0) vs. PD Stage 2 (Class 2) vs. PD Stage 2.5 (Class 2.5), and PD Stage 3 (Class 3)) and two-class classification (CO vs. PD (Class 2, Class 2.5, and Class 3)) Summary using several CNN classifiers (AlexNet, ResNet-50, ResNet-101, and GoogLeNet) with 10-fold cross-validation for all datasets.
DatasetvGRF SignalCNN
Classifier
Time WindowFrequency RangeClassification TaskEvaluation Parameters
Sen (%)Spec (%)Acc (%)AUC
AllLFResNet-10110 s0.83–1.95 HzMulti-ClassClass 094.27 *97.68 *96.63 *0.9597 *
ResNet-10110 s0.83–1.95 HzClass 287.63 *94.06 *91.72 *0.9085 *
ResNet-5010 s0.83–1.95 HzClass 2.591.45 *93.35 *92.89 *0.9240 *
GoogLeNet10 s1.95–50 HzClass 393.1498.1797.740.9566
ResNet-10110 s0.83–1.95 HzTwo-Class95.2896.5296.070.9915
RFResNet-5010 s0.83–1.95 HzMulti-ClassClass 094.9796.3495.920.9566
ResNet-10110 s0.83–1.95 HzClass 287.8893.7291.590.9080
ResNet-5015 s0.83–1.95 HzClass 2.589.5394.8293.550.9218
AlexNet10 s1.95–50 HzClass 394.95 *97.30 *97.09 *0.9612 *
ResNet-10110 s0.83–1.95 HzTwo-Class94.46 *97.69 *96.63 *0.9949 *
CFResNet-5010 s0.83–1.95 HzMulti-ClassClass 091.6596.2194.810.9393
ResNet-5010 s0.83–1.95 HzClass 287.3791.1989.800.8928
ResNet-5010 s0.83–1.95 HzClass 2.585.2095.5593.040.9038
AlexNet10 s1.95–50 HzClass 392.0698.6198.050.9534
ResNet-10110 s0.83–1.95 HzTwo-Class92.7195.1194.370.9856
Note: * denotes the best classification result and was selected using Youden’s index criteria.
Table 24. Multi-class classification results of comparisons with existing literature.
Table 24. Multi-class classification results of comparisons with existing literature.
Literature (Year)
(Cross-Validation)
Ga Dataset Acc (%)Ju Dataset Acc (%)Si Dataset Acc (%)
COPD
Stage 2
PD
Stage 2.5
PD
Stage 3
COPD
Stage 2
PD
Stage 2.5
PD
Stage 3
COPD
Stage 2
PD
Stage 2.5
Zhao et al. (2018) [14] (10foldCV)10093.3310010010010092.3110010096.55100
Proposed Method (10foldCV)99.0397.8596.879998.8497.8697.3299.1198.8397.5398.44
Table 25. Two-class classification results of comparisons with existing literature for all datasets.
Table 25. Two-class classification results of comparisons with existing literature for all datasets.
Literature (Year)Cross-ValidationEvaluation Parameters
Sen (%)Spec (%)Acc (%)AUC
Maachi et al. [39]
(2020)
10foldCV98.1010098.70-
Wu et al. [40]
(2017)
LOOCV72.4196.5584.480.9049
Ertugrul et al. [41]
(2016)
10foldCV88.9082.2088.89-
Zeng et al. [42]
(2016)
5foldCV96.7795.8996.39-
Daliri [43]
(2013)
50% training
50% testing
91.7189.9291.20-
Proposed Method10foldCV94.4697.6996.630.9949
Table 26. Two-class classification results of comparisons with existing literature for each sub-dataset.
Table 26. Two-class classification results of comparisons with existing literature for each sub-dataset.
Literature (Year)Cross-ValidationGa Dataset Acc (%)Ju Dataset Acc (%)Si Dataset Acc (%)
Khoury et al. [44] (2019)LOOCV86.0590.9182.81
Khoury et al. [45] (2018)10foldCV93.5797.5287.22
Proposed Method10foldCV99.1199.0198.56
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Setiawan, F.; Lin, C.-W. Implementation of a Deep Learning Algorithm Based on Vertical Ground Reaction Force Time–Frequency Features for the Detection and Severity Classification of Parkinson’s Disease. Sensors 2021, 21, 5207. https://doi.org/10.3390/s21155207

AMA Style

Setiawan F, Lin C-W. Implementation of a Deep Learning Algorithm Based on Vertical Ground Reaction Force Time–Frequency Features for the Detection and Severity Classification of Parkinson’s Disease. Sensors. 2021; 21(15):5207. https://doi.org/10.3390/s21155207

Chicago/Turabian Style

Setiawan, Febryan, and Che-Wei Lin. 2021. "Implementation of a Deep Learning Algorithm Based on Vertical Ground Reaction Force Time–Frequency Features for the Detection and Severity Classification of Parkinson’s Disease" Sensors 21, no. 15: 5207. https://doi.org/10.3390/s21155207

APA Style

Setiawan, F., & Lin, C. -W. (2021). Implementation of a Deep Learning Algorithm Based on Vertical Ground Reaction Force Time–Frequency Features for the Detection and Severity Classification of Parkinson’s Disease. Sensors, 21(15), 5207. https://doi.org/10.3390/s21155207

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop