Next Article in Journal
A Single Coronary Artery Originating from the Right Coronary Sinus with a Typical Course of the Right Coronary Artery and the Interarterial Course of the Left Main, Left Anterior Descending, and Left Circumflex as an Example of a Rare Case of High-Risk Coronary Anomaly
Previous Article in Journal
Ultrasound Imaging of Acquired Myometrial Pseudoaneurysm: The Role of Manipulators as an Unusual Cause during Laparoscopic Surgery
Previous Article in Special Issue
Comprehensive Survey of Using Machine Learning in the COVID-19 Pandemic
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Bias Investigation in Artificial Intelligence Systems for Early Detection of Parkinson’s Disease: A Narrative Review

1
Department of Biomedical Engineering, North Eastern Hill University, Shillong 793022, India
2
Department of CSE, International Institute of Information Technology, Bhuneshwar 751003, India
3
Department of Radiology, University of Cagliari, 09121 Cagliari, Italy
4
Department of Neurology, University Medical Centre Maribor, 1262 Maribor, Slovenia
5
Department of Radiology, Harvard Medical School, Boston, MA 02115, USA
6
Neurology Department, Fortis Hospital, Bangalore 560010, India
7
Stroke Monitoring and Diagnostic Division, AtheroPoint™, Roseville, CA 95661, USA
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(1), 166; https://doi.org/10.3390/diagnostics12010166
Submission received: 3 December 2021 / Revised: 27 December 2021 / Accepted: 1 January 2022 / Published: 11 January 2022

Abstract

:
Background and Motivation: Diagnosis of Parkinson’s disease (PD) is often based on medical attention and clinical signs. It is subjective and does not have a good prognosis. Artificial Intelligence (AI) has played a promising role in the diagnosis of PD. However, it introduces bias due to lack of sample size, poor validation, clinical evaluation, and lack of big data configuration. The purpose of this study is to compute the risk of bias (RoB) automatically. Method: The PRISMA search strategy was adopted to select the best 39 AI studies out of 85 PD studies closely associated with early diagnosis PD. The studies were used to compute 30 AI attributes (based on 6 AI clusters), using AP(ai)Bias 1.0 (AtheroPointTM, Roseville, CA, USA), and the mean aggregate score was computed. The studies were ranked and two cutoffs (Moderate-Low (ML) and High-Moderate (MH)) were determined to segregate the studies into three bins: low-, moderate-, and high-bias. Result: The ML and HM cutoffs were 3.50 and 2.33, respectively, which constituted 7, 13, and 6 for low-, moderate-, and high-bias studies. The best and worst architectures were “deep learning with sketches as outcomes” and “machine learning with Electroencephalography,” respectively. We recommend (i) the usage of power analysis in big data framework, (ii) that it must undergo scientific validation using unseen AI models, and (iii) that it should be taken towards clinical evaluation for reliability and stability tests. Conclusion: The AI is a vital component for the diagnosis of early PD and the recommendations must be followed to lower the RoB.

1. Introduction

Parkinson’s disease (PD) is a neurodegenerative disorder; James Parkinson first portrayed it in 1817 [1,2]. Globally, over 2% of the population is more than 65 years of age, and around 5–20 people per 100,000 each year are affected by this illness, demonstrating its predominance and frequency rate with maturity [3,4,5]. The registered PD cases reported in the UK were more than 1.45 million [6]. In India, approximately one million cases have had similar experiences for symptoms of PD [7]. Besides these challenges, the pharmaceutical industry has been slow in producing PD drugs. The last invention in this area was in 1967 [8].
PD illness is described by the disturbing dopaminergic cycle of the nerve cells of substantianigra [9,10,11]. A piece of the mind can create neurotransmitters such as “dopamine,” which fills in as a synapse for controlling developments in various body segments. The degenerative interaction begins from the foundation of the mind that prompts the annihilation of olfactory bulbs [12]. It is trailed by the lower cerebrum stem, affecting the susbstantianigra and mid-cerebrum [13]. Ultimately, it obliterates the limbic framework and front-facing neocortex, worsening physical and mental side effects.
The symptoms related to PD can be categorized in two ways (i) verifying the patient’s PD biomarkers, and (ii) by physically observing the differential response from the patient’s body parts [14,15]. Examples of PD indications are the forced closure of eyelids during eye tests [16], lack of breathing during lung tests [17], muscle stiffness during muscle tests [2], and movement of patients while walking [10]. Figure 1 shows various PD symptoms, namely, constipation problems, feelings of anxiety, depression, and abnormalities in breathing [18]. Other symptoms include difficulty in speaking [5], voice tone changes [17], and difficulty in swallowing food [19].
Artificial Intelligence (AI) has recently dominated healthcare, particularly in medical imaging [20,21,22]. Machine learning (ML) has further enhanced the ability to accurately and swiftly make the decisions in the diagnosis of several diseases such as diabetes [23,24], stroke [25,26,27], coronary artery disease prediction [28], and cancer detection in the thyroid [29,30] liver [31], prostate [32,33], and ovaries [34,35]. Recently, there have been attempts to diagnose PD early using AI, especially using ML and DL algorithms [9,11,18,36,37]. The ML/DL algorithms are sensitive to the sample size during the training model generation, and further, due to lack of (i) scientific validation, (ii) clinical evaluation of these AI strategies, and (iii) big data configuration [36], leads to bias in the AI. Thus, when PD symptoms (or risk factors) are considered as input to the AI model, one must ensure that the AI system is reliable, accurate, and has minimal AI bias. Therefore, the primary objective is to automatically identify the AI studies that have bias. In the secondary objective, the goal is to automatically detect the studies that lie in the three categories of bias, such as low, moderate, and high bias. Further, there is a need to understand the AI architectures used in these studies and link them with the AI attributes for different categories of AI bias. Lastly, we need to identify the RoB in these AI studies and suggest possible reduction recommendations. Further, we note that the scope does not involve developing the correlation between PD and other medical conditions.
Figure 1. Symptoms of PD disease [37].
Figure 1. Symptoms of PD disease [37].
Diagnostics 12 00166 g001
Our strategy is to score the 39 AI studies using 30 AI attributes per study with the help of an AI expert that has more than 15 years of AI experience, and then compute the mean aggregate score. Moderate-Low (ML) cutoff was determined using the intersection of the frequency plot of mean score vs. the cumulative frequency plot, where Moderate-Low (ML) cutoff was determined. Further, the second High-Moderate (HM) cutoff was computed based on the transition of slopes. The studies in low-, moderate-, and high-bias were then analyzed for recommendations to reduce the RoB.
The layout of this review is as follows. Section 2 presents the PRISMA model for selecting studies along with the statistical distributions of the parameters. Section 3 presents the AI architecture for PD diagnosis, while Section 4 presents the strategy for the computation of bias and the ranking of the studies used for bias analysis and its analysis. Finally, the critical discussions are presented in Section 5, leading to conclusions in Section 6.

2. Search Strategy and Statistical Distribution

2.1. PRISMA Model

An end-to-end writing search was performed utilizing PubMed, IEEE Xplore, Science Direct, and Google scholar. The significant watchwords utilized for choosing these studies were PD disease, neurodegenerative disease and symptoms, AI, machine learning, and differential finding of the neurodegenerative disease. The research articles selected for the studies consist of various parameters like detections of the PD by using machine learning, deep learning, hybrid learning, and AI. These research articles have also shown the classification of the normal vs. PD-affected people, the demographic analysis of the PD-affected patients, and the classification of the PD by considering the input parameter alternative assessment as one method to detect PD [9]. Studies unrelated to the symptomatic observation of PD are eliminated in published papers for many reasons [11,12,38]. Therefore, studies that are not related to the symptomatic observation of PD are excluded [39].
Figure 2 shows the PRISMA model for the selection strategy of the research articles. The identification phase shows that nearly (246) articles were searched from the identified sources, and 186 studies were searched from the other sources. A total of 396 study articles were removed as they cross the study objective and have duplications. Considering the feasibility of the objective of a selection strategy (396 studies), the articles were screened. The non-AI-based total (168 studies) articles were removed. Many discuss irrelevant information other than the objective of the search strategy. Most of the articles do not fulfill domain criteria like lack of data, lack of information, and poor presentation of the articles. Hence, the total (103 studies) studies are referred to for the analysis [40].
A study that does not include input parameter analysis, performance optimization, attributes analysis, and benchmarking was also not evaluated. Alzheimer’s disease, Huntington’s disease, Motor neuron disease, and Adrenoleukodystrophy (ALD) disease are not categorized as PD. Studies not performed in humans (rat, monkey, etc.) were excluded, as well as studies that do not have a huge sufficient dataset for analysis. The primary objective was to automatically identify the AI studies that have bias. In the secondary objective, the goal was to automatically detect the studies that lie in the three categories of bias, such as low, moderate, and high bias. Other exclusion criteria included having no correlation between Parkinson’s disease with other neurological diseases mentioned in the manuscript, and if the article was written in a different language other than English [1,41]. The information considered for the PD studies’ data extraction was (i) author name, (ii) year of publication, (iii) objective of the studies, (iv) demographic discussion, (v) data types, (vi) data source, (vii) diagnosis method, (viii) bias studies, and (ix) attribute studies. The selected studies were evaluated with the novel and unique implementation of the AI, hybrid AI, twin diagnosis approach, telemedicine approach, and biomarker-based approach for diagnosing PD. Every study was evaluated with feasibility analysis and cross verified with scientific validation [42].
Figure 3 represents the year of publications with reference to the impact factor. From the analysis point of view, we considered publications from the period 2016 to 2021. While observing Figure 3, it is clear that in 2019, the maximum publications are related to the early detection of PD and have a good impact factor. The use of datasets from open-source repositories to minimize research costs also leads to improved performance and overall applicability of the selected model. There is the risk of the bias coming out as High-Moderate (MH) if the model fails to adopt the appropriateness of the open-source data.

2.2. Statistical Distribution

The study objectives include the term exposure of “Parkinson’s disease.” The statistical distribution of the selected studies separates the main AI terms into ML, DL, and HDL [11,12]. The majority of the studies used ML for PD detection, and this accounts for 74%, while 9% use SDL [43,44,45] and 17% use HDL. The performance indicator of the selected algorithm plays a crucial role in bias estimation. Even though accuracy is good, there are chances of the existence of bias or inclusion bias due to the non-clinical validation of AI-based predictions [46,47,48].

Performance Metrics

Symptoms (or risk factors) of PD are considered as input to the AI model. It is important to ensure that the AI system is reliable, accurate, and has minimal AI distortion. The ML/DL algorithm is sensitive to sample size during training model generation and lacks scientific validation and the clinical evaluation of these AI strategies, resulting in a bias in the model.
The SDL (2 studies) architectures were used to detect the PD, showing an average overall accuracy of 97.83%. The maximum accuracy was 98.28% and the minimum was 97.38% for the SDL architecture. The HDL (4 studies) represent the average accuracy of 94.42%, while the maximum accuracy was 87.90% and the minimum was 97.68%. The ML-based model (17 studies) showed an average accuracy of 85.41%. The maximum observed was 94.86% and the minimum was 62.99%. Figure 4a,b, respectively, indicate the average accuracy of the studies and minimum and maximum accuracy of the individual studies.
It is clear from the analysis of AI-based studies that the DL models provide the highest accuracy, and then comes the HDL and ML-based studies [14,20,49,50,51]. Various models are accessed to evaluate the performance of the studies. Most studies comment on the model’s accuracy. Few of them represent the sensitivity, specificity, area-under-the-curve (AUC), net present value (NPV), and F1-score. Figure 5 represents the graph of the performance metrics versus the number of studies.
Table 1 represents the 22 studies’ comments on the accuracy parameter, with eight studies representing evaluation in terms of sensitivity and specificity, and the parameters AUC (4 studies), MCC (3 studies), NPV (2 studies), F1 (one study) were mentioned in the research articles.
Out of a total of 29 studies, 9 studies (33%) used voice as input parameter for the detection of the PD, 5 studies (19%) used tremor data, 4 studies (15%) used sketch as the input parameter, 2 studies (7%) used EEG, and 2 studies (7%) uses telemedicine for diagnosis. From the studies, it is fair that the input parameter is the crucial factor for diagnosing the disease [3,52,53,54]. Figure 6a indicates the various distributions of the input dataset features for the diagnosis of PD. Figure 6b refers to the statistical distribution of the selected studies, which separates the main AI terms into the machine, deep, and hybrid studies. The input element for early predication of PD is important for the reckoning of bias in the studies.

3. Biology of Parkinson’s Disease

The declension of nerve cells in the substantianigra region of the brain causes PD. This part of the brain is responsible for producing a neurotransmitter called dopamine, which is originated by nerve cells. The role of dopamine is to act as a mediator between the brain and the elements of the sensory organs that govern and regulate physical movements [20]. The abundance of dopamine in the brain is lowered when these neurons die or become injured. This indicates that the part of the brain that controls movement cannot function correctly, resulting in slow, unwanted, and irregular movements of the body parts [55]. The death of nerve cells is a gradual process. When somewhere around 80% of the nerve cells in the substantianigra are damaged, signs of PD begin to appear [56]. Figure 7 depicts the clinical biology of Parkinson’s disease [38,57]. Although additional research is desired to find the exact cause for the loss of nerve cells associated with PD, there are no proper explanations for why it happens [3,11]. The cause of the disease is currently linked with a mix of environmental factors and genetic mutations. Several hereditary variables have been demonstrated to enhance a person’s risk of getting PD, while it is unidentified how these factors make certain people more susceptible to the disease [53,55].
The abnormal genes are transferred down from parents to children, and PD can run in families. However, this is a rare kind of legacy for the condition. According to some experts, environmental variables may also enhance a person’s risk of PD [57]. The use of pesticides and herbicides in agriculture and industrial pollution and traffic have been suggested as impactable causes to trigger PD. The data relating external factors with PD, on the other hand, are ambiguous [41,58].
The motor symptoms (risk factors) of PD can be used for classification (PD vs. Non-PD) using the AI-based model. The dataset was generated while evaluating the patients and can be easily put in a matrix form to develop the training model. The huge PD symptomatic data are generated including motor and non-motor PD risk factors. The symptomatic data cannot be statistically resolved, but an ML/DL/TL/HDL can be used to better understand both data classifications, leading to better PD detection [59]. While analyzing the symptomatic biology of PD, in-feature AI is the best option to quickly predict PD.

4. Artificial Intelligence Architectures

The Artificial Intelligence (AI)-based detection of the PD can be achieved by using symptoms (or risk factors) as an input parameter for the algorithm. The majority of the studies explain voice as a risk factor for the diagnosis of PD [52]. Tremor data are also an important (risk factor) in detecting PD [6]. The hybrid model includes two risk factors, which were also explained in a few articles [14,60].

4.1. A Note on Assumptions for Adaptation of the ML Algorithms

Different input parameters brought different assumptions. When the input is a tremor, if the shaking is prevalent in one body part (say the uncontrolled movement of hand), HDL such as ANN was preferred. Since NN could handle the augment, scale, and normalization, it preferred HDL. In the case when the input data was a voice, ML was preferred. In the case of voice datasets, the main assumption for the application of ML was to help diagnose the early and subtle signs of PD. In other cases, since the gold standard was available, the assumption was that the training models can be very powerful for the early diagnosis of PD. A certain set of ML algorithms such as principal component analysis was adopted due to a reduction in the dimension of the input datasets.
The features of the voice database can be better analyzed using decision tree or k-mean clustering methods, and such classifiers can be better suited for voice data classification for control vs. PD. Since the voice data were violating the data in components, it was assumed that by breaking the voice data into components and then feeding it into ML algorithms, such as hidden Markov models, then the learning of the voice data can be the superior method, followed by the detection process. A deep convolutional neural network classifier with transfer learning and data augmentation techniques can be used to identify the risk of the PD. The usage of handwriting data for the prediction of the PD faces a severe classification challenge at the preliminary stages due to the small size of data. The use of ImageNet and MNIST datasets were used as input sources independently to achieve good accuracy. For accurate identification of PD, other parallel PD symptoms data such as voice, freezing, and gait can be used

4.2. Architecture Based on Voice and Sketch Input

Anitha et al. [38] proposed a methodology (Figure 8) to predict PD using a clustering and classification algorithm. On the voice dataset, k-means, clustering and decision tree-based ML algorithms are evaluated using R-studio. Python is used to analyze the patient’s spiral artwork. Principal component analysis (PCA) is used to extract features from these illustrations. X, Y, Z, Tension, Grip Angle, Timestamp, and Test ID variables are derived from the spiral drawings. For comparison, two factors are used for the UCI dataset and drawing the data. In the study, the accuracy was demonstrated to be 76% and 91%, respectively. In comparison to other literature, the accuracy is low. It is feasible to improve the model’s accuracy by combining DL with an existing algorithm [38].

4.3. Architecture Based on Tremor

Bala et al. [6] have proposed architecture for early detection of PD by using ML-based classification methodology (Figure 9). Two types of data elements used for analysis were the tremor dataset and speech dataset. Data of 77 (PD) patients were used for experimentation purposes. By using the computer algorithm Multi-Dimensional Voice Program (MDVP), 33 acoustic parameters of a voice sample were calculated. The program that can calculate various algorithms such as K-mean, Random Forest, SVM, NB, and KNN is applied to the dataset. In both cases, accuracy was calculated for speech signal using NB (88.05%) and for tremor by using KNN (85.67%). The detailed design does not discuss any standard database used [6].

4.4. Architecture Based Speech Input with Information Gain Parameter

To predict PD, Cleick et al. [61] presented a variety of classification methods, including Regression analysis, Support Vector Machine, Extra Trees, Gradient Boosting, and Random Forest (Figure 10). In the classification stage, a total of 1208 voice data sizes were employed, with 26 features gathered from PD patients and non-patients. Classification results obtained using enlarged features beat classification obtained results using the data’s unique features. Random forest was used to get an IG accuracy of 72.69 percent [57].

5. Ranking of Selected Studies

Since some studies offer better AI model designs than others for early PD detection, it is important to understand which studies are more suitable for early PD detection. For this objective, one must rank these studies and evaluate the bias in their AI models. These studies can then be partitioned into certain bias bins, which can have their own AI characteristics. Note that the AI model performance is governed by the AI architecture and its components (so-called AI attributes). Thus, a study must have an evaluation criterion by which one can grade these AI attributes, which can then be used for evaluation or ranking.
The various architectures in the studies explain the role of AI in the detection of PD [55,59]. If the components of the AI architecture used for early detection of the PD have low performance, then the AI models under-performs, leading to lower grading of that study [57]. Attribute studies, combinations of the input parameter, and benchmarks associated with the clusters of the studies are essential factors that decide the ranking of the studies [55,56,58]. The detailed subsection explains the various parameters related to the raking of the studies.

5.1. Grading, Scoring, and Ranking of the Studies

Every study graded correlated with the attributes; a total of 30 attributes were considered for evaluation purposes and clustered into six sections. The cluster (C1) is related to publication and citation, (C2) is about the objective of the studies, (C3) explains the types of AI architecture used in the model, (C4) demonstrates optimization of the AI algorithms, (C4) analyzes the performance and evaluation of various AI models, (C6) is about clinical evaluation, scientific validation, and benchmarking. Every attribute in the respective cluster was evaluated for the evaluation purpose grading score method, as explained in Table A1 (Appendix A).
After interpreting the results of every cluster of the associated studies (26 studies) mean value, the absolute score cumulative score was computed. According to the mean value, absolute score, and cumulative score of the concerned studies, the ranking of the studies was finalized. The ranking studies are mentioned in Table 2 [61,62]. The green, yellow, and red flags indicate the impact of low-bias, moderate-bias, and high-bias on individual cluster cells.

5.2. Bias Cutoff Computation

About 26 studies were selected for the bias analysis that was closely associated with early detection of PD. Using AP(ai)Bias 1.0 (AtheroPointTM, Roseville, CA, USA), bias analysis was carried out. Studies were ranked into three AI bias categories (low moderate (ML) and high moderate (MH)) by computing the mean score and cumulative score for each study, taken for the AI attributes. The comparative analysis with various AI algorithms was carried out to determine the bias cutoff and to understand the architecture of these studies [59,63,64].
It is seen that many of the AI models show high accuracy, but the data size used for the testing and training of the algorithm is small, and the model fails to explain scientific validation. Hence, it results in High-Moderate (HM) in the studies [1,5,9,37,62,65]. The cumulative cutoff for the studies was determined by using various factors such as (i) associated studies of the PD, (ii) impact factor, (iii) the selected data, (iv) performance indicators, (v) clinical trials, etc. After analyzing the selected studies (26 studies), the cutoff was finalized for the high-bias <0.064 (8 studies), moderate-bias <0.078 (8 studies), and low-bias >0.078 (7 studies).
The Low-Moderate (LM) studies [1,5,9,37,62,65] observations are the articles containing information such as (i) high data count of the PD vs. normal; (ii) performance measures; (iii) comparative analysis with various ML, DL, and HDL algorithms; (iv) explanations of the benchmarking studies. The Moderate-bias studies [1,5,9,37,62,65,66] observations were (i) sufficient data, (ii) average impact factor, and (iii) comparison of the input parameters. The High-Moderate (HM) studies [3,6,54,60,67,68] observations associated with the articles were (i) a smaller number of data, (ii) insufficient dissuasion on the selected model, (iii) improper explanation of the algorithm, (iv) insufficient performance analysis, (v) lack of demographic discussion, and (vi) insufficient discussion on clinical evaluation. Based on the attribute analysis, every cluster was marked. The benchmarking and attribute analysis were not done. The algorithm with classifier optimization was not explained [15]. There are several explanations as to why and how the articles were frittered away for the research [63]. Figure 11 shows the cumulative cutoff score for the evaluation of the selected studies.
While noting the ranking studies, it is clear that selecting the architecture model for the proximate input is essential. It is linked with the performance of the model and RoB [37,69,70]. In the case that more than one input was taken for the diagnosis of the PD, the architecture paradigm and the performance of the model would change [49,68]. Hence, it is essential to discuss the linking of the architecture concerning input parameters for diagnosing PD [18,71,72]. Table A3 (Appendix C) discusses twelve studies linked with AI models’ performance parameters and compared them with input risk factors.

5.3. Linking of Bias with AI Architectures

The various databases contain the resultant features of the voice, sketch, tremor, face, EEG, and a biomarker of the PD patients concerning the normal [73]. UCI, PubMed, IEEE, and MJFox are the few names of the database providers. Some of the articles also include local datasets for the analysis of PD [60,74]. Figure 12a represents the various algorithms used for the detection of the PD studies. The SVM algorithm, along with Decision Tree, Naive Bias, and Random Forest, was used. Few articles compare various algorithms with each other and compare their performance evolutions [12,70,75].
Table A2 (Appendix B) explains the various statistical significance of the input features selection for the diagnosis of the PD and the performance parameter of various AI architectures [19,76]. The architecture uses a model with a classifier. Optimization was discussed in the third cluster. The fourth cluster related to evaluating the performance includes parameters such as accuracy, AUC, MCC, and F1. The evaluation and benchmarking sections discussed seen unseen data, as well as conformability of the data. Table A3 (Appendix C) represents the attribute analysis [67]. The basic model of AI consists of (a) PD vs. normal training and (b) risk label forecasting (risk possibilities) on test scenarios. As a result, these learning methods were categorized according to the type of results (scoring element) of the models, the category of classifiers, the clusters of predictor variables (risk factors), the predictive unbiased for the short or long term, the type of cross-validation procedure, scientific validation, and the outcome diagnosis. These aspects are crucial in determining performance as well as hazards that lead to bias.

5.4. Bias Distribution in AI Attributes

The tri-color scheme was implemented to represent the scientific analysis for low, moderate, and high-bias in the various attributes of the clusters. The Low-Moderate (LM) observations were done for articles containing information such as (i) high data count of the PD vs. normal; (ii) performance measures; (iii) comparative analysis with various ML, DL, and HDL algorithms; (iv) explanations of the benchmarking studies; (v) Implantation of the PRISMA model search strategy. The High-Moderate (HM) studies [3,6,54,60,67,68] observations associated with the articles were (i) less numbers of data, (ii) insufficient dissuasion on the selected model, (iii) improper explanation of the algorithm, (iv) insufficient performance analysis, (v) lack of demographic discussion, (vi) no comments on clinical evaluation, and (vii) unmentioned benchmarking of the attribute. It is observed in the bias distribution studies plot that most of the articles do not discuss the clinical evaluation and benchmarking, which lead to an increase in the high bias of the selected studies [3,6,54,60,67,68].
The insufficient optimization of the AI architectures with many inputs also leads to high bias. The good accuracy of the AI model but with failed test clinical validation results also leads to high bias. The comparative analysis with various AI algorithms was carried out to determine the bias cutoff and to understand the architecture of these studies. The cluster-wise bias distribution plot is shown in Figure 13.

5.5. Recommendations for Bias Reduction

The recommendation is an integral part of the study evaluation. We summarize the key recommendations, which can potentially improve the bias in AI for early PD detection, namely (i) Validation: the AI-based PD detection should be scientifically validated and clinically evaluated [39,52,77]; (ii) Fusion of covariates: is recommend that the AI model uses combinations of risk factors as an input parameter for the detection of PD [40] to ensure non-linearity is detected; (iii) Continental databases for AI generalization: use of the “continental multiethnic categorized dataset” and usage of power analysis (in big data framework), which will lead to improving true accuracy of early PD predication [72,78,79,80]; (iv) Non-motorized symptoms: “non-motor validated data” for PD (risk factors) data (Figure 7) are important risk factors for the AI models and must be included [58,81,82,83]; (v) Comorbidities: the PD risk factors due to “comorbidities” like COVID-19 [59,84,85,86,87], diabetes [23,24], and liver [88,89,90], thyroid [91,92], coronary [32,93,94], prostate [95], ovarian [96], and skin cancer [97,98] must also be considered.

6. Discussion

6.1. Principal Findings

PD is a non-curable disease, but at an early stage with a correct and precise diagnosis, we can control the progression of the disease. AI is a good option to detect an early-stage PD compared to the conventional PD detection approaches. However, there is a risk of bias in AI models due to lack of AI design attributes, which also includes gold standards (risk factors) of PD. This proposed review is the first to discuss AI bias analysis in the early detection of PD. As a result of this study, the outcomes are (i) Usage of computing 30 AI attributes (based on 6 AI clusters) scored by an AI expert, and computation of mean aggregate score; (ii) Computation of two cutoffs (Moderate-Low (ML) and High-Moderate (MH)) and determination of three bins: low-, moderate-, and high-bias. Additionally, (iii) it is seen that many of the AI models show high accuracy but the sample size used for the testing and training of the algorithm is relatively small. Further, the model fails to explain scientific validation; hence, it results in High-Moderate (HM) bias in the studies. (iv) For an AI system to be reliable, accurate, and to have a minimal AI distortion, the bias must be minimal. (v) AI architecture such as deep layered neural network models and such as the ANN model were neglected in clinical design and decisions (e.g., voice, tremor, sketch) and indicate Moderate-Low (ML) bias in the ranking [13,62,99].

6.2. Benchmarking

Table 3 shows the benchmarking analysis of the eight selected AI studies. We have also mentioned various important aspects of the review related to early PD detection by using AI [100,101]. The demographic analysis of the PD is mentioned in column (B3). While analyzing demographics, we can find important factors such as the continent/country that is leading and lagging in major/minor cases of PD patients [43,102]. The (B4) column benchmarking table represents the objective of the studies. Most of the studies represent the comparative analysis of a normal person to a person diagnosed with PD [54]. Column (B4) is related to the inclusion and exclusion criteria of the studies. As per the disease symptoms point of view, most of the symptoms under the tree of neurodegenerative diseases such as Alzheimer’s, Huntington’s, Adrenoleukodystrophy (ALD), and PD are similar [37,43,58]. Few symptoms of the disease among them are different. When selecting the articles for the proposed study, we tried focusing on the symptoms related to PD [103,104]. As mentioned in column (B5), the data extraction criteria from the various sources are important to focus on the area of interest in the study [51,104]. The various AI models used in the studies are mentioned in column (B6), and it has been seen that in most of the article, ML algorithms were used to detect PD. The performance of various AI models is shown in Table A3 (Appendix C) [63]. The studies using the PRISMA model strategy for selecting the article were verified and are shown in column (B7) [4,32]. Column (B9) represents the risk factor as an input parameter analysis for the early detection of the PD. The early symptoms of PD are compulsiveness in movement, voice changes, and movement problems [4,16,17]. It is easy to predict the disease by observing the change in motion of the body parts such as freezing of the shoulder [6,14]. The column (B10, B11, B12, B13, and B14) represents benchmarking observations, cross-validation, bias studies, and scientific validation, respectively, but most of the selected studies failed to explain those terminologies. The last row depicts “Proposed,” which is about the current study. Note that we indicated “√” in places of solitary benefaction in the review.

6.3. A Short Note on Bias in ML

PD is a non-curable disease, even though the treatment cost of PD is very high. To avoid death and economic loss due to the late diagnosis of PD, early diagnosis of PD is very important. AI is a good option to detect the early stage of PD compared to the conventional PD detection approach, but compared to the conventional PD detection approach, there is a risk of implementing an AI model. An AI model is evaluated based on accuracy only, but the model fails to explain scientific validation and clinical validation. Further, there is a lack of evidence on the generalization of AI models; hence, it results in High-Moderate (HM) bias in the AI model. Many of the AI models show high accuracy, but the data size used for the testing and training of the algorithm is small; thus, it results in High-Moderate (HM) bias in the AI model. The AI-based detection of PD can be achieved by using symptoms (or risk factors) as an input parameter for the algorithm. The majority of the studies explain voice, tremor, gait, and sketches as risk factors for the diagnosis of PD [14,52,60]. It is seen that the AI model uses combinations of risk factors as the input parameter for the detection of PD, having a Low-Moderate bias.
The studies were used to compute 30 AI attributes (based on 6 AI clusters). The PD risk is intensifying due to existing comorbidities with PD; hence, it results in High-Moderate (HM) bias in the AI model. By adding more attributes such as comorbidities with PD, gender studies of PD patients, and clinical validation of AI-assisted PD detection, the grading score of the studies will be improved. Therefore, there is scope to minimize the High-Moderate bias in the AI model [83,102].

6.4. A Short Note PD Database and Gender Studies

Figure 12c shows the demographic distribution of the various continents, American (60 years), Europe (61 years), Australian (55 years), and Asian (56 years), and the average age of the PD patients in these respective continents [102,105,106].
Furthermore, their risk of granularities of a database to predict the PD results High-Moderate (HM) bias in the AI model. As lifestyle, environmental conditions, and human factors vary with the continents, attributes of the dataset will also vary. Thus, the unavailability of the continental categorized dataset of the PD AI model leads to High-Moderate (HM) risk. The average age of the PD patient is 57.77 years, and most of the database contains the age group of the patients between 50 to 60 [8,102]. Hence, the majority of risk factors are probably affecting PD patients in the age group of 50 to 60 years [59,75,107]. The PD risk is intensifying due to existing comorbidities. If we eliminate the associated comorbidities with PD to train the model, it results in High-Moderate (HM) bias in the AI model. There is no study of Age/Gender in certain ethnicities, and without this, the bias will erupt. Such a system that has not included the diversity in age will fail in the prediction models if the training is not also correct, so there is the risk of generating high bias in the model.

6.5. Role of Human-Computer Interface in Early Detection of the PD

Human-computer interaction (HCI) studies the interaction among humans and computers, providing indicators that may be used to assess a user’s physiological, behavioral, and psychological states, for example. Computers, cellphones, tablets, gaming platforms, and wearable technologies all fall under the heading of human-computer interaction (HCI). By using HCI, it is easy to predict early PD motor symptoms, for example, by monitoring the keyboard or touch screen of smartphone operating response from the user. There seem to be a variety of features present during typing on a keypad, according to current studies on PD diagnosis through different motor symptoms, including reaction speed to messages, uneven movement of the figures, typing pattern, degradation of repetitive movements, stiffness in figures, indications of sidedness, deterioration in repetitive motion and typing of sequences of letters, changes of motion and signs of hand and finger muscle spasms, and Jerkiness of movement. Therefore, the HCI parameters can be considered for the early detection of PD [1].

6.6. Strengths, Weakness, and Extensions of Our Study

The main strength of the study is the ability to automatically compute the RoB given the scored AI attributes by an expert in the AI field. These attributes were an amalgamation of demographics, AI architecture, performance evaluation, scientific validation, clinical evaluation, and big data analysis, framed into six clusters [108,109]. The second component was to compute the aggregate score for each of the AI studies, followed by an estimation of two cutoffs (Moderate-Low (ML) and High-Moderate (MH)) to classify the studies into three bins: low-, moderate-, and high-bias. The study further provides new insight into the building blocks of AI-based early PD detection such as architectural differences, input risk factors, and limited databases, which are the key elements responsible for RoB in the AI model. Further, the study presented a set of key recommendations for improving the RoB. The studies lacked discussions on database size, comorbidities with PD, gender information in PD, continental databases, and clinical validations of AI-assisted PD detection. By adding relevant, meaningful, and quality attributes to benchmarking, the RoB of the AI model may also be improved [7,20,86,110]. Some studies may help to observe the PD study of problem-solving and executive function.
Due to a lack of research funding and the non-involvement of the leading worldwide groups in the field of AI, the benchmarking section was compromised in quality. Even though it was a pilot study, due to a lack of AI participation in the PD field, the RoB has the potential for exhaustive analysis. Further, due to the COVID-19 pandemic, the PD research funds are limited and, therefore, PD research has been less attractive [86,111].
We expect to see more systematic reviews using DL and HDL models. Further, other neurological diseases such as Alzheimer’s and Adrenoleukodystrophy (A.L.D.) [112,113], when aligned to PD, can be explored for more robust scoring, ranking, and classification using advanced neural imaging tools [69,114,115]. Currently, the world is facing a COVID-19 pandemic, where 26 million people are affected and 5.2 million have died due to the coronavirus. COVID-19 has strongly affected neurological diseases due to its brain pathway [11,116]. Further, several comorbidities like diabetes, renal disease, and coronary artery disease have intensified in COVID-19 patients, causing pulmonary embolism [59,111]. Several AI tools have been researched and recommended for COVID-19 applications [86,117]. Just like one can characterize the lung or pulmonary COVID-19 data [110,118], there can be PD neurological imaging data on COVID-19 patients that can be analyzed. Recently, bias estimation on COVID-19 patients was designed and developed [59]. In the future, we anticipate more systematic reviews on PD-based RoB with comorbidities focusing on the COVID-19 virus [59,84,85,86,87].

7. Conclusions

To our knowledge, this is the unique review that contains RoB elements selected from all 26 research articles that used machine-learning, solo deep-learning, and hybrid-learning algorithms to diagnose PD. We shared our findings, which included studies in a high-level summary, such as (i) the AI is an essential component for the diagnosis of the early PD detection and the recommendations must be followed to lower the RoB; (ii) the studies were ranked and two cutoffs (Moderate-Low (ML) and High-Moderate (MH)) were determined to segregate the studies into three bins: low-, moderate-, and high-bias); (iii) clinical, behavioral, and biomarker data categories were useful while verifying symptoms of the PD; (iv) possible patients biomarkers and physical indicators that are very important for making a more accurate diagnosis for helping healthcare decision-making. We recommend (i) the usage of power analysis in big data framework, (ii) that it must undergo scientific validation using unseen AI models, and (iii) further adaptation in clinical evaluation for reliability and stability tests.
The accomplishment of AI-assisted PD diagnosis holds great promise for a more systematic clinical decision-making system, and the use of innovative biomarkers would lower the bias and make it easier to understand drugs. Diagnosis of PD at an early onset will be feasible and faster with the help of AI techniques. Approaches to AI may give clinicians more valuable information for screening, detection, and diagnosis techniques towards the early detection of PD disease.

Author Contributions

Conceptualization, S.P., M.M. and J.S.S.; Methodology and software, M.M. and J.S.S.; Validation, M.M., M.T., P.R.K. and J.S.S.; Investigation, S.P., S.S. and J.S.S.; Resources, S.P.; Data curation, M.M, L.S., P.R.K. and J.S.S.; Writing—original draft preparation, M.M., L.S. and J.S.S.; Writing—review and editing, S.P., M.M., S.S., L.S., M.T., M.K., P.R.K. and J.S.S. Visualization, S.P. and J.S.S.; Supervision, S.P. and J.S.S.; Project administration, S.P. and J.S.S. All authors have read and agreed to the published version of the manuscript.

Funding

We would like to acknowledge the Department of Science and Technology, Government of India for sponsoring the project under the scheme IMPRINT-2 vide file no: IMP/2018/000034, Dated: 28/03/2019.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No data availability.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

SNAbb*DefinitionSNAbb*Definition
1AIArtificial intelligence25HMIHuman-machine interface
2AAAttribute Analysis26IECInclusive and Exclusive criteria
3AUCArea under curve27IPAInput parameter analysis
4ACCAccuracy28InParmInput parameter
5ARCArchitecture29MAEMean absolute error
6BABenchmarking analysis30MEModel Evaluation
7BNBatch normalization31MLMachine learning
8BSBias studies32NPVNegative predictive value
9BIBrain interface33MSEMean square error
10OBObjective34PCAPrincipal component analysis
11CVCross-validation35PEPerformance evaluation indicator
12CTClassifier type36RoBRisk of bias
13CNNConvolution neural network37RNNRecurrent neural network
14CONVConvolution38RFRandom forest
15CVDCardiovascular disease39SENSensitivity
16DLDeep learning40SVScientific validation
17DTDecision tree41SDLSolo deep learning
18DEData Extraction42SPESpecificity
19DDDemographic discussion43SVMSupport vector machine
20DSData set44PPrecision
21DSEDataset Size45RRecall
22ETEthnicity46RSReference studies
23EEGElectroencephalography47PDParkinson’s Disease
24HARHuman activity recognition48Abb*Abbreviations

Appendix A

Every study graded correlated with the attributes; a total of 30 attributes were considered for evaluation purposes and clustered into six sections. The interpret grading is applied to every cluster, according to the explanation, and every cluster was evaluated.
Table A1. Grading sheet.
Table A1. Grading sheet.
Cluster TypeA#Name of Attributes#A/CGrading Scheme (G*)
Cluster 1
(Publications Details)
A1Citation35 (G = 3);
3 (G < 3);
1 (G < 2)
A2Year of Publication
A3Impact Factor
Cluster 2
(Objective)
A4Objective45 (G = 4);
4 (G < 3);
3 (G < 2);
1 (G = 1)
A5Dataset Used
A6Dataset Size
A7Diagnosis Method
Cluster 3
(AI Architecture)
A8AI Type75 (G = 7);
4 (G < 5);
3 (G < 4);
2 (G< 3);
1 (G < 2);
0 (G = 1);
A9Architecture Used
A10Internal Layers Used
A11Type of Classifiers
A12Data Pre-Processing
A13Feature Extraction
A14Activation Function
Cluster 4
(Optimization)
A15Learning/Optimization Algorithm35 (G = 3);
4 (G < 2);
2 (G = 1);
0 (G = 0)
A16Evaluation Metrics Used for Classification
A17Comparison With
Cluster 5
(Performance)
A18Accuracy75 (G = 7);
4 (G < 5);
3 (G < 4);
2 (G < 3);
0 (G = 1)
A19Sensitivity
A20Specificity
A21AUC
A22MCC
A23NPV
A24F1
Cluster 6
Clinical Evaluation and Benchmarking
A25Demographic65 (G = 6);
4 (G < 5);
3 (G < 4);
2 (G < 3);
0 (G = 1);
A26Age
A27Ethnicity
A28Validation
A29Seen Vs. Unseen
A30Treatment
A#: Attribute Number, A/C: # of attributes per cluster; G*: # of Qualifying attributes per cluster.

Appendix B

Attributes studies of 11 articles for the early detection of PD by using AI. To interpret the results of every study, a systematic approach of attributes analysis and performance indication was completed.
Table A2. Studies vs. Attributes.
Table A2. Studies vs. Attributes.
SNA0A1A2A3A4A5A6A7A8
CitationsDSDSEETAge
(yrs)
InPramArchCTACC (%)
1Alzubaidi et al. [30] (2021)ACM1011Asian60TremorHDLSVM, CNN87.9
2Ahmed et al. [31] (2021)UCI104Asian60VoiceMLRNN95.8
3Mei et al. [17] (2021)PubMed, IEEE209Europe60VoiceMLSVM83.07
4Singamaneni et al. [22] (2021)UCI410Asian50VoiceMLLDR94.86
5Jaichandran et al. [20] (2020)UCI129Asian60VoiceMLSVM, ET, K-Mean78.34
6Anitha et al. [6] (2020)UCI467Asian50VoiceMLCNN90.21
7Maitín et al. [15] (2020)ACM, IEEE780America60EEGMLSVM62.99
8Poorjam et al. [13] (2019)PPMI24Australia50VoiceHDLiHMM96.00
9Aseer et al. [1] (2019)MNIST255Asian65HandwritingSDLCNN98.28
10Naghsh et al. [35] (2019)ELAB20Asian50EEGSDLICA, SVM, K-Mean97.38
11Wang et al. [6] (2017)PPMI584Asian50BiomarkerHDLCNN, SVM, RF, NB, BT96.12
DS: Dataset, DSE: Dataset Size, ET: Ethnicity, InParm: Input Parameter, ARC: Architecture, CT: Classifier Type, ACC: Accuracy.

Appendix C

Performance parameters of 12 studies aligned with the type of input and AI architectures. The AI-based detection of the PD can be achieved by using symptoms as an input parameter for the algorithm. The majority of the studies explain voice as an input parameter for the diagnosis of PD. Tremor, EEG, sketch, and biomarker (chemical) data are also important input parameters to detect the PD.
Table A3. Twelve studies showing input data type, AI architecture, and performance parameters.
Table A3. Twelve studies showing input data type, AI architecture, and performance parameters.
Attributes (Left to Right)A0A1A2A3A4A5A6A7
CitationsIPAIACCSENSPEAUCMCCF1
Alzubaidi et al. [30] (2021)TremorHDL87.9---89.341.17
Ahmed et al. [31] (2021)VoiceML95.890.2492.3-92.0396
Mei et al. [17] (2021)VoiceML83.07--0.91--
Singamaneni et al. [22] (2021)VoiceML94.86-----
Jaichandran et al. [20] (2020)VoiceML78.34-----
Anitha et al. [6] (2020)VoiceML90.211.84.392.49 1.17
Maitín et al. [15] (2020)EEGML62.990.90670.981---
Poorjam et al. [13] (2019)VoiceHDL96.00-----
Aseer et al. [1] (2019)HandwritingSDL98.28----
Naghsh et al. [35] (2019)EEGSDL97.380.98910.987---
Wang et al. [6] (2017)BiomarkerHDL96.12-----
IP: Input Parameter, ACC: Accuracy, SEN: Sensitivity, SPE: Specificity, MCC: Matthew’s correlation coefficient.

References

  1. Aal, H.A.A.E.; Taie, S.A.; El-Bendary, N. An optimized RNN-LSTM approach for parkinson’s disease early detection using speech features. Bull. Electr. Eng. Inform. 2021, 10, 2503–2512. [Google Scholar] [CrossRef]
  2. Priya, S.J.; Rani, A.J.; Subathra, M.; Mohammed, M.A.; Damaševičius, R.; Ubendran, N. Local pattern transformation based feature extraction for recognition of Parkinson’s disease based on gait signals. Diagnostics 2021, 11, 1395. [Google Scholar] [CrossRef]
  3. Bhat, S.; Acharya, U.R.; Hagiwara, Y.; Dadmehr, N.; Adeli, H. Parkinson’s disease: Cause factors, measurable indicators, and early diagnosis. Comput. Biol. Med. 2018, 102, 234–241. [Google Scholar] [CrossRef]
  4. Sabeena, B.; Sivakumari, S.; Amudha, P. A technical survey on various machine learning approaches for Parkinson’s disease classification. Mater. Today Proc. 2020. [Google Scholar] [CrossRef]
  5. Naseer, A.; Rani, M.; Naz, S.; Razzak, M.I.; Imran, M.; Xu, G. Refining Parkinson’s neurological disorder identification through deep transfer learning. Neural Comput. Appl. 2019, 32, 839–854. [Google Scholar] [CrossRef] [Green Version]
  6. Neharika, D.B.; Anusuya, S. Machine Learning Algorithms for Detection of Parkinson’s Disease using Motor Symptoms: Speech and Tremor. IJRTE 2020, 8, 47–50. [Google Scholar]
  7. Sriram, T.V.; Rao, M.V.; Narayana, G.S.; Kaladhar, D.; Vital, T.P.R. Intelligent Parkinson disease prediction using machine learning algorithms. Int. J. Eng. Innov. Technol. 2013, 3, 1568–1572. [Google Scholar]
  8. Liu, R.; Umbach, D.M.; Peddada, S.D.; Xu, Z.; Tröster, A.I.; Huang, X.; Chen, H. Potential sex differences in nonmotor symptoms in early drug-naive Parkinson disease. Neurology 2015, 84, 2107–2115. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Alzubaidi, M.; Shah, U.; Zubaydi, H.D.; Dolaat, K.; Abd-Alrazaq, A.; Ahmed, A.; Househ, M. The Role of Neural Network for the Detection of Parkinson’s Disease: A Scoping Review. Healthcare 2021, 9, 740. [Google Scholar] [CrossRef] [PubMed]
  10. Moore, S.T.; MacDougall, H.; Ondo, W.G. Ambulatory monitoring of freezing of gait in Parkinson’s disease. J. Neurosci. Methods 2008, 167, 340–348. [Google Scholar] [CrossRef] [PubMed]
  11. Ahlrichs, C.; Lawo, M. Parkinson’s Disease Motor Symptoms in Machine Learning: A Review. Health Inform.-Int. J. 2013, 2, 1–18. [Google Scholar] [CrossRef]
  12. Maitín, A.M.; García-Tejedor, A.J.; Muñoz, J.P.R. Machine Learning Approaches for Detecting Parkinson’s Disease from EEG Analysis: A Systematic Review. Appl. Sci. 2020, 10, 8662. [Google Scholar] [CrossRef]
  13. Mei, J.; Desrosiers, C.; Frasnelli, J. Machine Learning for the Diagnosis of Parkinson’s Disease: A Review of Literature. Front. Aging Neurosci. 2021, 13, 633752. [Google Scholar] [CrossRef]
  14. Isaacs, D. Artificial intelligence in health care. J. Paediatr. Child Health 2020, 56, 1493–1495. [Google Scholar] [CrossRef]
  15. Anitha, R.; Nandhini, T.; Raj, S.S.; Nikitha, V. Early detection of parkinson’s disease using machine learning. IEEE Access 2020, 8, 147635–147646. [Google Scholar]
  16. Challa, K.N.R.; Pagolu, V.S.; Panda, G.; Majhi, B. An improved approach for prediction of Parkinson’s disease using machine learning techniques. In Proceedings of the 2016 International Conference on Signal Processing, Communication, Power and Embedded System (SCOPES), Paralakhemundi, India, 3–5 October 2016; pp. 1446–1451. [Google Scholar] [CrossRef] [Green Version]
  17. Wroge, T.J.; Özkanca, Y.; Demiroglu, C.; Si, D.; Atkins, D.C.; Ghomi, R.H. Parkinson’s disease diagnosis using machine learning and voice. In Proceedings of the 2018 IEEE Signal Processing in Medicine and Biology Symposium (SPMB), Philadelphia, PA, USA, 1–7 December 2018. [Google Scholar]
  18. Shimoda, A.; Li, Y.; Hayashi, H.; Kondo, N. Dementia risks identified by vocal features via telephone conversations: A novel machine learning prediction model. PLoS ONE 2021, 16, e0253988. [Google Scholar] [CrossRef] [PubMed]
  19. Wibawa, M.S.; Nugroho, H.A.; Setiawan, N.A. Performance evaluation of combined feature selection and classification methods in diagnosing parkinson disease based on voice feature. In Proceedings of the 2015 International Conference on Science in Information Technology (ICSITech) Yogyakarta, Indonesia, 27–28 October 2015; pp. 126–131. [Google Scholar] [CrossRef]
  20. Saba, L.; Biswas, M.; Kuppili, V.; Godia, E.C.; Suri, H.S.; Edla, D.R.; Omerzu, T.; Laird, J.R.; Khanna, N.N.; Mavrogeni, S.; et al. The present and future of deep learning in radiology. Eur. J. Radiol. 2019, 114, 14–24. [Google Scholar] [CrossRef]
  21. Tandel, G.S.; Biswas, M.; Kakde, O.G.; Tiwari, A.; Suri, H.S.; Turk, M.; Laird, J.R.; Asare, C.K.; Ankrah, A.A.; Khanna, N.N.; et al. A Review on a Deep Learning Perspective in Brain Cancer Classification. Cancers 2019, 11, 111. [Google Scholar] [CrossRef] [Green Version]
  22. Biswas, M.; Kuppili, V.; Saba, L.; Edla, D.R.; Suri, H.S.; Cuadrado-Godia, E.; Laird, J.R.; Marinhoe, R.T.; Sanches, J.M.; Nicolaides, A.J.F.B. State-of-the-art review on deep learning in medical imaging. Front. Biosci. 2019, 24, 392–426. [Google Scholar]
  23. Maniruzzaman, M.; Kumar, N.; Abedin, M.M.; Islam, M.S.; Suri, H.S.; El-Baz, A.S.; Suri, J.S. Comparative approaches for classification of diabetes mellitus data: Machine learning paradigm. Comput. Methods Programs Biomed. 2017, 152, 23–34. [Google Scholar] [CrossRef]
  24. Maniruzzaman, M.; Rahman, J.; Hasan, A.M.; Suri, H.S.; Abedin, M.; El-Baz, A.; Suri, J.S. Accurate Diabetes Risk Stratification Using Machine Learning: Role of Missing Value and Outliers. J. Med. Syst. 2018, 42, 92. [Google Scholar] [CrossRef] [Green Version]
  25. Saba, L.; Sanagala, S.S.; Gupta, S.K.; Koppula, V.K.; Johri, A.M.; Khanna, N.N.; Mavrogeni, S.; Laird, J.R.; Pareek, G.; Miner, M.; et al. Multimodality carotid plaque tissue characterization and classification in the artificial intelligence paradigm: A narrative review for stroke application. Ann. Transl. Med. 2021, 9, 1206. [Google Scholar] [CrossRef]
  26. Jamthikar, A.D.; Gupta, D.; Mantella, L.E.; Saba, L.; Laird, J.R.; Johri, A.M.; Suri, J.S. Multiclass machine learning vs. conventional calculators for stroke/CVD risk assessment using carotid plaque predictors with coronary angiography scores as gold standard: A 500 participants study. Int. J. Cardiovasc. Imaging 2020, 37, 1171–1187. [Google Scholar] [CrossRef] [PubMed]
  27. Jamthikar, A.; Gupta, D.; Saba, L.; Khanna, N.N.; Araki, T.; Viskovic, K.; Mavrogeni, S.; Laird, J.R.; Pareek, G.; Miner, M.; et al. Cardiovascular/stroke risk predictive calculators: A comparison between statistical and machine learning models. Cardiovasc. Diagn. Ther. 2020, 10, 919–938. [Google Scholar] [CrossRef] [PubMed]
  28. Banchhor, S.K.; Londhe, N.D.; Araki, T.; Saba, L.; Radeva, P.; Laird, J.R.; Suri, J.S. Wall-based measurement features provides an improved IVUS coronary artery risk assessment when fused with plaque texture-based features during machine learning paradigm. Comput. Biol. Med. 2017, 91, 198–212. [Google Scholar] [CrossRef] [PubMed]
  29. Acharya, U.R.; Faust, O.; Sree, S.V.; Molinari, F.; Garberoglio, R.; Suri, J.S. Cost-effective and non-invasive automated benign & malignant thyroid lesion classification in 3D contrast-enhanced ultrasound using combination of wavelets and textures: A class of ThyroScan™ algorithms. Technol. Cancer Res. Treat. 2011, 10, 371–380. [Google Scholar]
  30. Acharya, U.R.; Faust, O.; Sree, S.V.; Molinari, F.; Suri, J.S. ThyroScreen system: High resolution ultrasound thyroid image characterization into benign and malignant classes using novel combination of texture and discrete wavelet transform. Comput. Methods Programs Biomed. 2011, 107, 233–241. [Google Scholar] [CrossRef]
  31. Kuppili, V.; Biswas, M.; Sreekumar, A.; Suri, H.S.; Saba, L.; Edla, D.R.; Marinhoe, R.T.; Sanches, J.M.; Suri, J.S. Extreme learning machine framework for risk stratification of fatty liver disease using ultrasound tissue characterization. J. Med. Syst. 2017, 41, 152. [Google Scholar] [CrossRef]
  32. Pareek, G.; Acharya, U.R.; Sree, S.V.; Swapna, G.; Yantri, R.; Martis, R.J.; Saba, L.; Krishnamurthi, G.; Mallarini, G.; El-Baz, A.J. Prostate tissue characterization/classification in 144 patient population using wavelet and higher order spectra features from transrectal ultrasound images. Technol. Cancer Res. Treat. 2013, 12, 545–557. [Google Scholar] [CrossRef]
  33. McClure, P.; Elnakib, A.; El-Ghar, M.A.; Khalifa, F.; Soliman, A.; El-Diasty, T.; Suri, J.S.; Elmaghraby, A.; El-Baz, A. In-Vitro and In-Vivo Diagnostic Techniques for Prostate Cancer: A Review. J. Biomed. Nanotechnol. 2014, 10, 2747–2777. [Google Scholar] [CrossRef]
  34. Acharya, U.R.; Saba, L.; Molinari, F.; Guerriero, S.; Suri, J.S. Ovarian tumor characterization and classification: A class of GyneScan™ systems. In Proceedings of the 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Diego, CA, USA, 28 August–1 September 2012. [Google Scholar]
  35. Acharya, U.R.; Sree, S.V.; Saba, L.; Molinari, F.; Guerriero, S.; Suri, J.S. Ovarian Tumor Characterization and Classification Using Ultrasound—A New Online Paradigm. J. Digit. Imaging 2012, 26, 544–553. [Google Scholar] [CrossRef] [Green Version]
  36. El-Baz, A.; Suri, J.S. Big Data in Multimodal Medical Imaging; CRC Press: New York, NY, USA, 2019. [Google Scholar] [CrossRef]
  37. Jo, T.; Nho, K.; Saykin, A.J. Deep Learning in Alzheimer’s Disease: Diagnostic Classification and Prognostic Prediction Using Neuroimaging Data. Front. Aging Neurosci. 2019, 11, 220. [Google Scholar] [CrossRef] [Green Version]
  38. Watts, J.; Khojandi, A.; Shylo, O.; Ramdhani, R.A. Machine learning’s application in deep brain stimulation for Parkinson’s disease: A review. Brain Sci. 2020, 10, 809. [Google Scholar] [CrossRef] [PubMed]
  39. Pereira, C.R.; Pereira, D.R.; da Silva, F.A.; Hook, C.; Weber, S.A.; Pereira, L.A.; Papa, J.P. A Step Towards the Automated Diagnosis of Parkinson’s Disease: Analyzing Handwriting Movements. In Proceedings of the 2015 IEEE 28th International Symposium on Computer-Based Medical Systems, Sao Carlos, Brazil, 22–25 June 2015; pp. 171–176. [Google Scholar] [CrossRef]
  40. Adams, W.R. High-accuracy detection of early Parkinson’s Disease using multiple characteristics of finger movement while typing. PLoS ONE 2017, 12, e0188226. [Google Scholar] [CrossRef] [Green Version]
  41. Gill, S.; Mouches, P.; Hu, S.; Rajashekar, D.; MacMaster, F.P.; Smith, E.E.; Forkert, N.D.; Ismail, Z. Using Machine Learning to Predict Dementia from Neuropsychiatric Symptom and Neuroimaging Data. J. Alzheimer’s Dis. 2020, 75, 277–288. [Google Scholar] [CrossRef] [Green Version]
  42. Liu, X.; Rivera, S.C.; Moher, D.; Calvert, M.J.; Denniston, A.K.; Chan, A.-W.; Darzi, A.; Holmes, C.; Yau, C.; Ashrafian, H.; et al. Reporting guidelines for clinical trial reports for interventions involving artificial intelligence: The CONSORT-AI extension. Nat. Med. 2020, 26, 1364–1374. [Google Scholar] [CrossRef]
  43. Alroobaea, R.; Mechti, S.; Haoues, M.; Rubaiee, S.; Ahmed, A.; Andejany, M.; Bragazzi, N.L.; Sharma, D.K.; Kolla, B.P.; Sengan, S. Alzheimer’s Disease Early Detection Using Machine Learning Techniques. Front. Neurosci. 2020. Available online: https://assets.researchsquare.com/files/rs-624520/v1/b83914f7-3a09-4ff1-9456-8288ae815f20.pdf?c=1631885103 (accessed on 2 December 2021).
  44. Battineni, G.; Chintalapudi, N.; Amenta, F. Comparative Machine Learning Approach in Dementia Patient Classification using Principal Component Analysis. In Proceedings of the 12th International Conference on Agents and Artificial Intelligence, Valletta, Malta, 22–24 February 2020. [Google Scholar]
  45. Pondal, M.; Marras, C.; Miyasaki, J.; Moro, E.; Armstrong, M.J.; Strafella, A.P.; Shah, B.B.; Fox, S.; Prashanth, L.K.; Phielipp, N.; et al. Clinical features of dopamine agonist withdrawal syndrome in a movement disorders clinic. J. Neurol. Neurosurg. Psychiatry 2012, 84, 130–135. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Poorjam, A.H.; Kavalekalam, M.S.; Shi, L.; Raykov, J.P.; Jensen, J.R.; Little, M.A.; Christensen, M.G. Automatic quality control and enhancement for voice-based remote Parkinson’s disease detection. Speech Commun. 2020, 127, 1–16. [Google Scholar] [CrossRef]
  47. Arroyo-Gallego, T.; Ledesma-Carbayo, M.J.; Sanchez-Ferro, A.; Butterworth, I.; Mendoza, C.S.; Matarazzo, M.; Montero, P.; Lopez-Blanco, R.; Puertas-Martin, V.; Trincado, R.; et al. Detection of Motor Impairment in Parkinson’s Disease Via Mobile Touchscreen Typing. IEEE Trans. Biomed. Eng. 2017, 64, 1994–2002. [Google Scholar] [CrossRef] [PubMed]
  48. Khan, S.; Gill, S.; Mooney, L.; White, P.; Whone, A.; Brooks, D.; Pavese, N. Combined pedunculopontine-subthalamic stimulation in Parkinson disease. Neurology 2012, 78, 1090–1095. [Google Scholar] [CrossRef] [PubMed]
  49. Deepa, S.P.C.N. A Deep Learning Method on Medical Image Dataset Predicting Early Dementia in Patients Alzheimer’s Disease using Convolution Neural Network (CNN). Int. J. Recent Technol. Eng. 2019, 8, 604–609. [Google Scholar]
  50. Araki, T.; Ikeda, N.; Shukla, D.; Jain, P.K.; Londhe, N.D.; Shrivastava, V.K.; Banchhor, S.K.; Saba, L.; Nicolaides, A.; Shafique, S. PCA-based polling strategy in machine learning framework for coronary artery disease risk assessment in intravascular ultrasound: A link between carotid and coronary grayscale plaque morphology. Comput. Methods Programs Biomed. 2016, 128, 137–158. [Google Scholar] [CrossRef] [PubMed]
  51. Prashanth, R.; Roy, S.D. Early detection of Parkinson’s disease through patient questionnaire and predictive modelling. Int. J. Med. Inform. 2018, 119, 75–87. [Google Scholar] [CrossRef] [Green Version]
  52. Al-Wahishi, A.; Belal, N.; Ghanem, N. Diagnosis of Parkinson’s Disease by Deep Learning Techniques Using Handwriting Dataset. In Proceedings of the International Symposium on Signal Processing and Intelligent Recognition Systems, Chennai, India, 14–17 October 2020. [Google Scholar]
  53. Rao, K.M.M.; Reddy, M.S.N.; Teja, V.R.; Krishnan, P.; Aravindhar, D.J.; Sambath, M. Parkinson’s Disease Detection Using Voice and Spiral Drawing Dataset. In Proceedings of the 2020 Third International Conference on Smart Systems and Inventive Technology (ICSSIT), Online, 20–22 August 2020. [Google Scholar]
  54. Eskofier, B.M.; Lee, S.I.; Daneault, J.-F.; Golabchi, F.N.; Ferreira-Carvalho, G.; Vergara-Diaz, G.; Sapienza, S.; Costante, G.; Klucken, J.; Kautz, T. Recent machine learning advancements in sensor-based mobility analysis: Deep learning for Parkinson’s disease assessment. In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016. [Google Scholar]
  55. Jena, B.; Saxena, S.; Nayak, G.K.; Saba, L.; Sharma, N.; Suri, J.S. Artificial intelligence-based hybrid deep learning models for image classification: The first narrative review. Comput. Biol. Med. 2021, 137, 104803. [Google Scholar] [CrossRef]
  56. Raees, P.C.M.; Thomas, V. Automated detection of Alzheimer’s Disease using Deep Learning in MRI. J. Physics: Conf. Ser. 2021, 1921, 012024. [Google Scholar] [CrossRef]
  57. Oriol, J.D.V.; Vallejo, E.E.; Estrada, K.; Peña, J.G.T.; Initiative, T.A.D.N. Benchmarking machine learning models for late-onset alzheimer’s disease prediction from genomic data. BMC Bioinform. 2019, 20, 709–717. [Google Scholar] [CrossRef]
  58. Antor, M.B.; Jamil, A.; Mamtaz, M.; Monirujjaman, M.; Khan; Aljahdali, S.; Kaur, M.; Singh, P.; Masud, M.A. Comparative Analysis of Machine Learning Algorithms to Predict Alzheimer’s Disease. J. Healthc. Eng. 2021, 2021, 9917919. Available online: https://www.hindawi.com/journals/jhe/2021/9917919/ (accessed on 2 December 2021).
  59. Suri, J.S.; Agarwal, S.; Gupta, S.K.; Puvvula, A.; Viskovic, K.; Suri, N.; Alizad, A.; El-Baz, A.; Saba, L.; Fatemi, M.; et al. Systematic Review of Artificial Intelligence in Acute Respiratory Distress Syndrome for COVID-19 Lung Patients: A Biomedical Imaging Perspective. IEEE J. Biomed. Health Inform. 2021, 25, 4128–4139. [Google Scholar] [CrossRef]
  60. Jaichandran, R.; Leelavathy, S.; Usha, K.S.; Goutham, K.; Mevin, J.M.; Jomon, B. Machine learning technique based parkinson’s disease detection from spiral and voice inputs. EJMCM 2020, 7, 2815–2820. [Google Scholar]
  61. Celik, E.; Omurca, S.I. Improving Parkinson’s Disease Diagnosis with Machine Learning Methods. In Proceedings of the 2019 Scientific Meeting on Electrical-Electronics & Biomedical Engineering and Computer Science (EBBT), Istanbul, Turkey, 24–26 April 2019. [Google Scholar]
  62. Khedr, E.M.; El Fetoh, N.A.; Khalifa, H.E.; Ahmed, M.A.; El Beh, K.M. Prevalence of non motor features in a cohort of Parkinson’s disease patients. Clin. Neurol. Neurosurg. 2013, 115, 673–677. [Google Scholar] [CrossRef]
  63. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef]
  64. Sounderajah, V.; Ashrafian, H.; Rose, S.; Shah, N.H.; Ghassemi, M.; Golub, R.; Kahn, C.E.; Esteva, A.; Karthikesalingam, A.; Mateen, B.; et al. A quality assessment tool for artificial intelligence-centered diagnostic test accuracy studies: QUADAS-AI. Nat. Med. 2021, 27, 1663–1665. [Google Scholar] [CrossRef]
  65. Prashanth, R.; Roy, S.D.; Mandal, P.K.; Ghosh, S. High-Accuracy Detection of Early Parkinson’s Disease through Multimodal Features and Machine Learning. Int. J. Med. Inform. 2016, 90, 13–21. [Google Scholar] [CrossRef]
  66. Su, C.; Tong, J.; Wang, F. Mining genetic and transcriptomic data using machine learning approaches in Parkinson’s disease. NPJ Park. Dis. 2020, 6, 24. [Google Scholar] [CrossRef]
  67. Dias, A.E.; Limongi, J.C.; Barbosa, E.R.; Hsing, W.T. Voice telerehabilitation in Parkinson’s disease. Codas 2016, 28, 176–181. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  68. Hu, M.; Shu, X.; Yu, G.; Wu, X.; Välimäki, M.; Feng, H. A Risk Prediction Model Based on Machine Learning for Cognitive Impairment Among Chinese Community-Dwelling Elderly People With Normal Cognition: Development and Validation Study. J. Med. Internet Res. 2021, 23, e20298. [Google Scholar] [CrossRef] [PubMed]
  69. Mandal, I.; Sairam, N. New machine-learning algorithms for prediction of Parkinson’s disease. Int. J. Syst. Sci. 2012, 45, 647–666. [Google Scholar] [CrossRef]
  70. Battineni, G.; Chintalapudi, N.; Amenta, F.; Traini, E. A Comprehensive Machine-Learning Model Applied to Magnetic Resonance Imaging (MRI) to Predict Alzheimer’s Disease (AD) in Older Subjects. J. Clin. Med. 2020, 9, 2146. [Google Scholar] [CrossRef]
  71. Dashtipour, K.; Taylor, W.; Ansari, S.; Zahid, A.; Gogate, M.; Ahmad, J.; Assaleh, K.; Arshad, K.; Imran, M.A.; Abbai, Q. Detecting Alzheimer’s disease using machine learning methods. In Proceedings of the EAI BODYNETS 2021, Glasgow, UKingdo, 25–26 October 2021; Available online: https://hal.archives-ouvertes.fr/hal-03381752/ (accessed on 2 December 2021).
  72. Bind, S.; Tiwari, A.K.; Sahani, A.K.; Koulibaly, P.; Nobili, F.; Pagani, M.; Sabri, O.; Borght, T.; Laere, K.; Tatsch, K. A survey of machine learning based approaches for Parkinson disease prediction. IJCSIT 2015, 6, 1648–1655. [Google Scholar]
  73. Cao, X.; Song, J.; Zhang, C. Using Principal Component Analysis And Choqet Integral To Establish A Diagnostic Model of Parkinson Disease. Phys. Procedia 2012, 24, 1573–1581. [Google Scholar] [CrossRef] [Green Version]
  74. Naghsh, E.; Sabahi, M.F.; Beheshti, S. Spatial analysis of EEG signals for Parkinson’s disease stage detection. Signal Image Video Process 2019, 14, 397–405. [Google Scholar] [CrossRef]
  75. Parisi, L.; Ravichandran, N.; Manaog, M.L. Feature-driven machine learning to improve early diagnosis of Parkinson’s disease. Expert Syst. Appl. 2018, 110, 182–190. [Google Scholar] [CrossRef]
  76. Billah, M. Symptom Analysis of Parkinson Disease Using SVM-SMO and Ada-Boost Classifiers. Ph.D. Thesis, BRAC University, Dhaka, Bangladesh, 2014. Available online: https://dspace.bracu.ac.bd/bitstream/handle/10361/2938/10101002.pdf?sequence=1 (accessed on 2 December 2021).
  77. Khatamino, P.; Cantürk, İ.; Özyılmaz, L. A deep learning-CNN based system for medical diagnosis: An application on Parkinson’s disease handwriting drawings. In Proceedings of the 2018 6th International Conference on Control Engineering & Information Technology (CEIT), Istanbul, Turkey, 25–27 October 2018. [Google Scholar]
  78. Nagasubramanian, G.; Sankayya, M. Multi-Variate vocal data analysis for Detection of Parkinson disease using Deep Learning. Neural Comput. Appl. 2021, 33, 4849–4864. [Google Scholar] [CrossRef]
  79. Anila, M.; Pradeepini, G.D. A Review on Parkinson’s Disease Diagnosis using Machine Learning Techniques. IJERT 2020, 9, 330–334. [Google Scholar] [CrossRef]
  80. Mathur, R.; Pathak, V.; Bandil, D. Parkinson Disease Prediction Using Machine Learning Algorithm. In Advances in Intelligent Systems and Computing; Springer: Singapore, 2018; pp. 357–363. [Google Scholar]
  81. Fang, E.; Ann, C.N.; Maréchal, B.; Lim, J.X.; Tan, S.Y.Z.; Li, H.; Gan, J.; Tan, E.K.; Chan, L.L. Differentiating Parkinson’s disease motor subtypes using automated volume-based morphometry incorporating white matter and deep gray nuclear lesion load. J. Magn. Reson. Imaging 2020, 51, 748–756. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  82. Hardy, J. Genetic Analysis of Pathways to Parkinson Disease. Neuron 2010, 68, 201–206. [Google Scholar] [CrossRef] [Green Version]
  83. Antonini, A.; Leta, V.; Teo, J.; Chaudhuri, K.R. Outcome of Parkinson’s disease patients affected by COVID-19. Mov. Disord. 2020, 35, 905–908. [Google Scholar] [CrossRef] [PubMed]
  84. Salari, M.; Zali, A.; Ashrafi, F.; Etemadifar, M.; Sharma, S.; Hajizadeh, N.; Ashourizadeh, H. Incidence of Anxiety in Parkinson’s Disease During the Coronavirus Disease (COVID-19) Pandemic. Mov. Disord. 2020, 35, 1095–1096. [Google Scholar] [CrossRef]
  85. Suri, J.S.; Puvvula, A.; Biswas, M.; Majhail, M.; Saba, L.; Faa, G.; Singh, I.M.; Oberleitner, R.; Turk, M.; Chadha, P.S.; et al. COVID-19 pathways for brain and heart injury in comorbidity patients: A role of medical imaging and artificial intelligence-based COVID severity classification: A review. Comput. Biol. Med. 2020, 124, 103960. [Google Scholar] [CrossRef]
  86. Tipton, P.W.; Wszolek, Z.K. What can Parkinson’s disease teach us about COVID-19? Neurol. Neurochir. Polska 2020, 54, 204–206. [Google Scholar] [CrossRef] [Green Version]
  87. Saba, L.; Dey, N.; Ashour, A.; Samanta, S.; Nath, S.S.; Chakraborty, S.; Sanches, J.; Kumar, D.; Marinho, R.; Suri, J.S. Automated stratification of liver disease in ultrasound: An online accurate feature classification paradigm. Comput. Methods Programs Biomed. 2016, 130, 118–134. [Google Scholar] [CrossRef] [PubMed]
  88. Acharya, U.R.; Sree, S.V.; Ribeiro, R.; Krishnamurthi, G.; Marinho, R.; Sanches, J.; Suri, J.S. Data mining framework for fatty liver disease classification in ultrasound: A hybrid feature extraction paradigm. Med. Phys. 2012, 39, 4255–4264. [Google Scholar] [CrossRef] [Green Version]
  89. Acharya, U.R.; Sree, S.V.; Krishnan, M.M.R.; Molinari, F.; Garberoglio, R.; Suri, J.S. Non-invasive automated 3D thyroid lesion classification in ultrasound: A class of ThyroScan™ systems. Ultrasonics 2012, 52, 508–520. [Google Scholar] [CrossRef]
  90. Jamthikar, A.D.; Puvvula, A.; Gupta, D.; Johri, A.M.; Nambi, V.; Khanna, N.N.; Saba, L.; Mavrogeni, S.; Laird, J.R.; Pareek, G.; et al. Cardiovascular disease and stroke risk assessment in patients with chronic kidney disease using integration of estimated glomerular filtration rate, ultrasonic image phenotypes, and artificial intelligence: A narrative review. Int. Angiol. 2020, 40, 150–164. [Google Scholar] [CrossRef] [PubMed]
  91. Murgia, A.; Balestrieri, A.; Crivelli, P.; Suri, J.S.; Conti, M.; Cademartiri, F.; Saba, L. Cardiac computed tomography radiomics: An emerging tool for the non-invasive assessment of coronary atherosclerosis. Cardiovasc. Diagn. Ther. 2020, 10, 2005–2017. [Google Scholar] [CrossRef]
  92. Acharya, U.R.; Sree, S.V.; Krishnan, M.M.R.; Krishnananda, N.; Ranjan, S.; Umesh, P.; Suri, J.S. Automated classification of patients with coronary artery disease using grayscale features from left ventricle echocardiographic images. Comput. Methods Programs Biomed. 2013, 112, 624–632. [Google Scholar] [CrossRef] [PubMed]
  93. Acharya, U.R.; Sree, S.V.; Kulshreshtha, S.; Molinari, F.; Koh, J.E.W.; Saba, L.; Suri, J.S. GyneScan: An Improved Online Paradigm for Screening of Ovarian Cancer via Tissue Characterization. Technol. Cancer Res. Treat. 2014, 13, 529–539. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  94. Shrivastava, V.K.; Londhe, N.D.; Sonawane, R.S.; Suri, J.S. Exploring the color feature power for psoriasis risk stratification and classification: A data mining paradigm. Comput. Biol. Med. 2015, 65, 54–68. [Google Scholar] [CrossRef]
  95. Shrivastava, V.; Londhe, N.D.; Sonawane, R.S.; Suri, J.S. Computer-aided diagnosis of psoriasis skin images with HOS, texture and color features: A first comparative study of its kind. Comput. Methods Programs Biomed. 2016, 126, 98–109. [Google Scholar] [CrossRef]
  96. Chen, R.; Kumar, S.; Garg, R.R.; Lang, A. Impairment of motor cortex activation and deactivation in Parkinson’s disease. Clin. Neurophysiol. 2001, 112, 600–607. [Google Scholar] [CrossRef]
  97. Bodis-Wollner, I.; Yahr, M.D. Measurements of Visual Evoked Potentials in Parkinson’s Disease. Brain 1978, 101, 661–671. [Google Scholar] [CrossRef]
  98. Kaur, S.; Aggarwal, H.; Rani, R. Diagnosis of Parkinson’s Disease Using Principle Component Analysis and Deep Learning. J. Med. Imaging Health Inform. 2019, 9, 602–609. [Google Scholar] [CrossRef]
  99. Zappia, M.; Annesi, G.; Nicoletti, G.; Arabia, G.; Annesi, F.; Messina, D.; Pugliese, P.; Spadafora, P.; Tarantino, P.; Carrideo, S. Sex differences in clinical and genetic determinants of levodopa peak-dose dyskinesias in Parkinson disease: An exploratory study. Arch. Neurol. 2005, 62, 601–605. [Google Scholar] [CrossRef] [PubMed]
  100. Lang, E.A.; Lozano, M.A. Parkinson’s disease. N. Engl. J. Med. 1998, 339, 1130–1143. [Google Scholar] [CrossRef] [PubMed]
  101. Martinez-Martin, P.; Rodriguez-Blazquez, C.; Abe, K.; Bhattacharyya, K.B.; Bloem, B.R.; Carod-Artal, F.J.; Prakash, R.; Esselink, R.; Falup-Pecurariu, C.; Gallardo, M.; et al. International study on the psychometric attributes of the Non-Motor Symptoms Scale in Parkinson disease. Neurology 2009, 73, 1584–1591. [Google Scholar] [CrossRef] [PubMed]
  102. Weernink, M.G.; Groothuis-Oudshoorn, C.G.; Ijzerman, M.J.; van Til, J.A. Valuing Treatments for Parkinson Disease Incorporating Process Utility: Performance of Best-Worst Scaling, Time Trade-Off, and Visual Analogue Scales. Value Health 2016, 19, 226–232. [Google Scholar] [CrossRef] [Green Version]
  103. Senturk, Z.K. Early diagnosis of Parkinson’s disease using machine learning algorithms. Med. Hypotheses 2020, 138, 109603. [Google Scholar] [CrossRef]
  104. Tsoukra, P.; Velakoulis, D.; Wibawa, P.; Malpas, C.B.; Walterfang, M.; Evans, A.; Farrand, S.; Kelso, W.; Eratne, D.; Loi, S.M. The Diagnostic Challenge of Young-Onset Dementia Syndromes and Primary Psychiatric Diseases: Results From a Retrospective 20-Year Cross-Sectional Study. J. Neuropsychiatry Clin. Neurosci. 2021. [Google Scholar] [CrossRef]
  105. Pohar, S.L.; Jones, C.A. The burden of Parkinson disease (PD) and concomitant comorbidities. Arch. Gerontol. Geriatr. 2009, 49, 317–321. [Google Scholar] [CrossRef]
  106. Janghel, R.R.; Shukla, A.; Rathore, C.P.; Verma, K.; Rathore, S. A comparison of soft computing models for Parkinson’s disease diagnosis using voice and gait features. Netw. Model. Anal. Health Inform. Bioinform. 2017, 6, 14. [Google Scholar] [CrossRef]
  107. Agarwal, M.; Saba, L.; Gupta, S.K.; Carriero, A.; Falaschi, Z.; Paschè, A.; Danna, P.; El-Baz, A.; Naidu, S.; Suri, J.S. A Novel Block Imaging Technique Using Nine Artificial Intelligence Models for COVID-19 Disease Classification, Characterization and Severity Measurement in Lung Computed Tomography Scans on an Italian Cohort. J. Med Syst. 2021, 45, 1–30. [Google Scholar] [CrossRef] [PubMed]
  108. Cau, R.; Pacielli, A.; Fatemeh, H.; Vaudano, P.; Arru, C.; Crivelli, P.; Stranieri, G.; Suri, J.S.; Mannelli, L.; Conti, M.; et al. Complications in COVID-19 patients: Characteristics of pulmonary embolism. Clin. Imaging 2021, 77, 244–249. [Google Scholar] [CrossRef]
  109. Saba, L.; Tiwari, A.; Biswas, M.; Gupta, S.K.; Godia-Cuadrado, E.; Chaturvedi, A.; Turk, M.; Suri, H.S.; Orru, S.; Sanches, J.M. Wilson’s disease: A new perspective review on its genetics, diagnosis and treatment. Front. Biosci. 2019, 11, 166–185. [Google Scholar]
  110. Porcu, M.; Cocco, L.; Puig, J.; Mannelli, L.; Yang, Q.; Suri, J.S.; Defazio, G.; Saba, L. Global Fractional Anisotropy: Effect on Resting-state Neural Activity and Brain Networking in Healthy Participants. Neuroscience 2021, 472, 103–115. [Google Scholar] [CrossRef]
  111. Alqahtani, E.J.; Alshamrani, F.H.; Syed, H.F.; Olatunji, S.O. Classification of Parkinson’s Disease Using NNge Classification Algorithm. In Proceedings of the 2018 21st Saudi Computer Society National Computer Conference (NCC), Riyadh, Saudi Arabia, 25–26 April 2018; pp. 1–7. [Google Scholar]
  112. Porcu, M.; Cocco, L.; Cocozza, S.; Pontillo, G.; Operamolla, A.; Defazio, G.; Suri, J.S.; Brunetti, A.; Saba, L.J. The association between white matter hyperintensities, cognition and regional neural activity in healthy subjects. Eur. J. Neurosci. 2021, 54, 5427–5443. [Google Scholar] [CrossRef]
  113. Saba, L.; Gerosa, C.; Fanni, D.; Marongiu, F.; La Nasa, G.; Caocci, G.; Barcellona, D.; Balestrieri, A.; Coghe, F.; Orru, G.; et al. Molecular pathways triggered by COVID-19 in different organs: ACE2 receptor-expressing cells under attack? A review. Eur. Rev. Med. Pharmacol. Sci. 2020, 24, 12609–12622. [Google Scholar]
  114. Cau, R.; Bassareo, P.P.; Mannelli, L.; Suri, J.S.; Saba, L. Imaging in COVID-19-related myocardial injury. Int. J. Cardiovasc. Imaging 2021, 37, 1349–1360. [Google Scholar] [CrossRef]
  115. Suri, J.S.; Puvvula, A.; Majhail, M.; Biswas, M.; Jamthikar, A.D.; Saba, L.; Faa, G.; Singh, I.M.; Oberleitner, R.; Turk, M.; et al. Integration of cardiovascular risk assessment with COVID-19 using artificial intelligence. Rev. Cardiovasc. Med. 2020, 21, 541–560. [Google Scholar] [CrossRef]
  116. Saba, L.; Agarwal, M.; Patrick, A.; Puvvula, A.; Gupta, S.K.; Carriero, A.; Laird, J.R.; Kitas, G.D.; Johri, A.M.; Balestrieri, A.; et al. Six artificial intelligence paradigms for tissue characterization and classification of non-COVID-19 pneumonia against COVID-19 pneumonia in computed tomography lungs. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 423–434. [Google Scholar] [CrossRef]
  117. Souza, C.D.O.; Voos, M.C.; Barbosa, A.F.; Chen, J.; Francato, D.C.V.; Milosevic, M.; Popovic, M.; Fonoff, E.T.; Chien, H.F.; Barbosa, E.R. Relationship Between Posturography, Clinical Balance and Executive Function in Parkinson´s Disease. J. Mot. Behav. 2019, 51, 212–221. [Google Scholar] [CrossRef] [PubMed]
  118. Mc Kinlay, A.; Grace, R.C. Characteristic of Cognitive Decline in Parkinson’s Disease: A 1-Year Follow-Up. Appl. Neuropsychol. 2011, 18, 269–277. [Google Scholar] [CrossRef] [PubMed]
Figure 2. PRISMA model for early detection of the PD by using AI.
Figure 2. PRISMA model for early detection of the PD by using AI.
Diagnostics 12 00166 g002
Figure 3. Year of article publication vs. impact factor.
Figure 3. Year of article publication vs. impact factor.
Diagnostics 12 00166 g003
Figure 4. (a) Average accuracy of the various architectures for PD. (b) Minimum (red) and maximum (green) accuracy of the different PD architecture (AI: Artificial Intelligence, SDL: Solo deep learning, ML: Machine learning, HDL: Hybrid deep learning).
Figure 4. (a) Average accuracy of the various architectures for PD. (b) Minimum (red) and maximum (green) accuracy of the different PD architecture (AI: Artificial Intelligence, SDL: Solo deep learning, ML: Machine learning, HDL: Hybrid deep learning).
Diagnostics 12 00166 g004
Figure 5. The performance metrics vs. the number of studies (ACC: Accuracy, SEN: Sensitivity, SPE: Specificity, MCC: Matthew’s correlation coefficient, NPV: Net present value, F1: Dice similarity coefficient).
Figure 5. The performance metrics vs. the number of studies (ACC: Accuracy, SEN: Sensitivity, SPE: Specificity, MCC: Matthew’s correlation coefficient, NPV: Net present value, F1: Dice similarity coefficient).
Diagnostics 12 00166 g005
Figure 6. (a) Various kinds of input for the detection of the PD, (b) types of AI architectures, and (c) demographics of PD patients in four continents (SDL: Solo deep learning; ML: Machine learning; HDL: Hybrid deep learning).
Figure 6. (a) Various kinds of input for the detection of the PD, (b) types of AI architectures, and (c) demographics of PD patients in four continents (SDL: Solo deep learning; ML: Machine learning; HDL: Hybrid deep learning).
Diagnostics 12 00166 g006
Figure 7. Symptomatic biology of PD.
Figure 7. Symptomatic biology of PD.
Diagnostics 12 00166 g007
Figure 8. Proposed methodology for early detection of PD by Anitha et al. [37].
Figure 8. Proposed methodology for early detection of PD by Anitha et al. [37].
Diagnostics 12 00166 g008
Figure 9. Proposed architecture for early detection of PD [6].
Figure 9. Proposed architecture for early detection of PD [6].
Diagnostics 12 00166 g009
Figure 10. Proposed Architecture for early detection of PD by Cleick et al. [61].
Figure 10. Proposed Architecture for early detection of PD by Cleick et al. [61].
Diagnostics 12 00166 g010
Figure 11. The cumulative cutoff score for evaluation for the selected studies. (LM: Low-Moderate, MH: High-Moderate).
Figure 11. The cumulative cutoff score for evaluation for the selected studies. (LM: Low-Moderate, MH: High-Moderate).
Diagnostics 12 00166 g011
Figure 12. (a) Various types of architectures were mentioned in the studies; (b) classifier used; (c) optimization studies.
Figure 12. (a) Various types of architectures were mentioned in the studies; (b) classifier used; (c) optimization studies.
Diagnostics 12 00166 g012
Figure 13. Low vs. Moderate vs. High-Bias of AI Attributes (C1: Citation, C2: Objective and Design Methodology, C3: AI Architecture, C4: Optimization of AI Model, C5: Performance Evaluation of AI Models, C6: Clinical Evaluation and Benchmarking).
Figure 13. Low vs. Moderate vs. High-Bias of AI Attributes (C1: Citation, C2: Objective and Design Methodology, C3: AI Architecture, C4: Optimization of AI Model, C5: Performance Evaluation of AI Models, C6: Clinical Evaluation and Benchmarking).
Diagnostics 12 00166 g013
Table 1. Performance metrics of selected studies.
Table 1. Performance metrics of selected studies.
Performance MetricsACCSENSPEAUCMCCNPVF1
Number of Studies22884321
ACC: Accuracy, SEN: Sensitivity, SPE: Specificity, MCC: Matthew’s correlation coefficient, NPV: Negative predictor value, F1: Dice similarity coefficient.
Table 2. Ranking of the selected studies.
Table 2. Ranking of the selected studies.
Low-BiasModerate-BiasHigh-Bias
SNAuthorC1C2C3C4C5C6MeanAbsolute ScoreCDFRank
1Aseer et al. [1] (2019)3454544.17250.941
2Adams et al. [2] (2017)4444433.83230.882
3Prashantha et al. [3] (2018)3444433.67220.843
4Alzubaidi et al. [4] (2021)3443343.50210.784
5Ahmed et al. [5] (2021)2344443.50210.785
6Wang et al. [6] (2020)3344433.50210.786
7Wang et al. [7] (2017)3344433.50210.787
8Naghsh et al. [8] (2020)3344333.33200.718
9Prashanth et al. [9] (2018)3343433.33200.719
10Moore et al. [10] (2018)3343433.33200.7110
11Fang et al. [11] (2020)2244433.17190.6411
12Celik et al. [12] (2019)4443223.17190.6412
13Poorjam et al. [13] (2019)4342333.17190.6413
14Anitha et al. [14] (2020)3342333.00180.5614
15Maitín et al. [15] (2019)3443223.00180.5615
16Gallego et al. [16] (2017)2344233.00180.5616
17Mei et al. [17] (2021)2333332.83170.4817
18Wroge et al. [18] (2010)3342232.83170.4818
19White et al. [19] (2018)4332132.67160.4019
20Jaichandran et al. [20] (2020)4340122.33140.2520
21Lee et al. [21] (2021)4400042.00120.1421
22Singamaneni et al. [22] (2021)1332122.00120.1422
23Hu et al. [23] (2019)1221231.83110.1023
24Bhat et al. [22] (2019)3330021.83110.1024
25Bala et al. [24] (2020)1221121.5090.0525
26Dias et al. [10] (2016)1100020.6740.0026
C1: Citation, C2: Objective and Design Methodology, C3: AI Architecture, C4: Optimization of AI Model, C5: Performance Evaluation of AI Models, C6: Clinical Evaluation and Benchmarking, CDF: Cumulative score.
Table 3. Benchmarking scheme for selected and proposed studies.
Table 3. Benchmarking scheme for selected and proposed studies.
B0B1B2B3B4B5B6B7B8B9CB0B11B12B13B14
SNCitation (Year)OBDDIECDEMEPMAAIPABACVBSSVRS
1Ahlrichs et al. [25] (2013)Ns vs. PD×××××××××72
2Bind et al. [26] (2015)Ns vs. PD××××××××××52
3Maitín et al. [15] (2020)Ns vs. PD×××××37
4Anila et al. [27] (2020)Ns vs. PD××××××××37
5Watts et al. [28] (2020)Ns vs. PD××××××××××109
6Garg et al. [29] (2021)Ns vs. PD×××××××××××15
7Mei et al. [17] (2021)Ns vs. PD××××78
8Alzubaidi et al. [4] (2021)Ns vs. PD×××××108
9ProposedNs vs. PD105
B1: Citation, B2: Objective, B3: Demographic discussion, B4: Inclusive and Exclusive criteria, B5: Data Extraction, B6: Model Evaluation, B7: PRISMA Model, B8: Attribute Analysis, B9: Input parameter analysis, B10: Benchmarking analysis, B11: Cross-validation, B12: Bias studies, B13: Scientific validation, B14: Reference studies, “√” article includes particular benchmark, “×” article does not includes particular benchmark.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Paul, S.; Maindarkar, M.; Saxena, S.; Saba, L.; Turk, M.; Kalra, M.; Krishnan, P.R.; Suri, J.S. Bias Investigation in Artificial Intelligence Systems for Early Detection of Parkinson’s Disease: A Narrative Review. Diagnostics 2022, 12, 166. https://doi.org/10.3390/diagnostics12010166

AMA Style

Paul S, Maindarkar M, Saxena S, Saba L, Turk M, Kalra M, Krishnan PR, Suri JS. Bias Investigation in Artificial Intelligence Systems for Early Detection of Parkinson’s Disease: A Narrative Review. Diagnostics. 2022; 12(1):166. https://doi.org/10.3390/diagnostics12010166

Chicago/Turabian Style

Paul, Sudip, Maheshrao Maindarkar, Sanjay Saxena, Luca Saba, Monika Turk, Manudeep Kalra, Padukode R. Krishnan, and Jasjit S. Suri. 2022. "Bias Investigation in Artificial Intelligence Systems for Early Detection of Parkinson’s Disease: A Narrative Review" Diagnostics 12, no. 1: 166. https://doi.org/10.3390/diagnostics12010166

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop