Next Article in Journal
Cloud Service Broker Using Ontology-Based System
Previous Article in Journal
Drought Risks Assessment Using Standardized Precipitation Index
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Early Detection of Alzheimer’s Disease: An Extensive Review of Advancements in Machine Learning Mechanisms Using an Ensemble and Deep Learning Technique †

by
Renjith Prabhavathi Neelakandan
1,*,
Ramesh Kandasamy
2,
Balasubramani Subbiyan
3 and
Mariya Anto Bennet
4
1
School of Computer Science and Engineering, Vellore Institute of Technology, Chennai Campus, Chennai 603103, Tamilnadu, India
2
Department of Computer Science and Engineering, Sri Krishna College of Engineering and Technology, Coimbatore 641008, Tamilnadu, India
3
Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation, Vaddeswaram 522502, Andrapradesh, India
4
Department of Electronics and Communication Engineering, Vel Tech Rangaranjan Dr. Sagunthala R & D Institute of Science and Technology, Chennai 600062, Tamilnadu, India
*
Author to whom correspondence should be addressed.
Presented at the International Conference on Recent Advances on Science and Engineering, Dubai, United Arab Emirates, 4–5 October 2023.
Eng. Proc. 2023, 59(1), 10; https://doi.org/10.3390/engproc2023059010
Published: 11 December 2023
(This article belongs to the Proceedings of Eng. Proc., 2023, RAiSE-2023)

Abstract

:
Alzheimer’s disease (AD) is the most common form of dementia in senior individuals. It is a progressive neurological ailment that predominantly affects memory, cognition, and behavior. An early AD diagnosis is essential for effective disease management and timely intervention. Due to its complexity and heterogeneity, AD is, however, difficult to diagnose precisely. This paper investigates the integration of disparate machine learning algorithms to improve AD diagnostic accuracy. The used dataset includes instances with missing values, which are effectively managed by employing appropriate imputation techniques. Several feature selection algorithms are applied to the dataset to determine the most relevant characteristics. Moreover, the Synthetic Minority Oversampling Technique (SMOTE) is employed to address class imbalance issues. The proposed system employs an Ensemble Classification algorithm, which integrates the outcomes of multiple predictive models to enhance diagnostic accuracy. The proposed method has superior disease prediction capabilities in comparison to existing methods. The experiment employs a robust AD dataset from the UCI machine learning repository. The findings of this study contribute significantly to the field of AD diagnoses and pave the way for more precise and efficient early detection strategies.

1. Introduction

Alzheimer’s disease is a progressive neurological illness that affects memory, cognition, and behavior in senior individuals and is the most common cause of dementia [1,2,3]. An AD diagnosis should be accurate and timely to ensure effective disease management and timely intervention, resulting in improved patient care and potential therapeutic interventions. For clinicians, the precise diagnosis of AD is a challenging task due to its complexity and heterogeneity. In recent years, machine learning has emerged as a powerful tool in medical diagnoses, offering the potential to augment traditional diagnostic approaches and improve accuracy [4,5]. Figure 1 represents the IoT-based patient monitoring system. Through the integration of disparate machine learning algorithms, this study aims to improve AD diagnosis accuracy by leveraging the capabilities of machine learning. Multi-algorithm approaches strive to overcome the limitations of individual models by harnessing the predictive capabilities of multiple algorithms. This study aims to enhance Alzheimer’s disease (AD) diagnoses using machine learning techniques.
The method involves working with a dataset containing missing values. To address this, imputation techniques are employed, ensuring dataset integrity and analysis quality [6,7,8]. Additionally, a feature selection algorithm identifies key dataset characteristics crucial for accurate AD prediction. Class imbalance, common in medical datasets, is tackled using the Synthetic Minority Oversampling Technique (SMOTE) [9,10]. This ensures unbiased predictive models and more reliable disease prediction. To further improve AD prediction accuracy, an automated method using machine learning is proposed. An Ensemble Classification algorithm is applied, combining multiple predictive models [11]. This approach enhances AD detection reliability by amalgamating results from various algorithms. Feature extraction extracts fresh data features, while feature selection picks significant characteristics. Appropriate feature selection algorithms are vital for accurate prediction [12]. Algorithms like Mutual Information Scores, Relief, and Recursive Feature Elimination efficiently extract essential features. A Univariate Analysis assigns significance scores to each feature. Class distribution is balanced with SMOTE [13], addressing the unbalanced dataset issue by increasing minority class rows, thereby boosting minority class classifier accuracy [14]. Other studies have used advanced machine learning methods such as deep learning and convolutional neural networks for AD diagnoses [15,16]. Novel biomarkers and multimodal data integration [17,18] have shown promise in predicting AD. This study utilizes a robust AD dataset from UCI’s machine learning repository, ensuring findings’ reliability and generalizability. The proposed approach outperforms existing methods in disease prediction after rigorous evaluation. Beyond an AD diagnosis, this study holds broader implications. It offers insights into machine learning applications in medical diagnoses, facilitating precise early detection across medical domains. As AD and neurodegenerative disorders rise, accurate diagnostic methods are vital for patient outcomes and healthcare management. Moreover, this research advances machine learning’s role in medical research, beyond an AD diagnosis. The remainder of the paper is arranged as follows: Section 2 explains the materials and methodology, Section 3 presents the experimentation and performance assessment, Section 4 demonstrates Experimental Configuration, Results, and Discussion, and Section 5 concludes with a discussion of future work.

2. Materials and Methods/Methodology

2.1. Literature Review

Ensemble learning enhances an AD diagnosis through 3D convolutional neural networks and MRI. Such networks can distinguish between healthy individuals, those with mild cognitive impairment, and AD patients [1,2]. Transfer learning aids in detecting AD with these networks [3]. Techniques like deep aggregation learning and stacking-based ensemble learning with genetic hyperparameter tweaking improve diagnostic accuracy [4,5]. Using the ADNI dataset, MRI-based ensemble learning achieves 95.2% accuracy distinguishing AD from NC and 77.8% accuracy distinguishing sMCI from pMCI [11]. However, this dataset’s limitations include sample size and lack of method comparison with other state-of-the-art techniques. Additionally, the research does not assess the method’s interpretability, potentially limiting its application [11]. Another study introduces an ensemble learning architecture using 2D CNNs for an AD diagnosis [12]. This method trains on grey matter density maps and uses ensemble models to improve prediction accuracy. However, its limitations include reliance on 2D MRI images and the need for testing on larger datasets [12]. Research on machine learning for AD diagnoses using neuroimaging data explored techniques like Support Vector Machines and CNNs [13,14]. While some methods achieve significant accuracy, they often face challenges with real-world healthcare data or require further testing on more extensive datasets [15].
A study using a stacking-genetic algorithm ensemble learning model reached a high accuracy, precision, recall, and F1-score in early AD diagnoses [15]. Nevertheless, issues like variable dataset validation and clinical interpretability remain. Combining MRI classifiers offers reliable AD detection, but its applicability requires further exploration [16]. On the other hand, Random Forest achieves high accuracy predicting AD using limited features from MRI scans [17]. Deep learning has shown potential in AD diagnoses, especially when studying complex disease pathways [18]. Still, its reliability in predicting AD progression needs rigorous testing across various imaging modalities and larger datasets. The use of a deep CNN for a stage-based AD diagnosis shows promise, but a comprehensive methodology comparison and general applicability assessment are essential [19]. Other methods, such as high-pressure liquid chromatography with AI algorithms, offer insights into predicting Alzheimer’s medication properties [20]. Deep learning techniques integrating expert knowledge and multi-source data have outperformed many ensemble methods [21]. However, the system might need substantial computational resources and could vary across datasets. Research on ensemble learning with Conformal Predictors indicates improved categorization, but a broader dataset is essential for validation [22]. Hierarchical ensemble learning addresses some deep learning challenges, providing enhanced classification accuracy with pre-trained neural networks [23]. However, this may require substantial training datasets and high-quality MRI scans. Lastly, ensemble learning for regression problems shows potential in predicting medication effects, but needs expansion for broader applications [14,24]. Ensemble learning and advanced algorithms demonstrate significant promise in AD diagnoses [25,26]. However, broader dataset validations, methodology comparisons, and evaluations of real-world applicability are crucial.

2.2. Proposed Work

In the initial phase of the system, categorical attributes are converted into numeric attributes (0 s and 1 s). The absent values in the dataset are then handled using the median value. Feature extraction creates new data features. Next, feature selection is used to find disease–diagnosis-relevant traits. Accurate prediction requires this step. Several feature selection techniques are researched to choose the most useful characteristics for an AD diagnosis. After declaring a set number of features, Recursive Feature Elimination (RFE) removes them. A Univariate Analysis evaluates each attribute numerically. PCA reduces dimensionality while maintaining useful data. Mutual Information Scores and Relief automatically choose relevant features to accelerate a diagnosis. SMOTE is used to oversample the minority class in AD datasets to address class imbalance. Fair categorization datasets result. Ensemble classification mixes model predictions to efficiently handle textual characteristics. Aggregating label forecasts and forecasting the majority vote improve classification accuracy. A comprehensive and sophisticated ensemble-based model aims to improve AD diagnoses. The model extracts critical characteristics and uses a balanced dataset by merging several feature selection techniques and SMOTE for class imbalance, boosting illness prediction accuracy. The ensemble classification strengthens the model’s textual feature management. The planned study will enhance AD detection and diagnoses, improving patient outcomes and healthcare management.
A. 
Pre-processing
The pre-processing phase prepares the raw AD dataset for an analysis. Categorical attributes are transformed to numeric for compatibility with machine learning. Median imputation addresses missing values, ensuring data completeness. Feature extraction enriches the dataset, while feature selection pinpoints the most informative attributes. Recursive Feature Elimination (RFE) removes less vital features iteratively. A Univariate Analysis ranks features based on their importance. A Principal Component Analysis (PCA) compresses data without losing critical information. This rigorous preparation creates a solid foundation for the ensemble-based AD diagnosis model.
B. 
Extraction and Selection of Features
In the ensemble-based model for an AD diagnosis, feature extraction transforms the raw AD dataset to capture essential patterns, enhancing its richness for better prediction. Feature selection then identifies the most critical characteristics within this dataset. Several algorithms assess which features most influence diagnostic accuracy. Recursive Feature Elimination (RFE) methodically removes less important features to streamline the model, while a Univariate Analysis ranks each feature’s significance in classification. A Principal Component Analysis (PCA) compresses data, retaining essential variance for a concise representation. By using these feature extraction and selection methods, the model highlights the AD dataset’s key aspects, improving prediction accuracy and supporting early disease detection for improved patient results.
Given a dataset X of size n × m (n samples, m features),
X   R ( n × m )
Compute the mean of each feature and subtract it from the corresponding feature in X, resulting in a zero-mean dataset X c e n t e r e d .
X c e n t e r e d = s = X 1 n × ( i = 1 n × n · X i )
Calculate the covariance matrix.
C ( X c e n t e r e d ) = X c e n t e r e d T × X c e n t e r e d ( n 1 )
Compute the eigen values (λ) and eigenvectors (v) of the covariance matrix C.
C = X c e n t e r e d T × X c e n t e r e d ( n 1 )
C × v i = 1 m λ i × v i
V   of   size   m   ×   k :   V = [ v 1 ,   v 2 ,   v k ]
X n e w   o f   size   n   ×   k   : X new   =   X centered   ×   V
C. 
Synthetic Minority Oversampling Technique (SMOTE)
The proposed ensemble-based model for AD diagnoses uses the Synthetic Minority Oversampling Technique (SMOTE) to tackle class imbalance often found in medical datasets, including AD. Class imbalance can lead to biased learning, favoring the larger class and reducing accuracy. SMOTE addresses this by creating synthetic samples for the underrepresented class, enhancing its presence in the dataset. By adding these samples, the model better understands minority class patterns, leading to better AD diagnosis accuracy. Using SMOTE ensures a balanced dataset, enhancing the model’s prediction accuracy for both classes.
Algorithm: synthetic samples depending on the minority-majority class ratio.
Input
  • Minority class samples: M
  • k (number of nearest neighbors to consider)
Output
Synthetic samples: S
1.
Create an empty synthetic sample list: S = []
2.
Calculate the number of synthetic samples (n_synthetic) depending on the minority-majority class ratio.
3.
Each minority class sample m in M:
  • Find k closest neighbors of m from minority class samples, omitting m.
  • Randomly choose one of the k neighbours (nn).
  • Difference vector diff = nn − m.
  • Add a random proportion of diff to ‘m’ to create n_synthetic samples.
4.
Add all newly synthesized samples to S.
5.
Return synthetic sample list S.
6.
The method identifies AD efficiently and correctly. Healthy and AD patients are first separated. SMOTE fakes minority class samples for dataset balance. Representing both groups promotes learning. Splitting the balanced dataset into training and testing sets preserves class distribution.
7.
Logistic Regression, Random Forest, or SVM predict AD. The chosen model learns from training set characteristics and labels. The testing set evaluates the model’s accuracy, precision, recall, and F1-score.
8.
Successful classification models can discover AD in new data. Fresh instance features suggest AD.
SMOTE builds synthetic samples along line segments linking a minority class sample and its k nearest neighbors, extending the minority class in feature space. Logistic Regression, Random Forest, or a Support Vector Machine are used to predict AD. The selected model learns features and annotations from the training set. To measure the model’s efficacy, the accuracy, precision, recall, and F1-score are used on the assessment set. The trained classification model is able to detect AD in new, unlabeled data if its performance is adequate. By feeding the model the characteristics of new instances, it can precisely predict the presence of AD. With careful consideration of dataset quality, feature selection, and model selection, this algorithm provides a promising strategy for early and accurate AD detection. Utilizing SMOTE to resolve class imbalance and advanced classification techniques, the algorithm improves patient outcomes by facilitating timely diagnoses and intervention.
D. 
AD Prediction Using SMOTE
The proposed method efficiently classifies AD. The dataset, initially divided into healthy and AD patients, is balanced using SMOTE. This enhances learning by representing both classes equally. The data are then split for training and testing with equal class distribution. The model, using Logistic Regression, Random Forest, or a Support Vector Machine, learns from the training set and is evaluated based on the accuracy, precision, recall, and F1-score. Once trained, the model can predict AD in new data. By addressing class imbalances with SMOTE and using advanced techniques, this approach promises early and accurate AD detection, improving patient outcomes.
E. 
Classification Procedure
A Support Vector Machine (SVM) is a key classification tool with significant potential for an AD diagnosis. It is a versatile supervised learning algorithm suited for both linear and nonlinear tasks. Especially useful for complex medical datasets like AD, SVM identifies the best hyperplane to separate classes. After refining features, SVM can discern complex patterns and relationships in the dataset. Its ability to handle nonlinear relationships through various kernel functions and resist outliers ensures reliable predictions. When trained on a balanced dataset from SMOTE, SVM offers high sensitivity and specificity, vital for early AD detection.

3. Experimentation and Performance Assessment

This study evaluates the ensemble-based model’s efficacy in an AD diagnosis. Using advanced feature extraction and selection, it trains on a dataset balanced via SMOTE, ensuring equal representation of healthy and AD subjects. Through cross-validation, the model’s predictive accuracy is assessed against metrics like the precision, recall, and F1-score. This ensemble-based model excels in comparison to SVM and Logistic Regression. This study further examines the impact of SMOTE samples and optimal hyperparameters on performance. The results indicate superior diagnostic precision, highlighting the model’s potential for early AD detection and improved healthcare management. The experimental design begins with collecting comprehensive AD data, which undergoes pre-processing and feature augmentation. SMOTE ensures dataset balance, which is then divided for training and testing. Model efficacy is evaluated using metrics like the accuracy, precision, recall, F1-score, and AUC-ROC, alongside a confusion matrix.
Accuracy = (True Positives + True Negatives)/Total Instances
Precision = (True Positives)/((True Positives + False Positives))
Recall = (True Positives)/((True Positives + False Negatives))
F1 Score = (2× (Precision×Recall))/((Precision + Recall))
TP (True Positive) refers to instances correctly predicted as positive (having the disease), whereas FP (False Positive) refers to instances incorrectly predicted as positive, despite not having the disease. TN (True Negative) indicates instances that were accurately predicted as negative (not having the disease), whereas FN (False Negative) indicates instances that were incorrectly predicted as negative but actually included the disease.

4. Results and Discussion

Feature selection and data resampling are critical for enhancing machine learning model performance, especially with imbalanced datasets. Feature selection chooses relevant features from the initial set, eliminating unimportant or redundant ones. This enhances model efficiency and interpretability, and reduces overfitting. Methods like Recursive Feature Elimination (RFE), Univariate Feature Selection (UFS), and a Principal Component Analysis (PCA) help identify key features for accurate predictions. Data resampling adjusts the dataset’s distribution, particularly when class imbalances exist. Oversampling, like the SMOTE method, creates synthetic samples for the minority class, while under-sampling removes instances from the dominant class. However, under-sampling can lead to information loss. By integrating feature selection and resampling, models are trained on balanced and pertinent datasets, improving accuracy and real-world applicability. These techniques effectively address challenges like class imbalances and high-dimensional feature spaces.
A. 
Tuning hyperparameters
Hyperparameter optimization is essential for enhancing the ensemble-based AD diagnosis model. This process finds the best values for key model parameters like learning rate, depth of decision trees, or number of neighbors in K-Nearest Neighbors (KNNs). Methods like Grid Search or Random Search are used, combined with cross-validation to avoid overfitting. The model’s performance is tested with various hyperparameter combinations using metrics like the accuracy, F1-score, or AUC-ROC. After identifying the optimal hyperparameters, the model is validated on unseen data to confirm its reliability. This thorough tuning ensures the model’s peak accuracy in an AD diagnosis, benefiting patient care and disease management.
B. 
Grid-Search tuning
During the model’s hyperparameter tuning, various parameter combinations were explored to optimize performance. We examined the ‘Bootstrap’ parameter using both ‘True’ and ‘False’. Handling missing data, the maximum model depth was tested with values of 5 and 7, and the maximum features were assessed with options of 3 and 4. We also evaluated the impact of minimum sample leaf values of 3 and 4. The decision trees’ minimum sample split values were tried at 3, 5, and 7, while the number of estimators was tested with 200, 400, and 600. By assessing these combinations, we identified the optimal configuration for the best model performance. This rigorous tuning improved the model’s predictive capabilities.
C. 
Optimal RF hyperparameters
During hyperparameter tuning, we adjusted several parameters to enhance the model’s performance. We set the “Bootstrap” to “False”, the “Maximum depth” to seven layers, and “Maximum features” to four. The “Minimum samples leaf” was fixed at three, while “Minimum samples split” required seven samples. The model employed 200 estimators as indicated by the “n_estimators” value. These adjustments optimized the model’s performance, ensuring more accurate predictions. Proper hyperparameter tuning is vital for improved model results and capabilities.
D. 
Effectiveness Evaluation
For the study article on AD diagnoses, the ensemble-based model must be evaluated and compared to different classification methods at several significance levels. To correct class imbalance, the dataset is prepared, preprocessed, and balanced using SMOTE. Cross-validation divides the balanced dataset into training and testing sets for proper evaluation. Figure 2 represents the relationship with various models. The ensemble-based model is trained using optimized hyperparameters and SVM and Logistic Regression classifiers.
E. 
Precision
It quantifies the model’s True Positive predictions. A high accuracy score means the model reliably predicts positive situations, whereas a low score means it produces many erroneous positive predictions. Figure 3 represents the precision score for different models.
Precision = (True Positives)/((True Positives + False Positives))
The precision graph represented in Figure 3 clearly illustrates the varying precision scores of predictive algorithms. SVM stands out with an impressive 96%, indicating accurate positive predictions. Extra Tree shows a lower 76%, while the decision tree, Logistic Regression, and XG Boost perform moderately at 81%. SVM’s dominance is evident.
F. 
Recall
Recall is a performance statistic for binary classification models. It tests the model’s ability to identify all positive occurrences from the dataset’s total positive instances. Recall is sensitivity and the True Positive rate (TPR). Figure 4 represents the recall score for different models. The ratio of True Positive predictions (properly recognized positive cases) to the total of True Positive and False Negative predictions (positive instances mistakenly forecasted as negative) is used.
Recall = (True Positives)/((True Positives + False Negatives))
A high recall score suggests that the model can properly identify a significant proportion of positive cases, meaning few False Negatives. A low recall score means the model misses many positive examples, resulting in more False Negatives. Recall is critical in medical diagnoses (to detect illnesses) and fraud detection (to detect fraudulent transactions) to accurately identify positive instances. However, optimizing one statistic might affect other metrics in a classification assignment; therefore, it is important to balance recall and other metrics like accuracy.
The recall scores for various models were evaluated to quantify their ability to accurately identify positive cases in the dataset. The SVM algorithm exhibited an outstanding recall score of 97%, correctly identifying 97% of the positive cases. Surprisingly, the KNN algorithm surpassed even the SVM, obtaining a recall score of 95%, demonstrating its effectiveness in correctly identifying positive cases. In contrast, the decision tree algorithm achieved a lower recall score of 84%, indicating that it missed a considerable portion of the positive cases. The Naive Bayes model achieved a recall score of 75%, while the Logistic Regression model performed relatively better with a recall score of 81%. Overall, the results highlight the superior performance of the KNN model in identifying positive cases compared to the other four algorithms. The models’ diagnostic skills on the testing set are assessed using the accuracy, precision, recall, F1-score, and AUC-ROC. The confusion matrix also assesses true, false, positive, and negative predictions. External validation on a different dataset assesses the model’s capacity to generalize to unobserved data by comparing the proposed model’s performance to the baseline classifiers and using statistical tests to discover performance differences.
G. 
F1-Score
A model with a high F1-score is one that effectively balances precision and recall. Evaluating the F1-scores of the aforementioned models would provide a more thorough comprehension of their overall effectiveness and potential trade-offs between precision and recall. Figure 5 represents a confusion matrix on prediction of Alzheimer’s.
Visualization methods like ROC curves and precision–recall curves show the model’s discrimination performance, whereas a feature significance analysis shows feature contributions. Figure 5 illustrates an AD prediction confusion matrix. True Positive (TP) occurrences are accurately predicted as including the disease, while False Positive (FP) instances are wrongly forecasted as positive. True Negative (TN) occurrences were accurately predicted as negative (not having the disease), while False Negative (FN) examples were mistakenly forecasted as negative but included the disease. This research study evaluates the suggested ensemble-based model for an AD diagnosis to improve medical data analytics and patient care by revealing its accuracy and efficacy.

5. Conclusions

This research offers an in-depth study of an AD diagnosis through machine learning. Using feature selection and data resampling, our proposed ensemble-based model effectively differentiates between healthy individuals and AD patients. It outperforms baseline classifiers like SVM and Logical Regression in accuracy, precision, and other metrics. Relevant features enhance the model’s clarity and effectiveness, while SMOTE balancing addresses class imbalance. This work contributes significantly to AD diagnoses, promoting early detection and better patient outcomes. Future studies could explore deep learning techniques, such as CNNs and RNNs, for improved brain imaging pattern recognition. Combining varied data sources, like genetics and clinical data, might refine the diagnosis. A longitudinal patient data analysis can track disease progression and risk prediction. Collaborating with medical experts for real-world validation, improving model interpretability, and integrating it into clinical systems will further its potential in AD diagnoses and treatment.

Author Contributions

Experiment Design and Data Pre-processing, R.P.N. and B.S.; Design, R.K.; Review and Interpretation, M.A.B.; Data Analysis and Interpreted Result, R.P.N.; Writing—Review and Editing, R.P.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be provided on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Grueso, S.; Viejo-Sobera, R. Machine learning methods for predicting progression from mild cognitive impairment to AD dementia: A systematic review. Alzheimer’s Res. Ther. 2021, 13, 1–29. [Google Scholar]
  2. El-Sappagh, S.; Saleh, H.; Ali, F.; Amer, E.; Abuhmed, T. Two-stage deep learning model for AD detection and prediction of the mild cognitive impairment time. Neural Comp. Appl. 2022, 34, 14487–14509. [Google Scholar] [CrossRef]
  3. Iddi, S.; Li, D.; Aisen, P.S.; Rafii, M.S.; Thompson, W.K.; Donohue, M.C. AD Neuroimaging Initiative. Predicting the course of Alzheimer’s progression. Brain Inform. 2019, 6, 1–18. [Google Scholar] [CrossRef]
  4. Yuen, S.C.; Liang, X.; Zhu, H.; Jia, Y.; Leung, S.W. Prediction of differentially expressed microRNAs in blood as potential biomarkers for AD by meta-analysis and adaptive boosting ensemble learning. Alzheimer’s Res. Ther. 2021, 13, 1–30. [Google Scholar]
  5. Naz, S.; Ashraf, A.; Zaib, A. Transfer learning using freeze features for Alzheimer neurological disorder detection using ADNI dataset. Multi. Syst. 2022, 28, 85–94. [Google Scholar] [CrossRef]
  6. Wang, S.; Du, Z.; Ding, M.; Rodriguez-Paton, A.; Song, T. KG-DTI: A knowledge graph based deep learning method for drug-target interaction predictions and AD drug repositions. Appl. Intell. 2022, 52, 846–857. [Google Scholar] [CrossRef]
  7. Bermudez, C.; Graff-Radford, J.; Syrjanen, J.A.; Stricker, N.H.; Algeciras-Schimnich, A.; Kouri, N.; Vemuri, P. Plasma biomarkers for prediction of AD neuropathologic change. Acta Neuropathol. 2023, 146, 13–29. [Google Scholar] [CrossRef]
  8. Diogo, V.S.; Ferreira, H.A.; Prata, D. AD Neuroimaging Initiative. Early diagnosis of AD using machine learning: A multi-diagnostic, generalizable approach. Alzheimer’s Res. Ther. 2022, 14, 107. [Google Scholar] [CrossRef]
  9. Venkataramana, L.Y.; Jacob, S.G.; Prasad, V.; Athilakshmi, R.; Priyanka, V.; Yeshwanthraa, K.; Vigneswaran, S. Geometric SMOTE-Based Approach to Improve the Prediction of Alzheimer’s and Parkinson’s Diseases for Highly Class-Imbalanced Data. In AI, IoT, and Blockchain Breakthroughs in E-Governance; IGI Global: Hershey, PA, USA, 2023; pp. 114–137. [Google Scholar]
  10. Zhang, P.; Lin, S.; Qiao, J.; Tu, Y. Diagnosis of AD with ensemble learning classifier and 3D convolutional neural network. Sensors 2021, 21, 7634. [Google Scholar] [CrossRef]
  11. Rao, K.N.; Gandhi, B.R.; Rao, M.V.; Javvadi, S.; Vellela, S.S.; Basha, S.K. Prediction and Classification of AD using Machine Learning Techniques in 3D MR Images. In Proceedings of the 2023 International Conference on Sustainable Computing and Smart Systems (ICSCSS), Coimbatore, India, 14–16 June 2023; pp. 85–90. [Google Scholar]
  12. Khoei, T.T.; Labuhn, M.C.; Caleb, T.D.; Hu, W.C.; Kaabouch, N. A stacking-based ensemble learning model with genetic algorithm for detecting early stages of AD. In Proceedings of the 2021 IEEE International Conference on Electro Information Technology (EIT), Mt. Pleasant, MI, USA, 14–15 May 2021; pp. 215–222. [Google Scholar]
  13. Tambe, P.; Saigaonkar, R.; Devadiga, N.; Chitte, P.H. Deep Learning techniques for effective diagnosis of AD using MRI images. ITM Web Conf. 2021, 40, 03021. [Google Scholar] [CrossRef]
  14. Ghali, U.M.; Usman, A.G.; Chellube, Z.M.; Degm, M.A.A.; Hoti, K.; Umar, H.; Abba, S.I. Advanced chromatographic technique for performance simulation of anti-Alzheimer agent: An ensemble machine learning approach. SN Appl. Sci. 2020, 2, 1–12. [Google Scholar] [CrossRef]
  15. An, N.; Ding, H.; Yang, J.; Au, R.; Ang, T.F. Deep ensemble learning for AD classification. J. Biomed. Inform. 2020, 105, 103411. [Google Scholar]
  16. Pereira, T.; Cardoso, S.; Silva, D.; Guerreiro, M.; de Mendonça, A.; Madeira, S.C. Ensemble learning with Conformal Predictors: Targeting credible predictions of conversion from Mild Cognitive Impairment to AD. arXiv 2018, arXiv:1807.01619. [Google Scholar]
  17. Wang, R.; Li, H.; Lan, R.; Luo, S.; Luo, X. Hierarchical Ensemble Learning for AD Classification. In Proceedings of the 2018 7th International Conference on Digital Home (ICDH), Guilin, China, 30 November–1 December 2018; pp. 224–229. [Google Scholar]
  18. Orhobor, O.I.; Soldatova, L.N.; King, R.D. Federated ensemble regression using classification. In Proceedings of the Discovery Science: 23rd International Conference, DS 2020, Thessaloniki, Greece, 19–21 October 2020; Springer International Publishing: Berlin/Heidelberg, Germany; Volume 23, pp. 325–339. [Google Scholar]
  19. Kang, W.; Lin, L.; Zhang, B.; Shen, X.; Wu, S. AD Neuroimaging Initiative. Multi-model and multi-slice ensemble learning architecture based on 2D convolutional neural networks for AD diagnosis. Comput. Biol. Med. 2021, 136, 104678. [Google Scholar]
  20. Mirzaei, G.; Adeli, H. Machine learning techniques for diagnosis of Alzheimer disease, mild cognitive disorder, and other types of dementia. Biomed. Sign. Process. Control 2022, 72, 103293. [Google Scholar] [CrossRef]
  21. Nguyen, D.K.; Lan, C.H.; Chan, C.L. Deep ensemble learning approaches in healthcare to enhance the prediction and diagnosing performance: The workflows, deployments, and surveys on the statistical, image-based, and sequential datasets. Int. J. Environ. Res. Public Health 2021, 18, 10811. [Google Scholar] [CrossRef]
  22. Shaikh, T.A.; Ali, R. Enhanced computerised diagnosis of AD from brain MRI images using a classifier merger strategy. Int. J. Inform. Technol. 2021, 14, 1–13. [Google Scholar]
  23. Song, M.; Jung, H.; Lee, S.; Kim, D.; Ahn, M. Diagnostic classification and biomarker identification of AD with random forest algorithm. Brain Sci. 2021, 11, 453. [Google Scholar] [CrossRef]
  24. Hemalatha, B.; Renukadevi, M. Analysis of Alzheimer disease prediction using machine learning techniques. Inf. Technol. Ind. 2021, 9, 519–525. [Google Scholar]
  25. Alamro, H.; Thafar, M.A.; Albaradei, S.; Gojobori, T.; Essack, M.; Gao, X. Exploiting machine learning models to identify novel AD biomarkers and potential targets. Sci. Rep. 2023, 13, 4979. [Google Scholar] [CrossRef]
  26. Albahri, A.S.; Alwan, K.J.; Taha, Z.K.; Ismail, S.F.; Hamid, R.A.; Zaidan, A.A.; Albahri, O.S.; Zaidan, B.B.; Alamoodi, A.H.; Alsalem, M.A. IoT-based telemedicine for disease prevention and health promotion: State-of-the-Art. J. Netw. Comput. Appl. 2021, 173, 102873. [Google Scholar] [CrossRef]
Figure 1. IoT-based patient monitoring system.
Figure 1. IoT-based patient monitoring system.
Engproc 59 00010 g001
Figure 2. A representation of the relationship between various models.
Figure 2. A representation of the relationship between various models.
Engproc 59 00010 g002
Figure 3. A representation of precision scores for different models.
Figure 3. A representation of precision scores for different models.
Engproc 59 00010 g003
Figure 4. A representation of recall score for different model.
Figure 4. A representation of recall score for different model.
Engproc 59 00010 g004
Figure 5. Confusion matrix on prediction of Alzheimer’s.
Figure 5. Confusion matrix on prediction of Alzheimer’s.
Engproc 59 00010 g005
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Neelakandan, R.P.; Kandasamy, R.; Subbiyan, B.; Bennet, M.A. Early Detection of Alzheimer’s Disease: An Extensive Review of Advancements in Machine Learning Mechanisms Using an Ensemble and Deep Learning Technique. Eng. Proc. 2023, 59, 10. https://doi.org/10.3390/engproc2023059010

AMA Style

Neelakandan RP, Kandasamy R, Subbiyan B, Bennet MA. Early Detection of Alzheimer’s Disease: An Extensive Review of Advancements in Machine Learning Mechanisms Using an Ensemble and Deep Learning Technique. Engineering Proceedings. 2023; 59(1):10. https://doi.org/10.3390/engproc2023059010

Chicago/Turabian Style

Neelakandan, Renjith Prabhavathi, Ramesh Kandasamy, Balasubramani Subbiyan, and Mariya Anto Bennet. 2023. "Early Detection of Alzheimer’s Disease: An Extensive Review of Advancements in Machine Learning Mechanisms Using an Ensemble and Deep Learning Technique" Engineering Proceedings 59, no. 1: 10. https://doi.org/10.3390/engproc2023059010

Article Metrics

Back to TopTop