Machine Learning Algorithms for Depression: Diagnosis, Insights, and Research Directions
Abstract
:1. Introduction
2. Related Work
3. Methodology for Depression Diagnosis
3.1. Pre-Processing Algorithms
- (1)
- Linear Discriminant Analysis (LDA): LDA is a dimensionality reduction approach that removes redundant features by transforming them from a spatial space onto a lower-dimensional space. LDA reduces the dimensions in each dataset, retains the most important features, and achieves higher class separability [31].
- (2)
- Synthetic Minority Oversampling Technique (SMOTE): SMOTE is a statistical oversampling technique to obtain a synthetically class-balanced dataset. It provides a balanced class distribution that develops synthetic patterns from the minority class [32].
- (3)
- Linguistic Inquiry and Word Count (LIWC): LIWC is a text analysis technique for understanding different emotional, subjective, and structural components present in the spoken and written speech patterns [33].
- (4)
- Hidden Markov Model (HMM): HMM is a probabilistic model used to capture and describe information from observable sequential symbols. In HMM, the observed data are modeled as a series of outputs generated by several internal states [34].
3.2. Feature Extraction Methods
- (1)
- SelectKBest: SelectKBest is a feature extraction approach that retains relevant features and drops unwanted features in the given input data. It is a univariate feature selection approach based on the univariate statistical analysis. It combines the univariate statistical test with selecting the K-number of features based on the statistical result between the variables.
- (2)
- Particle Swarm Optimization (PSO): PSO is a computational process for optimizing nonlinear functions by developing the candidate solution in a repetitive pattern based on a defined quality measure. The general concept of the PSO algorithm is inspired by the swarm actions of birds, flocking, and schooling in nature [35].
- (3)
- Maximum Relevance Minimum Redundancy (mRMR): mRMR is a feature selection approach that manages multivariate temporal data without compressing previous data. The algorithm selects features with the most relevant class and the least correlation between redundant classes. It provides significantly improved class predictions in extensive datasets [36].
- (4)
- Boruta: Boruta is a feature selection approach designed around a Random Forest classification. Boruta is used for extracting all the relevant variables by removing less relevant features, using the statistical analysis iteratively [37].
- (5)
- RELIEFF: RELIEFF algorithm is one of the most successful filtering feature selection methods. RELIEFF algorithm is used to eliminate the redundant features [38].
3.3. Supervised Learning Classifiers
3.3.1. Classification
- (1)
- Naïve Bayes Classifier: A Naive Bayes classifier is dependent on applying Bayes’ hypothesis with strong independence assumptions. This classifier depends on basic learning strategies assembled by similitudes that utilize Bayes’ hypothesis of probability to build ML models, particularly those identified with report order and disease prediction [39].
- (2)
- KNN Classifier: KNN is used for data regression and classification based on the count of k neighbors [40].
- (3)
- Support Vector Machine Classifier (SVM): SVM [41] is a supervised ML model investigating regression analysis and classification data. It also uses the classification for two-group classification problems. SVM are nonparametric classifiers. For the training set, inputs and outputs are paired in SVM. Decision functions are attained through the input–output pairs that classify the input variables into the new and test datasets. i. Multikernel SVM: Multikernel SVM [42] is a feature selection approach based on oversampling and a hybrid algorithm for improving the classification of binary imbalanced classes.
- (4)
- Decision Tree (DT) Classifier: A DT [43] is a tree-like graph used as a decision support tool. It works with discrete-valued parameters and an inductive philosophy for decision tree is “A good decision tree should be as small as possible”.Decision Tree Ensembles:
- Bagging (RF, DF): Bagging is an ensemble algorithm [8]. It adapts various algorithms on different fragments of a training dataset. The predictions from all algorithms are then combined. Random Forest (RF), an extension of bagging, selects the features fragments in random patterns from the given dataset.
- Boosting (GBDT, XGBoost): Gradient Boosting is an ensemble classifier used for supervised ML tasks. It considers the individual algorithms and forms a collective model.
3.3.2. Regression
- (1)
- Logistic Regression: When the dependent variable is dichotomous, logistic regression is the best regression technique to use (binary). Logistic regression is employed to describe and explain the connection between one dependent binary variable and one or more nominal, ordinal, interval, or ratio-level independent variables.
- (2)
- Lasso Regression: Lasso regression is a form of shrinkage-based linear regression. Data values are shrunk toward a center; mean in shrinkage. Simple, sparse models are encouraged by the lasso method.
- (3)
- Elastic Net: Elastic net is a regularized linear regression that incorporates two well-known penalties, the L1 and L2 penalty functions.
- (4)
- SVR: Support vector regression (SVR) allows the flexibility to define how much error is acceptable in each model and find an appropriate line to fit the data.
3.3.3. Deep Learning
- (1)
- Convolutional Neural Network: ConvNet, also known as CNN, is a deep learning (DL) method that can take an input picture and assign significance (learnable weights and biases) to various aspects/objects in the image, as well as distinguish one from the other. Compared to other classification methods, the amount of pre-processing needed by a ConvNet is much less. While filters are hand-engineered in basic techniques, ConvNets can learn these filters/characteristics with enough training.
- (2)
- Artificial Neural Network (ANN): ANNs depend on the structure of many interaction units. The preparing unit receives signals from different neurons, consolidates, transforms them, and creates an outcome. The cycle units are generally compared to genuine neurons, giving the artificial neural networks.
- (3)
- DNN: An artificial neural network (ANN) having many layers between the input and output layers is known as a deep neural network (DNN) [9]. Various neural networks have different components, but they all have the same components: neurons, synapses, weights, biases, and functions. These components work in the same way as the human brain and can be taught just like any other machine learning algorithm.
- (4)
- DCNN: A deep convolutional neural network (DCNN) comprises many layers of neural networks. Convolutional and pooling layers are usually alternated in most cases. From left to right in the network, the depth of each filter rises. The final level is usually made up of one or more layers that are completely linked.
- (5)
- RNN: Recurrent neural networks (RNNs) are utilized in language modeling applications because input may flow in either direction. For this purpose, long short-term memory is especially useful. Long short-term memory (LSTM) is a recurrent neural network design utilized in deep learning. LSTM contains feedback connections, unlike conventional feedforward neural networks. It can handle large data sequences as well as single data points.
- (6)
- AiME (Novel Model): An artificial intelligence mental evaluation (AiME) framework for detecting symptoms of depression using multimodal deep networks-based human–computer interactive evaluation.
4. Depression Detection Models
4.1. Classification Models
Discussion of Classification Models
4.2. Deep Learning Models
Discussion of Deep Learning Models
4.3. Ensemble Models
Discussion of Ensemble Models
5. Future Research Possibilities
- (1)
- A larger data sample is required:
- (2)
- Learning method(s):
- (3)
- Clinical application:
- (4)
- Collaboration of research groups:
- (5)
- Availability of databases:
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Vitriol, V.; Cancino, A.; Weil, K.; Salgado, C.; Asenjo, M.A.; Potthoff, S. Depression and psychological trauma: An overview integrating current research and specific evidence of studies in the treatment of depression in public mental health services in chile. Depress. Res. Treat. 2014, 2014, 608671. [Google Scholar] [CrossRef] [PubMed]
- World Mental Health Day: An Opportunity to Kick-Start a Massive Scale-Up in Investment in Mental Health. Available online: https://www.who.int/news/item/27-08-2020-world-mental-health-day-an-opportunity-to-kick-start-a-massive-scale-up-in-investment-in-mental-health (accessed on 20 February 2022).
- GBD 2017 Disease and Injury Incidence and Prevalence Collaborators. Global, regional, and national incidence, prevalence, and years lived with disability for 354 diseases and injuries for 195 countries and territories, 1990–2017: A systematic analysis for the Global Burden of Disease Study 2017. Lancet 2018, 392, 1789–1858. [Google Scholar] [CrossRef] [Green Version]
- Strunk, D.R.; Pfeifer, B.J.; Ezawa, I.D. Depression. In Handbook of Cognitive Behavioral Therapy: Applications; Wenzel, A., Ed.; American Psychological Association: Washington, DC, USA, 2021; pp. 3–31. [Google Scholar] [CrossRef]
- World Health Organization. Depression and Other Common Mental Disorders: Global Health Estimates. Available online: https://apps.who.int/iris/bitstream/handle/10665/254610/WHO-MSD-MER-2017.2-eng.pdf (accessed on 20 February 2022).
- ÇELİK, Ö.; Altunaydin, S.S. A research on machine learning methods and its applications. J. Educ. Technol. Online Learn. 2018, 1, 25–40. [Google Scholar] [CrossRef] [Green Version]
- Shalev-Shwartz, S.; Ben-David, S. Decision Trees. Understanding Machine Learning; Cambridge University Press: Cambridge, UK, 2014. [Google Scholar]
- Kotsiantis, S.B. Supervised machine learning: A review of classification techniques. Informatica 2007, 31, 249–268. [Google Scholar]
- Alloghani, M.A.; Al-Jumeily, D.; Mustafina, J.; Hussain, A.; Aljaaf, A.J. A systematic review on supervised and unsupervised machine learning algorithms for data science. In Supervised and Unsupervised Learning for Data Science; Berry, M., Mohamed, A., Yap, B., Eds.; Springer: Cham, Switzerland, 2020. [Google Scholar]
- Van Engelen, J.E.; Hoos, H.H. A survey on semi-supervised learning. Mach. Learn. 2020, 109, 373–440. [Google Scholar] [CrossRef] [Green Version]
- Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT press: Cambridge, MA, USA, 2018. [Google Scholar]
- Dwyer, D.B.; Falkai, P.; Koutsouleris, N. Machine learning approaches for clinical psychology and psychiatry. Annu. Rev. Clin. Psychol. 2018, 14, 91–118. [Google Scholar] [CrossRef]
- Sidey-Gibbons, J.A.; Sidey-Gibbons, C.J. Machine learning in medicine: A practical introduction. BMC Med. Res. Methodol. 2019, 19, 1–18. [Google Scholar] [CrossRef] [Green Version]
- Thieme, A.; Belgrave, D.; Doherty, G. Machine Learning in Mental Health: A Systematic Review of the HCI Literature to Support the Development of Effective and Implementable ML Systems. ACM Trans. Comput.-Hum. Interact. 2020, 27, 34. [Google Scholar] [CrossRef]
- Chen, X.; Sykora, M.D.; Jackson, T.W.; Elayan, S. What about mood swings: Identifying depression on twitter with temporal measures of emotions. In Proceedings of the the Web Conference, Lyon, France, 23–27 April 2018; pp. 1653–1660. [Google Scholar]
- Joshi, D.J.; Makhija, M.; Nabar, Y.; Nehete, N.; Patwardhan, M.S. Mental health analysis using deep learning for feature extraction. In Proceedings of the ACM India Joint International Conference on Data Science and Management of Data, Goa, India, 11–13 January 2018; pp. 356–359. [Google Scholar]
- Aldabbas, H.; Albashish, D.; Khatatneh, K.; Amin, R. An Architecture of IoT-Aware Healthcare Smart System by Leveraging Machine Learning. Int. Arab J. Inf. Technol. 2022, 19, 160–172. [Google Scholar] [CrossRef]
- Amin, R.; Al Ghamdi, M.A.; Almotiri, S.H.; Alruily, M. Healthcare Techniques Through Deep Learning: Issues, Challenges and Opportunities. IEEE Access 2021, 9, 98523–98541. [Google Scholar]
- Morshed, M.B.; Saha, K.; Li, R.; D’Mello, S.K.; De Choudhury, M.; Abowd, G.D.; Plötz, T. Prediction of mood instability with passive sensing. ACM 2019, 3, 1–21. [Google Scholar] [CrossRef]
- Diederich, J.; Al-Ajmi, A.; Yellowlees, P. Ex-ray: Data mining and mental health. Appl. Soft Comput. 2007, 7, 923–928. [Google Scholar] [CrossRef]
- Adamou, M.; Antoniou, G.; Greasidou, E.; Lagani, V.; Charonyktakis, P.; Tsamardinos, I. Mining free-text medical notes for suicide risk assessment. In Proceedings of the 10th Hellenic Conference on Artificial Intelligence, Patras, Greece, 9–12 July 2018; pp. 1–8. [Google Scholar]
- Tran, T.; Phung, D.; Luo, W.; Harvey, R.; Berk, M.; Venkatesh, S. An integrated framework for suicide risk prediction. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 11–14 August 2013; pp. 1410–1418. [Google Scholar]
- Nobles, A.L.; Glenn, J.J.; Kowsari, K.; Teachman, B.A.; Barnes, L.E. Identification of imminent suicide risk among young adults using text messages. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, Montreal, QC, Canada, 21–26 April 2018; pp. 1–11. [Google Scholar]
- Patel, M.J.; Khalaf, A.; Aizenstein, H.J. Studying depression using imaging and machine learning methods. NeuroImage Clin. 2016, 10, 115–123. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Garcia-Ceja, E.; Riegler, M.; Nordgreen, T.; Jakobsen, P.; Oedegaard, K.J.; Tørresen, J. Mental health monitoring with multimodal sensing and machine learning: A survey. Pervasive Mob. Comput. 2018, 51, 1–26. [Google Scholar] [CrossRef]
- Gao, S.; Calhoun, V.D.; Sui, J. Machine learning in major depression: From classification to treatment outcome prediction. CNS Neurosci. Ther. 2018, 24, 1037–1052. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Cho, G.; Yim, J.; Choi, Y.; Ko, J.; Lee, S.-H. Review of Machine Learning Algorithms for Diagnosing Mental Illness. Psychiatry Investig. 2019, 16, 262–269. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Mahdy, N.; Magdi, D.A.; Dahroug, A.; Rizka, M.A. Comparative Study: Different Techniques to Detect Depression Using Social Media. In Internet of Things-Applications and Future; Springer: Singapore, 2020; pp. 441–452. [Google Scholar] [CrossRef]
- Liu, G.-D.; Li, Y.-C.; Zhang, W.; Zhang, L. A Brief Review of Artificial Intelligence Applications and Algorithms for Psychiatric Disorders. Engineering 2020, 6, 462–467. [Google Scholar] [CrossRef]
- He, L.; Niu, M.; Tiwari, P.; Marttinen, P.; Su, R.; Jiang, J.; Guo, C.; Wang, H.; Ding, S.; Wang, Z.; et al. Deep learning for depression recognition with audiovisual cues: A review. Inf. Fusion 2021, 80, 56–86. [Google Scholar] [CrossRef]
- Tharwat, A.; Gaber, T.; Ibrahim, A.; Hassanien, A.E. Linear discriminant analysis: A detailed tutorial. AI Commun. 2017, 30, 169–190. [Google Scholar] [CrossRef] [Green Version]
- Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
- Wegner, D.M. The Illusion of Conscious Will; MIT Press: Cambridge, MA, USA, 2017. [Google Scholar]
- Rabiner, L.; Juang, B.-H. Fundamentals of Speech Recognition; Prentice Hall: Hoboken, NJ, USA, 1993. [Google Scholar]
- Koohi, I.; Groza, V.Z. Optimizing particle swarm optimization algorithm. In Proceedings of the 2014 IEEE 27th Canadian conference on Electrical and Computer Engineering (CCECE), Toronto, ON, Canada, 4–7 May 2014. [Google Scholar]
- Bugata, P.; Drotar, P. On some aspects of minimum redundancy maximum relevance feature selection. Sci. China Inf. Sci. 2020, 63, 1–15. [Google Scholar] [CrossRef] [Green Version]
- Kursa, M.B.; Jankowski, A.; Rudnicki, W.R. Boruta—A system for feature selection. Fundam. Inform. 2010, 101, 271–285. [Google Scholar] [CrossRef]
- Urbanowicz, R.J.; Meeker, M.; La Cava, W.; Olson, R.S.; Moore, J.H. Relief-based feature selection: Introduction and review. J. Biomed. Inform. 2018, 85, 189–203. [Google Scholar] [CrossRef] [PubMed]
- Rish, I. An empirical study of the naive Bayes classifier. IJCAI 2001, 3, 41–46. [Google Scholar]
- Cunningham, P.; Delany, S.J. k-Nearest neighbour classifiers: 2nd Edition (with Python examples). arXiv 2020, arXiv:2004.04523. [Google Scholar]
- Cristianini, N.; Ricci, E. Encyclopedia of Algorithms; Springer: Boston, MA, USA, 2008. [Google Scholar]
- Chen, Z.; Li, J. A multiple kernel support vector machine scheme for simultaneous feature selection and rule-based classification. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Nanjing, China, 22–25 May 2007; pp. 441–448. [Google Scholar]
- Patel, H.H.; Prajapati, P. Study and Analysis of Decision Tree Based Classification Algorithms. Int. J. Comput. Sci. Eng. 2018, 6, 74–78. [Google Scholar] [CrossRef]
- Mathers, C.D.; Loncar, D. Projections of global mortality and burden of disease from 2002 to 2030. PLoS Med. 2006, 3, e442. [Google Scholar] [CrossRef] [Green Version]
- Miyajima, A.; Tanaka, M.; Itoh, T. Stem/progenitor cells in liver development, homeostasis, regeneration, and reprogramming. Cell Stem Cell 2014, 14, 561–574. [Google Scholar] [CrossRef] [Green Version]
- Zulfiker, M.S.; Kabir, N.; Biswas, A.A.; Nazneen, T.; Uddin, M.S. An in-depth analysis of machine learning approaches to predict depression. Curr. Res. Behav. Sci. 2021, 2, 100044. [Google Scholar] [CrossRef]
- Na, K.-S.; Cho, S.-E.; Geem, Z.W.; Kim, Y.-K. Predicting future onset of depression among community dwelling adults in the Republic of Korea using a machine learning algorithm. Neurosci. Lett. 2020, 721, 134804. [Google Scholar] [CrossRef]
- Choudhury, A.A.; Khan, R.H.; Nahim, N.Z.; Tulon, S.R.; Islam, S.; Chakrabarty, A. Predicting Depression in Bangladeshi Undergraduates using Machine Learning. In Proceedings of the 2019 IEEE Region 10 Symposium (TENSYMP), Kolkata, India, 7–9 June 2019; pp. 789–794. [Google Scholar] [CrossRef] [Green Version]
- Li, X.; Zhang, X.; Zhu, J.; Mao, W.; Sun, S.; Wang, Z.; Xia, C.; Hu, B. Depression recognition using machine learning methods with different feature generation strategies. Artif. Intell. Med. 2019, 99, 101696. [Google Scholar] [CrossRef] [PubMed]
- Priya, A.; Garg, S.; Tigga, N.P. Predicting Anxiety, Depression and Stress in Modern Life using Machine Learning Algorithms. Procedia Comput. Sci. 2020, 167, 1258–1267. [Google Scholar] [CrossRef]
- Fatima, I.; Mukhtar, H.; Ahmad, H.F.; Rajpoot, K. Analysis of user-generated content from online social communities to characterise and predict depression degree. J. Inf. Sci. 2018, 44, 683–695. [Google Scholar] [CrossRef]
- Sharma, A.; Verbeke, W.J.M.I. Improving Diagnosis of Depression with XGBOOST Machine Learning Model and a Large Biomarkers Dutch Dataset (n = 11,081). Front. Big Data 2020, 3, 15. [Google Scholar] [CrossRef]
- Hilbert, K.; Lueken, U.; Muehlhan, M.; Beesdo-Baum, K. Separating generalized anxiety disorder from major depression using clinical, hormonal, and structural MRI data: A multimodal machine learning study. Brain Behav. 2017, 7, e00633. [Google Scholar] [CrossRef] [Green Version]
- Guo, H.; Qin, M.; Chen, J.; Xu, Y.; Xiang, J. Machine-Learning Classifier for Patients with Major Depressive Disorder: Multifeature Approach Based on a High-Order Minimum Spanning Tree Functional Brain Network. Comput. Math. Methods Med. 2017, 2017, 4820935. [Google Scholar] [CrossRef] [Green Version]
- Nagpal, G.; Chaudhary, K.; Agrawal, P.; Raghava, G.P.S. Computer-aided prediction of antigen presenting cell modulators for designing peptide-based vaccine adjuvants. J. Transl. Med. 2018, 16, 181. [Google Scholar] [CrossRef] [Green Version]
- Mumtaz, W.; Qayyum, A. A deep learning framework for automatic diagnosis of unipolar depression. Int. J. Med Inform. 2019, 132, 103983. [Google Scholar] [CrossRef]
- Erguzel, T.T.; Sayar, G.H.; Tarhan, N. Artificial intelligence approach to classify unipolar and bipolar depressive disorders. Neural Comput. Appl. 2016, 27, 1607–1616. [Google Scholar] [CrossRef]
- Cong, Q.; Feng, Z.; Li, F.; Xiang, Y.; Rao, G.; Tao, C. XA-BiLSTM: A deep learning approach for depression detection in imbalanced data. In Proceedings of the 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 1624–1627. [Google Scholar]
- Zogan, H.; Wang, X.; Jameel, S.; Xu, G. Depression detection with multi-modalities using a hybrid deep learning model on social media. CoRR 2020, abs/2007.02847. [Google Scholar]
- He, L.; Cao, C. Automated depression analysis using convolutional neural networks from speech. J. Biomed. Inform. 2018, 83, 103–111. [Google Scholar] [CrossRef] [PubMed]
- Li, X.; La, R.; Wang, Y.; Hu, B.; Zhang, X. A Deep Learning Approach for Mild Depression Recognition Based on Functional Connectivity Using Electroencephalography. Front. Neurosci. 2020, 14, 192. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Shah, F.M.; Ahmed, F.; Joy, S.K.S.; Ahmed, S.; Sadek, S.; Shil, R.; Kabir, H. Early Depression Detection from Social Network Using Deep Learning Techniques. In Proceedings of the 2020 IEEE Region 10 Symposium (TENSYMP), Dhaka, Bangladesh, 5–7 June 2020; pp. 823–826. [Google Scholar] [CrossRef]
- Orabi, A.H.; Buddhitha, P.; Orabi, M.H.; Inkpen, D. Deep learning for depression detection of twitter users. In Proceedings of the Fifth Workshop on Computational Linguistics and Clinical Psychology: From Keyboard to Clinic, New Orleans, LA, USA, 5 June 2018; pp. 88–97. [Google Scholar]
- Tennenhouse, L.G.; Marrie, R.A.; Bernstein, C.N.; Lix, L.M. Machine-learning models for depression and anxiety in individuals with immune-mediated inflammatory disease. J. Psychosom. Res. 2020, 134, 110126. [Google Scholar] [CrossRef]
- Su, D.; Zhang, X.; He, K.; Chen, Y. Use of machine learning approach to predict depression in the elderly in China: A longitudinal study. J. Affect. Disord. 2021, 282, 289–298. [Google Scholar] [CrossRef] [PubMed]
- Tao, X.; Chi, O.; Delaney, P.J.; Li, L.; Huang, J. Detecting depression using an ensemble classifier based on Quality of Life scales. Brain Inform. 2021, 8, 2. [Google Scholar] [CrossRef]
- Karoly, P.; Ruehlman, L.S. Psychological “resilience” and its correlates in chronic pain: Findings from a national community sample. Pain 2006, 123, 90–97. [Google Scholar] [CrossRef] [PubMed]
- Zhao, M.; Feng, Z. Machine Learning Methods to Evaluate the Depression Status of Chinese Recruits: A Diagnostic Study. Neuropsychiatr. Dis. Treat. 2020, 16, 2743–2752. [Google Scholar] [CrossRef]
- Ma, Y.; Ji, J.; Huang, Y.; Gao, H.; Li, Z.; Dong, W.; Zhou, S.; Zhu, Y.; Dang, W.; Zhou, T.; et al. Implementing machine learning in bipolar diagnosis in China. Transl. Psychiatry 2019, 9, 305. [Google Scholar] [CrossRef]
Ref. | Year | Area Focused | Mental Health Domain | Algorithms under Review | Limitation |
---|---|---|---|---|---|
[24] | 2016 | Studying depression using imaging and ML approaches | MDD | SVM (linear kernel), SVM (nonlinear kernel), relevance vector regression | No comprehensive comparison of algorithms and did not mention depression screening scales. |
[25] | 2018 | Review of research work on mental health monitoring systems (MHMS) based on multimodal sensing and machine learning | Depression, anxiety, bipolar disorder, migraine, and stress | Supervised, unsupervised, semi-supervised, reinforcement, and transfer learning | No extensive review of the defined domains. No comparative evaluation of models or algorithms was presented. |
[26] | 2018 | ML-based classification and prediction studies of MDD combined with MRI data | MDD and BD | SVM, LDA, GPC, DT, RVM, NN, LR | Depression screening scales used in different studies are not mentioned. Only focuses on MDD and BD-based research studies. |
[27] | 2019 | Reviewed different ML techniques and recommends the working of ML methods in the practical world | PTSD, schizophrenia, depression, ASD, bipolar | SVM, GBM, Random Forest, KNN, Naïve Bayes | The limited number of algorithms utilized as compared to other researches. |
[28] | 2020 | Analysis of Facebook data to detect depression-relevant factors using supervised ML algorithms and linguistic approach | MDD | SVM, CNN, DT, KNN, LR, and RF | Limited attributed of LIWC Software. No scope for semi-supervised learning and DL. The study does not identify the individuals but only assesses the Facebook data. |
[29] | 2020 | Review of EEG, MRI, and Kinesics techniques with related AI applications and algorithms | Psychiatric disorders | DL, Naïve Bayes, LR, DT, SVM | No comprehensive comparison of algorithms and only considered classic shallow learning algorithms. |
[30] | 2021 | Extract depression cues from audio and video for automatic depression estimation | MDD, BD, and Other Mood Disorders (OMD) | DCNN, RNN, LTMS | N/A |
Ref. | Objective | Sample Size | Method/ML Classifier | Model Limitation | Depression Screening Scale | Result |
---|---|---|---|---|---|---|
[45] | Instantaneous mood assessment using voice samples, mobile and social media data | 202 (training), 335 (testing) participant’s data | Moodable Application with SVM, KNN, and RF | Not feasible for larger datasets | PHQ-9 | 76.6% Acc |
[46] | Diagnosis of depression using various psychosocial and socio-demographic factors | 604 Bangladeshi citizens | KNN, AdaBoost, GB, XGBoost, Bagging, Weighted Voting with SelectKBest, mRMR, Boruta feature selection, and SMOTE | No use of any biological marker and only BDC was considered as ground truth for diagnosis | Burns Depression Checklist (BDC) | 92.56% Acc (AdaBoost with SelectKBest) |
[47] | An ML-based predictive model for early depression detection | 6588 Korean citizens (6067 non-depression and 521 depression) | RF with SMOTE, 10-fold cross-validation, AUROC | Biomarkers were not included in the dataset | CES-D-11 | 86.20% Acc |
[51] | Use of linguistic and sentiment analysis with ML to distinguish depressive and non-depressive social media content | 4026 social media posts | RF with RELIEFF feature extractor, LIWC text-analysis tool, and Hierarchical Hidden Markov Model (HMM) and ANEW scale | All depression categories are taken as a single class for classification | Hamilton Depression Rating Scale | Acc% |
90% depressive posts classification | ||||||
92% depression degree classification | ||||||
95% depressive communities’ classification | ||||||
[52] | XGBoost is less expensive computationally than neural network and easy to implement | 11,081 | XGBoost | Need to use more datasets that are accepted by other countries and ethnic groups | - | 90% |
[53] | Classifying GAD and MD subjects based on the incremental value | 14 MD, 19 GAD, 24 healthy | SVM | Need to improve accuracy rate, small sample size, and unbalanced classes | STAI-T, PSWQ, BDI, and IUS-12 | 90.10% Acc |
[54] | Improve the accuracy of MDD diagnosis | 38 MDD, 28 healthy | Multikernel SVM with MST and Kolmogorov–Smirnov test for feature selection | Limited dataset | - | 97.54% Acc |
[50] | To identify symptoms of depression, anxiety, and scale using ML algorithms | 348 participants | DT, RF, Naïve Bayes, SVM, KNN | Imbalanced classes and smaller dataset | DASS 21 | Acc% |
85.50 (NB) | ||||||
79.80 (RF) | ||||||
77.80 (DT) | ||||||
80.30 (SVM) | ||||||
72.10 (KNN) | ||||||
[48] | Predicting depression in university students by identifying related features | 577 Bangladeshi undergraduate students | RF, SVM, KNN | Smaller dataset | BDI-II, DASS-21-BV | Acc% |
75 (RF) | ||||||
73 (SVM) | ||||||
67 (KNN) | ||||||
[49] | Recognition of depression using transformation of EEG features | EEG data of 14 depression patients, 14 normal subjects | Ensemble and DL model with DF and SVM | Limited dataset | Mini International Neuropsychiatric Interview (MINI) | Acc% |
89.02 (Ensemble model) | ||||||
84.75 (DL) |
Ref. | Objective | Sample Size | Method/ML Classifier | Model Limitation | Depression Screening Scale | Result |
---|---|---|---|---|---|---|
[55] | AI-based framework for depression detection with minimum human interaction. | 671 US citizens | AiME with multimodal deep networks with LSTM | The behavioral results of participants conducted at a specific time period may be emotionally influenced by an immediate event and not particularly associated with depression. | PHQ-9 | Acc: 69.23% |
Specificity: 87.77% | ||||||
Sensitivity: 86.81% | ||||||
[56] | An EEG-based DL model for diagnosing unipolar depression. | 30 healthy controls and 33 MDD patients | 1DCNN, 1DCNN with LSTM and 10-fold cross-validation | The process needs a GUI to be used in a clinical environment. Dataset is smaller. The use of anti-depressants, caffeine, and smoking may have negative effects on the classification results of the model. | BDI-II, HADS | 1DCNN: 98.32% Acc |
1DCNN with LSTM: 95.97% Acc | ||||||
[57] | A hybridized a methodology using PSO and ANN to discriminate unipolar and bipolar disorders based on EEG recordings. | 89 subjects (31 bipolar and 58 unipolar) | ANN with PSO for feature selection | Smaller dataset. | DSM-IV, SCID-I, HDRS, YMRS | 89.89% Acc |
[58] | DL based depression detection in imbalanced social media data. | Reddit posts of 9000 users and 107,000 control users | DL model X-A-BiLSTM with XGBoost and Attention-BiLSTM | No use of depression screening scale. | None | Precision: 69% |
Recall: 53% | ||||||
F1: 60% | ||||||
[63] | Diagnosing depression from Twitter data by using an effective DNN architecture and by optimizing word embedding. | 1145 Twitter users | CNN With Max, Multi-Channel CNN, Multi-Channel Pooling CNN, and Bidirectional LSTM with NLTK Tweet tokenizer, Word2Vec word embeddings (Skip-gram, CBOW, Rand) | The word embedding models do not perform efficiently with larger datasets. No use of depression screening scale. | None | Acc% |
CLPsych 2015: 87% | ||||||
Bell Let’s Talk: 83% | ||||||
[59] | Detection of depressed users on social media using hybrid DL model. | 4208 users (2159 depressed and 2049 healthy) | Multimodal Depression Detection with Hierarchical Attention Network (MDHAN) with Latent Dirichlet Allocation (LDA) and bidirectional Gated Recurrent Unit (BiGRU) word encoder | No use of a standard dataset of Twitter users; therefore, the social media data used in the research may be vague and can manipulate the experimental outcome. | DSM-IV | 89.5% Acc |
[60] | Proposed DCNN to boost the depression recognition performance. | 100 training, 100 development, 100 testing | Deep convolutional neural networks (DCNN) with LLD and MRELBP texture descriptor | The experimental results are based only on audio data. | BDI-II | MAE: 8.1901 |
RMSE: 9.8874 | ||||||
[61] | Diagnosis of mild depression by processing EEG signals using CNN. | 24 healthy participants, 24 participants with mild depression | CNN classification model with 24-fold cross-validation and 4 functional connectivity metrics (coherence, correlation, PLV, and PLI) | Only functional connectivity matrices are used in the research, other metrics should be used for evaluation. | BDI-II | 80.74% Acc using Coherence functional connectivity matric |
[62] | Early depression diagnoses by analyzing posts of Reddit users using a DL-based hybrid model. | 401 (for testing) and 486 (for training) with 531,453 posts | BiLSTM with Glove, Word2Vec, and Fastext embedding techniques, Meta-Data features, and LIWC | Imbalanced dataset. The time duration for depression classification is very elongated. | BDI | Word2VecEmbed + Meta feature Set: |
F1 Score: 0.81 | ||||||
Precision: 0.78 | ||||||
Recall: 0.86 |
Ref. | Objective | Sample Size | Method/ML Classifier | Model Limitation | Depression Screening Scale | Result |
---|---|---|---|---|---|---|
[64] | To utilize ML to predict MDD and anxiety disorder in IMID patients. | 637 IMID patients | LR, NN, RF with AUROC with 10-fold cross-validation, Brier scores | Participants in the study have different IMID conditions. No use of PROM instruments and separate testing dataset. | SCID, PROMs | For LR: AUC: 0.90 Brier score: 0.07 |
For NN: AUC: 0.90 Brier score: 0.07 | ||||||
For RF: AUC: 0.91 Brier score: 0.07 | ||||||
[65] | To predict depression among elderly people of China using ML. | 1538 elderly Chinese participants | LR, LR with lasso regularization, RF, GBDT, SVM, and DNN with LSTM | Retrospective waves in the LSTM need to be increased. No use of depression screening scale. | N/A | 0.629 AUC (LR with lasso regularization) |
[66] | MDD diagnosis using an ensemble binary classifier. | NHANES dataset | Ensemble model with DT, AAN, KNN, SVM | No use of rich online social media sources for feature extraction. Dataset range is not defined. | PHQ-9, SF-20 QOLS | 95.4% Acc |
[67] | Development of an algorithm to distinguish between MDD and bipolar disorder (BD) patients based on clinical variables. | 103 MDD and 52 BD patients | LR with Elastic Net and XGBoost | Small and unbalanced sample. Lack of external sample validation. Some misclassifications of classes. Lesser evaluation features. | Brazilian version of TCI, BDI, STAI, PANAS | 78% Acc for LR with Elastic Net Model |
[68] | Evaluating the depression status of Chinese recruits using ML. | 1000 participants | NN, SVM, DT | Need to include complex socio-demographic variables and career variables into the model. | BDI-II | Acc% |
86 (SVM) | ||||||
86 (NN) | ||||||
73 (DT) | ||||||
[69] | Diagnosis of bipolar disorder among Chinese by developing a Bipolar Diagnosis Checklist in Chinese (BDCC) by using ML algorithms. | 255 MDD, 360 BPD, 228 healthy | SVR, RF, LASSO, LR, and LDA | Require large datasets and need to enhance its cross-sectional nature. | N/A | 92% (MDD) 92% (BPD) |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Aleem, S.; Huda, N.u.; Amin, R.; Khalid, S.; Alshamrani, S.S.; Alshehri, A. Machine Learning Algorithms for Depression: Diagnosis, Insights, and Research Directions. Electronics 2022, 11, 1111. https://doi.org/10.3390/electronics11071111
Aleem S, Huda Nu, Amin R, Khalid S, Alshamrani SS, Alshehri A. Machine Learning Algorithms for Depression: Diagnosis, Insights, and Research Directions. Electronics. 2022; 11(7):1111. https://doi.org/10.3390/electronics11071111
Chicago/Turabian StyleAleem, Shumaila, Noor ul Huda, Rashid Amin, Samina Khalid, Sultan S. Alshamrani, and Abdullah Alshehri. 2022. "Machine Learning Algorithms for Depression: Diagnosis, Insights, and Research Directions" Electronics 11, no. 7: 1111. https://doi.org/10.3390/electronics11071111
APA StyleAleem, S., Huda, N. u., Amin, R., Khalid, S., Alshamrani, S. S., & Alshehri, A. (2022). Machine Learning Algorithms for Depression: Diagnosis, Insights, and Research Directions. Electronics, 11(7), 1111. https://doi.org/10.3390/electronics11071111