Next Article in Journal
Contrast Volume-to-Estimated Glomerular Filtration Rate Ratio as a Predictor of Short-Term Outcomes Following Transcatheter Aortic Valve Implantation
Previous Article in Journal
The Clinical Outcome of Cardiac Rehabilitation in Coronary Artery Disease Patients with Regard to the Presence of Left Ventricular Systolic Dysfunction
Previous Article in Special Issue
Cognitive Effects of Reducing First-Generation Antipsychotic Dose Compared to Switching to Ziprasidone in Long-Stay Patients with Schizophrenia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

External Validation of a Machine Learning Model for Schizophrenia Classification

1
Department of Public Health, Fujita Health University School of Medicine, Toyoake 470-1192, Japan
2
Department of Psychiatry, Fujita Health University School of Medicine, Toyoake 470-1192, Japan
3
Department of Public Health and Health Systems, Nagoya University Graduate School of Medicine, Nagoya 466-8550, Japan
4
Department of Public Health, School of Medicine, Kurume University, Kurume 830-0011, Japan
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2024, 13(10), 2970; https://doi.org/10.3390/jcm13102970
Submission received: 1 April 2024 / Revised: 13 May 2024 / Accepted: 15 May 2024 / Published: 17 May 2024
(This article belongs to the Special Issue Clinical Diagnosis and Treatment of Schizophrenia)

Abstract

:
Background and Objective: Excellent generalizability is the precondition for the widespread practical implementation of machine learning models. In our previous study, we developed the schizophrenia classification model (SZ classifier) to identify potential schizophrenia patients in the Japanese population. The SZ classifier has exhibited impressive performance during internal validation. However, ensuring the robustness and generalizability of the SZ classifier requires external validation across independent sample sets. In this study, we aimed to present an external validation of the SZ classifier using outpatient data. Methods: The SZ classifier was trained by using online survey data, which incorporate demographic, health-related, and social comorbidity features. External validation was conducted using an outpatient sample set which is independent from the sample set during the model development phase. The model performance was assessed based on the sensitivity and misclassification rates for schizophrenia, bipolar disorder, and major depression patients. Results: The SZ classifier demonstrated a sensitivity of 0.75 when applied to schizophrenia patients. The misclassification rates were 59% and 55% for bipolar disorder and major depression patients, respectively. Conclusions: The SZ classifier currently encounters challenges in accurately determining the presence or absence of schizophrenia at the individual level. Prior to widespread practical implementation, enhancements are necessary to bolster the accuracy and diminish the misclassification rates. Despite the current limitations of the model, such as poor specificity for certain psychiatric disorders, there is potential for improvement if including multiple types of psychiatric disorders during model development.

1. Introduction

Models trained using machine learning approaches have been applied in various fields of medical research [1,2]. They utilize multiple features to detect or estimate an individual’s existence or likelihood of a medical condition [3]. During the training process, it is essential to evaluate the model’s performance in terms of discriminating individuals who have the outcome from those who do not or to accurately estimate the individual’s risk. This can be tested through internal validation, such as bootstrapping, using the same dataset that was used in the training process. However, before deploying these pretrained models in real-world scenarios, the generalizability should be thoroughly confirmed through external validation. This involves testing the pretrained model using one or more independent sample sets. A model that deserves promotion and can be encapsulated as a tool for commercial use should exhibit excellent generalizability.
From 2021 to 2023, we conducted a research program focused on developing a prevalence estimation method for out-of-hospital schizophrenia and schizophrenia-related disorders in the general population. The program used large-scale epidemiological study data collected from an online survey. It involved information on demographic details, health-related backgrounds, physical and psychiatric comorbidities, and social comorbidities [4]. Based on that information, we developed a machine-learning model for classifying the presence or absence of schizophrenia (SZ classifier), in order to identify potential patients and estimate the prevalence of schizophrenia within the general population. The research design of the program is the first of its kind in Japan.
We envision that if there is a tool that can accurately detect the presence or absence of schizophrenia, it could contribute to the accurate estimation of the prevalence. However, the exact prevalence among the general population is poorly reported. Charlson et al. estimated the global age-standardized point prevalence of schizophrenia in 2016 to be 0.28%, but the actual prevalence could be higher, considering population growth and aging [5]. In Japan, an age–period–cohort analysis reported that the point prevalence was approximately 0.7% [6]. Challenges in the accurate estimation are considered, as (1) some patients may not know that they have the disease. Those who are sick may not always visit the doctor in time; (2) patients with mild disease may not seek medical attention; and (3) some diagnosed cases are for prescriptions to pass a medical insurance review. Hence, the estimated prevalence based solely on medical institution data may not represent the actual incidence in the population. Detecting these “potential” patients in the general population will be the key to estimating the prevalence.
The SZ classifier has provided a promising solution. In the internal validation, it achieved an area under the receiver operating characteristic curve (AUC) score of 0.86 [7]. This indicates that our SZ classifier successfully captures the differences between patients and healthy controls in demographic details, health-related backgrounds, physical and psychiatric comorbidities, and social comorbidities. For example, compared to healthy controls, individuals with schizophrenia demonstrate poor subjective well-being, happiness, and life satisfaction [8,9]. They are also more likely to have chronic diseases, such as poor oral health, noncommunicable diseases, and sleep disorders [8,9]. They also tend to have a lower socioeconomic status [8,9].
The SZ classifier was developed using artificial neural networks. Typically, machine learning validation reports rely on internal validation, employing a subset of the training sample set to assess the model performance. However, this approach carries the risk of inflating the model performance due to overfitting [10]. To accurately evaluate model generalizability, external validation is essential, utilizing an independent sample set [11]. Ideally, the sample set should comprise real-world data, which may provide evidence for medical application [12]. Therefore, in the current study, we aimed to present the external validation results of the SZ classifier. To further assess the external validation, we sampled 150 patients from those who visited the Psychiatric Department of Fujita Health University Hospital. The majority of these patients are from Central Japan, where the hospital is located. We anticipate that the SZ classifier can serve as a tool for detecting “potential” patients within the regional community, thereby allowing for estimating the incidence rate, and the subsequent promotion for broader population-level applications.

2. Materials and Methods

2.1. Development of the Schizophrenia Classification Model (SZ Classifier)

The SZ classifier was developed in our previous study [7] using a sample set collected from an internet research agency’s pooled panel (Rakuten Insight, Inc., Tokyo, Japan) [13]. It comprised 223 schizophrenia patients and 1776 healthy controls. Each sample’s data comprised the presence of schizophrenia, serving as the response variable, along with 76 features. These features included demographic details, health-related backgrounds, physical and psychiatric comorbidities, and social comorbidities. Details of the inclusion criteria have been published elsewhere [9], and the feature definitions are described in Supplementary Materials. The SZ classifier was constructed based on an artificial neural network, considering the significance of capturing nonlinear relationships and higher-order interactions. Specifically, it was structured with five hidden layers, with 128, 64, 32, 16, and 8 neurons in each layer, respectively. The details of the SZ classifier development have been published elsewhere [7] and described in Supplementary Materials. The model performance was examined using the AUC score.
In the internal validation results, 20% of the whole samples were left out as the test set to assess the model performance. The SZ classifier achieved an AUC of 0.86, accuracy of 0.93, sensitivity of 0.56, and specificity of 0.96 [7]. Also, the results of 10-fold cross-validation exhibited that models trained by artificial neural networks could achieve an AUC of around 0.81 [7].

2.2. Samples for External Validation

We targeted patients who visited the Psychiatric Department of Fujita Health University Hospital between January 2022 and May 2023. The patients aged between 20 and 75 years and were diagnosed with schizophrenia, major depressive disorder (MDD), bipolar disorder (BD), or obsessive–compulsive disorder (OCD) were selected. The authors, KS and TK, explained the study’s purpose and methods to the patients. Patients voluntarily decided whether to participate in our study. All the participants provided written informed consent to participate in this study. We confirmed that they had not participated in our research team’s online survey (the aforementioned SZ classifier study).
Participants were asked to complete a survey identical to that used in our previous SZ classifier study. The answers were collected face to face. When necessary, trained assistants helped them complete the survey. These assistants provided only technical help and did not influence the answers. Notably, all participant diagnoses in this study were provided by experienced psychiatrists who had access to comprehensive treatment records, ensuring the sample quality for external validation. In total, we collected 150 samples from 61 schizophrenia patients, 56 MDD patients, 32 BD patients, and 1 OCD patient. Using these samples, we assessed the external validity of the SZ classifier. The primary reason for selecting MDD, BD, and OCD as the control group is that these disorders are theoretically easier to distinguish from schizophrenia, despite often presenting with overlapping clinical features that require careful differential diagnosis. The study design is also considered to possibly simulate the complexities encountered in actual clinical scenarios.
The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Bioethics Review Committee of Fujita Health University (HM21-408).

2.3. Statistical Analysis

Statistical analyses in the current study were performed using Python (version 3.7; Python Software Foundation). The computational code editor used was the Jupyter Notebook 7 (Project Jupyter).

3. Results

Of the 61 schizophrenia patients, 46 were accurately classified, and the sensitivity of the model reached 75%. Among the 89 non-schizophrenia patients, 50 were incorrectly classified as having schizophrenia (Table 1). Of the 56 MDD patients, 31 (55%) were incorrectly classified as having schizophrenia, and of the 32 BD patients, 19 (59%) were incorrectly classified as having schizophrenia (Table 2).

4. Discussion

This study conducted an external validation of our predeveloped SZ classifier using outpatient data. The SZ classifier achieved a sensitivity of 0.75 (46 out of 61 patients) when applied to individuals diagnosed with schizophrenia. However, when applied to patients diagnosed with BD or MDD, the misclassification rates were 59% (19 out of 32) and 55% (31 out of 56), respectively.
The sensitivity achieved in the external validation exceeded that of the internal test (0.56) [7]. This disparity may stem from the fact that the external validation sample set relies on diagnoses by clinicians and regular medical visits, which are deemed more reliable than self-reported diagnoses obtained through online surveys. Hohmann emphasized that the dependency of the output is reliant on the accuracy of the big data and the sample size [14]. The model performance is limited by poor data quality and insufficient confounding variables [14]. Moreover, the sample set used for developing the SZ classifier was imbalanced. We adopted a well-defined algorithm—Synthetic Minority Over-sampling Technique (SMOTE)—to solve this problem [15]. Nevertheless, SMOTE does not always work perfectly when solving such a problem. Hence, it is reasonable to speculate that if the data used for developing the model were of higher quality, the performance of the SZ classifier would likely improve.
The SZ classifier demonstrated a misclassification rate of >50% when applied to MDD and BD patients. This could be attributed to the similarities between schizophrenia, MDD, and BD. The top five features, ranked according to their importance, of the SZ classifier were the frequency of sleep medication use, age, household income, type of employment, and bedtime [7]. Table S1 lists the characteristics of the external validation sample set based on these five features. The distributions between schizophrenia and MDD, as well as between schizophrenia and BD, were similar (see Supplementary Materials, Table S1). Clinically, MDD, BD, and OCD often present with symptoms similar to those of schizophrenia; thus, a careful differential diagnosis is required. Since our SZ classifier was not built using samples of MDD, BD, and OCD, the poor specificity observed in the external validation results was anticipated.
Although the SZ classifier achieved a respectable sensitivity of 0.75, the misclassification rate exceeding 50% indicates that the model still requires significant improvements before it can be widely promoted across various communities or populations. We found no prior studies discussing external validation for machine learning models in schizophrenia classification. However, some relevant research based on machine learning techniques could be referenced and utilized. For instance, Bae used content from social media and compared four machine learning methods using natural language processing techniques for classifying schizophrenia [16]. The Random Forest model obtained the best AUC of 0.97, while Bae’s study did not involve artificial neural networks. Bracher-Smith trained schizophrenia classification models using genetic and demographic factors from the UK Biobank, where the neural network model achieved an AUC of only 0.67 [17], which is lower than the 0.86 AUC from our study (internal validation). Shanarova designed a machine learning-based diagnostic for schizophrenia using event-related potential data, achieving a sensitivity and specificity of 91% and 90.8%, respectively [18]. Overall, the prospects for machine learning in schizophrenia research are promising, primarily because it can analyze speech, behavior, image, and people’s creativity [19]. It has been shown that various machine learning-based models can assist specialists in predicting and diagnosing schizophrenia using medical history, genetic data, and even epigenetic information [19].
The primary advantage of using machine learning to construct a classification model is its ability to extract valuable information from extensive datasets, particularly those related to feature engineering, without explicit instructions [20]. For example, machine learning techniques have great potential for quantitatively distinguishing the differences among similar feature expressions. We have previously reported that schizophrenia patients have an increased likelihood of developing diabetes, cardiovascular disease, worsening of metabolic syndrome, depression, and sleep disorders [8]. Similar findings were reported for MDD [21]. Additionally, schizophrenia patients are observed to exhibit elevated rates of smoking, drinking, and drug consumption [8], and a similar pattern is also observed in MDD and BD patients [21,22]. Schizophrenia patients typically exhibit lower health literacy [8], while older BD patients experience a faster decline in health perception [23]. However, previous studies have rarely addressed the distinctions among these similar feature expressions for those diseases. Given that our survey questionnaire delved into the aforementioned issues (namely, the information was collected quantitatively), we have reason to believe that introducing samples of MDD, BD, and OCD during the model development could improve the SZ classifier’s performance.
This study had several limitations. First, the sample set used for the external validation was collected only from Central Japan and may not be representative of the broader population. Additionally, the sample size was limited. Future studies should include more patients from diverse regions, as well as healthy controls, to enhance the reliability. Second, biases may arise from the online study design. Some individuals may not use the internet frequently, and patients with severe illnesses may struggle to complete the questionnaire owing to difficulties in concentration. However, in light of the changes in people’s habits following the Coronavirus pandemic in recent years, online methods such as apps have become increasingly applicable for conducting epidemiological research [24]. Third, relying on nonpsychiatric symptoms to detect the presence of mental disorders may seem arbitrary, despite yielding favorable results. Therefore, the findings of such studies should be interpreted and promoted with caution.

5. Conclusions

The SZ classifier currently faces challenges in accurately determining the presence or absence of schizophrenia at the individual level. The performance of the SZ classifier would likely improve if the data used during model development were of higher quality. Prior to widespread practical implementation, enhancements are necessary to bolster the accuracy and diminish misclassification rates. Considering the results of external validation and the inherent characteristics of machine learning techniques, there exists significant potential to improve the performance of the SZ classifier if samples of multiple diseases are included during the model development phase.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/jcm13102970/s1. References [25,26,27,28,29,30,31,32,33,34,35,36] are cited in the supplementary materials.

Author Contributions

Conceptualization: Y.H. and A.O.; methodology: Y.H.; formal analysis: Y.H., A.O. and M.M.; investigation: K.S., T.K., A.O. and M.M.; resources: A.O.; data curation: A.O. and M.M.; writing—original draft preparation: Y.H.; writing—review and editing: K.S., T.K., Y.L., M.M., S.T., N.I. and A.O.; project administration: A.O.; funding acquisition: A.O. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Ministry of Health, Labour and Welfare, Japan (grant number JPMH21GC1018 to A.O.).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Bioethics Review Committee of Fujita Health University (HM21-408, 24 January 2022).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available from the corresponding author upon reasonable request and after appropriate procedures, including approval from the institutional ethics review committee.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. National Cancer Center Institute for Cancer Control. Development and Validation of a 5-Year Predictive Model for Type 2 Diabetes Risk. Available online: https://epi.ncc.go.jp/jphc/outcome/9198.html (accessed on 29 February 2024). (In Japanese).
  2. Gray, E.; Donten, A.; Payne, K.; Hall, P.S. Survival estimates stratified by the Nottingham Prognostic Index for early breast cancer: A systematic review and meta-analysis of observational studies. Syst. Rev. 2018, 7, 142. [Google Scholar] [CrossRef]
  3. Snell, K.I.E.; Archer, L.; Ensor, J.; Bonnett, L.J.; Debray, T.P.A.; Phillips, B.; Collins, G.S.; Riley, R.D. External validation of clinical prediction models: Simulation-based sample size calculations were more reliable than rules-of-thumb. J. Clin. Epidemiol. 2021, 135, 79–89. [Google Scholar] [CrossRef]
  4. Development of a Prevalence Estimation Method for Out-of-Hospital Schizophrenia and Schizophrenia-Related Disorders in the General Population Using Large-Scale Epidemiological Study Data and Medical Reimbursement Statement Data. Ministry of Health, Labour and Welfare Grants System. Available online: https://mhlw-grants.niph.go.jp/project/158063 (accessed on 29 February 2024). (In Japanese).
  5. Charlson, F.J.; Ferrari, A.J.; Santomauro, D.F.; Diminic, S.; Stockings, E.; Scott, J.G.; McGrath, J.J.; Whiteford, H.A. Global Epidemiology and Burden of Schizophrenia: Findings from the Global Burden of Disease Study 2016. Schizophr. Bull. 2018, 44, 1195–1203. [Google Scholar] [CrossRef]
  6. Okui, T. An age-period-cohort analysis for prevalence of common psychiatric disorders in Japan, 1999–2017. Soc. Psychiatry Psychiatr. Epidemiol. 2021, 56, 639–648. [Google Scholar] [CrossRef]
  7. He, Y.; Matsunaga, M.; Li, Y.; Kishi, T.; Tanihara, S.; Iwata, N.; Tabuchi, T.; Ota, A. Classifying Schizophrenia Cases by Artificial Neural Network Using Japanese Web-Based Survey Data: Case-Control Study. JMIR Form. Res. 2023, 7, e50193. [Google Scholar] [CrossRef] [PubMed]
  8. He, Y.; Tanaka, A.; Kishi, T.; Li, Y.; Matsunaga, M.; Tanihara, S.; Iwata, N.; Ota, A. Recent findings on subjective well-being and physical, psychiatric, and social comorbidities in individuals with schizophrenia: A literature review. Neuropsychopharmacol. Rep. 2022, 42, 430–436. [Google Scholar] [CrossRef]
  9. Matsunaga, M.; Li, Y.; He, Y.; Kishi, T.; Tanihara, S.; Iwata, N.; Tabuchi, T.; Ota, A. Physical, Psychiatric, and Social Comorbidities of Individuals with Schizophrenia Living in the Community in Japan. Int. J. Environ. Res. Public Health 2023, 20, 4336. [Google Scholar] [CrossRef] [PubMed]
  10. Park, S.H.; Han, K. Methodologic Guide for Evaluating Clinical Performance and Effect of Artificial Intelligence Technology for Medical Diagnosis and Prediction. Radiology 2018, 286, 800–809. [Google Scholar] [CrossRef] [PubMed]
  11. England, J.R.; Cheng, P.M. Artificial Intelligence for Medical Image Analysis: A Guide for Authors and Reviewers. AJR Am. J. Roentgenol. 2019, 212, 513–519. [Google Scholar] [CrossRef]
  12. Thrall, J.H.; Fessell, D.; Pandharipande, P.V. Rethinking the Approach to Artificial Intelligence for Medical Image Analysis: The Case for Precision Diagnosis. J. Am. Coll. Radiol. 2021, 18, 174–179. [Google Scholar] [CrossRef]
  13. Rakute Disease Panel. Available online: https://insight.rakuten.com/wordpress/wp-content/uploads/Rakuten_Insight_Panel_Book.pdf (accessed on 29 February 2024).
  14. Hohmann, E. Editorial Commentary: Big Data and Machine Learning in Medicine. Arthroscopy 2022, 38, 848–849. [Google Scholar] [CrossRef] [PubMed]
  15. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  16. Bae, Y.J.; Shim, M.; Lee, W.H. Schizophrenia Detection Using Machine Learning Approach from Social Media Content. Sensors 2021, 21, 5924. [Google Scholar] [CrossRef] [PubMed]
  17. Bracher-Smith, M.; Rees, E.; Menzies, G.; Walters, J.T.R.; O’Donovan, M.C.; Owen, M.J.; Kirov, G.; Escott-Price, V. Machine learning for prediction of schizophrenia using genetic and demographic factors in the UK biobank. Schizophr. Res. 2022, 246, 156–164. [Google Scholar] [CrossRef] [PubMed]
  18. Shanarova, N.; Pronina, M.; Lipkovich, M.; Ponomarev, V.; Müller, A.; Kropotov, J. Application of Machine Learning to Diagnostics of Schizophrenia Patients Based on Event-Related Potentials. Diagnostics 2023, 13, 509. [Google Scholar] [CrossRef] [PubMed]
  19. Gashkarimov, V.R.; Sultanova, R.I.; Efremov, I.S.; Asadullin, A.R. Machine learning techniques in diagnostics and prediction of the clinical features of schizophrenia: A narrative review. Consort. Psychiatr. 2023, 4, 43–53. [Google Scholar] [CrossRef] [PubMed]
  20. Ahmed, S.F.; Alam, M.S.B.; Hassan, M.; Rozbu, M.R.; Ishtiak, T.; Rafa, N.; Mofijur, M.; Ali, A.B.M.S.; Gandomi, A.H. Deep learning modelling techniques: Current progress, applications, advantages, and challenges. Artif. Intell. Rev. 2023, 56, 13521–13617. [Google Scholar] [CrossRef]
  21. Arnaud, A.M.; Brister, T.S.; Duckworth, K.; Foxworth, P.; Fulwider, T.; Suthoff, E.D.; Werneburg, B.; Aleksanderek, I.; Reinhart, M.L. Impact of Major Depressive Disorder on Comorbidities: A Systematic Literature Review. J. Clin. Psychiatry 2022, 83, 43390. [Google Scholar] [CrossRef] [PubMed]
  22. Lipski, H.; Baylé, F. Bipolar disorders and psychiatric comorbidity. Rev. Prat. 2005, 55, 501–506. (In French) [Google Scholar]
  23. Beunders, A.J.M.; Kok, A.A.L.; Kosmas, P.C.; Beekman, A.T.F.; Sonnenberg, C.M.; Schouws, S.N.T.M.; Kupka, R.W.; Stek, M.L.; Dols, A. Physical comorbidity in Older-Age Bipolar Disorder (OABD) compared to the general population—A 3-year longitudinal prospective cohort study. J. Affect. Disord. 2021, 288, 83–91. [Google Scholar] [CrossRef]
  24. Torous, J.; Keshavan, M. COVID-19, mobile health and serious mental illness. Schizophr. Res. 2020, 218, 36–37. [Google Scholar] [CrossRef] [PubMed]
  25. Kusama, T.; Kiuchi, S.; Takeuchi, K.; Ikeda, T.; Nakazawa, N.; Kinugawa, A.; Osaka, K.; Tabuchi, T. Information usage and compliance with preventive behaviors for COVID-19: A longitudinal study with data from the JACSIS 2020/JASTIS 2021. Healthcare 2022, 10, 521. [Google Scholar] [CrossRef] [PubMed]
  26. Wakabayashi, M.; Sugiyama, Y.; Takada, M.; Kinjo, A.; Iso, H.; Tabuchi, T. Loneliness and increased hazardous alcohol use: Data from a nationwide Internet survey with 1-year follow-up. Int. J. Environ. Res. Public Health 2022, 19, 12086. [Google Scholar] [CrossRef]
  27. Sasaki, R.; Ota, A.; Yatsuya, H.; Tabuchi, T. Gender difference in fear and anxiety about and perceived susceptibility to COVID-19 in the third wave of pandemic among the Japanese general population: A nationwide web-based cross-sectional survey. Int. J. Environ. Res. Public Health 2022, 19, 16239. [Google Scholar] [CrossRef]
  28. The Ministry of Health, Labour and Welfare. The Standards for the Degree of Independent Living for Elderly and Disabled People. 1991. Available online: https://www.mhlw.go.jp/file/06-Seisakujouhou-12300000-Roukenkyoku/0000077382.pdf (accessed on 17 October 2023). (In Japanese).
  29. DeSalvo, K.B.; Bloser, N.; Reynolds, K.; He, J.; Muntner, P. Mortality prediction with a single general self-rated health question. J. Gen. Intern. Med. 2006, 21, 267–275. [Google Scholar] [CrossRef] [PubMed]
  30. Radloff, L.S. The CES-D scale. Appl. Psychol. Meas. 1977, 1, 385–401. [Google Scholar] [CrossRef]
  31. Kohout, F.J.; Berkman, L.F.; Evans, D.A.; Cornoni-Huntley, J. Two shorter forms of the CES-D Depression Symptoms index. J. Aging Health 1993, 5, 179–193. [Google Scholar] [CrossRef] [PubMed]
  32. Warttig, S.L.; Forshaw, M.; South, J.; White, A. New, normative, English-sample data for the Short Form Perceived Stress Scale (PSS-4). J. Health Psychol. 2013, 18, 1617–1628. [Google Scholar] [CrossRef]
  33. The ENRICHD Investigators. Enhancing recovery in coronary heart disease (ENRICHD) study intervention: Rationale and design. Psychosom. Med. 2001, 63, 747–755. [Google Scholar]
  34. Enhancing Recovery in Coronary Heart Disease Patients (ENRICHD): Study design and methods. Am. Heart J. 2000, 139, 1–9. [CrossRef]
  35. Grootaert, G.; Narayan, D.; Jones, V.N.; Woolcock, M. Measuring Social Capital: An Integrated Questionnaire; World Bank Publications: Washington, DC, USA, 2004. [Google Scholar]
  36. Murayama, H.; Fujiwara, Y.; Kawachi, I. Social Capital and Health: A review of Prospective Multilevel studies. J. Epidemiol. 2012, 22, 179–187. [Google Scholar] [CrossRef] [PubMed]
Table 1. Confusion matrix of external validation for the SZ classifier: results of all cases.
Table 1. Confusion matrix of external validation for the SZ classifier: results of all cases.
Observed Cases
SchizophreniaMajor Depressive Disorder, Bipolar, and Obsessive-Compulsive Disorder
Predicted casesSchizophrenia4650
Major depressive disorder, bipolar, and obsessive-compulsive disorder1539
Total6189
Sensitivity0.75
Specificity0.44
Table 2. Confusion matrix of external validation for the SZ classifier: results of the non-schizophrenia cases.
Table 2. Confusion matrix of external validation for the SZ classifier: results of the non-schizophrenia cases.
Observed Cases
Major Depressive DisorderBipolarObsessive-Compulsive Disorder
Predicted casesSchizophrenia31190
Non-schizophrenia25131
Total56321
Misclassification rate0.550.590
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, Y.; Sakuma, K.; Kishi, T.; Li, Y.; Matsunaga, M.; Tanihara, S.; Iwata, N.; Ota, A. External Validation of a Machine Learning Model for Schizophrenia Classification. J. Clin. Med. 2024, 13, 2970. https://doi.org/10.3390/jcm13102970

AMA Style

He Y, Sakuma K, Kishi T, Li Y, Matsunaga M, Tanihara S, Iwata N, Ota A. External Validation of a Machine Learning Model for Schizophrenia Classification. Journal of Clinical Medicine. 2024; 13(10):2970. https://doi.org/10.3390/jcm13102970

Chicago/Turabian Style

He, Yupeng, Kenji Sakuma, Taro Kishi, Yuanying Li, Masaaki Matsunaga, Shinichi Tanihara, Nakao Iwata, and Atsuhiko Ota. 2024. "External Validation of a Machine Learning Model for Schizophrenia Classification" Journal of Clinical Medicine 13, no. 10: 2970. https://doi.org/10.3390/jcm13102970

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop