Next Article in Journal
Association of Inflammatory Metabolic Activity of Psoas Muscle and Acute Myocardial Infarction: A Preliminary Observational Study with 18F-FDG PET/CT
Previous Article in Journal
Amyotrophic Lateral Sclerosis and Frontotemporal Lobar Degenerations: Similarities in Genetic Background
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Explainable Machine Learning Model for Glaucoma Diagnosis and Its Interpretation

1
Software Science, College of Software Convergence, Jukjeon Campus, Dankook University, Yongin 16890, Korea
2
Department of Ophthalmology, College of Medicine, Dankook University, 119, Dandae-ro, Dongnam-gu, Cheonan-si, Chungnam 31116, Korea
3
Department of Ophthalmology, Institute of Health Sciences, Gyeongsang National University School of Medicine and Gyeongsang National University Hospital, Jinju 52727, Korea
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(3), 510; https://doi.org/10.3390/diagnostics11030510
Submission received: 20 February 2021 / Revised: 8 March 2021 / Accepted: 10 March 2021 / Published: 13 March 2021
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
The aim is to develop a machine learning prediction model for the diagnosis of glaucoma and an explanation system for a specific prediction. Clinical data of the patients based on a visual field test, a retinal nerve fiber layer optical coherence tomography (RNFL OCT) test, a general examination including an intraocular pressure (IOP) measurement, and fundus photography were provided for the feature selection process. Five selected features (variables) were used to develop a machine learning prediction model. The support vector machine, C5.0, random forest, and XGboost algorithms were tested for the prediction model. The performance of the prediction models was tested with 10-fold cross-validation. Statistical charts, such as gauge, radar, and Shapley Additive Explanations (SHAP), were used to explain the prediction case. All four models achieved similarly high diagnostic performance, with accuracy values ranging from 0.903 to 0.947. The XGboost model is the best model with an accuracy of 0.947, sensitivity of 0.941, specificity of 0.950, and AUC of 0.945. Three statistical charts were established to explain the prediction based on the characteristics of the XGboost model. Higher diagnostic performance was achieved with the XGboost model. These three statistical charts can help us understand why the machine learning model produces a specific prediction result. This may be the first attempt to apply “explainable artificial intelligence” to eye disease diagnosis.

1. Introduction

Glaucoma is the leading cause of irreversible blindness worldwide and affects the optic nerve progressively [1]. It is diagnosed currently via four examinations: (1) detection of elevated intraocular pressure (IOP), (2) assessment of damage to the optic disc by calculating the cup-to-disc ratio (CDR), (3) identifying decreased retinal nerve fiber layer (RNFL) thickness, and (4) detection of characteristic visual field defects. Structural and functional clinical modalities such as optical coherence tomography (OCT) and visual field (VF) test provide indicators or values that can be used to diagnose glaucoma [2]. However, these conventional methods of glaucoma detection have higher chances of being misdiagnosed [2]. Hence, artificial intelligence (AI) systems are necessary to prevent misdiagnosis [3]. Machine learning has recently become one of the core technologies in the fields of science and technology, including life science and medicine. Classification is a major technology used in medical applications because it can be applied to prediction (diagnosis). The machine learning approach is also useful for predicting glaucoma. In the near future, machine learning will be an essential tool for predicting and treating glaucoma.
Despite the success of machine learning, machine learning has a fatal drawback for clinical applications. Most machine learning predictors do not explain the grounds of their individual prediction. As a result, they are known as a “black box.” In medical applications, this is a serious disadvantage. Medical personnel cannot trust a machine learning diagnosis without a reasonable explanation. We may trust a predictor based on a deep neural network model because it can show a high prediction accuracy against large test data. However, trusting individual prediction results is a different issue because the prediction may be incorrect. Even if the prediction is correct, it becomes difficult to trust without a reasonable basis.
The Defense Advanced Research Projects Agency has recently initiated a new project called Explainable Artificial Intelligence, which shows that the explainability of AI has become a hot issue. Researchers have developed a large number of prediction models (predictors) based on machine learning. Most of them are mainly based on “deep learning” and use image data for prediction [4,5,6,7,8]. Traditional machine learning models have also been applied for glaucoma prediction [9,10,11]. Some papers suggested segmentation of areas with abnormalities on optical images [12,13,14,15]. However, interpretations of “individual prediction” for glaucoma diagnosis remain unexplored. Recently, interpretable machine learning has been gaining attention to explain “prediction models” and “individual prediction” [16,17].
The partial dependence plot (PDP) shows the marginal effect that one or two features (variables) have on the predicted outcome of a machine learning model [16]. A PDP can show whether the relationship between the target and a feature (variable) is linear, monotonic, or more complex. When features interact with each other in a prediction model, the prediction cannot be expressed as the sum of the feature effects because the effect of one feature depends on the value of the other feature [16]. H-statistic, variable interaction networks (VINs), and partial dependence-based feature interaction have been suggested to measure the feature interactions [18]. Feature importance in a prediction model can be measured in various ways. Local interpretable model-agnostic explanations (LIME) [19], Shapley Values [20], and SHapley Additive ExPlanations (SHAP) [21] have been suggested to explain individual predictions. Microsoft researchers published a unified framework for machine-learning interpretability [21]. They integrated previous research outcomes related to machine-learning interpretability into an open-source library called ‘interpretML.’
In this paper, we propose a machine learning model for glaucoma prediction with explanation functions for individual prediction. We mainly focus on explaining individual predictions. For this purpose, we attempt to build an “understandable prediction model” instead of a “highly accurate model”, based on the XGboost algorithm [22,23]. Five features from a visual field (VF) examination, RNFL OCT, and IOP test are used for glaucoma prediction. To explain individual predictions, gauge, radar, and SHAP charts are used. An explanation of glaucoma prediction provides a basis for the ophthalmologist to determine whether to trust the predicted results. Furthermore, ophthalmologists may obtain clinical insight from the explanation. Details are described in the Discussion section.

2. Materials and Methods

2.1. Participants

We collected the medical records of patients who underwent RNFL OCT and VF examinations at Gyeongsang National University Hospital between January 2012 and November 2020. To conduct the study, all patients underwent comprehensive ophthalmological examinations, which included slit-lamp bio-microscopy, best corrected visual acuity (BCVA), autorefraction (KR8800, Topcon, Tokyo, Japan), central corneal thickness (CCT) measurement (Pentacam, Oculus GmbH, Wetzlar, Germany), Goldmann applanation tonometry (Haag-Streit AG, Bern, Switzerland), a dilated fundus examination, and fundus and red-free fundus photography (Canon, Tokyo, Japan). An automated VF test was conducted using the 30–2 program Swedish interactive threshold algorithm standard on a Humphrey 740 visual field analyzer (Carl Zeiss Meditec Inc., Dublin, CA). Spectral-domain OCT (SD-OCT) images, obtained using the Spectralis® (Heidelberg Engineering GmbH, Heidelberg, Germany) platform were used to measure the peripapillary RNFL (pRNFL) thickness.
In total, 975 eyes (of 430 patients) with glaucoma (POAG (primary open angle glaucoma) or NTG) and 649 eyes (of 377 patients) without glaucoma were included. The inclusion criteria for normal eyes were a BCVA of 20/40, normal anterior segment on a slit-lamp examination, no RNFL defects in red-free fundus photographs, no visual field defects, and an intraocular pressure of ≤21 mmHg. The inclusion criteria for glaucomatous eyes were as follows: BCVA of 20/40 or better, a normal anterior segment on a slit-lamp examination, and diagnosis of glaucoma by three glaucoma specialists. The glaucoma diagnosis was based on characteristic glaucomatous structural changes to the optic disc accompanied by glaucomatous visual field defects. The criteria for a glaucomatous visual field defect were as follows: glaucoma hemifield test [24] outside the normal limit, pattern standard deviation with a p value of <5%, or a cluster of <3 points in the pattern deviation plot in a single hemifield (superior or inferior) with a p value of <5%, one of which must have a p value of <1%. Any of the preceding criteria, if repeatable, were considered sufficient evidence of a glaucomatous visual field defect.
In addition to those not meeting the inclusion criteria, the exclusion criteria were as follows: history of ocular inflammation or trauma, and the presence of concurrent retinal disease (i.e., vascular disorder or macular degeneration), optic nerve disease other than glaucoma, or a brain disorder that could influence the visual field results.
The inclusion criteria for normal eyes were a BCVA of 20/40, normal anterior segment on a slit-lamp examination, no RNFL defects in red-free fundus photographs, no visual field defects, and an intraocular pressure of ≤21 mmHg. Table 1 summarizes the characteristics of the participants.

2.2. Prediction Model

Figure 1 describes the prediction model development procedure. First, we collected clinical data of the patients based on a visual field test, an RNFL OCT test, general examination including an IOP test, and a Fundus image test. The first three are numerical data. In the case of fundus images, we build a prediction model based on a convolutional neural network (CNN), and the supported degree of glaucoma for each observation form feature data. As a result, our base dataset contains 22 features and a class label. To build an understandable model, we tried to choose a small number of features as quickly as possible. We tested 22 features, as shown in Table 2.
First, we selected 10 informative features from the whole range of features using a chi-square feature selection measure. Then, a combination test was conducted to find a set of features that produces the best accuracy. Finally, we obtained five features that are described in Table 3.
Figure 2 shows a box plot of the five features. All features show a large difference in the median value between glaucoma and healthy controls. In particular, there is a big difference between Healthy control and glaucoma in pattern standard deviation (PSD).
To build a prediction model for glaucoma, we tested various classification algorithms such as a support vector machine (SVM), random forest, C5.0, and XGboost. We use R packages to test the classification algorithms. To evaluate and compare the prediction models, we chose four criteria: receiver operating characteristic (ROC) plot, area under the curve (AUC), sensitivity, and specificity. The sensitivity of a prediction model refers to the ability of the model to correctly identify patients with a disease. The specificity of a prediction model refers to the ability of the test to correctly identify patients without a disease. The ROC plot expresses the relationship between sensitivity and 1 –specificity. The closer the ROC curve is to the upper-left hand corner, the better the model. AUC expresses the area under the ROC curve and can have any value between zero and 1. It is a widely used indicator of the goodness of a binary classification model.
From the comparison results, we confirm that XGboost shows the best performance with the hyperparameters described in Table 4. Therefore, we chose the XGboost prediction model for a glaucoma diagnosis. After model selection, we build a model explanation system, more exactly an explanation system for individual prediction results. The details are described in the next section.

2.3. Explanations of an Individual Prediction

If we input the values of the five features in Table 2 into the prediction model, either “glaucoma” or “healthy” is the output. To explain why the model produces the decision, we suggest three graphical charts, i.e., gauge, radar, and SHAP charts. The gauge and radar charts show the position of the input values among the complete distributions of the values. The SHAP chart shows the roles of each value in the decision.
Figure 3 (Left) shows an example of a gauge chart. In the dial, 0 and 170 indicate the minimum and maximum values of RNFL_S in a distribution of training data. In the dial, 64 and 155 are the boundary values of the overlapped range between glaucoma and healthy individuals in a distribution of the training data. In Figure 3 (Left), the value of 86 at the center of the bottom line indicates the value of RNFL_S for a patient, and determines the angle of the needle on the dial. Red and green zones in a dial indicate the range of glaucoma and healthy individuals, respectively. The yellow zone indicates the overlapped range of glaucoma and healthy individuals.
A radar chart is a visualization method of multivariate data in the form of a two-dimensional chart of three or more quantitative variables represented on axes starting from the same point, as shown in Figure 4. We basically use a radar chart to express the distribution of five feature values for a patient who we want to predict. The second purpose of a radar chart is to observe the pattern of a polygon inside the chart. We can expect that the patterns of glaucoma and healthy patients will be different. Figure 4 is extracted from the average feature values of glaucoma and healthy patients. As we can see, glaucoma patients have higher PSD/IOP feature values and lower RNFL_ S/RNFL _ I/RNFL _T feature values than healthy patients. This is helpful for clinicians in making a glaucoma diagnosis and understand the prediction result of the proposed predictor.
An SHAP chart [21] is based on Shapley values [20]. In the Shapley values theory, a prediction can be explained by assuming that each feature value of the instance is a “player” in a game, where the prediction is the payout. Shapley values tell us how to fairly distribute the “payout” among the features (players) [16]. Figure 5 shows an example of a SHAP chart. In a SHAP cart, the Y-axis represents features and their values for the target patient. The X-axis in the SHAP chart represents the degree of support for glaucoma. Positive values (red bars) support a glaucoma diagnosis whereas negative values (green bars) support a healthy diagnosis. The longer the bar is, the stronger the support. The title of a SHAP chart represents the result of the prediction and its degree of certainty. In the case of Figure 5, the prediction result is “Glaucoma,” and its degree of certainty is 0.96 (96%).
In Figure 5, the PSD value of 10.51 strongly supports a glaucoma diagnosis, as does a RNFL_T value of 47. An RNFL_I value of 119 supports the diagnosis of a healthy individual. The RNFL_S and IOP values weakly support a glaucoma diagnosis. These indicators show why the model predicts the target patient as suffering from glaucoma.

3. Results

We divided 1624 cases into 80% training set and 20% test set. A total of 1306 cases are used for developing the prediction models, and 318 cases are used for evaluating the models. Four well-known classification algorithms are tested to build the prediction models, and xgboost shows the best accuracy. Table 5 describes a performance comparison of the four classification models. In the AUC, the proposed xgboost model has a 0.945 rate. The sensitivity of the proposed model is 0.941. This means that the proposed model accurately predicts glaucoma patients with an accuracy of 0.947. It also means that it shows a small false negative ratio. The specificity of the proposed model was 0.950, which shows a predictive power against the healthy controls.
In the machine learning prediction model, no features provide an equal contribution towards a prediction. Figure 6 describes the importance of the features in the proposed model. Here, importance refers to the contribution of each feature to the prediction task. In the proposed prediction model, RNFL_S, RNFL_I, and PSD have a stronger influence than RNFL_T and IOP.
In the prediction task, the roles of each feature are not independent, and they cooperate with each other. In other words, they contribute to the prediction task through an interaction. Figure 7 shows a feature interaction chart [25] for the five features of the proposed model. In the interaction table, the (i-th, j-th) cell shows the feature interaction between features Fi and Fj. The interaction value between IOP and RNFL_I was 0.0054. This indicates that their interaction increases the prediction accuracy by 0.0054. In other words, the combination of IOP and RNFL_I provides synergy to predict glaucoma. By contrast, IOP negatively interacts with RNFL_S and RNFL_T, which decrease the prediction accuracy by −0.0028 and −0.0003, respectively.
In this paper, we introduced a glaucoma prediction model and described its application. The use of our three explanation charts for a specific prediction are discussed in the next section.

4. Discussion

4.1. Diagnosis of Glaucoma

Glaucoma is an ocular disease that causes damage to the optic nerve because of the elevation of IOP and eventually a progressive visual field loss [26]. When ophthalmologists diagnose glaucoma, they measure the patient’s IOP and check the CDR through fundus examination. In addition, cpRNFL and/or mGCIPL thickness is measured using OCT, a standard modality for evaluating glaucomatous structural damage to the optic disc, and if there are some regions with reduced RNFL thickness on OCT, glaucoma is diagnosed by checking whether visual field abnormalities appear in the corresponding region in the visual field test, and by monitoring visual function in glaucoma [26]. Most of the glaucoma and AI-related articles searched are subject to screening and diagnosis of glaucoma [27]. Among them, there are many papers that diagnose glaucoma by detecting increased CDR or loss of neuro-retinal rim in the optic disc using fundus photography [28]. In addition, studies to diagnose glaucoma by extracting data from the OCT and VF tests have also been reported [29]. In recent times, machine learning models that use two or more of these diagnostic modalities in combination have been reported. Yohei et al. suggested that visual field abnormalities within the central 10° were predicted using SD-OCT in glaucoma patients, and Christopher et al. developed a deep learning system to identify glaucomatous visual field defects and predict the severity of VF defects using SD OCT [30,31].
Our study differs from previous reports because it maximizes the number of clinical values used to diagnose glaucoma. These include demographic and ocular factors that can potentially influence the diagnosis of glaucoma, such as age, IOP, spherical equivalent, axial length, and central corneal thickness. In addition, four values were obtained from the quadrant cpRNFL thickness map measured by OCT, and indices such as MD, PSD, VFI, and GHT were extracted in the VF test. Ultimately, this machine learning model was operated in the same situation that clinicians treat glaucoma patients, and it has improved to a level that ophthalmologists can refer to when they examine patients.
For clinicians to use it when actually examining and treating patients, the machine learning prediction model must have high reliability, that is, be highly accurate. In addition, the prediction model should provide users with clear information based on the results obtained. However, many machine learning models currently do not provide a clear rationale, and this is a barrier, i.e., that machine learning skills cannot be used in clinical settings. In fact, no evidence has been found in previous reports that have studied glaucoma and AI. To solve the “black box,” which is the fatal shortcoming of other machine learning models, our approach presented the basis of judgment with gauge, radar, and SHAP charts. By presenting the evidence of this judgment as glaucoma or healthy, our approach was able to move a little closer towards “explainable AI for glaucoma.”

4.2. Case Analysis of Prediction Results

We examined the results of our prediction model in several cases. Table 6 summarizes these cases. Details of the test results are described in the supplementary material. Cases 1 (Supplemental Figure S1) and 2 (Supplemental Figure S2) are typical healthy and glaucoma patients, respectively, and our model accurately predicted them. All three charts consistently indicated healthy and glaucoma, respectively. On the other hand, in case 3 (Supplemental Figure S3), CDR was increased, and RNFL defects were observed in the supero-temporal area. Therefore, we made a clinical diagnosis of pre-perimetric glaucoma, but the prediction model predicted it to be healthy. In case 4 (Supplemental Figure S4), the CDR was healthy and the RNFL defect was not present. In a patient diagnosed as healthy, our prediction model predicted glaucoma. By reviewing the three charts, it can be observed that the decrease in the superior cpRNFL thickness has an effect on the diagnosis of glaucoma. This patient had high myopia of −11.0D, resulting in a decrease in cpRNFL thickness overall.
Finally, case 5 (Supplemental Figure S5) was diagnosed with glaucoma owing to increased CDR, RNFL defect, and VF defect in the corresponding area. However, the proposed model predicted it to be normal. When examining the rationale for deriving these results based on the three charts, it seems that normal pRNFL thickness and VF test influenced this. In fact, in optic disc shape and red-free fundus photography, this patient can be clinically diagnosed with superior segmental optic hypoplasia (SSOH) accompanied by early glaucoma. However, our model was determined to be normal based on RNFL thickness and VF test indices. A review of all patients showed that AI has difficulty diagnosing glaucoma, pre-perimetric glaucoma, early glaucoma (no reduction in RNFL thickness in OCT or no VF defect in VF test), high myopia, and it was a case of accompanying optic nerve abnormalities such as SSOH. In fact, glaucoma specialists often have difficulty diagnosing glaucoma in these patients immediately at their first visit. In this case, if we use the proposed approach to check the basis of judgment by referring to the SHAP chart, we will be able to receive help for the diagnosis of glaucoma.

4.3. Analysis of Missed Predicted Cases

The proposed prediction model produces 5.3% of missed predicted cases. Table 7 shows the confusion matrix of the proposed model. We tested the proposed model using all cases, and the results are described in Table 7. False positive was observed in 28 cases and false negative in 59 cases. We analyzed the mean of feature values against correct and missed predictions for glaucoma cases. In Table 8, missed prediction implies that the prediction is “healthy,” but the actual one is “glaucoma.” We can confirm that glaucoma cases with low PSD and high RNFL_S and RNFL_I values can easily be mispredicted as they are the characteristics of healthy cases. Table 9 shows that healthy cases with low RNFL_S and RNFL_I values can easily be mispredicted as glaucoma. We need to consider more information to avoid misprediction. This is a further research topic.
A feature interaction is a well-known phenomenon in machine learning prediction models. In a specific case of prediction, the features influence each other. In Figure 8, for example, two cases have the same PSD value of 1.39 and prediction results are “Healthy.” However, the degrees of support for Cases A and B are 0.84 and 0.93, respectively. In case B, RNFL_S and RNFL_I support “Glaucoma,” whereas they support “Healthy” in Case A; however, degree of support is increased from 0.84 to 0.93. This is because the degree of support of RNFL_T and IOP for “Healthy” is increased. The PSD may be more strongly influenced by RNFL_T and IOP than by RNFL_S and RNFL_I. Figure 9 shows the distributions of the degree of support according to the feature values. A wide distribution of such support for a feature value indicates a high influence from the other feature values. Five features show different non-linear distribution patterns. Furthermore, the distributions are different according to the range of values in a feature. There are points where the distribution changes rapidly in IOP and RNFL_T. We do not know the full meaning of these facts at this point, but we believe the facts imply an important hint toward a glaucoma diagnosis. This is a further topic of research.

5. Conclusions

We implemented our prediction model and three explanation charts on the web. The name of the system is Magellan (Supplemental Figure S6). If users select a patient case, Magellan shows the prediction results and explanation charts. A standalone version of Magellan is also being developed. Ophthalmologists may refer to Magellan and obtain additional evidence before making a final decision. Magellan is particularly helpful when the results of IOP, fundus photography, OCT, and VF are not consistent. If the ophthalmologist uploads patient’s results to this system, it can help determine whether a patient has glaucoma. The diagnosis evidence provided by Magellan, an explainable AI system, improves reliability by giving clues for diagnosis to ophthalmologists who are not glaucoma specialists. By providing a basis for judgement when explaining to patients, Magellan will be also a useful tool for communicating with patients.

Supplementary Materials

The following are available online at https://www.mdpi.com/2075-4418/11/3/510/s1 (accessed on 19 February 2021), Figure S1: Prediction case 1, Figure S2: Prediction case 2, Figure S3: Prediction case 3, Figure S4: Prediction case 4, Figure S5: Prediction case 5, Figure S6: Screen shot of Magellan system.

Author Contributions

Conceptualization, S.O. and S.J.K.; methodology, S.O. and S.J.K.; software, S.O.; validation, Y.P., K.J.C. and S.J.K.; formal analysis, S.O.; investigation, Y.P., K.J.C., S.J.K., and S.O.; resources, K.J.C. and S.J.K.; data curation, S.O.; writing—original draft preparation, S.O. and S.J.K.; writing—review and editing, Y.P. and K.J.C.; visualization, S.O.; supervision, S.O.; project administration, S.O.; funding acquisition, S.O. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the ICT and RND program of MIST/IITP [2018-0-00242, Development of AI ophthalmologic diagnosis and smart treatment platform based on big data].

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Gyeongsang National University Hospital Institutional Review Board, Korea (GNUH-2019-11-026-004).

Informed Consent Statement

Informed consent of the participants was waived by the Gyeongsang National University Hospital Institutional Review Board.

Data Availability Statement

NA.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shaikh, Y.; Yu, F.; Coleman, A.L. Burden of undetected and untreated glaucoma in the United States. Am. J. Ophthalmol. 2014, 158, 1121–1129.e1. [Google Scholar] [CrossRef] [PubMed]
  2. Taketani, Y.; Murata, H.; Fujino, Y.; Mayama, C.; Asaoka, R. How many visual felds are required to precisely predict future test results in glaucoma patients when using diferent trend analyses? Investig. Ophthalmol. Vis. Sci. 2015, 56, 4076–4082. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Asaoka, R.; Murata, H.; Iwase, A.; Araie, M. Detecting preperimetric glaucoma with standard automated perimetry using a deep learning classifer. Ophthalmology 2016, 123, 1974–1980. [Google Scholar] [CrossRef]
  4. An, G.; Omodaka, K.; Hashimoto, K.; Tsuda, S.; Shiga, Y.; Takada, N.; Kikawa, T.; Yokota, H.; Akiba, M.; Nakazawa, T. Glaucoma diagnosis with machine learning based on optical coherence tomography and color fundus images. J. Healthc. Eng. 2019, 1, 10. [Google Scholar] [CrossRef]
  5. Wang, P.; Shen, J.; Chang, R.; Moloney, M.; Torres, M.; Burkemper, B.; Jiang, X.; Rodger, D.; Varma, R.; Richter, G.M. Machine learning models for diagnosing glaucoma from retinal nerve fiber layer thickness maps. Ophthalmol. Glaucoma 2019, 2, 422–428. [Google Scholar] [CrossRef] [PubMed]
  6. Asaoka, R.; Murata, H.; Hirasawa, K.; Fujino, Y.; Matsuura, M.; Miki, A.; Kanamoto, T.; Ikeda, Y.; Mori, K.; Iwase, A.; et al. Using deep learning and transfer learning to accurately diagnose early-onset glaucoma from macular optical coherence tomography images. Am. J. Ophthalmol. 2019, 198, 136–145. [Google Scholar] [CrossRef]
  7. Barros, D.M.; Moura, J.C.; Freire, C.R.; Taleb, A.C.; Valentim, R.A.; Morais, P.S. Machine learning applied to retinal image processing for glaucoma detection: Review and perspective. BioMed Eng. OnLine 2020, 19, 1–21. [Google Scholar] [CrossRef] [Green Version]
  8. Thomas, P.B.; Chan, T.; Nixon, T.; Muthusamy, B.; White, A. Feasibility of simple machine learning approaches to support detection of non-glaucomatous visual fields in future automated glaucoma clinics. Eye 2019, 33, 1133–1139. [Google Scholar] [CrossRef]
  9. Lee, S.D.; Lee, J.H.; Choi, Y.G.; You, H.C.; Kang, J.H.; Jun, C.H. Machine learning models based on the dimensionality reduction of standard automated perimetry data for glaucoma diagnosis. Artif. Intell. Med. 2019, 94, 110–116. [Google Scholar] [CrossRef]
  10. Renukalatha, S.; Suresh, K.V. Classification of glaucoma using simplified-multiclass support vector machine. Biomed. Eng. 2019, 31, 1950039. [Google Scholar] [CrossRef]
  11. Maetschke, S.; Antony, B.; Ishikawa, H.; Wollstein, G.; Schuman, J.; Garnavi, R. A feature agnostic approach for glaucoma detection in OCT volumes. PLoS ONE 2019, 14, e0219126. [Google Scholar] [CrossRef]
  12. Mehta, P.; Petersen, C.; Wen, J.C.; Banitt, M.R.; Chen, P.P.; Bojikian, K.D.; Rokem, A. Automated detection of glaucoma with interpretable machine learning using clinical data and multi-modal retinal images. BioRxiv 2020, 1–20. [Google Scholar] [CrossRef] [Green Version]
  13. Liao, W.; Zou, B.; Zhao, R.; Chen, Y.; He, Z.; Zhou, M. Clinical Interpretable Deep Learning Model for Glaucoma Diagnosis. IEEE J. Biomed. Health 2019, 24, 1405–1412. [Google Scholar] [CrossRef] [PubMed]
  14. MacCormick, I.J.; Williams, B.M.; Zheng, Y.; Li, K.; Al-Bander, B.; Czanner, S.; Cheeseman, R.; Willoughby, C.E.; Brown, E.N.; Spaeth, G.L.; et al. Accurate, fast, data efficient and interpretable glaucoma diagnosis with automated spatial analysis of the whole cup to disc profile. PLoS ONE 2019, 14, e0209409. [Google Scholar]
  15. Mojab, N.; Noroozi, V.; Yu, P.; Hallak, J. Deep Multi-task Learning for Interpretable Glaucoma Detection. In Proceedings of the IEEE 20th International Conference on Information Reuse and Integration for Data Science (IRI), Los Angeles, CA, USA, 30 July–1 August 2019; pp. 167–174. [Google Scholar]
  16. Interpretable Machine Learning. 2019. Available online: https://christophm.github.io/interpretable-ml-book/ (accessed on 7 January 2021).
  17. Du, M.; Liu, N.; Hu, X. Techniques for interpretable machine learning. Commun. ACM 2019, 63, 68–77. [Google Scholar] [CrossRef] [Green Version]
  18. Hooker, G. Discovering additive structure in black box functions. In Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 22–25 August 2004; pp. 575–580. [Google Scholar]
  19. Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
  20. Messalas, A.; Kanellopoulos, Y.; Makris, C. Model-agnostic interpretability with shapley values. In Proceedings of the 2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA), Achaia, Greece, 15–17 July 2019; pp. 1–7. [Google Scholar]
  21. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 4765–4774. [Google Scholar]
  22. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  23. Xgboost: Extreme Gradient Boosting. Available online: https://cran.r-project.org/web/packages/xgboost/xgboost.pdf (accessed on 3 December 2020).
  24. Wang, M.; Pasquale, L.R.; Shen, L.Q.; Boland, M.V.; Wellik, S.R.; De Moraes, C.G.; Myers, J.S.; Wang, H.; Baniasadi, N.; Li, D.; et al. Reversal of glaucoma hemifield test results and visual field features in glaucoma. Ophthalmology 2018, 125, 352–360. [Google Scholar] [CrossRef]
  25. Oh, S. Feature Interaction in Terms of Prediction Performance. Appl. Sci. 2019, 23, 5191. [Google Scholar] [CrossRef] [Green Version]
  26. Yadav, K.S.; Rajpurohit, R.; Sharma, S. Glaucoma: Current treatment and impact of advanced drug delivery systems. Life Sci. 2019, 221, 362–376. [Google Scholar] [CrossRef] [PubMed]
  27. Yadav, K.S.; Sharma, S.; Londhe, V.Y. Bio-tactics for neuroprotection of retinal ganglion cells in the treatment of glaucoma. Life Sci. 2020, 243, 117303. [Google Scholar] [CrossRef]
  28. Al-Aswad, L.A.; Kapoor, R.; Chu, C.K.; Walters, S.; Gong, D.; Garg, A.; Gopal, K.; Patel, V.; Sameer, T.; Rogers, T.W.; et al. Evaluation of a Deep Learning System For Identifying Glaucomatous Optic Neuropathy Based on Color Fundus Photographs. J. Glaucoma 2019, 28, 1029–1034. [Google Scholar] [CrossRef]
  29. Devalla, S.K.; Liang, Z.; Pham, T.H.; Boote, C.; Strouthidis, N.G.; Thiery, A.H.; Girard, M.J. Glaucoma management in the era of artificial intelligence. Br. J. Ophthalmol. 2019, 104, 301–311. [Google Scholar] [CrossRef] [PubMed]
  30. Hashimoto, Y.; Asaoka, R.; Kiwaki, T.; Sugiura, H.; Asano, S.; Murata, H.; Fujino, Y.; Matsuura, M.; Miki, A.; Mori, K.; et al. Deep learning model to predict visual field in central 10° from optical coherence tomography measurement in glaucoma. Br. J. Ophthalmol. 2020, 1–7. [Google Scholar] [CrossRef] [PubMed]
  31. Christopher, M.; Bowd, C.; Belghith, A.; Goldbaum, M.H.; Weinreb, R.N.; Fazio, M.A.; Girkin, C.A.; Liebmann, J.M.; Zangwill, L.M. Deep Learning Approaches Predict Glaucomatous Visual Field Damage from Optical Coherence Tomography Optic Nerve Head Enface Images and Retinal Nerve Fiber Layer Thickness Maps. Ophthalmology 2020, 127, 346–356. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Procedure of prediction model development.
Figure 1. Procedure of prediction model development.
Diagnostics 11 00510 g001
Figure 2. Box plots for selected features (H, healthy control; G, glaucoma).
Figure 2. Box plots for selected features (H, healthy control; G, glaucoma).
Diagnostics 11 00510 g002
Figure 3. An example of a gauge chart and a boxplot for RNFL_T. (Left: gauge chart of RNFL_S with a value of 86, Right: boxplot and statistics of distributions for RNFL_S).
Figure 3. An example of a gauge chart and a boxplot for RNFL_T. (Left: gauge chart of RNFL_S with a value of 86, Right: boxplot and statistics of distributions for RNFL_S).
Diagnostics 11 00510 g003
Figure 4. Radar charts for typical glaucoma and healthy patients.
Figure 4. Radar charts for typical glaucoma and healthy patients.
Diagnostics 11 00510 g004
Figure 5. An example of a Shapley Additive Explanations (SHAP) chart.
Figure 5. An example of a Shapley Additive Explanations (SHAP) chart.
Diagnostics 11 00510 g005
Figure 6. Feature importance of proposed model.
Figure 6. Feature importance of proposed model.
Diagnostics 11 00510 g006
Figure 7. Feature interaction chart of the proposed model.
Figure 7. Feature interaction chart of the proposed model.
Diagnostics 11 00510 g007
Figure 8. SHAP chart that shows feature interaction. The same PSD value of 1.39 has different weights in case A and case B.
Figure 8. SHAP chart that shows feature interaction. The same PSD value of 1.39 has different weights in case A and case B.
Diagnostics 11 00510 g008
Figure 9. Distributions of degree of support according to the feature values.
Figure 9. Distributions of degree of support according to the feature values.
Diagnostics 11 00510 g009
Table 1. Characteristics of the participants.
Table 1. Characteristics of the participants.
PatientNormal
Group
Glaucoma
Group
Totalp-Values *
Number of participants377430807-
Gender (male/female)201/176260/170461/3460.04061
Age (mean ± SD)51.7 ± 16.560.3 ± 14.156.9 ± 15.7<0.001
Number of eyes5646801244-
Number of cases6499751624-
* Normal group vs. Glaucoma group.
Table 2. List of candidate features for Model building.
Table 2. List of candidate features for Model building.
NoFeatureGlaucoma
Mean (SD 1)
Healthy
Mean (SD)
p-Value
1Sex ---
2Age60.3 (14.13)51.7 (16.45)<0.001
3GHT 24.28 (1.28)2.10 (1.53)<0.001
4VFI 372.3 (32.24)95.7 (5.55)<0.001
5MD 4−10.24 (9.72)−2.33 (2.55)<0.001
6Pattern standard deviation6.76 (4.27)2.49 (1.03)<0.001
7RNFL 5 superior82.0 (27.71)112.6 (19.89)<0.001
8RNFL Nasal56.1 (33.84)64.6 (16.64)<0.001
9RNFL inferior79.9 (29.76)117.6 (20.28)<0.001
10RNFL temporal58.6 (18.48)71.6 (15.01)<0.001
11Mean of RNFL thickness68.5 (18.82)91.6 (12.60)<0.001
12Intraocular pressure18.7 (8.69)15.7 (3.10)<0.001
13Cornea thickness527.2 (34.15)530.1 (34.01)<0.001
14BCVA 60.63 (0.31)0.73 (0.31)0.002
15Spherical equivalent−1.63 (2.88)−1.42 (3.08)0.12
16Axial length24.1 (1.81)24.1 (1.42)0.92
17Neuro-retinal rim0.79 (0.28)1.06 (0.21)<0.001
18Cup 0.47 (0.23)0.38 (0.43)0.16
19Disc1.97 (0.23)2.09 (0.43)0.25
20Mean of cup/disc ratio0.74 (0.11)0.65 (0.12)<0.001
21vertical_cup/disc ratio0.73 (0.10)0.62 (0.16)<0.001
22CNN 7 degree0.69 (0.18)0.53 (0.21)<0.001
1 standard deviation; 2 glaucoma hemifield test; 3 visual field index; 4 mean deviation; 5 retinal nerve fiber layer; 6 best-corrected visual acuity; 7 convolutional neural network.
Table 3. Final features list for building the prediction model.
Table 3. Final features list for building the prediction model.
NoFeatureAbbreviationSource
1pattern standard deviationPSDVF
2RNFL superiorRNFL_SRNFL optical coherence tomography (OCT)
3RNFL inferiorRNFL_I RNFL OCT
4RNFL temporalRNFL_TRNFL OCT
5intraocular pressureIOPIOP test
Table 4. Hyper parameters of proposed XGboost prediction model.
Table 4. Hyper parameters of proposed XGboost prediction model.
NoHyper Parameter *Value
1booster“gbtree”
2eta0.7
3max_depth8
4gamma3
5subsample0.8
6colsample_bytree0.5
7objective “multi:softprob”
8eval_metric“merror”
9num_class2
* We use default values for other hyper-parameters that are not listed in the table.
Table 5. Final features list for building the prediction model.
Table 5. Final features list for building the prediction model.
MetricSupport Vector Machine (SVM)C50Random Forest (RF)xgboost
Accuracy0.9250.9030.9370.947
Sensitivity0.9330.8740.9240.941
Specificity0.9200.920.9450.950
AUC0.9450.8970.9450.945
Table 6. Cases of prediction by the proposed model.
Table 6. Cases of prediction by the proposed model.
CasePSDRNFL_SRNFL_IRNFL_TIOPDiagnosisPrediction
11.921421539413HealthyHealthy
211.8583415514GlaucomaGlaucoma
31.53731077118GlaucomaHealthy
42.7681957318HealthyGlaucoma
52.31981306012GlaucomaHealthy
Table 7. Confusion matrix of the proposed prediction model.
Table 7. Confusion matrix of the proposed prediction model.
Predict
GlaucomaHealthy
ActualGlaucoma91659
Healthy28621
Table 8. Analysis of prediction: glaucoma cases.
Table 8. Analysis of prediction: glaucoma cases.
CasePSDRNFL_SRNFL_IRNFL_TIOP
Correct prediction7.9975.3172.8657.9820.73
Miss prediction2.65111.12122.6971.2515.78
p-value2.23 × 10−552.74 × 10−241.22 × 10−343.09 × 10−101.39 × 10−17
Table 9. Analysis of prediction: healthy cases.
Table 9. Analysis of prediction: healthy cases.
CasePSDRNFL_SRNFL_IRNFL_TIOP
Correct prediction2.23117.0126.5977.6915.72
Miss prediction3.8490.4091.2161.8915.68
p-value1.64 × 10−31.74 × 10−82.66 × 10−111.27 × 10−69.44 × 10−1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Oh, S.; Park, Y.; Cho, K.J.; Kim, S.J. Explainable Machine Learning Model for Glaucoma Diagnosis and Its Interpretation. Diagnostics 2021, 11, 510. https://doi.org/10.3390/diagnostics11030510

AMA Style

Oh S, Park Y, Cho KJ, Kim SJ. Explainable Machine Learning Model for Glaucoma Diagnosis and Its Interpretation. Diagnostics. 2021; 11(3):510. https://doi.org/10.3390/diagnostics11030510

Chicago/Turabian Style

Oh, Sejong, Yuli Park, Kyong Jin Cho, and Seong Jae Kim. 2021. "Explainable Machine Learning Model for Glaucoma Diagnosis and Its Interpretation" Diagnostics 11, no. 3: 510. https://doi.org/10.3390/diagnostics11030510

APA Style

Oh, S., Park, Y., Cho, K. J., & Kim, S. J. (2021). Explainable Machine Learning Model for Glaucoma Diagnosis and Its Interpretation. Diagnostics, 11(3), 510. https://doi.org/10.3390/diagnostics11030510

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop