Next Article in Journal
A Spatially Guided Machine-Learning Method to Classify and Quantify Glomerular Patterns of Injury in Histology Images
Next Article in Special Issue
Retinal Microvasculature Image Analysis Using Optical Coherence Tomography Angiography in Patients with Post-COVID-19 Syndrome
Previous Article in Journal
Threshold-Based BRISQUE-Assisted Deep Learning for Enhancing Crack Detection in Concrete Structures
Previous Article in Special Issue
Multi-Fundus Diseases Classification Using Retinal Optical Coherence Tomography Images with Swin Transformer V2
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluating Retinal Disease Diagnosis with an Interpretable Lightweight CNN Model Resistant to Adversarial Attacks

1
Department of Science and Technology, Samriddhi College, Bhaktapur 44800, Nepal
2
School of Engineering and Technology, Central Queensland University, Norman Gardens, Rockhampton, QLD 4701, Australia
3
Central Department of Computer Science and IT, Tribhuvan University, Kathmandu 44600, Nepal
*
Author to whom correspondence should be addressed.
J. Imaging 2023, 9(10), 219; https://doi.org/10.3390/jimaging9100219
Submission received: 1 September 2023 / Revised: 29 September 2023 / Accepted: 9 October 2023 / Published: 11 October 2023
(This article belongs to the Special Issue Advances in Retinal Image Processing)

Abstract

:
Optical Coherence Tomography (OCT) is an imperative symptomatic tool empowering the diagnosis of retinal diseases and anomalies. The manual decision towards those anomalies by specialists is the norm, but its labor-intensive nature calls for more proficient strategies. Consequently, the study recommends employing a Convolutional Neural Network (CNN) for the classification of OCT images derived from the OCT dataset into distinct categories, including Choroidal NeoVascularization (CNV), Diabetic Macular Edema (DME), Drusen, and Normal. The average k-fold (k = 10) training accuracy, test accuracy, validation accuracy, training loss, test loss, and validation loss values of the proposed model are 96.33%, 94.29%, 94.12%, 0.1073, 0.2002, and 0.1927, respectively. Fast Gradient Sign Method (FGSM) is employed to introduce non-random noise aligned with the cost function’s data gradient, with varying epsilon values scaling the noise, and the model correctly handles all noise levels below 0.1 epsilon. Explainable AI algorithms: Local Interpretable Model-Agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP) are utilized to provide human interpretable explanations approximating the behaviour of the model within the region of a particular retinal image. Additionally, two supplementary datasets, namely, COVID-19 and Kidney Stone, are assimilated to enhance the model’s robustness and versatility, resulting in a level of precision comparable to state-of-the-art methodologies. Incorporating a lightweight CNN model with 983,716 parameters, 2.37 × 10 8 floating point operations per second (FLOPs) and leveraging explainable AI strategies, this study contributes to efficient OCT-based diagnosis, underscores its potential in advancing medical diagnostics, and offers assistance in the Internet-of-Medical-Things.

1. Introduction

The retina, situated at the posterior aspect of the ocular globe, comprises photoreceptor cells adept at transducing luminous stimuli into intricate electrical signals, subsequently dispatched to the cerebral cortex via the optic nerve. This intricate process serves as the foundation for human visual perception, wherein the brain deciphers these electrical transmissions as coherent visual representations. Retinal diseases can seriously affect vision and in some cases, can lead to permanent blindness [1] which is a big problem for the general health of the public. Getting a prompt and accurate diagnosis with the help of automated tools is a great assist for medical specialists in making wise medical decisions. The advancement of digital medical imaging has brought about a significant change in ophthalmology as it has introduced effective technologies that help in the detection of such diseases. By improving early detection through image analysis and identifying minuscule anomalies, Artificial Intelligence (AI) has considerably coped with retinal diseases. AI has also enhanced treatment planning by analyzing patient data and enabling tailored care. Additionally, AI-driven systems help track the development of diseases, resulting in therapies that are more successful [2].
Different Machine Learning (ML) and Convolutional Neural Networks (CNNs) are efficient at analyzing images and are particularly incredible at recognizing complex patterns in medical images [3]. Their ability to diagnose complicated retinal diseases is efficient without a doubt, but in medical practice, using CNNs depends not only on how well they can diagnose the issues but also on how useful they are in places with limited computational resources. Not only CNN, but different variants of CNN like ResNet [4], VGG [5] and more have produced good accuracies statistically. These CNNs and their variants have a very high number of training parameters, and many layers which make it time-consuming in real-time predictions [6] and integration with the Internet-of-Medical-Things (IoMT) [7].
As AI technology advances, it has become essential to not only achieve better diagnostic abilities but also to understand how these AI systems make predictions and decisions [8,9,10]. As these models can be hard to understand because of their statistical nature making them black boxes [11], the addition of Explainable Artificial Intelligence (XAI) into these models can solve the problem. The combination of small and efficient CNN models in IoMT devices with XAI, as a bio-marker, helps retinal disease diagnosis to be more accurate and more accessible for medical experts, practitioners, and even ordinary people.
To resolve all of these issues, this study aims to achieve the following three key objectives:
  • To develop an efficient CNN model with minimal parameters for detecting retinal abnormalities such as CNV, DME, and Drusen using OCT datasets.
  • To incorporate Explainable Artificial Intelligence (XAI) methodologies such as Local Interpretable Model-agnostic Explanations (LIME) and Shapley additive explanations (SHAP) into the realm of clinical interpretation, with the aim of comprehending the prediction by the Convolutional Neural Network (CNN) model.
  • To generalize the model’s reliability and applicability; two new additional datasets were trained and evaluated for the model.

2. Related Works

Researchers in health informatics are leveraging the predictive power of Deep Learning (DL) to address the automated diagnosis of various diseases such as COVID-19 [12], monkeypox [13], kidney stone [14] and so on. Here, we summarise the recent DL methods that have been employed for retinal disease diagnosis using various image modalities. These methods can be categorized into two broad classes: pre-trained DL models (that leverage the transfer learning strategies) and custom-designed CNN (which needs training from scratch).
Subramanian et al. [15] utilised four CNN models such as VGG16, DenseNet-201, Inception-V3, and Xception, to classify seven different retinal diseases. Moreover, Bayesian optimization was employed to fine-tune the hyperparameters of these CNN models, coupled with image augmentation techniques to enhance their ability to generalize. The use of DenseNet-201 in classifying retinal diseases on the Retinal OCT Image dataset resulted in an accuracy exceeding 99%, demonstrating superior performance compared to alternative methods. Puneet et al. [16] implemented the combination of Attention and Transfer Learning approaches into a DCNN for categorizing retinal diseases such as CNV, DME, and Drusen using OCT images. Their proposal achieved notable results, attaining accuracies of 97.79% during training and 95.6% during testing. Kayadibi et al. [17] implemented a hybrid fine-tuned CNN for retinal disease classification from OCT images. They utilized PCA to reduce the feature size and enhance the performance. The benchmarking outcomes for two OCT datasets demonstrated a high level of promise in terms of accuracy. Specifically, the UCSD dataset yielded an impressive accuracy of 99.70% according to Kermany et al.’s study  [18], while the Duke dataset achieved a perfect 100% accuracy as reported by Srinivasan et al. [19]. In their research, Kim and colleagues [20] harnessed a variety of Convolutional Neural Networks (CNNs) like VGG16, ResNet50, DenseNet121, and Inception-v3 as feature extractors. Subsequently, they employed these features to develop binary OCT image classification models. A binary classifier model is developed for each category (CNV, DME, Drusen and Normal) and the VGG-16-based model for CNV vs. other classes achieved 98.6% accuracy. They achieve high accuracy using the pre-trained DL models. However, their proposal needs the training of individual models for each class which incurs high computational complexity. A pre-trained VGG-16 network was implemented by Li et al. [21] for retinal image classification on OCT images. They validated the model’s performance on 1000 independent OCT images. Their work revealed that the transfer learning with the VGG-16 model has a promising accuracy of 98.6%, sensitivity of 97.8%, and specificity of 00.4%. With such commendable performance of the model, deep learning can automate the diagnosis of retinal disease. Li et al. [22] adopted the ensemble models for retinal disease classification using OCT images. They trained four DL- models based on improved ResNet50 to build the ensemble and achieved the highest accuracy of 96.3%, sensitivity of 96.6%, and specificity of 98.7%. However, the ResNet50 model is itself the heavyweight model.
In addition to employing pre-trained deep learning models, only a limited number of researchers have created custom CNNs for the classification of retinal images. For example, a deep CNN with six convolution blocks (including the Relu, batch normalization, and pooling operation) was implemented by Sujina et al. [23]. Their proposal achieved a promising accuracy of 99.69% with a low misclassification rate. However, the generalisability of the CNN on additional datasets is not reported. Altan et al. [24] implemented a deep learning architecture to detect the macular edema on OCT images and reported an accuracy of 99.20%.
Hybrid deep learning models for retinal image classification have also been proposed recently. For instance, a hybrid deep learning model for OCT image classification was implemented by Khan et al. [25]. They extracted retinal features from OCT images using three pre-trained deep learning models (DenseNet121, InceptionV3, and ResNet50), and ant colony optimization was used for best feature selection. Finally, the SVM and KNN were employed for classification. Their proposal achieved high performance on OCT image classification. However, the approach is not applicable to end-to-end training of the model.

3. Materials and Methods

The entire material and methods adopted in the study are depicted in Figure 1, which includes the stages ranging from data preparation to model evaluation.

3.1. Dataset

3.1.1. Disease Description

Figure 2 [26] shows the different diseases considered in this study. CNV, depicted in Figure 2a, arises from the emergence of fresh blood vessels in proximity to the choroid. CNV is caused by flaws in the innermost section of the choroid known as Bruch’s membrane, along with conditions like severe nearsightedness and heightened vascular endothelial growth. DME, Figure 2b, primarily affects individuals with diabetes. It leads to vision distortion as fluid accumulates in the macula. This accumulation impairs cone cells’ light-sensing abilities, causing blurred vision. DME arises from the expansion of blood vessels at the posterior region. In Figure 2c, we can observe Drusen, a condition primarily linked to the aging process. It involves the accumulation of yellow extracellular particles between the Bruch’s membrane and the retinal pigment in the eye. Drusen has the potential to hinder the transport system, which could lead to a deprivation of oxygen for the cone cells responsible for colour vision, ultimately resulting in their deterioration.

3.1.2. Dataset Description

The publicly accessible dataset [27] encompasses detailed cross-sectional images of living patients’ retinas, which have been classified into four distinct categories: Normal, CNV, Drusen, and DME. These categories are visually represented in Figure 3. The dataset comprises a grand total of 84,492 images, distributed as follows: CNV contains 37,457 images, Normal contains 26,567 images, DME includes 11,600 images, and Drusen encompasses 8868 images.

3.2. Dataset Pre-Processing

The originally imbalanced dataset was transformed into a balanced one, where each category contained exactly 8868 images. In this balanced dataset, all images were resized to dimensions of 180 × 180 and then normalized to fall within the range [0, 1]. To ensure representative sampling, a stratified approach was employed, allocating 80% of the data for training, and the remaining 20% for testing, with half of the testing data reserved for validation.

3.3. CNN Model

The proposed lightweight model reported in Table 1 provides a concise overview of the CNN architecture and Figure 4 shows the graphical result. The model holds only five convolution layers to perform convolution operations on the input image, with increasing filter depths (16, 32, 64, 128, 256) to capture hierarchical features. Each convolution layer is followed by max-pooling layers to down-sample the feature maps, aiding in information compression. Dropout layers help mitigate over-fitting by randomly deactivating neurons during training with values of 0.2 on each. The final dense layers (256, 4) process flattened features for classification, culminating in four output classes. The model contains around 983,716 total trainable parameters, contributing to its complexity and predictive capacity.

3.4. Implementation Setup

The CNN model and XAI algorithms in this study were implemented using Python 3.10.12, along with Keras version 2.13.1 and TensorFlow version 2.13.0. The computational resources employed for this research included the runtime environment provided by Google Colab, supported by a robust NVIDIA K80 GPU with an impressive 12 GB of RAM. To assess the performance of the CT dataset, a cross-validation technique with a designated value of K = 10 [28] was employed, which entailed distinct and random allocations for both training and testing subsets. For the purpose of regularization, an early stopping strategy was employed, which relied on monitoring the validation loss for 10 consecutive epochs.

3.5. Performance Evaluation Metrics

3.5.1. Accuracy

The performance of classification models is typically evaluated using the metric of accuracy. Out of all the examples in a dataset, it calculates the percentage of accurately predicted instances. Mathematically, accuracy ( A c c ) is calculated as Equation (1).
A c c = Number of Correct Predictions Total Number of Predictions

3.5.2. Precision

Precision is a performance measure that assesses the correctness of a model’s positive predictions. It is determined by the ratio of true positive predictions ( T P ) to the sum of true positive and false positive predictions ( T P + F P ).

3.5.3. Recall

Recall evaluates a capacity to accurately identify every positive occurrence present in the dataset. It is described as the proportion of genuine positives (positives that were correctly detected) to all real positives.

3.5.4. F1-Score

The F1 Score is a classification task evaluation metric that balances precision and recall. It is calculated as the harmonic mean of precision and recall, offering a single measure of model performance that takes both false positives and false negatives into account.

3.5.5. ROC Curve

The Receiver Operating Characteristic (ROC) curve is a graph that shows how well a model works. It shows the trade-off between two things: how often the model correctly says “yes” when it should (Sensitivity), and how often it incorrectly says “yes” when it shouldn’t (1-Specificity). This graph helps us see how good the model is at telling things apart in different settings. To make the ROC curve, we draw a graph of the True Positive Rate (TPR) against the False Positive Rate (FPR) for different settings. The area under the ROC curve (AUC-ROC) gives us a single number that tells us how well the model is doing.

3.5.6. FLOPS

Algorithm 1 calculates the Floating Point Operations (FLOPs) for the CNN model [29]. It defines two functions, CalculateCNNLayerFLOPs, and CalculateDenseLayerFLOPs, to compute FLOPs for Conv2D and Dense layers, respectively, based on their parameters. The CalculateTotalFLOPs function iterates through the model’s layers, identifying Conv2D and Dense layers, and accumulates their respective FLOPs. This provides an estimate of the total computational complexity of the CNN model. The algorithm is valuable for assessing the computational efficiency of the CNN in terms of the number of operations needed for inference.
Algorithm 1 Calculate FLOPs for CNN Model [29]
1:
function CalculateCNNLayerFLOPs(layer)
2:
      Input: CNN layer
3:
      Output: FLOPs for the layer
4:
      return  2 × layer . filters × layer . kernel_size [ 0 ] × layer . kernel_size [ 1 ] × layer . input_shape [ 1 ] × layer . output_shape [ 1 ] × layer . output_shape [ 2 ]
5:
end function
6:
function CalculateDenseLayerFLOPs(layer)
7:
      Input: Dense layer
8:
      Output: FLOPs for the layer
9:
      return  2 × layer . input_shape [ 1 ] × layer . output_shape [ 1 ]
10:
end function
11:
function CalculateTotalFLOPs(model)
12:
      Input: CNN model
13:
      Output: Total FLOPs for the model
14:
      total_flops ← 0
15:
      for layer in model.layers do
16:
          if layer is Conv2D then
17:
                total_flops + = CalculateCNNLayerFLOPs ( l a y e r )
18:
          else if layer is Dense then
19:
                total_flops + = CalculateDenseLayerFLOPs ( l a y e r )
20:
          end if
21:
      end for
22:
      return total_flops
23:
end function

3.5.7. Explainable AI

Although there are certain challenges associated with XAI, such as its sensitivity to individual cases, the trade-off involving complexity, and the assumption of highly interdependent features, XAI delves into the visual computational approach of Deep Learning models. Consequently, the study incorporates the use of LIME and SHAP.
  • LIME
    In the pursuit of enhancing the transparency and interpretability of modern machine learning models, LIME has emerged as a powerful technique. LIME addresses the challenge of understanding complex black-box models’ predictions by approximating their behaviour through locally interpretable models. This approach allows us to shed light on how specific features influence predictions, especially in contexts involving intricate data types such as images.
    LIME operates by selecting a target instance x, model f and generating perturbed instances x i in its vicinity. The model’s predictions f ( x ) and f ( x i ) are obtained, and interpretable features z i are extracted from the perturbed instances. An interpretable CNN model g ( z ) , was trained using pairs ( z i , f ( x i ) ) to approximate the complex behavior of f in the local neighborhood of x. To analyse g ( z ) , coefficients β i in g ( z ) , the importance of the corresponding features z i in influencing the predictions were reflected. Larger absolute values of β i indicate stronger influences [30].
  • SHAP
    For the retinal input retinal image x with N number of pixels and f prediction, SHAP values for each pixel in the image were calculated. The values of SHAP show the contribution of the model to define how much each pixel i in the retinal image x contributes and is calculated using Equation (2).
    ϕ i ( x ) = S { 1 , 2 , , N } { i } | S | ! ( N | S | 1 ) ! N ! [ f ( x S { i } ) f ( x S ) ]
    S represents a subset of pixels excluding pixel i, x S is the image with the pixels in subset S unchanged, f ( x S { i } ) is the model’s prediction when pixel i is included in the subset S and f ( x S ) is the model’s prediction when pixel i is excluded from the subset S [31].
    These SHAP values provide insights into the contribution of each pixel to the model’s prediction. Positive SHAP values indicate that a pixel’s presence positively influenced the forecast, while negative values suggest the opposite.

4. Results

4.1. Statistical Results

We analyzed classic statistical validation measures, which included the model’s performance in terms of error and correctness throughout training, as well as across the validation and test datasets. Furthermore, we incorporated precision, F1-score, recall, confusion matrix, and k-fold validation into our evaluation.
Among ten different folds, 10th fold, stopped early in 15th epoch, holding the lowest accuracy, and the same is considered to plot the evaluation metrics. The training process spanned 20 epochs, with each fold configured to terminate early if the validation loss persisted for five consecutive epochs.
Figure 5, shows the training and validation accuracy of 10th fold where training phase yielded 95.64% and a loss of 0.1201, and the validation accuracy stood at 94.12% with a corresponding loss of 0.2185.
Figure 6 shows the performance matrix of 10th fold. Among 892 Drusen samples, 843 were accurately predicted. Only 52 normal samples were misclassified, out of which 18 were predicted as Drusen, 4 were predicted as CNV, and 30 were predicted as DME. Out of 900 samples, a total of 807 samples were predicted correctly for the CNV class. Considering 876 samples, 843 samples were predicted correctly for DME.
Following 10 folds, the model demonstrated an average training accuracy of 96.33%, a validation accuracy of 94.12%, a training loss of 0.1073, and a validation loss of 0.1927. The average testing accuracy and testing loss stood at 94.29% and 0.2002 respectively (Table 2).
Table 3 shows the classification report where “Drusen” and “Normal” show high Precision (0.94–0.95) indicating accurate positive predictions and DME shows high Recall (0.96), capturing most positives. F1-Score ranges from 0.94 to 0.95 showing the harmony between Precision and Recall.
As the imbalanced datasets were made balanced, the ROC curve (for 10th fold) was plotted as shown in Figure 7 to calculate the area under the curve and evaluate the model performance.
The model showed outstanding performance in distinguishing between positive and negative categories, as proven by its impressive AUC score of 0.99 in all areas. The ROC curve, which depends on TPR and FPR, displayed the model’s predictions on the test dataset having an exceptionally high TPR, covering the entire range of AUC values. This demonstrates the model’s exceptional effectiveness.
Table 4 presents sensitivity and specificity values of four classes in 10th fold. DME and Normal classes exhibit high sensitivity (>0.98), indicating accurate detection of relevant cases. Drusen has slightly lower sensitivity, while CNV has the highest specificity (>0.97), suggesting strong performance in distinguishing its class.

4.2. Explainable Results

4.2.1. SHAP

Because it is hard to understand how the CNN model predicted the output, XAI techniques are used to explain it [32]. The testing images are on the left, and each explanation has a transparent grey background (see Figure 8, Figure 9, Figure 10 and Figure 11).
The red pixels in the first explanation image (refer to Figure 8) increase the probability of predicting a Drusen. In the second explanation, the model somehow attempts to indicate that the image is normal, but the red pixels’ concentration in the first explanation is higher. Third, and fourth explanation images do not contain any red or blue pixels, so the probability of classifying the input image as CNV and DME is low.
In Figure 9, the presence of concentrated red pixels suggests that the image is normal. Conversely, in Figure 10 and Figure 11, the absence of concentrated red pixels indicates that they correspond to CNV and DME, respectively.

4.2.2. LIME

LIME is employed to extract the features and perturbations from the training dataset. These perturbations are randomly generated from a standardized image, and subsequent operations involving mean-centering and scaling are conducted. The simple linear iterative clustering (SLIC) [33] is used to compute the initial three characteristics which delineate the most influential boundaries and incorporate them into the image. The second column of Table 5 displays the original test images corresponding to each category. The segmented image segment in the third column of Table 5 represents the segmentation obtained through LIME. As illustrated in Table 5, LIME furnishes visual explanations of the model’s decision-making process, spotlighting the image regions that make a significant contribution to a specific class prediction.

5. FLOPS Calculation

The proposed model’s Floating-point Operations (FLOPs) were determined by considering all arithmetic operations involving floating-point values, such as addition, subtraction, division, multiplication, and any other relevant operations. The model executed a total of 2.37 × 10 8 operations, and this calculation was accomplished using Algorithm 1.

6. Generation of Fast Gradient Sign Method for Adversarial Examples

To evaluate the model’s robustness, we conducted tests using adversarial examples. We computed the gradient of the loss function with respect to the input images to capture subtle variations. We introduced epsilon as a hyperparameter to quantify the perturbation’s intensity, which was generated by taking the sign of the gradient and adjusting its magnitude. Subsequently, we incorporated this perturbation into the image and forwarded it for prediction. Our findings indicate that the model exhibited resilience up to an epsilon value of 0.1, as illustrated in Table 6 for DRUSEN and CNV, whereas all categories remained stable below the threshold of 0.1.

7. Generalizability Investigation

To see if the proposed model could be used to diagnose other common datasets with the same number of categories, the model was trained under the same constraints as before and the results were analysed for two additional datasets, COVID-19 and Kidney Stone.

7.1. COVID-19

Publicly available COVID-19 dataset [34], with 4273 pneumonia samples, 1583 normal samples, 703 tuberculosis samples, and 576 COVID-19 samples was balanced with an equal number of 576 images for each category. With a training accuracy of 97.18%, a test accuracy of 92.54%, and a validation accuracy of 94.78% as shown in Figure 12a, the model achieved a training loss of 0.0804, test loss of 0.2180 and validation loss of 0.1960 as shown in Figure 12b.
Figure 13a represents the confusion matrix. Here, six instances of Tuberculosis samples were inaccurately predicted out of a comprehensive pool of 56 samples. Seven mispredictions were observed among pneumonia samples out of 55. Similarly, for COVID-19 cases, seven errors emerged out of 66 samples, whereas all 54 samples categorized as normal were accurately predicted. To offer a more detailed insight into the findings, it is worth noting that the AUC-ROC values for tuberculosis, pneumonia, COVID-19, and normal cases stand at 0.99, 0.98, 0.97, and 1.00, respectively, as shown in Figure 13b.
Figure 14, Figure 15, Figure 16 and Figure 17 show the SHAP explanation of tuberculosis, pneumonia, COVID-19, and Normal samples respectively.
Figure 18 shows the LIME segmented results for individual categories of the COVID-19 dataset. The segmented results highlight the infected regions in respective images.
The proposed model shows better results in comparison with other SOTA approaches and the results are tabulated in Table 7 showing training accuracy of 97.18% and number of parameters (0.983 million).

7.2. Kidney Stone

The publicly available Kidney Stone dataset [14], with 5077 normal samples, 3709 cyst samples, 2283 tumor samples and 1377 stone samples was balanced with an equal number of 1377 images for each category. With a training accuracy of 99.70%, a test accuracy of 99.64%, and a validation accuracy of 99.82% as shown in Figure 19a, the model achieved a training loss of 0.0056, test loss of 0.0345 and validation loss of 0.0078 as shown in Figure 19b.
Figure 20a represents the confusion matrix for the kidney stone dataset. All cyst and stone samples were correctly classified. One tumor sample was misclassified as normal, and one normal sample was misclassified as stone. The AUCROC values for all categories are 1.00, which indicates that the model has perfect accuracy as shown in Figure 20b.
Figure 21, Figure 22, Figure 23, and Figure 24 show the SHAP explanation of tumor, cyst, and Normal samples respectively.
Figure 25 shows the LIME segmented results for individual categories of the kidney stone dataset. The segmented results highlight the infected regions in respective images.
The proposed model shows the competitive results in comparison with other state-of-the-art methods as tabulated in Table 8 in terms of training accuracy (99.70%) and number of parameters (0.983 million).

8. Conclusions

The study presents a significant advancement in OCT-based diagnostic methodologies to address the labor-intensive nature of manual anomaly classification in retinal images. Achieving remarkable average accuracies and minimal losses across training, validation and test sets, the proposed model demonstrates its efficacy in classifying CNV, DME, Drusen, and Normal retinal conditions. The integration of XAI techniques provides interpretable insights into the model’s decision-making process. Moreover, the model’s robustness and generalizability are substantiated by its consistent performance on additional datasets related to COVID-19 and Kidney Stone conditions. With a focus on efficiency and lightweight, the model can play a significant role in IoMT devices.
With more real-time datasets, augmentation, and generative adversarial networks, other lightweight transfer learning models like MobileNet can be tested further as real-time sensations.

Author Contributions

M.B. came up with the research topic, designing the methodology, and writing the code. M.B. and T.B.S. wrote the first version of the document. M.B., T.B.S. and A.N. were responsible for gathering resources, examining the outcome, and making final changes/corrections. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

The research investigation employed a readily accessible dataset for its analysis, devoid of any direct participation from actual patients in the study.

Data Availability Statement

The data used in the study is publicly accessed on 29 June 2023 https://data.mendeley.com/datasets/rscbjbr9sj/2. The COVID-19 dataset and Kidney Stone dataset are publicly available and cited with URLs in respective sections. Codes will be made available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Varela, M.D.; Conti, G.M.; Malka, S.; Vaclavik, V.; Mahroo, O.A.; Webster, A.R.; Tran, V.; Michaelides, M. Coats-like vasculopathy in inherited retinal disease: Prevalence, characteristics, genetics and management. Ophthalmology 2023. [Google Scholar] [CrossRef]
  2. Tariq, A.; Gill, A.Y.; Hussain, H.K. Evaluating the Potential of Artificial Intelligence in Orthopedic Surgery for Value-based Healthcare. Int. J. Multidiscip. Sci. Arts 2023, 2, 27–35. [Google Scholar] [CrossRef]
  3. Mall, P.K.; Singh, P.K.; Srivastav, S.; Narayan, V.; Paprzycki, M.; Jaworska, T.; Ganzha, M. A comprehensive review of deep neural networks for medical image processing: Recent developments and future opportunities. Healthc. Anal. 2023, 4, 100216. [Google Scholar] [CrossRef]
  4. Kumari, A.A.; Bhagat, A.; Henge, S.K.; Mandal, S.K. Automated Decision Making ResNet Feed-Forward Neural Network based Methodology for Diabetic Retinopathy Detection. Int. J. Adv. Comput. Sci. Appl. 2023, 14, 303–314. [Google Scholar] [CrossRef]
  5. Udayaraju, P.; Jeyanthi, P.; Sekhar, B. A hybrid multilayered classification model with VGG-19 net for retinal diseases using optical coherence tomography images. Soft Comput. 2023, 27, 12559–12570. [Google Scholar] [CrossRef]
  6. Narayan, V.; Mall, P.K.; Alkhayyat, A.; Abhishek, K.; Kumar, S.; Pandey, P. Enhance-Net: An Approach to Boost the Performance of Deep Learning Model Based on Real-Time Medical Images. J. Sensors 2023, 2023, 8276738. [Google Scholar] [CrossRef]
  7. Sinha, A.; Garcia, D.W.; Kumar, B.; Banerjee, P. Application of Big Data Analytics and Internet of Medical Things (IoMT) in Healthcare with View of Explainable Artificial Intelligence: A Survey. In Interpretable Cognitive Internet of Things for Healthcare; Springer: Berlin/Heidelberg, Germany, 2023; pp. 129–163. [Google Scholar]
  8. Carter, S.M.; Rogers, W.; Win, K.T.; Frazer, H.; Richards, B.; Houssami, N. The ethical, legal and social implications of using artificial intelligence systems in breast cancer care. Breast 2020, 49, 25–32. [Google Scholar] [CrossRef] [PubMed]
  9. Nahavandi, S. Industry 5.0—A human-centric solution. Sustainability 2019, 11, 4371. [Google Scholar] [CrossRef]
  10. Abou El Houda, Z.; Brik, B.; Khoukhi, L. “why should i trust your ids?”: An explainable deep learning framework for intrusion detection systems in internet of things networks. IEEE Open J. Commun. Soc. 2022, 3, 1164–1176. [Google Scholar] [CrossRef]
  11. Adadi, A.; Berrada, M. Peeking inside the black-box: A survey on explainable artificial intelligence (XAI). IEEE Access 2018, 6, 52138–52160. [Google Scholar] [CrossRef]
  12. Bhandari, M.; Shahi, T.B.; Siku, B.; Neupane, A. Explanatory classification of CXR images into COVID-19, Pneumonia and Tuberculosis using deep learning and XAI. Comput. Biol. Med. 2022, 150, 106156. [Google Scholar] [CrossRef]
  13. Sitaula, C.; Shahi, T.B. Monkeypox virus detection using pre-trained deep learning-based approaches. J. Med. Syst. 2022, 46, 78. [Google Scholar] [CrossRef]
  14. Islam, M.N.; Hasan, M.; Hossain, M.; Alam, M.; Rabiul, G.; Uddin, M.Z.; Soylu, A. Vision transformer and explainable transfer learning models for auto detection of kidney cyst, stone and tumor from CT-radiography. Sci. Rep. 2022, 12, 11440. [Google Scholar] [CrossRef] [PubMed]
  15. Subramanian, M.; Kumar, M.S.; Sathishkumar, V.; Prabhu, J.; Karthick, A.; Ganesh, S.S.; Meem, M.A. Diagnosis of retinal diseases based on Bayesian optimization deep learning network using optical coherence tomography images. Comput. Intell. Neurosci. 2022, 2022, 8014979. [Google Scholar] [CrossRef] [PubMed]
  16. Puneet; Kumar, R.; Gupta, M. Optical coherence tomography image based eye disease detection using deep convolutional neural network. Health Inf. Sci. Syst. 2022, 10, 13. [Google Scholar] [CrossRef] [PubMed]
  17. Kayadibi, İ.; Güraksın, G.E.; Köse, U. A Hybrid R-FTCNN based on principal component analysis for retinal disease detection from OCT images. Expert Syst. Appl. 2023, 230, 120617. [Google Scholar] [CrossRef]
  18. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.; Liang, H.; Baxter, S.L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F.; et al. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 2018, 172, 1122–1131. [Google Scholar] [CrossRef] [PubMed]
  19. Srinivasan, P.P.; Kim, L.A.; Mettu, P.S.; Cousins, S.W.; Comer, G.M.; Izatt, J.A.; Farsiu, S. Fully automated detection of diabetic macular edema and dry age-related macular degeneration from optical coherence tomography images. Biomed. Opt. Express 2014, 5, 3568–3577. [Google Scholar] [CrossRef] [PubMed]
  20. Kim, J.; Tran, L. Retinal disease classification from oct images using deep learning algorithms. In Proceedings of the 2021 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Melbourne, Australia, 13–15 October 2021; pp. 1–6. [Google Scholar]
  21. Li, F.; Chen, H.; Liu, Z.; Zhang, X.; Wu, Z. Fully automated detection of retinal disorders by image-based deep learning. Graefes Arch. Clin. Exp. Ophthalmol. 2019, 257, 495–505. [Google Scholar] [CrossRef]
  22. Li, F.; Chen, H.; Liu, Z.; Zhang, X.D.; Jiang, M.S.; Wu, Z.Z.; Zhou, K.Q. Deep learning-based automated detection of retinal diseases using optical coherence tomography images. Biomed. Opt. Express 2019, 10, 6204–6226. [Google Scholar] [CrossRef]
  23. Sunija, A.; Kar, S.; Gayathri, S.; Gopi, V.P.; Palanisamy, P. Octnet: A lightweight cnn for retinal disease classification from optical coherence tomography images. Comput. Methods Programs Biomed. 2021, 200, 105877. [Google Scholar]
  24. Altan, G. DeepOCT: An explainable deep learning architecture to analyze macular edema on OCT images. Eng. Sci. Technol. Int. J. 2022, 34, 101091. [Google Scholar] [CrossRef]
  25. Khan, A.; Pin, K.; Aziz, A.; Han, J.W.; Nam, Y. Optical coherence tomography image classification using hybrid deep learning and ant colony optimization. Sensors 2023, 23, 6706. [Google Scholar] [CrossRef] [PubMed]
  26. Bhowmik, A.; Kumar, S.; Bhat, N. Eye disease prediction from optical coherence tomography images with transfer learning. In Proceedings of the Engineering Applications of Neural Networks: 20th International Conference, EANN 2019, Xersonisos, Crete, Greece, 24–26 May 2019; Proceedings 20. Springer: Berlin/Heidelberg, Germany, 2019; pp. 104–114. [Google Scholar]
  27. Kermany, D.; Zhang, K.; Goldbaum, M. Labeled Optical Coherence Tomography (OCT) and Chest X-ray Images for Classification. Mendeley Data 2018, 2, 651. [Google Scholar] [CrossRef]
  28. Wong, T.T.; Yeh, P.Y. Reliable accuracy estimates from k-fold cross validation. IEEE Trans. Knowl. Data Eng. 2019, 32, 1586–1594. [Google Scholar] [CrossRef]
  29. Patel, C.; Bhatt, D.; Sharma, U.; Patel, K.; Patel, R.; Patel, A.; Bhatt, U.; Pandya, S.; Modi, K.; Cholli, N.; et al. DBGC: Dimension Based Generic Convolution Block for Object Recognition. Sensors 2022, 22, 1780. [Google Scholar] [CrossRef] [PubMed]
  30. Gramegna, A.; Giudici, P. SHAP and LIME: An evaluation of discriminative power in credit risk. Front. Artif. Intell. 2021, 4, 752558. [Google Scholar] [CrossRef] [PubMed]
  31. Zhang, Y.; Xu, F.; Zou, J.; Petrosian, O.L.; Krinkin, K.V. XAI Evaluation: Evaluating Black-Box Model Explanations for Prediction. In Proceedings of the 2021 II International Conference on Neural Networks and Neurotechnologies (NeuroNT), Saint Petersburg, Russia, 16 June 2021; pp. 13–16. [Google Scholar]
  32. Banerjee, P.; Barnwal, R.P. Methods and Metrics for Explaining Artificial Intelligence Models: A Review. Explain. AI Found. Methodol. Appl. 2023, 232, 61–88. [Google Scholar]
  33. Sharma, V.; Mir, R.N.; Rout, R.K. Towards secured image steganography based on content-adaptive adversarial perturbation. Comput. Electr. Eng. 2023, 105, 108484. [Google Scholar] [CrossRef]
  34. JTIPTJ. Chest X-ray (Pneumonia, COVID-19, Tuberculosis). Available online: https://www.kaggle.com/datasets/jtiptj/chest-xray-pneumoniacovid19tuberculosis (accessed on 29 June 2023).
  35. Khan, A.I.; Shah, J.L.; Bhat, M.M. CoroNet: A deep neural network for detection and diagnosis of COVID-19 from chest X-ray images. Comput. Methods Programs Biomed. 2020, 196, 105581. [Google Scholar] [CrossRef]
  36. Qaqos, N.N.; Kareem, O.S. COVID-19 diagnosis from chest X-ray images using deep learning approach. In Proceedings of the 2020 International Conference on Advanced Science and Engineering (ICOASE), Duhok, Iraq, 23–24 December 2020; pp. 110–116. [Google Scholar]
  37. Sitaula, C.; Hossain, M.B. Attention-based VGG-16 model for COVID-19 chest X-ray image classification. Appl. Intell. 2021, 51, 2850–2863. [Google Scholar] [CrossRef]
  38. Qadir, A.M.; Abd, D.F. Kidney Diseases Classification using Hybrid Transfer-Learning DenseNet201-Based and Random Forest Classifier. Kurd. J. Appl. Res. 2023, 7, 131–144. [Google Scholar] [CrossRef]
  39. Rajinikanth, V.; Vincent PM, D.R.; Srinivasan, K.; Ananth Prabhu, G.; Chang, C.Y. Framework to Distinguish Healthy/Cancer Renal CT Images using Fused Deep Features. Front. Public Health 2023, 11, 1109236. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Methodology.
Figure 1. Methodology.
Jimaging 09 00219 g001
Figure 2. Representative images for diseases.
Figure 2. Representative images for diseases.
Jimaging 09 00219 g002
Figure 3. Illustrative examples from the retinal image dataset are presented. Figure (a) showcasing Choroidal NeoVascularization, characterized by the presence of neovascular membranes (indicated by white arrowheads) along with associated sub-retinal fluid (marked by arrows). Figure (b) illustrates Diabetic Macular Edema, which manifests as intra-retinal fluid associated with retinal thickening (denoted by arrows). Figure (c) displays multiple instances of drusen (highlighted by arrowheads), while Figure (d) illustrates a normal, pristine retina with an undisturbed foveal structure and no signs of retinal fluid or edema.
Figure 3. Illustrative examples from the retinal image dataset are presented. Figure (a) showcasing Choroidal NeoVascularization, characterized by the presence of neovascular membranes (indicated by white arrowheads) along with associated sub-retinal fluid (marked by arrows). Figure (b) illustrates Diabetic Macular Edema, which manifests as intra-retinal fluid associated with retinal thickening (denoted by arrows). Figure (c) displays multiple instances of drusen (highlighted by arrowheads), while Figure (d) illustrates a normal, pristine retina with an undisturbed foveal structure and no signs of retinal fluid or edema.
Jimaging 09 00219 g003
Figure 4. Graphical visualization of the proposed model.
Figure 4. Graphical visualization of the proposed model.
Jimaging 09 00219 g004
Figure 5. Training and Validation Result. (a) represents training and validation accuracy (b) shows training and validation loss.
Figure 5. Training and Validation Result. (a) represents training and validation accuracy (b) shows training and validation loss.
Jimaging 09 00219 g005
Figure 6. Class-wise performance of CNN model.
Figure 6. Class-wise performance of CNN model.
Jimaging 09 00219 g006
Figure 7. ROC-AUC curve.
Figure 7. ROC-AUC curve.
Jimaging 09 00219 g007
Figure 8. The model determined the presence of Drusen in the image by analyzing the OCT image and noting a significant concentration of red pixels (scattered in central regions) in the explanatory image (second in a row), which is located in the second column.
Figure 8. The model determined the presence of Drusen in the image by analyzing the OCT image and noting a significant concentration of red pixels (scattered in central regions) in the explanatory image (second in a row), which is located in the second column.
Jimaging 09 00219 g008
Figure 9. The model’s examination of the OCT image revealed a significant number of red pixels in the explanatory image (third in a row) to suggest the presence of a healthy eye.
Figure 9. The model’s examination of the OCT image revealed a significant number of red pixels in the explanatory image (third in a row) to suggest the presence of a healthy eye.
Jimaging 09 00219 g009
Figure 10. The model determined that the OCT image exhibited indications of CNV, which is a retinal disorder, due to the significant abundance of red pixels in the third explanatory image located in the fourth column.
Figure 10. The model determined that the OCT image exhibited indications of CNV, which is a retinal disorder, due to the significant abundance of red pixels in the third explanatory image located in the fourth column.
Jimaging 09 00219 g010
Figure 11. The model determined that the OCT image indicated the presence of DME due to the significant number of red pixels observed in the fifth column of the fourth explanatory image.
Figure 11. The model determined that the OCT image indicated the presence of DME due to the significant number of red pixels observed in the fifth column of the fourth explanatory image.
Jimaging 09 00219 g011
Figure 12. Training and Validation Results for COVID-19.
Figure 12. Training and Validation Results for COVID-19.
Jimaging 09 00219 g012
Figure 13. Confusion Matrix and ROC for COVID-19.
Figure 13. Confusion Matrix and ROC for COVID-19.
Jimaging 09 00219 g013
Figure 14. The analysis conducted by the model on the X-ray image indicated elevated concentrations of red pixels in the initial explanatory image (located in the second column). These red pixels are likely to represent regions of the image that are suggestive of tuberculosis.
Figure 14. The analysis conducted by the model on the X-ray image indicated elevated concentrations of red pixels in the initial explanatory image (located in the second column). These red pixels are likely to represent regions of the image that are suggestive of tuberculosis.
Jimaging 09 00219 g014
Figure 15. The model’s prediction is based on the presence of a substantial number of red pixels in the second explanatory image, situated in the third column, suggesting that the X-ray image depicts pneumonia.
Figure 15. The model’s prediction is based on the presence of a substantial number of red pixels in the second explanatory image, situated in the third column, suggesting that the X-ray image depicts pneumonia.
Jimaging 09 00219 g015
Figure 16. The model’s forecast of pneumonia was substantiated by the elevated density of red pixels in the third explanatory image.
Figure 16. The model’s forecast of pneumonia was substantiated by the elevated density of red pixels in the third explanatory image.
Jimaging 09 00219 g016
Figure 17. The model determined that the X-ray image was classified as “Normal” because there was a notable concentration of red pixels in the fourth explanatory image.
Figure 17. The model determined that the X-ray image was classified as “Normal” because there was a notable concentration of red pixels in the fourth explanatory image.
Jimaging 09 00219 g017
Figure 18. Results for LIME segmentation for Figure 14, Figure 15, Figure 16, and Figure 17 respectively.
Figure 18. Results for LIME segmentation for Figure 14, Figure 15, Figure 16, and Figure 17 respectively.
Jimaging 09 00219 g018
Figure 19. Training and Validation Result for Kidney Stone.
Figure 19. Training and Validation Result for Kidney Stone.
Jimaging 09 00219 g019
Figure 20. Confusion Matrix and ROC for COVID-19.
Figure 20. Confusion Matrix and ROC for COVID-19.
Jimaging 09 00219 g020
Figure 21. The model found that there were a lot of red pixels in the first explanation image (second column) of the CT scan. These red pixels are likely to be areas of the image that are indicative of a tumor.
Figure 21. The model found that there were a lot of red pixels in the first explanation image (second column) of the CT scan. These red pixels are likely to be areas of the image that are indicative of a tumor.
Jimaging 09 00219 g021
Figure 22. The model’s prediction that the CT image is a cyst is supported by the high concentration of red pixels in the explanation image in the third column.
Figure 22. The model’s prediction that the CT image is a cyst is supported by the high concentration of red pixels in the explanation image in the third column.
Jimaging 09 00219 g022
Figure 23. The model’s prediction of Stone was supported by the high concentration of red pixels in the third explanation image.
Figure 23. The model’s prediction of Stone was supported by the high concentration of red pixels in the third explanation image.
Jimaging 09 00219 g023
Figure 24. The model found that the CT image is predicted as Normal as the high concentration of red pixels is located in the fourth explanation image.
Figure 24. The model found that the CT image is predicted as Normal as the high concentration of red pixels is located in the fourth explanation image.
Jimaging 09 00219 g024
Figure 25. Results for LIME segmentation for Figure 21, Figure 22, Figure 23, and Figure 24, respectively.
Figure 25. Results for LIME segmentation for Figure 21, Figure 22, Figure 23, and Figure 24, respectively.
Jimaging 09 00219 g025
Table 1. The proposed lightweight CNN Model Architecture. Note the “Param #.” represents the parameters involved in the given CNN.
Table 1. The proposed lightweight CNN Model Architecture. Note the “Param #.” represents the parameters involved in the given CNN.
Layer (Type)Output ShapeParam #
Conv2D(None, 178, 178, 16)448
MaxPooling2D(None, 89, 89, 16)0
Dropout(None, 89, 89, 16)0
Conv2D(None, 87, 87, 32)4640
MaxPooling2D(None, 43, 43, 32)0
Dropout(None, 43, 43, 32)0
Conv2D(None, 41, 41, 64)18,496
MaxPooling2D(None, 20, 20, 64)0
Dropout(None, 20, 20, 64)0
Conv2D(None, 18, 18, 128)73,856
MaxPooling2D(None, 9, 9, 128)0
Dropout(None, 9, 9, 128)0
Conv2D(None, 7, 7, 256)295,168
MaxPooling2D(None, 3, 3, 256)0
Dropout(None, 3, 3, 256)0
Flatten(None, 2304)0
Dense(None, 256)590,080
Dense(None, 4)1028
Total params.: 983,716
Trainable params.: 983,716
Non-trainable params.: 0
Table 2. Performance Metrics for Different Folds. Symbols: TA, TL, TP, TR, VA, VL, VP, VR, TeA, TeL, TeP, and TeR represent training accuracy, training loss, training precision, training recall, validation accuracy, validation loss, validation precision, validation recall, test accuracy, test loss, test precision, and test recall, respectively in percentages.
Table 2. Performance Metrics for Different Folds. Symbols: TA, TL, TP, TR, VA, VL, VP, VR, TeA, TeL, TeP, and TeR represent training accuracy, training loss, training precision, training recall, validation accuracy, validation loss, validation precision, validation recall, test accuracy, test loss, test precision, and test recall, respectively in percentages.
KTATLTPTRVAVLVPVRTeATeLTePTeR
196.310.101196.4996.1394.110.182094.3793.9994.280.197494.4394.17
296.220.101096.3996.0494.010.181994.2793.9094.180.197294.3394.07
396.310.101196.4996.1394.110.182094.3793.9994.280.197494.4394.17
496.410.101296.5996.2394.200.182294.4694.0994.370.197694.5294.26
596.510.101396.6996.3394.300.182494.5694.1994.470.197894.6294.36
696.310.113596.4996.1394.100.190894.3693.9994.270.202994.4294.16
796.810.101296.9996.6394.600.178194.8694.4994.770.201594.9294.66
896.140.117696.3295.9693.930.216094.1993.8294.100.210194.2593.99
996.640.115196.8296.4694.430.213594.6994.3294.600.197594.7594.49
1095.640.120195.8295.4693.430.218593.6993.3293.600.202593.7593.49
Average96.330.107396.5196.1594.120.192794.3894.0194.290.200294.4494.18
Table 3. Classification Report of 10th fold.
Table 3. Classification Report of 10th fold.
ClassPrecisionRecallF1-Score
Drusen0.940.950.94
Normal0.950.940.95
CNV0.950.900.92
DME0.900.960.93
Accuracy 0.94
Macro Avg0.940.940.94
Weighted Avg0.940.940.94
Table 4. Sensitivity and Specificity of 10th fold.
Table 4. Sensitivity and Specificity of 10th fold.
ClassSensitivitySpecificity
DME0.9899630.916084
NORMAL0.990960.948488
DRUSEN0.9742540.925115
CNV0.9667810.974166
Table 5. Interpretation of LIME results alongside input image and segmented image.
Table 5. Interpretation of LIME results alongside input image and segmented image.
CategoryOCT ImageLIME Interpretation
DRUSENJimaging 09 00219 i001Jimaging 09 00219 i002
NormalJimaging 09 00219 i003Jimaging 09 00219 i004
CNVJimaging 09 00219 i005Jimaging 09 00219 i006
DMEJimaging 09 00219 i007Jimaging 09 00219 i008
Table 6. Adversarial Examples.
Table 6. Adversarial Examples.
ClassOriginal ImageEpsilon = 0.1Adversarial
Prediction
DrusenJimaging 09 00219 i009Jimaging 09 00219 i010Drusen
NormalJimaging 09 00219 i011Jimaging 09 00219 i012Drusen
CNVJimaging 09 00219 i013Jimaging 09 00219 i014CNV
DMEJimaging 09 00219 i015Jimaging 09 00219 i016CNV
Table 7. Comparative Analysis of COVID-19 Result to other SOTA methods.
Table 7. Comparative Analysis of COVID-19 Result to other SOTA methods.
RefAlgorithmAccuracy (%)Parameters (Millions)
[35]CNN-based CoroNet89.633.97
[36]Custom CNN94.5334.73
[37]Attention based VGG85.43VGG-16 = 18
VGG-19 = 21.2
[12]Custom CNN95.943.7
ProposedCustom CNN97.180.983
Table 8. Comparative Analysis of the Kidney Stone dataset.
Table 8. Comparative Analysis of the Kidney Stone dataset.
RefAlgorithmAccuracy (%)Parameters (Millions)
[14]Inception V361.622.32
VGG1698.214.74
Resnet73.823.71
EANet77.026
Swin Transformers99.34.12
CCT96.544.07
[38]DenseNet201-Random Forest99.4420
[39]VGG16NB96.2614.74
DenseNet121-KNN96.6420
VGG-DN-KNN10014.74
ProposedCustom CNN99.700.983
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bhandari, M.; Shahi, T.B.; Neupane, A. Evaluating Retinal Disease Diagnosis with an Interpretable Lightweight CNN Model Resistant to Adversarial Attacks. J. Imaging 2023, 9, 219. https://doi.org/10.3390/jimaging9100219

AMA Style

Bhandari M, Shahi TB, Neupane A. Evaluating Retinal Disease Diagnosis with an Interpretable Lightweight CNN Model Resistant to Adversarial Attacks. Journal of Imaging. 2023; 9(10):219. https://doi.org/10.3390/jimaging9100219

Chicago/Turabian Style

Bhandari, Mohan, Tej Bahadur Shahi, and Arjun Neupane. 2023. "Evaluating Retinal Disease Diagnosis with an Interpretable Lightweight CNN Model Resistant to Adversarial Attacks" Journal of Imaging 9, no. 10: 219. https://doi.org/10.3390/jimaging9100219

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop