Next Article in Journal
Machine Learning Prediction of Prediabetes in a Young Male Chinese Cohort with 5.8-Year Follow-Up
Previous Article in Journal
Machine Learning for Short-Term Mortality in Acute Decompensation of Liver Cirrhosis: Better than MELD Score
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CSDNet: A Novel Deep Learning Framework for Improved Cataract State Detection

by
Lahari P.L
1,
Ramesh Vaddi
1,
Mahmoud O. Elish
2,3,
Venkateswarlu Gonuguntla
4,* and
Siva Sankar Yellampalli
1,*
1
Department of Electronics and Communication Engineering, SRM University AP, Andhra Pradesh, India
2
Computer Science Department, Gulf University for Science and Technology, Hawally 32093, Kuwait
3
GUST Engineering and Applied Innovation Research Center, Gulf University for Science and Technology, Hawally 32093, Kuwait
4
Symbiosis Centre for Medical Image Analysis, Symbiosis International (Deemed University), Pune, India
*
Authors to whom correspondence should be addressed.
Diagnostics 2024, 14(10), 983; https://doi.org/10.3390/diagnostics14100983
Submission received: 31 March 2024 / Revised: 22 April 2024 / Accepted: 25 April 2024 / Published: 8 May 2024
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Cataracts, known for lens clouding and being a common cause of visual impairment, persist as a primary contributor to vision loss and blindness, presenting notable diagnostic and prognostic challenges. This work presents a novel framework called the Cataract States Detection Network (CSDNet), which utilizes deep learning methods to improve the detection of cataract states. The aim is to create a framework that is more lightweight and adaptable for use in environments or devices with limited memory or storage capacity. This involves reducing the number of trainable parameters while still allowing for effective learning of representations from data. Additionally, the framework is designed to be suitable for real-time or near-real-time applications where rapid inference is essential. This study utilizes cataract and normal images from the Ocular Disease Intelligent Recognition (ODIR) database. The suggested model employs smaller kernels, fewer training parameters, and layers to efficiently decrease the number of trainable parameters, thereby lowering computational costs and average running time compared to other pre-trained models such as VGG19, ResNet50, DenseNet201, MIRNet, Inception V3, Xception, and Efficient net B0. The experimental results illustrate that the proposed approach achieves a binary classification accuracy of 97.24% (normal or cataract) and an average cataract state detection accuracy of 98.17% (normal, grade 1—minimal cloudiness, grade 2—immature cataract, grade 3—mature cataract, and grade 4—hyper mature cataract), competing with state-of-the-art cataract detection methods. The resulting model is lightweight at 17 MB and has fewer trainable parameters (175, 617), making it suitable for deployment in environments or devices with constrained memory or storage capacity. With a runtime of 212 ms, it is well-suited for real-time or near-real-time applications requiring rapid inference.

1. Introduction

The eye is prone to various disorders, notably cataracts, causing vision loss if untreated [1,2]. Factors like aging, diabetes, UV radiation, genetics, pollutants, eye trauma, and habits contribute to cataracts [3]. Diagnosis involves a comprehensive eye exam, and treatment includes removing the clouded lens and implanting an artificial one. Cataracts are graded based on cloudiness and vision impact, with stages as follows: grade 1, characterized by an incipient cataract with minimal vision impact; grade 2, an immature cataract featuring noticeable cloudiness and slight vision blurring; grade 3, a mature cataract significantly impairing vision, particularly at night; and grade 4, a hyper mature cataract, exhibiting lens changes, reduced vision, and potential complications such as glaucoma.
Deep learning has revolutionized eye medical imaging across various tasks [4,5,6,7,8,9], automating cataract detection and classification for precise diagnosis [9]. It aids in diabetic retinopathy screening, identifying severity for timely intervention [10], and enables early glaucoma detection through optic nerve analysis [11]. Deep learning segments retinal vessels and optic discs, assisting in vascular change analysis and glaucoma diagnosis [12]. It identifies age-related macular degeneration early and monitors eye conditions via OCT [13], analyzes corneal diseases through corneal images [14], and guides surgeries like cataract surgery precisely [15]. In telemedicine, it monitors changes for timely intervention and facilitates algorithm development through synthetic image generation [16].
Enhanced cataract state classification and detection are crucial for precise diagnosis and treatment, potentially improving patient outcomes and healthcare efficiency. Convolutional neural networks (CNNs) are commonly used for image classification and detection, benefiting from data preparation, model selection, and augmentation for improved generalization [17,18].
Several deep learning methods in the literature aim to advance cataract state detection, as outlined below. In ref. [19], the authors achieve a high training accuracy of 99.47% and validation accuracy of 97.94% using NasNet Mobile for cataract classification. Ref. [20] categorizes cataracts into normal and cataract groups, with average accuracies ranging from 91.06% to 93.50% using models like VGG19 and ResNet. Transfer learning is applied in [21], where an ensemble technique combining VGG19, ResNet101V2, and InceptionV3 achieves an F1_score of 95.90% on a test dataset. In ref. [22], the EYENET model, serving as a self-diagnosis tool for five eye disorders, achieves an accuracy of 92.3%, potentially alleviating doctor burden and enabling the rapid detection of ailments. In the literature, several pre-trained CNN models such as VGG [23], ResNet [24], DenseNet [25], Inception [26], MIRnet [27], Xception [28,29], and EfficientNet [30] have been utilized to train large image datasets, as outlined in Table 1.
However, efforts to improve cataract classification and detection have faced challenges in detecting early stages accurately, reducing overall accuracy. Variations in image quality and patient characteristics further complicate classification. Simpler models may face challenges in understanding complex data patterns, which can lead to lower performance in tasks requiring intricate relationships. Underfitting becomes a concern when the model fails to represent the data patterns effectively. Nonetheless, simpler models offer advantages in memory-constrained environments, with fewer parameters yet still-capable representation learning. This feature renders them suitable for real-time applications where rapid inference is essential.
The depicted model in Figure 1 introduces an enhanced framework of improved cataract state classification, addressing the aforementioned issues in detecting cataracts and their severity by utilizing a combination of convolutional layers, dense layers, and dropout layers. The framework comprises three main stages: dataset preparation for binary (normal or cataract) and multi-class (normal, grade 1—minimal cloudiness, grade 2—immature cataract, grade 3—mature cataract, and grade 4—hyper mature cataract) classification, data augmentation and preprocessing, and the Cataract States Detection Network (CSDNet) for the precise classification and detection of cataract states.
This proposed architecture enhances the model’s capability to discern intricate patterns within data, capturing subtle cataract indications and accurately identifying cataracts while predicting their severity. It achieves this by learning stage-specific features, thus enabling the categorization of cataract progression from early to severe stages. Evaluation metrics such as accuracy, precision, F1_score, and recall are employed for cataract state classification. Furthermore, the model is designed to be lightweight and suitable for deployment in memory-constrained environments or devices with limited storage capacity, maintaining a reasonable number of parameters for learning from data while being efficient in memory consumption and computational resources. These attributes enable the model to excel in tasks like image recognition and classification, rendering it suitable for real-time applications.
The remainder of this paper is organized as follows: The next section provides a detailed explanation of the proposed model for enhanced classification and detection, while Section 3 presents the detailed results and metrics evaluation. Section 4 offers detailed discussions and conclusions of the paper.

2. Proposed Model

The CSDNet model is designed for efficient deployment on memory-constrained devices, featuring tailored customization for enhanced detection capabilities. It employs global average pooling and reshaping layers to capture essential global features, while dropout regularization prevents overfitting and ReLu activation functions aid convergence.
The proposed model consists of 14 layers, including a flattened layer and a sigmoid activation function, providing a balance between accuracy and complexity for efficient deployment. The input layer is optimized for 224 × 224 images with three channels (RGB). Functional, dense, and dropout layers are integrated, leveraging pre-trained models for feature extraction. Functional layers integrate pre-trained models or complex sub-networks into the architecture, enabling the reuse of established architectures as building blocks. Dense layers, also known as fully connected layers, perform linear transformations on input data followed by non-linear activation functions, connecting each neuron to every neuron in the previous layer. Dropout layers prevent overfitting by randomly deactivating neurons during training, encouraging the learning of robust features and improving generalization. Activation functions introduce non-linearity to the model. They are, in general, employed to acquire high-level representations and patterns within the data. Typically found in the hidden layers of feedforward neural networks, they play a common role in capturing intricate features.
The model, as illustrated in Figure 2, begins by passing the input through a functional layer for feature extraction. The number of filters in a CNN is crucial and influenced by multiple factors. More filters enhance the model’s capacity to detect intricate patterns, beneficial for tasks needing fine detail for accurate detection. Moreover, employing more filters helps the network learn diverse feature representations per layer, facilitating the extraction of hierarchical features necessary for understanding input data. Starting with a larger number of filters allows the network to capture various low-level features initially, which are then abstracted into higher-level features in subsequent layers.
To select the optimal model configuration, a series of experiments were conducted, varying the number of blocks and filters (presented in the results section as Table 2. Model accuracy with various combinations of blocks and filters). The findings revealed that the most effective model consists of four blocks with varying filter sizes (64, 128, 256, 512) and is the proposed model. The proposed model comprises four convolutional layers, with ascending filter dimensions of 64, 128, 256, and 512, respectively. It accepts input of shape (224, 224, 3), with each convolutional layer followed by ReLU activation to maintain spatial information. Consequent to the convolutional layers, a functional layer executes convolution operations, resulting in an output shape of (7, 7, 512). Following this, three dense layers are integrated, with decreasing neuron quantities: 256, 128, and 64. Each dense layer is paired with a dropout layer (with dropout rates of 0.5, 0.2, and 0.1, respectively) to mitigate overfitting, resulting in an output (7, 7, 64). This output is then processed through a flattened layer, converting the previous output shape (7, 7, 64) into a one-dimensional array of size 3136. Finally, a dense layer with sigmoid activation, suitable for binary classification tasks, is employed. For the detection of cataract states, softmax is utilized.

3. Results and Discussion

In this section, a detailed description of the dataset chosen, data preparation, implementation details, evaluation metrics, and results of the proposed model in comparison with the existing models are discussed in detail.
Data: This study uses the Ocular Disease Intelligent Recognition (ODIR) database [31], containing structured ophthalmic data from 5000 patients, including age details and color fundus images of both eyes. Each patient record includes diagnostic keywords provided by medical professionals. The database mirrors real-world patient data collected from various hospitals and medical facilities across China by Shanggong Medical Technology Co., Ltd. Fundus images exhibit resolution variations due to different camera usage. Trained human readers meticulously annotated the dataset to ensure quality control. Patients are classified into 8 distinct categories: Normal, Diabetes, Glaucoma, Cataract, Age-related macular degeneration, Hypertension, Pathological myopia, and other diseases/abnormalities.
As the emphasis is on enhancing cataract state classification and detection, only categories Cataract (C, 1168 images) and Normal (N, 1000 images), 2168 images in total, were chosen for further analysis. On the removal of the noisy and blurry images—data cleaning, the total number was images is 2000 (C—1100 (grade 1—230; grade 2—137; grade 3—469; grade 4—264) and N—900 images).
Preprocessing: The preprocessing procedures involve the following steps: Image resizing, which standardizes all images to a uniform size to reduce computational complexity and ensure compatibility with the model’s architecture. Normalization, which adjusts pixel values of the images to a standardized range (typically 0 to 1) to stabilize the training process and potentially improve model convergence. Finally, data augmentation encompasses rotating, flipping, and shifting images to produce supplementary samples, thereby enlarging the dataset. As a result of the data augmentation, the dataset is expanded to 8000 images and is included in further analysis.
Data for training, testing, and validation: The preprocessed dataset comprising 8000 color fundus images, reflecting diverse eye conditions and demographic profiles, underwent a meticulous process of preparation and partitioning for training, testing, and validation. Through label encoding, diagnostic terms were numerically represented, with 0 denoting normal eyes and 1 indicating the presence of cataracts in binary classification. In this analysis, we explored different splitting strategies (90/05/05; 80/10/10; 70/15/15; 60/20/20) to assess classification accuracy. For instance, following an 80/10/10 split strategy, the dataset was divided, allocating 80% (6400 images) for training, 10% (800 images) for testing, and another 10% (800 images) for validation purposes.
Implementation details: All experiments were conducted on a computer with the following properties: Intel(R) core i3 processor, 7th Generation, and 8 GB RAM. We utilized Google Collab with a T4 GPU to accelerate our model training and experiments, resulting in faster processing. Preprocessing, augmentation, VGG19, ResNet 50, DenseNet 201, inception v3, MIRNet, Xception, and Efficient net B0 models were implemented using Python (Google Collab), Keras environment. The proposed model was optimized using Adam Optimizer with a learning rate of 0.001, and a batch size of 15 for memory efficiency.
Evaluation criteria: In this paper, accuracy, recall (also known as the sensitivity or true positive rate), precision (the model’s ability to properly identify positive cases), and F1_score (harmonic mean of precision and recall to provide a balanced assessment of a model’s performance) are the metrics used as evaluation indicators. They are defined as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N
R e c a l l = T P T P + F N
P r e c i s i o n = T P T P + F P
F 1 _ s c o r e = 2 ( P r e c i s i o n R e c a l l ) ( P r e c i s i o n + R e c a l l )
where T P is true positive, T N is true negative, F P is false positive, and F N is false negative.
Performance of the proposed model: In this section, we extensively analyze the performance of the proposed model. We evaluate our model alongside eight other pre-trained models for cataract classification and detection using identical datasets. We compare the experimental outcomes with state-of-the-art methods for cataract classification and detection.
First, to select the most suitable model configuration, a series of experiments were conducted using an 80/10/10 split strategy, varying the number of blocks and filters, which are tabulated in Table 2. The results show that the most effective model consists of four blocks with varying filter sizes (64, 128, 256, 512) and is the proposed model.
The proposed CSDNet, along with several pre-trained CNN models such as VGG19, ResNet50, DenseNet201, InceptionV3, MIRNet, Xception, and EfficientNetB0, were developed and tested for binary classification (cataract or normal) using a dataset comprising 8000 images (4400 cataract and 3600 normal) after preprocessing. The evaluation metrics and accuracies of all models were compared and are tabulated in Table 3—metrics comparison of existing and proposed model for the classification of cataract and normal data. Various training, testing, and validation set combinations (90/05/05; 80/10/10; 70/15/15; 60/20/20) were employed. The results indicate that the 80-10-10 split ratio consistently yielded better accuracy, precision, recall, and F1_score across most models, including the proposed CSDNet, thus becoming the chosen splitting strategy.
With the chosen model and split strategy, the cataract states were now classified using the same model, transitioning from binary classification to multi-class classification (replacing sigmoid with softmax). Each class is represented by the following number of preprocessed images: grade 1—920; grade 2—548; grade 3—1876; grade 4—1056; and Normal (N)—3600. To address class imbalance, 548 images per class (totaling 2740 images) were randomly selected, given that grade 2 had only 548 images after preprocessing. Using the 80-10-10 split ratio, the evaluation metrics and accuracies for all models, including the proposed CSDNet, in the cataract state detection task were tabulated in Table 4—metrics comparison of existing and proposed model for cataract state detection. The results indicate that the proposed CSDNet outperformed others, achieving an average cataract state detection accuracy of 98.17%.
The collective experimental findings demonstrate that the proposed method attained a binary classification accuracy of 97.24% (distinguishing normal from cataract) and an average cataract state detection accuracy of 98.17% (encompassing normal, as well as grades 1 through 4 of cataract severity). This performance matches top cataract detection methods and consistently surpasses other models in all evaluation aspects.
To assess efficiency in terms of model size, layer count, and average running time, all pretrained models were compared alongside the proposed model, and the results were tabulated in Table 5—model comparison of existing and proposed models. The findings highlight several advantages of the proposed model: 1. Model Size—it has a relatively smaller size compared to most other models listed, with only 17 MB. This compactness renders it more lightweight and adaptable for deployment in memory-limited environments or devices with constrained storage capacity; 2. Trainable parameters—despite having fewer trainable parameters than most models, it maintains a reasonable count of parameters (1, 75, 617), enabling effective learning from data; 3. Average run time—with an average run time of 212 ms, it is well-suited for real-time or near-real-time applications requiring quick inference. These characteristics make the proposed model efficient in terms of both memory consumption and computational resources while still maintaining good performance in various tasks like image recognition, classification, or other related tasks.
The confusion matrix, model accuracy, and model loss plots obtained by using the CSDNet for binary classification are depicted in Figure 3. Additionally, Figure 4 provides an example of actual versus predicted outcomes for normal and cataract classifications using the CSDNet. The training and validation accuracy are plotted over 30 epochs, displaying the model’s training progress alongside validation results for accuracy and loss metrics across epochs. The proposed model achieved a classification accuracy of 97.24% and a loss of 0.1368.
Similarly, the confusion matrix, model accuracy, and model loss plots for cataract state detection using the CSDNet are presented in Figure 5. Furthermore, Figure 6 showcases an instance of actual versus predicted states for normal and cataract conditions employing the CSDNet. The training and validation accuracy are illustrated over 30 epochs, with the curve representing the model’s training and validation performance concerning accuracy and loss metrics across epochs. The proposed model achieved an accuracy of 98.17%, with a loss of 0.0983.

4. Conclusions

In this study, we proposed a framework for classifying cataracts and detecting their states. Our proposed CSDNet achieved an accuracy of 98.17% in detecting cataract states. Comparison with pre-trained models revealed that the proposed method has improved the accuracy of cataract state detection. Furthermore, we compared our proposed CSDNet for cataract state detection with existing models and summarized the results in Table 6. The findings indicate that our model either matches or surpasses existing ones. Additionally, the lightweight nature of the CSDNet, requiring only 17 MB of memory, significantly lowers the barriers to deploying advanced diagnostic tools in low-resource settings. With a rapid inference time of 212 ms, the CSDNet holds promise for real-time diagnostic capabilities. Its high detection accuracy despite limited trainable parameters illustrates the feasibility of achieving precise diagnostics without extensive computational resources. Future work will focus on real-time implementation for the detection of cataracts and its evolution into a portable kit aimed at early cataract detection.

Author Contributions

Conceptualization, S.S.Y. and V.G.; methodology, L.P., V.G. and R.V.; software, V.G.; validation, S.S.Y. and M.O.E.; formal analysis, R.V. and L.P.; investigation, L.P.; resources, V.G.; data curation, R.V. and L.P. writing—original draft preparation, L.P. and V.G.; writing—review and editing, V.G. and R.V.; visualization, S.S.Y., V.G. and M.O.E.; supervision, S.S.Y. and V.G.; project administration, S.S.Y. and M.O.E.; funding acquisition, V.G. and M.O.E. All authors have read and agreed to the published version of the manuscript.

Funding

Financial support was provided by the “Symbiosis Centre for Medical Image Analysis, Symbiosis International (Deemed University)” and “GUST Engineering and Applied Innovation Research Center at Gulf University for Science and Technology, Kuwait” for the publication of this research.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset utilized in this research was obtained from a publicly accessible source. Access to the dataset can be obtained via the reference provided in this paper. All other findings made in this study are fully detailed within the paper itself.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Pratap, T.; Kokil, P. Computer-Aided Diagnosis of Cataract Using Deep Transfer Learning. Biomed. Signal Process. Control 2019, 53, 101533. [Google Scholar] [CrossRef]
  2. Xu, X.; Zhang, L.; Li, J.; Guan, Y.; Zhang, L. A Hybrid Global-Local Representation CNN Model for Automatic Cataract Grading. IEEE J. Biomed. Health Inform. 2019, 24, 556–567. [Google Scholar] [CrossRef] [PubMed]
  3. Varma, N.; Yadav, S.; Kant, J. A Reliable Automatic Cataract Detection Using Deep Learning. Int. J. Syst. Assur. Eng. Manag. 2023, 14, 1089–1102. [Google Scholar] [CrossRef]
  4. Ting, D.S.W.; Pasquale, L.R.; Peng, L.; Campbell, J.P.; Lee, A.Y.; Raman, R.; Tan, G.S.W.; Schmetterer, L.; Keane, P.A.; Wong, T.Y. Artificial Intelligence and Deep Learning in Ophthalmology. Br. J. Ophthalmol. 2018, 103, 167–175. [Google Scholar] [CrossRef] [PubMed]
  5. Tong, Y.; Lu, W.; Yu, Y.; Shen, Y. Application of Machine Learning in Ophthalmic Imaging Modalities. Eye Vis. 2020, 7, 22. [Google Scholar] [CrossRef] [PubMed]
  6. Balyen, L.; Peto, T. Promising Artificial Intelligence-Machine Learning-Deep Learning Algorithms in Ophthalmology. Asia-Pac. J. Ophthalmol. 2019, 8, 264–272. [Google Scholar] [CrossRef] [PubMed]
  7. Razzak, M.I.; Naz, S.; Zaib, A. Deep Learning for Medical Image Processing: Overview, challenges and the future. Lect. Notes Comput. Vis. Biomech. 2017, 26, 323–350. [Google Scholar] [CrossRef] [PubMed]
  8. Lv, Q.; Zhang, S.; Wang, Y. Deep Learning Model of Image Classification Using Machine Learning. Adv. Multimed. 2022, 2022, 3351256. [Google Scholar] [CrossRef]
  9. Junayed, M.S.; Islam, M.B.; Sadeghzadeh, A.; Rahman, S. CataractNet: An Automated Cataract Detection System Using Deep Learning for Fundus Images. IEEE Access 2021, 9, 128799–128808. [Google Scholar] [CrossRef]
  10. Yadav, S.; Kant, J. Automatic Cataract Severity Detection and Grading Using Deep Learning. J. Sens. 2023, 2023, 2973836. [Google Scholar] [CrossRef]
  11. Yadav, J.K.P.S.; Yadav, S. Computer-Aided Diagnosis of Cataract Severity Using Retinal Fundus Images and Deep Learning. Comput. Intell. 2022, 38, 1450–1473. [Google Scholar] [CrossRef]
  12. Akter, N.; Fletcher, J.; Perry, S.; Simunovic, M.P.; Briggs, N.; Roy, M. Glaucoma Diagnosis Using Multi-Feature Analysis and a Deep Learning Technique. Sci. Rep. 2022, 12, 8064. [Google Scholar] [CrossRef]
  13. Ran, A.R.; Tham, C.C.; Chan, P.P.; Cheng, C.-Y.; Tham, Y.-C.; Rim, T.H.; Cheung, C.Y. Deep Learning in Glaucoma with Optical Coherence Tomography: A Review. Eye 2020, 35, 188–201. [Google Scholar] [CrossRef] [PubMed]
  14. Keenan, T.D.L.; Chen, Q.; Agrón, E.; Tham, Y.-C.; Goh, J.H.L.; Lei, X.; Ng, Y.P.; Liu, Y.; Xu, X.; Cheng, C.-Y.; et al. DeepLensNet: Deep Learning Automated Diagnosis and Quantitative Classification of Cataract Type and Severity. Ophthalmology 2022, 129, 571–584. [Google Scholar] [CrossRef] [PubMed]
  15. Tham, Y.-C.; Goh, J.H.L.; Anees, A.; Lei, X.; Rim, T.H.; Chee, M.-L.; Wang, Y.X.; Jonas, J.B.; Thakur, S.; Teo, Z.L.; et al. Detecting Visually Significant Cataract Using Retinal Photograph-Based Deep Learning. Nat. Aging 2022, 2, 264–271. [Google Scholar] [CrossRef]
  16. Schünke, L.C.; Mello, B.; da Costa, C.A.; Antunes, R.S.; Rigo, S.J.; de Oliveira Ramos, G.; da Rosa Righi, R.; Scherer, J.N.; Donida, B. A Rapid Review of Machine Learning Approaches for Telemedicine in the Scope of COVID-19. Artif. Intell. Med. 2022, 129, 102312. [Google Scholar] [CrossRef] [PubMed]
  17. Rawat, W.; Wang, Z. Deep Convolutional Neural Networks for Image Classification: A Comprehensive Review. Neural. Comput. 2017, 29, 2352–2449. [Google Scholar] [CrossRef] [PubMed]
  18. Valente, J.; António, J.; Mora, C.; Jardim, S. Developments in Image Processing Using Deep Learning and Reinforcement Learning. J. Imaging 2023, 9, 207. [Google Scholar] [CrossRef] [PubMed]
  19. El-Baz, A. Classification of Retinal Diseases Based on OCT Images. Front. Biosci. 2018, 23, 247–264. [Google Scholar] [CrossRef] [PubMed]
  20. Thakur, K.; Kaur, M.; Kumar, Y. A Comprehensive Analysis of Deep Learning-Based Approaches for Prediction and Prognosis of Infectious Diseases. Arch. Comput. Methods Eng. 2023, 30, 4477–4497. [Google Scholar] [CrossRef]
  21. Kora, P.; Ooi, C.P.; Faust, O.; Raghavendra, U.; Gudigar, A.; Chan, W.Y.; Meenakshi, K.; Swaraja, K.; Plawiak, P.; Rajendra Acharya, U. Transfer Learning Techniques for Medical Image Analysis: A Review. Biocybern. Biomed. Eng. 2022, 42, 79–107. [Google Scholar] [CrossRef]
  22. Liesegang, T.J.; Hoskins, H.D.; Albert, D.M.; O’Day, D.H.; Spivey, B.E.; Sadun, A.A.; Parke, D.W.; Mondino, B.J. Ophthalmic Education: Where Have We Come From, and Where Are We Going? Am. J. Ophthalmol. 2003, 136, 114–121. [Google Scholar] [CrossRef] [PubMed]
  23. Very Deep Convolutional Networks for Large-Scale Image Recognition. ar5iv. Available online: https://ar5iv.labs.arxiv.org/html/1409.1556 (accessed on 10 September 2023).
  24. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.-A. Automated Melanoma Recognition in Dermoscopy Images via Very Deep Residual Networks. IEEE Trans. Med. Imaging 2017, 36, 994–1004. [Google Scholar] [CrossRef] [PubMed]
  25. Lodhi, B.; Kang, J. Multipath-DenseNet: A Supervised Ensemble Architecture of Densely Connected Convolutional Networks. Inf. Sci. 2019, 482, 63–72. [Google Scholar] [CrossRef]
  26. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar] [CrossRef]
  27. Zhang, Z.; Li, G.; Wang, L. Seismic Random Noise Suppression Based on Mirnet with Dense Feature Fusion. IEEE Geosci. Remote Sens. Lett. 2022, 19, 8028005. [Google Scholar] [CrossRef]
  28. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar] [CrossRef]
  29. Chen, L.; Li, S.; Bai, Q.; Yang, J.; Jiang, S.; Miao, Y. Review of Image Classification Algorithms Based on Convolutional Neural Networks. Remote Sens. 2021, 13, 4712. [Google Scholar] [CrossRef]
  30. Zhou, A.; Ma, Y.; Ji, W.; Zong, M.; Yang, P.; Wu, M.; Liu, M. Multi-Head Attention-Based Two-Stream EfficientNet for Action Recognition. Multimed. Syst. 2022. [Google Scholar] [CrossRef]
  31. Available online: https://www.kaggle.com/code/matthewmaddock/ocular-disease-recognition-model-training/input (accessed on 7 August 2023).
Figure 1. The proposed framework of improved cataract state classification.
Figure 1. The proposed framework of improved cataract state classification.
Diagnostics 14 00983 g001
Figure 2. CSDNet architecture.
Figure 2. CSDNet architecture.
Diagnostics 14 00983 g002
Figure 3. Confusion matrix, model accuracy, and model loss plots using the CSDNet for binary classification.
Figure 3. Confusion matrix, model accuracy, and model loss plots using the CSDNet for binary classification.
Diagnostics 14 00983 g003
Figure 4. Actual and predicted classification of normal and cataracts using the CSDNet for binary classification.
Figure 4. Actual and predicted classification of normal and cataracts using the CSDNet for binary classification.
Diagnostics 14 00983 g004
Figure 5. Confusion matrix, model accuracy, and model loss plots using the CSDNet for cataract state detection.
Figure 5. Confusion matrix, model accuracy, and model loss plots using the CSDNet for cataract state detection.
Diagnostics 14 00983 g005
Figure 6. Actual and detected cataract states using the CSDNet for cataract state detection.
Figure 6. Actual and detected cataract states using the CSDNet for cataract state detection.
Diagnostics 14 00983 g006
Table 1. Pre-trained CNN models for large image datasets (ALs—architecture layers; P—parameters (millions); INT1A—ImageNet Top-1 Accuracy (%); INT5A—ImageNet Top-5 Accuracy (%); CE—computational efficiency; R—reliability; PE—power efficiency; I—interpretability; TL—transfer learning; DA—data augmentation; HPs—hyper parameters; RTs—regularization techniques; L—low; M—moderate; H—high; NS—not specified).
Table 1. Pre-trained CNN models for large image datasets (ALs—architecture layers; P—parameters (millions); INT1A—ImageNet Top-1 Accuracy (%); INT5A—ImageNet Top-5 Accuracy (%); CE—computational efficiency; R—reliability; PE—power efficiency; I—interpretability; TL—transfer learning; DA—data augmentation; HPs—hyper parameters; RTs—regularization techniques; L—low; M—moderate; H—high; NS—not specified).
ModelALsPINT1AINT5ACERPEITLDAHPsRTs
VGG19 [23]19143.67~71~89LMLHYesYesYesYes
ResNet 50 [24]5025.6~76~93MHMMYesYesYesYes
DenseNet 201 [25]20120.0~77~93MHMMYesYesYesYes
Inception V3 [26]Variable23.8~77~93MMMHYesYesYesYes
MIRNet [27]VariableNSNSNSMMMHYesYesYesYes
Xception [28,29]Variable22.9~79~94MHMHYesYesYesYes
EfficientNet B0 [30]205.3~77~93HHHMYesYesYesYes
Table 2. Model accuracy with various combinations of blocks and filters.
Table 2. Model accuracy with various combinations of blocks and filters.
BlocksFilter SetAccuracy
316, 32, 6490.56%
432, 64, 128, 25695.41%
464, 128, 256, 51297.24%
532, 64, 128, 256, 51294.03%
Table 3. Metrics comparison of existing and proposed models for the classification of cataract and normal data.
Table 3. Metrics comparison of existing and proposed models for the classification of cataract and normal data.
Normal vs. CataractsModelAccuracyPrecisionRecall F1_Score
Training set (90%), testing set (5%), and validation set (5%)Vgg1995.410.940.970.96
ResNet 5094.030.970.920.95
DenseNet 20193.110.920.950.94
InceptionV394.030.930.960.94
MIRNet95.410.980.940.96
Xception95.410.940.980.96
EfficientNet B095.870.970.940.96
CSDNet96.790.980.970.97
Training set (80%), testing set (10%), and validation set (10%)Vgg1995.870.950.970.96
ResNet 5097.241.000.950.97
DenseNet 20190.820.861.000.93
InceptionV397.240.951.000.98
MIRNet94.500.980.910.94
Xception91.740.910.950.93
EfficientNet B098.160.971.000.98
CSDNet97.240.970.990.98
Training set (70%), testing set (15%), and validation set (15%) Vgg1992.040.920.920.92
ResNet 5092.660.900.960.93
DenseNet 20195.100.960.960.96
InceptionV395.720.930.990.96
MIRNet92.660.950.910.93
Xception91.130.930.900.91
EfficientNet B095.710.950.970.96
CSDNet92.350.890.980.93
Training set (60%), testing set (20%), and validation set (20%)Vgg1995.070.950.970.96
ResNet 5094.030.940.950.94
DenseNet 20194.270.930.980.95
InceptionV392.430.950.910.93
MIRNet93.810.950.930.94
Xception87.610.880.890.88
EfficientNet B094.180.940.950.95
CSDNet95.180.940.970.96
Table 4. Metrics comparison of existing and proposed models for cataract state detection.
Table 4. Metrics comparison of existing and proposed models for cataract state detection.
Normal, Grade 1 to 4ModelAccuracyPrecisionRecallF1_Score
Training set (80%), testing set (10%), and validation set (10%)Vgg1997.04Normal:0.98
Grade 1:0.96
Grade 2:0.96
Grade 3.0.96
Grade 4:0.99
0.99
0.96
0.97
0.96
0.97
0.99
0.96
0.97
0.96
0.98
ResNet 5096.81Normal:0.99
Grade 1:0.97
Grade 2:0.95
Grade 3.0.96
Grade 4:0.97
0.98
0.97
0.98
0.95
0.96
0.98
0.97
0.97
0.95
0.96
DenseNet 20190.74Normal:0.95
Grade 1:0.87
Grade 2:0.88
Grade 3:0.90
Grade 4:0.94
0.93
0.90
0.91
0.89
0.90
0.94
0.89
0.89
0.89
0.92
InceptionV396.17Normal:0.98
Grade 1:0.98
Grade 2:0.93
Grade 3:0.95
Grade 4:0.97
0.97
0.96
0.98
0.95
0.95
0.97
0.97
0.96
0.95
0.96
MIRNet93.54Normal:0.97
Grade 1:0.91
Grade 2:0.92
Grade 3:0.93
Grade 4:0.95
0.95
0.95
0.96
0.90
0.92
0.96
0.93
0.94
0.91
0.93
Xception91.23Normal:0.93
Grade 1:0.91
Grade 2:0.88
Grade 3:0.90
Grade 4:0.95
0.94
0.90
0.91
0.90
0.91
0.94
0.90
0.89
0.90
0.93
EfficientNet B097.81Normal:0.98
Grade 1:0.97
Grade 2:0.99
Grade 3.0.95
Grade 4:0.98
0.98
0.96
0.99
0.97
0.97
0.98
0.97
0.99
0.96
0.97
CSDNet98.17Normal:1.00
Grade 1:0.97
Grade 2:0.98
Grade 3.0.97
Grade 4:0.98
0.99
0.98
0.99
0.97
0.97
0.99
0.98
0.99
0.97
0.97
Table 5. Model comparison of existing and proposed models.
Table 5. Model comparison of existing and proposed models.
ModelLayersTrainable ParametersModel SizeAverage Run Time
VGG1919200, 49, 473574 MB267 ms
ResNet 50503, 01, 05798 MB118 ms
DenseNet 201201488, 85, 50580 MB185 ms
MIRnet453, 93, 217128 MB276 ms
Inception V34851, 20192 MB226 ms
Xception7110, 49, 60188 MB207 ms
EfficientNet B02146, 56, 38529 MB291 ms
CSDNet141, 75, 61717 MB212 ms
Table 6. Accuracy comparisons of existing and proposed models for cataract state detection.
Table 6. Accuracy comparisons of existing and proposed models for cataract state detection.
ModelDatasetAccuracy
Cataract detection using deep learning [3]Mixed databases and internet images92.7%
Early cataract detection using deep learning [10]1600 images93.10%
Computer-aided cataract severity diagnosis using pre-trained CNNs for feature extraction [11]Online platforms96%
CSDNet—proposed model with fewer layers and runtimeOpen-source dataset [31]98.17%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

P.L, L.; Vaddi, R.; Elish, M.O.; Gonuguntla, V.; Yellampalli, S.S. CSDNet: A Novel Deep Learning Framework for Improved Cataract State Detection. Diagnostics 2024, 14, 983. https://doi.org/10.3390/diagnostics14100983

AMA Style

P.L L, Vaddi R, Elish MO, Gonuguntla V, Yellampalli SS. CSDNet: A Novel Deep Learning Framework for Improved Cataract State Detection. Diagnostics. 2024; 14(10):983. https://doi.org/10.3390/diagnostics14100983

Chicago/Turabian Style

P.L, Lahari, Ramesh Vaddi, Mahmoud O. Elish, Venkateswarlu Gonuguntla, and Siva Sankar Yellampalli. 2024. "CSDNet: A Novel Deep Learning Framework for Improved Cataract State Detection" Diagnostics 14, no. 10: 983. https://doi.org/10.3390/diagnostics14100983

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop