Next Article in Journal
Colposcopy Accuracy and Diagnostic Performance: A Quality Control and Quality Assurance Survey in Italian Tertiary-Level Teaching and Academic Institutions—The Italian Society of Colposcopy and Cervico-Vaginal Pathology (SICPCV)
Next Article in Special Issue
Clinical Research of Lupus Retinopathy: Quantitative Analysis of Retinal Vessels by Optical Coherence Tomography Angiography in Patients with Systemic Lupus Erythematosus
Previous Article in Journal
An Adaptive Early Stopping Technique for DenseNet169-Based Knee Osteoarthritis Detection Model
Previous Article in Special Issue
Deep Learning in Optical Coherence Tomography Angiography: Current Progress, Challenges, and Future Directions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Counteracting Data Bias and Class Imbalance—Towards a Useful and Reliable Retinal Disease Recognition System

by
Adam R. Chłopowiec
1,†,
Konrad Karanowski
1,*,†,
Tomasz Skrzypczak
2,
Mateusz Grzesiuk
1,
Adrian B. Chłopowiec
1 and
Martin Tabakov
1
1
Department of Artificial Intelligence, Wroclaw University of Science and Technology, Wybrzeże Wyspianskiego 27, 50-370 Wroclaw, Poland
2
Faculty of Medicine, Wroclaw Medical University, Wybrzeże Ludwika Pasteura 1, 50-367 Wroclaw, Poland
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2023, 13(11), 1904; https://doi.org/10.3390/diagnostics13111904
Submission received: 27 March 2023 / Revised: 22 May 2023 / Accepted: 25 May 2023 / Published: 29 May 2023
(This article belongs to the Special Issue Deep Learning Applications in Ophthalmology)

Abstract

:
Multiple studies presented satisfactory performances for the treatment of various ocular diseases. To date, there has been no study that describes a multiclass model, medically accurate, and trained on large diverse dataset. No study has addressed a class imbalance problem in one giant dataset originating from multiple large diverse eye fundus image collections. To ensure a real-life clinical environment and mitigate the problem of biased medical image data, 22 publicly available datasets were merged. To secure medical validity only Diabetic Retinopathy (DR), Age-Related Macular Degeneration (AMD) and Glaucoma (GL) were included. The state-of-the-art models ConvNext, RegNet and ResNet were utilized. In the resulting dataset, there were 86,415 normal, 3787 GL, 632 AMD and 34,379 DR fundus images. ConvNextTiny achieved the best results in terms of recognizing most of the examined eye diseases with the most metrics. The overall accuracy was 80.46 ± 1.48. Specific accuracy values were: 80.01 ± 1.10 for normal eye fundus, 97.20 ± 0.66 for GL, 98.14 ± 0.31 for AMD, 80.66 ± 1.27 for DR. A suitable screening model for the most prevalent retinal diseases in ageing societies was designed. The model was developed on a diverse, combined large dataset which made the obtained results less biased and more generalizable.

1. Introduction

According to the first World Report on Vision issued by the World Health Organization (WHO) in 2019, approximately 2.2 billion people had vision impairment or blindness, globally [1]. This number is expected to rise because of the growth of the global population and the changes in its age structure [2]. The soaring work effort associated with the ageing population is an overwhelming problem for the limited number of eye care providers [3,4]. Efficiency and effectiveness enhancements should be a fundamental response to a projected undersupply of eye care providers [4].
Recent research has proved that deep learning systems could be useful in delivering patient care in a real-world setting [5]. Multiple satisfactory performances of artificial intelligence models for the automated detection of ocular diseases were reported [5,6,7,8,9]. Clinically useful models should differentiate the most distressing diseases: diabetic retinopathy (DR), glaucoma (GL) and age-related macular degeneration (AMD) [2,10] from a healthy fundus, with high sensitivity and specificity. These diseases are prevalent in ageing populations, which makes them suitable targets for a screening system [1,2,10]. Recently, there have been several multiclass models published that at least partially meet these conditions [11,12,13,14,15,16,17,18]. However, all these models had multiple limitations.
Most of the published multiclass models were developed on a single dataset [11,13,15,16,17], mainly the Ocular Disease Intelligent Recognition (ODIR) database [13,15,17,19]. This could lead to a potential bias in the development of machine learning models. A single database is often a survey of a certain population, collected with a small number of cameras in several medical centers by a limited number of investigators. Data gathered in similar environments, or a single process, may not apply to other clinics due to different cameras, ethnicity, or an image acquisition technique. These models are not generalizable to the overall patient population. One of the most effective strategies to mitigate these biases is to compile a large-scale, multiethnic dataset that would be representative and would simulate a real-world environment for model training [20]. The collection of such a dataset would contribute to better accuracy and fairness in the decision-making process. Such an approach was partially adopted by previous works [12,14] although the clarity of data selection and quantity of merged datasets could still be improved.
Class imbalanced datasets occur in many real-world applications where class distributions of data are highly imbalanced [21]. Many classification learning algorithms have lower predictive accuracy for infrequent classes [21]. Models misclassify diseases with lower prevalence in the retinal images database. Merging multiple different datasets could even potentiate this issue. Due to this imbalance, the accuracy of detection or classification of disease is relatively low [15]. Most of the published studies [11,12,14,16,17] did not address the problem, which could influence the results. Common techniques for handling class imbalance problems involve reweighting, resampling and other algorithmic solutions [22,23]. Applying them to a large dataset could help in the recognition of less prevalent diseases.
Three out of the eight most recently published works [9,11,14] utilized private datasets. These are often formally available upon correspondence with a reasonable request. Potentially, these data could never be made available to the public. Research transparency could be put under question, as these studies may not be reproducible due to data unavailability.
Almost all published models included cataracts as a retinal disease [11,12,13,14,15,17,18]. A cataract is a cloudification of the natural intraocular lens and is not classified as a retinal disorder by the medical literature [24] In a cataract, the fundus image is not visible or is heavily distorted when photographed [24]. It seems reasonable to assume that the usage of such images in multiclass model development aimed at retinal diseases has influenced the results and has no utility in the screening and diagnostic process.
Although the assessment of the retinal fundus in both myopia and hypertensive retinopathy may have some usefulness in routine patient screening, no medical guidelines recommend this in clinical practice. Inclusion of these diseases in multiclass models developed by multiple previous investigators [11,13,14,15] could lead to unnecessary class proliferation, influence results and lead to lower screening utility. Similarly, the inclusion of relatively rare diseases like retinitis pigmentosa had a limited purpose in model development. We assumed that the perfect screening multiclass model should be focused on the most common retinal diseases that distress whole nations.
The primary aim of this study was to create an image recognition model for retinal disease screening in ageing, developed countries. The model was developed on one cumulative dataset and differentiated DR, AMD and GL from a normal eye fundus for the best clinical utility. The created database utilized multiple types of fundus cameras, evaluated by various retinal experts, and represents multiple nationalities, which approximates to the true real-world environment. This results in mitigation of the data bias problem. The utilized data had clear selection criteria and usage of only publicly available datasets made our experiment reproducible. The secondary aim was to address the problem of class imbalance which is a result of merging multiple different large datasets. To achieve that we proposed to combine transfer learning, loss function weighting and two-stage learning techniques.

2. Materials and Methods

To ensure our database minimizes the problem of biased medical image data, we collected and merged 22 publicly available fundus datasets containing images of any of the diseases classified in our paper. We selected only the data strictly related to the diagnostic process of the pathologies considered. Such a dataset consists of fundus images obtained from multiple hospitals and clinical centers all around the world, providing data from various ethnic and demographic groups. Such data contain noise, overexposure, underexposure and other visual artifacts, as well as perfectly prepared fundus images. Similar artifacts may be commonly encountered in hospitals due to human or hardware errors. The images were taken with various cameras, mydriatic and non-mydriatic. Such a wide range of images provides the least biased and most real-world adjusted clinical usage database that has been collected in the studies up to the present, consisting only of public data, which has been properly filtered for selected pathologies and their diagnostic process. Therefore, unlike studies using single public or private datasets, which are possibly biased, we provide the most reliable results for the task of classification of fundus diseases.
In our experiments, we had to tackle important problems related to medical image classification in general. We used state-of-the-art models (ConvNext [25] RegNet [26] ResNet [27]) employed in computer vision and verified their accuracy on biomedical data. Further, we present data augmentation methods used to avoid overfitting which is a common problem in the domain [28,29,30]. To address the problem of class imbalance we split the dataset into two parts: pre-training and fine-tuning. Splitting the data into train, validation and test sets is described in the section Fine-Tuning. Our study workflow is presented in Figure 1, Figure 2 and Figure 3.

2.1. Models

We chose Convolutional Neural Networks (CNNs) for fundus image classification as widely used and well-performing models in image recognition tasks. CNNs consist of two main parts:
  • A feature extractor built mostly with convolutional layers, used to capture increasingly abstract image features, which are then compressed into a vector, called feature-embedding, during the process.
  • A classifier containing mainly dense, fully connected layers, responsible for the classification of a feature-embedding vector.
In our experiments we decided to use recently published state-of-the-art models of CNNs: ConvNext [25] and RegNet [26] and compare their performance to the most-used architecture in image classification tasks—ResNet [13,27,31,32]. ConvNext architecture was inspired by Hierarchical Transformers [33]. It modernizes ResNet by employing various macro and micro design choices from Transformers and other popular CNNs like ResNext [34] or MobileNetV2 [35]. RegNet architectures are a family of CNNs that come from the progressively designed RegNet design space. They have proved to be effective in many computational regimes [26,36]. Current state-of-the-art architectures are understudied in the biomedical domain, although recent studies prove their potential in diverse applications [36,37,38,39,40]. Therefore, we found it valuable to verify their superiority over commonly used architectures in medical image classification. The most widely used architecture of ResNet is ResNet50, which we found suitable for the data we collected. To match its size, we chose ConvNextTiny and RegNetY3_2gf. All architectures were imported from the torchvision package [41].

2.2. Data Augmentation

Lack of data is a common problem for applications of deep learning techniques in the biomedical domain [42,43,44,45]. Therefore, we decided to use data augmentation, with a library provided by Buslaev et al. [46], to cover a larger space of possible inputs to our networks to increase robustness. Fundus images in real-world cases are transformed affinely—images are often rotated or inverted. Moreover, a natural characteristic of medical images is underexposure or overexposure due to hardware or human mistakes [47]. Images from different databases come with a range of resolutions, so there was a need to standardize their size. Taking such features into consideration, we decided to use the transformations described in Table 1. We have additionally used cutouts for regularization [48]. No data augmentation was used during the validation or testing phase.

2.3. Model Training

Data imbalance is a common problem in medical image classification. Naturally, some diseases are rare and difficult to classify, or data are collected from limited sources due to data collection costs or law-related issues. In such cases, data imbalance occurs. The problem is potentiated when compiling a large and diverse dataset from many smaller datasets. We proposed to use transfer learning and two-stage learning to better adjust our models to fundus images classification task. A two-stage learning procedure consists of pre-training a model on excess domain data and fine-tuning on thresholded data. The procedure described is similar to the two-phase learning reported by Johnson et al. [23]. Although it differs in the way it defines both stages, two-phase learning first pre-trains a model with thresholded data and then fine-tunes it using all data. For the pre-training part, we selected an excessive amount of normal and diabetic retinopathy images over the threshold of the cardinality of glaucoma images. The fine-tuning part consisted of the remaining normal, diabetic retinopathy, AMD, and glaucoma images. The summary of the data split is presented in Table 2. Such data division allows us to adjust a model to the domain problem with excess data from the major classes, reducing general overfitting to them during fine-tuning, by matching their cardinality with minor classes. The pre-training dataset was used in the pre-training phase and the fine-tuning dataset was used in the fine-tuning phase. We used Weights & Biases [49] for experiment tracking and visualizations to develop insights for this paper.

2.4. Pre-Training

In the pre-training phase, we used ImageNet-1K pre-trained models. We removed the fully-connected layer and replaced it with a new, randomly initialized one which had only two outputs—for diabetic retinopathy and normal image predictions. We froze half of the CNN layers to use the pre-trained feature-extraction abilities. Next, we trained each model with early stopping with patients of 5 epochs, monitoring validation-set loss. For the optimizer we chose Radam [50] with a learning rate 3−4, batch size 32 and weight decay of 1−5. To further tackle the problem of class imbalance we decided to use weighted cross entropy loss with weights 1 and 2 for Normal and Diabetic Retinopathy classes, respectively. We used a cosine-annealing learning rate scheduler [51], with Tmax = 20, ηmin = 10−5 and ηmax = 3 × 10−4.

2.5. Fine Tuning

From the model obtained in the fine-tuning phase, we removed the fully connected layer and replaced it with a new, randomly initialized one which had four outputs, unlike in pre-training. Similarly, we froze half of the convolutional layers of the model. To perform an unbiased evaluation, we trained our models in a 10-fold cross-validation process. We trained each model 10 times, every time choosing a different part of the dataset for the test set, another for the validation set and the rest for the train set. The experiments were performed with the same hyperparameters as in the pre-training phase, except for the weights used in cross-entropy—here they are equal to 1, 0.9, 1.5, 1.2 for normal, glaucoma, AMD, and diabetic retinopathy classes, respectively. We report the average results of all runs for each model.

2.6. Verification of Other Resampling Methods

Resampling methods are widely used in the literature on class imbalance [23,52,53]. To present a fair comparison and verify the results provided on data with mitigated bias we performed experiments using other resampling methods, namely: random minority oversampling (ROS) and random majority undersampling (RUS). Similarly to the procedure described in the Fine Tuning section, we trained models in a 10-fold cross-validation process. To maintain comparability with the outcomes of two-stage learning and align the class ratios for validation and test sets with our Fine Tuning phase, we applied a threshold during each cross-validation iteration for validation and test folds. This threshold ensured that the number of normal and diabetic retinopathy images matched the cardinality of glaucoma images. We used the same hyperparameters as in our Fine Tuning phase. All experiments were performed using the ConvNextTiny architecture pretrained on ImageNet-1K.

3. Results

3.1. Dataset

In the resulting dataset, there are 86,415 normal, 3787 glaucoma, 632 AMD and 34,379 diabetic retinopathy fundus images. The summary and medical characteristics of the datasets are presented in Table 3.
Most datasets were annotated by experts, except four for which the data acquisition process was not described: APTOS 2019 Blindness Detection Dataset [55], Cataract [56], Machine learning for glaucoma [60] and BAIDU: iChallenge-AMD [62].

3.2. Evaluation Criteria

In our experiments, to leverage the advantage of a diverse real-world dataset we report mean and standard deviation over 10 runs in a 10-fold cross-validation process, therefore ensuring that every part of the dataset was used for evaluation. We used 5 metrics for every class: Accuracy, F1-Score, Sensitivity, Specificity, and AUC, and then we also averaged them across classes and reported the overall accuracy. For class-specific metrics, we used the one-versus-rest technique. Such a wide set of metrics allows a thorough examination of the models’ performance with respect to every disease [73,74].

3.3. Performance

In Table 4 we present the results of our experiments. ConvNextTiny achieved the best results in terms of recognizing most of the eye diseases examined with the most metrics.
It specifically excels over ResNet50 in the F1-Score for AMD with a difference of 1.2 pp. This proves the purposefulness of choosing modern state-of-the-art architectures for medical experiments. The ResNet50 model achieved the best results at recognizing glaucoma. RegNetY3_2gf scored the worst results at recognizing every disease with respect to the most metrics. Figure 4 summarizes the performance of each model with ROC curves for all diseases with their respective standard deviation. These curves show similar trends for all diseases across all models. ConvNextTiny achieved higher results than ResNet50 with an AUC of 90.64 and 91.65 for normal and diabetic retinopathy images, respectively.

3.4. Comparison of Resampling Methods

The results of experiments with other resampling methods are presented in Table 5. As in the previous experiments, we report mean and std over the cross-validation process. ROS performed the best with respect to most metrics. Most notably it achieved a difference of 0.7 pp. for the average F1-Score over two-stage learning and of 7.81 pp. over RUS. RUS achieved the worst results. Worth noting also is the difference in the average AUC between the tested methods. Two-stage learning achieved the same results as ROS and a 2.89 pp. higher score than RUS. Random minority oversampling is a technique that requires a lot of computer power due to the increased size of the training set. Therefore, it may not be feasible in all scenarios, especially for a hyperparameter search procedure. Two-stage learning, while still performing well, requires the model to be pre-trained on excess data only once and then a hyperparameter search can be performed using the thresholded dataset. Random majority undersampling requires less computer power, although because of voluntarily discarding data it achieves worse results in comparison.

3.5. Comparison to Other Recent Models

In Table 6 we compare the results of our experiments to other works. Goals and test sets used across the works make the results not directly comparable. Previous studies reported the results with different metrics, which made them difficult to compare with each other and our study.

4. Discussion

The authors presented a model trained for retinal disease screening in the ageing societies of developed countries. The best utilized architecture (ConvNextTiny) reached 80.46 ± 1.48 overall accuracy, with average 81.20 ± 2.26 sensitivity and 92.96 ± 0.55 specificity. It was reported that ophthalmic consultants detect retinal diseases with 89% sensitivity and 86% specificity when relying on eye fundus photographs [75]. The presented model had a lower sensitivity and higher specificity than ophthalmologists, however these benchmarks proved its potential clinical utility. An average AUC of 95.10 ± 0.36 classified our model as an acceptable screening method with excellent classification performance [76]. The utilized dataset potentially contributed little to the result. Most of the database consisted of poor-quality retinal images, often blurred or distorted, annotated by different experts according to various guidelines. This could lead to ambiguous interpretation for each class in the dataset. However, this diversity gave a better approximation of a real clinical setting. The results were more reliable and generalizable to the true screening process.
Despite a lack of certain comparability, the authors attempted to compare the AUC between the presented model and the most recent studies. Although AUC remains the most reliable measure of the learning algorithm’s performance [77], only three studies reported this benchmark [11,14,16]. The model outperformed all three models in GL and AMD classification. The AUC for DR was higher than presented by Han et al. [14] and Bulut et al. [11], but Li et al. [16] obtained a better result.
Our model not only presented an acceptable performance, but also was the first that truly approximates to the real-world environment. Based on one of the largest datasets, it made the received results less biased and more generalizable than in previously published papers. The authors merged multiple datasets from around the world into one cumulative dataset. This minimized bias from the single-image acquisition process, ethnicity, or limited cameras models. The created model could be used in multiple clinics, located in distant places and which could be using different equipment, without a need for additional fine-tuning or calibration. This approach was partially adopted by Chellaswamy et al. [12] and Han et al. [14]. Chellaswamy et al. [12] merged 5 publicly available datasets and extracted with an undescribed method a limited number of fundus images for each of the analyzed diseases. As a result, the final dataset was relatively small and potentially biased due to the unknown method of image selection. Han et al. [14] combined 6 publicly available with 2 private datasets and achieved a large and diverse ophthalmic image collection. However, there was the potential for an even more diverse large-scale database, as 94 public, open-access, fully downloadable ocular image collections were available at the time when Han Y et al. [14] conducted their research [15]. Han et al. [14] utilized 2 proprietary image collections, thus full transparency of the model’s development could not be guaranteed. Neither the composition nor the collection process of the datasets was described in this paper. Moreover, these 2 private databases were collected from the Chinese population, which did not increase the overall ethnical diversity.
The presented study is the first study that only utilizes multiple public data sets. This made the presented findings fully reproductible by the scientific community. Previous models trained on multiple datasets have always encompassed at least one proprietary image collection. While Han Y et al. [14] mixed public and private collections, Bulut et al. [11] and Li et al. [16] developed their models exclusively with private datasets.
To date, this has been first study to address the problem of class imbalance in a large-scale database of retinal fundus images. The ODIR dataset, the most frequently utilized retinal image collection in published multiclass models [12,13,15,17] has severe class imbalance problems [13]. It seems reasonable to assume that merging multiple different datasets into one large one could even potentiate this issue. Our model exhibited an AUC > 90 in all included classes, despite large discrepancies in the number of images. The highest sensitivity and specificity were received for AMD and GL. Significantly lower benchmarks were reported for normal eye fundus and DR. Normal and DR had the highest shares in the final dataset, significantly greater than AMD and GL. The potential explanation of these findings is that the vast majority of normal eye fundus and DR images come from the EyePACS dataset, which describes its images as real-world data that may include noise in both the images and labels, artifacts, under- and overexposure [61]. Therefore, robust classification for this data may have proven the most challenging. Yet these conditions and the extensive cross-validation process in the fine-tuning stage of the model’s development made the received results the most reliable among recently published models. Gour et al. [13] partially approached the difficulty of class imbalance in the ODIR dataset. Although Gour et al. [13] supported their research with an analysis of class-wise performance, the developed model still showed higher sensitivity and accuracy for diseases with the highest prevalence in the dataset [13]. The model correctly classified fundus images of healthy retinas and glaucoma but failed to recognize other classes such as diabetic retinopathy or AMD [13]. Only one study aimed to address the problem of class imbalance in a dataset of retinal images [15]. Khan et al. [15] with unknown selection criteria created a balanced training set for the VGG-19 architecture model and utilized only the ODIR database. It cannot be excluded that the extraction process was biased, e.g., by selecting images with the highest quality and aimed to achieve the highest possible model performance.
Aside from advances in computer science, the presented model brought some novelty into the medical field. This was the first model targeted at the most distressing retinal diseases in ageing societies. Excellent AUC values for GL, AMD and DR proved its potential screening utility. The overall accuracy of 80.46 ± 1.48 meets the performance requirements for routine screening tests in medicine [72]. This model has been the first to be trained on medically relevant diseases. Apart from cataracts, which are not a retinal disease [24], the authors did not include rare diseases or ones irrelevant to screening such as retinitis pigmentosa. The inclusion of multiple less prevalent diseases in previous research [11,13,14,15] potentially decreased the screening utility of those models. Creating a model with multiple various diseases may be a curious academic problem. However, due to limited data availability and the tedious process of its collection, the creation of a real-world deep learning model with real clinical application should be restricted only to the most prevalent and distressing diseases, such as GL, AMD, and DR.
The present study has multiple limitations. Firstly, the developed model lacks a class that would signify “other” conditions—elderly people could suffer multiple other retinal diseases than GL, AMD, DR, and diseases could overlap with each other. However, public datasets include a limited number of classes of retinal disorders. Due to the almost infinite possibilities of “other” diseases, the model was simplified to these three most distressing diseases. Furthermore, the model’s performance was not validated by ophthalmologists. It is still uncertain whether the presented performance is comparable to that from a healthcare professional. The authors could not assume that retinal images from different datasets had consistent image classification. Retinal classification guidelines vary between countries, and even partial assessment of the final dataset by an experienced physician could be beneficial. Finally, authors did not have access to some datasets, which limited the number of images utilized. This could influence the final performance of the developed model.

5. Conclusions

This work presents classification results for the most distressing and screening-relevant retinal diseases: diabetic retinopathy, glaucoma and age-related macular degeneration, on the basis of multiple publicly available datasets, without performing an evaluation of private datasets gathered in controlled environments. Availability of the data and clear selection criteria ensured reproducibility of the results. The achieved results classified the developed model as a useful screening method and the data utilized made it more reliable. Merging multiple datasets mitigated the data bias problem. A class imbalance problem, potentiated because of dataset merging, was addressed via transfer learning, loss function weighting and two-stage learning procedures. Such a model can enhance the efficiency and effectiveness of eye care providers. This research fills the gap in the literature on multiclass models and contributes to improving the diagnosis and treatment of retinal diseases.

Author Contributions

Conceptualization, T.S. and K.K.; methodology, T.S., K.K., A.R.C. and A.B.C.; software, K.K., A.B.C., A.R.C., M.G. and M.T.; validation, K.K., A.B.C., A.R.C., M.G. and M.T.; formal analysis, K.K., A.R.C., M.G. and A.B.C.; investigation, K.K., A.R.C., M.G. and A.B.C.; resources, K.K., T.S., A.R.C., M.G. and A.B.C.; data curation, A.R.C., A.B.C. and M.G.; writing—original draft preparation, T.S., K.K., A.R.C., M.G. and A.B.C.; writing—review and editing, T.S., K.K., A.R.C., M.G. and A.B.C.; visualization, K.K. and M.G.; supervision, M.T.; project administration, K.K., A.R.C. and M.G.; funding acquisition M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the statutory funds of the Department of Artificial Intelligence, Wroclaw University of Science and Technology.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Utilized datasets were public, while preparing this manuscript. To access any dataset please refer to relevant citation for link or guideline. All data enquires should be mailed to [email protected].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. WHO. Blindness and Vision Impairment. 2021. Available online: https://www.who.int/news-room/fact-sheets/detail/blindness-and-visual-impairment (accessed on 13 October 2022).
  2. Bourne, R.R.; Flaxman, S.R.; Braithwaite, T.; Cicinelli, M.V.; Das, A.; Jonas, J.B.; Zheng, Y. Magnitude, temporal trends, and projections of the global prevalence of blindness and distance and near vision impairment: A systematic review and meta-analysis. Lancet Glob. Health 2017, 5, e888–e897. [Google Scholar] [CrossRef]
  3. Buchan, J.C.; Norman, P.; Shickle, D.; Cassels-Brown, A.; MacEwen, C. Failing to plan and planning to fail. Can we predict the future growth of demand on UK Eye Care Services? Eye 2019, 33, 1029–1031. [Google Scholar] [CrossRef] [PubMed]
  4. Lee, P.P.; Hoskins, H.D., Jr.; Parke, D.W., III. Access to Care: Eye Care Provider Workforce Considerations in 2020. Arch. Ophthalmol. 2007, 125, 406–410. [Google Scholar] [CrossRef] [PubMed]
  5. Lin, D.; Xiong, J.; Liu, C.; Zhao, L.; Li, Z.; Yu, S.; Wu, X.; Ge, Z.; Hu, X.; Wang, B.; et al. Application of Comprehensive Artificial intelligence Retinal Expert (CARE) system: A national real-world evidence study. Lancet Digit. Health 2021, 3, e486–e495. [Google Scholar] [CrossRef] [PubMed]
  6. Burlina, P.M.; Joshi, N.; Pacheco, K.D.; Liu, T.Y.A.; Bressler, N.M. Assessment of deep generative models for high-resolution synthetic retinal image generation of age-related macular degeneration. JAMA Ophthalmol. 2019, 137, 258–264. [Google Scholar] [CrossRef]
  7. Gulshan, V.; Rajan, R.; Widner, K.; Wu, D.; Wubbels, P.; Rhodes, T.; Whitehouse, K.; Coram, M.; Corrado, G.; Ramasamy, K.; et al. Performance of a deep-learning algorithm vs manual grading for detecting diabetic retinopathy in India. JAMA Ophthalmol. 2019, 137, 987–993. [Google Scholar] [CrossRef] [PubMed]
  8. Milea, D.; Najjar, R.P.; Jiang, Z.; Ting, D.; Vasseneix, C.; Xu, X.; Biousse, V. Artificial intelligence to detect papilledema from ocular fundus photographs. N. Engl. J. Med. 2020, 382, 1687–1695. [Google Scholar] [CrossRef]
  9. Son, J.; Shin, J.Y.; Kim, H.D.; Jung, K.H.; Park, K.H.; Park, S.J. Development and validation of deep learning models for screening multiple abnormal findings in retinal fundus images. Ophthalmology 2020, 127, 85–94. [Google Scholar] [CrossRef]
  10. Taylor, H.R.; Keeffe, J.E. World blindness: A 21st century perspective. Br. J. Ophthalmol. 2001, 85, 261–266. [Google Scholar] [CrossRef]
  11. Bulut, B.; Kalın, V.; Güneş, B.B.; Khazhin, R. Deep Learning Approach For Detection Of Retinal Abnormalities Based On Color Fundus Images. In Proceedings of the 2020 Innovations in Intelligent Systems and Applications Conference (ASYU), Istanbul, Turkey, 15–17 October 2020; pp. 1–6. [Google Scholar]
  12. Chellaswamy, C.; Geetha, T.S.; Ramasubramanian, B.; Abirami, R.; Archana, B.; Bharathi, A.D. Optimized Convolutional Neural Network based Multiple Eye Disease Detection and Information Sharing System. In Proceedings of the 2022 6th International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 25–27 May 2022; pp. 1105–1113. [Google Scholar]
  13. Gour, N.; Khanna, P. Multi-class multi-label ophthalmological disease detection using transfer learning based convolutional neural network. Biomed. Signal. Process. Control. 2021, 66, 102329. [Google Scholar] [CrossRef]
  14. Han, Y.; Li, W.; Liu, M.; Wu, Z.; Zhang, F.; Liu, X.; Tao, L.; Li, X.; Guo, X. Application of an Anomaly Detection Model to Screen for Ocular Diseases Using Color Retinal Fundus Images: Design and Evaluation Study. J. Med. Internet Res. 2021, 23, e27822. [Google Scholar] [CrossRef]
  15. Khan, S.; Tafshir, N.; Alam, K.N.; Dhruba, A.R.; Khan, M.M.; Albraikan, A.A.; Almalki, F.A. Deep Learning for Ocular Disease Recognition: An Inner-Class Balance. Comput. Intell. Neurosci. 2022, 2022, 1–12. [Google Scholar] [CrossRef] [PubMed]
  16. Li, B.; Chen, H.; Zhang, B.; Yuan, M.; Jin, X.; Lei, B.; Xu, J.; Gu, W.; Wong, D.C.S.; He, X.; et al. Development and evaluation of a deep learning model for the detection of multiple fundus diseases based on colour fundus photography. Br. J. Ophthalmol. 2022, 106, 1079–1086. [Google Scholar] [CrossRef] [PubMed]
  17. Muthukannan, P. Optimized convolution neural network based multiple eye disease detection. Comput. Biol. Med. 2022, 146, 105648. [Google Scholar]
  18. Rathakrishnan, N.; Raja, D. Optimized convolutional neural network-based comprehensive early diagnosis method for multiple eye disease recognition. J. Electron. Imaging 2022, 31, 043016. [Google Scholar] [CrossRef]
  19. Shanggong Medical Technology Co., Ltd. ODIR-5K. Available online: https://odir2019.grand-challenge.org/dataset/ (accessed on 3 October 2022).
  20. Vokinger, K.N.; Feuerriegel, S.; Kesselheim, A.S. Mitigating bias in machine learning for medicine. Commun. Med. 2021, 1, 1–3. [Google Scholar] [CrossRef] [PubMed]
  21. Ling, C.X.; Sheng, V.S. Class Imbalance Problem. In Encyclopedia of Machine Learning; Sammut, C., Webb, G.I., Eds.; Springer: Boston, MA, USA, 2011; Volume 10, p. 978. [Google Scholar]
  22. Lee, H.; Park, M.; Kim, J. Plankton classification on imbalanced large scale database via convolutional neural networks with transfer learning. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3713–3717. [Google Scholar]
  23. Johnson, J.M.; Khoshgoftaar, T.M. Survey on deep learning with class imbalance. J. Big Data 2019, 6, 1–54. [Google Scholar] [CrossRef]
  24. Brad, M.D.; Feldman, H.; Alpa, S. Cataract. Available online: https://eyewiki.aao.org/Cataract (accessed on 3 October 2022).
  25. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 11976–11986. [Google Scholar]
  26. Radosavovic, R.P.; Kosaraju, R.; Girshick, K.; Dollár, P.H. Designing network design spaces. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 14–19 June 2020; pp. 10428–10436. [Google Scholar]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  28. Horta, A.; Joshi, N.; Pekala, M.; Pacheco, K.D.; Kong, J.; Bressler, N.; Freund, D.E.; Burlina, P. A hybrid approach for incorporating deep visual features and side channel information with applications to AMD detection. In Proceedings of the 2017 16th IEEE International Conference on Machine Learning and Applications (ICMLA), Cancun, Mexico, 18–21 December 2017; pp. 716–720. [Google Scholar]
  29. Islam, M.T.; Imran, S.A.; Arefeen, A.; Hasan, M.; Shahnaz, C. Source and camera independent ophthalmic disease recognition from fundus image using neural network. In Proceedings of the 2019 IEEE International Conference on Signal Processing, Information, Communication & Systems (SPICSCON), Dhaka, Bangladesh, 28–30 November 2019; pp. 59–63. [Google Scholar]
  30. Tan, J.H.; Bhandary, S.V.; Sivaprasad, S.; Hagiwara, Y.; Bagchi, A.; Raghavendra, U.; Rao, A.K.; Raju, B.; Shetty, N.S.; Gertych, A.; et al. Age-related macular degeneration detection using deep convolutional neural network. Future Gener. Comput. Syst. 2018, 87, 127–135. [Google Scholar] [CrossRef]
  31. Ikechukwu, A.V.; Murali, S.; Deepu, R.; Shivamurthy, R. ResNet-50 vs VGG-19 vs training from scratch: A comparative analysis of the segmentation and classification of Pneumonia from chest X-ray images. Glob. Transit. Proc. 2021, 2, 375–381. [Google Scholar] [CrossRef]
  32. Reddy, S.B.; Juliet, D.S. Transfer learning with ResNet-50 for malaria cell-image classification. In Proceedings of the 2019 International Conference on Communication and Signal Processing (ICCSP), Chennai, India, 4–6 April 2019; pp. 945–949. [Google Scholar]
  33. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Nashville, TN, USA, 19–25 June 2021; pp. 10012–10022. [Google Scholar]
  34. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  35. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  36. Goyal, P.; Duval, Q.; Seessel, I.; Caron, M.; Misra, I.; Sagun, L.; Bojanowski, P. Vision models are more robust and fair when pretrained on uncurated images without supervision. arXiv 2022, arXiv:2202.08360. [Google Scholar]
  37. Azzuni, H.; Ridzuan, M.; Xu, M.; Yaqub, M. Color Space-based HoVer-Net for Nuclei Instance Segmentation and Classification. arXiv 2022, arXiv:2203.01940. [Google Scholar]
  38. Lihacova, I.; Bondarenko, A.; Chizhov, Y.; Uteshev, D.; Bliznuks, D.; Kiss, N.; Lihachev, A. Multi-Class CNN for Classification of Multispectral and Autofluorescence Skin Lesion Clinical Images. J. Clin. Med. 2022, 11, 2833. [Google Scholar] [CrossRef] [PubMed]
  39. Touvron, H.; Bojanowski, P.; Caron, M.; Cord, M.; El-Nouby, A.; Grave, E.; Jégou, H. Resmlp: Feedforward networks for image classification with data-efficient training. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 5314–5321. [Google Scholar] [CrossRef] [PubMed]
  40. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual Event, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
  41. Marcel, S.; Rodriguez, Y. Torchvision the machine-vision package of torch. In Proceedings of the 18th ACM international conference on Multimedia, Firenze, Italy, 25–29 October 2010; pp. 1485–1488. [Google Scholar]
  42. Althnian, A.; AlSaeed, D.; Al-Baity, H.; Samha, A.; Bin Dris, A.; Alzakari, N.; Elwafa, A.A.; Kurdi, H. Impact of dataset size on classification performance: An empirical evaluation in the medical domain. Appl. Sci. 2021, 11, 796. [Google Scholar] [CrossRef]
  43. Perez, L.; Wang, J. The effectiveness of data augmentation in image classification using deep learning. arXiv 2017, arXiv:1712.04621. [Google Scholar]
  44. Sajjad, M.; Khan, S.; Muhammad, K.; Wu, W.; Ullah, A.; Baik, S.W. Multi-grade brain tumor classification using deep CNN with extensive data augmentation. J. Comput. Sci. 2019, 30, 174–182. [Google Scholar] [CrossRef]
  45. Sedigh, P.; Sadeghian, R.; Masouleh, M.T. Generating synthetic medical images by using GAN to improve CNN performance in skin cancer classification. In Proceedings of the 2019 7th International Conference on Robotics and Mechatronics (ICRoM), IEEE, Tehran, Iran, 20–21 November 2019; pp. 497–502. [Google Scholar]
  46. Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and flexible image augmentations. Information 2020, 11, 125. [Google Scholar] [CrossRef]
  47. Decencière, E.; Cazuguel, G.; Zhang, X.; Thibault, G.; Klein, J.-C.; Meyer, F.; Marcotegui, B.; Quellec, G.; Lamard, M.; Danno, R.; et al. TeleOphta: Machine learning and image processing methods for teleophthalmology. IRBM 2013, 34, 196–203. [Google Scholar] [CrossRef]
  48. DeVries, T.; Taylor, G.W. Improved regularization of convolutional neural networks with cutout. arXiv 2017, arXiv:1708.04552. [Google Scholar]
  49. Biewald, L. Experiment Tracking with Weights and Biases. Software available from wandb.com. 2020. Available online: https://www.wandb.com/ (accessed on 24 May 2023).
  50. Liu, L.; Jiang, H.; He, P.; Chen, W.; Liu, X.; Gao, J.; Han, J. On the variance of the adaptive learning rate and beyond. arXiv 2019, arXiv:1908.03265. [Google Scholar]
  51. Loshchilov, I.; Hutter, F. Sgdr: Stochastic gradient descent with warm restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar]
  52. Chawla, N.V.; Japkowicz, N.; Kotcz, A. Special issue on learning from imbalanced data sets. ACM SIGKDD Explor. Newsl. 2004, 6, 1–6. [Google Scholar] [CrossRef]
  53. Van Hulse, J.; Khoshgoftaar, T.M.; Napolitano, A. Experimental perspectives on learning from imbalanced data. In Proceedings of the 24th International Conference on MACHINE Learning, Corvalis, OR, USA, 20–24 June 2007; pp. 935–942. [Google Scholar]
  54. Diaz-Pinto, A.; Morales, S.; Naranjo, V.; Köhler, T.; Mossi, J.M.; Navea, A. CNNs for automatic glaucoma assessment using fundus images: An extensive validation. Biomed. Eng. Online 2019, 18, 1–19. [Google Scholar] [CrossRef]
  55. Abràmoff, M.D.; Lou, Y.; Erginay, A.; Clarida, W.; Amelon, R.; Folk, J.C.; Niemeijer, M. Improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning. Investig. Ophthalmol. Vis. Sci. 2016, 57, 5200–5206. [Google Scholar] [CrossRef]
  56. Adal, K.M.; van Etten, P.G.; Martinez, J.P.; van Vliet, L.J.; Vermeer, K.A. Accuracy assessment of intra-and intervisit fundus image registration for diabetic retinopathy screening. Investig. Ophthalmol. Vis. Sci. 2015, 56, 1805–1812. [Google Scholar] [CrossRef]
  57. Holm, S.; Russell, G.; Nourrit, V.; McLoughlin, N. DR HAGIS—A fundus image database for the automatic extraction of retinal surface vessels from diabetic patients. J. Med. Imaging 2017, 4, 014503. [Google Scholar] [CrossRef] [PubMed]
  58. Pires, R.; Jelinek, H.F.; Wainer, J.; Valle, E.; Rocha, A. Advancing bag-of-visual-words representations for lesion classification in retinal images. PLoS ONE 2014, 9, e96814. [Google Scholar] [CrossRef]
  59. Drive. Digital Retinal Images for Vessel Extraction. Available online: https://drive.grand-challenge.org/ (accessed on 12 December 2022).
  60. Mahdi, H.; El Abbadi, N. Glaucoma Diagnosis Based on Retinal Fundus Image: A Review. Iraqi J. Sci. 2022, 63, 4022–4046. [Google Scholar] [CrossRef]
  61. Kaggle, E. Kaggle Diabetic Retinopathy Detection. 2015. Available online: https://www.kaggle.com/c/diabetic-retinopathy-detection/data (accessed on 1 December 2022).
  62. Almazroa, A.A.; Alodhayb, S.; Osman, E.; Ramadan, E.; Hummadi, M.; Dlaim, M.; Alkatee, M.; Raahemifar, K.; Lakshminarayanan, V. Retinal fundus images for glaucoma analysis: The RIGA dataset. In Proceedings of the Medical Imaging 2018: Imaging Informatics for Healthcare, Research, and Applications, SPIE, Houston, TX, USA, 10–15 February 2018; pp. 55–62. [Google Scholar]
  63. Orlando, J.I.; Fu, H.; Breda, J.B.; van Keer, K.; Bathula, D.R.; Diaz-Pinto, A.; Fang, R.; Heng, P.-A.; Kim, J.; Lee, J.; et al. REFUGE Challenge: A unified framework for evaluating automated methods for glaucoma assessment from fundus photographs. Med. Image Anal. 2020, 59, 101570. [Google Scholar] [CrossRef]
  64. Takahashi, H.; Tampo, H.; Arai, Y.; Inoue, Y.; Kawashima, H. Applying artificial intelligence to disease staging: Deep learning for improved staging of diabetic retinopathy. PLoS ONE 2017, 12, e0179790. [Google Scholar] [CrossRef]
  65. Abràmoff, M.D.; Folk, J.; Han, D.P.; Walker, J.D.; Williams, D.F.; Russell, S.; Massin, P.; Cochener, B.; Gain, P.; Tang, L.; et al. Automated analysis of retinal images for detection of referable diabetic retinopathy. JAMA Ophthalmol. 2013, 131, 351–357. [Google Scholar] [CrossRef] [PubMed]
  66. Li, L.; Xu, M.; Wang, X.; Jiang, L.; Liu, H. Attention based glaucoma detection: A large-scale database and CNN model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10571–10580. [Google Scholar]
  67. Batista, F.J.F.; Diaz-Aleman, T.; Sigut, J.; Alayon, S.; Arnay, R.; Angel-Pereira, D. Rim-one dl: A unified retinal image database for assessing glaucoma using deep learning. Image Anal. Stereol. 2020, 39, 161–167. [Google Scholar] [CrossRef]
  68. Niemeijer, M.; van Ginneken, B.; Cree, M.J.; Mizutani, A.; Quellec, G.; Sanchez, C.I.; Zhang, B.; Hornero, R.; Lamard, M.; Muramatsu, C.; et al. Retinopathy online challenge: Automatic detection of microaneurysms in digital color fundus photographs. IEEE Trans. Med. Imaging 2009, 29, 185–195. [Google Scholar] [CrossRef]
  69. Hoover, D.; Kouznetsova, V.; Goldbaum, M. Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response. IEEE Trans. Med. Imaging 2000, 19, 203–210. [Google Scholar] [CrossRef]
  70. Farnell, D.; Hatfield, F.; Knox, P.; Reakes, M.; Spencer, S.; Parry, D.; Harding, S. Enhancement of blood vessels in digital fundus photographs via the application of multiscale line operators. J. Frankl. Inst. 2008, 345, 748–765. [Google Scholar] [CrossRef]
  71. Pachade, S.; Porwal, P.; Thulkar, D.; Kokare, M.; Deshmukh, G.; Sahasrabuddhe, V.; Giancardo, L.; Quellec, G.; Mériaudeau, F. Retinal fundus multi-disease image dataset (RFMiD): A dataset for multi-disease detection research. Data 2021, 6, 14. [Google Scholar] [CrossRef]
  72. Hicks, S.A.; Strümke, I.; Thambawita, V.; Hammou, M.; Riegler, M.A.; Halvorsen, P.; Parasa, S. On evaluation metrics for medical applications of artificial intelligence. Sci. Rep. 2022, 12, 1–9. [Google Scholar] [CrossRef]
  73. Korotcov, A.; Tkachenko, V.; Russo, D.P.; Ekins, S. Comparison of deep learning with multiple machine learning methods and metrics using diverse drug discovery data sets. Mol. Pharm. 2017, 14, 4462–4475. [Google Scholar] [CrossRef]
  74. Harding, S.P.; Broadbent, D.M.; Neoh, C.; White, M.C.; Vora, J. Sensitivity and specificity of photography and direct ophthalmoscopy in screening for sight threatening eye disease: The Liverpool diabetic eye study. BMJ 1995, 311, 1131. [Google Scholar] [CrossRef]
  75. Santini, A.M.; Voidăzan, S. Accuracy of Diagnostic Tests. J. Crit. Care Med. 2021, 7, 241–248. [Google Scholar] [CrossRef]
  76. Huang, J.; Ling, C.X. Using AUC and accuracy in evaluating learning algorithms. IEEE Trans. Knowl. Data Eng. 2005, 17, 299–310. [Google Scholar] [CrossRef]
  77. Alberg, J.; Park, J.W.; Hager, B.W.; Brock, M.V.; Diener-West, M. The use of “overall accuracy” to evaluate the validity of screening or diagnostic tests. J. Gen. Intern. Med. 2004, 19, 460–465. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Data processing pipeline. We gathered 22 databases and merged them into one large dataset. We resized all images to 224 × 224 and normalized them by applying mean and std values derived from ImageNet-1K. Further, we split the data into two subgroups—the one used in pre-training and the other used in fine-tuning. During the training process, we dynamically augmented the images with fixed probabilities.
Figure 1. Data processing pipeline. We gathered 22 databases and merged them into one large dataset. We resized all images to 224 × 224 and normalized them by applying mean and std values derived from ImageNet-1K. Further, we split the data into two subgroups—the one used in pre-training and the other used in fine-tuning. During the training process, we dynamically augmented the images with fixed probabilities.
Diagnostics 13 01904 g001
Figure 2. Pre-training flow. We split the pre-training data into parts: train and validation set. We pre-train each model, performing evaluation every epoch while monitoring validation loss and applying early stopping with patients of five epochs.
Figure 2. Pre-training flow. We split the pre-training data into parts: train and validation set. We pre-train each model, performing evaluation every epoch while monitoring validation loss and applying early stopping with patients of five epochs.
Diagnostics 13 01904 g002
Figure 3. Fine-tuning workflow. We performed 10-fold cross-validation, selecting one-fold for a validation set and one-fold for a test set at each cross-validation step. We fine-tuned each model on a training set while monitoring validation loss and applying early stopping with patients of five epochs. In each cross-validation step, each fine-tuned model was evaluated on a test set, providing a set of metrics for that step. These performance metrics were further averaged.
Figure 3. Fine-tuning workflow. We performed 10-fold cross-validation, selecting one-fold for a validation set and one-fold for a test set at each cross-validation step. We fine-tuned each model on a training set while monitoring validation loss and applying early stopping with patients of five epochs. In each cross-validation step, each fine-tuned model was evaluated on a test set, providing a set of metrics for that step. These performance metrics were further averaged.
Diagnostics 13 01904 g003
Figure 4. Performance of our models shown with ROC plots. Each disease average-classification metric is shown with a different color. The area surrounding each line represents a standard deviation of ROC.
Figure 4. Performance of our models shown with ROC plots. Each disease average-classification metric is shown with a different color. The area surrounding each line represents a standard deviation of ROC.
Diagnostics 13 01904 g004
Table 1. Image augmentations used during model training with the probability of their application.
Table 1. Image augmentations used during model training with the probability of their application.
NameProbabilityValues
Rotate0.8[−90°, 90°]
Horizontal flip0.5-
Vertical flip0.5-
Random brightness contrast0.5Brightness limit: 0.1
Contrast limit: 0.15
Cutout0.5Number of holes: 20
Maximum height of hole: 11 px
Maximum width of hole: 11 px
Table 2. Summary of datasets used for pre-training, fine-tuning and ROS/RUS experiments.
Table 2. Summary of datasets used for pre-training, fine-tuning and ROS/RUS experiments.
Split# of Normal# of Glaucoma# of AMD# of Diabetic RetinopathyTotal # of Samples
Pre-training82,6260030,592113,218
Fine-tuning37873787632378711,993
ROS/RUS86,415378763234,379125,211
Table 3. Summary and characteristics of the datasets.
Table 3. Summary and characteristics of the datasets.
DatasetNGLAMDDRCamera ModelsAnnotators
ACRIMA [54]30939600Topcon → TRC → non-mydriaticTwo glaucoma experts with 8 years of experience
APTOS 2019 Blindness Detection Dataset [55]3733001857Variety of camerasNot available
Cataract [56]3001010100Not availableNot available
DR HAGIS [57]0101010Topcon TRC-NW6s non-mydriatic, Topcon TRC-NW8 non-mydriatic or Canon CR DGi non-mydriaticExpert grader
DR1, DR2 [58]895001118Topcon → TRC-50X mydriaticRespectively, three and two medical specialists
DRIVE [59]33005Canon → CR5 → non-mydriatic 3CCDOphthalmology expert
Machine learn for glaucoma [60]78875600Not availableNot available
e-optha [47]11600121Not availableOphthalmology experts
Kaggle: EyePACS [61]65,3430023,359Variety of camerasA panel of medical specialists
BAIDU: iChallenge-AMD [62]3110890Not availableNot available
REFUGE [63]3604000Zeiss Visucam 500 non-mydriaticSeven glaucoma specialists
Davis Grading of One and Concatenated Figures [64]6561003378Nidek AFC-230 non-mydriaticSpecialist grader
Longitudinal diabetic retinopathy screening data [65]0001120Topcon → TRC-NW65 non-mydriaticTwo graders
Messidor-2 [66]101700731Topcon TRC NW6Medical expert
ODIR-5K [19]30983122801697Various cameras such as Canon, Zeiss, KowaTrained human readers with
Quality control management
LAG [67]3147171100Not availableGlaucoma specialists
RIGA [62]028900Topcon → TRC → 50DX mydriaticSix experienced ophthalmologists
RIM-ONE DL [68]31317200Kowa WX 3D stereo non-mydriatic or Nidek AFzC-210 non-mydriatic with a Canon EOS 5D Mark II bodyThree experts
ROC [69]000100Topcon NW 100, NW 200, or Canon CR5-45NMRetinal experts
STARE [70]3606192TOPCON TRV-50Ophthalmology experts
ARIA [71]6102359Zeiss FF450+ mydriaticRetinal expert
RFMID [72]6690169632TOPCON 3D OCT-2000, Kowa VX-10alfa mydriatic and non-mydriatic two in one, and TOPCON TRC-NW300 non-mydriaticTwo ophthalmologists
TOTAL86,415378763234,379
N: Normal fundus image; GL: Glaucoma; AMD: Age-related macular degeneration; DR: Diabetic Retinopathy.
Table 4. Performance metrics for each model with standard deviation computed over the ten cross validation folds. Values in bold are the best results obtained.
Table 4. Performance metrics for each model with standard deviation computed over the ten cross validation folds. Values in bold are the best results obtained.
ClassMetricResNet50RegNetY3_2gfConvNextTiny
NormalF1-Score72.61 ± 1.8672.15 ± 2.3272.97 ± 2.60
Sensitivity73.75 ± 3.4973.75 ± 6.6474.57 ± 3.94
Specificity86.46 ± 1.6485.99 ± 2.9086.27 ± 1.76
AUC90.53 ± 0.7690.19 ± 0.7790.64 ± 0.56
Accuracy82.50 ± 1.2782.17 ± 0.8880.01 ± 1.10
GlaucomaF1-Score95.22 ± 0.8094.42 ± 0.8394.83 ± 0.96
Sensitivity95.64 ± 1.0295.11 ± 1.0995.54 ± 1.22
Specificity97.57 ± 0.6497.06 ± 0.5197.25 ± 0.81
AUC99.44 ± 0.1899.30 ± 0.2399.32 ± 0.17
Accuracy92.78 ± 0.4196.92 ± 0.6697.20 ± 0.66
AMDF1-Score81.78 ± 4.3579.25 ± 4.2182.98 ± 3.50
Sensitivity84.01 ± 8.1383.23 ± 6.8684.02 ± 6.37
Specificity98.82 ± 0.4398.51 ± 0.4798.97 ± 0.49
AUC99.25 ± 0.3498.99 ± 0.5198.79 ± 0.83
Accuracy97.91 ± 0.2998.13 ± 0.3198.14 ± 0.31
Diabetic RetinopathyF1-Score72.32 ± 1.2871.60 ± 2.2172.96 ± 1.78
Sensitivity70.56 ± 2.2369.11 ± 6.6770.69 ± 3.33
Specificity88.65 ± 1.6589.09 ± 3.3089.36 ± 1.91
AUC91.15 ± 0.6590.97 ± 0.6491.65 ± 0.87
Accuracy82.63 ± 1.0983.04 ± 0.9480.66 ± 1.27
AverageAccuracy89.88 ± 7.5390.15 ± 7.4388.99 ± 8.74
F1-Score80.48 ± 1.5179.36 ± 1.6080.93 ± 1.61
Sensitivity80.99 ± 2.1380.30 ± 2.0381.20 ± 2.26
Specificity92.88 ± 0.4092.66 ± 0.3992.96 ± 0.55
AUC95.09 ± 0.3994.87 ± 0.4195.10 ± 0.36
OverallAccuracy79.76 ± 1.3979.53 ± 1.0780.46 ± 1.48
Table 5. Performance metrics for each resampling method with standard deviation computed over the ten cross validation folds. All experiments were performed using the ConvNextTiny architecture. Values in bold are the best results obtained.
Table 5. Performance metrics for each resampling method with standard deviation computed over the ten cross validation folds. All experiments were performed using the ConvNextTiny architecture. Values in bold are the best results obtained.
Class Metric Two-Stage Learning (Our)RUS ROS
NormalF1-Score 72.97 ± 2.60 63.52 ± 3.5073.84 ± 1.36
Sensitivity 74.57 ± 3.9463.27 ± 8.6980.34 ± 5.50
Specificity 86.27 ± 1.7683.75 ± 3.7682.83 ± 3.79
AUC 90.64 ± 0.5685.59 ± 0.6790.23 ± 0.87
GlaucomaF1-Score 94.83 ± 0.9692.36 ± 1.0095.34 ± 1.28
Sensitivity 95.54 ± 1.2291.08 ± 3.0392.45 ± 3.07
Specificity 97.25 ± 0.8197.17 ± 0.9199.33 ± 0.45
AUC 99.32 ± 0.1798.69 ± 0.1899.46 ± 0.12
AMD F1-Score 82.98 ± 3.5076.16 ± 3.8484.44 ± 2.55
Sensitivity 84.02 ± 6.3793.01 ± 3.9075.40 ± 4.18
Specificity 98.97 ± 0.4997.13 ± 0.7299.82 ± 0.28
AUC 98.79 ± 0.8399.05 ± 0.3999.28 ± 0.27
Diabetic RetinopathyF1-Score 72.96 ± 1.7863.23 ± 2.8772.89 ± 1.59
Sensitivity 70.69 ± 3.3362.43 ± 7.7470.05 ± 4.84
Specificity 89.36 ± 1.9184.09 ± 4.3789.82 ± 2.74
AUC 91.65 ± 0.8785.52 ± 0.8491.41 ± 0.71
AverageF1-Score 80.93 ± 1.6173.82 ± 1.4381.63 ± 1.31
Sensitivity 81.20 ± 2.2677.45 ± 1.2979.56 ± 1.38
Specificity 92.96 ± 0.5590.54 ± 0.4392.95 ± 0.46
AUC 95.10 ± 0.3692.21 ± 0.4195.10 ± 0.40
Overall Accuracy 80.46 ± 1.4873.35 ± 1.2680.65 ± 1.30
TechnicalRuntime [s] 1196.437037,194.9
Table 6. Summary of the results obtained by our model and in related works. The results are difficult to compare, because each study had different aims, questions to answer and used different test sets. Because we used the most diverse dataset our metrics were the most reliable in terms of developing a model applicable in clinical screening.
Table 6. Summary of the results obtained by our model and in related works. The results are difficult to compare, because each study had different aims, questions to answer and used different test sets. Because we used the most diverse dataset our metrics were the most reliable in terms of developing a model applicable in clinical screening.
PaperClassF1-ScoreSensitivitySpecificityAUCAccuracy
Ours (ConvNextTiny)Normal72.9774.5786.2790.6480.01
Glaucoma94.8395.5497.2599.3297.20
AMD82.9884.0298.9798.7998.14
Diabetic Retinopathy72.9670.6989.3691.6580.66
Han et al. [14]Normal-----
Glaucoma-83.7084.0091.6083.89
AMD-77.6178.7586.7078.37
Diabetic Retinopathy-80.3680.5089.1080.39
Bulut et al. [11]Normal-----
Glaucoma---81.10-
AMD---96.30-
Diabetic Retinopathy---87.10-
Gour et al. [13]Normal-77.0021.00-40.00
Glaucoma-40.0060.00-54.00
AMD-06.0093.00-88.00
Diabetic Retinopathy-05.0094.00-89.00
Chellaswamy et al. [12]Normal96.3995.9991.27-95.00
Glaucoma96.4394.9596.32-96.00
AMD93.9699.0194.98-96.38
Diabetic Retinopathy-----
Muthukannan et al. [17]Normal94.0995.6598.56-99.20
Glaucoma97.0497.7799.28-97.80
AMD95.4994.9899.01-98.40
Diabetic Retinopathy94.9894.3198.92-97.90
Khan et al. [15]Normal-----
Glaucoma92.0097.00---
AMD88.0092.00---
Diabetic Retinopathy89.0092.00---
Li et al. [16]Normal-94.5095.7098.90-
Glaucoma-80.4093.4095.30-
AMD-85.8093.9097.60-
Diabetic Retinopathy-80.4089.7095.00-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chłopowiec, A.R.; Karanowski, K.; Skrzypczak, T.; Grzesiuk, M.; Chłopowiec, A.B.; Tabakov, M. Counteracting Data Bias and Class Imbalance—Towards a Useful and Reliable Retinal Disease Recognition System. Diagnostics 2023, 13, 1904. https://doi.org/10.3390/diagnostics13111904

AMA Style

Chłopowiec AR, Karanowski K, Skrzypczak T, Grzesiuk M, Chłopowiec AB, Tabakov M. Counteracting Data Bias and Class Imbalance—Towards a Useful and Reliable Retinal Disease Recognition System. Diagnostics. 2023; 13(11):1904. https://doi.org/10.3390/diagnostics13111904

Chicago/Turabian Style

Chłopowiec, Adam R., Konrad Karanowski, Tomasz Skrzypczak, Mateusz Grzesiuk, Adrian B. Chłopowiec, and Martin Tabakov. 2023. "Counteracting Data Bias and Class Imbalance—Towards a Useful and Reliable Retinal Disease Recognition System" Diagnostics 13, no. 11: 1904. https://doi.org/10.3390/diagnostics13111904

APA Style

Chłopowiec, A. R., Karanowski, K., Skrzypczak, T., Grzesiuk, M., Chłopowiec, A. B., & Tabakov, M. (2023). Counteracting Data Bias and Class Imbalance—Towards a Useful and Reliable Retinal Disease Recognition System. Diagnostics, 13(11), 1904. https://doi.org/10.3390/diagnostics13111904

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop