Next Article in Journal
Apple Consumption Protects against Acute Ethanol-Induced Liver Injury in Rats
Next Article in Special Issue
Transfer Learning for Diabetic Retinopathy Detection: A Study of Dataset Combination and Model Performance
Previous Article in Journal
Leveraging Dialogue State Tracking for Zero-Shot Chat-Based Social Engineering Attack Recognition
Previous Article in Special Issue
MIU-Net: MIX-Attention and Inception U-Net for Histopathology Image Nuclei Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Survey on Diabetic Retinopathy Lesion Detection and Segmentation

Department of Computer Science and Engineering, Qatar University, Doha P.O. Box 2713, Qatar
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(8), 5111; https://doi.org/10.3390/app13085111
Submission received: 25 February 2023 / Revised: 6 April 2023 / Accepted: 11 April 2023 / Published: 19 April 2023
(This article belongs to the Special Issue The Application of Machine Learning in Medical Image Processing)

Abstract

:
Diabetes is a global problem which impacts people of all ages. Diabetic retinopathy (DR) is a main ailment of the eyes resulting from diabetes which can result in loss of eyesight if not detected and treated on time. The current process of detecting DR and its progress involves manual examination by experts, which is time-consuming. Extracting the retinal vasculature, and segmentation of the optic disc (OD)/fovea play a significant part in detecting DR. Detecting DR lesions like microaneurysms (MA), hemorrhages (HM), and exudates (EX), helps to establish the current stage of DR. Recently with the advancement in artificial intelligence (AI), and deep learning(DL), which is a division of AI, is widely being used in DR related studies. Our study surveys the latest literature in “DR segmentation and lesion detection from fundus images using DL”.

1. Introduction

The number of people diagnosed with diabetes has been increasing at an alarming rate during the past two decades. This accounts for almost half a billion people in the world as per the IDF Diabetes Atlas [1], which makes it a global health concern. This includes people of all ages. By 2045, this figure is supposed to reach seven hundred million and by the year 2040, one in three persons with this ailment will have DR, which is an ailment characterized by the occurrence of damaged blood vessels in the back of the retina. If it is not detected in a timely manner, it can lead to serious problems, like loss of sight. This makes DR a topic of the utmost importance. Doctors must physically analyze fundus images to determine the presence of this disease and to grade it. Fundus images offer graphical records that detail the current ophthalmic form of an indvidual’s retina. This is a very time-consuming process, and in many countries the number of available trained professionals may not be proportionate to the number of patients. This may hinder providing timely treatment to many patients. Diabetes patients are instructed by doctors to have a routine medical screening of their fundus. Still many DR cases are left unnoticed until they reach a critical state. Due to this reason, it is advantageous to have computerized systems that assist in the diagnosis of DR.
Researchers are developing automated systems for DR diagnosis by identifying the existence of certain DR lesions in fundus images. Few of the DR lesions include MA, HM, EX consisting of soft exudates (SE) and hard exudates(HE). DR grades are also established with the help of lesions present in the fundus. For example, the preliminary stage of DR is marked by the beginning of the growth of new blood vessels whose walls weaken, and tiny bulges or MA protrude from them. Later, when DR progresses to proliferative DR, other lesions like HE and SE may develop. HE are little white/yellowish-white deposits with crisp boundaries whereas, SE/Cotton-wool spots are light yellow or white areas which have unclear edges. Additionally, abnormal blood vessels may grow and bleed causing HM. Hence, some researchers are focusing on the detection and segmentation of DR lesions like MA, HM, and EX which will aid in developing better DR diagnoses/grading for automated systems. For example, Figure 1 shows each of these DR lesions, whereas Figure 2 provides a visualization of a healthy retina vs an unhealthy retina.
There are many challenges to developing such automated systems. The size of these lesions vary in a temporal manner and the OD, and fovea may be mistaken for lesions. Due to this reason, some researchers are focusing on OD/fovea detection which can also help to improve such systems. Another category of studies that are related to DR include retinal blood vessel segmentation studies. These are important to detect several medical disorders like DR and hypertensive retinopathy. Atypical variations in the retinal vessel will give an idea of the severity of several non-ophthalmic ailments like DR, diabetes mellitus, hypertension, arteriosclerosis, cardiovascular disease, and stroke [4]. Hence, it is important for diagnosing these diseases, especially DR.
Several DL methods were used by researchers to accomplish these types of DR studies. We present a review of the current literature in this field by emphasising how DL has been applied to DR segmentation and lesion detection from fundus images. Deep learning is a section of artificial intelligence which uses artificial neural networks with many processing layers for steadily extracting high-level features from the data.
This article is organized as follows: the studies that were reviewed associated with segmentation and lesion detection are shown in Section 2. Section 3 illustrates the datasets used. These are followed by an evaluation and discussion of these methods in Section 4. After this we have compiled a list of some possible future directions in this area of study in Section 5. Lastly, a conclusion is provided in Section 6.

2. Related Works

The DR studies reviewed in this article have been classified into two categories which include traditional-based approaches and DL-based studies.
Different traditional methods are being used for retinal blood vessel segmentation as well as for lesion detection from fundus images. Detection and counting of MA were done by the authors in [5] by using morphological operations, histogram equalization, and intensity transformation functions. HM segmentation was performed by the authors of [6] using histogram equalization and morphological operations on the green channel image of fundus images.
Several traditional methods have been used for EX detection like thresholding, morphology, edge detection, mixture modeling, and region growing. Grayscale morphology was used along with active contour-based segmentation and region-wise classification by the authors of [7] for automatic EX detection. The authors used a blend of region growing and edge detection in [8]. They used the Luv color space to get the color difference image of objects. A mean squared Wiener filter in two dimensions was used to remove noise which was trailed by edge detection and an enhancement of the region growing method. Gray level thresholding was used by the authors in [9] since EX have high local and gray levels. The local properties of EX were exploited and a blend of global as well as local histogram thresholding was used by them. Sanchez et al. [10] used an EX detection method built using mixture models for obtaining image histograms. This helped them get a dynamic threshold for every image. Some machine learning approaches have also been used in the literature. Niemeijer et al. [11] used the k-nearest neighbor classifier for detecting red lesions from fundus images. A ranking SVM-boosted convolutional neural network (CNN) was utilized for EX detection by the authors of [12]. Damaged or falsely detected OD and HM were also detected by them.
Retinal blood vessel segmentation using a top-hat morphological transform followed by hybrid median filtering was performed by the authors in [13]. Quinn and Krishnan [14] performed a contrast adjustment of the green channel of the image trailed by curvelet transform and morphological reconstruction for retinal blood vessel segmentation. A mathematical morphology-based fuzzy algorithm method was used for OD elimination followed by retinal blood vessel extraction by the authors of [15].
The DL-based DR studies reviewed in this study have been categorized into two types which include segmentation-based studies and lesion detection-based studies. Figure 3 captures these two distinct types of DR studies.

2.1. Diabetic Retinopathy Lesion Detection

The studies that were devoted to the detection of DR lesions like MA, HM, and EX fall under this category. Another group of studies detected multiple lesions as well as OD. A set of proposed methods are discussed in this section as well as a summarization of them in terms of techniques and datasets used in Table 1 summarizes.
Some studies reviewed, focused on detecting EX. Raja and Balaji [16] used adaptive histogram equalization for preprocessing followed by blood vessel segmentation using CNN. After this, EX segmentation was performed using fuzzy c-means clustering. Finally, DR classification was achieved using SVM. Adem [17], combined a circular Hough transform with CNNs to detect EX. The circular Hough transform helps to ignore OD. EX detection was performed after segmentation of EX. SVM based on the ranking was implemented along with CNN architecture to accurately segment fundus images by the authors in [12]. They also detected EX after segmentation. HE detection with deep CNN and multi-feature joint representation was performed by Wang et al. [18] using a new methodology. They utilized combined features that included deep features as well as handcrafted features. HE detection using SVM blended with a speedier region-based CNN (RCNN) was performed by the authors of [19]. Wang et al. [20] performed EX classification. They used a CNN to investigate if the classification performance improves if more importance was given to the image characteristic inside an EX. The outcomes showed that the performance improved. Recurrent neural network (RNN) based semantic features were used for EX segmentation by the authors of [21] for detecting DR. Huang et al. [22] used a super-pixel segmentation algorithm, followed by multi-feature extraction and patch generation. Finally, CNN was used for HE detection. RCNN with MobileNet was utilized as a feature extractor by the authors of [23] for EX detection. Mateen et al. [24] used pre-trained Inception-v3, ResNet-50, and VGG-19 to perform feature extraction before EX classification using the softmax classifier. Another study by the authors of [25] used three different models including SqueezeNet, ResNet50, and GoogLeNet for EX detection. Among these three, ResNet50 was found to provide the best results. A CNN with residual skip connection was used for EX detection by the authors in [26]. The architecture used was called RTC-net.
Few studies focused on MA detection. Multi-scale residual network (MSRNet) was used for segmentation by the authors of [27] to detect tiny MA. It was based on U-Net architecture. After this, classification was achieved using MS-EfficientNet (multi-scale efficient network). Qiao et al. [28] used the prognosis of MA for DR grading into three, which includes early NPDR, moderate NPDR, and severe NPDR. Deep CNN was used to perform semantic segmentation of fundus images. Gaussian answer Laplacian (LOG) and maximum matching filter response (MFR) filters were used for lesion identification. Feature extraction using VGG and Inception V3 backbones, alongside transfer learning was utilized for MA detection by the authors of [29]. After this, several machine learning classifiers were used for classifying input images. Qomariah et al. [30] modified the U-Net model and used it for MA segmentation. Residual units with modified identity mapping were used in this new model known as MResUNet. Subhasree et al. [31] made use of three different transfer learning-based CNN models for MA detection. These consisted of VGG-16, Inception V3, and ResNet50. They also performed the same task by making use of a random forest approach.
Some studies did a general detection of DR lesions. Conditional generative adversarial network (cGAN) was used for pixel-level lesion segmentation by combining with holistically-nested edge detection (HEDNet) by the authors in [32], and it was found that adversarial loss advances the lesion segmentation performance. An approach known as super-pixel-based segmentation was used by the authors of [33] to compare the performance of three different algorithms under a single unified framework in performing multi-lesion detection. Out of these, a compacted watershed algorithm was found to be faster, whereas a linear spectral clustering (LSC) algorithm generated better segmentation results. After this, a single-layer CNN was used to detect lesions. Dai et al. [34] introduced a novel system called DeepDR comprising of three sub-networks (based on ResNet and Mask-RCNN) to grade DR using a large dataset collected between 2014 and 2017. Lesion segmentation was performed before DR grading. The Javeria segmentation (JSeg) model, a novel model with the ResNet-50 model as the backbone, was utilized for segmentation by the authors in [35]. In their next phase, Resnet-101 was used for feature extraction built on the equilibrium optimizer (EO). After this, support vector machine (SVM) and neural network classifiers were used for grading lesions. The YOLOv3 model was used for lesion localization by the authors in [36]. It was fused with the CNN512 model to grade DR and localize DR lesions. Wang et al. [37] introduced a new model called adaptive multi-target weighting network (AMWNet) to segment DR lesions. It consisted of an encoder, decoder and an adaptive multi-target weighting module. After this reverse data recovery network (RRN) was used to replicate the cycle perception as part of the hierarchy. DR lesion detection using a CNN-based one-shot detector was used by the authors in [38]. The YOLOv4 deep neural network (DNN) architecture was used along with the Darknet framework.
Many studies reviewed, detected multiple DR lesions. Pixel-wise segmentation of MA, EX, and OD were achieved by the authors of [39] using a novel membrane system termed “dynamic membrane system” and mask R-CNN. MA and EX segmentation was performed by the authors of [40] by using a revised U-Net architecture which was grounded on the residual network. U-Net with incremental learning was used by the authors in [41] for EX and HM segmentation. A new incremental learning standard was introduced to excerpt the acquaintance gained by the preceding model to advance the present model. Another study that detected EX and HM was performed by the authors in [42]. They performed a semantic segmentation with three classes and used a color space transformation along with U-Net. Ananda et al. [43] used one individual network per one type of lesion related to DR (e.g., one network for MA, one network for HE, etc.). They found that this modification yielded better results for U-Net to classify MA, SE, and HM and detect DR. U-Net segmentation followed by classification using support vector machine was achieved by the authors of [44] for grading DR by detecting some DR lesions. Blood vessels, EX, MA, and HM were segmented for this purpose. A nested U-Net architecture known as U-Net++ was used for red lesion segmentation (MA and HM segmentation) from fundus images by the authors in [45]. This was followed by sub-image classification using ResNet-18 to reduce false positives. Another study that focused on red lesion detection was one by the authors of [46]. For this purpose, they fine-tuned various pre-trained CNN models with an augmented set of image patches and found ResNet50 to be the best. They modified it with reinforced skip connections to get a new model called DR-ResNet50.
The authors in [47] used MA, EX, and HM diagnosis to detect DR by using regions with CNNs (R-CNN). Feature extraction using the VGG19 model trialed by classification using different machine learning classifiers by the authors of [48]. They identified and classified DR lesions into MA, SE, HE, and HM. A single-scale CNN (SS-CNN) followed by a multi-scale CNN (MS-CNN) was used by the authors of [49] to perform pixel-level detection of MA and HM. Four DR lesions, including MA, HE, SE, and HM, were segmented by the authors of [50] using a relation transformer network (RTNet). In this, a relation transformer block was incorporated into attention mechanisms in two levels. Another new model was used to detect these same four lesions by the authors in [51]. It was called faster region-based CNN (RCNN) and was a combination of a regional proposal network (RPN) and fast-RCNN. Another new model known as deeply-supervised multiscale attention U-Net (Mult-Attn-U-Net) was used by the authors of [52] for detecting MA, EX, and HM. It was based on the U-Net architecture. Dual-input attentive RefineNet (DARNet) was used by the authors in [53] to detect four DR lesions including HE, SE, HM, and MA. DARNet included a global image encoder, local image encoder, and attention refinement decoder. The same authors used a cascade attentive RefineNet for detecting the same four DR lesions in another study [54]. CARNet included a global image encoder, a patch image encoder, and an attention refinement decoder. A YOLOv5 model along with a genetic algorithm was used by the authors in [55] to detect MA, SE, HE, and HM. Another study by the authors of [56] detected MA, EX, and HM with the help of a CNN along with a SVM. Prior to this, they used a U-Net to segment OD and retinal blood vessels for feature extraction.
Three studies among the reviewed works, focused on the segmentation of the DR lesion called HM. Skouta et al. [57] made use of a U-Net architecture for detecting the presence of HM. Aziz et al. [58] utilized a novel CNN model known as HE network (HemNet) for HM detection. Quality enhancement, seed points extraction, image calibration, and smart window-based adaptive thresholding were performed prior to training this model. A 3D CNN deep learning framework with feature fusion was used to perform HM detection by the authors in [59]. A variation of the traditional CLAHE method was for pre-processing to achieve better edge detection.
Table 1. Deep Learning based approaches for diabetic retinopathy lesion detection.
Table 1. Deep Learning based approaches for diabetic retinopathy lesion detection.
TypeMethodYearTechniqueDataset
ExudatesAdem [17]2018Adaptive histogram equalization, CNNDIARETDB0, DIARETDB1
Raja and Balaji [16]2019CNN and fuzzy c-means clustering, SVMDRIVE, STARE
Wang et al. [20]2019CNNSelf-collected
Wang et al. [18]2020Deep CNN, multi-feature joint representationE-Optha, DIARETDB1
Mateen et al. [24]2020Inception-v3, ResNet-50 and VGG-19DIARETDB1, E-ophtha
Ghosh and Ghosh [12]2021rSVM, CNNSTARE, DIARETDB0, DIARETDB1
Kurilová et al. [19]2021SVM, RCNN, ResNet-50E-ophtha, DIARETDB1, MESSIDOR
Huang et al. [22]2021Patch-based deep CNNDIARETDB1, E-ophtha, IDRiD
Bibi et al. [23]2021RCNNDIARETDB1, E-ophtha
Cincan et al. [25]2021SqueezeNet, ResNet50 and GoogLeNet-
Sivapriya et al. [21]2022RNNMESSIDOR
Manan et al. [26]2022RTC-netDIARETDB1, E-ophtha
MicroaneurysmsQiao et al. [28]2020Deep CNNIDRiD
Qomariah et al. [30]2021MResUNetDIARETDB1, IDRiD
Xia et al. [27]2021MSRNet, MS-EfficientNetE-Ophtha, IDRiD, DRIVE, CHASE DB1
Subhasree et al. [31]2022VGG-16, Inception V3, and ResNet50E-Ophtha, DIARETDB1
Gupta et al. [29]2022VGG, Inception V3ROC
Multiple LesionsXiao et al. [32]2019cGAN, HEDNetIDRiD
Ananda et al. [43]2019U-Net, SegNetIDRiD, MESSIDOR
Praveena and Lavanya [33]2019Super-pixel based segmentationDIARETDB1, DRiDB, HRF
Xue et al. [39]2019CNNIDRiD, E-Optha, MESSIDOR
Gupta et al. [48]2019VGG-19IDRiD
Nazir et al. [51]2020RCNNDIARETDB1, MESSIDOR
Sambyal et al. [40]2020Modified U-Net architectureIDRiD, E-Optha
He et al. [41]2020U-Net, Incremental learning-
Dai et al. [34]2021ResNet, Mask-RCNNSelf-collected
Alyoubi et al. [36]2021YOLOv3, CNN512DDR
Wang et al. [37]2021AMWNetIDRiD
Abdelmaksoud et al. [44]2021U-Net, SVMDRIVE, STARE, CHASEDB1, E-ophtha
Li et al. [49]2021SS-CNN, MS-CNNDIARETDB1
Basu and Mitra [52]2021Mult-Attn-U-NetIDRiD
Santos et al. [38]2021YOLOv4 with darknetDDR
Amin et al. [35]2022JSeg modelIDRiD
Kundu et al. [45]2022U-Net++, ResNet-18MESSIDOR, DIARETDB1
Latchoumi et al. [47]2022R-CNN-
Huang et al. [50]2022RTNetIDRiD, DDR
Guo and Peng [53]2022DARNetIDRiD, DDR, E-ophtha
Guo and Peng [54]2022CARNetIDRiD, DDR, E-ophtha
Ashraf et al. [46]2022DR-ResNet50DIARETDB1, E-ophtha, IDRiD
Santos et al. [55]2022YOLOv5DDR
Selçuk et al. [42]2022U-NetMESSIDOR, DIARETDB1
Jena et al. [56]2023U-Net, CNNAPTOS, MESSIDOR
HemorrhageMaqsood et al. [59]20213D CNN with feature fusionDIARETDB0, DIARETDB1, MESSIDOR, HRF, DRIVE, STARE
Skouta et al. [57]2022U-NetIDRiD, DIARETDB1
Aziz et al. [58]2023HemNetDIARETDB0, DIARETDB1

2.2. Diabetic Retinopathy Retinal Blood Vessel/Optic Disc Segmentation

This category includes those studies which performed retinal blood vessel segmentation as well as those which performed OD/region of interest (ROI) segmentation. Most studies fall under the first type, which helps to study the vascular tree and can thus help assess the current medical condition of the patient’s eyes. Other studies focused on optic disc (OD)/fovea segmentation and ROI segmentation. These studies can aid the feature extraction process required for developing automatic DR diagnosis systems by helping to remove unnecessary features before doing them [60].
One of the most preferred DL methods for segmentation is to use CNNs [61,62,63,64,65]. The presence of neovascularization was used to spot proliferative DR by the authors in [66]. Different layers of CNNs were modeled together with VGG-16 net architecture for this. They performed blood vessel segmentation to remove the extra features before doing the classification. Chala et al. [67] used a multi-encoder decoder architecture with two encoder units, as well as a decoder unit for retinal blood vessel segmentation. Aujih et al. [68] studied how the U-Net model performs for the same purpose. Dropout and batch normalization with various settings were utilized for this. Batch normalization was found to make U-Net learn faster but degraded the performance after epoch thirty. They also studied how the presence/absence of retinal blood vessel segmentation affected the DR classification performance using Inception v1. Burewar et al. [69] made use of U-Net for retinal segmentation with region merging. After this, CNN with ReLU activation function and max-pooling was utilized to perform DR grading. Yadav [70] applied a dual-tree discrete Ridgelet transform (DT-DRT) to fundus images for feature extraction from ROIs and used a U-Net-based method for retinal blood vessel segmentation after this. The U-Net++ architecture was used to extract the retinal vasculature by the authors in [71]. They used the extracted features to predict DR in the next stage of their experiment. A context-involved U-Net was used to perform retinal blood vessel segmentation in the study by the authors in [72]. They made use of patch-based loss weight mapping to improve the segmentation of thin blood vessels. Another study by the authors in [73] used an encoder enhanced atrous U-Net to extract the retinal vasculature from fundus images. They introduced two additional layers to the encoder at every stage to enhance the extraction of edge information. Jebaseeli et al. [74] utilized contrast limited adaptive histogram equalization (CLAHE) to perform pre-processing. After this, feature generation and classification were performed. The former step was accomplished using the tandem pulse coupled neural network model and the latter was achieved using DL based support vector machine (DLBSVM). DLBSVM also extracted the blood vessel tree from the fundus images. Similar work was done by the authors of [75]. Another study by the authors of [76] made use of a pool-less residual segmentation network (PLRS-Net) for retinal blood vessel segmentation, which could segment smaller vessels better. They used two pool-less approaches called PLRS-Net and PLS-Net (pool-less segmentation network) which did not require any pre/post-processing. A multilevel deep neural network with a three-plane preprocessing method was used by the authors of [77] for retinal blood vessel segmentation. The first four convolutional layers of VGG-16 were fine-tuned for blood vessel feature extraction. Jin et al. [78] performed retinal blood vessel segmentation using a new model known as deformable U-Net (DUNet). This model combined the benefits of the U-Net architecture with those of a deformable convolutional network. Another new network followed, a network called N F N + was used by the authors of [79] to segment retinal blood vessels. A distinct cascaded design, along with inter-network connections made this model unique. Distinctive segmentation of thick as well as thin retinal blood vessels was performed by the authors of [80] by using a three-stage DL model. These stages consisted of the segmentation of thick vessels, the segmentation of thin vessels, and the fusion of vessels. Another model known as multi-scale CNN with attention mechanisms (MSCNN-AM) was used by the authors in [81] to segment retinal blood vessels using various dimensions. Atrous separable convolutions which had differing dilation rates were used to seize global and multi-scale vessel data more effectively. Li et al. [82] used an enhanced U-Net architecture in order to segment retinal blood vessels. They made use of data from the West China Hospital of Sichuan University to perform their experiments. Three DL models consisting of SegNet, U-Net, and CNN were used by the authors of [83] for the same task. The SegNet model was found to be better than the other two models in segmenting both thin and thick blood vessels. A back propagation neural network was utilized by the researchers in [84] to extract the retinal vasculature from fundus images. They could obtain a reduction in operation time in addition to improved accuracy. A pre-processing AlexNet block followed by an encoder-decoder U-Net model was utilized by the authors in [85] to extract the retinal vasculature. Their study aimed to support the diagnosis of several ophthalmological diseases including DR. Another similar study that performed extraction of the retinal vasculature was performed by the authors in [86] and used a multi-scale residual CNN known as MSR-Net, along with a generative adversarial network (GAN). The generator of the GAN was a segmentation network, whereas the discriminator was a classification network. Another GAN called Pix2Pix GAN was used by the authors in [87] for retinal blood vessel segmentation to help studies that detect different ophthalmological diseases. A novel model known as D-MNet with multi-scale attention and a residual mechanism, was used by the authors in [88] for retinal blood vessel segmentation. It was used along with an enhanced version of pulse-coupled neural network to achieve this task.
Some studies performed OD segmentation. This was achieved by using an encoder-decoder with the VGG-16 backbone by the authors of [89]. Convolutional long short term memory was merged into the encoder to improve accuracy. Kumar et al. [90] used a mathematical morphology operation before extracting the retinal blood vessels from fundus images. OD segmentation was done with the help of watershed transform. Finally, classification was done using radial basis function neural network (RBFNN). Yeh et al. [91] used a preprocessing technique for OD segmentation called local differential filter (LDF). Good prediction results were obtained when LDF was used for U-Net model training data to differentiate between OD and background regions. Retinal blood vessel segmentation, as well as extraction of the optic disc as well as fovea centers, were performed by the authors in [92] using an end-to-end encoder-decoder network known as DRNet. Shankar et al. [93] introduced a deep CNN model named Synergic DL model for retinopathy grading. They also performed histogram-based OD segmentation prior to this.
Another set of studies in this category focused on ROI segmentation. An algorithm named “adaptive-learning-rate-based” enhanced GMM was used for ROI segmentation in fundus images by Mansour [94]. After this, they used connected component analysis (CCA) to process the input to AlexNet architecture. This was followed by feature selection and grading using a polynomial kernel-based SVM classifier. Active DL was used in a seven-layer CNN architecture (ADL-CNN) by Qureshi et al. [95] to automatically grade the images into the five stages of DR after retinal blood vessel segmentation as well as ROI localization. Table 2 summarizes the studies that were presented in this section.

3. Datasets Used for Diabetic Retinopathy Segmentation/Lesion Detection

The datasets being employed have a significant role in the accomplishment of the DL studies reviewed here. Some factors like the quality of these datasets, as well as the precision of their annotations, impact the results of such studies. Therefore, it will be useful to have a list of a few generally utilized fundus image datasets for DR segmentation/lesion detection. We have presented such a list in Table 3.
Some of the generally utilized, publicly accessible datasets in such studies include DRIVE, STARE, IDRiD, E-ophtha, MESSIDOR, DIARET DB1, and DIARET DB0. Images gathered during a DR assessment program in the Netherlands were employed to create the dataset known as Digital Retinal Images for Vessel Extraction (DRIVE). It consists of twenty images each for training and testing sets [96]. The STARE dataset was collected as part of a project known as STARE (Structured Analysis of the Retina). It provides forty hand-labeled images for blood vessel segmentation [97]. The Indian DR Image Dataset (IDRiD) provides groundtruth for DR and Diabetic Macular Edema (DME). It contains annotations for DR grading, DR lesions, and optic disc [98]. The E-Optha dataset comprises of two sub-datasets called E-Optha-MA and E-Optha-EX [99]. Just as their names indicate these datasets have been annotated for EX and MA correspondingly. Another dataset known as DIARETDB1 has been annotated by four medical experts for DR lesions like MA and EX [100]. The MESSIDOR dataset was part of the MESSIDOR project and has been publicly available since 2008 [101]. The DDR dataset was collected from 9598 patients and 757 images with DR have been annotated for DR lesions [102].
Out of these datasets, DRIVE and STARE datasets are the highly applied datasets for retinal blood vessel segmentation. However, since manual annotation of the retinal blood vessels is time consuming and tedious, these datasets have very few images. Datasets like IDRiD, E-Optha MA, DIARET DB0, and DIARET DB1 have been annotated for MA detection. E-Optha EX, IDRiD, DIARET DB0, and DIARET DB1 have been annotated for EX detection. Whereas the annotation for OD is available for MESSIDOR and DIARET DB1 datasets. It was noted that only very few datasets are annotated for retinal blood vessel segmentation and for lesion detection. This may be because marking retinal blood vessels manually will require a lot of time, as well as effort, and annotating lesions demands intra-annotator agreement and high annotation precision. Due to this reason, the majority of the studies utilized more than one dataset for training as well as validation.

4. Discussion

Several DL architectures were used to effectively perform DR segmentation and lesion detection tasks in the studies reviewed in the previous section. Some examples of these architectures/models are ResNet, Inception, EfficientNet, U-Net, and Encoder-Decoder. Apart from these, most studies used different pre-processing techniques for gaining better performance and for image enhancement. This can include different steps like intensity conversion, image variation attenuation, contrast enhancement, and denoising [103]. Since there will be a huge disparity in the retinal color of different patients, attenuation of fundus images may have to be done. Additionally, to make the features in an image distinctly visible, intensity conversion can be used. Another pre-processing step is the denoising of fundus images. This is necessary, since throughout the image acquisition procedure plenty of noise may have been introduced in these images. Lastly, since there will be highest contrast at the image centre for fundus images, which steadily reduces when moving away from the centre, contrast enhancement may have to be performed. Additional pre-processing steps consist of image resizing, grayscale conversion, gaussian filtering, applying CLAHE [74], and doing several image augmentations using techniques like rotation, flip, and zoom. Some morphological operations like erosion, dilation, top-hat transformation, and bottom-hat transformation were accomplished on the green channel images of the fundus images as pre-processing in [90]. Hasan et al. [92] resized and standardized the images as part of pre-processing. They also performed image augmentation and generated 2D heatmaps as pre-processing.
The studies under review can be compared using several metrics like accuracy, sensitivity, and specificity. These are some of the most commonly used metrics in computer vision. This part presents the results attained per dataset by utilizing the cited methods for segmentation or lesion detection. This has been done by reporting the results with the help of tables and figures to identify the best-performing deep-learning architectures for these tasks. Table 4, Table 5 and Table 6 present an evaluation of the results achieved by a few studies which were assessed in this study. A visual comparison of some studies that performed retinal blood vessel segmentation, MA detection, and HE detection have been provided in Figure 4 and Figure 5. Finally, Figure 6 provides a comparison of the results attained by researchers for deep-learning-based HE and MA detection on the IDRiD dataset, which is one of the biggest and most highly used datasets preferred by researchers for DR segmentation/lesion detection.

4.1. Diabetic Retinopathy Segmentation

DR segmentation-based studies include those studies that segment retinal blood vessels, OD, or ROI from fundus images. Table 4 and Table 5 provides a comparison of the reviewed DR retinal blood vessel segmentation studies grouped according to the datasets used by them. The segmented results obtained will be helpful for extended research in the area of DR. According to these tables, it is obvious that most of the studies obtained good accuracies but the specificity and sensitivity values are unavailable for some studies. It is clear that on the STARE dataset, the authors of [74] attained the highest accuracy value of 99.7% which is better by 0.6% than the second highest accuracy value of 99.1%, obtained by the authors in [75] and 1.1% better than the accuracy values obtained by other methods. They could also achieve the highest specificity value compared to other methods. Whereas, when sensitivity is taken into account, the highest value was obtained by the authors in [66].
When we look at the values obtained for the DRIVE dataset in the table, the same study that achieved the highest accuracy value for the STARE dataset also obtained the best value for specificity as well as the second-best value for accuracy. Whereas, the highest values for accuracy and sensitivity, 99.9% and 96.7% were obtained by the authors in [83,92], respectively. Figure 4 provides a further comparison of retinal blood vessel segmentation results using the DRIVE dataset by the researchers in [75,76]. We can observe that the former study clearly segments the retinal blood vessels without any trace of OD and fovea, whereas in the second study parts of the OD and fovea are present in the obtained result.
One of the major challenges which researchers find while performing retinal blood vessel segmentation, is the segmention of tiny blood vessels. Hence, more studies are required that can effectively detect these similar to the study of [76]. It would also be beneficial to classify the detected blood vessels into arteries/veins as well as to determine the width of these blood vessels, which will provide valuable information that can be used to determine the present condition of the patient’s eyes. When it comes to OD segmentation, the major challenge is that other DR lesions may be wrongly segmented as OD due to their similar shape, size, and color.

4.2. Lesion Detection

DR lesion detection-based studies include studies that detected DR lesions like MA, EX, and HM. This section attempts to compare these studies using three metrics which include accuracy, sensitivity, and specificity. The values for these are stated in many studies. Table 6 provides a comparison of DR lesion-detection-based studies. The studies have been grouped according to the type of lesions they detected as well as according to the datasets used by them. In the case of EX, the maximum number of studies used the DIARET DB1, and the highest value for accuracy was obtained by the authors in [24] which is 0.2% higher than the second-best accuracy value obtained by the authors in [44]. The highest sensitivity value was obtained by the authors in [17], whereas the authors in [26] obtained the highest specificity value. From the table, it can also be seen that the E-Optha dataset has been used by many studies for HE and MA detection. In the first case, the highest accuracy and specificity values of 98.4% and 98.8%, respectively, were achieved by the authors in [39]. They also achieved the best accuracy and specificity values of 99.2% and 99.8%, respectively, for MA detection also. Whereas, the highest values of sensitivity for HE and MA detection on the E-Optha dataset were achieved by the authors in [22,53,54]. In the case of HM, the maximum accuracy of 99.4% on the MESSIDOR dataset was obtained by the authors in [59]. Whereas, the maximum accuracy value of 97.2% on the DIARET DB1 dataset was obtained by the authors in [58].
Figure 5 provides a comparison of MA detection results on the IDRiD dataset by the researchers in [35,43] and a comparison of HE detection results on the IDRiD dataset by the researchers in [35,43]. As seen in the figure, the authors in [43] could effectively detect HE using U-Net and MA using a modified U-Net. Whereas, the authors in [35] could obtain similar results using the JSeg model.
Figure 6 provide a further comparison of the results obtained by researchers for deep-learning-based HE and MA detection on the IDRiD dataset. It can be clearly seen that the authors in Sambyal et al. have obtained the best value of 99.9% for all three metrics for both HE and MA detection on the IDRiD dataset. This is followed by the study by the authors in Xue et al., which achieved very close values for accuracy and specificity. In the case of HE detection, all the studies could achieve accuracy values greater than 98% except for the study by the authors in Gupta et al. [48].
Just like DR segmentation, there are several challenges to performing DR lesion detection. If we take the case of MA detection, the main challenge is their very small size. But MA detection is very important to determine the onset of DR. EX detection is also challenging since they may not have well-defined boundaries. Additionally, as mentioned before, the OD and fovea may be mistakenly detected as lesions due to their similar shape, size, and color. Finally, the size of DR lesions, like EX, vary according to the severity of DR which makes it even more challenging.

5. Future Directions

Diabetic retinopathy is a topic in which a large amount of research is happening lately. In this section, we would like to present some possible directions for future work related to DR segmentation and DR lesion detection.
  • The number of studies which performed optic disc/fovea segmentation are relatively few in the reviewed literature. Hence, more research studies are needed in this area.
  • Similarly, in the case of DR lesion detection, more studies need to be performed for HM/SE detection.
  • Due to the serious complications that can result from proliferative DR, it is desirable to detect DR at an early stage. Hence, it will be beneficial to have more studies which focus on non-proliferative DR lesions.
  • Security is of prime importance for any proposed method for DR segmentation/lesion detection. Hence, studies like the recent one by the authors in [104] are required. They have used adversarial training and feature fusion for DR detection.
  • DR patients are at the risk of developing other conditions like glaucoma. Hence, more studies related to such conditions similar to the latest study by the authors in [105] are needed. They performed glaucoma detection by using a novel deep CNN.

6. Conclusions

The latest literature in the field of “diabetic retinopathy segmentation and lesion detection from fundus images using deep learning” was surveyed. The studies in this field can be grouped as blood vessel segmentation based studies, lesion detection based studies, and OD segmentation-based studies. Lesion detection-based studies can further be classified into MA detection, EX (soft and hard) detection, and HM detection.
We found that almost all recent DL networks/architectures have been utilized effectively for DR segmentation and lesion detection in the studies that were reviewed. It was also observed that there is a spike in the number of these kinds of studies recently. We also created a table of the generally utilized retinal fundus image datasets for DR retinal blood vessel segmentation and lesion detection. Finally, we performed a comparison based on the performance of studies from each type of DR study reviewed here. As future work, we may do a systematic literature review in this field.

Author Contributions

Conceptualization, A.S., O.E. and S.A.-M.; data curation, A.S.; formal analysis, A.S.; methodology, A.S., O.E. and S.A.-M.; project administration, S.A.-M. and N.A.; supervision, S.A.-M. and N.A.; validation, A.S., O.E., S.A.-M. and N.A.; visualization, A.S. and O.E.; writing—original draft, A.S.; writing—review and editing, A.S., O.E., S.A.-M. and N.A. All authors have read and agreed to the published version of the manuscript.

Funding

Open Access funding provided by the Qatar National Library.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This publication was supported by Qatar University Internal Grant QUHI-CENG22/23-548. The findings achieved herein are solely the responsibility of the authors.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

DRDiabetic retinopathy
DLDeep learning
AIArtificial intelligence
CNNConvolutional neural network
ODOptic disc
MAMicroaneurysms
HMHemorrhages
EXExudates
SESoft exudates
HEHard exudates

References

  1. IDF Diabetes Atlas. 9th Edition. Available online: https://diabetesatlas.org/atlas/ninth-edition/ (accessed on 30 March 2023).
  2. Understanding Diabetic Retinopathy and How to Reverse It. Available online: https://neoretina.com/blog/diabetic-retinopathy-can-it-be-reversed/ (accessed on 30 March 2023).
  3. Automated Retinal Image Analysis (ARIA) Data Set—Damian JJ Farnell. Available online: https://www.damianjjfarnell.com/?page_id=276 (accessed on 30 March 2023).
  4. Thangaraj, S.; Periyasamy, V.; Balaji, R. Retinal vessel segmentation using neural network. IET Image Process. 2018, 12, 669–678. [Google Scholar] [CrossRef]
  5. Patwari, M.B.; Manza, R.R.; Rajput, Y.M.; Saswade, M.; Deshpande, N. Detection and counting the microaneurysms using image processing techniques. Int. J. Appl. Inform. Syst 2013, 6, 11–17. [Google Scholar]
  6. Raja, S.S.; Vasuki, S. Screening diabetic retinopathy in developing countries using retinal images. Appl. Med. Inform. 2015, 36, 13–22. [Google Scholar]
  7. Harangi, B.; Hajdu, A. Automatic exudate detection by fusing multiple active contours and regionwise classification. Comput. Biol. Med. 2014, 54, 156–171. [Google Scholar] [CrossRef] [PubMed]
  8. Li, H.; Chutatape, O. Automated feature extraction in color retinal images by a model based approach. IEEE Trans. Biomed. Eng. 2004, 51, 246–254. [Google Scholar] [CrossRef] [PubMed]
  9. Garcia, M.; Sanchez, C.I.; Poza, J.; López, M.I.; Hornero, R. Detection of hard exudates in retinal images using a radial basis function classifier. Ann. Biomed. Eng. 2009, 37, 1448–1463. [Google Scholar]
  10. Sanchez, C.I.; Mayo, A.; Garcia, M.; Lopez, M.I.; Hornero, R. Automatic image processing algorithm to detect hard exudates based on mixture models. In Proceedings of the 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, New York, NY, USA, 30 August–3 September 2006; pp. 4453–4456. [Google Scholar]
  11. Niemeijer, M.; Van Ginneken, B.; Staal, J.; Suttorp-Schulten, M.S.; Abràmoff, M.D. Automatic detection of red lesions in digital color fundus photographs. IEEE Trans. Med. Imaging 2005, 24, 584–592. [Google Scholar] [CrossRef] [PubMed]
  12. Ghosh, S.K.; Ghosh, A. A novel retinal image segmentation using rSVM boosted convolutional neural network for exudates detection. Biomed. Signal Process. Control 2021, 68, 102785. [Google Scholar]
  13. Gour, N.; Khanna, P. Blood vessel segmentation using hybrid median filtering and morphological transformation. In Proceedings of the 2017 13th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), Jaipur, India, 4–7 December 2017; pp. 151–157. [Google Scholar]
  14. Quinn, E.A.E.; Krishnan, K.G. Retinal blood vessel segmentation using curvelet transform and morphological reconstruction. In Proceedings of the 2013 IEEE International Conference ON Emerging Trends in Computing, Communication and Nanotechnology (ICECCN), Tirunelveli, India, 25–26 March 2013; pp. 570–575. [Google Scholar]
  15. Bhardwaj, C.; Jain, S.; Sood, M. Automated optical disc segmentation and blood vessel extraction for fundus images using ophthalmic image processing. In Proceedings of the International Conference on Advanced Informatics for Computing Research, Gurugram, India, 18–19 December 2018; pp. 182–194. [Google Scholar]
  16. Raja, C.; Balaji, L. An automatic detection of blood vessel in retinal images using convolution neural network for diabetic retinopathy detection. Patt. Recogn. Image Anal. 2019, 29, 533–545. [Google Scholar]
  17. Adem, K. Exudate detection for diabetic retinopathy with circular Hough transformation and convolutional neural networks. Expert Syst. Appl. 2018, 114, 289–295. [Google Scholar] [CrossRef]
  18. Wang, H.; Yuan, G.; Zhao, X.; Peng, L.; Wang, Z.; He, Y.; Qu, C.; Peng, Z. Hard exudate detection based on deep model learned information and multi-feature joint representation for diabetic retinopathy screening. Comput. Methods Prog. Biomed. 2020, 191, 105398. [Google Scholar] [CrossRef]
  19. Kurilová, V.; Goga, J.; Oravec, M.; Pavlovičová, J.; Kajan, S. Support vector machine and deep-learning object detection for localisation of hard exudates. Sci. Rep. 2021, 11, 16045. [Google Scholar] [CrossRef]
  20. Wang, L.; Huang, Y.; Lin, B.; Wu, W.; Chen, H.; Pu, J. Automatic Classification of Exudates in Color Fundus Images Using an Augmented Deep Learning Procedure. In Proceedings of the Third International Symposium on Image Computing and Digital Medicine, Xi’an, China, 24–26 August 2019; pp. 31–35. [Google Scholar]
  21. Sivapriya, G.; Praveen, V.; Gowri, P.; Saranya, S.; Sweetha, S.; Shekar, K. Segmentation of Hard exudates for the detection of Diabetic Retinopathy with RNN based sematic features using fundus images. Mater. Today Proc. 2022, 64, 693–701. [Google Scholar] [CrossRef]
  22. Huang, C.; Zong, Y.; Ding, Y.; Luo, X.; Clawson, K.; Peng, Y. A new deep learning approach for the retinal hard exudates detection based on superpixel multi-feature extraction and patch-based CNN. Neurocomputing 2021, 452, 521–533. [Google Scholar] [CrossRef]
  23. Bibi, N.; Nida, N.; Irtaza, A.; Anwar, S.M. Automatic Detection of Exudates for Daignosis of Non-proliferative Diabetic Retinopathy using Region-based Convolutional Neural Networks. In Proceedings of the 2021 International Conference on Frontiers of Information Technology (FIT), Islamabad, Pakistan, 11–12 December 2021; pp. 218–223. [Google Scholar]
  24. Mateen, M.; Wen, J.; Nasrullah, N.; Sun, S.; Hayat, S. Exudate detection for diabetic retinopathy using pretrained convolutional neural networks. Complexity 2020, 2020, 5801870. [Google Scholar] [CrossRef]
  25. Cincan, R.G.; Popescu, D.; Ichim, L. Exudate Detection in Diabetic Retinopathy Using Deep Learning Techniques. In Proceedings of the 2021 25th International Conference on System Theory, Control and Computing (ICSTCC), Iași, Romania, 20–23 October 2021; pp. 473–477. [Google Scholar]
  26. Manan, M.A.; Khan, T.M.; Saadat, A.; Arsalan, M.; Naqvi, S.S. A Residual Encoder-Decoder Network for Segmentation of Retinal Image-Based Exudates in Diabetic Retinopathy Screening. arXiv 2022, arXiv:2201.05963. [Google Scholar]
  27. Xia, H.; Lan, Y.; Song, S.; Li, H. A multi-scale segmentation-to-classification network for tiny microaneurysm detection in fundus images. Knowl.-Based Syst. 2021, 226, 107140. [Google Scholar] [CrossRef]
  28. Qiao, L.; Zhu, Y.; Zhou, H. Diabetic retinopathy detection using prognosis of microaneurysm and early diagnosis system for non-proliferative diabetic retinopathy based on deep learning algorithms. IEEE Access 2020, 8, 104292–104302. [Google Scholar] [CrossRef]
  29. Gupta, S.; Panwar, A.; Kapruwan, A.; Chaube, N.; Chauhan, M. Real Time Analysis of Diabetic Retinopathy Lesions by Employing Deep Learning and Machine Learning Algorithms using Color Fundus Data. In Proceedings of the 2022 International Conference on Innovative Trends in Information Technology (ICITIIT), Kottayam, India, 12–13 February 2022; pp. 1–5. [Google Scholar]
  30. Qomariah, D.; Nopember, I.; Tjandrasa, H.; Fatichah, C. Segmentation of microaneurysms for early detection of diabetic retinopathy using MResUNet. Int. J. Intell. Eng. Syst. 2021, 14, 359–373. [Google Scholar] [CrossRef]
  31. Subhasree, A.; Princess, J.B. Analysis and Automatic Detection of Microaneurysms in Diabetic Retinopathy using transfer learning. In Proceedings of the 2022 6th International Conference on Computation System and Information Technology for Sustainable Solutions (CSITSS), Bangalore, India, 21–23 December 2022; pp. 1–7. [Google Scholar]
  32. Xiao, Q.; Zou, J.; Yang, M.; Gaudio, A.; Kitani, K.; Smailagic, A.; Costa, P.; Xu, M. Improving Lesion Segmentation for Diabetic Retinopathy using Adversarial Learning. In Proceedings of the International Conference on Image Analysis and Recognition, Waterloo, ON, USA, 27–29 August 2019; pp. 333–344. [Google Scholar]
  33. Praveena, S.; Lavanya, R. Superpixel based segmentation for multilesion detection in diabetic retinopathy. In Proceedings of the 2019 3rd International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, 23–25 April 2019; pp. 314–319. [Google Scholar]
  34. Dai, L.; Wu, L.; Li, H.; Cai, C.; Wu, Q.; Kong, H.; Liu, R.; Wang, X.; Hou, X.; Liu, Y.; et al. A deep learning system for detecting diabetic retinopathy across the disease spectrum. Nat. Commun. 2021, 12, 3242. [Google Scholar] [CrossRef] [PubMed]
  35. Amin, J.; Anjum, M.A.; Sharif, M. Fused information of DeepLabv3+ and transfer learning model for semantic segmentation and rich features selection using equilibrium optimizer (EO) for classification of NPDR lesions. Knowl.-Based Syst. 2022, 249, 108881. [Google Scholar] [CrossRef]
  36. Alyoubi, W.L.; Abulkhair, M.F.; Shalash, W.M. Diabetic retinopathy fundus image classification and lesions localization system using deep learning. Sensors 2021, 21, 3704. [Google Scholar] [CrossRef]
  37. Wang, L.; Chen, Z.; Wang, M.; Wang, T.; Zhu, W.; Chen, X. Cycle Adaptive Multi-Target Weighting Network For Automated Diabetic Retinopathy Segmentation. In Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), Nice, France, 13–16 April 2021; pp. 1141–1144. [Google Scholar]
  38. Santos, C.; de Aguiar, M.S.; Welfer, D.; Belloni, B.M. Detection of Fundus Lesions through a Convolutional Neural Network in Patients with Diabetic Retinopathy. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Guadalajara, Mexico, 26 July 2021; pp. 2692–2695. [Google Scholar]
  39. Xue, J.; Yan, S.; Qu, J.; Qi, F.; Qiu, C.; Zhang, H.; Chen, M.; Liu, T.; Li, D.; Liu, X. Deep membrane systems for multitask segmentation in diabetic retinopathy. Knowl.-Based Syst. 2019, 183, 104887. [Google Scholar] [CrossRef]
  40. Sambyal, N.; Saini, P.; Syal, R.; Gupta, V. Modified U-Net architecture for semantic segmentation of diabetic retinopathy images. Biocyber. Biomed. Eng. 2020, 40, 1094–1109. [Google Scholar] [CrossRef]
  41. He, W.; Wang, X.; Wang, L.; Huang, Y.; Yang, Z.; Yao, X.; Zhao, X.; Ju, L.; Wu, L.; Wu, L.; et al. Incremental learning for exudate and hemorrhage segmentation on fundus images. Inform. Fusion 2021, 73, 157–164. [Google Scholar] [CrossRef]
  42. Selçuk, T.; Beyoğlu, A.; Alkan, A. Automatic detection of exudates and hemorrhages in low-contrast color fundus images using multi semantic convolutional neural network. Concurr. Comput. Pract. Exp. 2022, 34, e6768. [Google Scholar] [CrossRef]
  43. Ananda, S.; Kitahara, D.; Hirabayashi, A.; Reddy, K.U.K. Automatic Fundus Image Segmentation for Diabetic Retinopathy Diagnosis by Multiple Modified U-Nets and SegNets. In Proceedings of the 2019 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Lanzhou, China, 18–21 November 2019; pp. 1582–1588. [Google Scholar]
  44. Abdelmaksoud, E.; El-Sappagh, S.; Barakat, S.; Abuhmed, T.; Elmogy, M. Automatic diabetic retinopathy grading system based on detecting multiple retinal lesions. IEEE Access 2021, 9, 15939–15960. [Google Scholar] [CrossRef]
  45. Kundu, S.; Karale, V.; Ghorai, G.; Sarkar, G.; Ghosh, S.; Dhara, A.K. Nested U-Net for Segmentation of Red Lesions in Retinal Fundus Images and Sub-image Classification for Removal of False Positives. J. Digit. Imag. 2022, 35, 1111–1119. [Google Scholar] [CrossRef]
  46. Ashraf, M.N.; Hussain, M.; Habib, Z. Deep Red Lesion Classification for Early Screening of Diabetic Retinopathy. Mathematics 2022, 10, 686. [Google Scholar] [CrossRef]
  47. Latchoumi, T.; Kumar, A.S.D.; Raja, J.Y. Detection of Diabetic Retinopathy with Ground-Truth Segmentation using Fundus Image. In Proceedings of the 2022 7th International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 22–24 June 2022; pp. 1770–1774. [Google Scholar]
  48. Gupta, S.; Panwar, A.; Goel, S.; Mittal, A.; Nijhawan, R.; Singh, A.K. Classification of lesions in retinal fundus images for diabetic retinopathy using transfer learning. In Proceedings of the 2019 International Conference on Information Technology (ICIT), Shanghai, China, 20–23 December 2019; pp. 342–347. [Google Scholar]
  49. Li, Q.; Peng, C.; Ma, Y.; Du, S.; Guo, B.; Li, Y. Pixel-level Diabetic Retinopathy Lesion Detection Using Multi-scale Convolutional Neural Network. In Proceedings of the 2021 IEEE 3rd Global Conference on Life Sciences and Technologies (LifeTech), Nara, Japan, 9–11 March 2021; pp. 438–440. [Google Scholar]
  50. Huang, S.; Li, J.; Xiao, Y.; Shen, N.; Xu, T. RTNet: Relation transformer network for diabetic retinopathy multi-lesion segmentation. IEEE Trans. Med. Imaging 2022, 41, 1596–1607. [Google Scholar] [CrossRef]
  51. Nazir, T.; Irtaza, A.; Rashid, J.; Nawaz, M.; Mehmood, T. Diabetic retinopathy lesions detection using faster-RCNN from retinal images. In Proceedings of the 2020 First International Conference of Smart Systems and Emerging Technologies (SMARTTECH), Riyadh, Saudi Arabia, 3–5 November 2020; pp. 38–42. [Google Scholar]
  52. Basu, S.; Mitra, S. Segmentation in Diabetic Retinopathy using Deeply-Supervised Multiscalar Attention. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Guadalajara, Mexico, 26 July 2021; pp. 2614–2617. [Google Scholar]
  53. Guo, Y.; Peng, Y. Multiple lesion segmentation in diabetic retinopathy with dual-input attentive RefineNet. Appl. Intell. 2022, 52, 14440–14464. [Google Scholar] [CrossRef]
  54. Guo, Y.; Peng, Y. CARNet: Cascade attentive RefineNet for multi-lesion segmentation of diabetic retinopathy images. Complex Intell. Syst. 2022, 8, 1681–1701. [Google Scholar] [CrossRef]
  55. Santos, C.; Aguiar, M.; Welfer, D.; Belloni, B. A New Approach for Detecting Fundus Lesions Using Image Processing and Deep Neural Network Architecture Based on YOLO Model. Sensors 2022, 22, 6441. [Google Scholar] [CrossRef] [PubMed]
  56. Jena, P.K.; Khuntia, B.; Palai, C.; Nayak, M.; Mishra, T.K.; Mohanty, S.N. A Novel Approach for Diabetic Retinopathy Screening Using Asymmetric Deep Learning Features. Big Data Cognit. Comput. 2023, 7, 25. [Google Scholar] [CrossRef]
  57. Skouta, A.; Elmoufidi, A.; Jai-Andaloussi, S.; Ouchetto, O. Hemorrhage semantic segmentation in fundus images for the diagnosis of diabetic retinopathy by using a convolutional neural network. J. Big Data 2022, 9, 1–24. [Google Scholar] [CrossRef]
  58. Aziz, T.; Charoenlarpnopparut, C.; Mahapakulchai, S. Deep learning-based hemorrhage detection for diabetic retinopathy screening. Sci. Rep. 2023, 13, 1479. [Google Scholar] [CrossRef] [PubMed]
  59. Maqsood, S.; Damaševičius, R.; Maskeliūnas, R. Hemorrhage detection based on 3D CNN deep learning framework and feature fusion for evaluating retinal abnormality in diabetic patients. Sensors 2021, 21, 3865. [Google Scholar] [CrossRef] [PubMed]
  60. Sebastian, A.; Elharrouss, O.; Al-Maadeed, S.; Almaadeed, N. A Survey on Deep-Learning-Based Diabetic Retinopathy Classification. Diagnostics 2023, 13, 345. [Google Scholar] [CrossRef]
  61. Elasri, M.; Elharrouss, O.; Al-Maadeed, S.; Tairi, H. Image Generation: A Review. Neur. Process. Lett. 2022, 54, 4609–4646. [Google Scholar] [CrossRef]
  62. Al-Mohannadi, A.; Al-Maadeed, S.; Elharrouss, O.; Sadasivuni, K.K. Encoder-decoder architecture for ultrasound IMC segmentation and cIMT measurement. Sensors 2021, 21, 6839. [Google Scholar] [CrossRef]
  63. Elharrouss, O.; Akbari, Y.; Almaadeed, N.; Al-Maadeed, S. Backbones-review: Feature extraction networks for deep learning and deep reinforcement learning approaches. arXiv 2022, arXiv:2206.08016. [Google Scholar]
  64. Riahi, A.; Elharrouss, O.; Al-Maadeed, S. BEMD-3DCNN-based method for COVID-19 detection. Comput. Biol. Med. 2022, 142, 105188. [Google Scholar] [CrossRef] [PubMed]
  65. Elharrouss, O.; Al-Maadeed, S.; Subramanian, N.; Ottakath, N.; Almaadeed, N.; Himeur, Y. Panoptic segmentation: A review. arXiv 2021, arXiv:2111.10250. [Google Scholar]
  66. Saranya, P.; Prabakaran, S.; Kumar, R.; Das, E. Blood vessel segmentation in retinal fundus images for proliferative diabetic retinopathy screening using deep learning. Vis. Comput. 2022, 38, 977–9926. [Google Scholar] [CrossRef]
  67. Chala, M.; Nsiri, B.; El yousfi Alaoui, M.H.; Soulaymani, A.; Mokhtari, A.; Benaji, B. An automatic retinal vessel segmentation approach based on Convolutional Neural Networks. Expert Syst. Appl. 2021, 184, 115459. [Google Scholar] [CrossRef]
  68. Aujih, A.; Izhar, L.; Mériaudeau, F.; Shapiai, M.I. Analysis of retinal vessel segmentation with deep learning and its effect on diabetic retinopathy classification. In Proceedings of the 2018 International conference on intelligent and advanced system (ICIAS), Kuala Lumpur, Malaysia, 13–14 August 2018; pp. 1–6. [Google Scholar]
  69. Burewar, S.; Gonde, A.B.; Vipparthi, S.K. Diabetic retinopathy detection by retinal segmentation with region merging using CNN. In Proceedings of the 2018 IEEE 13th International Conference on Industrial and Information Systems (ICIIS), Rupnagar, India, 1–2 December 2018; pp. 136–142. [Google Scholar]
  70. Yadav, N. A deep data-driven approach for enhanced segmentation of blood vessel for diabetic retinopathy. Int. J. Imaging Syst. Technol. 2022, 32, 1696–1708. [Google Scholar] [CrossRef]
  71. Gargari, M.S.; Seyedi, M.H.; Alilou, M. Segmentation of Retinal Blood Vessels Using U-Net++ Architecture and Disease Prediction. Electronics 2022, 11, 3516. [Google Scholar] [CrossRef]
  72. Zhang, Y.; He, M.; Chen, Z.; Hu, K.; Li, X.; Gao, X. Bridge-Net: Context-involved U-net with patch-based loss weight mapping for retinal blood vessel segmentation. Expert Syst. Appl. 2022, 195, 116526. [Google Scholar] [CrossRef]
  73. Sathananthavathi, V.; Indumathi, G. Encoder enhanced atrous (EEA) unet architecture for retinal blood vessel segmentation. Cognit. Syst. Res. 2021, 67, 84–95. [Google Scholar]
  74. Jebaseeli, T.J.; Durai, C.A.D.; Peter, J.D. Retinal blood vessel segmentation from diabetic retinopathy images using tandem PCNN model and deep learning based SVM. Optik 2019, 199, 163328. [Google Scholar] [CrossRef]
  75. Jebaseeli, T.J.; Durai, C.A.D.; Peter, J.D. Segmentation of retinal blood vessels from ophthalmologic diabetic retinopathy images. Comput. Electr. Eng. 2019, 73, 245–258. [Google Scholar] [CrossRef]
  76. Arsalan, M.; Haider, A.; Lee, Y.W.; Park, K.R. Detecting retinal vasculature as a key biomarker for deep Learning-based intelligent screening and analysis of diabetic and hypertensive retinopathy. Expert Syst. Appl. 2022, 200, 117009. [Google Scholar] [CrossRef]
  77. Samuel, P.M.; Veeramalai, T. Multilevel and multiscale deep neural network for retinal blood vessel segmentation. Symmetry 2019, 11, 946. [Google Scholar] [CrossRef]
  78. Jin, Q.; Meng, Z.; Pham, T.D.; Chen, Q.; Wei, L.; Su, R. DUNet: A deformable network for retinal vessel segmentation. Knowl.-Based Syst. 2019, 178, 149–162. [Google Scholar] [CrossRef]
  79. Wu, Y.; Xia, Y.; Song, Y.; Zhang, Y.; Cai, W. NFN+: A novel network followed network for retinal vessel segmentation. Neur. Netw. 2020, 126, 153–162. [Google Scholar] [CrossRef] [PubMed]
  80. Yan, Z.; Yang, X.; Cheng, K.T. A three-stage deep learning model for accurate retinal vessel segmentation. IEEE J. Biomed. Health Inform. 2018, 23, 1427–1436. [Google Scholar] [CrossRef]
  81. Fu, Q.; Li, S.; Wang, X. MSCNN-AM: A multi-scale convolutional neural network with attention mechanisms for retinal vessel segmentation. IEEE Access 2020, 8, 163926–163936. [Google Scholar] [CrossRef]
  82. Li, Q.; Fan, S.; Chen, C. An intelligent segmentation and diagnosis method for diabetic retinopathy based on improved U-NET network. J. Med. Syst. 2019, 43, 304. [Google Scholar] [CrossRef]
  83. Elaouaber, Z.; Feroui, A.; Lazouni, M.; Messadi, M. Blood vessel segmentation using deep learning architectures for aid diagnosis of diabetic retinopathy. Comput. Meth. Biomech. Biomed. Eng. Imaging Vis. 2022, 1–15. [Google Scholar] [CrossRef]
  84. Liu, Z. Construction and verification of color fundus image retinal vessels segmentation algorithm under BP neural network. J. Supercomput. 2021, 77, 7171–7183. [Google Scholar] [CrossRef]
  85. Prajna, Y.; Nath, M.K. Efficient blood vessel segmentation from color fundus image using deep neural network. J. Intell. Fuzzy Syst. 2022, 42, 3477–3489. [Google Scholar] [CrossRef]
  86. Kar, M.K.; Neog, D.R.; Nath, M.K. Retinal vessel segmentation using multi-scale residual convolutional neural network (MSR-Net) combined with generative adversarial networks. Circ. Syst. Sign. Process. 2023, 42, 1206–1235. [Google Scholar] [CrossRef]
  87. Popescu, D.; Deaconu, M.; Ichim, L.; Stamatescu, G. Retinal blood vessel segmentation using pix2pix gan. In Proceedings of the 2021 29th Mediterranean Conference on Control and Automation (MED), Bari, Italy, 22–25 June 2021; pp. 1173–1178. [Google Scholar]
  88. Deng, X.; Ye, J. A retinal blood vessel segmentation based on improved D-MNet and pulse-coupled neural network. Biomed. Sign. Process. Control 2022, 73, 103467. [Google Scholar] [CrossRef]
  89. Maiti, S.; Maji, D.; Dhara, A.K.; Sarkar, G. Automatic detection and segmentation of optic disc using a modified convolution network. Biomed. Sign. Process. Control 2022, 76, 103633. [Google Scholar] [CrossRef]
  90. Kumar, S.; Adarsh, A.; Kumar, B.; Singh, A.K. An automated early diabetic retinopathy detection through improved blood vessel and optic disc segmentation. Opt. Laser Technol. 2020, 121, 105815. [Google Scholar] [CrossRef]
  91. Yeh, H.; Lin, C.J.; Hsu, C.C.; Lee, C.Y. Deep-learning based automated segmentation of Diabetic Retinopathy symptoms. In Proceedings of the 2020 International Symposium on Computer, Consumer and Control (IS3C), Taichung, Taiwan, 13–16 November 2020; pp. 497–499. [Google Scholar]
  92. Hasan, M.K.; Alam, M.A.; Elahi, M.T.E.; Roy, S.; Martí, R. DRNet: Segmentation and localization of optic disc and fovea from diabetic retinopathy image. Artif. Intell. Med. 2021, 111, 102001. [Google Scholar] [CrossRef]
  93. Shankar, K.; Sait, A.R.W.; Gupta, D.; Lakshmanaprabu, S.; Khanna, A.; Pandey, H.M. Automated detection and classification of fundus diabetic retinopathy images using synergic deep learning model. Patt. Recogn. Lett. 2020, 133, 210–216. [Google Scholar] [CrossRef]
  94. Mansour, R.F. Deep-learning-based automatic computer-aided diagnosis system for diabetic retinopathy. Biomed. Eng. Lett. 2018, 8, 41–57. [Google Scholar] [CrossRef]
  95. Qureshi, I.; Ma, J.; Abbas, Q. Diabetic retinopathy detection and stage classification in eye fundus images using active deep learning. Multim. Tools Appl. 2021, 80, 11691–11721. [Google Scholar] [CrossRef]
  96. Staal, J.; Abràmoff, M.D.; Niemeijer, M.; Viergever, M.A.; Van Ginneken, B. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef]
  97. Structured Analysis of the Retina. Available online: https://cecas.clemson.edu/~ahoover/stare/ (accessed on 30 March 2023).
  98. Porwal, P.; Pachade, S.; Kamble, R.; Kokare, M.; Deshmukh, G.; Sahasrabuddhe, V.; Meriaudeau, F. Indian diabetic retinopathy image dataset (IDRiD): A database for diabetic retinopathy screening research. Data 2018, 3, 25. [Google Scholar] [CrossRef]
  99. E-ophtha. Available online: https://www.adcis.net/en/third-party/e-ophtha/ (accessed on 30 March 2023).
  100. Kauppi, T.; Kalesnykiene, V.; Kamarainen, J.K.; Lensu, L.; Sorri, I.; Raninen, A.; Voutilainen, R.; Uusitalo, H.; Kälviäinen, H.; Pietilä, J. The diaretdb1 diabetic retinopathy database and evaluation protocol. In Proceedings of the BMVC, Warwick, UK, 10–13 September 2007; Volume 1, pp. 1–10. [Google Scholar]
  101. Decencière, E.; Zhang, X.; Cazuguel, G.; Lay, B.; Cochener, B.; Trone, C.; Gain, P.; Ordonez, R.; Massin, P.; Erginay, A.; et al. Feedback on a publicly distributed image database: The Messidor database. Image Anal. Stereol. 2014, 33, 231–234. [Google Scholar] [CrossRef]
  102. Li, T.; Gao, Y.; Wang, K.; Guo, S.; Liu, H.; Kang, H. Diagnostic assessment of deep learning algorithms for diabetic retinopathy screening. Inform. Sci. 2019, 501, 511–522. [Google Scholar] [CrossRef]
  103. Bhardwaj, C.; Jain, S.; Sood, M. Appraisal of pre-processing techniques for automated detection of diabetic retinopathy. In Proceedings of the 2018 Fifth International Conference on Parallel, Distributed and Grid Computing (PDGC), Solan, India, 20–22 December 2018; pp. 734–739. [Google Scholar]
  104. Lal, S.; Rehman, S.U.; Shah, J.H.; Meraj, T.; Rauf, H.T.; Damaševičius, R.; Mohammed, M.A.; Abdulkareem, K.H. Adversarial attack and defence through adversarial training and feature fusion for diabetic retinopathy recognition. Sensors 2021, 21, 3922. [Google Scholar] [CrossRef] [PubMed]
  105. Mahum, R.; Rehman, S.U.; Okon, O.D.; Alabrah, A.; Meraj, T.; Rauf, H.T. A novel hybrid approach based on deep CNN to detect glaucoma using fundus imaging. Electronics 2021, 11, 26. [Google Scholar] [CrossRef]
Figure 1. Diabetic Retinopathy Lesions [2].
Figure 1. Diabetic Retinopathy Lesions [2].
Applsci 13 05111 g001
Figure 2. Visualization of a healthy retina and an unhealthy retina [3].
Figure 2. Visualization of a healthy retina and an unhealthy retina [3].
Applsci 13 05111 g002
Figure 3. Categorisation of Diabetic Retinopathy studies reviewed.
Figure 3. Categorisation of Diabetic Retinopathy studies reviewed.
Applsci 13 05111 g003
Figure 4. Retinal blood vessel segmentation results on the DRIVE dataset using [75,76] methods, respectively. Left: Original Image. Middle: Ground-truth, Right: Segmentation Result.
Figure 4. Retinal blood vessel segmentation results on the DRIVE dataset using [75,76] methods, respectively. Left: Original Image. Middle: Ground-truth, Right: Segmentation Result.
Applsci 13 05111 g004
Figure 5. Comparison of deep-learning-based diabetic retinopathy microaneurysm detection and hard exudates detection results performed by [35,43] respectively on the IDRiD dataset.
Figure 5. Comparison of deep-learning-based diabetic retinopathy microaneurysm detection and hard exudates detection results performed by [35,43] respectively on the IDRiD dataset.
Applsci 13 05111 g005
Figure 6. Performance comparison of deep-learning based diabetic retinopathy with hard exudate and microaneurysm detection studies which employed the IDRiD dataset [22,28,31,33,38,40,41,50,53,54].
Figure 6. Performance comparison of deep-learning based diabetic retinopathy with hard exudate and microaneurysm detection studies which employed the IDRiD dataset [22,28,31,33,38,40,41,50,53,54].
Applsci 13 05111 g006
Table 2. Deep-learning based approaches for diabetic retinopathy segmentation.
Table 2. Deep-learning based approaches for diabetic retinopathy segmentation.
StudyYearMethodDataset
Aujih et al. [68]2018U-Net, InceptionV1MESSIDOR
Burewar et al. [69]2018U-Net segmentationDRIVE
Yan et al. [80]2018CNNSTARE, DRIVE, CHASE-DB1
Jebaseeli et al. [74]2019TPCNN, DLBSVMSTARE, DRIVE, HRF, REVIEW, DRIONS
Li et al. [82]2019U-Net-
Jebaseeli et al. [75]2019TPCNN, DLBSVMSTARE, DRIVE, HRF, REVIEW, DRIONS
Samuel and Veeramalai [77]2019DNNSTARE, DRIVE, HRF
Jin et al. [78]2019DUNetSTARE, DRIVE, HRF, CHASE-DB1
Yeh et al. [91]2020U-Net architectureSelf-collected
Kumar et al. [90]2020RBFNNDIARET DB0,DIARET DB1
Shankar et al. [93]2020Deep CNN model-SDLMESSIDOR
Wu et al. [79]2020 N F N + STARE, DRIVE, CHASE-DB1
Fu et al. [81]2020MSCNN-AMSTARE, DRIVE, CHASE-DB1
Hasan et al. [92]2020DRNetIDRiD, HRF, DRIVE
Liu [84]2021Back propagation neural networkMESSIDOR, DRIVE
Qureshi et al. [95]2021ADL-CNNEyePACS
Saranya et al. [66]2021VGG-16 net architectureDRIVE, STARE
Chala et al. [67]2021Multi-encoder decoder architectureSTARE, DRIVE
Popescu et al. [87]2021Pix2Pix GANCHASE DB1, DRIVE, STARE
Sathananthavathi and Indumathi [73]2021Encoder enhanced atrous U-NetCHASE DB1, DRIVE, STARE, HRF
Deng and Ye [88]2022D-MNet, pulse-coupled neural networkCHASE DB1, DRIVE, STARE, HRF
Yadav [70]2022U-Net architectureCHASE DB1, DRIVE, STARE
Gargari et al. [71]2022U-Net++ architectureDRIVE, MESSIDOR
Zhang et al. [72]2022Context-involved U-NetDRIVE, CHASE-DB1, STARE, HRF
Arsalan et al. [76]2022PLRS-Net, PLS-NetDRIVE, CHASE-DB1, STARE
Maiti et al. [89]2022Encoder-decoder with VGG-16DRIVE, DIARETDB0, CHASE-DB1, STARE
Elaouaber et al. [83]2022SegNet, U-Net, CNNDRIVE, CHASE-DB1, HRF
Prajna and Nath [85]2022AlexNet, U-NetDRIVE, ARIA, MESSIDOR
Kar et al. [86]2022MSR-Net, GANDRIVE, ARIA, CHASE-DB1, STARE, HRF
Table 3. Datasets used for Diabetic Retinopathy Segmentation/Lesion Detection.
Table 3. Datasets used for Diabetic Retinopathy Segmentation/Lesion Detection.
DatasetNo. of ImagesImage Size
DIARETDB01301500 × 1152
DIARETDB1891500 × 1152
STARE400700 ×605
DRIVE40768 × 584
IDRiD5164288 × 2848
E-Ophtha463Different sizes
MESSIDOR1200Different sizes
HRF453504 × 2336
REVIEW161360 × 1024
DRIONS110600× 400
CHASE DB128999 × 960
DDR13,673Different sizes
ARIA143768 × 576
Table 4. Performance comparison of deep-learning based diabetic retinopathy segmentation.The bold and underlined fonts correspondingly show the first and second place.
Table 4. Performance comparison of deep-learning based diabetic retinopathy segmentation.The bold and underlined fonts correspondingly show the first and second place.
DatasetMethodAccuracySensitivitySpecificity
STAREJebaseeli et al. [74]0.9970.8060.997
Chala et al. [67]0.9650.810.984
Yadav [70]0.9860.812-
Saranya et al. [66]0.950.9380.99
Arsalan et al. [76]0.9720.8640.98
Jebaseeli et al. [75]0.9910.7070.996
Samuel and Veeramalai [77]0.9650.8980.97
Jin et al. [78]0.964--
Wu et al. [79]0.9670.9860.796
Yan et al. [80]0.9640.7740.986
Fu et al. [81]0.9660.8410.981
Kar et al. [86]0.9490.8440.964
Popescu et al. [87]0.9150.8180.942
Zhang et al. [72]0.9670.80.986
Sathananthavathi and Indumathi [73]0.9450.8020.956
Deng and Ye [88]0.9730.8270.985
DRIVEJebaseeli et al. [74]0.990.8030.998
Chala et al. [67]0.9720.8210.988
Yadav [70]0.9770.821-
Hasan et al. [92]0.9990.962-
Burewar et al. [69]0.933--
Arsalan et al. [76]0.9680.8270.982
Jebaseeli et al. [75]0.990.7010.995
Samuel and Veeramalai [77]0.9610.8280.974
Jin et al. [78]0.957--
Wu et al. [79]0.9580.7990.981
Yan et al. [80]0.9540.7630.982
Fu et al. [81]0.9560.8340.973
Elaouaber et al. [83]0.9770.9670.996
Liu [84]0.9580.8140.906
Gargari et al. [71]0.9890.9410.988
Prajna and Nath [85]0.9060.9180.905
Kar et al. [86]0.9740.8940.988
Popescu et al. [87]0.9210.8340.961
Zhang et al. [72]0.9570.7850.982
Sathananthavathi and Indumathi [73]0.9580.7920.971
Deng and Ye [88]0.9680.8360.981
Table 5. Performance comparison of deep-learning based diabetic retinopathy segmentation. The bold and underlined fonts correspondingly show the first and second place.
Table 5. Performance comparison of deep-learning based diabetic retinopathy segmentation. The bold and underlined fonts correspondingly show the first and second place.
DatasetMethodAccuracySensitivitySpecificity
REVIEWJebaseeli et al. [74]0.9990.8090.988
Jebaseeli et al. [75]0.9950.6930.989
HRFJebaseeli et al. [74]0.990.8080.997
Jebaseeli et al. [75]0.990.8260.994
Samuel and Veeramalai [77]0.8530.8660.852
Jin et al. [78]0.965--
Elaouaber et al. [83]0.9790.9790.995
Kar et al. [86]0.9770.8890.985
Zhang et al. [72]0.960.8570.969
Sathananthavathi and Indumathi [73]0.9240.6590.98
Deng and Ye [88]0.9670.7540.985
DRIONS
Jebaseeli et al. [74]0.9990.8050.998
Jebaseeli et al. [75]0.9930.7950.997
CHASE DB1Yadav [70]0.960.802-
Arsalan et al. [76]0.9730.830.984
Jin et al. [78]0.961--
Wu et al. [79]0.9690.80.988
Yan et al. [80]0.9610.7640.981
Fu et al. [81]0.9640.8130.982
Elaouaber et al. [83]0.9550.8370.982
Kar et al. [86]0.9870.9340.984
Popescu et al. [87]0.9340.9170.969
Zhang et al. [72]0.9670.8130.984
Sathananthavathi and Indumathi [73]0.9340.6460.965
Deng and Ye [88]0.9710.8540.979
IDRiDWang et al. [37]0.9830.4850.999
Hasan et al. [92]0.9970.899-
MESSIDORAujih et al. [68]0.84--
Liu [84]0.9620.8520.911
Prajna and Nath [85]0.8340.7970.844
ARIAKar et al. [86]0.9630.7180.984
Prajna and Nath [85]0.9250.5660.961
Table 6. Performance comparison of deep-learning based diabetic retinopathy lesion-detection. The bold and underlined fonts correspondingly show the first and second place.
Table 6. Performance comparison of deep-learning based diabetic retinopathy lesion-detection. The bold and underlined fonts correspondingly show the first and second place.
TypeDatasetMethodAccuracySensitivitySpecificity
ExudatesIDRiDSambyal et al. [40]0.9970.8060.997
Abdelmaksoud et al. [44]111
MESSIDORSelçuk et al. [42]0.9580.9870.915
Sivapriya et al. [21]0.970.950.99
E-OpthaBibi et al. [23]-0.960.99
Manan et al. [26]0.990.92-
Mateen et al. [24]0.984--
DIARET DB1Abdelmaksoud et al. [44]0.9870.970.962
Manan et al. [26]0.980.95-
Cincan et al. [25]0.9790.960.96
Adem [17]-0.9920.979
Bibi et al. [23]-0.980.94
Manan et al. [26]0.980.950.99
Selçuk et al. [42]0.9490.9790.905
Mateen et al. [24]0.989--
Hard ExudatesDDRGuo and Peng [53]0.9740.9660.978
Guo and Peng [54]0.9740.9660.978
E-OpthaXue et al. [39]0.9840.8460.988
Wang et al. [18]-0.899-
Guo and Peng [53]0.9750.9650.985
Guo and Peng [54]0.9750.9650.985
Huang et al. [22]0.9760.9830.912
MESSIDORSivapriya et al. [21]0.9730.950.99
MicroaneurysmE-OpthaXue et al. [39]0.9920.6720.998
Guo and Peng [53]0.9490.9350.958
Guo and Peng [54]0.9490.9350.958
Subhasree et al. [31]0.960.90.85
Xia et al. [27]-0.740-
DDRGuo and Peng [53]0.9540.9360.965
Guo and Peng [54]0.9540.9360.965
DIARET DB1Abdelmaksoud et al. [44]0.980.960.98
Qomariah et al. [30]0.9970.8590.998
Li et al. [49]-0.88-
Subhasree et al. [31]0.910.940.88
HemorrhageDIARET DB0Maqsood et al. [59]0.955--
Aziz et al. [58]0.9710.910.98
DIARET DB1Maqsood et al. [59]0.964--
Selçuk et al. [42]0.9430.9720.896
Aziz et al. [58]0.9720.9420.975
MESSIDORMaqsood et al. [59]0.994--
Selçuk et al. [42]0.960.9870.917
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sebastian, A.; Elharrouss, O.; Al-Maadeed, S.; Almaadeed, N. A Survey on Diabetic Retinopathy Lesion Detection and Segmentation. Appl. Sci. 2023, 13, 5111. https://doi.org/10.3390/app13085111

AMA Style

Sebastian A, Elharrouss O, Al-Maadeed S, Almaadeed N. A Survey on Diabetic Retinopathy Lesion Detection and Segmentation. Applied Sciences. 2023; 13(8):5111. https://doi.org/10.3390/app13085111

Chicago/Turabian Style

Sebastian, Anila, Omar Elharrouss, Somaya Al-Maadeed, and Noor Almaadeed. 2023. "A Survey on Diabetic Retinopathy Lesion Detection and Segmentation" Applied Sciences 13, no. 8: 5111. https://doi.org/10.3390/app13085111

APA Style

Sebastian, A., Elharrouss, O., Al-Maadeed, S., & Almaadeed, N. (2023). A Survey on Diabetic Retinopathy Lesion Detection and Segmentation. Applied Sciences, 13(8), 5111. https://doi.org/10.3390/app13085111

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop