Next Article in Journal
Updates on Triple-Negative Breast Cancer in Type 2 Diabetes Mellitus Patients: From Risk Factors to Diagnosis, Biomarkers and Therapy
Previous Article in Journal
A Two-Step Single Plex PCR Method for Evaluating Key Colonic Microbiota Markers in Young Mexicans with Autism Spectrum Disorders: Protocol and Pilot Epidemiological Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Supervised Contrastive Learning with Angular Margin for the Detection and Grading of Diabetic Retinopathy

1
Academy for Engineering & Technology, Fudan University, Shanghai 200433, China
2
School of Information Science and Technology, Fudan University, Shanghai 200433, China
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(14), 2389; https://doi.org/10.3390/diagnostics13142389
Submission received: 6 June 2023 / Revised: 6 July 2023 / Accepted: 13 July 2023 / Published: 17 July 2023
(This article belongs to the Section Medical Imaging and Theranostics)

Abstract

:
Many researchers have realized the intelligent medical diagnosis of diabetic retinopathy (DR) from fundus images by using deep learning methods, including supervised contrastive learning (SupCon). However, although SupCon brings label information into the calculation of contrastive learning, it does not distinguish between augmented positives and same-label positives. As a result, we propose the concept of Angular Margin and incorporate it into SupCon to address this issue. To demonstrate the effectiveness of our strategy, we tested it on two datasets for the detection and grading of DR. To align with previous work, Accuracy, Precision, Recall, F1, and AUC were selected as evaluation metrics. Moreover, we also chose alignment and uniformity to verify the effect of representation learning and UMAP (Uniform Manifold Approximation and Projection) to visualize fundus image embeddings. In summary, DR detection achieved state-of-the-art results across all metrics, with Accuracy = 98.91, Precision = 98.93, Recall = 98.90, F1 = 98.91, and AUC = 99.80. The grading also attained state-of-the-art results in terms of Accuracy and AUC, which were 85.61 and 93.97, respectively. The experimental results demonstrate that Angular Margin is an excellent intelligent medical diagnostic algorithm, performing well in both DR detection and grading tasks.

1. Introduction

According to the International Diabetes Federation’s (IDF) Diabetes Atlas, 10th edition, 537 million adults over the age of 20 have diabetes [1]. Diabetes-related eye disease affects at least one-third of all diabetics, of which diabetic retinopathy (DR) is the most common [2]. DR occurs when excessive blood sugar levels damage the capillaries and blood vessels in the retina [3]. If the eye is treated early in the disease’s progression, it can significantly reduce the chance of blindness by up to 98% [4,5]. Therefore, regular ophthalmic screening and timely diagnoses for diabetic patients are necessary [6]. An early clinical diagnosis mainly relies on a fundus examination. Color Fundus Photography (CFP) is a rapid, noninvasive, well-tolerated, and widely used imaging technique to assess the grading of DR [7]. However, the manual CFP-based diagnosis of DR requires the expertise and efforts of specialized ophthalmologists. Especially in underdeveloped areas with large populations, it is difficult to have sufficient medical resources to meet local needs [8]. These conditions highlight the need for the development of intelligent computer-aided diagnosis systems.
With improvements in computational resources and capabilities, deep-learning-related technologies are widely used. A large number of intelligent medical diagnostic technologies are constantly being proposed. In particular, the popular contrastive learning technology has left a deep impression on academia in the past two years [9,10]. Supervised contrastive learning (SupCon) currently achieves great performance in the detection and grading of DR [11]. SupCon is essentially an embedding representation learning method. It makes up for the problem that the cross-entropy loss function is sensitive to noisy data and hyperparameter changes [12]. However, there is certainly an opportunity for development with this method. SupCon adds the supervision signal of the labels to the loss function and treats the same-label data and the augmented data as same-level positives in a coarse-grained manner. But actually, when taking an image as an anchor, the augmented data should be closer to the anchor than the same-label data in the embedding space. For this consideration, our research proposes Angular Margin on the basis of SupCon, which enhances the representation embedding of the fundus image by setting different margins (Figure 1). The experimental results reveal that the Angular Margin approach achieves remarkable performance on the APTOS 2019 dataset for DR detection, with Accuracy = 98.91, Precision = 98.93, Recall = 98.90, F1 = 98.91, and AUC = 99.80. Moreover, for DR grading on the same dataset, it achieves Accuracy = 85.61, Precision = 78.44, Recall = 68.32, F1 = 71.67, and AUC = 93.97, which are significantly superior to the results of other methods. Furthermore, when directly comparing these performance metrics with the SupCon method on the Messidor-2 dataset, the Angular Margin approach demonstrates a substantial improvement across all aforementioned metrics.
For a better theoretical understanding of the role of Angular Margin, we introduced alignment and uniformity [13] into the training process of SupCon. In the embedding space, alignment measures the semantic correlation between positive samples, and uniformity measures the distribution of all samples. The comparison experiment shows that the addition of Angular Margin can improve SupCon’s performance in the above two analysis metrics. Moreover, we also conducted a visual analysis on both SupCon and Angular Margin to strengthen the experimental conclusions. UMAP (Uniform Manifold Approximation and Projection) [14] is a advanced manifold learning technique for dimensionality reduction. It is competitive with t-SNE in terms of visualization quality and preserves more global structure with an excellent runtime performance. In this work, we used UMAP to reduce the dimensionality of the image embeddings to a two-dimensional plane for visualization and compared Angular Margin with the original SupCon.
To sum up, the main innovations of this study are as follows:
(1)
We propose a novel intelligent medical diagnostic method called “Angular Margin”, which, within the framework of SupCon, enhances the Accuracy of fundus image representation by additionally introducing Angular Margin. This method achieves state-of-the-art performance in both DR detection and grading tasks.
(2)
Alignment and uniformity are innovatively introduced as evaluation metrics to assess the representational capacity of CFP embeddings.
(3)
UMAP is introduced to project the embeddings of image representations onto a two-dimensional plane for visualization, enhancing the observation of fundus embeddings.

2. Related Work

The development of SupCon is inseparable from self-supervised representation learning, metric distance learning, and supervised learning. Excellent self-supervised representation learning algorithms [9,15] have greatly contributed to academia by enabling noise contrastive estimation and N-pair losses [16,17]. Generally, the loss function is connected to the network’s last layer. But at inference time, it is more likely to use the embeddings of the previous layer for downstream tasks. The loss based on metric distance learning with triplets is similar to contrastive learning [18,19]. The positive and negative logarithms of each data anchor are what distinguishes triplet loss from contrastive loss. Triplet loss uses only one positive and negative pair per anchor [20,21,22]. However, contrastive loss has only one positive pair, but the number of negative pairs is very large. This makes hard-negative mining unnecessary in contrastive loss. Later, inspired by [23,24,25], SupCon [10] was derived, which normalizes embeddings and replaces Euclidean distances with inner products. At the same time, the model capability is further enhanced by using various data augmentation methods. Similar to self-supervision, in the classification task, SupCon adopts a two-stage training method (contrastive learning and cross-entropy) and discards the contrastive head in the second stage.
In recent years, in the field of fundus diagnosis, a large number of deep-learning-related works have also emerged. DR binary and multiclassification works by [26,27,28] were based on the traditional cross-entropy loss function with convolutional neural networks. Quite a few attention mechanism networks [29,30,31,32] and ensemble algorithms [33,34] have also been applied to DR diagnosis. Moreover, several recent studies [35,36,37] have employed more advanced CNN models or introduced additional auxiliary models to enhance the grading of DR. And some approaches [38,39] have utilized a multi-stage training methodology to progressively improve the model’s performance. All these methods have involved good innovations in the structure of the model and achieved the state-of-the-art results at that time. However, the network structures in most of these works are complex and large. IsIam et al. [11] recently brought DR detection and classification to a new level without adding additional modules by introducing SupCon. This shows the potential of SupCon for the application of fundus imaging. In addition, Huang et al. [40] proposed a self-supervised model for lesion-based contrastive learning for DR grading, and it performs well on the EyePACS dataset. Cai et al. [41] used optical coherence tomography (OCT) and CFP for multi-modality supervised contrastive learning to diagnose glaucoma. In two papers, Cheng et al. [42,43] used two contrastive learning methods to improve the imaging quality of fundus images.

3. Methodology

This section mainly introduces the datasets used in the experiments and the details of our methods. The innovative methods proposed in this section have yielded excellent results, which can be reviewed in Section 4.

3.1. Datasets

Currently, there are many datasets for DR detection and grading, but to align with previous methods, we chose APTOS 2019 (download link: https://www.kaggle.com/c/aptos2019-blindness-detection (accessed on 5 March 2022)) and Messidor-2 [44,45].
APTOS 2019 was provided by Aravind Eye Hospital in India and includes five grades to identify and assess the severity of DR: no DR, mild, moderate, severe, and proliferative diabetic retinopathy (PDR). The number of training samples in the dataset is 3662, and the testing part consists of 1928 samples. The image sizes are 2416 × 1736, 819 × 614, and 3216 × 2136. Although test samples are available, their labels are not public and are only available for online submission for testing. Therefore, our work does not consider the test part.
The Messidor-2 dataset consists of more than 1700 images of various sizes, such as 1440 × 960, 2240 × 1488, and 2340 × 1536, and is used for DR grading. The partial dataset (Messidor-Original) was kindly provided by the partners of the Messidor project, and the rest (Messidor-Extension) was provided by the University Hospital of Brest, France.
We further split the above dataset into training (70%), validation (15%), and testing (15%) sets for DR detection and severity grading. The data have the problem of being highly imbalanced, so they need to be sampled evenly when splitting. Multiclass classification is performed for the five-stage grading of DR severity, and binary classification is carried out for DR detection. Table 1 summarizes the specifics of the dataset.

3.2. Contrastive Learning Framework

SupCon is an advanced deep learning method that aims to improve the discriminative power of learned representations. By leveraging labeled data, SupCon uses a contrastive loss function to encourage similar representations for data instances from the same class while pushing apart representations from different classes. This approach facilitates the extraction of meaningful and discriminative features for tasks such as classification and clustering. SupCon maximizes the agreement between augmented views of the same sample and minimizes the agreement between different samples, promoting the learning of informative representations. It has shown success in various domains and offers a promising direction for representation learning by utilizing unlabeled data.
Our framework is primarily composed of a data augmentation module ( A u g ( · ) ), an encoder network ( E n c ( · ) ), a projection layer ( P r o j ( · ) ), and a classifier ( C l a s s ( · ) ). The pipeline of the whole framework is shown in Figure 2 in detail.
For each input data x, we randomly generate two augmentations x ^ = A u g ( x ) . Each augmentation represents a different view of the image and contains some subset information of the original data. The image augmentation methods operate on the principle of making simple direct modifications to an image, such as cropping, rotating, adjusting the color, and changing the image contrast. These techniques aim to generate another image that is visually similar to the original one. We used RandomResizedCrop, RandomHorizontalFlip, and RandomVerticalFlip (the augmentation methods come from the transforms module of torchvision; official link: https://pytorch.org/vision/stable/transforms.html (accessed on 10 May 2022)) to construct the data augmentation module.
The encoder is a convolutional neural network (CNN)-structured backbone that maps x to a vector r = E n c ( x ) R D E as the representation embedding. To obtain a pair of representation embeddings, two augmentations are fed into the same encoder. We follow the works of [10,13,19] to normalize r to the unit hypersphere in R D E . The normalization utilized here involves scaling each component of the embedding to fall within the range between 0 and 1. In this work, ResNet [46] was chosen as the encoder. ResNet is a convolutional neural network architecture that addresses the degradation problem in deep networks. Proposed in 2015, it introduces residual learning and utilizes residual blocks with skip connections to learn residual mappings. ResNet models, such as ResNet-18 and ResNet-50, have achieved impressive performance in computer vision tasks. Its widespread adoption has made it a benchmark in the field and influenced subsequent network designs.
The projection transforms r to the vector z = P r o j ( r ) R D P . P r o j ( · ) is instantiated as a 2-layer perceptron [47] with a 2048 hidden layer size and a 128 output layer size. Previous studies [10,11] also normalized the P r o j ( · ) result to lie on the unit hypersphere so that the inner product could be used to measure distances in a projective space. At the end of training, we drop P r o j ( · ) so that the contrastive learning framework has the same number of parameters as the cross-entropy model.
The classifier’s objective is to compute the cross-entropy loss. Because our main idea in this research is the classification of DR, even with Angular Margin loss, it is still necessary to train a classifier for classification.

3.3. Angular Margin Loss Function

To better demonstrate the Angular Margin loss, we need to introduce a family of contrastive learning losses. For a set of N randomly sampled data, { x k , y k } k = 1 , . . . , N , so there are 2N pairs of corresponding batches actually used for training after augmentation, { x ^ l , y ^ l , } l = 1 , . . . , 2 N . The collection of N data is referred to as a “batch”, and the set of 2N augmented data is referred to as a “double-viewed batch”. This setting is common to all contrastive learning losses.
Within a double-viewed batch, i I { 1 , . . . , 2 N } represents the index of any data, and j ( i ) is the corresponding augmented data. In the self-supervised paradigm, the loss usually takes the following form:
L s e l f = i I L i s e l f = i I log exp ( z i · z j ( i ) / τ ) a A ( i ) exp ( z i · z a / τ )
In the formula, z l = P r o j ( E n c ( x ^ l ) ) R D P , the symbol · is the vector dot product, τ R + represents the temperature parameter, and A ( i ) I { i } . The index i can be regarded as an anchor, j ( i ) is the corresponding positive, and all other samples in the double-viewed batch are regarded as negatives. So, in the self-supervised paradigm [9,20], each sample has 1 positive pair and 2 N 2 negative pairs. However, the label information of the data is not taken into account, and the coarse-grained thinking that there is only one positive pair in a double-view batch is also problematic.
To address this issue, Khosla et al. [10] proposed a supervised paradigm. Equation (2) extends the contrastive learning loss to multiple positive pairs by introducing label information:
L s u p = i I L i s u p = i I 1 | P ( i ) | p P ( i ) log exp ( z i · z p / τ ) a A ( i ) exp ( z i · z a / τ )
P ( i ) is the set of positives in the double-viewed batch. In this way, in the framework of contrastive learning, an anchor can have multiple positives and negatives. But at the same time, multiple positive pairs also expose the potential problem that not every positive is at the same similarity level for the anchor. As shown in Figure 1, the augmented data from anchor data should naturally be closer to the anchor than the same-label data. Therefore, we propose the Angular Margin loss to increase the degree of discrimination between positives.
To add the angle calculation to the loss, we need to replace the previous vector dot product with cosine similarity z i · z p z i z p . Let us assume that θ i , p has the following form:
θ i , p = arccos ( z i · z p z i z p )
Through Equation (3), we can convert the dot product of vectors into an angle representation so that we can add margins to increase the discrimination between positives.
L a m = i I L i a m L i a m = i I 1 | P ( i ) | p P ( i ) log exp ( cos ( θ i , p + I U p m u + I V p m v ) / τ ) a A ( i ) exp ( cos ( θ i , a + I U a m u + I V a m v ) / τ )
U indicates the augmented positives in the double-viewed batch, and V indicates the same-label positives; both belong to P ( i ) ( | U | + | V | = | P ( i ) | ). I { 0 , 1 } is the indicator function, which is equal to 1 if and only if the superscript belongs to the subscript. To strengthen the discriminative power, we add m u and m v , corresponding to augmented positives and same-label positives, respectively (Figure 3). Their values are expressed in radians.
The final loss function (Equation (6)) is obtained by combining Equation (4) and the cross-entropy loss function (Equation (5)) through the hyperparameter λ .
L c e = i = 1 k y i log ( p i ) , m u l t i c l a s s [ y log ( p ) + ( 1 y ) log ( 1 p ) ] , b i n a r y
L = L c e + λ L a m

3.4. Alignment and Uniformity

Wang et al. [13] discovered two important properties related to contrastive learning, alignment and uniformity. Given the normalized distribution of positive pairs p p o s , alignment computes the expectation of the distance between embeddings of the pairs:
L a l i g n E ( x , y ) p p o s f ( x ) f ( y ) 2
At the same time, given a dataset p d a t a , uniformity measures the degree of uniform distribution of the embeddings.
L u n i f o r m log E ( x , y ) i . i . d . p d a t a e 2 f ( x ) f ( y ) 2
For the above two metrics, the smaller the value, the better the learning of the embedding representation. In this work, we use them to demonstrate that our method is better than the standard SupCon.

4. Experiments and Results

4.1. Experimental Protocol

To pick the best model for the test set, we trained the network for 20 epochs and tested it on the validation set at the conclusion of each epoch. The size of the input image was 224 × 224. The augmentation methods were RandomResizedCrop (scale = (0.8, 1.0)), RandomHorizontalFlip (p = 0.5), and RandomVerticalFlip (p = 0.5). The batch size was 64, the learning rate was 2 × 10 4 , and the optimizer used was Adam. These hyperparameters were obtained through empirical means. Among other hyperparameters, τ was 0.05, m u was 0.2, m v was 0.1, and λ was 1. The device we used was NVIDIA A4000 GPU.

4.2. Metrics for Classification

There are five evaluation metrics, namely, Accuracy, Precision, Recall, F1-score, and AUC. Accuracy refers to the proportion of correctly predicted samples to total samples. The F1-score (Equation (11)) is a combination of Precision and Recall. It will perform well only if both Precision and Recall scores are high. Precision (Equation (9)) and Recall (Equation (10)) consist of True Positives (TP), False Positives (FP), and False Negatives (FN). AUC (Area Under the Curve) is defined as the area contained by the ROC curve and the coordinate axis. The values of the above metrics are all between 0 and 1. For convenience, we express them in percentiles (%).
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l

4.3. Results

We compared our approach with state-of-the-art methods on the APTOS 2019 dataset (Table 2). The APTOS 2019 dataset consists of a large number of fundus images, and significant research efforts have been devoted to training and testing on this dataset. Therefore, the comparative experiments presented in this study were conducted on the APTOS 2019 dataset. For the binary classification task, our method obtains state-of-the-art performance across all metrics, improving by more than 0.5% over the already achieved 98% baseline. But in multiclassification tasks, our method only achieves state-of-the-art results for Accuracy and AUC. A possible reason may be that Bodapati et al. [26] used more data (their training set accounts for 80% of APTOS 2019, whereas ours is only 70%) with their method and introduced more fully connected network layers (additional FC1 and FC2).
To assess the effectiveness of Angular Margin in its entirety, the experiment also compared differences in alignment and uniformity between SupCon and Angular Margin. The lower the values of the two metrics, the better the representation effect. We took the first 1500 steps of the training process and calculated the alignment and uniformity of the current batch every 10 steps. Figure 4 and Figure 5 depict alignment and uniformity, respectively. In Figure 4, all lines gradually decrease, indicating that the alignment gradually decreases with training until it becomes stable. After 200 to 600 steps, the line of Angular Margin is obviously located below that of SupCon, showing that Angular Margin can bring the positives closer. However, in Figure 5, the value of uniformity either increases gradually or oscillates repeatedly, failing to drop to lower values like alignment does. We estimate that there may be a lot of data with the same labels in a batch because the number of classes is too small (only two classes for detection and five classes for grading). Therefore, after optimizing the alignment, it is difficult for the embeddings to maintain a standard uniform distribution. Even so, from the performance in the figure, the effect of Angular Margin is still better than that of SupCon.
To understand the advantages and characteristics of Angular Margin more specifically, we used UMAP [14] for visualization. UMAP is a dimensionality reduction technique used to visualize high-dimensional data in a lower-dimensional space. It preserves both the local and global structures of the data by constructing a topological representation. UMAP is computationally efficient, can handle large datasets, and does not require labeled data. It has gained popularity in machine learning, bioinformatics, and data visualization for capturing complex data structures and revealing patterns. UMAP concatenates 2048-dimensional representation embeddings and projects them onto a 2-dimensional plane. Figure 6 shows the results of SupCon and Angular Margin on the APTOS 2019 and Messidor-2 test sets. The binary classification task is relatively simple, so we chose the multiclassification task for visual representation. This allows a clearer gap to be seen in the visualization. Judging from the results in Figure 6, since the classification results of APTOS 2019 are better than those of Messidor-2, the discrimination shown in Figure 6a,b is better than that in Figure 6c,d. On the same dataset, the distinguishing ability of the representation of Angular Margin on the two-dimensional projection is stronger than that of SupCon. Figure 6a shows that SupCon is able to clearly divide the embeddings into normal and DR, but the grading of DR is still vague. Figure 6b introduces Angular Margin, which can more clearly distinguish between severe (3.0) and PDR (4.0). Similarly, although the clustering of embeddings in Figure 6c,d is not very clear, separate independent clusters are more easily discerned in Figure 6d than in Figure 6c.

4.4. Ablation Study

Ablation experiments based on the original SupCon and Angular Margin were conducted on two datasets, APTOS 2019 and Messidor-2. Table 3 records the results of experiments on APTOS 2019. It can be seen that each method works well in the task of binary classification, but Angular Margin is still one percentage point higher on average. The advantage of Angular Margin is amplified in the multiclassification task, with an average improvement of more than 4 percentage points. Table 4 lists the results of Messidor-2 and also depicts a similar situation. There is an average improvement of more than 4 percentage points in binary classification and multiclassification. The results in the two tables reflect that Angular Margin can indeed make a big improvement compared to SupCon.
There are two particularly important hyperparameters in the Angular Margin loss function, m u and m v . Because these two angle values are directly related to the relative angle relationship between positives and anchor data, they will affect the distinguishing effect of representation embedding. For this reason, this study deliberately conducted detailed ablation experiments to explore its specific impact on the results. The experiments were compared on the basis of multiclassification tasks on the Messidor-2 and APTOS 2019 datasets. In Figure 7, each polyline has a certain value of m v , and there are five values in total: 0.02, 0.05, 0.1, 0.2, and 0.4. The vertical axis is the value of AUC on the test set. Compared with other evaluation indicators, AUC is more balanced and comprehensive. The horizontal axis represents the value of m u , which is a multiple of m v , because it is easier to find a reliable rule than arbitrary values. The minimum multiple of m u is 1.0, which represents the special case where m u and m v have the same value. Angular Margin will degenerate to a SupCon-like form. The highest multiple of m u is 3.0, because the angle value calculated by cosine similarity has a range, and too large an angle value will lead to more errors. In fact, the results in Figure 7 also reflect this. The five polylines at the position of 1.0 are relatively close, because they all have a structure similar to SupCon. The position at 3.0 is the worst performance of each polyline, and some are even lower than the results of SupCon, because the excessive m u has led to errors in the loss calculation. From the perspective of the two subgraphs, m v is 0.1, and m u is twice as high as m v , which is the best setting.
The augmentation module of contrastive learning has always been a key point and has a very large impact on the results. In this study, six commonly used image augmentation methods were selected, and the relationship between them is described with a heatmap. The six augmentation methods are RandomResizedCrop, RandomHorizontalFlip, RandomGrayscale, RandomVerticalFlip, ColorJitter, and RandomRotation. In Figure 8, the values in the heatmap are the results for AUC on the test set. Because the horizontal and vertical axes on the main diagonal refer to the same method, the augmentation module contains only one augmentation method. The parts other than the main diagonal refer to two augmentation methods. From the results in the figure, it can be seen that RandomResizedCrop, RandomHorizontalFlip, and RandomVerticalFlip have the best compatibility. They contribute the most to the improvement in results.

5. Conclusions

Inspired by SupCon, this paper proposes an improved method called Angular Margin based on it. To test the effect of this training strategy, in this study, the theory and formula were first deduced, and then detailed tests were conducted on APTOS 2019 and Messidor-2. Angular Margin could better distinguish positives with the anchor and achieved state-of-the-art results according to multiple classification metrics by combining the cross-entropy loss function. Alignment and uniformity were also introduced in the experiment, and the embedding effect of Angular Margin was found to be better than that of ordinary SupCon through comparison. In addition to these numerical comparisons, we also include visualizations of representational embeddings. A visualization of embeddings projected to 2D space by UMAP clearly shows that the results for Angular Margin are more spatially clustered. Finally, various ablation experiments were performed to investigate the role of Angular Margin loss, the effect of the angular value, and the difference between different augmentation methods as comprehensively as possible. To summarize, this paper confirms that Angular Margin performs better than SupCon in the intelligent medical diagnosis of DR.
The results of the experiments indicate that the proposed method is relatively successful, but there are still some areas worthy of improvement. First of all, the improvement scheme that we propose must not be optimal. Determining how to better use the contrastive learning framework and apply it to actual scenarios is a very valuable direction. Second, the presented experiments were all conducted on public datasets. In the future, we look forward to actively seeking cooperation with hospitals, hoping to obtain more data to conduct some clinical experiments. Finally, Angular Margin is a mechanism for representing images in general. We believe that it has the potential to be applied to other elements of ophthalmological medical diagnostics, as well as the diagnosis of other human tissues. This part of the work can be left to other researchers for further refinement.

Author Contributions

Conceptualization, D.Z. and A.G.; methodology, D.Z. and A.G.; software, D.Z.; validation, D.Z., X.C., Q.W., J.W. and S.L.; formal analysis, X.C.; investigation, D.Z.; resources, D.Z.; data curation, D.Z.; writing—original draft preparation, D.Z.; writing—review and editing, A.G.; visualization, X.C.; supervision, A.G.; project administration, A.G.; funding acquisition, A.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Yiwu Research Institute of Fudan University (grant number 20-1-20) and the recipient is Aiming Ge.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found here: https://www.kaggle.com/c/aptos2019-blindness-detection (accessed on 5 March 2022).

Acknowledgments

Thanks to all medical institutions or units that published the datasets. Relevant links and citations are indicated in the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. International Diabetes Federation. International Diabetes Federation Diabetes Atlas, 10th ed.; IDF: Brussels, Belgium, 2021; Available online: https://diabetesatlas.org (accessed on 12 February 2022).
  2. Jenkins, A.J.; Joglekar, M.V.; Hardikar, A.A.; Keech, A.C.; O’Neal, D.N.; Januszewski, A.S. Biomarkers in diabetic retinopathy. Rev. Diabet. Stud. RDS 2015, 12, 159. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Boyd, K. American Academy of Ophthalmology—What Is Diabetic Retinopathy; American Academy of Ophthalmology: San Francisco, CA, USA, 2021. [Google Scholar]
  4. Bresnick, G.H.; Mukamel, D.B.; Dickinson, J.C.; Cole, D.R. A screening approach to the surveillance of patients with diabetes for the presence of vision-threatening retinopathy. Ophthalmology 2000, 107, 19–24. [Google Scholar] [CrossRef] [PubMed]
  5. Hill, L.; Makaroff, L.E. Early detection and timely treatment can prevent or delay diabetic retinopathy. Diabetes Res. Clin. Pract. 2016, 120, 241–243. [Google Scholar] [CrossRef] [PubMed]
  6. Logue, J. Management of Obesity: A National Clinical Guideline; NHS Quality Improvement Scotland: Edinburgh, UK, 2010. [Google Scholar]
  7. Kwan, C.C.; Fawzi, A.A. Imaging and biomarkers in diabetic macular edema and diabetic retinopathy. Curr. Diabetes Rep. 2019, 19, 1–10. [Google Scholar] [CrossRef] [PubMed]
  8. Resnikoff, S.; Felch, W.; Gauthier, T.M.; Spivey, B. The number of ophthalmologists in practice and training worldwide: A growing gap despite more than 200,000 practitioners. Br. J. Ophthalmol. 2012, 96, 783–787. [Google Scholar] [CrossRef] [Green Version]
  9. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A Simple Framework for Contrastive Learning of Visual Representations. In Proceedings of the 37th International Conference on Machine Learning, Virtual, 13–18 July 2020; Volume 119, pp. 1597–1607. [Google Scholar]
  10. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; Krishnan, D. Supervised Contrastive Learning. Adv. Neural Inf. Process. Syst. 2020, 33, 18661–18673. [Google Scholar]
  11. Islam, M.R.; Abdulrazak, L.F.; Nahiduzzaman, M.; Goni, M.O.F.; Anower, M.S.; Ahsan, M.; Haider, J.; Kowalski, M. Applying supervised contrastive learning for the detection of diabetic retinopathy and its severity levels from fundus images. Comput. Biol. Med. 2022, 146, 105602. [Google Scholar] [CrossRef]
  12. Zhang, Z.; Sabuncu, M. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels. In Advances in Neural Information Processing Systems; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; NIPS: Grenada, Spain, 2018; Volume 31. [Google Scholar]
  13. Wang, T.; Isola, P. Understanding Contrastive Representation Learning through Alignment and Uniformity on the Hypersphere. In Proceedings of the 37th International Conference on Machine Learning, Virtual, 13–18 July 2020; Volume 119, pp. 9929–9939. [Google Scholar]
  14. McInnes, L.; Healy, J.; Saul, N.; Grossberger, L. UMAP: Uniform Manifold Approximation and Projection. J. Open Source Softw. 2018, 3, 861. [Google Scholar] [CrossRef]
  15. Wu, Z.; Xiong, Y.; Yu, S.X.; Lin, D. Unsupervised feature learning via non-parametric instance discrimination. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3733–3742. [Google Scholar]
  16. Gutmann, M.; Hyvärinen, A. Noise-contrastive estimation: A new estimation principle for unnormalized statistical models. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; Volume 9, pp. 297–304. [Google Scholar]
  17. Sohn, K. Improved Deep Metric Learning with Multi-class N-pair Loss Objective. In Advances in Neural Information Processing Systems; Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R., Eds.; NIPS: Grenada, Spain, 2016; Volume 29. [Google Scholar]
  18. Weinberger, K.Q.; Blitzer, J.; Saul, L. Distance metric learning for large margin nearest neighbor classification. In Advances in Neural Information Processing Systems; Weiss, Y., Schölkopf, B., Platt, J., Eds.; NIPS: Grenada, Spain, 2005; Volume 18. [Google Scholar]
  19. Schroff, F.; Kalenichenko, D.; Philbin, J. Facenet: A unified embedding for face recognition and clustering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 815–823. [Google Scholar]
  20. Hjelm, R.D.; Fedorov, A.; Lavoie-Marchildon, S.; Grewal, K.; Bachman, P.; Trischler, A.; Bengio, Y. Learning deep representations by mutual information estimation and maximization. arXiv 2018, arXiv:1808.06670. [Google Scholar] [CrossRef]
  21. Henaff, O. Data-Efficient Image Recognition with Contrastive Predictive Coding. In Proceedings of the 37th International Conference on Machine Learning, Virtual, 13–18 July 2020; Volume 119, pp. 4182–4192. [Google Scholar]
  22. Tian, Y.; Krishnan, D.; Isola, P. Contrastive multiview coding. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 776–794. [Google Scholar] [CrossRef]
  23. Wu, Z.; Efros, A.A.; Yu, S.X. Improving generalization via scalable neighborhood component analysis. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 685–701. [Google Scholar]
  24. Frosst, N.; Papernot, N.; Hinton, G. Analyzing and Improving Representations with the Soft Nearest Neighbor Loss. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; Volume 97, pp. 2012–2020. [Google Scholar]
  25. Kamnitsas, K.; Castro, D.; Folgoc, L.L.; Walker, I.; Tanno, R.; Rueckert, D.; Glocker, B.; Criminisi, A.; Nori, A. Semi-Supervised Learning via Compact Latent Space Clustering. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Volume 80, pp. 2459–2468. [Google Scholar]
  26. Bodapati, J.D.; Shaik, N.S.; Naralasetti, V. Composite deep neural network with gated-attention mechanism for diabetic retinopathy severity classification. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 9825–9839. [Google Scholar] [CrossRef]
  27. Dondeti, V.; Bodapati, J.D.; Shareef, S.N.; Veeranjaneyulu, N. Deep Convolution Features in Non-linear Embedding Space for Fundus Image Classification. Rev. Intell. Artif. 2020, 34, 307–313. [Google Scholar] [CrossRef]
  28. Gangwar, A.K.; Ravi, V. Diabetic retinopathy detection using transfer learning and deep learning. In Evolution in Computational Intelligence; Springer: Singapore, 2021; Volume 1176, pp. 679–689. [Google Scholar] [CrossRef]
  29. Li, X.; Hu, X.; Yu, L.; Zhu, L.; Fu, C.W.; Heng, P.A. CANet: Cross-disease attention network for joint diabetic retinopathy and diabetic macular edema grading. IEEE Trans. Med. Imaging 2019, 39, 1483–1493. [Google Scholar] [CrossRef] [Green Version]
  30. Mahmoud, M.H.; Alamery, S.; Fouad, H.; Altinawi, A.; Youssef, A.E. An automatic detection system of diabetic retinopathy using a hybrid inductive machine learning algorithm. Pers. Ubiquitous Comput. 2021, 27, 751–765. [Google Scholar] [CrossRef]
  31. Afrin, R.; Shill, P.C. Automatic lesions detection and classification of diabetic retinopathy using fuzzy logic. In Proceedings of the 2019 International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST), Dhaka, Bangladesh, 10–12 January 2019; pp. 527–532. [Google Scholar] [CrossRef]
  32. Lal, S.; Rehman, S.U.; Shah, J.H.; Meraj, T.; Rauf, H.T.; Damaševičius, R.; Mohammed, M.A.; Abdulkareem, K.H. Adversarial attack and defence through adversarial training and feature fusion for diabetic retinopathy recognition. Sensors 2021, 21, 3922. [Google Scholar] [CrossRef]
  33. Zhang, W.; Zhong, J.; Yang, S.; Gao, Z.; Hu, J.; Chen, Y.; Yi, Z. Automated identification and grading system of diabetic retinopathy using deep neural networks. Knowl.-Based Syst. 2019, 175, 12–25. [Google Scholar] [CrossRef]
  34. Kaushik, H.; Singh, D.; Kaur, M.; Alshazly, H.; Zaguia, A.; Hamam, H. Diabetic retinopathy diagnosis from fundus images using stacked generalization of deep models. IEEE Access 2021, 9, 108276–108292. [Google Scholar] [CrossRef]
  35. Sugeno, A.; Ishikawa, Y.; Ohshima, T.; Muramatsu, R. Simple methods for the lesion detection and severity grading of diabetic retinopathy by image processing and transfer learning. Comput. Biol. Med. 2021, 137, 104795. [Google Scholar] [CrossRef]
  36. Minarno, A.E.; Mandiri, M.H.C.; Azhar, Y.; Bimantoro, F.; Nugroho, H.A.; Ibrahim, Z. Classification of diabetic retinopathy disease using convolutional neural network. JOIV Int. J. Inform. Vis. 2022, 6, 12–18. [Google Scholar] [CrossRef]
  37. Kobat, S.G.; Baygin, N.; Yusufoglu, E.; Baygin, M.; Barua, P.D.; Dogan, S.; Yaman, O.; Celiker, U.; Yildirim, H.; Tan, R.S.; et al. Automated diabetic retinopathy detection using horizontal and vertical patch division-based pre-trained DenseNET with digital fundus images. Diagnostics 2022, 12, 1975. [Google Scholar] [CrossRef]
  38. Oulhadj, M.; Riffi, J.; Chaimae, K.; Mahraz, A.M.; Ahmed, B.; Yahyaouy, A.; Fouad, C.; Meriem, A.; Idriss, B.A.; Tairi, H. Diabetic retinopathy prediction based on deep learning and deformable registration. Multimed. Tools Appl. 2022, 81, 28709–28727. [Google Scholar] [CrossRef]
  39. Zhang, C.; Lei, T.; Chen, P. Diabetic retinopathy grading by a source-free transfer learning approach. Biomed. Signal Process. Control 2022, 73, 103423. [Google Scholar] [CrossRef]
  40. Huang, Y.; Lin, L.; Cheng, P.; Lyu, J.; Tang, X. Lesion-based contrastive learning for diabetic retinopathy grading from fundus images. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2021, Proceedings of the 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; Volume 12902, pp. 113–123. [Google Scholar] [CrossRef]
  41. Cai, Z.; Lin, L.; He, H.; Tang, X. Corolla: An efficient multi-modality fusion framework with supervised contrastive learning for glaucoma grading. In Proceedings of the 2022 IEEE 19th International Symposium on Biomedical Imaging (ISBI), Kolkata, India, 28–31 March 2022; pp. 1–4. [Google Scholar] [CrossRef]
  42. Cheng, P.; Lin, L.; Huang, Y.; Lyu, J.; Tang, X. I-secret: Importance-guided fundus image enhancement via semi-supervised contrastive constraining. In Medical Image Computing and Computer Assisted Intervention–MICCAI 2021, Proceedings of the 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; Volume 12908, pp. 87–96. [Google Scholar] [CrossRef]
  43. Cheng, P.; Lin, L.; Huang, Y.; Lyu, J.; Tang, X. Prior guided fundus image quality enhancement via contrastive learning. In Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), Nice, France, 13–16 April 2021; pp. 521–525. [Google Scholar] [CrossRef]
  44. Abràmoff, M.D.; Folk, J.C.; Han, D.P.; Walker, J.D.; Williams, D.F.; Russell, S.R.; Massin, P.; Cochener, B.; Gain, P.; Tang, L.; et al. Automated analysis of retinal images for detection of referable diabetic retinopathy. JAMA Ophthalmol. 2013, 131, 351–357. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Decencière, E.; Zhang, X.; Cazuguel, G.; Lay, B.; Cochener, B.; Trone, C.; Gain, P.; Ordonez, R.; Massin, P.; Erginay, A.; et al. Feedback on a publicly distributed image database: The Messidor database. Image Anal. Stereol. 2014, 33, 231–234. [Google Scholar] [CrossRef] [Green Version]
  46. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  47. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning; Springer Series in Statistics; Springer: New York, NY, USA, 2001. [Google Scholar]
  48. Bodapati, J.D.; Naralasetti, V.; Shareef, S.N.; Hakak, S.; Bilal, M.; Maddikunta, P.K.R.; Jo, O. Blended multi-modal deep convnet features for diabetic retinopathy severity prediction. Electronics 2020, 9, 914. [Google Scholar] [CrossRef]
  49. Kassani, S.H.; Kassani, P.H.; Khazaeinezhad, R.; Wesolowski, M.J.; Schneider, K.A.; Deters, R. Diabetic retinopathy classification using a modified xception architecture. In Proceedings of the 2019 IEEE international symposium on signal processing and information technology (ISSPIT), Ajman, United Arab Emirates, 10–12 December 2019; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. The figure shows the role of Angular Margin in SupCon from fundus images. In the traditional SupCon, both the augmented data and the same-label data in the batch belong to positive samples. But in practice, it is reasonable that the augmented data should be closer to the anchor in the embedding space, because they are from the same image. The addition of Angular Margin can solve this problem well, making the representation of fundus images more accurate in the embedding space.
Figure 1. The figure shows the role of Angular Margin in SupCon from fundus images. In the traditional SupCon, both the augmented data and the same-label data in the batch belong to positive samples. But in practice, it is reasonable that the augmented data should be closer to the anchor in the embedding space, because they are from the same image. The addition of Angular Margin can solve this problem well, making the representation of fundus images more accurate in the embedding space.
Diagnostics 13 02389 g001
Figure 2. The whole model training process is from left to right. A fundus image can generate two sub-images with different views after passing through the data augmentation module. Afterward, through the encoder network, sub-images can be turned into embeddings. Finally, the Angular Margin loss and cross-entropy loss are calculated through the output of the projection layer and the classifier, respectively. The two losses are combined to form a completed training framework.
Figure 2. The whole model training process is from left to right. A fundus image can generate two sub-images with different views after passing through the data augmentation module. Afterward, through the encoder network, sub-images can be turned into embeddings. Finally, the Angular Margin loss and cross-entropy loss are calculated through the output of the projection layer and the classifier, respectively. The two losses are combined to form a completed training framework.
Diagnostics 13 02389 g002
Figure 3. The figure shows the angle relation of representation embedding. We assume that z i is an anchor, z a is the corresponding negative, and z p is the positive. After introducing Angular Margin, z u corresponds to the augmented positive, and z v corresponds to the same-label positive. Because m u and m v are set to different sizes, z u is pushed closer to the anchor than z v .
Figure 3. The figure shows the angle relation of representation embedding. We assume that z i is an anchor, z a is the corresponding negative, and z p is the positive. After introducing Angular Margin, z u corresponds to the augmented positive, and z v corresponds to the same-label positive. Because m u and m v are set to different sizes, z u is pushed closer to the anchor than z v .
Diagnostics 13 02389 g003
Figure 4. The figure shows the situation of alignment for SupCon and Angular Margin. Graphs in (a,b) represent binary classification and multiclassification on APTOS 2019. Similarly, (c,d) represent the two classification tasks on Messidor-2.
Figure 4. The figure shows the situation of alignment for SupCon and Angular Margin. Graphs in (a,b) represent binary classification and multiclassification on APTOS 2019. Similarly, (c,d) represent the two classification tasks on Messidor-2.
Diagnostics 13 02389 g004
Figure 5. The figure shows the situation of uniformity for SupCon and Angular Margin. Graphs in (a,b) represent binary classification and multiclassification on APTOS 2019. Similarly, (c,d) represent the two classification tasks on Messidor-2.
Figure 5. The figure shows the situation of uniformity for SupCon and Angular Margin. Graphs in (a,b) represent binary classification and multiclassification on APTOS 2019. Similarly, (c,d) represent the two classification tasks on Messidor-2.
Diagnostics 13 02389 g005
Figure 6. (a,b) The results of SupCon and Angular Margin on APTOS 2019. (c,d) The results of SupCon and Angular Margin on Messidor-2. 0.0 to 4.0, which indicate normal to PDR, and the larger the number, the darker the color, and the more serious the DR.
Figure 6. (a,b) The results of SupCon and Angular Margin on APTOS 2019. (c,d) The results of SupCon and Angular Margin on Messidor-2. 0.0 to 4.0, which indicate normal to PDR, and the larger the number, the darker the color, and the more serious the DR.
Diagnostics 13 02389 g006
Figure 7. Comparison of ablation experiments with angle values, m u and m v . (a) The case of the Messidor-2 dataset and (b) the case of the APTOS 2019 dataset.
Figure 7. Comparison of ablation experiments with angle values, m u and m v . (a) The case of the Messidor-2 dataset and (b) the case of the APTOS 2019 dataset.
Diagnostics 13 02389 g007
Figure 8. The relationship heatmap of 6 commonly used augmentation modules. (a) The results of Messidor-2 and (b) the results of APTOS 2019. The effects of pairwise combinations of augmentation methods on the experimental results are described in the form of diagonal matrices.
Figure 8. The relationship heatmap of 6 commonly used augmentation modules. (a) The results of Messidor-2 and (b) the results of APTOS 2019. The effects of pairwise combinations of augmentation methods on the experimental results are described in the form of diagonal matrices.
Diagnostics 13 02389 g008
Table 1. The table covers each component of APTOS 2019 and Messidor-2 after division. In the multiclass part, there are serious data imbalances in both datasets.
Table 1. The table covers each component of APTOS 2019 and Messidor-2 after division. In the multiclass part, there are serious data imbalances in both datasets.
ClassificationDR StageTraining PartValidation PartTesting Part
APTOS 2019Messidor-2APTOS 2019Messidor-2APTOS 2019Messidor-2
MulticlassNo DR1263711271153271153
Mild25918956405541
Moderate6992431505215052
Severe1355229122911
PDR20725445445
BinaryNo DR1263711271153271153
DR1300509279109278109
Table 2. Our method compared with previous state-of-the-art methods in binary and multiclassification tasks on APTOS 2019. The data in the table are from the results in the original papers of the respective authors. Boldface type indicates the best-performing result among the metrics.
Table 2. Our method compared with previous state-of-the-art methods in binary and multiclassification tasks on APTOS 2019. The data in the table are from the results in the original papers of the respective authors. Boldface type indicates the best-performing result among the metrics.
MethodAccuracyPrecisionRecallF1AUC   
Binary class
Bodapati et al. [48]97.3497.0097.0097.00/
Bodapati et al. [26]97.8298.0098.0098.0098.00
Islam et al. [11]98.3698.3798.3698.3798.50
This study98.9198.9398.9098.9199.80
Multiclass
Dondeti et al. [27]77.9076.0077.0075.00/
Bodapati et al. [48]79.7378.0078.0078.00/
Kassani et al. [49]79.59/82.35//
Bodapati et al. [26]82.5482.0083.0082.0079.00
Kobat et al. [37]84.9074.3268.5370.75/
Islam et al. [11]84.3670.5173.8470.4993.83
This study85.6178.4468.3271.6193.97
Table 3. Experimental results of SupCon and Angular Margin (A-M) on the APTOS 2019 dataset.
Table 3. Experimental results of SupCon and Angular Margin (A-M) on the APTOS 2019 dataset.
MetricsBinary ClassMulticlass
SupConwith A-MSupConwith A-M
Accuracy96.9098.9181.0985.61
Precision96.9498.9374.5878.44
Recall96.9298.9061.7268.32
F196.9098.9167.2671.61
AUC99.5899.8089.7593.97
Table 4. Experimental results of SupCon and Angular Margin (A-M) on the Messidor-2 dataset.
Table 4. Experimental results of SupCon and Angular Margin (A-M) on the Messidor-2 dataset.
MetricsBinary ClassMulticlass
SupConwith A-MSupConwith A-M
Accuracy75.9580.5367.9472.14
Precision75.8981.3551.8770.09
Recall76.6478.4550.3756.40
F175.7779.1750.4159.46
AUC87.0187.3187.2389.50
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, D.; Ge, A.; Chen, X.; Wang, Q.; Wu, J.; Liu, S. Supervised Contrastive Learning with Angular Margin for the Detection and Grading of Diabetic Retinopathy. Diagnostics 2023, 13, 2389. https://doi.org/10.3390/diagnostics13142389

AMA Style

Zhu D, Ge A, Chen X, Wang Q, Wu J, Liu S. Supervised Contrastive Learning with Angular Margin for the Detection and Grading of Diabetic Retinopathy. Diagnostics. 2023; 13(14):2389. https://doi.org/10.3390/diagnostics13142389

Chicago/Turabian Style

Zhu, Dongsheng, Aiming Ge, Xindi Chen, Qiuyang Wang, Jiangbo Wu, and Shuo Liu. 2023. "Supervised Contrastive Learning with Angular Margin for the Detection and Grading of Diabetic Retinopathy" Diagnostics 13, no. 14: 2389. https://doi.org/10.3390/diagnostics13142389

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop