Next Article in Journal
Development of 3D Cell-Based Fluorescent Reporter Assay for Screening of Drugs Downregulating Telomerase Reverse Transcriptase
Next Article in Special Issue
Multimodal Classification of Alzheimer’s Disease Using Longitudinal Data Analysis and Hypergraph Regularized Multi-Task Feature Selection
Previous Article in Journal
Relationship Between Implant Connection and Implant Fracture: Systematic Review
Previous Article in Special Issue
Detection of Disease Features on Retinal OCT Scans Using RETFound
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Foveal Avascular Zone Segmentation in Optical Coherence Tomography Angiography Across Multiple Eye Diseases Using Knowledge Distillation

1
Doheny Image Analysis Laboratory, Doheny Eye Institute, 150 North Orange Grove Blvd, Pasadena, CA 91103, USA
2
Doheny Image Reading and Research Laboratory, Doheny Eye Institute, 150 North Orange Grove Blvd, Pasadena, CA 91103, USA
3
Hedgefog Research Inc., 1891 N Gaffey St. Ste 224, San Pedro, CA 90731, USA
4
Wilmer Eye Institute, Department of Ophthalmology, Johns Hopkins University, 1800 Orleans St, Baltimore, MD 21287, USA
*
Author to whom correspondence should be addressed.
Bioengineering 2025, 12(4), 334; https://doi.org/10.3390/bioengineering12040334
Submission received: 14 February 2025 / Revised: 15 March 2025 / Accepted: 21 March 2025 / Published: 23 March 2025
(This article belongs to the Special Issue AI in OCT (Optical Coherence Tomography) Image Analysis)

Abstract

:
Optical coherence tomography angiography (OCTA) is a noninvasive imaging technique used to visualize retinal blood flow and identify changes in vascular density and enlargement or distortion of the foveal avascular zone (FAZ), which are indicators of various eye diseases. Although several automated FAZ detection and segmentation algorithms have been developed for use with OCTA, their performance can vary significantly due to differences in data accessibility of OCTA in different retinal pathologies, and differences in image quality in different subjects and/or different OCTA devices. For example, data from subjects with direct macular damage, such as in age-related macular degeneration (AMD), are more readily available in eye clinics, while data on macular damage due to systemic diseases like Alzheimer’s disease are often less accessible; data from healthy subjects may have better OCTA quality than subjects with ophthalmic pathologies. Typically, segmentation algorithms make use of convolutional neural networks and, more recently, vision transformers, which make use of both long-range context and fine-grained detail. However, transformers are known to be data-hungry, and may overfit small datasets, such as those common for FAZ segmentation in OCTA, to which there is limited access in clinical practice. To improve model generalization in low-data or imbalanced settings, we propose a multi-condition transformer-based architecture that uses four teacher encoders to distill knowledge into a shared base model, enabling the transfer of learned features across multiple datasets. These include intra-modality distillation using OCTA datasets from four ocular conditions: healthy aging eyes, Alzheimer’s disease, AMD, and diabetic retinopathy; and inter-modality distillation incorporating color fundus photographs of subjects undergoing laser photocoagulation therapy. Our multi-condition model achieved a mean Dice Index of 83.8% with pretraining, outperforming single-condition models (mean of 83.1%) across all conditions. Pretraining on color fundus photocoagulation images improved the average Dice Index by a small margin on all conditions except AMD (1.1% on single-condition models, and 0.1% on multi-condition models). Our architecture demonstrates potential for broader applications in detecting and analyzing ophthalmic and systemic diseases across diverse imaging datasets and settings.

1. Introduction

Optical coherence tomography (OCT) is a high-resolution, noninvasive imaging technique for visualizing tissue structures. Optical coherence tomography angiography (OCTA) adds the capability to visualize blood flow and retinal vasculature. OCTA has become a preferred tool for retinal imaging due to its ability to visualize the microvasculature of the retina and choroid without the need for exogenous dye injections. OCT captures detailed 3D images of the retina, while OCTA generates en face images—2D projections of specific vascular layers within 3D images [1,2].
The retinal layers contain the superficial vascular complex (SVC) in the nerve fiber and ganglion cell layers and the deep vascular complex (DVC) in the inner nuclear and outer plexiform layers. The foveal avascular zone (FAZ) is a small, capillary-free region located at the center of the fovea, a specialized area of the retina with densely packed cone photoreceptors that enable high-resolution vision. The FAZ forms a roughly circular, shallow depression, free of blood vessels, to maximize light reaching the foveola—the center of the fovea, where cone cells are most concentrated. The FAZ presents at multiple retinal layers/complexes. The area and shape of the FAZ and the density of retinal vasculature in the surrounding area are markers of visual acuity and can also indicate the presence and progression of disease processes.
FAZ is a sensitive indicator of retinal microcirculation and ischemic events. Its enlargement signals the loss of retinal vasculature, disrupting the supply of essential nutrients and oxygen to retinal tissues, potentially causing ischemia, inflammation, and ultimately leading to retinal neovascularization. Enlargement and irregular shape of the FAZ have been observed in various diseases affecting the retina, including Alzheimer’s disease (ALZ) [3,4], diabetic retinopathy (DR) [5], age-related macular degeneration (AMD) [6], macular branch retinal vein occlusion (BRVO) [7,8], and Stargardt disease [9]. Morphological changes in the FAZ generally reduce visual acuity by lowering the efficiency of light transmission to the foveola. A decline in visual acuity has been found to correlate with enlargement of the FAZ in multiple retinal diseases [10,11]. Decreased vascular perfusion near the FAZ has been observed in age-related macular degeneration [12]. The size of the FAZ is strongly positively correlated with capillary non-perfusion in patients with diabetic retinopathy [13]. The size of the FAZ correlates with disease severity in DR and can serve as a diagnostic marker. Diabetic patients without clinically detectable DR show microvascular changes, including an enlarged FAZ, which can serve as an early indicator of DR, even before the clinical signs of the disease appear [14]. OCTA FAZ area measurements reveal early retinal microvascular changes, including neovascularization and microaneurysms, in DR patients with hypertension, preceding detection by spectral domain OCT (SD-OCT) and conventional retinography [15]. FAZ enlargement has also been noted in post-COVID patients, where underlying mechanisms such as hypoxia, impaired diffusion, and undernutrition are implicated [16]. Other conditions in which changes in the FAZ area have been observed include diabetic macular edema [17], quiescent posterior uveitis and panuveitis [18], ischemic stroke, and Behҫet’s disease [19]. Studies have identified significant decreases in the FAZ area following panretinal photocoagulation (PRP) treatment in DR patients [20], after intravitreal anti-VEGF treatment in DR patients [17,21], and after glaucoma surgery [22].
The shape of the FAZ can also be an indicator of retinal disorders. For instance, diabetic eyes, even without the presence of clinically detectable diabetic retinopathy, have been found to exhibit irregular morphology and borders, including significant differences in circularity, roundness, and solidity, and an increased axial ratio [23,24,25]. Studies have reported an increased FAZ perimeter and a decreased circularity index in glaucoma, suggesting that changes in the FAZ circularity index may reflect disruption of the parafoveal capillary network [26]. The FAZ perimeter has demonstrated strong diagnostic ability in distinguishing eyes with central visual field defects from normal eyes and may serve as a biomarker for detecting individuals with central visual field defects in glaucoma [27]. Studies have found that the circularity of the FAZ is lower in subjects with high myopia [28,29], and that it is larger and more irregular in subjects with pathological myopia [30].
The automated segmentation and analysis of the FAZ in OCTA are garnering significant interest due to their potential to improve the diagnosis, monitoring, and management of various eye diseases. In this paper, we develop segmentation models for automated detection of the FAZ in retinal images, with a focus on enhancing performance across multiple conditions (Healthy, ALZ, AMD, DR). We use multi-condition models and pretraining to improve model accuracy and generalization, particularly for small datasets, to develop more reliable diagnostic tools for retinal and systemic diseases.

2. Methods

2.1. Background

Recent advances in OCTA imaging allow for the detailed visualization of the FAZ and retinal vasculature. Figure 1 shows examples of OCT and OCTA en face images obtained from the SVC, where FAZ presents in the foveal regions in OCTA (Spectralis OCT2, Heidelberg Eng.). Figure 2 shows an annotated SVC en face map of the FAZ. Vessel density (SVC VD) can be estimated from every OCTA image (for precise definitions, see: [31]). Machine learning models have been employed for OCT image processing to improve diagnostic speed and accuracy [32], to classify retinal diseases using OCT images [33], and to accurately segment retinal layers, which is useful in diagnosing various conditions [34]. Automated segmentation of the FAZ can assist in medical diagnosis and management of various eye diseases, as discussed above [35]. Several machine learning methods have been proposed for characterizing the FAZ and surrounding vasculature in the OCTA en face maps of the macular region. These include a method based on gradient boosting to diagnose ALZ using features extracted from FAZ segmentations and convolutional neural networks (CNNs) for automated segmentation [36,37]. U-Net, a CNN-based model commonly used for image segmentation, particularly in biomedical applications, has been employed for segmentation of the FAZ and surrounding vasculature [38,39]. A 2022 study proposed a modified U-Net for FAZ segmentation, which utilizes attention modules in the skip connections so the model can better integrate features at different scales [40]. A segmentation pipeline based on Mask R-CNN was also proposed [41,42].
More recently, vision transformers (ViTs) have achieved state of the art performance on various vision tasks through the use of self-attention, rather than convolution, which allows them to efficiently aggregate information globally, capture rich contextual relationships, and dynamically adjust the region of focus depending on the input [43,44]. ViTs have achieved competitive performance on semantic segmentation tasks including in automated segmentation of retinal lesions in OCT images [45,46]. ViTs’ greater generality enables them to often outperform CNNs on large datasets [47]. Conversely, the strong inductive bias conferred by the convolution operation in CNNs biases the latter toward learning local spatial relationships and hierarchical structure in images, which helps them learn relevant features with fewer parameters and generalize well from smaller datasets [47]. The Swin Transformer addresses some of the problems with the original ViT by incorporating a shifted window attention mechanism and a hierarchical design that progressively reduces spatial dimensions while increasing feature depth (much like in U-Net) [48].
Biomedical imaging data are often scarce due to the high costs of data acquisition and annotation, the need for specialized expertise, and patient privacy concerns [49]. Deep neural networks trained on small datasets are prone to overfitting, leading to poor generalization. In many medical imaging datasets, some data classes have significantly more instances. Data imbalance can pose significant challenges, including bias towards the majority classes, negative knowledge transfer, and poor generalization [50,51,52,53]. Disparities in data accessibility hence affect both the training of deep learning models and their clinical applicability. This is compounded by challenges such as visual artifacts and variability inherent in retinal imaging, which often complicates accurate FAZ segmentation [54].
Many techniques have been proposed to address data scarcity and imbalance [55,56]: Data augmentation improves generalization by applying transformations like rotation, flipping, or noise injection to existing samples, increasing diversity [57]. Synthetic data generation expands datasets by creating artificial samples, while semi-supervised learning leverages both labeled and unlabeled data, often through pseudo-labeling or consistency regularization [58]. Self-supervised learning generates supervisory signals from the structure or properties of unlabeled data itself, allowing models to learn meaningful representations without requiring explicit labels [59]. A common approach is to do self-supervised pretraining of large foundation models on natural images, and then fine-tune them on medical data [60]. Large datasets of radiological images, such as RadImageNet, have been created for pretraining foundation models [61]. A 2024 study found that a foundation model trained using a database of medical images from different imaging types, including tomography, microscopy, and X-ray images, improved performance over pretraining on non-medical images [62]. Balancing techniques, such as over-sampling and under-sampling, address class imbalances by adjusting the data distribution, though they risk noise or information loss [63]. Approaches such as class weighting prioritize minority-class performance but require careful tuning to prevent overcompensation [64].
Ensemble methods improve robustness by combining predictions from multiple models [65]. Transfer learning enables effective training on limited data by fine-tuning pre-trained models from large, related datasets [60]. Transfer learning techniques allow us to integrate shared information across different diseases and can improve performance on under-represented data. While transfer learning adapts knowledge across tasks, model distillation or knowledge distillation addresses challenges posed by scarce and unbalanced data by allowing a larger, well-trained teacher model to impart knowledge to a smaller student model, enabling it to generalize better and make more accurate predictions despite limited or imbalanced training samples [52,53]. While transfer learning retains the original model’s architecture, leveraging its learned features, knowledge distillation results in a compressed model that retains much of the performance of the larger model while being faster and more efficient. Studies have shown that model distillation can also help mitigate overfitting [53]. Model distillation has been shown to improve model efficiency on small medical image datasets [66,67]. Qin et al. introduce one such architecture, which enhances a medical image segmentation model via knowledge distillation, enabling the smaller network to better represent the differences between tissue regions [68]. Xing et al. apply knowledge distillation to improve medical image classification, and use a mean-teacher model and contrastive loss to address intra-class variance and class imbalance [69]. Du et al. propose MDViT, a multi-domain vision transformer with a universal network for shared knowledge and “peer” branches for domain-specific learning. Each self-attention block includes a “domain adapter” that adjusts attention heads based on the task, promoting head specialization [70]. Benech et al. used a deep learning model based on Mask R-CNN to show that automated segmentation of the FAZ can be improved by incorporating auxiliary tasks for superficial vascular complex FAZ (SVC FAZ) boundary and vessel segmentation [41,71].

2.2. Overview

We propose a neural network architecture for FAZ segmentation that addresses both single-condition and multi-condition scenarios (Figure 3). For single-condition segmentation, the encoder-decoder backbone combines CNN and Transformer blocks to balance local feature extraction with global context modeling. Residual connections and hierarchical attention mechanisms ensure efficient feature representation and accurate reconstruction. In the multi-condition setting, condition-specific teacher decoders guide the shared backbone using knowledge distillation, enabling the base model to generalize across multiple datasets while maintaining robust performance on individual conditions. Sparse attention mechanisms are applied to improve generalization, and a hybrid loss function is used to regularize the model when working with limited data.

2.3. Materials

The study cohort includes 102 eyes in total from 67 patients: among them, 17 eyes (9 OS-left eye, 8 OD-right eye) from 9 elderly adults; 33 eyes (16 OS, 17 OD) from 18 patients with ALZ; 31 eyes (12 OS, 19 OD) from 28 patients with AMD; and 21 eyes (11 OS, 10 OD) from 12 patients with DR, respectively. OCTA macular images were captured with Spectralis OCT2 (Heidelberg Eng.) using scan pattern 3 mm × 3 mm (10° × 10°; 256 × 256 pixels). Both the superficial vascular complex (SVC) and deep vascular complex (DVC) may show FAZ abnormalities in the retinal diseases mentioned above; however, in this pilot study focusing on multi-domain segmentation, we concentrate on the automated segmentation of the FAZ in the SVC. The ground truth binary masks of the FAZ boundary were produced by two trained human annotators with expertise in manual grading of retinal images for clinical trial and research studies.

2.4. Single-Condition FAZ Segmentation

Our base model consists of alternating CNN and Transformer blocks in an encoder-decoder structure, with residual connections between the encoder and decoder. The encoder progressively reduces spatial dimensions with convolutions while increasing feature depth, and the decoder reverses this process with deconvolutions. Convolution and deconvolution layers integrate spatially local information, with lower layers capturing long-range dependencies and higher layers capturing finer details. Residual connections across layers help retain information lost in max pooling. Transformer encoder layers placed between each step of convolution and deconvolution allow image patch feature embeddings to exchange information globally. The result is a hierarchical arrangement of multi-head attention layers, with transformers in the first layers of the encoder and last layers of the decoder attending to many smaller image patches with low-dimensional embeddings, and intermediate layers attending to larger image patches with higher-dimensional embeddings. This allows attention to be applied to both more abstract representations (based on larger image patches) and more fine-grained spatial information (based on smaller image patches). Residual connections are integrated with upstream information using cross attention. A fully connected network connects the encoder and decoder.
To address the challenge of training our model on a very small dataset, we incorporate the following inductive biases into our transformer encoder layers:
(1.) We replace the fully connected layers in the transformer encoders with sparse autoencoders, an architecture commonly used in neural network interpretability research, because they tend to produce “disentangled” interpretable representations [72]. Intermediate layers in a sparse autoencoder are of much higher dimension than the input layer. An L1 weight penalty, a standard loss penalizing the absolute value of the weights, which encourages sparsity and is also robust to outliers, is applied to ensure that only a small number of neurons remain active. This is thought to create an information bottleneck that promotes efficient encoding and discourages memorization. For simplicity, we apply an L1 loss to all attention weights.
(2.) We add to each attention matrix (a) a positional attention matrix, which does not depend on the patch embeddings but only on their absolute locations in the image to encode the spatial relationships between different patches, regardless of their content, and (b) a localized attention matrix which computes each attention score using an elliptical Gaussian kernel in the two-dimensional space of image patches, with learnable major and minor axes. The positional attention matrix captures global spatial relationships based on the absolute positions of patches, while the localized attention matrix gives more weight to nearby patches with a learnable Gaussian function, with an attention score Lij between patches i and j given by:
L i j   e x p ( x i x j ) 2 σ x i ( y i y j ) 2 σ y i
(where σx and σy are learnable parameters). Unlike CNNs, which rely on fixed convolutional filters to process local patch interactions, the Gaussian kernel-based attention adapts based on relative positions, allowing for more flexible spatial dependencies. Both positional and localized attention mechanisms require far fewer weights than the full dot-product attention mechanism, at the expense of losing the flexibility and contextual understanding provided by input-dependent attention weights in the full attention mechanism. We apply an L1 penalty on the attention weights to encourage sparse representations, guiding the model to prioritize simple positional attention mechanisms while reducing reliance on more complex global or context-dependent relationships.
The input to the network is a single-channel FAZ OCTA scan of a retinal region that includes the FAZ and the output is a ground truth binary mask of the FAZ boundary, produced by a human annotator. Let Enc and Dec denote the encoder and decoder models, and let X ϵ R n × n be an input image with n × n pixels, Y ϵ { 0,1 } n × n the corresponding ground truth binary mask, of the same dimension, and Y ^ ϵ [ 0 , 1 ] n × n the output of our model, that is, Y ^ = D e c ( E n c ( X ) ) . Each loss term is a weighted sum of binary cross entropy (BCE) loss (a standard loss function for binary classification) and a soft dice loss (another standard loss for segmentation tasks): L = c 1   L B C E + c 2   L D i c e , where
L D i c e = 1 2   i ( Y Y ^ ) i + δ 1 i ( Y + Y ^ ) i + δ 2
Here, denotes the Hadamard product, and c1, c2, δ1, δ2 are positive constants (we set c1 = 1, c2 = ½, δ1 = δ2 = 1). The presence of the dice loss improves model accuracy while the BCE loss improves training stability.

2.5. Multi-Condition FAZ Segmentation via Knowledge Distillation for Intra-Modality Data

To enable training on four separate datasets, we employ model distillation. In this technique, separate networks are trained on each dataset, and each specialized “Teacher” network helps train a larger “Base” model. The model distillation literature shows that larger teacher networks can improve the performance of smaller distilled networks and reduce overfitting, especially on small datasets [53]. In our case, the teacher networks are the same size as the distilled network, allowing training over many separate modalities, while each teacher network dedicates all of its parameters to a single condition. Well-trained teacher networks provide richer information to the base model than the target images alone because the former are binary masks while the latter are real-valued probabilities for each pixel.
Our architecture includes the following components: (a) a multi-condition “Base” model, and (b) a series of “Teacher” decoders, each specializing in a single condition (Healthy, AMD, ALZ, or DR). The base model is used across all conditions, while separate teacher decoders are used for each of the four conditions. The encoder is shared across base and teacher models, in order to learn feature embeddings that generalize across conditions. The base and teacher models are trained simultaneously and end-to-end, with gradient information flowing from both base and teacher decodes back through the base encoder. At test time, the teacher decoders are discarded and only the base model is used for inference.
The base decoder receives a one-hot encoding denoting a particular condition, which is projected into the dimension of the image patch feature embeddings using a fully connected network. To account for the possibility that condition information is not available, one can simply add an additional slot to the one-hot encoding and train the base model to make condition-independent predictions as an additional task. We explored several methods for integrating image and condition information, including using attention, but found the most effective approach was to add the condition embedding to each patch embedding and let the attention mechanism learn how best to integrate them. Our architecture is shown in Figure 3.
The loss function for condition j includes three terms:
L j = α L ( Y ^ b a s e   , Y ) + β L ( Y ^ t e a c h e r   j   , Y ) + γ ( Y ^ b a s e   , Y t e a c h e r   j )
where the first term compares the output of the base model to the ground truth image, the second term compares the output of the jth teacher model to ground truth, and the third term compares the outputs of the base model and jth teacher model. Here, α, β, γ are positive constant hyperparameters, with α > β > γ (we set α = 1, β = ½, γ = ¼), so that the model prioritizes the base model over the teacher model, and the fidelity to the ground truth over the comparison between models.
Information flow between the teacher networks and the base network is limited by the fact that gradients must flow through both networks from end to end, which means there is no means of comparing internal representations directly. This limitation can be addressed by comparing the outputs of the networks layer by layer. We add terms to the loss function that mask out residual connections upstream of a given layer, forcing the model to represent as much of the output as possible in the lower layers. Since the size of the layers decreases with depth, the inclusion of these loss terms should force the network to produce compressed representations at each layer, which we expect to generalize better. Hence, our total loss for condition j becomes:
L j ,   t o t = i a i L ( D e c < i ( E n c ( Y ) ) , Y )
where D e c < i denotes that every residual layer above layer i in our decoder model is masked out. Here, each ai is a positive constant, with ai < aj for i > j so that the loss prioritizes more complete models. We set a i = 2 i so that each term is greater than the sum of all subsequent terms. The single-condition model has 8 attention heads, while the multi-condition model is somewhat larger (16 attention heads) to capture relationships across the four datasets.

2.6. Knowledge Distillation for Inter-Modality Data

Due to the small size of our OCTA FAZ dataset, we explore pretraining our models on a medical segmentation dataset of fundus images from patients undergoing laser photocoagulation (PC) therapy, a specialized ophthalmic procedure involving laser treatment for conditions like diabetic retinopathy. The dataset consists of 298 images, with masks of annotated regions where laser treatment has been applied. An example image from this dataset is pictured in Figure 4.
Pretraining provides the model with a foundation for recognizing structural and tissue distinctions within retinal images, enabling it to handle diverse medical imaging tasks. Color fundus images, despite being from a different domain, capture fundamental retinal features—such as tissue structures, vessels, and abnormalities—shared across ophthalmic conditions. These features help the model learn generalizable representations for segmenting the FAZ in OCTA images, and valuable context for differentiating between normal and pathological structures [73,74,75]. Many large datasets of fundus images are publicly available, making them a good choice for pretraining foundation models [76,77].
Our multi-condition model also supports training using multiple ground truth annotations per image. By interpolating between different annotations, the dataset can be effectively expanded, mitigating systematic biases by some annotators. For instance, when masks by two annotators are available for each image, we can use an expanded one-hot encoding to represent each condition-annotator pair and provide the model with linear combinations of the images and one-hot encodings.

3. Results

The multi-condition model has 8 attention heads and embedding dimensions of 128, 256, and 512 in subsequent layers. We use the Adam optimizer with a learning rate of 1E-4. The input and output data are each tensors of dimension NB × NC × NH × NW, together with a one-hot-encoding vector of size NB × ND, where NB is the batch size, NC is the channel dimension, NH and NW are the image height and width, and ND is the number of conditions. (We set: NB = 10, NC = 1, NH = NW = 256, ND = 4.) We cycle through conditions during training, training each for one epoch at a time. Data augmentation included random shifts, rotations, flips, color jitter, and additive zero-mean Gaussian white noise. The model was trained for 500 epochs, approximately 24 hours on one NVIDIA GeForce RTX 3090 GPU. The model is evaluated using two-fold cross-validation (training on one set and testing on the other, then swapping the sets), with the same random partition applied to each condition. This process is repeated for each condition to ensure a robust evaluation across all datasets. Example model inputs/outputs for each condition are shown in Figure 5.
The following tables show two-fold cross validation dice indices on the test set for single-condition and multi-condition models across different retinal conditions. Table 1 shows the effect of PC pretraining on single-condition and multi-condition model performance for the Healthy, ALZ, AMD, and DR datasets. We pretrain both our base and teacher models on PC color fundus images and then finetune on our FAZ dataset. Table 2 compares the effect of using two annotations during training under the same pretraining conditions. As the small size of the test set results in somewhat variable performance, we compute an average of the test accuracy over the final 100 training epochs, providing a more stable measure of model performance. The average test dice score on the single-condition PC pretraining data is 52.5%.
The results in Table 1 show that multi-condition models typically outperform single-condition models. Notably, multi-condition models benefit the most from PC pretraining, achieving the highest overall mean dice index, with consistent improvements across all conditions but AMD. Although single-condition models also show slight improvements with pretraining, the gains are smaller, suggesting that multi-condition models more effectively leverage pretraining to enhance generalization.
While PC pretraining enhances performance for healthy and DR cases in both single-condition and multi-condition models, it appears to have no significant effect on ALZ segmentation and a negative effect on AMD segmentation, particularly in the multi-condition model. This decline suggests that the pretraining may introduce domain-specific biases that are less effective for certain pathological features, which suggests the need for targeted adjustments to pretraining or changes to model architecture to mitigate negative knowledge transfer. Additionally, the smaller gains from pretraining in single-condition models imply that multi-condition configurations are better suited for transferring learned features across diverse conditions. The largest improvement in PC pretraining is observed in healthy eyes (the smallest dataset).
To further analyze model performance, a paired t-test was conducted to assess the significance of the observed improvements on the test data over the final 100 epochs of training (Table 3). A paired t-test is appropriate in this case because it compares the mean differences in performance between two conditions (e.g. pretraining vs. no pretraining) for the same set of models. This test controls for individual model variability by focusing on the differences within paired observations, allowing us to assess whether the observed improvements are statistically significant. We also calculated Cohen’s d-values, which measure the effect size (Table 4).
PC pretraining and the inclusion of two annotations showed large effect sizes for Healthy eyes in both single-condition and multi-condition models but exhibited variable effects across conditions, with notable improvements in DR and moderate gains in ALZ. In contrast, the AMD dataset experienced mixed results, including, in one case, a small decline in performance, likely due to negative knowledge transfer or the presence of domain-specific biases that the model failed to adequately address. This indicates that AMD’s unique pathological features may conflict with the shared features learned across conditions in the multi-condition approach, which suggests the need for further refinement in the way the model represents shared features, a greater ability to specialize to conditions, and targeted pretraining strategies.
The relatively small improvement from multi-condition training suggests that the differences in the FAZ between these conditions is relatively small. The small improvement from pretraining on PC fundus images, and slightly worse performance for AMD, suggests that the general features of retinal structure learned from this data are of limited value, and more specialized FAZ data is necessary to improve performance. Inconsistency or ambiguity in FAZ annotations also limits performance.

4. Discussion and Conclusions

In this study, we presented a multi-condition transformer-based architecture designed to improve automated segmentation of the FAZ in a variety of eye diseases, such as AMD, ALZ, and DR. Our approach addresses the challenges posed by limited and imbalanced datasets by leveraging knowledge distillation from multiple retinal diseases. The results demonstrate substantial improvements in segmentation performance in scenarios with very small training sets, and the model’s ability to generalize across different diseases. The multi-condition model generally outperformed single-condition models, demonstrating its ability to leverage shared features across conditions, though certain pathological features in AMD require further attention.
Adaptive pretraining strategies, additional methods of selective feature sharing, or condition-specific regularization could mitigate domain-specific biases seen in AMD, while tailored data augmentation strategies and synthetic data generation techniques that simulate condition-relevant pathological variations could provide more representative training data.
The proposed FAZ segmentation framework may be of use in the study of systemic diseases that manifest through retinal vascular abnormalities, such as diabetes, hypertension, and neurodegenerative disorders like ALZ, including for identifying biomarkers for early detection, disease progression monitoring, and treatment efficacy assessment. Large-scale implementation of such models could facilitate population-level screening and early intervention for high-risk groups, supporting public health initiatives and epidemiological research. Integrating this framework with other diagnostic tools, such as AI-based fundus analysis or blood biomarker studies, could enhance clinical workflows.
Extensions of this work could focus on cross-modal integration with other imaging techniques, combining OCTA with techniques such as fundus photography or fluorescein angiography. Expanding the applications of the model to include a wider range of systemic conditions would increase its utility in medical practice and research. Multi-condition segmentation methods could also be combined with next-frame prediction techniques for longitudinal data to track disease progression over time.

Author Contributions

Conceptualization, Z.H.; Methodology, P.R.; Investigation, P.R.; Evaluation, P.R.; Visualization, P.R.; Data, G.C., S.R.S., G.M. and Y.M.P.; Other resources, Z.H.; Project administration, Z.H.; Data curation and ground truth, A.A., N.V.P. and Z.W.; Writing—original draft preparation, P.R.; Writing—review and editing, P.R., Z.H., S.R.S., Y.M.P. and G.M.; Supervision, Z.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Eye Institute of the National Institutes of Health under Award Number R21EY030619.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki. The OCTA datasets were approved by the Institutional Review Board of the University of California, Los Angeles. The PC dataset was from Kaggle competition which was provided by EyePACS. There is no restriction for research usage for the PC data per EyePACS (https://www.kaggle.com/c/diabetic-retinopathy-detection/data). This dataset was accessed 1 May 2020. Please note, for this study, all data are priorly de-identified.

Informed Consent Statement

Not applicable because de-identified data (non-human subjects’ data) were used in this study.

Data Availability Statement

The datasets are available from the corresponding author on reasonable request. The code generated during the study is accessible from the corresponding author based on reasonable request and subject to the regulations of the institute.

Acknowledgments

We thank Jianhua (Jay) Wang in Bascom Palmer Eye Institute at the University of Miami Miller School of Medicine for the discussion and input of the OCTA vascular complexes.

Conflicts of Interest

Gary Mikaelian is an employee of Hedgefog Research Inc. The other authors declare no conflicts of interest.

References

  1. Spaide, R.F.; Fujimoto, J.G.; Waheed, N.K.; Sadda, S.R.; Staurenghi, G. Optical Coherence Tomography Angiography. Prog. Retin. Eye Res. 2018, 64, 1–55. [Google Scholar] [CrossRef] [PubMed]
  2. Javed, A.; Khanna, A.; Palmer, E.; Wilde, C.; Zaman, A.; Orr, G.; Kumudhan, D.; Lakshmanan, A.; Panos, G.D. Optical Coherence Tomography Angiography: A Review of the Current Literature. J. Int. Med. Res. 2023, 51, 03000605231187933. [Google Scholar] [CrossRef]
  3. Rifai, O.M.; McGrory, S.; Robbins, C.B.; Grewal, D.S.; Liu, A.; Fekrat, S.; MacGillivray, T.J. The Application of Optical Coherence Tomography Angiography in Alzheimer’s Disease: A Systematic Review. Alzheimers Dement. 2021, 13, e12149. [Google Scholar] [CrossRef]
  4. Yap, T.E.; Balendra, S.I.; Almonte, M.T.; Cordeiro, M.F. Retinal Correlates of Neurological Disorders. Ther. Adv. Chronic Dis. 2019, 10, 2040622319882205. [Google Scholar] [CrossRef] [PubMed]
  5. Takase, N.; Nozaki, M.; Kato, A.; Ozeki, H.; Yoshida, M.; Ogura, Y. Enlargement of Foveal Avascular Zone in Diabetic Eyes Evaluated by En Face Optical Coherence Tomography Angiography. Retina 2015, 35, 2377–2383. [Google Scholar] [CrossRef]
  6. Lee, S.C.; Rusakevich, A.M.; Amin, A.; Tran, S.; Emami-Naeini, P.; Moshiri, A.; Park, S.S.; Yiu, G. Long-Term Retinal Vascular Changes in Age-Related Macular Degeneration Measured Using Optical Coherence Tomography Angiography. Ophthalmic Surg. Lasers Imaging Retin. 2022, 53, 529–536. [Google Scholar] [CrossRef]
  7. Parodi, M.; Visintin, F.; Della Rupe, P.; Ravalico, G. Foveal Avascular Zone in Macular Branch Retinal Vein Occlusion. Int. Ophthalmol. 1995, 19, 25–28. [Google Scholar] [CrossRef] [PubMed]
  8. Koulisis, N.; Kim, A.Y.; Chu, Z.; Shahidzadeh, A.; Burkemper, B.; de Koo, L.C.O.; Moshfeghi, A.A.; Ameri, H.; Puliafito, C.A.; Isozaki, V.L.; et al. Quantitative Microvascular Analysis of Retinal Venous Occlusions by Spectral Domain Optical Coherence Tomography Angiography. PLoS ONE 2017, 12, e0176404. [Google Scholar] [CrossRef]
  9. Reich, M.; Glatz, A.; Cakir, B.; Böhringer, D.; Lang, S.; Küchlin, S.; Joachimsen, L.; Lagreze, W.; Agostini, H.T.; Lange, C. Characterisation of Vascular Changes in Different Stages of Stargardt Disease Using Double Swept-Source Optical Coherence Tomography Angiography. BMJ Open Ophthalmol. 2019, 4, e000318. [Google Scholar] [CrossRef]
  10. Duffy, B.; Castellanos Canales, D.; Decker, N.; Yamaguchi, T.; Pearce, L.; Fawzi, A.A. Foveal Avascular Zone Enlargement Correlates with Visual Acuity Decline in Patients with Diabetic Retinopathy. Investig. Ophthalmol. Vis. Sci. 2024, 65, 955. [Google Scholar] [CrossRef]
  11. Balaratnasingam, C.; Inoue, M.; Ahn, S.; McCann, J.; Dhrami-Gavazi, E.; Yannuzzi, L.A.; Freund, K.B. Visual Acuity Is Correlated with the Area of the Foveal Avascular Zone in Diabetic Retinopathy and Retinal Vein Occlusion. Ophthalmology 2016, 123, 2352–2367. [Google Scholar] [CrossRef] [PubMed]
  12. Trinh, M.; Kalloniatis, M.; Nivison-Smith, L. Radial Peripapillary Capillary Plexus Sparing and Underlying Retinal Vascular Impairment in Intermediate Age-Related Macular Degeneration. Investig. Ophthalmol. Vis. Sci. 2021, 62, 2. [Google Scholar] [CrossRef] [PubMed]
  13. Bresnick, G.H.; Condit, R.; Syrjala, S.; Palta, M.; Groo, A.; Korth, K. Abnormalities of the Foveal Avascular Zone in Diabetic Retinopathy. Arch. Ophthalmol. Chic. 1984, 102, 1286–1293. [Google Scholar] [CrossRef] [PubMed]
  14. Park, Y.G.; Kim, M.; Roh, Y.J. Evaluation of Foveal and Parafoveal Microvascular Changes Using Optical Coherence Tomography Angiography in Type 2 Diabetes Patients without Clinical Diabetic Retinopathy in South Korea. J. Diabetes Res. 2020, 2020, 6210865. [Google Scholar] [CrossRef]
  15. Stana, D.; Potop, V.; Istrate, S.L.; Eniceicu, C.; Mihalcea, A.R. Foveal Avascular Zone Area Measurements Using OCT Angiography in Patients with Type 2 Diabetes Mellitus Associated with Essential Hypertension. Rom. J. Ophthalmol. 2019, 63, 354–359. [Google Scholar] [CrossRef]
  16. Koutsiaris, A.G. A Blood Supply Pathophysiological Microcirculatory Mechanism for Long COVID. Life 2024, 14, 1076. [Google Scholar] [CrossRef]
  17. Gill, A.; Cole, E.D.; Novais, E.A.; Louzada, R.N.; de Carlo, T.; Duker, J.S.; Waheed, N.K.; Baumal, C.R.; Witkin, A.J. Visualization of Changes in the Foveal Avascular Zone in Both Observed and Treated Diabetic Macular Edema Using Optical Coherence Tomography Angiography. Int. J. Retin. Vitr. 2017, 3, 19. [Google Scholar] [CrossRef]
  18. Agarwal, A.; Bhatt, S.; Keshari, S.; Erckens, R.J.; Berendschot, T.T.J.M.; Webers, C.a.B.; Agrawal, R.; Bansal, R.; Gupta, V. Retinal Microvascular Alterations in Patients with Quiescent Posterior and Panuveitis Using Optical Coherence Tomography Angiography. Ocul. Immunol. Inflamm. 2022, 30, 1781–1787. [Google Scholar] [CrossRef]
  19. Smid, L.M.; Vermeer, K.A.; Missotten, T.O.A.R.; van Laar, J.A.M.; van Velthoven, M.E.J. Parafoveal Microvascular Alterations in Ocular and Non-Ocular Behҫet’s Disease Evaluated With Optical Coherence Tomography Angiography. Investig. Ophthalmol. Vis. Sci. 2021, 62, 8. [Google Scholar] [CrossRef]
  20. Hoshiyama, K.; Hirano, T.; Hirabayashi, K.; Wakabayashi, M.; Tokimitsu, M.; Murata, T. Morphological Changes in the Foveal Avascular Zone after Panretinal Photocoagulation for Diabetic Retinopathy Using OCTA: A Study Focusing on Macular Ischemia. Medicina 2022, 58, 1797. [Google Scholar] [CrossRef]
  21. Liao, D.; Zhou, Z.; Wang, F.; Zhang, B.; Wang, Y.; Zheng, Y.; Li, J. Changes in Foveal Avascular Zone Area and Retinal Vein Diameter in Patients with Retinal Vein Occlusion Detected by Fundus Fluorescein Angiography. Front. Med. 2023, 10. [Google Scholar] [CrossRef]
  22. Shoji, T.; Kanno, J.; Weinreb, R.N.; Yoshikawa, Y.; Mine, I.; Ishii, H.; Ibuki, H.; Shinoda, K. OCT Angiography Measured Changes in the Foveal Avascular Zone Area after Glaucoma Surgery. Br. J. Ophthalmol. 2022, 106, 80–86. [Google Scholar] [CrossRef] [PubMed]
  23. Sijilmassi, O. Quantitative Analysis of Different Foveal Avascular Zone Metrics in Healthy and Diabetic Subjects. Diabetology 2024, 5, 246–254. [Google Scholar] [CrossRef]
  24. Tang, F.Y.; Ng, D.S.; Lam, A.; Luk, F.; Wong, R.; Chan, C.; Mohamed, S.; Fong, A.; Lok, J.; Tso, T.; et al. Determinants of Quantitative Optical Coherence Tomography Angiography Metrics in Patients with Diabetes. Sci. Rep. 2017, 7, 2575. [Google Scholar] [CrossRef]
  25. Krawitz, B.D.; Mo, S.; Geyman, L.S.; Agemy, S.A.; Scripsema, N.K.; Garcia, P.M.; Chui, T.Y.P.; Rosen, R.B. Acircularity Index and Axis Ratio of the Foveal Avascular Zone in Diabetic Eyes and Healthy Controls Measured by Optical Coherence Tomography Angiography. Vis. Res. 2017, 139, 177–186. [Google Scholar] [CrossRef]
  26. Choi, J.; Kwon, J.; Shin, J.W.; Lee, J.; Lee, S.; Kook, M.S. Quantitative Optical Coherence Tomography Angiography of Macular Vascular Structure and Foveal Avascular Zone in Glaucoma. PLoS ONE 2017, 12, e0184948. [Google Scholar] [CrossRef]
  27. Kwon, J.; Choi, J.; Shin, J.W.; Lee, J.; Kook, M.S. Glaucoma Diagnostic Capabilities of Foveal Avascular Zone Parameters Using Optical Coherence Tomography Angiography According to Visual Field Defect Location. J. Glaucoma 2017, 26, 1120. [Google Scholar] [CrossRef]
  28. Živković, M.L.J.; Lazić, L.; Zlatanovic, M.; Zlatanović, N.; Brzaković, M.; Jovanović, M.; Barišić, S.; Darabus, D.-M. The Influence of Myopia on the Foveal Avascular Zone and Density of Blood Vessels of the Macula—An OCTA Study. Medicina 2023, 59, 452. [Google Scholar] [CrossRef]
  29. Piao, H.; Guo, Y.; Zhang, H.; Sung, M.S.; Park, S.W. Acircularity and Circularity Indexes of the Foveal Avascular Zone in High Myopia. Sci. Rep. 2021, 11, 16808. [Google Scholar] [CrossRef]
  30. Sui, J.; Li, H.; Bai, Y.; He, Q.; Sun, Z.; Wei, R. Morphological Characteristics of the Foveal Avascular Zone in Pathological Myopia and Its Relationship with Macular Structure and Microcirculation. Graefes Arch. Clin. Exp. Ophthalmol. 2024, 262, 2121–2133. [Google Scholar] [CrossRef]
  31. Koutsiaris, A.G.; Batis, V.; Liakopoulou, G.; Tachmitzi, S.V.; Detorakis, E.T.; Tsironi, E.E. Optical Coherence Tomography Angiography (OCTA) of the Eye: A Review on Basic Principles, Advantages, Disadvantages and Device Specifications. Clin. Hemorheol. Microcirc. 2023, 83, 247–271. [Google Scholar] [CrossRef] [PubMed]
  32. Karn, P.K.; Abdulla, W.H. On Machine Learning in Clinical Interpretation of Retinal Diseases Using OCT Images. Bioengineering 2023, 10, 407. [Google Scholar] [CrossRef]
  33. Akinniyi, O.; Rahman, M.M.; Sandhu, H.S.; El-Baz, A.; Khalifa, F. Multi-Stage Classification of Retinal OCT Using Multi-Scale Ensemble Deep Architecture. Bioengineering 2023, 10, 823. [Google Scholar] [CrossRef]
  34. Xue, S.; Wang, H.; Guo, X.; Sun, M.; Song, K.; Shao, Y.; Zhang, H.; Zhang, T. CTS-Net: A Segmentation Network for Glaucoma Optical Coherence Tomography Retinal Layer Images. Bioengineering 2023, 10, 230. [Google Scholar] [CrossRef] [PubMed]
  35. Zheng, Y.; Gandhi, J.S.; Stangos, A.N.; Campa, C.; Broadbent, D.M.; Harding, S.P. Automated Segmentation of Foveal Avascular Zone in Fundus Fluorescein Angiography. Investig. Ophthalmol. Vis. Sci. 2010, 51, 3653–3659. [Google Scholar] [CrossRef] [PubMed]
  36. Yoon, J.M.; Lim, C.Y.; Noh, H.; Nam, S.W.; Jun, S.Y.; Kim, M.J.; Song, M.Y.; Jang, H.; Kim, H.J.; Seo, S.W.; et al. Enhancing Foveal Avascular Zone Analysis for Alzheimer’s Diagnosis with AI Segmentation and Machine Learning Using Multiple Radiomic Features. Sci. Rep. 2024, 14, 1841. [Google Scholar] [CrossRef]
  37. Guo, M.; Zhao, M.; Cheong, A.M.Y.; Dai, H.; Lam, A.K.C.; Zhou, Y. Automatic Quantification of Superficial Foveal Avascular Zone in Optical Coherence Tomography Angiography Implemented with Deep Learning. Vis. Comput. Ind. Biomed. Art 2019, 2, 21. [Google Scholar] [CrossRef]
  38. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. ISBN 978-3-319-24573-7. [Google Scholar]
  39. Liang, Z.; Zhang, J.; An, C. Foveal Avascular Zone Segmentation of Octa Images Using Deep Learning Approach with Unsupervised Vessel Segmentation. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 1200–1204. [Google Scholar]
  40. Meng, Y.; Lan, H.; Hu, Y.; Chen, Z.; Ouyang, P.; Luo, J. Application of Improved U-Net Convolutional Neural Network for Automatic Quantification of the Foveal Avascular Zone in Diabetic Macular Ischemia. J. Diabetes Res. 2022, 2022, 4612554. [Google Scholar] [CrossRef]
  41. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  42. Mirshahi, R.; Anvari, P.; Riazi-Esfahani, H.; Sardarinia, M.; Naseripour, M.; Falavarjani, K.G. Foveal Avascular Zone Segmentation in Optical Coherence Tomography Angiography Images Using a Deep Learning Approach. Sci. Rep. 2021, 11, 1031. [Google Scholar] [CrossRef]
  43. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Nice, France, 2017; Volume 30. [Google Scholar]
  44. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021. [Google Scholar]
  45. Thisanke, H.; Deshan, C.; Chamith, K.; Seneviratne, S.; Vidanaarachchi, R.; Herath, D. Semantic Segmentation Using Vision Transformers: A Survey. Eng. Appl. Artif. Intell. 2023, 126, 106669. [Google Scholar] [CrossRef]
  46. Philippi, D.; Rothaus, K.; Castelli, M. A Vision Transformer Architecture for the Automated Segmentation of Retinal Lesions in Spectral Domain Optical Coherence Tomography Images. Sci. Rep. 2023, 13, 517. [Google Scholar] [CrossRef]
  47. Raghu, M.; Unterthiner, T.; Kornblith, S.; Zhang, C.; Dosovitskiy, A. Do Vision Transformers See Like Convolutional Neural Networks? In Proceedings of the 35th Conference on Neural Information Processing Systems (NeurIPS 2021), Online, 6–14 December 2021. [Google Scholar]
  48. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  49. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A Survey on Deep Learning in Medical Image Analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef]
  50. He, H.; Garcia, E.A. Learning from Imbalanced Data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar] [CrossRef]
  51. Liu, Y.; Yang, G.; Qiao, S.; Liu, M.; Qu, L.; Han, N.; Wu, T.; Yuan, G.; Wu, T.; Peng, Y. Imbalanced Data Classification: Using Transfer Learning and Active Sampling. Eng. Appl. Artif. Intell. 2023, 117, 105621. [Google Scholar] [CrossRef]
  52. Hinton, G.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531v1. [Google Scholar]
  53. Habib, G.; Saleem, T.J.; Kaleem, S.M.; Rouf, T.; Lall, B. A Comprehensive Review of Knowledge Distillation in Computer Vision. Int. J. Innov. Res. Comput. Sci. Technol. 2024, 12, 106–112. [Google Scholar]
  54. Linderman, R.; Salmon, A.E.; Strampe, M.; Russillo, M.; Khan, J.; Carroll, J. Assessing the Accuracy of Foveal Avascular Zone Measurements Using Optical Coherence Tomography Angiography: Segmentation and Scaling. Transl. Vis. Sci. Technol. 2017, 6, 16. [Google Scholar] [CrossRef]
  55. Alzubaidi, L.; Bai, J.; Al-Sabaawi, A.; Santamaría, J.; Albahri, A.S.; Al-dabbagh, B.S.N.; Fadhel, M.A.; Manoufali, M.; Zhang, J.; Al-Timemy, A.H.; et al. A Survey on Deep Learning Tools Dealing with Data Scarcity: Definitions, Challenges, Solutions, Tips, and Applications. J. Big Data 2023, 10, 46. [Google Scholar] [CrossRef]
  56. Gani, H.; Naseer, M.; Yaqub, M. How to Train Vision Transformer on Small-Scale Datasets? In Proceedings of the 33rd British Machine Vision Conference, London, UK, 21–24 November 2022.
  57. Wang, Z.; Wang, P.; Liu, K.; Wang, P.; Fu, Y.; Lu, C.-T.; Aggarwal, C.C.; Pei, J.; Zhou, Y. A Comprehensive Survey on Data Augmentation. arXiv 2024, arXiv:2405.09591. [Google Scholar]
  58. Goyal, M.; Mahmoud, Q.H. A Systematic Review of Synthetic Data Generation Techniques Using Generative AI. Electronics 2024, 13, 3509. [Google Scholar] [CrossRef]
  59. Gui, J.; Chen, T.; Zhang, J.; Cao, Q.; Sun, Z.; Luo, H.; Tao, D. A Survey on Self-Supervised Learning: Algorithms, Applications, and Future Trends. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 9052–9071. [Google Scholar] [CrossRef]
  60. Kim, H.E.; Cosa-Linan, A.; Santhanam, N.; Jannesari, M.; Maros, M.E.; Ganslandt, T. Transfer Learning for Medical Image Classification: A Literature Review. BMC Med. Imaging 2022, 22, 69. [Google Scholar] [CrossRef]
  61. Mei, X.; Liu, Z.; Robson, P.M.; Marinelli, B.; Huang, M.; Doshi, A.; Jacobi, A.; Cao, C.; Link, K.E.; Yang, T.; et al. RadImageNet: An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning. Radiol. Artif. Intell. 2022, 4, e210315. [Google Scholar] [CrossRef]
  62. Schäfer, R.; Nicke, T.; Höfener, H.; Lange, A.; Merhof, D.; Feuerhake, F.; Schulz, V.; Lotz, J.; Kiessling, F. Overcoming Data Scarcity in Biomedical Imaging with a Foundational Multi-Task Model. Nat. Comput. Sci. 2024, 4, 495–509. [Google Scholar] [CrossRef] [PubMed]
  63. Jafarigol, E.; Trafalis, T. A Review of Machine Learning Techniques in Imbalanced Data and Future Trends. arXiv 2023, arXiv:2310.07917. [Google Scholar]
  64. Araf, I.; Idri, A.; Chairi, I. Cost-Sensitive Learning for Imbalanced Medical Data: A Review. Artif. Intell. Rev. 2024, 57, 80. [Google Scholar] [CrossRef]
  65. Ganaie, M.A.; Hu, M.; Malik, A.K.; Tanveer, M.; Suganthan, P.N. Ensemble Deep Learning: A Review. Eng. Appl. Artif. Intell. 2022, 115, 105151. [Google Scholar] [CrossRef]
  66. Zhao, L.; Qian, X.; Guo, Y.; Song, J.; Hou, J.; Gong, J. MSKD: Structured Knowledge Distillation for Efficient Medical Image Segmentation. Comput. Biol. Med. 2023, 164, 107284. [Google Scholar] [CrossRef]
  67. Li, M.; Cui, C.; Liu, Q.; Deng, R.; Yao, T.; Lionts, M.; Huo, Y. Dataset Distillation in Medical Imaging: A Feasibility Study. arXiv 2024, arXiv:2407.14429. [Google Scholar]
  68. Qin, D.; Bu, J.; Liu, Z.; Shen, X.; Zhou, S.; Gu, J.; Wang, Z.; Wu, L.; Dai, H. Efficient Medical Image Segmentation Based on Knowledge Distillation. IEEE Trans. Med. Imaging 2021, 40, 3820–3831. [Google Scholar] [CrossRef]
  69. Xing, X.; Hou, Y.; Li, H.; Yuan, Y.; Li, H.; Meng, M.Q.-H. Categorical Relation-Preserving Contrastive Knowledge Distillation for Medical Image Classification. In Proceedings of the 24th International Conference on Medical Image Computing and Computer Assisted Intervention, Strasbourg, France, 27 September–1 October 2021. [Google Scholar]
  70. Du, S.; Bayasi, N.; Hamarneh, G.; Garbi, R. MDViT: Multi-Domain Vision Transformer for Small Medical Image Segmentation Datasets. In Proceedings of the 26th International Conference on Medical Image Computing and Computer Assisted Intervention, Vancouver, BC, Canada, 8–12 October 2023. [Google Scholar]
  71. Camino Benech, A.; Wang, Z.; Basu Bal, A.; Benmansour, F.; Carano, R.; Ferrara, D. Deep Learning Segmentation of Foveal Avascular Zone (FAZ) in Optical Coherence Tomography Angiography (OCTA) of Nonproliferative Diabetic Retinopathy. Investig. Ophthalmol. Vis. Sci. 2023, 64, 1125. [Google Scholar]
  72. Cunningham, H.; Ewart, A.; Riggs, L.; Huben, R.; Sharkey, L. Sparse Autoencoders Find Highly Interpretable Features in Language Models. arXiv 2023, arXiv:2309.08600. [Google Scholar]
  73. Alam, M.N.; Yamashita, R.; Ramesh, V.; Prabhune, T.; Lim, J.I.; Chan, R.V.P.; Hallak, J.; Leng, T.; Rubin, D. Contrastive Learning-Based Pretraining Improves Representation and Transferability of Diabetic Retinopathy Classification Models. Sci. Rep. 2023, 13, 6047. [Google Scholar] [CrossRef] [PubMed]
  74. Hervella, Á.S.; Rouco, J.; Novo, J.; Ortega, M. Multimodal Image Encoding Pre-Training for Diabetic Retinopathy Grading. Comput. Biol. Med. 2022, 143, 105302. [Google Scholar] [CrossRef] [PubMed]
  75. Sükei, E.; Rumetshofer, E.; Schmidinger, N.; Mayr, A.; Schmidt-Erfurth, U.; Klambauer, G.; Bogunović, H. Multi-Modal Representation Learning in Retinal Imaging Using Self-Supervised Learning for Enhanced Clinical Predictions. Sci. Rep. 2024, 14, 26802. [Google Scholar] [CrossRef]
  76. Zhou, Y.; Chia, M.A.; Wagner, S.K.; Ayhan, M.S.; Williamson, D.J.; Struyven, R.R.; Liu, T.; Xu, M.; Lozano, M.G.; Woodward-Court, P.; et al. A Foundation Model for Generalizable Disease Detection from Retinal Images. Nature 2023, 622, 156–163. [Google Scholar] [CrossRef]
  77. Wu, R.; Zhang, C.; Zhang, J.; Zhou, Y.; Zhou, T.; Fu, H. MM-Retinal: Knowledge-Enhanced Foundational Pretraining with Fundus Image-Text Expertise. In Proceedings of the International Conference on Medical Image Computing and Computer Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024. [Google Scholar]
Figure 1. Illustration of OCT, OCTA, and FAZ. (a) OCT en face obtained from the SVC layer; (b) corresponding OCTA en face, i.e., the layer between the dotted red lines in the corresponding OCT (bottom) (Spectralis OCT2, Heidelberg Eng.). (c) OCT Ascan (Gray) and (d) OCT Bscan overlaid by OCTA signal (yellow).
Figure 1. Illustration of OCT, OCTA, and FAZ. (a) OCT en face obtained from the SVC layer; (b) corresponding OCTA en face, i.e., the layer between the dotted red lines in the corresponding OCT (bottom) (Spectralis OCT2, Heidelberg Eng.). (c) OCT Ascan (Gray) and (d) OCT Bscan overlaid by OCTA signal (yellow).
Bioengineering 12 00334 g001
Figure 2. (a) OCTA image of the retina superficial vascular complex (SVC) in a normal eye, displaying microvessel trees surrounding the foveal avascular zone (FAZ), the capillary-free region near the center of the image. (b) The same image with the FAZ marked in green.
Figure 2. (a) OCTA image of the retina superficial vascular complex (SVC) in a normal eye, displaying microvessel trees surrounding the foveal avascular zone (FAZ), the capillary-free region near the center of the image. (b) The same image with the FAZ marked in green.
Bioengineering 12 00334 g002
Figure 3. Top: Multi-condition model architecture with a shared Base encoder-decoder for all conditions (blue) and condition-specific Teacher decoders for each target condition (yellow). Bottom: Structure of a single base-encoder block, altering the standard transformer block to include a context-independent kernel attention matrix and sparse autoencoder for improved generalizability. (Decoder blocks are the same, but with a CNN Decoder placed after the multi-head attention block).
Figure 3. Top: Multi-condition model architecture with a shared Base encoder-decoder for all conditions (blue) and condition-specific Teacher decoders for each target condition (yellow). Bottom: Structure of a single base-encoder block, altering the standard transformer block to include a context-independent kernel attention matrix and sparse autoencoder for improved generalizability. (Decoder blocks are the same, but with a CNN Decoder placed after the multi-head attention block).
Bioengineering 12 00334 g003
Figure 4. (a) Fundus image of the retina in a patient undergoing laser PC therapy. (b) Fundus image with masks highlighting the areas treated with laser PC.
Figure 4. (a) Fundus image of the retina in a patient undergoing laser PC therapy. (b) Fundus image with masks highlighting the areas treated with laser PC.
Bioengineering 12 00334 g004
Figure 5. Predicted masks, overlaid on OCTA scans of SVC layer of FAZ, for each of four conditions: (a) Healthy (b) ALZ (c) AMD (d) DR. First Row: OCTA scan of SVC layer of FAZ. Second Row: Predicted binary masks.
Figure 5. Predicted masks, overlaid on OCTA scans of SVC layer of FAZ, for each of four conditions: (a) Healthy (b) ALZ (c) AMD (d) DR. First Row: OCTA scan of SVC layer of FAZ. Second Row: Predicted binary masks.
Bioengineering 12 00334 g005
Table 1. Two-fold cross-validation dice indices of trained models.
Table 1. Two-fold cross-validation dice indices of trained models.
ModelHealthyALZAMDDRMean
Single-condition ModelWithout pretraining86.480.682.879.682.0
With pretraining86.683.182.880.883.1
Multi-condition ModelWithout pretraining85.483.782.983.383.7
With pretraining88.083.981.384.483.8
Table 2. Two-fold cross-validation dice indices of model trained using two annotations.
Table 2. Two-fold cross-validation dice indices of model trained using two annotations.
ModelHealthyALZAMDDRMean
Single-condition ModelWithout pretraining85.381.383.981.281.1
With pretraining86.484.185.481.084.2
Multi-condition ModelWithout pretraining86.284.082.382.583.5
With pretraining87.184.182.383.383.9
Table 3. Paired t-test p-values comparing model performance. (<0.05: reject null hypothesis).
Table 3. Paired t-test p-values comparing model performance. (<0.05: reject null hypothesis).
ModelHealthyALZAMDDR
Single-condition ModelWith/without
pretraining
0.005<0.0010.94<0.001
With/without
2 annotations
<0.0010.006<0.001<0.001
Multi-condition ModelWith/without
pretraining
<0.0010.21<0.001<0.001
With/without
2 annotations
<0.0010.090.030.001
Single-condition/Multi-condition Model<0.001<0.001<0.001<0.001
Table 4. Cohen’s d-threshold. (|d|: small: ≤ 0.2, medium: 0.5, large: ≥ 0.8).
Table 4. Cohen’s d-threshold. (|d|: small: ≤ 0.2, medium: 0.5, large: ≥ 0.8).
ModelHealthyALZAMDDR
Single-condition ModelWith/without
pretraining
0.31.80.00.86
With/without
2 annotations
−1.10.30.81.2
Multi-condition ModelWith/without
pretraining
3.20.1−0.70.6
With/without
2 annotations
0.60.2−0.2−0.3
Single-condition/Multi-condition Model1.72.61.30.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Racioppo, P.; Alhasany, A.; Pham, N.V.; Wang, Z.; Corradetti, G.; Mikaelian, G.; Paulus, Y.M.; Sadda, S.R.; Hu, Z. Automated Foveal Avascular Zone Segmentation in Optical Coherence Tomography Angiography Across Multiple Eye Diseases Using Knowledge Distillation. Bioengineering 2025, 12, 334. https://doi.org/10.3390/bioengineering12040334

AMA Style

Racioppo P, Alhasany A, Pham NV, Wang Z, Corradetti G, Mikaelian G, Paulus YM, Sadda SR, Hu Z. Automated Foveal Avascular Zone Segmentation in Optical Coherence Tomography Angiography Across Multiple Eye Diseases Using Knowledge Distillation. Bioengineering. 2025; 12(4):334. https://doi.org/10.3390/bioengineering12040334

Chicago/Turabian Style

Racioppo, Peter, Aya Alhasany, Nhuan Vu Pham, Ziyuan Wang, Giulia Corradetti, Gary Mikaelian, Yannis M. Paulus, SriniVas R. Sadda, and Zhihong Hu. 2025. "Automated Foveal Avascular Zone Segmentation in Optical Coherence Tomography Angiography Across Multiple Eye Diseases Using Knowledge Distillation" Bioengineering 12, no. 4: 334. https://doi.org/10.3390/bioengineering12040334

APA Style

Racioppo, P., Alhasany, A., Pham, N. V., Wang, Z., Corradetti, G., Mikaelian, G., Paulus, Y. M., Sadda, S. R., & Hu, Z. (2025). Automated Foveal Avascular Zone Segmentation in Optical Coherence Tomography Angiography Across Multiple Eye Diseases Using Knowledge Distillation. Bioengineering, 12(4), 334. https://doi.org/10.3390/bioengineering12040334

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop