Next Article in Journal
Research Progress on Color Image Quality Assessment
Previous Article in Journal
FDMNet: A Multi-Task Network for Joint Detection and Segmentation of Three Fish Diseases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient Retinal Vessel Segmentation with 78K Parameters

School of Computer Science, Hunan University of Technology, Zhuzhou 412007, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2025, 11(9), 306; https://doi.org/10.3390/jimaging11090306
Submission received: 13 August 2025 / Revised: 26 August 2025 / Accepted: 2 September 2025 / Published: 8 September 2025
(This article belongs to the Section Image and Video Processing)

Abstract

Retinal vessel segmentation is critical for early diagnosis of diabetic retinopathy, yet existing deep models often compromise accuracy for complexity. We propose DSAE-Net, a lightweight dual-stage network that addresses this challenge by (1) introducing a Parameterized Cascaded W-shaped Architecture enabling progressive feature refinement with only 1% of the parameters of a standard U-Net; (2) designing a novel Skeleton Distance Loss (SDL) that overcomes boundary loss limitations by leveraging vessel skeletons to handle severe class imbalance; (3) developing a Cross-modal Fusion Attention (CMFA) module combining group convolutions and dynamic weighting to effectively expand receptive fields; and (4) proposing Coordinate Attention Gates (CAGs) to optimize skip connections via directional feature reweighting. Evaluated extensively on DRIVE, CHASE_DB1, HRF, and STARE datasets, DSAE-Net significantly reduces computational complexity while outperforming state-of-the-art lightweight models in segmentation accuracy. Its efficiency and robustness make DSAE-Net particularly suitable for real-time diagnostics in resource-constrained clinical settings.

1. Introduction

Retinal vessel segmentation [1] is a fundamental task in medical image analysis of diseases such as diabetic retinopathy and glaucoma, where its accuracy critically impacts the early diagnosis and treatment efficacy assessment [2]. Traditional approaches to achieving retinal vessel segmentation rely on manually selecting features from fundus images and classifying them at the pixel level based on these extracted features. However, manual methods are inherently subjective in feature selection, as these require experts with substantial domain expertise, and are ill-suited for large-scale segmentation tasks. The advent of artificial intelligence [3,4], particularly the rapid advancements in deep learning [5,6,7,8], has revolutionized this field. Deep learning-based vessel segmentation methods have continuously set new performance benchmarks on public retinal vessel segmentation datasets [9]. Nevertheless, existing models commonly confront two critical dilemmas: On the one hand, large models exemplified by vision transformers (ViTs) [10,11] could enhance segmentation accuracy through complex architectures. Although several efficient transformer variants—such as those using linear attention, directional connectivity [12], or hybrid designs [13]—have been developed in general vision tasks to alleviate computational burdens, their direct application to retinal vessel segmentation remains limited. These methods often still struggle to achieve an optimal balance between efficiency and the capacity to capture fine vascular details, particularly the delicate and complex structures of capillaries, under tight computational constraints. On the other hand, some lightweight designs could lower the barrier to deployment but often sacrifice the ability to accurately identify vessels, particularly capillaries. This trade-off is particularly acute in primary-level clinical settings, which frequently lack access to high-performance computing resources, yet the precise segmentation of capillaries in high-resolution fundus images imposes stringent demands on model capability.
In recent years, research into deep learning [14] has made significant progress in terms of semantic segmentation techniques. Convolutional Neural Network (CNN)-based methods, including Fully Convolutional Networks (FCN) [15], SegNet [16], the DeepLab series [17,18,19], and other related approaches [20,21,22,23], have been successively developed. Furthermore, various large general-purpose segmentation models, like the Segment Anything Model series [24,25], have achieved considerable success. However, in the medical domain, due to its specific characteristics, the application of these technologies faces constraints. For example, deploying large general-purpose models often entails prohibitive computational costs. Other industries might mitigate this by deploying models on the cloud or utilizing fine-tuned public models, though the medical field grapples with significant ethical and patient privacy concerns surrounding medical image data. Consequently, practical applications necessitate highly tailored solutions. Beyond large models, models for automatic semantic segmentation of medical images remain a substantial challenge. Medical images typically exhibit complex structures and contain minute targets, and boundaries between anatomical structures are often indistinct and overlap with surrounding tissue. Among existing architectures, U-Net, [26] along with its variants, has demonstrated distinct advantages in the field of medical image segmentation. By combining an encoder–decoder structure with skip connections, U-Net effectively integrates multi-scale features. However, the original U-Net has three significant limitations: First, the limited receptive field of the shallow convolutional blocks hinders their ability to capture the topology of long-range vascular details. Second, the coarse nature of skip connections leads to inefficient fusion of low-level features with high-level semantic information. Third, although variants like Attention U-Net [27] and TransUNet [28] incorporate attention mechanisms to address some issues, the redundancy of the introduced modules results in exponential growth in parameters. Some studies such as ConvUNeXt [29,30] have attempted to optimize feature extraction using large-kernel convolutions, while their computational complexity remains problematic for real-time diagnostic applications. Conversely, approaches like ULite [31] and U-Next [32] achieve significant speed improvements through depthwise separable convolutions but at the expense of segmentation accuracy. This series of problems highlights the core research challenge: how to achieve precise segmentation of vascular structures, especially micro-vessels, while simultaneously ensuring model lightweightness.
To address these challenges, this paper proposes the Dual-Stage Attention-Enhanced Network (DSAE-Net). The main contributions of this work are summarized as follows:
  • DSAE-Net Architecture: A novel cascaded lightweight architecture for retinal vessel segmentation, featuring a parameterized W-shaped topology that implements a dual-stage feature refinement mechanism for progressive segmentation. Experiments demonstrate that DSAE-Net ranks among the top-performing models within the category of highly efficient and compact networks.
  • Skeleton Distance Loss (SDL): A novel loss function that overcomes boundary loss limitations by leveraging vessel skeletons to handle severe class imbalance.
  • Convolutional Module with Feature Attention (CMFA): A novel module introduced for feature extraction, fusing grouped convolutions with feature weight redistribution. Grounded in channel grouping and weight redistribution concepts, CMFA effectively enlarges the receptive field while preserving spatial precision, helping to navigate the trade-off between lightweightness and performance.
  • Coordinate Attention Gate (CAG): A plug-and-play module is designed for direction-sensitive feature reweighting during fusion, effectively suppressing noise interference inherent to skip connections. Combined with the cascaded architecture’s refinement strategy, this significantly enhances model precision.

2. Methods

We proposed the Dual-Stage Attention-Enhanced Network (DSAE-Net), which innovatively cascades two modified U-Net architectures, termed Cross-modal Attention-Enhanced U-Net (CAE-UNet), to form a W-shaped network topology endowed with feature refinement capabilities. As illustrated in Figure 1, the overall architecture comprises two cascaded CAE-UNet modules, enabling progressive feature learning through a feature re-focusing mechanism. This section details the improvement from the standard U-Net to the DSAE-Net.

2.1. Dual U-Net Cascaded Architecture

The network adopts a W-shaped topology. It includes the primary feature extraction module, CAE-UNet1, which receives the raw input image and performs initial feature encoding, and the second feature extraction module, CAE-UNet2, which serves as the feature refinement stage. CAE-UNet2 concatenates the original input image with the preliminary prediction from the first stage along the channel dimension to achieve feature re-calibration. Thus during training, outputs from both modules are supervised ( Loss mid , Loss end ), and only the refined output from the second stage ( Loss end ) is retained during inference. The experimental findings demonstrate that the feature refinement strategy inherent to this cascaded architecture effectively enhances model precision.
We designed a parametric and scalable encoder–decoder architecture for the constituent CAE-UNet modules. The network topology is governed by a tuple T: ( k , m ), where
  • k N + denotes the network depth;
  • m N + denotes the number of kernels in the top convolutional layer.
Under this parametric design, a standard U-Net with a depth of 5 and 64 kernels in the top layer possesses approximately 7.76 M parameters, corresponding to the tuple (5, 64). While the conventional U-Net’s channel expansion rule (typically doubling channels per layer) has proven effective, increasing network depth(k)or doubling the kernel base (m)significantly escalates model size. Therefore, adjusting the values of ( k , m ) enables rapid structural compression or expansion of the U-shaped architecture. The experimental results identified that a CAE-UNet configured with (3, 8) achieves competitive performance while utilizing only 0.5% of the parameters of the standard U-Net.

2.2. Skeleton Distance Loss

To address the significant class imbalance inherent to retinal vessel segmentation, we propose a novel Skeleton Distance Loss (SDL) designed to overcome limitations observed in existing boundary loss. While boundary loss [33] utilizes distance maps for edge-aware optimization, it suffers from the generation of negative weights within large foreground mask regions (Figure 2B), which can impede effective learning during the early stages of training. This limitation stems from the formulation of boundary loss (Equation (1)):
1 2 Dist ( G , S ) Ω ϕ G ( q ) s ( q ) d q Ω ϕ G ( q ) g ( q ) d q
where, in the case of large foreground mask regions, the second integral term can dominate and cause the overall value of the formula to become negative, as visualized by the blue regions in the distance map. Our SDL is defined as an integral over the image domain Ω :
L S D L ( θ ) = Ω ϕ K ( q ) s θ ( q ) d q
Here, ϕ K : Ω R + represents the Euclidean distance transform computed from the vessel skeleton K, which is extracted directly from the ground truth mask g via skeletonization [34]. Crucially, ϕ K ( q ) denotes the non-negative distance from pixel q to the nearest point on the vessel skeleton (Figure 2E), ensuring that SDL avoids the detrimental negative weights associated with boundary loss in large foreground areas. The term s θ ( q ) corresponds to the model’s predicted vessel probability at pixel q, parameterized by weights θ . By penalizing errors proportionally to their distance from the vessel centerlines, SDL inherently prioritizes topological fidelity, particularly for thin structures like capillaries.
Since SDL, akin to boundary loss, is a linear function of the network’s softmax probability outputs, it can be seamlessly integrated with standard regional losses L r e g ( θ ) (e.g., Dice loss, cross-entropy loss) to form a composite segmentation loss:
L s ( θ ) = α L S D L ( θ ) + ( 1 α ) L r e g ( θ )
The hyperparameter α [ 0 , 0.7 ] governs the relative contribution of SDL. To balance global feature learning with boundary refinement, α is initially small, allowing L r e g to dominate during early training. As training progresses, α is gradually increased, shifting the focus towards precise boundary delineation guided by the vessel skeletons. This progressive strategy effectively enhances segmentation accuracy for vascular structures. Finally, the overall loss function for training the cascaded DSAE-Net is defined as
L e n h a n c e d ( y ^ , y ) = k = 0 1 w k L s ( y ^ k , y )
This combined loss incorporates the segmentation errors from both cascaded stages ( k 0 , 1 ), with stage-specific weights w k adjusted during training. This approach ensures the first stage produces a reasonably accurate base segmentation upon which the second stage performs refinement.

2.3. Lightweight Contextual Convolution Module

To enhance feature representation within a lightweight framework, we propose the Contextual Convolution Module with Feature Attention (CMFA), depicted in Figure 3D and Figure 4. CMFA replaces standard convolution blocks in the U-Net encoder/decoder. Its core innovation is the integration of a Contextual Transformer (CoT) mechanism with a multi-branch feature reuse strategy. The module operates as follows: (1) the input feature map F undergoes a 1 × 1 convolution followed by channel-wise splitting into two groups, F 1 and F 2 . F 1 preserves the original features, while F 2 feeds into the CoT module. (2) Within CoT, F 2 is processed by two sequential 3 × 3 convolutions for local feature aggregation, followed by dynamic feature reweighting via the CoT attention mechanism to produce F 3 . (3) The features F 1 , F 2 , and F 3 are concatenated channel-wise. A final 1 × 1 convolution facilitates cross-channel information fusion, yielding the output feature map. Mathematically, key steps are
[ F 1 , F 2 ] = Chunk Conv 1 × 1 ( F )
F 3 = CoT Conv 3 × 3 Conv 3 × 3 ( F 2 )
Out = Conv 1 × 1 Concat F 1 , F 2 , F 3
where Chunk denotes channel splitting, Conv represents convolution, and CoT implements the Contextual Transformer algorithm [35]. This design preserves original features ( F 1 ), enhances critical features via dynamic attention ( F 3 ), and enriches feature diversity through fusion. Within the overall CAE-UNet block structure (Figure 3D), CMFA is utilized as
Out = F BN ReLU CMFA BN ReLU CMFA ( F )
where ⊕ denotes element-wise addition, BN is batch normalization, and ReLU is the activation function. Experiments confirm that replacing standard convolutions with CMFA significantly improves performance while maintaining model efficiency.

2.4. Optimization for Decoder

To effectively capture both low-level and high-level semantic features in medical images, it is essential to integrate low-level semantics during upsampling. To refine U-Net’s direct skip connections, a Coordinate Attention Gate (CAG) is designed (Figure 5). It utilizes a coordinate attention mechanism [36] to decouple directional pooling, thereby capturing long-range dependencies along the vertical and horizontal directions. This output is then concatenated with upsampled features. Mathematically it can be shown as
Out = ResConvB CAG X , Skip
where CAG is the Coordinate Attention Gate module we proposed and ResConvB denotes a residual convolution block added to U-Net’s decoder for enhanced feature extraction. This design strengthens target localization capabilities, with experimental results confirming its significant improvements.

3. Results

3.1. Datasets

The proposed model was evaluated on four widely used public retinal vessel segmentation datasets, selected to assess different aspects of model performance:
  • DRIVE [37]: A foundational dataset containing 40 JPEG color fundus images (size: 565 × 584 pixels), including 7 cases with pathologies. It provides 20 training and 20 test images. The training set was split into 15 images for training and 5 for validation. Purpose: It serves as the primary benchmark for general segmentation performance and model comparison.
  • CHASE_DB1 [38]: A dataset comprising 28 retinal images (size: 999 × 960 pixels) captured from both eyes of 14 children. The dataset was randomly divided into 16 images for training, 6 for validation, and 6 for testing. Purpose: It can be used to validate model applicability and robustness across a younger demographic cohort.
  • HRF (High-Resolution Fundus) [39]: This dataset features high-resolution images (size: 3504 × 2336 pixels) from 15 healthy subjects, 15 diabetic retinopathy patients, and 15 glaucoma patients. Images were split into 20 for training, 5 for validation, and 20 for testing. Purpose: It can be used to test model performance on high-resolution imagery and is able to handle fine details like capillaries under varying health conditions.
  • STARE [40]: This dataset contains retinal vessel images (size: 605 × 700 pixels) exhibiting a variety of pathologies. The 20 images were split into 12 for training, 4 for validation, and 4 for testing. Purpose: It can be used to evaluate model robustness and segmentation accuracy in the presence of diverse pathological manifestations.
Table 1 summarizes the key characteristics and data splits of these datasets.

3.2. Details of Implementation

The Dice coefficient [41] served as the primary evaluation metric. All models were trained using the Adam optimizer [42] with an initial learning rate of λ = 10 2 , decaying via cosine annealing scheduling to λ = 10 8 . Training spanned 200 epochs (20 cycles of 10 epochs each) using cross-entropy loss. Input images were resized to 512 × 512 pixels to simulate primary healthcare computational constraints. Data augmentation included random resizing/cropping, horizontal/vertical flipping, and random adjustments to brightness, contrast, saturation, and hue. The batch size was set to 4 across all experiments. The highest validation AUC achieved by the model at the end of each cycle was retained for testing. Test-time augmentation (horizontal/vertical flipping) was applied during inference, with the final evaluation conducted at the original image resolutions.

3.3. Experimental Results

DSAE-Net and ten comparative models (UNet, Attention UNet, ResUNet++ [43], SA-UNet [44], ConvUNeXt, FR-Unet [45], PDF-UNet [46], ULite, LMFR-Net [47] and MSDANet [48]) were evaluated across the DRIVE, CHASE-DB, HRF, and STARE datasets. Performance was assessed using efficiency metrics (parameters, FLOPs, inference time) and segmentation metrics (accuracy (Acc), area under the ROC curve (AUC), Dice coefficient (Dice), Matthews correlation coefficient (MCC)and 95th percentile of Hausdorff distance (HD95)). Segmentation metrics are defined as
Acc = T P + T N T P + T N + F P + F N
Dice = 2 × T P 2 × T P + F P + F N
MCC = T P × T N F P × F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
HD 95 = percentile 95 % max min a A d ( a , B ) , min b B d ( b , A )
where T P = true positive, T N = true negative, F P = false positive, F N = false negative. In Equation (13), A and B represent the set of points on the boundary of the predicted segmentation and the ground truth, respectively, d ( a , B ) is the distance from point a to the closest point in set B, and the 95th percentile is used to mitigate sensitivity to outliers. MCC was prioritized to mitigate class imbalance effects, while HD95 provides a stringent assessment of the maximum segmentation boundary error.
Table 2 and Figure 6 present model efficiency metrics (calculated on 512 × 512 images). DSAE-Net demonstrates exceptional efficiency, achieving the second-lowest parameter count (0.08 M) and competitive FLOPs (4.76 G) while maintaining fast inference (0.67 s for 10 images). This represents a 99% reduction in parameters compared to the average comparator (7.54 M) and a 37% reduction in inference time (1.32 s avg.).
DSAE-Net demonstrates state-of-the-art or highly competitive performance compared to existing models across all benchmark datasets (see comparative results in Table 3). Furthermore, it exhibits consistently high and statistically robust performance over multiple training runs, as detailed by the average metrics and 95% confidence intervals in Table 4.
DRIVE: DSAE-Net ranks first in all metrics (Acc: 95.48%; AUC: 98.00%; Dice: 82.44%; MCC: 79.86%), surpassing MSDANet (second-best) by 0.02-0.03% while using only 0.14% of its parameters (0.08 M vs. 55.55 M). Notably, DSAE-Net achieved the lowest HD95 (12.12) on the DRIVE dataset, significantly outperforming all other models. This indicates that our model produces segmentation maps with the most precise boundaries, a critical factor for accurately capturing thin capillaries.
CHASE-DB: DSAE-Net achieves the highest Dice (81.46%) and MCC (79.54%), exceeding ResUNet++ (second-best) by 0.17% and 0.21%, demonstrating robustness across age groups.
HRF: While Attention UNet holds a slight lead (Dice: 80.46%; MCC: 78.46%), DSAE-Net delivers highly competitive accuracy (Dice: 80.12%; MCC: 78.10%), with only 0.2% of the parameters (0.08 M vs. 34.88 M), proving its exceptional efficiency for processing high-resolution images. We attribute the minor performance gap primarily to the resolution mismatch between training and test data. Furthermore, DSAE-Net attained a competitive HD95 of 126.25, on par with the much larger Attention UNet (127.14) and superior to other lightweight methods, demonstrating its robustness in segmenting fine vessels under pathological conditions.
STARE: DSAE-Net excels with the highest Dice (81.92%) and MCC (80.25%), outperforming ConvUNeXt (second-best) by 0.39% and 0.47%, confirming its effectiveness on pathological images.

3.3.1. Visual Comparison

Visual analysis in Figure 7 (rows 1–4: DRIVE, CHASE_DB1, HRF, STARE) confirms DSAE-Net’s superior vessel segmentation fidelity against six benchmarks (UNet, Attention UNet, SA-UNet, ConvUNeXt, ULite), demonstrating (1) 18% fewer capillary false negatives (FNs, blue) than UNet/AttUNet in sparse regions; (2) 28% fewer complex-branch false positives (FPs, green) versus ConvUNeXt/ULite on STARE; and (3) unprecedented micro-vessel continuity in HRF with higher precision than its lightweight peers. Crucially, while several industry-wide limitations persist—branch oversegmentation (ConvUNeXt/ULite exhibit 0.4× more FPs than DSAE-Net), capillary underdetection, and HRF resolution sensitivity (all models suffer from cross-scale degradation)—DSAE-Net achieves the optimal balance. Residual STARE branch FPs (SDL’s terminal-weighting bias) and HRF micro-FNs (training-test resolution gap) represent focal improvement targets, with planned multi-scale augmentation and SDL spatial-prior refinement poised to advance state-of-the-art robustness boundaries.

3.3.2. Parametric Design Study

To validate the effectiveness of the proposed parametric design, an ablation study was conducted on the DRIVE dataset by systematically varying the tuple T: (k, m) across configurations [(3, 4), (3, 8), (3, 16), (4, 8), (4, 16), (4, 32)]. As shown in Table 5, the results demonstrate robust performance across all parameterizations, with test Dice peaking at intermediate complexity levels. Notably, while increasing the depth (k) or width (m) of the model could improved Dice after validation (e.g., 0.8340 for (4, 16) vs. 0.8282 for (3, 8)), this did not translate to proportional test set gains (0.8268 vs. 0.8244). The configuration (k = 3, m = 8) achieved an optimal balance between computational efficiency (0.08 M parameters, 4.76 G FLOPs) and segmentation accuracy (82.44% test Dice), exhibiting negligible performance degradation compared to larger models while significantly reducing resource demands; thus, it was established as the optimal parameterization for DSAE-Net.

3.3.3. Ablation Study on Skeleton Distance Loss

To validate the efficacy of the proposed Skeleton Distance Loss (SDL), comprehensive experiments were conducted on the DRIVE dataset, comparing standard cross-entropy (CE) and Dice losses with their SDL-enhanced variants. As demonstrated in Figure 8, SDL integration consistently elevates final segmentation quality across both loss functions, resulting in a substantially improved mean Intersection over Union (mIoU). Crucially, SDL fundamentally enhances training stability: when combined with CE loss, it accelerates convergence while dramatically reducing performance oscillations throughout optimization. For Dice loss, SDL effectively mitigates characteristic early-stage instability, yielding smoother learning curves with significantly lower variance. These synergistic effects establish SDL as a universal training stabilizer that not only boosts final segmentation accuracy but also ensures more reliable and consistent model optimization across diverse loss functions.
Furthermore, we conducted a sensitivity analysis on the balancing factor α in the combined loss function (Equation (3)) to quantitatively assess its impact. Table 6 presents the performance on the DRIVE dataset with varying α values.
The analysis reveals that even a small SDL weight ( α = 0.3 ) produces a notable improvement in HD95 (from 23.05 to 20.45 on the test set), confirming that incorporating geometric constraints immediately refines segmentation boundaries. The optimal balance is achieved at α = 0.7 , which yields the highest Dice score and the lowest HD95 (12.68) on the test set, demonstrating a superior trade-off between overall segmentation accuracy and precise boundary delineation.
The performance at α = 1.0 offers crucial insight into the loss dynamics. While achieving the best HD95 on the validation set, its final Dice score on the validation set is marginally lower, and more importantly, its test set HD95 (15.35) is significantly worse than at α = 0.7 . This performance pattern can be attributed to altered training dynamics: relying solely on L S D L appears to slow the convergence rate, potentially preventing the model from reaching a fully optimized solution within the fixed training epochs. Furthermore, the increased gap between validation and test performance at α = 1.0 suggests that the pure SDL loss may overemphasize learning precise skeletal geometries from the training set, potentially at the expense of learning more generalized, robust features that are invariant to slight anatomical variations present in the test set.
These findings collectively establish SDL as a universal training stabilizer that not only boosts final segmentation accuracy but also ensures more reliable and consistent model optimization. The sensitivity analysis confirms that a balanced weighting ( α = 0.7 ) is crucial for achieving optimal and generalizable performance, as it synergistically leverages the fast convergence and robust features learned by the Dice loss with the geometric precision guidance of the SDL.

3.3.4. Ablation Study

Comprehensive ablation experiments on the DRIVE and CHASEDB1 datasets (Table 7) quantify the incremental contributions of each module within our lightweight cascaded framework. The baseline model (without CAG or CMFA) establishes fundamental segmentation capability. Significantly, integrating the CMFA module—while replacing standard convolutions with minimal parameter overhead—yields measurable gains, boosting DRIVE Dice by 0.02% and CHASEDB1 Dice by 1.92%. The CAG module then optimizes multi-scale feature fusion via skip connections, delivering substantial performance lifts: MCC increases by 1.33% on DRIVE and 2.51% on CHASEDB1 with only 0.02M additional parameters. Crucially, the full DSAE-Net configuration synergistically combines both modules, achieving peak metrics (82.44% Dice/79.86% MCC on DRIVE) while maintaining an ultra-lightweight design (0.08 M parameters). This ablation sequence confirms that CMFA and CAG collectively overcome parametric constraints through (i) efficient local–global feature integration (CMFA) and (ii) directionally optimized feature fusion (CAG).

4. Conclusions and Discussion

We proposed DSAE-Net, a lightweight dual-stage attention-enhanced network that achieves breakthrough parameter efficiency in retinal vessel segmentation through three core innovations: (1) a parametric cascaded W-shaped encoder, enabling progressive feature refinement with only 0.08 M parameters; (2) Skeleton Distance Loss (SDL), leveraging vessel topology priors to resolve severe class imbalance; and (3) coordinated attention mechanisms (CMFA + CAG) that synergistically expand receptive fields while optimizing multi-scale feature fusion.
Extensive validation on the DRIVE, CHASE_DB1, HRF, and STARE datasets demonstrates DSAE-Net’s superior trade-off between accuracy and complexity: it attains state-of-the-art or competitive segmentation metrics (Acc/AUC/Dice/MCC) while utilizing merely 1% of U-Net’s parameters and 2.2% of ConvUNeXt’s. Ablation studies confirm the critical synergies within our cascaded framework, where CMFA and CAG collectively enhance capillary recovery through global dependency modeling and directional feature reweighting.
Despite these advances, two limitations warrant further investigation: (1) occasional oversegmentation of complex vascular branches, attributed to SDL’s spatial weighting strategy, and (2) underdetection of subtle vessels in high-resolution HRF images, exacerbated by training–test resolution discrepancies. The clinical implications of these limitations are noteworthy. Oversegmentation could introduce false positives into automated screening systems, potentially leading to unnecessary referrals. More critically, the underdetection of micro-vessels, as observed in HRF, poses a greater risk, as it may result in missing early signs of diabetic retinopathy (e.g., microaneurysms) or impaired capillary perfusion indicative of glaucoma. Future work will focus on adaptive kernel selection and SDL’s spatial prior refinement—particularly optimizing its terminal vessel weighting—to enhance micro-vessel delineation without compromising efficiency. We will also explore comparisons with emerging lightweight vision transformers to further benchmark our architectural approach. These refinements are expected to solidify DSAE-Net’s applicability in clinical screening environments with diverse imaging protocols.

Author Contributions

Conceptualization, Z.Z.; Methodology, J.L. and Z.Z.; Software, J.L.; Validation, J.L. and K.L.; Investigation, J.L.; Resources, Z.Z.; Data curation, J.L. and K.L.; Writing—original draft, J.L.; Writing—review and editing, Z.Z., X.H., Y.Z. and X.Y.; Funding acquisition, X.Y. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of Hunan Province (Grant no. 2025JJ70071, 2025JJ70028, 2025JJ81178, 2024JJ9550) and the Scientific Research Project of the Education Department of Hunan Province (Grant no. 24A0401).

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the data used were obtained from the public databases.

Informed Consent Statement

Patient consent was waived due to the data used were obtained from the public databases.

Data Availability Statement

The Drive and CHASE-DB1 datasets, along with the HRF dataset, are publicly available at https://gitcode.com/open-source-toolkit/33a91 (accessed on 1 September 2025). The STARE dataset is available at http://cecas.clemson.edu/~ahoover/stare/ (accessed on 1 September 2025). The code for DSAE-Net is available at https://anonymous.4open.science/r/DSAE-Net-2CB1 (accessed on 1 September 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. The mean and standard deviation results from 5-fold cross-validation for UNet, Attention UNet, SA-UNet, SA-UNet, ConvUNeXt, ULite, and DSAE-Net.
Table A1. The mean and standard deviation results from 5-fold cross-validation for UNet, Attention UNet, SA-UNet, SA-UNet, ConvUNeXt, ULite, and DSAE-Net.
ModelAcc (%)AUC (%)Dice (%)MCC (%)HD95
UNet94.87 ± 0.0896.99 ± 0.1280.00 ± 0.4377.06 ± 0.4537.96 ± 2.96
Attention UNet95.42 ± 0.1398.04 ± 0.0782.32 ± 0.0979.74 ± 0.1120.03 ± 1.98
SA-UNet95.44 ± 0.0997.95 ± 0.1282.42 ± 0.0879.82 ± 0.1220.84 ± 1.24
ConvUNeXt95.46 ± 0.1297.76 ± 0.1082.31 ± 0.2079.71 ± 0.2615.63 ± 2.66
ULite95.40 ± 0.2797.48 ± 0.2981.77 ± 0.4479.14 ± 0.6121.83 ± 2.92
DSAE-Net95.48 ± 0.0298.00 ± 0.0482.44 ± 0.1079.86 ± 0.1112.12 ± 2.17

References

  1. Fathi, A.; Naghsh-Nilchi, A.R. Automatic wavelet-based retinal blood vessels segmentation and vessel diameter estimation. Biomed. Signal Process. Control 2013, 8, 71–80. [Google Scholar] [CrossRef]
  2. Qin, Q.; Chen, Y. A review of retinal vessel segmentation for fundus image analysis. Eng. Appl. Artif. Intell. 2024, 128, 107454. [Google Scholar] [CrossRef]
  3. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems 30; The MIT Press: Cambridge, MA, USA, 2017. [Google Scholar] [CrossRef]
  4. Wang, L.; Pan, Z.; Wang, J. A Review of Reinforcement Learning Based Intelligent Optimization for Manufacturing Scheduling. Complex Syst. Model. Simul. 2021, 1, 257–270. [Google Scholar] [CrossRef]
  5. Cui, Z.; Xue, F.; Cai, X.; Cao, Y.; Wang, G.; Chen, J. Detection of Malicious Code Variants Based on Deep Learning. IEEE Trans. Ind. Inf. 2018, 14, 3187–3196. [Google Scholar] [CrossRef]
  6. Zhang, K.; Su, Y.; Guo, X.; Qi, L.; Zhao, Z. MU-GAN: Facial Attribute Editing Based on Multi-Attention Mechanism. IEEE/CAA J. Autom. Sin. 2021, 8, 1614–1626. [Google Scholar] [CrossRef]
  7. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar] [CrossRef]
  8. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
  9. Chen, C.; Chuah, J.H.; Ali, R.; Wang, Y. Retinal Vessel Segmentation Using Deep Learning: A Review. IEEE Access 2021, 9, 111985–112004. [Google Scholar] [CrossRef]
  10. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
  11. Liu, Y.; Wang, H.; Chen, Z.; Huangliang, K.; Zhang, H. TransUNet+: Redesigning the skip connection to enhance features in medical image segmentation. Knowl.-Based Syst. 2022, 256, 109859. [Google Scholar] [CrossRef]
  12. Yang, Z.; Farsiu, S. Directional Connectivity-Based Segmentation of Medical Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 11525–11535. [Google Scholar] [CrossRef]
  13. Xie, Y.H.; Huang, B.S.; Li, F. UnetTransCNN: Integrating Transformers with Convolutional Neural Networks for Enhanced Medical Image Segmentation. Front. Oncol. 2025, 15, 1467672. [Google Scholar] [CrossRef] [PubMed]
  14. Yan, Y.; Qin, B. Research on Lung Nodule Segmentation Method Based on RSE-Vnet Convolutional Network. J. Hunan Univ. Technol. 2025, 39, 46–51. [Google Scholar] [CrossRef]
  15. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar] [CrossRef]
  16. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  17. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.P.; Yuille, A.L. Semantic Image Segmentation with Deep Convolutional Nets and Fully Connected CRFs. arXiv 2014, arXiv:1412.7062. [Google Scholar] [CrossRef]
  18. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  19. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar] [CrossRef]
  20. Sun, K.; Chen, Y.; Chao, Y.; Geng, J.; Chen, Y. A retinal vessel segmentation method based improved U-Net model. Biomed. Signal Process. Control 2023, 82, 104574. [Google Scholar] [CrossRef]
  21. Li, W.; Zhang, M.; Chen, D. Fundus Retinal Blood Vessel Segmentation Based on Active Learning. In Proceedings of the 2020 International Conference on Computer Information and Big Data Applications (CIBDA), Guiyang, China, 17–19 April 2020; pp. 264–268. [Google Scholar] [CrossRef]
  22. Uysal, E.; Güraksin, G.E. Computer-aided retinal vessel segmentation in retinal images: Convolutional neural networks. Multimed. Tools Appl. 2021, 80, 3505–3528. [Google Scholar] [CrossRef]
  23. Galdran, A.; Anjos, A.; Dolz, J.; Chakor, H.; Lombaert, H.; Ayed, I.B. State-of-the-art retinal vessel segmentation with minimalistic models. Sci. Rep. 2022, 12, 6174. [Google Scholar] [CrossRef]
  24. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment Anything. arXiv 2023, arXiv:2304.02643. [Google Scholar] [CrossRef]
  25. Ma, J.; He, Y.; Li, F.; Han, L.; You, C.; Wang, B. Segment anything in medical images. Nat. Commun. 2024, 15, 654. [Google Scholar] [CrossRef] [PubMed]
  26. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar] [CrossRef]
  27. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.J.; Heinrich, M.P.; Misawa, K.; Mori, K.; McDonagh, S.G.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-Net: Learning Where to Look for the Pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  28. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  29. Han, Z.; Jian, M.; Wang, G.-G. ConvUNeXt: An efficient convolution neural network for medical image segmentation. Knowl.-Based Syst. 2022, 253, 109512. [Google Scholar] [CrossRef]
  30. Yin, Y.; Han, Z.; Jian, M.; Wang, G.-G.; Chen, L.; Wang, R. AMSUnet: A neural network using atrous multi-scale convolution for medical image segmentation. Comput. Biol. Med. 2023, 162, 107120. [Google Scholar] [CrossRef]
  31. Dinh, B.D.; Nguyen, T.T.; Tran, T.T.; Pham, V.T. 1M parameters are enough? A lightweight CNN-based model for medical image segmentation. In Proceedings of the 2023 Asia Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Taipei, Taiwan, 31 October–3 November 2023; pp. 1279–1284. [Google Scholar] [CrossRef]
  32. Valanarasu, J.M.J.; Patel, V.M. UNeXt: MLP-Based Rapid Medical Image Segmentation Network. In Proceedings of the Medical Image Computing and Computer Assisted Intervention (MICCAI), Singapore, 18–22 September 2022; pp. 23–33. [Google Scholar] [CrossRef]
  33. Kervadec, H.; Bouchtiba, J.; Desrosiers, C.; Granger, E.; Dolz, J.; Ayed, I.B. Boundary loss for highly unbalanced segmentation. In Proceedings of the International Conference on Medical Imaging with Deep Learning (MIDL), London, UK, 8–10 July 2019; pp. 285–296. [Google Scholar]
  34. Ma, J.; Ren, X.; Tsviatkou, V.Y. A novel fully parallel skeletonization algorithm. Pattern Anal. Appl. 2022, 25, 169–188. [Google Scholar] [CrossRef]
  35. Li, Y.; Yao, T.; Pan, Y.; Mei, T. Contextual Transformer Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 1489–1500. [Google Scholar] [CrossRef]
  36. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar] [CrossRef]
  37. Staal, J.; Abramoff, M.D.; Niemeijer, M.; Viergever, M.A.; Ginneken, B.v. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef]
  38. Owen, C.G.; Rudnicka, A.R.; Nightingale, C.M.; Mullen, R.; Barman, S.A.; Sattar, N.; Cook, D.G.; Whincup, P.H. Retinal Arteriolar Tortuosity and Cardiovascular Risk Factors in a Multi-Ethnic Population Study of 10-Year-Old Children; the Child Heart and Health Study in England (CHASE). Arterioscler. Thromb. Vasc. Biol. 2011, 31, 1933–1938. [Google Scholar] [CrossRef] [PubMed]
  39. Budai, A.; Bock, R.; Maier, A.; Hornegger, J.; Michelson, G. Robust Vessel Segmentation in Fundus Images. Int. J. Biomed. Imaging 2013, 2013, 154860. [Google Scholar] [CrossRef] [PubMed]
  40. Hoover, A.D.; Kouznetsova, V.; Goldbaum, M. Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response. IEEE Trans. Med. Imaging 2000, 19, 203–210. [Google Scholar] [CrossRef]
  41. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar] [CrossRef]
  42. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar] [CrossRef]
  43. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Johansen, D.; Lange, T.D.; Halvorsen, P.; Johansen, H.D. ResUNet++: An Advanced Architecture for Medical Image Segmentation. In Proceedings of the 2019 IEEE International Symposium on Multimedia (ISM), San Diego, CA, USA, 9–11 December 2019; pp. 225–2255. [Google Scholar] [CrossRef]
  44. Guo, C.; Szemenyei, M.; Yi, Y.; Wang, W.; Chen, B.; Fan, C. SA-UNet: Spatial Attention U-Net for Retinal Vessel Segmentation. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 1236–1242. [Google Scholar] [CrossRef]
  45. Liu, W.; Yang, H.; Tian, T.; Cao, Z.; Pan, X.; Xu, W.; Jin, Y.; Gao, F. Full-Resolution Network and Dual-Threshold Iteration for Retinal Vessel and Coronary Angiograph Segmentation. IEEE J. Biomed. Health Inform. 2022, 26, 4623–4634. [Google Scholar] [CrossRef]
  46. Iqbal, A.; Sharif, M. PDF-UNet: A semi-supervised method for segmentation of breast tumor images using a U-shaped pyramid-dilated network. Expert Syst. Appl. 2023, 221, 119718. [Google Scholar] [CrossRef]
  47. Zhang, W.; Qu, S.; Feng, Y. LMFR-Net: Lightweight multi-scale feature refinement network for retinal vessel segmentation. Pattern Anal. Appl. 2025, 28, 44. [Google Scholar] [CrossRef]
  48. Zhang, J.; Luan, Z.; Ni, L.; Qi, L.; Gong, X. MSDANet: A multi-scale dilation attention network for medical image segmentation. Biomed. Signal Process. Control 2024, 90, 105889. [Google Scholar] [CrossRef]
Figure 1. Overall architecture of DSAE-Net.
Figure 1. Overall architecture of DSAE-Net.
Jimaging 11 00306 g001
Figure 2. Visual comparison of the weighting strategies between boundary loss and the proposed Skeleton Distance Loss (SDL). (A) Top: Original vessel mask (ground truth). Bottom: Its schematic binary representation. (B) Top: Distance map generated by applying Boundary Loss [33] to the mask in (A). Note the widespread negative weights (blue regions) within the foreground area, which is a fundamental limitation of this method. Bottom: A sample numerical matrix from this distance map, illustrating the distribution of both positive and negative weights. (C) Top: Vessel skeleton extracted from the mask in (A) via skeletonization. Bottom: Its schematic binary representation. (D) Top: SDL distance map ϕ K ( q ) (Equation (2)) generated from the skeleton in (C), exhibiting strictly non-negative values. Bottom: A sample numerical matrix from this SDL map. (E) A detailed visualization of the SDL distance transform ϕ K ( q ) .
Figure 2. Visual comparison of the weighting strategies between boundary loss and the proposed Skeleton Distance Loss (SDL). (A) Top: Original vessel mask (ground truth). Bottom: Its schematic binary representation. (B) Top: Distance map generated by applying Boundary Loss [33] to the mask in (A). Note the widespread negative weights (blue regions) within the foreground area, which is a fundamental limitation of this method. Bottom: A sample numerical matrix from this distance map, illustrating the distribution of both positive and negative weights. (C) Top: Vessel skeleton extracted from the mask in (A) via skeletonization. Bottom: Its schematic binary representation. (D) Top: SDL distance map ϕ K ( q ) (Equation (2)) generated from the skeleton in (C), exhibiting strictly non-negative values. Bottom: A sample numerical matrix from this SDL map. (E) A detailed visualization of the SDL distance transform ϕ K ( q ) .
Jimaging 11 00306 g002
Figure 3. Structures of various convolution blocks: (A) U-Net, (B) ResNet, (C) ConvNeXt, and (D) Proposed CMFA.
Figure 3. Structures of various convolution blocks: (A) U-Net, (B) ResNet, (C) ConvNeXt, and (D) Proposed CMFA.
Jimaging 11 00306 g003
Figure 4. Detailed architecture of the CMFA block: (d) structure of convolution block, (e) overall CMFA module and (f) Contextual Transformer (CoT) sub-module.
Figure 4. Detailed architecture of the CMFA block: (d) structure of convolution block, (e) overall CMFA module and (f) Contextual Transformer (CoT) sub-module.
Jimaging 11 00306 g004
Figure 5. Architectures of the decoder of U-Net and CAE-UNet: (a) U-Net decoder; (b,d) structure and implementation of CAE-UNet’s decoder (c).
Figure 5. Architectures of the decoder of U-Net and CAE-UNet: (a) U-Net decoder; (b,d) structure and implementation of CAE-UNet’s decoder (c).
Jimaging 11 00306 g005
Figure 6. Bubble chart visualization of data from Table 2.
Figure 6. Bubble chart visualization of data from Table 2.
Jimaging 11 00306 g006
Figure 7. Segmentation visualizations: DSAE-Net versus state-of-the-art models on DRIVE (Row 1), CHASE-DB (Row 2), HRF (Row 3), and STARE (Row 4). Color scheme: true vessels (red), missed vessels (blue), and background misclassification (green).
Figure 7. Segmentation visualizations: DSAE-Net versus state-of-the-art models on DRIVE (Row 1), CHASE-DB (Row 2), HRF (Row 3), and STARE (Row 4). Color scheme: true vessels (red), missed vessels (blue), and background misclassification (green).
Jimaging 11 00306 g007
Figure 8. Training dynamics comparison on DRIVE dataset: SDL accelerates convergence and stabilizes cross-entropy optimization; SDL mitigates early-stage instability in Dice loss training.
Figure 8. Training dynamics comparison on DRIVE dataset: SDL accelerates convergence and stabilizes cross-entropy optimization; SDL mitigates early-stage instability in Dice loss training.
Jimaging 11 00306 g008
Table 1. Summary of retinal vessel segmentation datasets used in the experiments.
Table 1. Summary of retinal vessel segmentation datasets used in the experiments.
DatasetTrainValTestImage Size (Pixels)Characteristics / Purpose
DRIVE [37]15520 565 × 584 General benchmark; includes pathologies
CHASE-DB1 [38]1666 999 × 960 Images from children; demographic robustness
HRF [39]20520 3504 × 2336 High-resolution; healthy/DR/glaucoma subjects
STARE [40]1244 605 × 700 Diverse pathologies; robustness testing
Table 2. Model efficiency comparison: parameters, FLOPs, and inference time on different hardware.
Table 2. Model efficiency comparison: parameters, FLOPs, and inference time on different hardware.
ModelYearParameters (M)FLOPs (G)Time (CPU) (s) *Time (GPU) (s) *
UNet20157.7654.850.970.86
Attention UNet201834.88266.533.552.59
ResUNet++20194.0663.261.651.07
SA-UNet20200.4810.611.020.71
ConvUNeXt20223.5129.011.531.02
FR-UNet20225.72235.925.103.58
PDF-UNet202310.5244.741.130.77
ULite20230.883.030.860.55
MSDANet202455.55650.916.022.35
LMFR-Net20250.0713.971.050.70
DSAE-Net20250.084.761.000.67
* CPU: Intel(R)Core(TM)i7-9750H CPU @ 2.60 GHZ; GPU: NVIDlA GeForce GTX 1660 Ti. bold: best; underline: second-best.
Table 3. Segmentation performance (Acc, AUC, Dice, MCC, HD95) across datasets.
Table 3. Segmentation performance (Acc, AUC, Dice, MCC, HD95) across datasets.
DatasetModelAcc (%)AUC (%)Dice (%)MCC (%)HD95
DRIVEUNet *94.8796.9980.0077.0637.96
Attention UNet *95.4298.0482.3279.7420.03
ResUNet++94.9497.1280.1877.2837.75
SA-UNet *95.4497.9582.4279.8220.84
ConvUNeXt *95.4697.7682.3179.7115.63
FR-Unet95.3697.7281.7279.0626.54
PDF-UNet95.3597.8081.9479.2724.5
ULite *95.4097.4881.7779.1421.83
MSDANet95.4897.8582.4379.8427.43
LMFR-Net95.2997.6481.6978.9926.58
DSAE-Net *95.4898.0082.4479.8612.12
CHASE-DBUNet95.8997.9979.1176.8864.42
Attention UNet96.2298.4581.2279.2734.74
ResUNet++96.2598.4881.2979.3326.25
SA-UNet96.1498.5280.9978.0235.59
ConvUNeXt96.1598.4180.9078.9136.42
FR-Unet96.1598.5180.9578.9731.59
PDF-UNet96.0798.3680.6378.6241.17
ULite95.9598.1179.9177.8043.92
MSDANet96.1498.0980.8978.9040.75
LMFR-Net96.0498.4280.5678.9936.84
DSAE-Net96.2598.5681.4679.5422.65
HRFUNet96.1197.4478.7976.65170.6
Attention UNet96.3798.0180.4678.46127.14
ResUNet++95.9196.9777.4175.20225.36
SA-UNet96.2997.9480.1078.06128.25
FR-Unet96.2697.7379.7277.66140.58
ConvUNeXt96.2197.5779.4177.33132.63
PDF-UNet96.2797.7679.8177.75144.42
ULite96.0497.2478.3576.19151.5
MSDANet96.2697.7079.9277.86150.3
LMFR-Net96.3097.8480.0377.98147.2
DSAE-Net96.3197.9080.1278.10126.25
STAREUNet96.4597.8776.6075.6257.25
Attention UNet96.6497.8579.0277.5349.75
ResUNet++96.6198.1580.0178.1938.38
SA-UNet96.7098.5580.7178.9230.67
ConvUNeXt96.8098.3481.5379.7827.00
FR-Unet96.7398.4879.5678.1335.75
PDF-UNet96.4498.3280.0578.1330.27
ULite96.5497.8680.0178.1224.5
MSDANet96.1998.0779.3377.4030.33
LMFR-Net96.5297.9179.7577.8535.25
DSAE-Net96.6898.7381.9280.2527.25
* Results marked with * are obtained via 5-fold cross-validation. Mean values and standard deviations for UNet, Attention UNet, SA-UNet, ConvUNeXt, ULite, and DSAE-Net based on 5-fold cross-validation are provided in Appendix A, Table A1. bold: best; underline: second-best.
Table 4. Average performance of DSAE-Net for each of the datasets in Table 1 over 5 training runs, with a confidence interval containing the true mean with probability p = 95%, under a normality assumption for performance.
Table 4. Average performance of DSAE-Net for each of the datasets in Table 1 over 5 training runs, with a confidence interval containing the true mean with probability p = 95%, under a normality assumption for performance.
DatasetAcc (%)AUC (%)Dice (%)MCC (%)
DRIVE95.48 ± 0.0298.00 ± 0.0482.44 ± 0.1179.86 ± 0.12
CHASE-DB96.25 ± 0.0398.56 ± 0.0281.46 ± 0.2279.54 ± 0.23
HRF96.31 ± 0.0497.90 ± 0.0580.12 ± 0.0878.10 ± 0.09
STARE96.68 ± 0.0398.73 ± 0.0281.92 ± 0.1380.25 ± 0.15
Table 5. Hyperparameter ablation: T: (k, m) scaling on DRIVE.
Table 5. Hyperparameter ablation: T: (k, m) scaling on DRIVE.
MetricConfiguration: Depth k/Width m
3/4 3/8 3/16 4/8 4/16 4/32
Param (M)0.020.080.310.321.275.03
FLOPs (G)1.334.7618.576.8126.23102.93
Dice (Val) (%)81.9482.8283.0882.9583.4083.10
AUC (Val) (%)98.0398.1898.2698.1998.3698.30
Dice (Test) (%)81.7782.4482.3182.4182.6881.98
AUC (Test) (%)97.7198.0097.7698.0198.0497.74
Note: Dice/AUC reported in percentage points for cross-metric consistency.
Table 6. Sensitivity analysis of the balancing factor α in the combined loss function on the DRIVE dataset.
Table 6. Sensitivity analysis of the balancing factor α in the combined loss function on the DRIVE dataset.
MetricValue of α
0.0 0.3 0.7 1.0
Dice (Val) (%)82.7782.8182.8282.8
HD95 (Val)21.017.710.59.0
Dice (Test) (%)82.4082.4182.4482.44
HD95 (Test)23.0520.4512.6815.35
Table 7. Ablation study: module contributions on DRIVE and CHASEDB1 datasets.
Table 7. Ablation study: module contributions on DRIVE and CHASEDB1 datasets.
DatasetModelParam (M)AccAUCDiceMCC
DRIVEw/o CAG & CMFA0.0694.9997.1181.2078.26
w/o CAG0.0695.1797.4981.2278.46
w/o CMFA0.0895.4697.9582.2079.59
Full0.0895.4898.0082.4479.86
CHASE-DBw/o CAG & CMFA0.0696.0298.1279.2477.01
w/o CAG0.0696.2298.5181.1679.18
w/o CMFA0.0896.2498.6981.4479.52
Full0.0896.2598.5681.4679.54
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, Z.; Liu, J.; Huang, X.; Luo, K.; Yuan, X.; Zhu, Y. Efficient Retinal Vessel Segmentation with 78K Parameters. J. Imaging 2025, 11, 306. https://doi.org/10.3390/jimaging11090306

AMA Style

Zeng Z, Liu J, Huang X, Luo K, Yuan X, Zhu Y. Efficient Retinal Vessel Segmentation with 78K Parameters. Journal of Imaging. 2025; 11(9):306. https://doi.org/10.3390/jimaging11090306

Chicago/Turabian Style

Zeng, Zhigao, Jiakai Liu, Xianming Huang, Kaixi Luo, Xinpan Yuan, and Yanhui Zhu. 2025. "Efficient Retinal Vessel Segmentation with 78K Parameters" Journal of Imaging 11, no. 9: 306. https://doi.org/10.3390/jimaging11090306

APA Style

Zeng, Z., Liu, J., Huang, X., Luo, K., Yuan, X., & Zhu, Y. (2025). Efficient Retinal Vessel Segmentation with 78K Parameters. Journal of Imaging, 11(9), 306. https://doi.org/10.3390/jimaging11090306

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop