Next Article in Journal
Ultrasound-Based Deep Learning Radiomics Models for Predicting Primary and Secondary Salivary Gland Malignancies: A Multicenter Retrospective Study
Previous Article in Journal
Pathomechanics of Early-Stage Lumbar Intervertebral Disc Degradation Leading to Discogenic Pain—A Narrative Review
Previous Article in Special Issue
Exploring Feature Selection with Deep Learning for Kidney Tissue Microarray Classification Using Infrared Spectral Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SwinDAF3D: Pyramid Swin Transformers with Deep Attentive Features for Automated Finger Joint Segmentation in 3D Ultrasound Images for Rheumatoid Arthritis Assessment

1
GE HealthCare Technology & Innovation Center, Niskayuna, NY 12309, USA
2
Department of Biomedical Engineering, University of Michigan, Ann Arbor, MI 48109, USA
3
GE Vernova Research, Niskayuna, NY 12309, USA
4
Division of Rheumatology, Department of Internal Medicine, University of Michigan, Ann Arbor, MI 48109, USA
5
Department of Radiology, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
*
Author to whom correspondence should be addressed.
Bioengineering 2025, 12(4), 390; https://doi.org/10.3390/bioengineering12040390
Submission received: 7 February 2025 / Revised: 1 April 2025 / Accepted: 3 April 2025 / Published: 5 April 2025

Abstract

:
Rheumatoid arthritis (RA) is a chronic autoimmune disease that can cause severe joint damage and functional impairment. Ultrasound imaging has shown promise in providing real-time assessment of synovium inflammation associated with the early stages of RA. Accurate segmentation of the synovium region and quantification of inflammation-specific imaging biomarkers are crucial for assessing and grading RA. However, automatic segmentation of the synovium in 3D ultrasound is challenging due to ambiguous boundaries, variability in synovium shape, and inhomogeneous intensity distribution. In this work, we introduce a novel network architecture, Swin Transformers with Deep Attentive Features for 3D segmentation (SwinDAF3D), which integrates Swin Transformers into a Deep Attentive Features framework. The developed architecture leverages the hierarchical structure and shifted windows of Swin Transformers to capture rich, multi-scale and attentive contextual information, improving the modeling of long-range dependencies and spatial hierarchies in 3D ultrasound images. In a six-fold cross-validation study with 3D ultrasound images of RA patients’ finger joints (n = 72), our SwinDAF3D model achieved the highest performance with a Dice Score (DSC) of 0.838 ± 0.013, an Intersection over Union (IoU) of 0.719 ± 0.019, and Surface Dice Score (SDSC) of 0.852 ± 0.020, compared to 3D UNet (DSC: 0.742 ± 0.025; IoU: 0.589 ± 0.031; SDSC: 0.661 ± 0.029), DAF3D (DSC: 0.813 ± 0.017; IoU: 0.689 ± 0.022; SDSC: 0.817 ± 0.013), Swin UNETR (DSC: 0.808 ± 0.025; IoU: 0.678 ± 0.032; SDSC: 0.822 ± 0.039), UNETR++ (DSC: 0.810 ± 0.014; IoU: 0.684 ± 0.018; SDSC: 0.829 ± 0.027) and TransUNet (DSC: 0.818 ± 0.013; IoU: 0.692 ± 0.017; SDSC: 0.815 ± 0.016) models. This ablation study demonstrates the effectiveness of combining a Swin Transformers feature pyramid with a deep attention mechanism, improving the segmentation accuracy of the synovium in 3D ultrasound. This advancement shows great promise in enabling more efficient and standardized RA screening using ultrasound imaging.

1. Introduction

Rheumatoid arthritis (RA) is an autoimmune condition, in which the immune system targets and damages the synovium which lines the joints. This immune reaction results in inflammation of the affected tissue, which may result in joint deformities, bone erosion, and functional impairment. RA impacts approximately 0.5–1% of the global population [1,2]. Early and accurate diagnosis of RA is essential for developing effective treatment strategies and mitigating complications [3,4].
Early RA can be identified using medical imaging modalities, including magnetic resonance imaging (MRI) and ultrasound (US) [5]. MRI has been shown to be capable of accurately detecting joint inflammation, compared with histological examination of synovial tissue [6,7]. However, MRI is limited by its availability, lengthy scan times, and cost [8]. US imaging shows potential in overcoming such limitations due to its capability for real-time visualization, affordability, portability and widespread accessibility. Power Doppler (PD) is commonly integrated with standard B-mode US imaging to map blood flow patterns in the affected joints and assess the degree of inflammation. However, US imaging involves limitations, including user dependency and variability among observers, posing challenges when it comes to establishing consistent diagnostic standards [8]. Recently, photoacoustic (PA) imaging has emerged as a promising modality for detecting RA. By integrating optical contrast with ultrasonic resolution, PA imaging exhibits high sensitivity to hemoglobin, which is the key biomarker in RA. This makes PA imaging highly appropriate for RA detection and has demonstrated potential in enhancing rheumatology diagnostics [9,10,11,12].
A critical part of assessing RA in US imaging is the process of isolating synovial tissue from B-mode images, which allows for the analysis of PA or PD blood flow measurements in the respective regions. The main challenges in synovium segmentation from US images include the complexity of interpreting the varying shades of gray in the images, the difficulty of accurately detecting bones due to noise and artifacts, and the need for precise localization of skin borders and joint regions [13]. These issues lead to false positives and inaccuracies in synovitis region identification, making the automation of this process particularly challenging. Manual segmentation heavily relies on the operator and can introduce high inter-observer variability.
To circumvent this shortcoming, automated segmentation approaches have been developed. A technique that uses confidence maps to automatically delineate the surface of the bone in hand joints in US scans has been previously proposed in [14]. Another approach for automated detection of multiple anatomical regions, including skin, bones and joints for RA grading in US, has been presented in [13]. Moreover, an approach for synovial tissue segmentation based on the active contour algorithm, combined with identification of skin border, bones, and joint center coordinates has been developed [15]. These existing methods focus on 2D US images, which present several challenges. Firstly, it requires more operator time and skill to position the probe in the optimal 2D plane for diagnostic-quality images of the finger joint. Furthermore, the PA signal may not be strong enough in the selected 2D plane, reducing the effectiveness of RA detection. Lastly, there can be high variability in the 2D plane selection due to the operator’s experience level. Therefore, 3D US images are more feasible, as they do not require the operator to have a high skill or experience level, and measurement variability can also be reduced. However, automatic segmentation of the synovium in 3D US remains challenging due to ambiguous boundaries, variability in synovium shape, and inhomogeneous intensity distribution (see Figure 1).
Recent advancements in deep learning (DL) have demonstrated significant promise in US image segmentation, markedly improving the accuracy and efficiency of identifying anatomical structures. DL algorithms leverage neural networks that can learn highly complex patterns of multiple structures, effectively addressing the specific challenges associated with US imaging, including inconsistencies in image quality and the presence of noise [16]. These algorithms have been successfully applied to various organs and anatomies, including the heart [17], lungs [18], liver [19], breast [20], blood vessels [21], kidneys [22], bones [23], nerves [24], and teeth [25], demonstrating great potential in diagnosing pathologies such as atherosclerosis, heart disease, liver fibrosis, kidney disease, RA, and various types of cancer.
Convolutional neural networks (CNNs) are highly effective DL models that have been widely utilized in the field of medical image segmentation. However, traditional CNN-based architectures, such as 3D UNet [26], tend to capture localized spatial information but may not effectively capture long-range dependencies and multi-scale contextual information, which are crucial for accurate 3D segmentation in US images. Recently, Swin Transformers [27,28] have been introduced as hierarchical vision transformers. Unlike conventional Vision Transformers (ViTs) [29], Swin Transformers partition images into non-overlapping patches and employ a hierarchical structure, facilitating efficient multi-scale feature extraction. At each hierarchical stage, Swin Transformer blocks process these patches at varying resolutions, effectively capturing both local and global contextual information. Moreover, the integration of a shifted window partitioning scheme significantly enhances the model’s capability to encode spatial dependencies by alternating the attention windows across layers, thus allowing interactions between adjacent non-overlapping windows [27,30]. These advancements have made Swin Transformers effective and adaptable for a wide range of downstream tasks, including medical image segmentation. Furthermore, Swin Transformers have demonstrated superior adaptability across different modalities of medical imaging [31,32,33,34]. Their hierarchical nature allows them to maintain computational efficiency while scaling to larger image resolutions, which is particularly beneficial for 3D US images where fine details and broader context must be simultaneously considered.
In this work, we introduce a novel network architecture, Swin Transformers with Deep Attentive Features for 3D segmentation (SwinDAF3D). This model builds upon the DAF3D architecture proposed by Wang et al. [35], by replacing the ResNeXt [36] backbone with a Swin Transformers Feature Pyramid Network (Swin FPN). The key novelties and contributions of our SwinDAF3D network include the following:
  • Integration of Feature Pyramid with Swin Transformers: SwinDAF3D integrates hierarchical Swin Transformers enhanced by a shifted window mechanism into a Feature Pyramid, outperforming other transformer-based models such as Swin UNETR [31], UNETR++ [37], TransUNet [38].
  • Utilization of Deep Attention Mechanism: SwinDAF3D uniquely integrates deep attention mechanisms with Swin FPN, allowing the model to focus on relevant features while suppressing irrelevant ones, thereby refining the overall feature extraction process in 3D US images.
  • Clinical Impact for RA Assessment: SwinDAF3D improves automated synovium segmentation accuracy in 3D US images compared to baseline models, a critical advancement for reliable RA assessment and monitoring.
Furthermore, an ablation study is conducted to demonstrate that the developed SwinDAF3D model effectively leverages the combined strengths of Swin FPN and deep attention mechanisms. This study compares the segmentation performance of SwinDAF3D against five baseline models: 3D UNet [26], DAF3D [35], Swin UNETR [31], UNETR++ [37], and TransUNet [38], demonstrating improved performance in synovium segmentation for 3D US images. These advancements show great promise in enhancing diagnostic accuracy and improving clinical outcomes for patients with RA.

2. Materials and Methods

2.1. Data Acquisition

A total of n = 19 patients with active synovitis were recruited for this study. Patient screening and recruitment was carried out in collaboration with the Division of Radiology at the University of Michigan Hospital. Informed consent was obtained from all participants, and the study was approved by the University of Michigan’s institutional review board (IRB) (protocol no. HUM00003693). The patients’ finger joints were scanned using a GEHC VE95 US unit with an L8-18i-D high-frequency linear probe (GE HealthCare, Chicago, IL, USA). Real-time 3D US imaging was performed using a robotic hand scanning system, which can provide simultaneous and co-localized US and PA imaging, as described by Peng et al. [39]. The robotic hand captured multiple 2D B-mode US images in a stop–move–scan manner using a step size of 0.2 mm. The acquired 2D data were subsequently stacked together in order to obtain 3D US images. A total of n = 102 3D US images of the patients’ finger joints were acquired, including metacarpophalangeal joints (MCPs), proximal interphalangeal joints (PIPs) and distal interphalangeal joints (DIPs), and each of them was interpolated on a Cartesian grid of 1083 × 277 × 116 samples, with spacings of 0.0277 mm, 0.0902 mm, and 0.1487 mm along the axial, lateral, and elevation directions, respectively. Further details on the acquisition setup are available in [39]. The synovial region was annotated in random 2D slices across the elevation direction of the 3D US images by manually delineating the hypoechoic area above the finger bones using ITK-SNAP software (version 4.0.2) [40].

2.2. Data Pre-Processing

To generate dense 3D synovium annotations from sparse manual annotations, we utilized our previously developed DL-based semi-automated sparse-to-dense annotation model [41]. This approach involved training a 2D UNet++ [42] model for slice-level segmentation and interpolating the expert 2D sparse annotations from 2D US slices. This sparse-to-dense annotation generation methodology has been qualitatively and quantitatively evaluated, and it demonstrated the capability to provide accurate dense 3D annotations, enhancing the continuity and consistency of the segmentation across all dimensions. Figure 2 shows an example of the 3D sparse-to-dense annotation, demonstrating that the dense 3D masks generated from the sparse annotations maintained consistent and continuous shapes. To further evaluate the reliability of the 3D annotations, an expert radiologist reviewed the generated masks across each consecutive slice within each US 3D volume and corroborated that the marked regions accurately depict the synovium region. The 3D US volumes were centrally cropped in a sagittal view to focus on the finger joint region, and then resized to 256 × 256 × 64 . Due to the limited availability of annotated 3D US images, data augmentation was crucial. To artificially enhance the size and diversity of training dataset, various offline data augmentation techniques from TorchIO [43], a Python library designed for medical imaging augmentation, were applied on the 3D US images and the corresponding dense 3D annotations. These techniques included horizontal/vertical flip, random rotation, random brightness/contrast adjustment, random Gaussian blur, and random noise.

2.3. Network Architectures

In this section, we explore and compare six network architectures for synovium segmentation in 3D US images of finger joints: 3D UNet, which extends 2D capabilities to handle volumetric data; DAF3D, which enhances accuracy with deep attention mechanisms; Swin UNETR, which integrates UNet with Swin Transformers to effectively capture local and long-range dependencies; UNETR++, which refines feature representations by incorporating efficient paired attention modules in both the encoder and decoder; TransUNet, which incorporates transformer-based encoders into the UNet architecture to capture global context; and our proposed SwinDAF3D, which combines Swin FPN with a deep attention mechanism for optimized feature extraction in synovium segmentation task.

2.3.1. 3D UNet

The implemented 3D UNet [26] model, depicted in Figure 3, consists of two main components: a downsampling path (encoder) and an upsampling path (decoder), both conducting operations in 3D. The downsampling path includes four convolutional encoder blocks, each comprising two sets of 3D convolution layers with a kernel size of 3 × 3 × 3 . Following each convolution, there is batch normalization and a rectified linear unit (ReLU) [44] activation function. Downsampling is achieved through 3D max pooling layers with a window size of 2 × 2 × 2 and strides of two in each dimension. To prevent bottlenecks and ensure an sufficient number of feature channels, the channel count is doubled before each downsampling step. In the upsampling path, there are also four convolutional decoder blocks. Each layer in this path starts with a 3D transposed convolution using a 2 × 2 × 2 kernel and strides of two in each dimension. Subsequently, two 3D convolutional layers with 3 × 3 × 3 kernels are applied, each followed by a ReLU activation. Skip connections are employed to link corresponding layers from the downsampling path to the upsampling path.
The 3D UNet’s relatively simple architecture has limitations when it comes to capturing complex and ambiguous structures in 3D US images of finger joints. Additionally, its convolutional operations, which excel at capturing localized 3D information, may not effectively capture global, long-range spatial relationships.

2.3.2. DAF3D

The DAF3D [35] network architecture is illustrated in Figure 4. Originally proposed for the 3D transrectal US prostate segmentation task, it is built around a four-level 3D feature pyramid network (FPN) with a ResNeXt [36] backbone enhanced with dilated convolutions to capture multi-scale contextual information. This architecture incorporates Multi-Layer Features (MLFs) and an innovative deep attention mechanism, which computes attention weights for these MLF, refining Single-Layer Features (SLF) through element-wise multiplication.
The attention module consists of three convolutional layers with group normalization and parametric rectified linear unit (PReLU) [45] activation, which collectively produce an attention map that enhances feature representation. Additionally, the 3D Atrous Spatial Pyramid Pooling (ASPP) [46] module further processes these features to address scale variability in synovium shapes.
The primary advantage of DAF3D lies in its ability to integrate multi-level features via attention mechanisms, enhancing both detailed and semantic features at various network depths. This results in more accurate and robust segmentation, effectively handling challenges such as ambiguous boundaries and inhomogeneous intensity distributions, which are common in 3D US imaging.

2.3.3. Swin UNETR

Figure 5 illustrates the network architecture of the Swin UNETR model [31], a hybrid design that integrates Swin Transformers as the encoder and a fully convolutional network as the decoder, similar to the 3D UNet [26]. It leverages the capabilities of transformers in capturing long-range dependencies and the efficiency of CNNs in extracting local features. The input to Swin UNETR is a 3D US image of finger joint X R H × W × D , where H , W , D are spatial dimensions of 3D US image. The input image is divided into non-overlapping patches of size 2 × 2 × 2 , and each patch is flattened and projected into an embedding space of dimension C, with C set to 48 in the implementation. The encoder comprises four stages, each containing multiple Swin Transformer blocks that compute self-attention within local windows with size 7 × 7 × 7 . Windows are also shifted between layers to enable cross-window connections. Within the Swin Transformer blocks, the output of the subsequent layers l and l + 1 (denoted as z l and z l + 1 ) are calculated as follows:
z ^ l = W - MSA ( LN ( z l 1 ) ) + z l 1 z l = MLP ( LN ( z ^ l ) ) + z ^ l z ^ l + 1 = SW - MSA ( LN ( z l ) ) + z l z l + 1 = MLP ( LN ( z ^ l + 1 ) ) + z ^ l + 1 .
Here, W-MSA and SW-MSA denote window-based multi-head self-attention using regular and shifted window partitioning configurations, respectively; z ^ l and z ^ l + 1 denote the outputs of W-MSA and SW-MSA; MLP and LN denote Multi-Layer Perceptron and layer normalization, respectively. Finally, the self-attention is computed as follows:
Attention ( Q , K , V ) = Softmax Q K d V .
In which Q , K , V are the queries, keys, and values, respectively, and d represents the dimension of the queries and keys [47].
At the end of each level, a patch merging layer reduces the spatial dimensions of the feature maps by a factor of 2 while increasing the channel dimension. Features extracted at different resolutions from the encoder are passed to the decoder via skip connections. The decoder comprises five convolutional residual blocks, each consisting of two 3 × 3 × 3 convolutional layers with instance normalization. These blocks refine the feature representations and generate the synovium segmentation.
The Swin UNETR architecture effectively combines the hierarchical feature extraction capabilities of Swin Transformers with the efficient local feature processing of convolutional networks, capturing both global and local context comprehensively.

2.3.4. UNETR++

The network architecture of the UNETR++ model is depicted in Figure 6. UNETR++ [37] builds on the original UNETR [48] framework by integrating a novel efficient paired-attention (EPA) block that simultaneously captures spatial and channel dependencies, thereby enhancing segmentation accuracy while reducing computational overhead [37]. The model employs an encoder–decoder architecture in which EPA blocks are embedded at every stage of both the encoder and the decoder. The key innovation of UNETR++ lies in its dual-branch EPA module, which comprises two complementary attention mechanisms with shared query and key mappings. This design enables the network to focus on both the spatial layout and the channel-wise relationships of the input data.
In the spatial branch, the EPA block reduces the complexity of traditional self-attention from quadratic to linear by projecting the keys and values into a lower-dimensional space. For a given input feature map, shared queries ( Q s h a r e d ) and keys are computed, and the keys are further projected to a lower-dimensional space ( K p r o j ). The spatial attention map is then obtained via a weighted sum of the projected values ( V ˜ spatial ), as defined by [37]:
X ^ s = Softmax Q s h a r e d K p r o j d · V ˜ s p a t i a l ,
where d is the dimensionality of the queries and keys.
Concurrently, the channel branch addresses the interdependencies between feature channels. Using the same shared queries and keys, the channel attention is computed by creating a dot product in the channel dimension, which emphasizes the correlations among different feature maps [37]:
X ^ c = V c h a n n e l · Softmax Q s h a r e d K s h a r e d d ,
with V c h a n n e l representing the channel-specific value projection.
The outputs of the spatial and channel attention modules are subsequently fused through a series of convolutional layers to generate an enriched feature representation. This fusion not only integrates complementary information from both branches but also refines the feature maps for improved segmentation performance.
Overall, the UNETR++—with its EPA attention mechanism and hierarchical structure—demonstrates significant potential for improving volumetric segmentation tasks by learning enriched inter-dependent spatial and channel features.

2.3.5. TransUNet

The network architecture of the TransUNet model is depicted in Figure 7. We implemented a 3D variant of the original TransUNet architecture proposed by Chen et al. [38] for medical image segmentation tasks. The TransUNet architecture integrates the strengths of the UNet’s ability to capture detailed local features with the Transformer’s self-attention mechanism for global context modeling. The architecture consists of three primary components: an encoder, a Transformer-based bottleneck, and a decoder enhanced with skip connections.
The model initially employs an encoder path structured similarly to a traditional 3D UNet, leveraging successive convolutional blocks and downsampling operations to extract hierarchical, high-dimensional features from volumetric input data. Each convolutional block consists of two 3D convolutional layers, each followed by instance normalization and ReLU activation. At the bottleneck, CNN-extracted features are reshaped into sequential token representations for processing by a 12-layer Transformer encoder. Positional embeddings are incorporated to preserve spatial information.
A series of Transformer encoder layers then apply multi-head self-attention mechanisms, enabling the model to capture long-range dependencies and global contextual information beyond the local receptive fields of conventional CNN-based encoders. The decoder adopts a CNN-based approach, using transposed convolutions for upsampling. Skip connections from corresponding encoder layers are concatenated at each decoding stage to integrate high-resolution features. Following each concatenation, convolutional blocks consisting of two sequential 3D convolutions, instance normalization, and ReLU activation further refine the fused features.
By effectively capturing long-range dependencies and global context, TransUNet can enhance segmentation accuracy. Its adaptive Transformer-based architecture offers flexibility across diverse medical segmentation tasks.

2.3.6. SwinDAF3D

Figure 8 shows the network architecture of our proposed SwinDAF3D model, which integrates the advantages of Swin UNETR [31] and DAF3D [35], aiming to leverage the strengths of both models for improved performance. The same Swin Transformer encoder from Swin UNETR is utilized to construct a four-level Swin FPN, capturing hierarchical multi-scale contextual information from 3D US images. Multi-scale features from different levels of the Swin FPN are extracted and concatenated to form the MLF.
Attention weights are then computed for the MLF to enhance relevant features and suppress irrelevant ones. The attention weight computation is defined as follows:
A i = σ ( f a ( ( F i , F M L F ) ; θ ) ) ,
where A i represents the computed attention weights at level i, F i denotes the SLF extracted from level i of Swin FPN, F M L F denotes the MLF, θ represents the parameters learned by f a , which contains three 3D convolutional layers, and σ is the Sigmoid activation function. The weighted MLF denoted as F M L F is computed by element-wise multiplication of F M L F and A i :
F M L F = A i F M L F .
Finally, the attentive features are computed by merging F M L F with corresponding F i by applying two 3 × 3 × 3 and one 1 × 1 × 1 convolutional layers. This process refines the layer-wise SLF and produces the final attentive features for the given Swin FPN level. The computed attentive features are then concatenated and go through 3D multi-layer ASPP and convolution blocks to generate the synovium segmentation.
SwinDAF3D combines Swin UNETR’s hierarchical and long-range dependency modeling capabilities with DAF3D’s feature refinement strengths via a deep attention mechanism. This integration provides an effective approach to capturing and enhancing multi-scale US features.

2.4. Design of Ablation Study

To rigorously assess the contribution of each architectural component in our proposed SwinDAF3D model, we conducted a comprehensive ablation study comparing it against five established baseline models (summarized in Table 1). Each baseline was chosen to isolate and evaluate the impact of the backbone type, attention mechanism, and the level at which transformer components are integrated. The objectives of each model comparison are described below:
  • 3D UNet: Serves as a standard baseline CNN model for evaluating segmentation performance.
  • DAF3D: Assesses the performance gains achieved by employing a hierarchical Swin Transformers backbone as opposed to a conventional CNN-based backbone like ResNeXt.
  • Swin UNETR: Assesses the incremental benefits obtained by integrating deep attention mechanisms into the hierarchical Swin Transformers backbone.
  • UNETR++: Compares the relative effectiveness of deep attention mechanisms against efficient paired attention (EPA) modules within a hierarchical transformer framework.
  • TransUNet: Assesses the advantages of fully integrating hierarchical Swin Transformers and deep attention mechanisms compared to architectures that use transformers only at the bottleneck stage.
In this study, we performed six-fold cross-validation with a dataset of clinically confirmed cases of RA (n = 72). Additionally, we tested the models on an independent test set, excluded from the training and validation sets, with clinically confirmed cases of RA (n = 30).

2.5. Model Training

For all the networks, including 3D UNet, DAF3D, Swin UNETR, UNETR++, TransUNet, and proposed SwinDAF3D, Dice loss and binary cross-entropy (BCE) loss are computed at the output of each network. The Dice loss is defined as follows:
L D i c e = 1 2 i = 1 N P i G i i = 1 N P i 2 + i = 1 N G i 2
where N is the number of voxels in the input 3D US volumes, P i is the predicted probability for voxel i, and G i is the ground truth label for voxel i. The binary cross-entropy loss is defined as follows:
L B C E = i = 1 N G i log ( P i ) + ( 1 G i ) log ( 1 P i )
For the 3D UNet, Swin UNETR, UNETR++, and TransUNet networks, which do not have deep attention mechanisms, the total loss denoted as L o u t is simply the sum of the Dice loss and the BCE loss at the output level of each network:
L o u t = L o u t , D i c e + L o u t , B C E
For DAF3D and SwinDA3D networks, which incorporate deep attention mechanisms, additional Dice loss and BCE loss are also computed for each SLF and output of each attention module:
L s l f = L s l f , D i c e + L s l f , B C E L a t t n = L a t t n , D i c e + L a t t n , B C E
The total loss, denoted as L t o t a l , for both DAF3D and the SwinDAF3D network is the sum of the loss across all layers, including the refined layers after applying the attention mechanisms and the output layer. It is defined as follows:
L total = i = 1 N w i L i , s l f + j = 1 M w j L j , a t t n + L o u t
Here, N = 4 and M = 4 represent the number of layers before and after applying the attention modules, respectively. w i and w j are the weights assigned to each layer’s loss. We use the same weighting as defined in [35], which weights ( w i = 1, 2, 3, 4, w j = 1, 2, 3, 4) empirically as (0.4, 0.5, 0.7, 0.8, 0.4, 0.5, 0.7, 0.8).
All networks in this work were implemented based on PyTorch [49] and MONAI [50], and they were trained on a single NVIDIA Tesla V100 GPU. We used the loss function defined in Equation (9) for 3D UNET, Swin UNETR, UNETR++, and TranUNet, and the loss function defined in Equation (11) for DAF3D and SwinDAF3D. We adopted the Adam optimizer [51] with a learning rate of 0.0001 and a weight decay of 0.0001. All models were trained for 50 epochs with a batch size of one. To enhance the robustness and generalizability of the model and to artificially expand the training dataset’s size and diversity, we applied extensive data augmentation strategies on the 3D US images, including random flip, random rotation, random brightness/contrast, random Gaussian blur, and random noise using the TorchIO library [43]. These transforms are intended to simulate different positioning and angles of the ultrasound probe, as well as varying settings of dynamic range, gain, and transmission power. Regularization was implicitly managed through weight decay, and performance was monitored using both BCE loss and Dice loss metrics. The best-performing model was selected based on the highest Dice coefficient achieved on the validation set.

2.6. Model Performance Assessment

The performance metrics for evaluating all networks include the Dice score (DSC), intersection over union (IoU), and the Surface Dice Score (SDSC). The Dice score measures the similarity between the predicted segmentation and the ground truth. It quantifies similarity as follows:
D S C = 2 × | A B | | A | + | B |
where A represents the set of predicted voxels and B represents the set of ground truth voxels. A higher Dice score indicates better overlap between the predicted and ground truth segmentation. IoU measures the ratio of the intersection of the predicted voxels and the ground truth to their union, and it is calculated as follows:
I o U = | A B | | A B |
IoU gives a straightforward measure of how much the predicted and actual segmentation overlap, with values closer to 1 indicating better segmentation accuracy.
In addition to volumetric-based metrics, the SDSC evaluates segmentation performance at the surface level, which is particularly important for assessing boundary accuracy. The surface of both the predicted segmentation S p and the ground truth segmentation S g is extracted using morphological operations such as binary erosion. The SDSC is then computed as follows:
S D S C = 2   ×   | S p S g | | S p | + | S g |
where | S p S g | denotes the number of surface voxels from S p that lie within a defined distance d of S g , and vice versa. In this formulation, the tolerance distance is set to d = 1.
In the cross-validation study, a Wilcoxon signed-rank test with Bonferroni correction was performed to evaluate the differences in DSC, IoU and SDSC values across the six validation folds for each model, determining whether these differences were statistically significant. The Wilcoxon signed-rank test is a non-parametric method that compares paired samples, while the Bonferroni correction adjusts the p-values to account for multiple comparisons.

3. Results

Figure 9 illustrates the training and validation Dice score accuracy curves of a single fold over 50 epochs for all networks. The results indicate that the 3D UNet has the lowest Dice score values for both training and validation, highlighting the limitation of CNNs in capturing global, long-range spatial relationships in complex 3D US images of finger joints. DAF3D, Swin UNETR, UNETR++, and TransUNet exhibit similar Dice scores, but the curve for DAF3D is more stable. This stability suggests that the dataset size may be a limitation for Transformer-based models, which typically require larger datasets for optimal training. It also indicates that the use of deep attention in DAF3D aids feature refinement, improving the model’s robustness. Our proposed SwinDAF3D network achieved the highest Dice scores for both training and validation, demonstrating the effectiveness of combining Swin FPN with a deep attention mechanism for capturing attentive hierarchical multi-scale contextual information.
In the feature map visualizations presented in Figure 10, we demonstrate the efficacy of integrating Swin FPN with a deep attention mechanism for feature extraction. Figure 10A illustrates the DA3D SLFs across levels 1 to 4, both before and after applying attention modules, culminating in the final segmentation output. Similarly, Figure 10B displays the SwinDAF3D SLFs from levels 1 to 4, also before and after applying attention modules, along with the final segmentation output. The results clearly indicate that Swin FPN significantly enhances feature representation across all levels comparing to ResNeXt FPN, providing a more detailed and granular feature set. Furthermore, the attention modules are effective in refining these features, leading to improved segmentation accuracy. This demonstrates the robustness of the combined Swin FPN and attention module approach in extracting and refining hierarchical features for better segmentation performance.
Performance metrics for the six-fold cross-validation (n = 72) and independent test (n = 30) of all networks are summarized in Table 2. We used the same Swin Transformer encoder configuration as proposed in SwinUNETR [31], with a feature size of 48 and window size of 7 × 7 × 7. During cross-validation, the highest evaluation DSC and IoU were selected for each fold. For each model type, the fold with the highest DSC and IoU was used to evaluate the independent test set. Overall, the proposed SwinDAF3D model achieved the highest performance, with a DSC of 0.838 ± 0.013, an IoU of 0.719 ± 0.019, and a SDSC of 0.852 ± 0.020 for cross-validation, and a DSC of 0.825, an IoU of 0.692, and a SDSC of 0.832 for the independent test set. The SwinDAF3D outperforms the 3D UNet (p < 0.03, n = 6), DAF3D (p < 0.05, n = 6), Swin UNETR (p < 0.05, n = 6), UNETR++ (p < 0.05, n = 6), and TransUNet (p < 0.05, n = 6). Additionally, we conducted a post hoc power analysis based on the DSC metrics of the SwinDAF3D model. Specifically, the effect size was estimated from the observed differences in DSC between SwinDAF3D and the baseline model with highest mean DSC (TransUNet). Using the Wilcoxon signed-rank test at an alpha level of 0.05, the analysis indicated that our cross-validation study achieved a power of approximately 85% in detecting the reported significant improvements.
In addition to the ablation studies comparing different network architectures, we conducted a parameter sensitivity analysis for the proposed SwinDAF3D model. As summarized in Table 3, varying the feature size while keeping the window size constant at seven revealed that the model’s performance improves with larger feature sizes, reaching an optimal performance at a feature size of 48 (DSC: 0.825, IoU: 0.692, SDSC: 0.832), with marginal gains observed at a feature size of 60. Similarly, experiments varying the local window size with a fixed feature size of 48 indicated that a window size of seven achieves the best performance, compared to smaller (three and five) or larger (nine) window sizes. These results demonstrate the robustness of the SwinDAF3D architecture and validate the selection of a feature size of 48 and a window size of seven as the optimal configuration for synovium segmentation in 3D US images.
Figure 11 presents example segmentation results of the 3D UNet, Swin UNETR, DAF3D, TransUNet, UNETR++, and SwinDAF3D models compared to their corresponding ground truth annotations. The 3D UNet model demonstrates the poorest performance in synovium segmentation, often failing to capture the synovium accurately. DAF3D, Swin UNETR, UNETR++, and TransUNet models show improved segmentation but exhibit some missing regions and holes along the synovium boundaries and within the synovium itself. In contrast, the SwinDAF3D model produces more detailed and complete segmentations, accurately representing the granular and comprehensive shape of the synovium.
All computation performance metrics were measured on a NVIDIA Tesla V100 GPU. Table 4 compares the computational cost (GFLOPs), memory usage (MB), and inference time (s) of the models evaluated in the ablation study, including 3D UNet, DAF3D, Swin UNETR, UNETR++, TransUNet, and SwinDAF3D. The UNETR++ model achieves the lowest computational burden, requiring only 262 GFLOPs, using 1165 MB of memory, and achieving an inference time of 0.085 s—substantially outperforming the simple CNN-based 3D UNet model, which operates at 1006 GFLOPs, 2536 MB, and 0.175 s. In comparison, DAF3D and TransUNet present moderate computational demands with 1263 and 1182 GFLOPs, memory usage of 1958 MB and 4125 MB, and inference times of 0.201 s and 0.189 s, respectively. Swin transformer-based models, particularly Swin UNETR and SwinDAF3D, incur higher computational cost (1522 GFLOPs and 1730 GFLOPs), increased memory consumption (7227 MB and 7597 MB), and longer inference times (0.343 s and 0.375 s), reflecting their increased complexity due to hierarchical Swin Transformers architecture.

4. Discussion

In this work, we presented the novel SwinDAF3D architecture, which demonstrated the feasibility to provide highly accurate, automatic segmentation of the synovium region in 3D US images, showing promise in establishing standardized clinical workflows for RA diagnosis and assessment. Our comparative analysis showed that the proposed architecture provided more accurate segmentation in terms of DSC and IoU, and SDSC as compared to baseline models: 3D UNet, DAF3D, Swin UNETR, UNETR++, and TransUNet. The improved performance demonstrated that the innovative integration of Swin FPN with a deep attention mechanism, enhances the model’s capability to capture rich, multi-scale contextual information and long-range dependencies.
The results from our six-fold cross-validation and independent test set evaluations indicate significant performance variations across these models, providing insights into their respective strengths and limitations. The 3D UNet consistently demonstrated the lowest DSC, IoU, and SDSC values, highlighting its limitations in accurately capturing the complex and intricate structures of the synovium in 3D US images. This observation is consistent with previous studies [52,53,54], suggesting that traditional CNNs may not effectively capture global context and long-range dependencies in 3D data, leading to suboptimal segmentation outcomes.
DAF3D, Swin UNETR, UNETR++, and TransUNet models showed better performance compared to 3D UNet. The DAF3D model, which incorporates a deep attention mechanism, exhibited a more stable Dice score curve during training and validation, indicating that DAF3D effectively refines multi-layer features, enhancing its robustness and segmentation accuracy. Conversely, Swin UNETR, UNETR++, and TransUNet, despite their competitive performance, displayed sensitivity to dataset size, likely due to the requirement for larger training datasets to train transformer-based architectures.
The proposed SwinDAF3D model achieved the highest DSC, IoU and SDSC metrics, outperforming other baseline models in the study. The integration of Swin FPN with a deep attention mechanism in SwinDAF3D effectively captures hierarchical multi-scale contextual information, leading to more accurate and comprehensive segmentation of the synovium. This model’s encouraging performance demonstrates the importance of combining advanced transformer-based architectures with deep attention mechanisms to enhance segmentation performance in 3D medical imaging tasks. The segmentation results illustrated in Figure 11 further substantiate our quantitative findings. The SwinDAF3D model not only achieved the highest accuracy but also produced more granular and complete synovium segmentation. This level of detail is crucial for clinical applications where precise delineation of anatomical structures is essential.
While the results of this study are promising, several limitations should be noted. The dataset used for training and validation was relatively small, potentially rendering the model prone to overfitting and limiting its generalization capabilities. To artificially increase the dataset’s size and variability, we applied random augmentation transforms, including random rotation, brightness, contrast, and noise addition, while weight decay was used for regularization of the training process. Additionally, we performed six-fold cross-validation and testing on an independent set to enhance the robustness of the model’s performance assessment. Ongoing efforts involve acquiring larger and more diverse datasets from different operators and varying US acquisition settings to further improve the model’s ability to handle variability in real-world clinical data.
Another matter warranting further exploration is the use of semi-automated generation of dense 3D annotations. Previous studies have mostly focused on training 2D segmentation models for synovium segmentation using fully manual 2D expert annotations as ground truth [55,56,57]. Due to the challenging nature of manually annotating 3D US data, we employed the DL-based methodology previously proposed and validated in [41], which utilizes manual 2D expert annotations to generate 3D dense annotations. To further verify the reliability of the 3D annotations, an expert radiologist confirmed that the generated annotations accurately depict the synovial region. Additional validation of the sparse-to-dense annotation methodology could further improve the robustness of this study, where the generated 3D annotations would be compared against ground truth labels from multiple experts in terms of inter-rater segmentation accuracy metrics.
In a clinical setting, the proposed 3D segmentation model is designed to operate locally, on a dedicated GPU within the US acquisition setup. The model will process B-mode US images immediately after acquisition. Another potential application involves integrating the model with PACS systems using standard protocols like DICOM to further enhance clinical workflows. This integration ensures that segmented images are readily accessible to radiologists and other healthcare professionals, facilitating seamless storage, retrieval, and sharing of imaging data. Moreover, the model’s near real-time inference capability would allow for the immediate processing of US images during examinations. This can provide clinicians with instant feedback, aiding in quicker decision-making.
Finally, the imaging setup used in this study enables simultaneous and co-localized acquisition of PA images. As the next step in our research, we will apply the segmentation model in 3D B-mode US images to identify regions for obtaining PD and PA measurements. Such measurements will be used to quantify hemoglobin levels in the synovial tissue, which is a crucial biomarker for RA and has the potential to significantly enhance the sensitivity of RA detection.

5. Conclusions

In conclusion, a novel deep learning architecture, SwinDAF3D, was presented. It combines Swin Transformers with a deep attention mechanism, enabling an enhanced capability to capture comprehensive, multiscale contextual information in 3D US images. The proposed model provided superior performance in segmenting complex synovium structures in 3D US images, as compared to four baseline DL architectures for 3D segmentation. This advancement shows great promise for enabling more efficient and standardized clinical workflows for RA diagnosis using US imaging.

Author Contributions

Conceptualization, J.Q., G.M.K., X.P., S.G. and Z.Y.; methodology, J.Q., G.M.K. and X.P.; software, J.Q., G.M.K. and X.P.; validation, J.Q., G.M.K., X.P., N.A., G.G. and X.W.; formal analysis, J.Q. and X.P.; investigation, J.Q., X.P. and X.W.; resources, X.W. and D.M.; data curation, X.P., N.A. and G.G.; writing—original draft preparation, J.Q. and G.M.K.; writing—review and editing, X.P., S.G., A.D., Z.X., J.J., S.R., G.X., X.W. and D.M.; visualization, J.Q., G.M.K. and X.P.; supervision, X.W. and D.M.; project administration, X.W. and D.M.; funding acquisition, X.W. and D.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Institute of Arthritis and Musculoskeletal and Skin Diseases (NIAMS) of the National Institutes of Health under award number R01AR060350.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of University of Michigan (protocol number HUM00003693).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Retrospective data were collected at University of Michigan and cannot be shared publicly.

Acknowledgments

The research reported in this publication was supported by the National Institute of Arthritis and Musculoskeletal and Skin Diseases (NIAMS) of the National Institutes of Health under award number R01AR060350. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Conflicts of Interest

Qiu, J.; Karageorgos, G.; Ghose, S.; Dentinger, A. and Mills, D. are employees of GE HealthCare. Yang, Z. is an employee of GE Vernova.

References

  1. Aletaha, D.; Smolen, J.S. Diagnosis and management of rheumatoid arthritis: A review. JAMA 2018, 320, 1360–1372. [Google Scholar] [CrossRef] [PubMed]
  2. Dougados, M.; Devauchelle-Pensec, V.; François Ferlet, J.; D’Agostino, M.A.; Backhaus, M.; Bentin, J.; Chalès, G.; Chary-Valckenaere, I.; Conaghan, P.; Wakefield, R.J.; et al. The ability of synovitis to predict structural damage in rheumatoid arthritis: A comparative study between clinical examination and ultrasound. Ann. Rheum. Dis. 2013, 72, 665–671. [Google Scholar] [CrossRef] [PubMed]
  3. Visser, H. Early diagnosis of rheumatoid arthritis. Best Pract. Res. Clin. Rheumatol. 2005, 19, 55–72. [Google Scholar] [CrossRef] [PubMed]
  4. Naredo, E.; Collado, P.; Cruz, A.; Palop, M.J.; Cabero, F.; Richi, P.; Carmona, L.; Crespo, M. Longitudinal power Doppler ultrasonographic assessment of joint inflammatory activity in early rheumatoid arthritis: Predictive value in disease activity and radiologic progression. Arthritis Care Res. Off. J. Am. Coll. Rheumatol. 2007, 57, 116–124. [Google Scholar] [CrossRef]
  5. Sudoł-Szopińska, I.; Jans, L.; Teh, J. Rheumatoid arthritis: What do MRI and ultrasound show. J. Ultrason. 2017, 17, 5–16. [Google Scholar]
  6. Savnik, A.; Malmskov, H.; Thomsen, H.S.; Graff, L.B.; Nielsen, H.; Danneskiold-Samsøe, B.; Boesen, J.; Bliddal, H. MRI of the wrist and finger joints in inflammatory joint diseases at 1-year interval: MRI features to predict bone erosions. Eur. Radiol. 2002, 12, 1203–1210. [Google Scholar]
  7. Østergaard, M.; Stoltenberg, M.; Løvgreen-Nielsen, P.; Volck, B.; Jensen, C.H.; Lorenzen, I. Magnetic resonance imaging-determined synovial membrane and joint effusion volumes in rheumatoid arthritis and osteoarthritis. Comparison with the macroscopic and microscopic appearance of the synovium. Arthritis Rheum. 1997, 40, 1856–1867. [Google Scholar] [CrossRef]
  8. Rowbotham, E.L.; Grainger, A.J. Rheumatoid arthritis: Ultrasound versus MRI. Am. J. Roentgenol. 2011, 197, 541–546. [Google Scholar]
  9. Jo, J.; Xu, G.; Zhu, Y.; Burton, M.; Sarazin, J.; Schiopu, E.; Gandikota, G.; Wang, X. Detecting joint inflammation by an LED-based photoacoustic imaging system: A feasibility study. J. Biomed. Opt. 2018, 23, 110501. [Google Scholar]
  10. Yang, M.; Zhao, C.; Wang, M.; Wang, Q.; Zhang, R.; Bai, W.; Liu, J.; Zhang, S.; Xu, D.; Liu, S.; et al. Synovial oxygenation at photoacoustic imaging to assess rheumatoid arthritis disease activity. Radiology 2023, 306, 220–228. [Google Scholar]
  11. Jo, J.; Xu, G.; Cao, M.; Marquardt, A.; Francis, S.; Gandikota, G.; Wang, X. A functional study of human inflammatory arthritis using photoacoustic imaging. Sci. Rep. 2017, 7, 15026. [Google Scholar]
  12. Peng, X.; Dentinger, A.; Kewalramani, S.; Xu, Z.; Gray, S.; Ghose, S.; Tan, Y.T.; Yang, Z.; Jo, J.; Chamberland, D.; et al. An Automatic 3D Ultrasound and Photoacoustic Combined Imaging System for Human Inflammatory Arthritis. IEEE Trans. Ultrason. Ferroelectr. Freq. Control 2023, 70, 1691–1702. [Google Scholar] [PubMed]
  13. Nurzynska, K.; Smolka, B. Segmentation of finger joint synovitis in ultrasound images. In Proceedings of the 2016 IEEE Sixth International Conference on Communications and Electronics (ICCE), Ha Long, Vietnam, 27–29 July 2016; IEEE: New York, NY, USA, 2016; pp. 335–340. [Google Scholar]
  14. Radlak, K.; Raddle, N.; Smolka, B. Automatic detection of bones based on the confidence map for rheumatoid arthritis analysis. Comput. Vis. Med. Image Process. 2015, 215–220. [Google Scholar]
  15. Hemalatha, R.; Vijaybaskar, V.; Thamizhvani, T. Automatic localization of anatomical regions in medical ultrasound images of rheumatoid arthritis using deep learning. Proc. Inst. Mech. Eng. Part H J. Eng. Med. 2019, 233, 657–667. [Google Scholar]
  16. Wang, Z. Deep learning in medical ultrasound image segmentation: A review. arXiv 2020, arXiv:2002.07703. [Google Scholar]
  17. Song, Y.; Ren, S.; Lu, Y.; Fu, X.; Wong, K.K. Deep learning-based automatic segmentation of images in cardiac radiography: A promising challenge. Comput. Methods Programs Biomed. 2022, 220, 106821. [Google Scholar]
  18. Xing, W.; Li, G.; He, C.; Huang, Q.; Cui, X.; Li, Q.; Li, W.; Chen, J.; Ta, D. Automatic detection of A-line in lung ultrasound images using deep learning and image processing. Med. Phys. 2023, 50, 330–343. [Google Scholar]
  19. Rahman, H.; Bukht, T.F.N.; Imran, A.; Tariq, J.; Tu, S.; Alzahrani, A. A deep learning approach for liver and tumor segmentation in CT images using ResUNet. Bioengineering 2022, 9, 368. [Google Scholar] [CrossRef]
  20. Xu, Y.; Wang, Y.; Yuan, J.; Cheng, Q.; Wang, X.; Carson, P.L. Medical breast ultrasound image segmentation by machine learning. Ultrasonics 2019, 91, 1–9. [Google Scholar]
  21. Huang, Q.; Tian, H.; Jia, L.; Li, Z.; Zhou, Z. A review of deep learning segmentation methods for carotid artery ultrasound images. Neurocomputing 2023, 545, 126298. [Google Scholar]
  22. Sharbatdaran, A.; Romano, D.; Teichman, K.; Dev, H.; Raza, S.I.; Goel, A.; Moghadam, M.C.; Blumenfeld, J.D.; Chevalier, J.M.; Shimonov, D.; et al. Deep learning automation of kidney, liver, and spleen segmentation for organ volume measurements in autosomal dominant polycystic kidney disease. Tomography 2022, 8, 1804–1819. [Google Scholar] [CrossRef]
  23. Ramakrishnan, A.H.; Rajappa, M.; Krithivasan, K.; Chockalingam, N.; Chatzistergos, P.E.; Amirtharajan, R. A concept for fully automated segmentation of bone in ultrasound imaging. Sci. Rep. 2025, 15, 8124. [Google Scholar] [CrossRef] [PubMed]
  24. Huang, A.; Jiang, L.; Zhang, J.; Wang, Q. Attention-VGG16-UNet: A novel deep learning approach for automatic segmentation of the median nerve in ultrasound images. Quant. Imaging Med. Surg. 2022, 12, 3138. [Google Scholar] [CrossRef] [PubMed]
  25. Song, W.; Wang, X.; Guo, Y.; Li, S.; Xia, B.; Hao, A. Centerformer: A novel cluster center enhanced transformer for unconstrained dental plaque segmentation. IEEE Trans. Multimed. 2024, 26, 10965–10978. [Google Scholar] [CrossRef]
  26. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016: 19th International Conference, Athens, Greece, 17–21 October 2016; Proceedings, Part II 19. Springer: Berlin/Heidelberg, Germany, 2016; pp. 424–432. [Google Scholar]
  27. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  28. Liu, Z.; Ning, J.; Cao, Y.; Wei, Y.; Zhang, Z.; Lin, S.; Hu, H. Video swin transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3202–3211. [Google Scholar]
  29. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  30. Rehman, M.U.; Nizami, I.F.; Ullah, F.; Hussain, I. IQA Vision Transformed: A Survey of Transformer Architectures in Perceptual Image Quality Assessment. IEEE Access 2024, 12, 183369–183393. [Google Scholar] [CrossRef]
  31. Hatamizadeh, A.; Nath, V.; Tang, Y.; Yang, D.; Roth, H.R.; Xu, D. Swin unetr: Swin transformers for semantic segmentation of brain tumors in mri images. In International MICCAI Brainlesion Workshop; Springer: Berlin/Heidelberg, Germany, 2021; pp. 272–284. [Google Scholar]
  32. Sun, W.; Chen, J.; Yan, L.; Lin, J.; Pang, Y.; Zhang, G. COVID-19 CT image segmentation method based on swin transformer. Front. Physiol. 2022, 13, 981463. [Google Scholar] [CrossRef]
  33. Yang, H.; Yang, D. CSwin-PNet: A CNN-Swin Transformer combined pyramid network for breast lesion segmentation in ultrasound images. Expert Syst. Appl. 2023, 213, 119024. [Google Scholar] [CrossRef]
  34. Li, G.Y.; Chen, J.; Jang, S.I.; Gong, K.; Li, Q. SwinCross: Cross-modal Swin transformer for head-and-neck tumor segmentation in PET/CT images. Med. Phys. 2024, 51, 2096–2107. [Google Scholar] [CrossRef]
  35. Wang, Y.; Dou, H.; Hu, X.; Zhu, L.; Yang, X.; Xu, M.; Qin, J.; Heng, P.A.; Wang, T.; Ni, D. Deep attentive features for prostate segmentation in 3D transrectal ultrasound. IEEE Trans. Med. Imaging 2019, 38, 2768–2778. [Google Scholar] [CrossRef]
  36. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
  37. Shaker, A.M.; Maaz, M.; Rasheed, H.; Khan, S.; Yang, M.H.; Khan, F.S. UNETR++: Delving into efficient and accurate 3D medical image segmentation. IEEE Trans. Med. Imaging 2024, 43, 3377–3390. [Google Scholar] [PubMed]
  38. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar]
  39. Peng, X.; Xu, Z.; Dentinger, A.; Kewalramani, S.; Jo, J.; Xu, G.; Chamberland, D.; Abdulaziz, N.; Gandikota, G.; Mills, D.; et al. Longitudinal volumetric assessment of inflammatory arthritis via photoacoustic imaging and Doppler ultrasound imaging. Photoacoustics 2023, 31, 100514. [Google Scholar] [CrossRef] [PubMed]
  40. Yushkevich, P.A.; Piven, J.; Cody Hazlett, H.; Gimpel Smith, R.; Ho, S.; Gee, J.C.; Gerig, G. User-Guided 3D Active Contour Segmentation of Anatomical Structures: Significantly Improved Efficiency and Reliability. Neuroimage 2006, 31, 1116–1128. [Google Scholar] [CrossRef]
  41. Karageorgos, G.M.; Qiu, J.; Peng, X.; Yang, Z.; Ghose, S.; Dentinger, A.; Xu, Z.; Jo, J.; Ragupathi, S.; Xu, G.; et al. Automated Deep Learning-Based Finger Joint Segmentation in 3-D Ultrasound Images with Limited Dataset. Ultrason. Imaging 2025, 47, 14–23. [Google Scholar] [CrossRef]
  42. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; Proceedings 4. Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  43. Pérez-García, F.; Sparks, R.; Ourselin, S. TorchIO: A Python library for efficient loading, preprocessing, augmentation and patch-based sampling of medical images in deep learning. Comput. Methods Programs Biomed. 2021, 208, 106236. [Google Scholar]
  44. Agarap, A.F. Deep learning using rectified linear units (relu). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  45. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
  46. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  47. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  48. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.; Roth, H.R.; Xu, D. Unetr: Transformers for 3d medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 574–584. [Google Scholar]
  49. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8024–8035. [Google Scholar]
  50. Cardoso, M.J.; Li, W.; Brown, R.; Ma, N.; Kerfoot, E.; Wang, Y.; Murrey, B.; Myronenko, A.; Zhao, C.; Yang, D.; et al. Monai: An open-source framework for deep learning in healthcare. arXiv 2022, arXiv:2211.02701. [Google Scholar]
  51. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  52. Shamshad, F.; Khan, S.; Zamir, S.W.; Khan, M.H.; Hayat, M.; Khan, F.S.; Fu, H. Transformers in medical imaging: A survey. Med. Image Anal. 2023, 88, 102802. [Google Scholar]
  53. Heidari, M.; Kolahi, S.G.; Karimijafarbigloo, S.; Azad, B.; Bozorgpour, A.; Hatami, S.; Azad, R.; Diba, A.; Bagci, U.; Merhof, D.; et al. Computation-Efficient Era: A Comprehensive Survey of State Space Models in Medical Image Analysis. arXiv 2024, arXiv:2406.03430. [Google Scholar]
  54. Vafaeezadeh, M.; Behnam, H.; Gifani, P. Ultrasound Image Analysis with Vision Transformers. Diagnostics 2024, 14, 542. [Google Scholar] [CrossRef]
  55. Yeung, P.H.; Tan, Y.K.; Xu, S. Automated synovium segmentation in doppler ultrasound images for rheumatoid arthritis assessment. In Proceedings of the Medical Imaging 2018: Computer-Aided Diagnosis, Houston, TX, USA, 10–15 February 2018; SPIE: Philadelphia, PA, USA, 2018; Volume 10575, pp. 135–142. [Google Scholar]
  56. du Toit, C.; Hutter, M.; Gyacskov, I.; Tessier, D.; Dima, R.; Fenster, A.; Lalone, E. Deep learning for synovial volume segmentation of the first carpometacarpal joint in osteoarthritis patients. Osteoarthr. Imaging 2024, 4, 100176. [Google Scholar] [CrossRef]
  57. Kuok, C.P.; Yang, T.H.; Tsai, B.S.; Jou, I.M.; Horng, M.H.; Su, F.C.; Sun, Y.N. Segmentation of finger tendon and synovial sheath in ultrasound image using deep convolutional neural network. Biomed. Eng. Online 2020, 19, 24. [Google Scholar]
Figure 1. Examples of US images of the synovium with corresponding segmentation overlays (green). The images illustrate the challenges posed by ambiguous boundaries, large variations in synovium shape, and inhomogeneous intensity distribution.
Figure 1. Examples of US images of the synovium with corresponding segmentation overlays (green). The images illustrate the challenges posed by ambiguous boundaries, large variations in synovium shape, and inhomogeneous intensity distribution.
Bioengineering 12 00390 g001
Figure 2. An example of a 3D manual sparse annotation (A) and its corresponding generated 3D dense annotation (B).
Figure 2. An example of a 3D manual sparse annotation (A) and its corresponding generated 3D dense annotation (B).
Bioengineering 12 00390 g002
Figure 3. Three-dimensional UNet network architecture for synovium segmentation in 3D US images of finger joints. Conv: convolution; BN: batch normalization; ReLu: ReLu activation.
Figure 3. Three-dimensional UNet network architecture for synovium segmentation in 3D US images of finger joints. Conv: convolution; BN: batch normalization; ReLu: ReLu activation.
Bioengineering 12 00390 g003
Figure 4. DAF3D network architecture for synovium segmentation in 3D US images of finger joints. FPN: Feature Pyramid Network; SLF: single-layer features; MLF: multi-layer features; ASPP: Atrous spatial pyramid pooling.
Figure 4. DAF3D network architecture for synovium segmentation in 3D US images of finger joints. FPN: Feature Pyramid Network; SLF: single-layer features; MLF: multi-layer features; ASPP: Atrous spatial pyramid pooling.
Bioengineering 12 00390 g004
Figure 5. Swin UNETR network architecture for synovium segmentation in 3D US images of finger joints. MLP: Multi-layer Perceptron; W-MSA: window-based Multi-Head Self-Attention; SW-MSA is shifted window-based Multi-Head Self-Attention.
Figure 5. Swin UNETR network architecture for synovium segmentation in 3D US images of finger joints. MLP: Multi-layer Perceptron; W-MSA: window-based Multi-Head Self-Attention; SW-MSA is shifted window-based Multi-Head Self-Attention.
Bioengineering 12 00390 g005
Figure 6. UNETR++ network architecture for synovium segmentation in 3D US images of finger joints. EPA: Efficient Paired-Attention.
Figure 6. UNETR++ network architecture for synovium segmentation in 3D US images of finger joints. EPA: Efficient Paired-Attention.
Bioengineering 12 00390 g006
Figure 7. TransUNet network architecture for synovium segmentation in 3D US images of finger joints. MLP: Multi-layer Perceptron.
Figure 7. TransUNet network architecture for synovium segmentation in 3D US images of finger joints. MLP: Multi-layer Perceptron.
Bioengineering 12 00390 g007
Figure 8. SwinDAF3D network architecture for synovium segmentation in 3D US images of finger joints. FPN: Feature Pyramid Network; SLF: single-layer features; MLF: multi-layer features; ASPP: Atrous spatial pyramid pooling.
Figure 8. SwinDAF3D network architecture for synovium segmentation in 3D US images of finger joints. FPN: Feature Pyramid Network; SLF: single-layer features; MLF: multi-layer features; ASPP: Atrous spatial pyramid pooling.
Bioengineering 12 00390 g008
Figure 9. Training (A) and validation (B) Dice score accuracy curves (one-fold) over 50 epochs for all models in the ablation study: 3D UNet, DAF3D, Swin UNETR, UNETR++, TransUNet, and SwinDAF3D.
Figure 9. Training (A) and validation (B) Dice score accuracy curves (one-fold) over 50 epochs for all models in the ablation study: 3D UNet, DAF3D, Swin UNETR, UNETR++, TransUNet, and SwinDAF3D.
Bioengineering 12 00390 g009
Figure 10. Example feature map visualizations demonstrating the efficacy of feature extraction by integrating Swin transformers feature pyramid (Swin FPN) with attention modules. (A) illustrates the DA3D single-layer features (SLFs) from levels 1 to 4 before and after applying attention modules, along with the final segmentation output. (B) shows the SwinDAF3D SLFs from levels 1 to 4 before and after applying attention modules, alongside the final segmentation output. We can observe that Swin FPN’s feature extraction provides superior feature representation at each level, and the attention module effectively refines these features.
Figure 10. Example feature map visualizations demonstrating the efficacy of feature extraction by integrating Swin transformers feature pyramid (Swin FPN) with attention modules. (A) illustrates the DA3D single-layer features (SLFs) from levels 1 to 4 before and after applying attention modules, along with the final segmentation output. (B) shows the SwinDAF3D SLFs from levels 1 to 4 before and after applying attention modules, alongside the final segmentation output. We can observe that Swin FPN’s feature extraction provides superior feature representation at each level, and the attention module effectively refines these features.
Bioengineering 12 00390 g010
Figure 11. Test set sample segmentation results from the 3D UNet, Swin UNETR, DAF3D, TransUNet, UNETR++, and SwinDAF3D models are compared against their corresponding ground truth annotations, with both segmentation and ground truth masks overlaid in green. (A) Ultrasound images with higher quality and larger synovium size; (B) challenging ultrasound images featuring a larger shadow, reduced contrast, and a smaller synovium.
Figure 11. Test set sample segmentation results from the 3D UNet, Swin UNETR, DAF3D, TransUNet, UNETR++, and SwinDAF3D models are compared against their corresponding ground truth annotations, with both segmentation and ground truth masks overlaid in green. (A) Ultrasound images with higher quality and larger synovium size; (B) challenging ultrasound images featuring a larger shadow, reduced contrast, and a smaller synovium.
Bioengineering 12 00390 g011
Table 1. Comparison of model architectures in the ablation study. This table provides a comprehensive overview of each model’s key components, including the the use of CNN backbones, transformer backbones, attention mechanisms, and the specific levels of transformer integration.
Table 1. Comparison of model architectures in the ablation study. This table provides a comprehensive overview of each model’s key components, including the the use of CNN backbones, transformer backbones, attention mechanisms, and the specific levels of transformer integration.
Model NameCNN BackboneTransformer BackboneAttention MechanismTransformer Integration Level
3D UNet [26]
DAF3D [35]✓(Deep Attention)
Swin UNETR [31]✓(Swin)Encoder
UNETR++ [37]✓(EPA)Encoder & Decoder
TransUNet [38]Bottleneck
SwinDAF3D✓(Swin)✓(Deep Attention)Encoder
Table 2. Segmentation performance statistics of six-fold cross-validation and independent test set in terms of dice score (DSC), intersection over union (IoU), and surface Dice score (SDSC) for all models in the ablation study: 3D UNet, DAF3D, Swin UNETR, UNETR++, TransUNet, and SwinDAF3D.
Table 2. Segmentation performance statistics of six-fold cross-validation and independent test set in terms of dice score (DSC), intersection over union (IoU), and surface Dice score (SDSC) for all models in the ablation study: 3D UNet, DAF3D, Swin UNETR, UNETR++, TransUNet, and SwinDAF3D.
3D UNetDAF3DSwin UNETRUNETR++TransUNetSwinDAF3D
MetricsDSCIoUSDSCDSCIoUSDSCDSCIoUSDSCDSCIoUSDSCDSCIoUSDSCDSCIoUSDSC
Fold 10.7230.5660.6390.8350.7150.8380.8290.7060.8630.8330.7110.8760.8420.7220.8230.8540.7450.887
Fold 20.7470.5960.6750.8140.6970.8260.7680.6280.7980.7910.6590.8010.8170.6900.7890.8150.6840.845
Fold 30.7840.6380.6380.8010.6730.8140.8020.6740.8230.8080.6850.8180.8050.6740.7960.8410.7230.821
Fold 40.7130.5550.6230.8330.7140.8210.8350.7040.8220.8010.6710.8110.8070.6770.8220.8320.7090.851
Fold 50.7210.5630.6870.7880.6530.7980.8310.7080.8690.8210.7010.8570.8240.7030.8310.8490.7330.864
Fold 60.7630.6170.7010.8110.6830.8050.7850.6450.7550.8070.6750.8140.8120.6830.8280.8350.7190.841
Mean0.7420.5890.6610.8130.6890.8170.8080.6780.8220.8100.6840.8290.8180.6920.8150.8380.7190.852
Std0.0250.0310.0290.0170.0220.0130.0250.0320.0390.0140.0180.0270.0130.0170.0160.0130.0190.020
Test Set0.6910.5710.6330.7680.6430.7930.7880.6540.8140.7790.6460.8110.7810.6510.7980.8250.6920.832
Table 3. Segmentation performance metrics on the independent test set for SwinDAF3D with varying feature size and local window size.
Table 3. Segmentation performance metrics on the independent test set for SwinDAF3D with varying feature size and local window size.
Feature Size (Window Size = 7)Window Size (Feature Size = 48)
SizeDSCIoUSDSCSizeDSCIoUSDSC
240.7930.6580.80130.7780.6410.793
360.8110.6760.81850.8080.6710.831
480.8250.6920.83270.8250.6920.832
600.8270.6950.83390.8240.6900.837
Table 4. Comparison of computational cost, memory usage, and inference time for all models in the ablation study: 3D UNet, DAF3D, Swin UNETR, UNETR++, TransUNet and SwinDAF3D.
Table 4. Comparison of computational cost, memory usage, and inference time for all models in the ablation study: 3D UNet, DAF3D, Swin UNETR, UNETR++, TransUNet and SwinDAF3D.
Model NameGFLOPsMemory Usage (MB)Inference Time (s)
3D UNet100625360.175
DAF3D126319580.201
Swin UNETR152272270.343
UNETR++26211650.085
TransUNet118241250.189
SwinDAF3D173075970.375
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qiu, J.; Karageorgos, G.M.; Peng, X.; Ghose, S.; Yang, Z.; Dentinger, A.; Xu, Z.; Jo, J.; Ragupathi, S.; Xu, G.; et al. SwinDAF3D: Pyramid Swin Transformers with Deep Attentive Features for Automated Finger Joint Segmentation in 3D Ultrasound Images for Rheumatoid Arthritis Assessment. Bioengineering 2025, 12, 390. https://doi.org/10.3390/bioengineering12040390

AMA Style

Qiu J, Karageorgos GM, Peng X, Ghose S, Yang Z, Dentinger A, Xu Z, Jo J, Ragupathi S, Xu G, et al. SwinDAF3D: Pyramid Swin Transformers with Deep Attentive Features for Automated Finger Joint Segmentation in 3D Ultrasound Images for Rheumatoid Arthritis Assessment. Bioengineering. 2025; 12(4):390. https://doi.org/10.3390/bioengineering12040390

Chicago/Turabian Style

Qiu, Jianwei, Grigorios M. Karageorgos, Xiaorui Peng, Soumya Ghose, Zhaoyuan Yang, Aaron Dentinger, Zhanpeng Xu, Janggun Jo, Siddarth Ragupathi, Guan Xu, and et al. 2025. "SwinDAF3D: Pyramid Swin Transformers with Deep Attentive Features for Automated Finger Joint Segmentation in 3D Ultrasound Images for Rheumatoid Arthritis Assessment" Bioengineering 12, no. 4: 390. https://doi.org/10.3390/bioengineering12040390

APA Style

Qiu, J., Karageorgos, G. M., Peng, X., Ghose, S., Yang, Z., Dentinger, A., Xu, Z., Jo, J., Ragupathi, S., Xu, G., Abdulaziz, N., Gandikota, G., Wang, X., & Mills, D. (2025). SwinDAF3D: Pyramid Swin Transformers with Deep Attentive Features for Automated Finger Joint Segmentation in 3D Ultrasound Images for Rheumatoid Arthritis Assessment. Bioengineering, 12(4), 390. https://doi.org/10.3390/bioengineering12040390

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop