Next Article in Journal
Gamification as an Innovative Approach for the Assessment of Procedural Knowledge
Previous Article in Journal
Calculations of Electrical Parameters of Cables in Wide Frequency Range
Previous Article in Special Issue
Frequency-Domain Masking and Spatial Interaction for Generalizable Deepfake Detection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Adaptive Cycle-GAN-Based Augmented LIME-Enabled Multi-Stage Transfer Learning Model for Improving Breast Tumor Detection Using Ultrasound Images

Department of Computer Science and Engineering, GITAM University, Hyderabad 502329, India
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2025, 14(8), 1571; https://doi.org/10.3390/electronics14081571
Submission received: 26 February 2025 / Revised: 2 April 2025 / Accepted: 8 April 2025 / Published: 13 April 2025
(This article belongs to the Special Issue Advances in Machine Learning for Image Classification)

Abstract

:
Breast cancer is recognized as an aggressive cancer with the highest rate of mortality. Ultrasound imaging is a non-invasive and cost-effective strategy which is most frequently utilized in clinical methods. Especially, in ultrasound scan, breast tumors may appear in blurred and unclear boundaries. Thus, there is a necessity to improve the quality of breast ultrasound images. In this work, we introduce a cycle generative adversarial network (GAN) for translating noisy breast ultrasound images to denoised images. Furthermore, translating denoised images to reconstructed images helps in preserving breast tumor boundaries for better efficacy. To accurately identify the augmented breast tumor images, we consider an ensemble model of pre-trained transfer learning models such as Inception-v3, Densenet121, and XceptionLike. Furthermore, we present an automated boundary extraction using Local Interpretable Model-agnostic Explanations (LIME), providing interpretability for boundary extraction in breast lesions from ultrasound images. Through experimentation, we have achieved 93% of accuracy for the proposed model, and LIME provides better interpretability for each pre-trained model. Furthermore, the proposed model outperforms Vison Transformer (ViT) models.

1. Introduction

Breast cancer has emerged as the most common malignancy in the globe, with more than 2 million new cases since 2020 [1]. The most accurate way to effectively handle breast cancer is to diagnose the cancer at an early stage. Magnetic resonance imaging (MRI), Breast ultrasound (BUS) images, and computed tomography (CT) are non-invasive strategies for early breast tumor identification. However, BUS is effective for soft tissues and less expensive than MRI and CT. In addition, BUS is an effective approach to initially diagnose benign and malignant breast tumors. Moreover, BUS does not require ionizing radiations and is more effective in diagnosing dense breast tissues, whereas mammograms are very sensitivity and less effective in dense tissues [2]. In the medical field, automated diagnostic systems have emerged as vital for breast cancer identification. Moreover, automated diagnostic systems may help physicians in diagnosing cancer patients and reduce the mortality rate through remote medical diagnosis [3]. For BUS images, image pre-processing and segmentation techniques are important for accurate classification. Most of the existing studies rely on texture changes and image segmentation techniques which may not retrieve essential features and suffer from overfitting [4,5]. However, BUS images may have blurred, low contrast, speckle noise and unclear boundaries, which makes early and accurate diagnosis challenging [6]. Thus, there is a necessity to improve the quality of breast ultrasound images.
Even though medical datasets are publicly available, the majority of datasets exhibit an imbalanced class distribution and are limited in size. However, acquiring health-related data is a time-consuming and expensive process that relies on collaborating with radiologists and clinical researchers [7]. Recently, for BUS images, generative learning has attained superior performance in comparison to supervised learning [8]. A new era of generative learning with the rise of generative adversarial networks (GANs) has gained attention, generating high quality synthetic BUS images. Some existing studies have adapted the GAN for medical image segmentation, generating realistic texture translation and medical image denoising [9,10]. In this paper, we have considered an adaptive cycle-GAN to eliminate unwanted distortion from BUS images. Moreover, the proposed adaptive cycle-GAN generates high-quality BUS images from unwanted distortion of low-quality BUS images in order to improve diagnostic accuracy. In addition, the performance of the proposed adaptive cycle-GAN is evaluated based on the loss function, which relies on cycle consistency loss, adversarial loss, and the residual loss of encoder and decoder networks for image denoising.
For deep learning applications, the datasets with a sufficient number of training samples and credible ground truth are essential factors for an accurate diagnosis [11]. For a medical dataset, acquiring these factors is a major difficult task. Thus, transfer learning helps in utilizing the knowledge that has been trained on a particular model in order to handle a similar task [12]. In addition, transfer learning also helps to improve the performance, especially when the datasets are limited in size. Recently, some existing studies have depicted that pre-trained convolutional neural network (CNN) models have attained higher accuracy, especially with limited datasets [13,14]. However, during transfer learning, identifying the layers of pre-trained model which require fine-tuning is a challenging task. This problem can be addressed with the consideration of an ensemble of transfer learning through an adaptive fine-tuning techniques (like freezing early layers) that can reduce the over-fitting and computational cost. In this work, we consider an ensemble of multi-stage transfer learning models such as DenseNet 121, Inception-V3, and XceptionLike for classifying breast tumor images. The predictions of the proposed ensemble-based multi-stage transfer learning model are explained using explainable artificial intelligence techniques such as LIME. For breast tumor diagnosis, LIME provides an individual predictions on the local interpretable representation of BUS images. Thus, with the consideration of LIME, the efficacy of the proposed model improves trustworthiness in diagnosing breast cancer patients. In this work, the major contributions are summarized as follows:
  • We have considered an adaptive cycle-GAN to eliminate unwanted distortion from BUS images.
  • We consider an ensemble of multi-stage transfer learning models such as DenseNet 121, Inception-V3, and XceptionLike for classifying breast tumor images.
  • Furthermore, we present an automated boundary extraction using Local Interpretable Model-agnostic Explanations (LIME) which highlights the boundary extraction of breast lesions from BUS images and provides the interpretable explanations on the predictions of the proposed multi-stage transfer learning model.
  • During experimentation, the proposed model’s efficacy has been assessed and compared with other existing approaches.

2. Related Work

In this section, GAN-based synthetic medical image augmentation, identification, and classification of breast tumor using deep techniques have been discussed.
Most of the current existing studies have employed generative adversarial networks (GANs) for medical image augmentation. Recently, some studies have adapted GAN-based image-to-image translation, cross-domain image translation and segmentation to image translation. Xue et al. [15] have considered two adversarial networks, namely, segmentor and critic for medical image segmentation. Dai et al. [16] have adapted the GAN to generate realistic segmentation that resembles the ground truth. In [17], the authors have employed semi-supervised the GAN for breast ultrasound image segmentation. Schlegl et al. [18] have applied the GAN on the retinal region for the identification of benign tissues. Furthermore, the authors have validated the GAN for the identification of anomalous data in retinal images. In [19], the authors have trained their model using the GAN for data augmentation to improve breast ultrasound images. However, the model cannot achieve better performance with unknown noisy data in real images. Zhong et al. [20] have adapted the GAN for image denoising using convolutional neural networks (CNNs). Moreover, neural networks often suffer from data drift and vanishing or exploding gradient problems. To avoid data drift, the residual loss of an autoencoder within a generator network of the GAN helps in extracting the features from the input BUS images and improves the training stability of the GAN. Han et al. [21] have adapted an unsupervised GAN model to detect medical anomalies, either from reconstruction loss or the latent feature space. The recent advances in variational autoencoders with different variants of GANs help in improving the efficacy of diagnosing breast cancer patients.
Physicians require a significant amount of time and clinical effort for the precise identification and diagnosis of breast cancer from medical imaging modalities. Moreover, the usage of current diagnostic techniques and software helps in reducing the costs in diagnosing breast cancer. Several existing studies have been proposed for breast cancer detection and diagnosis. In breast tumor diagnosis, deep learning-based frameworks have shown remarkable success in the last few decade. Feng et al. [6] have presented a boundary enhancement for BUS image segmentation using a transformer network that highlights breast lesion region boundaries. In [22], a boundary-rendering network is presented using graph convolutional networks for extracting breast lesions from ultrasound images. In [8], an attentional GAN model is presented for breast tumor identification using BUS images. However, for the multi-classification problem, there exists some challenging issues related to clinical similarities, as the inconsistent staining of biological samples and at different resolutions BUS images from the similar group may exhibit major variations for the interpretation and analysis of medical images. Zheng et al. [23] have presented a deep learning-assisted adaboost algorithm for the early diagnosis and identification of breast cancer. Yala et al. [24] have adapted a reinforcement learning model with a Markov decision process for the early identification of breast cancer. Furthermore, the authors have defined the reward value as assessing the imaging cost and time step between recommended and actual screening dates for early detection. However, deep learning models may not be well tuned due to the limited availability of labeled and annotated medical data. Thus, transfer learning can be applied to improve the performance of deep learning models.
In recent years, transfer learning has been utilized for medical image analysis as computer vision applications have shown to be more beneficial for precise classification. Ayana et al. [25] have adapted transfer learning by considering pre-trained CNN models for the identification of breast cancer from ultrasound images. The authors have utilized two different types of transfer learning models, namely cross-domain and cross-modal models, for breast cancer diagnosis. Wang et al. [26] have developed an automated computer-aided detection system by considering CNNs with a deep transfer learning model, trained on BUS images. The authors have claimed that CNNs with transfer learning have achieved a sensitivity of 89%. In [27], the authors have utilized transfer learning with GoogLeNet and AlexNet for improving mammographic breast lesion identification. In [28], an automated breast tumor identification system is developed by transferring Inceptionv3, ResNet50, and VGG-16. However, in the past decades, studies have yet to explore deep transfer learning in breast imaging. The main objective of this work is to improve the diagnosis of breast cancer with an ensemble of deep transfer learning models using breast imaging modalities.

3. LIME-Assisted Ensemble-Based Multi-Stage Deep Transfer Learning Model

3.1. Cycle GAN

The early and precise identification and diagnosis of cancer from BUS images is a major challenge due to the presence of random noise and blurred images. Hence, BUS image denoising is an essential pre-processing step that helps to improve the quality of the image. Recently, deep learning-based image denoising techniques have been presented for the removal of random noise from BUS imaging modalities. Most of the existing deep learning-based image denoising techniques rely on supervised learning, which requires realistic noisy images with the corresponding ground-truth images. Thus, it may be a difficult task to acquire real-time BUS images. In this work, we consider an unsupervised denoising technique for BUS imaging using a cycle-generative adversarial network (Cycle-GAN). Moreover, the cycle-GAN relies on cycle consistency and the adversarial loss function in order to avoid paired training samples.
As shown in Figure 1, the cycle-GAN mainly employs two generators and two discriminators for distinguishing among real and generated BUS images. The denoising generator network G P is trained to map from a noisy original medical image P to a distortion-free medical image Q. Moreover, the noisy generator network F Q is trained by mapping from distortion-free medical image Q to the reconstructed medical image P. As depicted in Figure 2, the generator network contains encoder, transformer, and decoder networks. In generator networks, convolution and deconvolution layers are constructed to represent encoder and decoder networks, respectively. Furthermore, transformers are build using residual blocks that leads to faster convergence by updating their weights. Moreover, each convolution layer is represented by the number of filters and size of filters with ReLU activation and batch normalization. In the deconvolution layer, reconstructed images are generated based on the extracted feature set.
The structure of the discriminator networks is depicted in Figure 3, where BN represents batch normalization. The two discriminator networks, namely D i s P and D i s Q , help in distinguishing whether the image contains any distorted data in the medical imaging modalities. The denoised discriminator network D i s P helps to differentiate images from the distortion-free domain, generated by the denoising generator network G P . In addition, the noisy discriminator network D i s Q helps to differentiate images from the reconstructed domain, generated by the noisy generator network F Q . Thus, adversarial loss assures that the generator network helps in generating more genuine BUS images. The generator network G P maps noisy BUS image X to G(q) (where q p r D represents data distribution) in a distortion-free BUS image. For a given generator network G ( q ) , the discriminator network D i s P distinguishes whether the image represents the original data in noisy image sample P. The Wasserstein adversarial loss function with two generator networks, G P and F Q , with gradient penalty is computed as
L o s s W G A N ( D i s P , D i s Q ) = E p [ D i s Q ( G P ( p ) ) ] E q [ D i s Q ( q ) ] + E q [ D i s P ( F Q ( q ) ) ] E p [ D i s P ( p ) ] + λ E p ^ p ^ D i s P ( p ^ ) 2 1 2 + λ E q ^ q ^ D i s Q ( q ^ ) 2 1 2
where the term E represents the expectation value under probabilistic distribution. The first and second terms in Equation (1) represent the adversarial Wasserstein loss function for G P and F Q , respectively. Furthermore, the third and fourth terms in Equation (1) represent the gradient penalty for improving stability during model training. However, training two generator networks, G P and F Q , with two adversarial loss functions may not preserve the original medical image, and some reconstructed images might still contain noise. In addition to the two adversarial loss functions, the cycle-consistency loss function is required to assure that the original image and denoised image (which is generated from the generator network) have different representations, by retaining the same content and making generator networks more stable during training process. This can restrict the generator network from creating random images and ensures that the image-to-image translation is reversible. Thus, the cycle-consistency loss function L o s s C ( F Q , G P ) is computed as
L o s s C ( F Q , G P ) = E p F Q ( G P ( p ) ) p L 1 + E q G P ( F Q ( q ) q L 1
where . represents L1-norm. In order to ensure generated images are similar to the real images, we consider L1 loss for image similarity, which is computed as follows:
L o s s L 1 = E p G P ( p ) p L 1 + E q F Q ( q ) q L 1
Thus, the loss function of the adaptive cycle-GAN contains three components, namely, the Wasserstein adversarial loss function with two generator networks with gradient penalty, cycle-consistency loss, and L1 loss (for image similarity), and it is computed as follows:
L o s s = α w L o s s W G A N ( D i s P , D i s Q ) + α c t L o s s C ( F Q , G P ) + α l L o s s L 1
where α w ,   α c t , and α l are adjusted during training. The terms α w and α c t represent the adaptive weighting strategies of adversarial loss and cycle-consistency loss, and they are computed as α w t = α c t = α 0 e λ t , where α 0 , λ , and t represent the initial weight, decay speed, and current epoch, respectively. Furthermore, the term α l represents the adaptive weighting strategy of the L1 loss, and it is computed as α l = α 0 + 1 1 + E [ L o s s W G A N ] . Moreover, low adversarial loss leads to a higher L1 weight. However, the adaptive mechanism in the cycle-GAN enhances the quality of image translation by dynamically adjusting with the consideration of the adaptive cycle consistency loss parameter and dynamically tuning adversarial loss with the gradient penalty weight. In Algorithm 1, the adaptive mechanism is implemented based on adaptive cycle consistency loss and adaptive Wasserstein loss with gradient penalty. Thus, adaptive weighting helps to improve the quality of reconstructed images and the stability of the training process through the gradient penalty, which requires the Lipschitz constraint. Furthermore, during training, two generators and discriminators are updated alternately in order to maintain a balance between the two generators and discriminators. Thus, alternate training promotes the generator networks to improve in generating realistic BUS images, and discriminators can effectively distinguish real images from generated BUS images. In addition, the cycle-GAN contains the cycle-consistency loss, preserving structural integrity and improving stability during training.
Algorithm 1 Training the cycle-GAN algorithm
  1:
Input:
 
         Input images, N number of iterations
    
Output:
 
         Trained two Generator networks G p and F Q , Discriminator networks D i s P and D i s Q
  2:
for i = 1 to N do
  3:
       for each j t h step do
  4:
             Generate minibatch of m samples to construct G P
  5:
             Generate minibatch of m samples to construct F Q
  6:
             Update Discriminator networks D i s P and D i s Q
  7:
             Compute Wasserstein adversarial loss using (1)
  8:
             Compute cycle-consistency loss function using (2)
  9:
             Compute identity loss using (3)
10:
             Update loss function of cycle-GAN using (4)
11:
       end for
12:
end for
13:
return C G G, G , D i s P and D i s Q

3.2. Ensemble-Based Multi-Stage Deep Transfer Learning Model

In deep neural networks (DNNs), increasing the number of layers can enhance the performance, but this requires more computational resources and processing time. Thus, the integration of DNNs with the transfer learning approach is an effective strategy to reduce the training costs. Through transfer learning, the learning efficacy of the new model increases by utilizing the model parameters of the previously trained model, rather than starting from scratch. In recent years, massive BUS image acquisition with data labeling is a difficult task. However, transfer learning can yield better performance, without the necessity of acquiring massive BUS images. For a sequential-based deep transfer learning model, deep convolutional layers are added, followed by a dropout with batch normalization appended to the flattened layer, for constructing an individual transfer learning model. For the efficient multi-class classification of the breast tumor identification problem, three neurons are used in the output layer of the model with softmax activation. The loss function for each individual transfer learning model is depicted in Equation (5), including the learning rate (as 0.0001) and Adam optimizer.
L o s s = l = 0 M y l l o g ( y ^ l )
where y l and y ^ l represent the actual and predicted model output at the l t h value. Based on the performance metrics of each model, we consider the three best transfer learning models, namely, InceptionV3, Densenet 121, and XceptionLike, with the same loss function and optimizer. InceptionV3 can achieve better performance with minimal available computational resources. Moreover, inceptionV3 incorporates auxiliary classifiers in order to stabilize the training process that leads a to fast convergence rate. Densenet 121 includes skip connections that can be utilized as a feature extractor for BUS images, and it can handle the vanishing gradient problem. Xception utilizes a depthwise separable convolution mechanism that enables the model to extract complex features from BUS images.
For better performance, we propose an ensemble of the three best sequential deep transfer learning models, namely, InceptionV3, Densenet 121, and XceptionLike. As depicted in Algorithm 2, the proposed cycle GAN with an ensemble based multi-stage deep transfer learning (EMDTL) model aggregates the predicted value from InceptionV3, Densenet 121, and XceptionLike. The transfer learning models are pre-trained on BUS images and may not provide precise results on the new data. Moreover, the proposed EMDTL model uses grouped regularization and is adaptable to the variational changes in the BUS imaging data that can effectively address the overfitting problem. However, ensemble models provide precise results rather than individual transfer learning models. Furthermore, the proposed EMDTL model can improve performance by providing adaptability and interpretability. This is because the proposed EMDTL model incorporates three different pre-trained models which are fine-tuned by different CNN architectures. Moreover, the proposed EMDTL model can handle noisy data with the incorporation of image augmentation using the cycle-GAN.
The proposed EMDTL model considers three combinations, namely, average, stacked, and hybrid approaches, in order to define the output class. For the average approach, the values of the softmax activation layer are averaged and the maximum value represents the output class label. For the stacked approach, the output of softmax layers are represented as a stacked vector, which is fed to Softmax layer in order to determine the output class label. For the hybrid approach, the last dense layer and softmax layer are summed to determine the output class label. These approaches help in optimizing the performance of the final class label. Furthermore, we consider Local Interpretable Model-Agnostic Explanations (LIME) to verify the validity of the proposed proposed EMDTL model. Algorithm 3 presents a cycle-GAN with LIME-assisted ensemble-based multi-stage deep transfer learning model (C-LEMDTL) which explores the interpretability of the transfer learning models for breast tumor identification from ultrasound images. In LIME, for each input image, multiple image perturbations are generated and trained using a surrogate model. Thus, LIME focuses on providing local approximations.
Algorithm 2 Ensemble-based multi-stage deep transfer learning model (EMDTL) algorithm
 
    Input:
 
         Trained augmented dataset A d using cycle-GAN, Learning Algorithms { L 1 , L 2 , , L n } , θ threshold parameter
 
    Output:
  1:
i 1 , halt ← false
  2:
while (halt ≠ true) do
  3:
      if (i == 1) then
  4:
             C i = Train_Classifier( L i , A d )
  5:
            Compute loss of C i using (5)
  6:
      else
  7:
            Learn( L i , A d )
  8:
             C i + 1 Transfer( C i ) Train_Classifier( L i , A d )
  9:
            Compute loss of C i using (5)
10:
            if  | C i + 1 | / | C i | θ  then
11:
                  halt = true
12:
                  Break
13:
            end if
14:
      end if
15:
      i = i + 1
16:
end while
17:
C ¯ ensemble ( C i ) i [ 1 , 2 , , n ]
Algorithm 3 Cycle-GAN with LIME-assisted ensemble-based multi-stage deep transfer learning model (C-LEMDTL) algorithm
 
  Input:
 
       Trained augmented dataset A d using cycle-GAN, Learning Algorithms { L 1 , L 2 , , L n } , θ threshold parameter
 
  Output:
1:
Z { }
2:
A d Train_Cycle-GAN(D)
3:
for each instance z A d  do
4:
       z vicinity(z)
5:
      f(z) ← Train_EMDTL(z)
6:
      Compute the explainability of input image z ν ( z ) using (6)
7:
       Z Z ν ( z )
8:
end for

3.3. LIME for an Improved Interpretability from BUS Images

Local Interpretable Model-Agnostic Explanations (LIME) rely on over-segmenting the input image into superpixels (i.e., a set of adjacent pixels with similar features.) based on the Slic [29], Felzenszwalb [30], and Quick-Shift [31] segmentation algorithms. In this work, choosing an optimal quantity of superpixels relies on the required level of segmentation. Let Z represent the sample in the BUS image domain and binary randomized vector Z = {1, 0, …, 1} be a modified perturbed sample from an explainable domain. While translating from an explainable domain Z to the BUS image domain Z, pixels contained inside the superpixels are marked as ‘0’, and the other superpixels remain constant. The perturb BUS images are fed to a machine learning model F, and the predicted output f ( z ) is recorded. Through repeating the above-mentioned process, a new BUS dataset is constructed with the set of perturbed BUS images Z and predicted output labels. Thus, the explainable model is trained on perturbed BUS image samples to train the surrogate model, and the weights assure that the surrogate model is more accurate. Moreover, higher weights are assigned to perturbed BUS images, which are similar to the input BUS image samples. The goal of LIME is to identify the surrogate model g that tries the interpretable machine learning model f, highlighting regions of the input BUS image. Mathematically, this is formulated as
ν ( z ) = a r g m i n g υ L f ( z ) , g ( z ) + α ψ ( g )
where ν ( z ) represents the explainability of input image z, and υ represents a set of explainable models. Let α be a regularization parameter that achieves an optimal balance between interpretability and precision. The term ψ represents the penalty of the surrogate model that restricts the utilization of highly complex models in order to retain interpretability. The loss function L uses the distance metric, which is based on the proximity of z to z and is computed as
L f ( z ) , g ( z ) = z , z f ( z ) g ( z ) 2 η ( z )
where η ( z ) assigns higher weights to the samples that exhibit reduced perturbation in contrast to input image sample z, and whose value lie between 0 and 1. LIME uses the Gaussian kernel function to assess the perturbed image z , which relies on the close proximity to the input image sample z.
η ( z ) = e x p z z 2 ) σ 2
where σ adjusts the dimensions of the exponential function that controls the amount of weights assigned to samples based on their proximity to z.
The main motivation behind the consideration of the explainability is to help physicians’ clinical decision making by incorporating LIME with the classification accuracy of the proposed C-LEMDTL model. Moreover, the transfer learning model mainly focuses on identifying important features of the input and not on the irrelevant background similarities. The proposed C-LEMDTL model helps in identifying a class that is vulnerable to background noise. Thus, the proposed C-LEMDTL model improves efficacy against such vulnerabilities.

4. Experimental Results

The breast ultrasound image (BUSI) dataset [2] contains images which are captured from 600 patients with 500 × 500 image dimensions. The BUSI image dataset is categorized into three classes with 487 benign, 210 malignant, and 133 normal images. During experimentation, 80% and 20% images are used to train and test the proposed C-LEMDTL model, respectively. During experimentation, K-fold cross-validation is performed to evaluate the generalization ability of the proposed C-LEMDTL model. We have conducted experimentation using a GPU-based server (which is manufactured by Tyrone Systems) with NVIDIA A100 and 80GB PCIe. Furthermore, CUDA 11.8 with cuDNN 11.8 was utilized with a batch size 16.
In this section, the performance of the proposed C-LEMDTL model and the existing methods are evaluated using a variety of metrics, such as classification reports, accuracy plots, and explainable AI techniques. For each model, the stability of the accuracy curve indicates how effectively the model performs as a classifier, and it depicts the maximum accuracy of the model. Thus, a relatively smooth curve improves classification accuracy. For the classification reports, each model efficacy is evaluated using three metrics, such as the f1-score, recall, and precision. The precision (P) represents the ratio of correctly classified outputs and the number of positively classified samples, and it is represented as P = ( T _ P / ( T _ P + F _ P ) ) . Moreover, the recall (R) is computed as a ratio of the number of predicted outcomes and all the evaluations of the actual class outcomes, and it is represented as R = ( T _ P + ( T _ P + F _ N ) ) , where T _ P represents true positive, F _ P represents false positive, and F _ N represents false negative. Furthermore, the f1-score (F1) is computed by calculating an average of the precision and recall, and it is represented as F 1 = 2 × P × R P + R .
Table 1 shows a comparison of the various pre-trained models for breast tumor detection. It can be observed that the XceptionLike, Inceptionv3, and DenseNet121 models have achieved the highest accuracy when compared to Resnet50 and MaxViT. To provide efficient treatment to patients, it is essential to determine the accuracy and reliability of each model using the medical image dataset. However, some models may identify and classify BUS images that are not within region of interest (RoI). As shown in Figure 4, we adopt a mask regional CNN model [32] to localize and identify the suspicious lesion in bounding boxes. In addition, Figure 4 depicts that two sample benign, malignant and normal images and masked images (named as Image 1 and Image 2) which play a major role in improving for tumor identification. In BUSI dataset, + symbol represents a reference point for lesion which helps in identifying bounding box around tumor. Later, each bounding box is fed as input to the proposed C-LEMDTL model to determine the malignancy probability value. Furthermore, there is a necessity that the proposed model must identify breast cancer regions using explainable AI techniques for the visualization of the proposed model predictions. In this work, the LIME model is utilized for the visualization of breast cancer regions.
Figure 5 shows the sample experimental results of the proposed C-LEMDTL model. Table 2 shows the comparison of the individual transfer learning models without pre-processing. It is observed that XceptionLike has achieved better precision, whereas Inception-V3 is the second-best classifier in terms of precision. On the other hand, MaxViT has demonstrated the worst classifier performance in terms of precision, with a value of 47.2%. Figure 6 depicts the LIME-based interpretability of the proposed model. Moreover, yellow lines highlight outline tumor boundaries that play a major role in classifying an ultrasound image as benign, malignant, or normal.
Figure 7 shows the comparison of confusion matrices for individual transfer learning models such as InceptionV3, DenseNet121, XceptionLike and the proposed C-LEMDTL model. The proposed C-LEMDTL model achieves an accuracy of 93%. The proposed C-LEMDTL model achieves better performance, while identifying more malignant tumors than the other individual transfer learning models. The proposed model reduces complexity with the consideration of the cycleGAN, which incorporates features from individual transfer learning models. The proposed C-LEMDTL model is validated using 5-fold cross validation and achieved an average of 93% accuracy on the BUSI dataset. Table 3 shows the comparison of the proposed C-LEMDTL model with other existing algorithms. It can be observed that the proposed C-LEMDTL model achieves the highest average accuracy. The average execution time for the proposed C-LEMDTL model with a batch size of 16 and 100 epochs is approximately 1.5–2 h. Moreover, the average execution time without a transfer learning model, with batch size of 16, and 100 epochs is approximately 3.5–4 h. This is due to the fact that transfer learning utilizes the pre-trained models, which enhances learning speed, improves stability, and reduces mode collapse.

5. Conclusions

Breast cancer medical imaging helps in early diagnosing breast cancer patients. Transfer learning with explainable artificial intelligence plays a vital role in the analysis and interpretability of breast cancer imaging. However, most of the existing studies rely on image segmentation techniques and suffer from overfitting. In this work, to avoid imbalanced data, a cycle-GAN is adopted for image augmentation, which synthesizes medical denoised images. In addition, to accurately identify augmented breast tumor images, we consider an ensemble model of pre-trained transfer learning models—Inception-v3, DenseNet121, and XceptionLike. Furthermore, LIME is adopted to provide interpretable explanations on the predictions of the proposed multi-stage transfer learning model. The experimental results have illustrated that the proposed C-LEMDTL model has achieved a 93% of accuracy and outperforms the other existing models.

Author Contributions

Conceptualization, N.S. and G.L.; Methodology, N.S. and G.L.; Software, N.S.; Validation, N.S. and G.L.; Formal analysis, N.S.; Investigation, N.S.; Resources, N.S.; Data curation, N.S.; Writing—original draft, N.S.; Writing—review & editing, G.L.; Supervision, G.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original data presented in the study are openly available in Kaggle at: https://www.kaggle.com/datasets/aryashah2k/breast-ultrasound-images-dataset.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Giaquinto, A.N.; Sung, H.; Miller, K.D.; Kramer, J.L.; Newman, L.A.; Minihan, A.; Jemal, A.; Siegel, R.L. Breast cancer statistics, 2022. CA Cancer J. Clin. 2022, 72, 524–541. [Google Scholar]
  2. Al-Dhabyani, W.; Gomaa, M.; Khaled, H.; Fahmy, A. Dataset of breast ultrasound images. Data Brief 2020, 28, 104863. [Google Scholar] [PubMed]
  3. Hekal, A.A.; Elnakib, A.; Moustafa, H.E.D. Automated early breast cancer detection and classification system. Signal Image Video Process. 2021, 15, 1497–1505. [Google Scholar]
  4. Rezaei, Z. A review on image-based approaches for breast cancer detection, segmentation, and classification. Expert Syst. Appl. 2021, 182, 115204. [Google Scholar]
  5. Ilesanmi, A.E.; Chaumrattanakul, U.; Makhanov, S.S. Methods for the segmentation and classification of breast ultrasound images: A review. J. Ultrasound. 2021, 24, 367–382. [Google Scholar]
  6. Feng, J.; Dong, X.; Liu, X.; Zheng, X. HTBE-Net: A hybrid transformer network based on boundary enhancement for breast ultrasound image segmentation. Displays 2024, 84, 102753. [Google Scholar]
  7. Desai, S.D.; Giraddi, S.; Verma, N.; Gupta, P.; Ramya, S. Breast cancer detection using gan for limited labeled dataset. In Proceedings of the 2020 12th International Conference on Computational Intelligence and Communication Networks (CICN), Bhimtal, India, 25–26 September 2020; pp. 34–39. [Google Scholar]
  8. Han, L.; Huang, Y.; Dou, H.; Wang, S.; Ahamad, S.; Luo, H.; Liu, Q.; Fan, J.; Zhang, J. Semi-supervised segmentation of lesion from breast ultrasound images with attentional generative adversarial network. Comput. Methods Programs Biomed. 2020, 189, 105275. [Google Scholar]
  9. Kaur, P.; Singh, G.; Kaur, P. A review of denoising medical images using machine learning approaches. Curr. Med. Imaging 2018, 14, 675–685. [Google Scholar]
  10. Ho, D.J.; Yarlagadda, D.V.; D’Alfonso, T.M.; Hanna, M.G.; Grabenstetter, A.; Ntiamoah, P.; Brogi, E.; Tan, L.K.; Fuchs, T.J. Deep multi-magnification networks for multi-class breast cancer image segmentation. Comput. Med. Imaging Graph. 2021, 88, 101866. [Google Scholar]
  11. Mohi ud din, N.; Dar, R.A.; Rasool, M.; Assad, A. Breast cancer detection using deep learning: Datasets, methods, and challenges ahead. Comput. Biol. Med. 2022, 149, 106073. [Google Scholar]
  12. Khan, S.; Islam, N.; Jan, Z.; Din, I.U.; Rodrigues, J.J.C. A novel deep learning based framework for the detection and classification of breast cancer using transfer learning. Pattern Recognit. Lett. 2019, 125, 1–6. [Google Scholar] [CrossRef]
  13. Gupta, K.; Chawla, N. Analysis of histopathological images for prediction of breast cancer using traditional classifiers with pre-trained CNN. Procedia Comput. Sci. 2020, 167, 878–889. [Google Scholar] [CrossRef]
  14. Masud, M.; Hossain, M.S.; Alhumyani, H.; Alshamrani, S.S.; Cheikhrouhou, O.; Ibrahim, S.; Muhammad, G.; Rashed, A.E.E.; Gupta, B. Pre-trained convolutional neural networks for breast cancer detection using ultrasound images. ACM Trans. Internet Technol. (TOIT) 2021, 21, 1–17. [Google Scholar] [CrossRef]
  15. Xue, Y.; Xu, T.; Zhang, H.; Long, L.R.; Huang, X. Segan: Adversarial network with multi-scale l 1 loss for medical image segmentation. Neuroinformatics 2018, 16, 383–392. [Google Scholar] [CrossRef]
  16. Dai, W.; Dong, N.; Wang, Z.; Liang, X.; Zhang, H.; Xing, E.P. Scan: Structure correcting adversarial network for organ segmentation in chest x-rays. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Cham, Switzerland, 2018; pp. 263–273. [Google Scholar]
  17. Zhai, D.; Hu, B.; Gong, X.; Zou, H.; Luo, J. ASS-GAN: Asymmetric semi-supervised GAN for breast ultrasound image segmentation. Neurocomputing 2022, 493, 204–216. [Google Scholar] [CrossRef]
  18. Schlegl, T.; Seeböck, P.; Waldstein, S.M.; Schmidt-Erfurth, U.; Langs, G. Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. In Information Processing in Medical Imaging; Springer: Cham, Switzerland, 2017; pp. 146–157. [Google Scholar]
  19. Jiménez-Gaona, Y.; Carrión-Figueroa, D.; Lakshminarayanan, V.; Rodríguez-Álvarez, M.J. Gan-based data augmentation to improve breast ultrasound and mammography mass classification. Biomed. Signal Process. Control. 2024, 94, 106255. [Google Scholar] [CrossRef]
  20. Zhong, Y.; Liu, L.; Zhao, D.; Li, H. A generative adversarial network for image denoising. Multimed. Tools Appl. 2020, 79, 16517–16529. [Google Scholar] [CrossRef]
  21. Han, C.; Rundo, L.; Murao, K.; Noguchi, T.; Shimahara, Y.; Milacski, Z.Á.; Koshino, S.; Sala, E.; Nakayama, H.; Satoh, S. MADGAN: Unsupervised medical anomaly detection GAN using multiple adjacent brain MRI slice reconstruction. BMC Bioinform. 2021, 22, 31. [Google Scholar] [CrossRef]
  22. Huang, R.; Lin, M.; Dou, H.; Lin, Z.; Ying, Q.; Jia, X.; Xu, W.; Mei, Z.; Yang, X.; Dong, Y.; et al. Boundary-rendering network for breast lesion segmentation in ultrasound images. Med. Image Anal. 2022, 80, 102478. [Google Scholar] [CrossRef]
  23. Zheng, J.; Lin, D.; Gao, Z.; Wang, S.; He, M.; Fan, J. Deep learning assisted efficient AdaBoost algorithm for breast cancer detection and early diagnosis. IEEE Access 2020, 8, 96946–96954. [Google Scholar] [CrossRef]
  24. Yala, A.; Mikhael, P.G.; Lehman, C.; Lin, G.; Strand, F.; Wan, Y.L.; Hughes, K.; Satuluru, S.; Kim, T.; Banerjee, I.; et al. Optimizing risk-based breast cancer screening policies with reinforcement learning. Nat. Med. 2022, 28, 136–143. [Google Scholar] [CrossRef] [PubMed]
  25. Ayana, G.; Dese, K.; Choe, S.w. Transfer learning in breast cancer diagnoses via ultrasound imaging. Cancers 2021, 13, 738. [Google Scholar] [CrossRef]
  26. Wang, Y.; Choi, E.J.; Choi, Y.; Zhang, H.; Jin, G.Y.; Ko, S.B. Breast cancer classification in automated breast ultrasound using multiview convolutional neural network with transfer learning. Ultrasound Med. Biol. 2020, 46, 1119–1132. [Google Scholar] [CrossRef]
  27. Yu, S.; Liu, L.; Wang, Z.; Dai, G.; Xie, Y. Transferring deep neural networks for the differentiation of mammographic breast lesions. Sci. China Technol. Sci. 2019, 62, 441–447. [Google Scholar] [CrossRef]
  28. Agarwal, R.; Diaz, O.; Lladó, X.; Yap, M.H.; Martí, R. Automatic mass detection in mammograms using deep convolutional neural networks. J. Med. Imaging 2019, 6, 031409. [Google Scholar] [CrossRef] [PubMed]
  29. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef]
  30. Felzenszwalb, P.F.; Huttenlocher, D.P. Efficient graph-based image segmentation. Int. J. Comput. Vis. 2004, 59, 167–181. [Google Scholar] [CrossRef]
  31. Vedaldi, A.; Soatto, S. Quick shift and kernel methods for mode seeking. In Computer Vision–ECCV 2008: Proceedings of the 10th European Conference on Computer Vision, Marseille, France, 12–18 October 2008; Part IV 10; Springer: Singapore, 2008; pp. 705–718. [Google Scholar]
  32. Bharati, P.; Pramanik, A. Deep learning techniques—R-CNN to mask R-CNN: A survey. In Computational Intelligence in Pattern Recognition: Proceedings of CIPR 2019; Springer: Singapore, 2020; pp. 657–668. [Google Scholar]
  33. Abbas, Q. DeepCAD: A computer-aided diagnosis system for mammographic masses using deep invariant features. Computers 2016, 5, 28. [Google Scholar] [CrossRef]
  34. Charan, S.; Khan, M.J.; Khurshid, K. Breast cancer detection in mammograms using convolutional neural network. In Proceedings of the 2018 International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Sukkur, Pakistan, 3–4 March 2018; pp. 1–5. [Google Scholar]
  35. Ting, F.F.; Tan, Y.J.; Sim, K.S. Convolutional neural network improvement for breast cancer classification. Expert Syst. Appl. 2019, 120, 103–115. [Google Scholar] [CrossRef]
Figure 1. Proposed architecture for breast tumor detection using a multi-stage transfer learning model.
Figure 1. Proposed architecture for breast tumor detection using a multi-stage transfer learning model.
Electronics 14 01571 g001
Figure 2. Architecture of the proposed generator network cycle-GAN.
Figure 2. Architecture of the proposed generator network cycle-GAN.
Electronics 14 01571 g002
Figure 3. Architecture of the discriminator network in this study.
Figure 3. Architecture of the discriminator network in this study.
Electronics 14 01571 g003
Figure 4. Identification of suspicious lesion in bounding boxes with BUS images.
Figure 4. Identification of suspicious lesion in bounding boxes with BUS images.
Electronics 14 01571 g004
Figure 5. Sample experimental results of the proposed C-LEMDTL model.
Figure 5. Sample experimental results of the proposed C-LEMDTL model.
Electronics 14 01571 g005
Figure 6. LIME-based interpretability of the proposed model.
Figure 6. LIME-based interpretability of the proposed model.
Electronics 14 01571 g006
Figure 7. Comparison of the proposed C-LEMDTL with other models.
Figure 7. Comparison of the proposed C-LEMDTL with other models.
Electronics 14 01571 g007
Table 1. Comparison of various pre-trained models after pre-processing.
Table 1. Comparison of various pre-trained models after pre-processing.
ModelMetricsBenignMalignantNormalAccuracy
XceptionLikePrecision0.900.860.830.88
Recall0.890.7741.00.88
f1-score0.900.80.910.88
Inceptionv3Precision0.860.730.890.83
Recall0.890.710.8500.84
f1-score0.880.720.870.83
Densenet121Precision0.880.860.890.88
Recall0.920.800.850.88
f1-score0.900.830.870.88
Resnet50Precision0.840.810.850.84
Recall0.890.700.850.84
f1-score0.870.760.850.84
MaxViTPrecision0.470.680.540.38
Recall0.850.750.540.53
f1-score0.600.710.540.45
Table 2. Comparison of various models before pre-processing.
Table 2. Comparison of various models before pre-processing.
ModelMetricsBenignMalignantNormalMacro Average
XceptionLikePrecision0.8060.7620.7940.779
Recall0.7920.7060.890.788
Inceptionv3Precision0.7820.7570.73330.765
Recall0.7710.6020.7500.759
Densenet121Precision0.7650.7330.7940.731
Recall0.7920.6090.7500.717
Resnet50Precision0.7400.7140.750.735
Recall0.7920.6090.750.717
MaxViTPrecision0.4720.6810.5420.384
Recall0.750.650.5420.533
Table 3. Comparison of the proposed C-LEMDTL with other existing algorithms.
Table 3. Comparison of the proposed C-LEMDTL with other existing algorithms.
MethodAccuracySensitivityPrecision
DeepCAD [33]91.4592-
CNN [34]65--
CNNI-BCC [35]9188.89-
Proposed C-LEMDTL939293
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sappa, N.; Lingam, G. An Adaptive Cycle-GAN-Based Augmented LIME-Enabled Multi-Stage Transfer Learning Model for Improving Breast Tumor Detection Using Ultrasound Images. Electronics 2025, 14, 1571. https://doi.org/10.3390/electronics14081571

AMA Style

Sappa N, Lingam G. An Adaptive Cycle-GAN-Based Augmented LIME-Enabled Multi-Stage Transfer Learning Model for Improving Breast Tumor Detection Using Ultrasound Images. Electronics. 2025; 14(8):1571. https://doi.org/10.3390/electronics14081571

Chicago/Turabian Style

Sappa, Neeraja, and Greeshma Lingam. 2025. "An Adaptive Cycle-GAN-Based Augmented LIME-Enabled Multi-Stage Transfer Learning Model for Improving Breast Tumor Detection Using Ultrasound Images" Electronics 14, no. 8: 1571. https://doi.org/10.3390/electronics14081571

APA Style

Sappa, N., & Lingam, G. (2025). An Adaptive Cycle-GAN-Based Augmented LIME-Enabled Multi-Stage Transfer Learning Model for Improving Breast Tumor Detection Using Ultrasound Images. Electronics, 14(8), 1571. https://doi.org/10.3390/electronics14081571

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop