Hybrid Optimization and Explainable Deep Learning for Breast Cancer Detection
Abstract
1. Introduction
- Development of a hybrid diagnostic framework using mammography and ultrasound images, constructing a diverse fusion dataset for binary breast cancer classification;
- Fine-tuning MobileNet using the Firefly Algorithm (FLA) and Dingo Optimization Algorithm (DOA) to optimize feature selection and improve classification performance;
- Comparative evaluation of multiple transfer learning models (CNN, MobileNet, Xception, DenseNet121, EfficientNet, ResNet50, and VGG16) to identify the most effective architecture for breast cancer diagnosis;
- Performance validation on three datasets: MIAS, Breast Ultrasound, and their fusion;
- Integration of Grad-CAM to generate interpretable visual explanations for model predictions;
- Demonstration of the framework’s superiority over existing methods in terms of accuracy, precision, recall, and F1-score.
2. Proposed Methodology
2.1. Data Description
2.1.1. MIAS Dataset Description
2.1.2. Breast Ultrasound Images Dataset Description
- Top Row (Benign): Three benign images show uniform textures with minimal irregularities, indicating non-cancerous conditions;
- Bottom Row (Malignant): Three malignant images exhibit distinct anomalies, suggesting potential malignancies.
2.1.3. Fusion Dataset Description
- Top Row (Benign): Three benign images show uniform textures with minimal irregularities, indicating non-cancerous conditions;
- Bottom Row (Malignant): Three malignant images exhibit distinct anomalies, suggesting potential malignancies.
2.2. Data Preprocessing
2.3. Transfer Learning Models
- MobileNet inspired a bespoke CNN that uses depth-wise separable convolutions to lower the number of parameters and computational cost without sacrificing speed. The four separate convolution blocks in this lightweight design are followed by batch normalizing and max pooling, which results in dense layers that are completely linked and include dropout regularization. It serves as a foundational architecture for comparison against more complex pre-trained models;
- MobileNet, a highly efficient network designed for mobile and embedded vision applications, further leverages depth-wise separable convolutions and a streamlined structure. In this work, the base MobileNet was frozen and extended with dense layers (1024, 512) and dropout to prevent overfitting. Its main advantage lies in its low latency and memory footprint while achieving strong performance on small medical datasets;
- The Xception model, built upon depth-wise separable convolutions and residual connections, expands on Inception modules by replacing them with extreme Inception blocks. In this approach, a stack of thick layers (2024, 1024, 512) with dropout was added for binary classification, while the ImageNet-trained layers were frozen. The strength of Xception is its ability to preserve parameter efficiency while capturing fine-grained spatial hierarchies in medical pictures;
- In order to encourage feature reuse and mitigate the vanishing gradient issue, DenseNet201, which is renowned for its dense connection across layers, was also used. DenseNet promotes gradient flow and increases representational power by feed-forwardly linking each layer to every other layer. Here, it was augmented with dense layers (2024 × 2, 1024) and dropout layers to tailor the model for binary decision-making in cancer detection;
- VGG16 was employed for its simplicity and consistency. VGG16’s consistent architecture of 3 × 3 convolution layers and 2 × 2 pooling layers makes it resilient and interpretable even if it lacks the sophisticated advances of more current models. In this study, VGG16’s convolutional base was frozen, and new dense layers (2048, 1024, 512) were stacked to adapt it for the classification task. The key strength of VGG16 lies in its deep hierarchical structure, which is particularly beneficial for capturing visual patterns in high-resolution mammographic and ultrasound images;
- ResNet50, another widely used model, addresses the degradation problem in very deep networks through the introduction of residual learning. The residual blocks al low gradients to flow directly through skip connections, enabling the training of deeper architectures. In this implementation, the ResNet50 model was extended with multiple dense layers (2024, 1024, 512, 256, 128) and dropout, providing a deep and expressive architecture that balances depth and training stability;
- The method was modified to use EfficientNetB0, a more modern architecture that scales depth, breadth, and resolution in a compound way. It is renowned for using fewer FLOPs and parameters to achieve excellent precision. EfficientNet’s efficient scaling strategy allows it to outperform traditional CNNs on various benchmarks. In this work, the EfficientNetB0 backbone was frozen, and dense layers (1024, 512, 256, 128) were added with dropout to optimize performance for the binary classification task. Its compact design makes it particularly suitable for real-time and resource-constrained deployment scenarios.
2.4. Optimization Algorithms
2.4.1. Algorithm for Dingo Optimization
- : current position of the i-th candidate;
- : positions of randomly selected attackers;
- : best-known solution so far;
- ∼U(−2, 2): stochastic scaling factor for update magnitude.
- : randomly selected peer position;
- ∼U(−1, 1): additional stochastic factor.
2.4.2. Firefly Algorithm
2.5. Explainable AI
3. Experimental Setup
- TP (True Positives): Number of correctly predicted positive cases;
- TN (True Negatives): Number of correctly predicted negative cases;
- FP (False Positives): Number of negative cases incorrectly predicted as positive;
- FN (False Negatives): Number of positive cases incorrectly predicted as negative.
4. Results and Discussion
4.1. Performance of Hybrid Models in the MIAS Dataset
4.2. Performance of Hybrid Models in the Breast Ultrasound Images Dataset
4.3. Performance of Hybrid Models in the Fusion Dataset
4.4. GRAD-CAM Explainability Analysis for Breast Cancer Identification
4.5. Comparison Results
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Tan, Y.; Sim, K.-S.; Ting, F.F. Breast cancer detection using convolutional neural networks for mammogram imaging system. In Proceedings of the 2017 International Conference on Robotics, Automation and Sciences (ICORAS), Melaka, Malaysia, 27–29 November 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–5. [Google Scholar]
- López-Cabrera, J.D.; Rodríguez, L.A.L.; Pérez-Díaz, M. Classification of breast cancer from digital mammography using deep learning. Intel. Artif. 2020, 23, 56–66. [Google Scholar] [CrossRef]
- Pang, J.; Ding, N.; Liu, X.; He, X.; Zhou, W.; Xie, H.; Feng, J.; Li, Y.; He, Y.; Wang, S.; et al. Prognostic Value of the Baseline Systemic Immune-Inflammation Index in HER2-Positive Metastatic Breast Cancer: Exploratory Analysis of Two Prospective Trials. Ann. Surg. Oncol. 2025, 32, 750–759. [Google Scholar] [CrossRef]
- World Health Organization. Breast Cancer. 2023. Available online: https://www.who.int/news-room/fact-sheets/detail/breast-cancer (accessed on 13 March 2024).
- Han, L.; Yin, Z. A hybrid breast cancer classification algorithm based on meta-learning and artificial neural networks. Front. Oncol. 2022, 12, 1042964. [Google Scholar] [CrossRef]
- Halim, A.; Andrew, A.M.; Yasin, M.N.M.; Rahman, M.A.A.; Jusoh, M.; Veeraperumal, V.; Rahim, H.A.; Illahi, U.; Karim, M.K.A.; Scavino, E. Existing and emerging breast cancer detection technologies and its challenges: A review. Appl. Sci. 2021, 11, 10753. [Google Scholar] [CrossRef]
- Vijayarajeswari, R.; Parthasarathy, P.; Vivekanandan, S.; Basha, A.A. Classification of mammogram for early detection of breast cancer using svm classifier and hough transform. Measurement 2019, 146, 800–805. [Google Scholar] [CrossRef]
- Sarosa, S.J.A.; Utaminingrum, F.; Bachtiar, F.A. Mammogram breast cancer classification using gray-level co-occurrence matrix and support vector machine. In Proceedings of the 2018 International Conference on Sustainable Information Engineering and Technology (SIET), Malang, Indonesia, 10–12 November 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 54–59. [Google Scholar]
- Ma, X.; Cheng, H.; Hou, J.; Jia, Z.; Wu, G.; Lü, X.; Li, H.; Zheng, X.; Chen, C. Detection of breast cancer based on novel porous silicon Bragg reflector surface-enhanced Raman spectroscopy-active structure. Chin. Opt. Lett. 2020, 18, 051701. [Google Scholar] [CrossRef]
- Torres, R.E.O.; Gutiérrez, J.R.; Jacome, A.G.L. Neutrosophic-based machine learning techniques for analysis and diag nosis the breast cancer. Int. J. Neutrosophic Sci. (IJNS) 2023, 21, 1. [Google Scholar]
- Yadav, A.R.; Vaegae, N.K. Development of an early prediction system for breast cancer using machine learning techniques. In Proceedings of the 2023 International Conference on Next Generation Electronics (NEleX), Vellore, India, 14–16 December 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
- Resch, D.; Gullo, R.L.; Teuwen, J.; Semturs, F.; Hummel, J.; Resch, A.; Pinker, K. Ai-enhanced mammography with digital breast tomosynthesis for breast cancer detection: Clinical value and comparison with human performance. Radiol. Imaging Cancer 2024, 6, e230149. [Google Scholar] [CrossRef]
- Shaaban, S.M.; Nawaz, M.; Said, Y.; Barr, M. Anefficient breast cancer segmentation system based on deep learning techniques. Eng. Technol. Appl. Sci. Res. 2023, 13, 12415–12422. [Google Scholar] [CrossRef]
- Gurumoorthy, R.; Kamarasan, M. Breast cancer classification from histopathological images using future search optimization algorithm and deep learning. Eng. Technol. Appl. Sci. Res. 2024, 14, 12831–12836. [Google Scholar] [CrossRef]
- Shen, D.; Wu, G.; Suk, H.-I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 2017, 19, 221–248. [Google Scholar] [CrossRef]
- Le Cun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 7553, 436–444. [Google Scholar] [CrossRef]
- Islam, U.; Al-Atawi, A.A.; Alwageed, H.S.; Mehmood, G.; Khan, F.; Innab, N. Detection of renal cell hydronephrosis in ultrasound kidney images: A study on the efficacy of deep convolutional neural networks. PeerJ Comput. Sci. 2024, 10, e1797. [Google Scholar] [CrossRef]
- Razavian, A.S.; Azizpour, H.; Sullivan, J.; Carlsson, S. Cnn features off-the-shelf: An astounding baseline for recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Columbus, OH, USA, 23–28 June 2014; pp. 806–813. [Google Scholar]
- Penatti, O.A.; Nogueira, K.; Santos, J.A.D. Do deep features generalize from everyday objects to remote sensing and aerial scenes domains? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 44–51. [Google Scholar]
- Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef]
- Nasir, I.M.; Raza, M.; Shah, J.H.; Khan, M.A.; Nam, Y.-C.; Nam, Y. Improved shark smell optimization algorithm for human action recognition. Comput. Mater. Contin 2023, 76, 2667–2684. [Google Scholar]
- Nasir, I.M.; Raza, M.; Ulyah, S.M.; Shah, J.H.; Fitriyani, N.L.; Syafrudin, M. Enga: Elastic net-based genetic algorithm for human action recognition. Expert Syst. Appl. 2023, 227, 120311. [Google Scholar] [CrossRef]
- Nasir, I.M.; Raza, M.; Shah, J.H.; Wang, S.-H.; Tariq, U.; Khan, M.A. Harednet: A deep learning based architecture for autonomous video surveillance by recognizing human actions. Comput. Electr. Eng. 2022, 99, 107805. [Google Scholar] [CrossRef]
- Wang, L.; Jiang, S.; Jiang, S. A feature selection method via analysis of relevance, redundancy, and interaction. Expert Syst. Appl. 2021, 183, 115365. [Google Scholar] [CrossRef]
- Tariq, J.; Alfalou, A.; Ijaz, A.; Ali, H.; Ashraf, I.; Rahman, H.; Armghan, A.; Mashood, I.; Rehman, S. Fast intra mode selection in hevc using statistical model. Comput. Mater. Contin 2022, 70, 3903–3918. [Google Scholar] [CrossRef]
- Gao, Y.; Wang, C.; Wang, K.; He, C.; Hu, K.; Liang, M. The effects and molecular mechanism of heat stress on spermatogenesis and the mitigation measures. Syst. Biol. Reprod. Med. 2022, 68, 331–347. [Google Scholar] [CrossRef] [PubMed]
- Shafipour, M.; Fadaei, S. Particle distance rank feature se lection by particle swarm optimization. Expert Syst. Appl. 2021, 185, 115620. [Google Scholar] [CrossRef]
- Nasir, I.M.; Rashid, M.; Shah, J.H.; Sharif, M.; Awan, M.Y.; Alkinani, M.H. An optimized approach for breast cancer classification for histopathological images based on hybrid feature set. Curr. Med. Imaging Rev. 2021, 17, 136–147. [Google Scholar] [CrossRef]
- Samieinasab, M.; Torabzadeh, S.A.; Behnam, A.; Aghsami, A.; Jolai, F. Meta-health stack: A new approach for breast cancer prediction. Health Care Anal. 2022, 2, 100010. [Google Scholar] [CrossRef]
- Mushtaq, I.; Umer, M.; Imran, M.; Nasir, I.M.; Muhammad, G.; Shorfuzzaman, M. Customer prioritization for medical supply chain dur ing COVID-19 pandemic. Comput. Mater. Contin. 2021, 70, 59–72. [Google Scholar]
- Nardin, S.; Mora, E.; Varughese, F.M.; D’Avanzo, F.; Vachanaram, A.R.; Rossi, V.; Saggia, C.; Rubinelli, S.; Gennari, A. Breast cancer survivor ship, quality of life, and late toxicities. Front. Oncol. 2020, 10, 864. [Google Scholar] [CrossRef]
- Nasir, I.M.; Raza, M.; Shah, J.H.; Khan, M.A.; Rehman, A. Human action recognition using machine learning in uncontrolled environment. In Proceedings of the 2021 1st International Conference on Artificial Intelligence and Data Analytics (CAIDA), Riyadh, Saudi Arabia, 6–7 April 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 182–187. [Google Scholar]
- Alkhanbouli, R.; Almadhaani, H.M.A.; Alhosani, F.; Simsekler, M.C.E. The role of explainable artificial intelligence in disease prediction: A systematic literature review and future research directions. BMC Med. Inform. Decis. Mak. 2025, 25, 110. [Google Scholar] [CrossRef] [PubMed]
- Teoh, J.R.; Hasikin, K.; Lai, K.W.; Wu, X.; Li, C. Enhancing early breast cancer diagnosis through automated microcalcification detection using an optimized ensemble deep learning framework. PeerJ Comput. Sci. 2024, 10, e2082. [Google Scholar] [CrossRef] [PubMed]
- Wang, W.; Li, Y.; Yan, X.; Xiao, M.; Gao, M. Breast cancer image classification method based on deep transfer learning. In Proceedings of the International Conference on Image Processing, Machine Learning and Pattern Recognition, Guangzhou, China, 13–15 September 2024; pp. 190–197. [Google Scholar]
- Sarwar, N.; Al-Otaibi, S.; Irshad, A. Optimizing breast cancer detection: Integrating few-shot and transfer learning for enhanced accuracy and efficiency. Int. J. Imaging Syst. Technol. 2025, 35, e70033. [Google Scholar] [CrossRef]
- Raghavan, K.; B, S.; v, K. Attention guided grad-cam: An improved explain able artificial intelligence model for infrared breast cancer detection. Multimed. Tools Appl. 2024, 83, 57551–57578. [Google Scholar] [CrossRef]
- Naz, A.; Khan, H.; Din, I.U.; Ali, A.; Husain, M. An efficient optimization system for early breast cancer diagnosis based on internet of medical things and deep learning. Eng. Technol. Appl. Sci. Res. 2024, 14, 15957–15962. [Google Scholar] [CrossRef]
- Hassan, M.M.; Hassan, M.M.; Yasmin, F.; Khan, M.A.R.; Zaman, S.; Islam, K.K.; Bairagi, A.K. A comparative assessment of machine learning algorithms with the least absolute shrinkage and selection operator for breast cancer detection and prediction. Decis. Anal. J. 2023, 7, 100245. [Google Scholar] [CrossRef]
- Malakouti, S.M.; Menhaj, M.B.; Suratgar, A.A. Ml: Early breast cancer diagnosis. Curr. Probl. Cancer Case Rep. 2024, 13, 100278. [Google Scholar] [CrossRef]
- Duan, H.; Zhang, Y.; Qiu, H.; Fu, X.; Liu, C.; Zang, X.; Xu, A.; Wu, Z.; Li, X.; Zhang, Q.; et al. Machine learning-based prediction model for distant metastasis of breast cancer. Comput. Biol. Med. 2024, 169, 107943. [Google Scholar] [CrossRef]
- Alhsnony, F.H.; Sellami, L. Advancing breast cancer detection with convolutional neural networks: A comparative analysis of mias and ddsm datasets. In Proceedings of the 2024 IEEE 7th International Conference on Advanced Technologies, Signal and Image Processing (ATSIP) 1, Sousse, Tunisia, 11–13 July 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 194–199. [Google Scholar]
- Thawkar, S. A hybrid model using teaching–learning-based optimization and salp swarm algorithm for feature selection and classification in digital mammography. J. Ambient Intell. Humaniz. Comput. 2021, 12, 8793–8808. [Google Scholar] [CrossRef]
- Sheth, P.; Patil, S. Improved jaya optimization algorithm for feature selection on cancer diagnosis data using evolutionary binary coded approach. Solid State Technol. 2020, 29, 992–1006. [Google Scholar]
- Johnson, D.S.; Johnson, D.L.L.; Elavarasan, P.; Karunanithi, A. Feature selection using flower pollination optimization to diagnose lung cancer from ct images. In Advances in Information and Communication: Proceedings of the 2020 Future of Information and Communication Conference (FICC), San Francisco, CA, USA, 5–6 March 2020; Springer: Berlin/Heidelberg, Germany, 2020; Volume 2, pp. 604–620. [Google Scholar]
- Singh, L.K.; Khanna, M.; Singh, R. Feature subset selection through nature inspired computing for efficient glaucoma classification from fundus images. Multimed. Tools Appl. 2024, 83, 77873–77944. [Google Scholar] [CrossRef]
- Houssein, E.H.; Oliva, D.; Samee, N.A.; Mahmoud, N.F.; Emam, M.M. Liver cancer algorithm: A novel bio-inspired optimizer. Comput. Biol. Med. 2023, 165, 107389. [Google Scholar] [CrossRef] [PubMed]
- Alnowami, M.R.; Abolaban, F.A.; Taha, E. A wrapper-based feature selection approach to investigate potential biomarkers for early detection of breast cancer. J. Radiat. Res. Appl. Sci. 2022, 15, 104–110. [Google Scholar] [CrossRef]
- Houssein, E.H.; Emam, M.M.; Ali, A.A. An optimized deep learning architecture for breast cancer diagnosis based on improved marine predators’ algorithm. Neural Comput. Appl. 2022, 34, 18015–18033. [Google Scholar] [CrossRef]
- Song, W.; Wang, X.; Guo, Y.; Li, S.; Xia, B.; Hao, A. CenterFormer: A Novel Cluster Center Enhanced Transformer for Unconstrained Dental Plaque Segmentation. IEEE Trans. Multimed. 2024, 26, 10965–10978. [Google Scholar] [CrossRef]
- Naas, M.; Mzoughi, H.; Njeh, I.; Slima, M.B. Deep learning-based computer aided diagnosis (cad) tool supported by explainable artificial intelligence for breast cancer exploration. Appl. Intell. 2025, 55, 679. [Google Scholar] [CrossRef]
- Fattahi, A.S.; Hoseini, M.; Dehghani, T.; Nia, R.G.N.N.; Naseri, Z.; Ebrahimzadeh, A.; Mahri, A.; Eslami, S. Explainable machine learning versus known nomogram for predicting non-sentinel lymph node metastases in breast cancer patients: A comparative study. Comput. Biol. Med. 2025, 184, 109412. [Google Scholar] [CrossRef]
- Murugan, T.K.; Karthikeyan, P.; Sekar, P. Efficient breast cancer detection using neural networks and explainable artificial intelligence. Neural Comput. Appl. 2025, 37, 3759–3776. [Google Scholar] [CrossRef]
- Manojee, K.S.; Kannan, A.R. Patho-net: Enhancing breast cancer classification using deep learning and explainable artificial intelligence. Am. J. Cancer Res. 2025, 15, 754. [Google Scholar] [CrossRef] [PubMed]
- Ariyametkul, A.; Tamang, S.; Paing, M.P. Explainable ai (xai) for breast cancer diagnosis. In Proceedings of the 2024 16th Biomedical Engineering International Conference (BMEiCON), Pattaya, Thailand, 21–24 November 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–5. [Google Scholar]
- Hussain, S.M.; Buongiorno, D.; Altini, N.; Berloco, F.; Prencipe, B.; Moschetta, M.; Bevilacqua, V.; Brunetti, A. Shape-based breast lesion classification using digital tomosynthesis images: The role of explainable artificial intelligence. Appl. Sci. 2022, 12, 6230. [Google Scholar] [CrossRef]
- Suckling, J.; Parker, J.; Dance, D.; Astley, S.; Hutt, I.; Boggis, C.; Ricketts, I.; Stamatakis, E.; Cerneaz, N.; Kok, S.; et al. Mammographic Image Analysis Society (MIAS) Database v1.21; Apollo-University of Cambridge Repository: Online, 2015. [Google Scholar] [CrossRef]
- Al-Dhabyani, W.; Gomaa, M.; Khaled, H.; Fahmy, A. Dataset of breast ultrasound images. Data Brief 2020, 28, 104863. [Google Scholar] [CrossRef]
- Peraza-Vázquez, H.; Peña-Delgado, A.F.; Echavarría-Castillo, G.; Morales-Cepeda, A.B.; Velasco-Álvarez, J.; Ruiz-Perez, F. A bio-inspired method for engineering design optimization inspired by dingoes hunting strategies. Math. Probl. Eng. 2021, 2021, 9107547. [Google Scholar] [CrossRef]
- Boyd, S.P.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
- Rashedi, E.; Nezamabadi-Pour, H.; Saryazdi, S. GSA: A gravitational search algorithm. Inf. Sci. 2009, 179, 2232–2248. [Google Scholar] [CrossRef]
- Yang, X.-S. Firefly algorithms for multimodal optimization. In Stochastic Algorithms: Foundations and Applications; SAGA: London, UK; Springer: Berlin/Heidelberg, Germany, 2009; LNCS 5792; pp. 169–178. [Google Scholar]
- Yang, X.-S.; Deb, S. Engineering optimisation by cuckoo search. Int. J. Math. Model. Numer. Optim. 2010, 1, 330–343. [Google Scholar] [CrossRef]
- Hashim, F.A.; Mostafa, R.R.; Hussien, A.G.; Mirjalili, S.; Sallam, K.M. Firefly algorithm: A physical law-based algorithm for numerical optimization. Knowl.-Based Syst. 2023, 260, 110146. [Google Scholar] [CrossRef]
- Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
- Holzinger, A.; Biemann, C.; Pattichis, C.S.; Kell, D.B. What do we need to build explainable ai systems for the medical domain? arXiv 2017, arXiv:1712.09923. [Google Scholar]
- Gunning, D.; Aha, D. Darpa’s explainable artificial intelligence (xai) program. AI Mag. 2019, 40, 44–58. [Google Scholar]
- Arrieta, A.B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; García, S.; Gil-López, S.; Molina, D.; Benjamins, R.; et al. Explainable artificial intelligence (xai): Concepts, taxonomies, opportunities and challenges toward responsible ai. Inf. Fusion 2020, 58, 115. [Google Scholar] [CrossRef]
- Hossin, M.; Sulaiman, M.N. A review on evaluation metrics for data classification evaluations. Int. J. Data Min. Knowl. Manag. Process 2015, 5, 1. [Google Scholar]
- van Rijsbergen, C.J.; Lalmas, M. Information calculus for information retrieval. J. Am. Soc. Inf. Sci. 1996, 47, 385–398. [Google Scholar] [CrossRef]
- Powers, D.M. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar]
Ref. | Methodology | Dataset or Application Context | Major Findings | Noted Limitations | Distinctive Contribution Relative to the Present Work |
---|---|---|---|---|---|
[35] | Ensemble CNNs (AlexNet, GoogLeNet, VGG16, ResNet-50) + Transfer Learning | Mammograms | Confidence: 0.9305 (microcalcifications), 0.8859 (normal) | Focus on microcalcifications; no multimodal input | Improved microcalcification detection using ensemble learning |
[36] | DenseNet with attention + Multi-level Transfer Learning | Augmented mammogram dataset | Accuracy > 84% | Performance dependent on attention layers | Robust classification via attention-enhanced DenseNet |
[37] | Few-Shot Learning (FSL) + Transfer Learning | Limited annotated mammograms | Accuracy: 95.6%, AUC: 0.970 | Specialized FSL models may not generalize broadly | Combined FSL with TL for better generalization in low-data settings |
[38] | Ensemble CNNs + Attention-Guided Grad-CAM | Infrared thermograms (DMR) | Accuracy: 98.04%, AUC: 0.97 | Domain-specific (infrared imaging only) | Enhanced interpretability using channel and spatial attention |
[39] | CNN + IoT Integration | Real-time breast cancer detection | Accuracy: 95% | Not evaluated on public benchmarks | Integrated IoT with CNN for scalable diagnostics |
[40] | LASSO + RF, KNN, MLP | Breast cancer classification | RF: Accuracy 90.68%, F1: 94.60%; KNN Recall: 98.80% | No DL used | Demonstrated impact of LASSO FS on ML model accuracy |
[41] | RF, AdaBoost, GNB, Logistic Regression + Grid Search | BC classification | RF: 99%, GNB: 91% | No XAI or interpretability included | Performance benchmarking across traditional ML algorithms |
[42] | XGBoost, RF, SVM, GBDT, Logistic Regression | Distant metastasis prediction | Accuracy: 93.6%, F1: 88.9%, AUC: 91.3% | Limited to metastasis focus | Ensemble model for metastasis risk prediction |
[43] | SSA + TLBO + ANN | 651 mammograms, UCI dataset | Accuracy: 98.46%, Sensitivity: 98.81%, AUC: 0.997 | Dependent on heuristic FS quality | Hybrid FS using SSA + TLBO for enhanced selection |
[44] | BinJOA-S (Binary Jaya Optimization with scalarization) | Breast cancer datasets | High classification accuracy and fitness score | Algorithm complexity not discussed | Multi-objective FS via scalarization with optimization |
[45] | Snake-spline segmentation + Flower Pollination Optimization | CT lung cancer dataset | Accuracy: 84% using 33 features | Focused on lung CT, not BC | Hybrid segmentation and FS using FPO for SVM |
[46] | Bacterial Foraging + Emperor Penguin Optimization (EPO), hGSAEPO | WDBC BC dataset | Accuracy: 98.31%, AUC > 0.998 | Computational complexity of hybrids | High-performing hybrid FS + XAI integration |
[47] | Liver Cancer Algorithm (LCA) + SVM optimization | MonoAmine Oxidase (MAO) dataset | Accuracy: 98.704% | Not breast cancer-specific | Novel evolutionary optimization inspired by tumor dynamics |
[48] | Wrapper-based FS + RF, SVM, DT | Early BC biomarker detection | AUC: 0.89–0.98, Sensitivity: 0.94, Specificity: 0.90 | Requires feature engineering | Sequential backward FS enhancing early biomarker identification |
[49] | DeepLabV3+ + Autoencoder + Grad-CAM + GLCM | Breast ultrasound datasets | Accuracy: 97.4%, Dice: 0.981 | Specific to ultrasound; not multimodal | Strong segmentation and interpretability using XAI |
[51] | Random Forest + SHAP (XAI) + Clinical features | NSLN metastases prediction | Accuracy: 72.2%, AUC: 0.77 | Limited dataset (183 cases) | Interpretable model using SHAP for clinical guidance |
[52] | CNN (VGG16, VGG19, ResNet) + XAI (LIME, SHAP, Saliency Maps) | Histopathology images | Accuracy: 92.59% (VGG19) | Downsampled input, lacks real-time validation | Evaluation of XAI methods for clinical interpretability |
[53] | Patho-Net (GRU + U-Net + GLCM + XAI) | BreakHis 100X histopathology | Accuracy: 98.90% | Focused on histopathology only | End-to-end DL with built-in interpretability and preprocessing |
[54] | CNNs + LIME, Grad-CAM, Grad-CAM++ | Mammography images | DenseNet201: Accuracy 99% | Limited multiclass analysis | Comprehensive use of multiple XAI methods |
[55] | 3D tomosynthesis + Grad-CAM + LIME + t-SNE + UMAP | Lesion-region images | AUC: 98.2% | High-dimensional visualization needed | Structural + visual interpretability in multiclass BC diagnosis |
[56] | Condensed CNN | MIAS, DDSM, Combined | Accuracy: 96.67% (combined dataset) | Limited XAI integration | Efficient CNN optimized for performance and speed |
[57] | ResNet50 + Improved Marine Predators Algorithm (IMPA) | MIAS, CBIS-DDSM | Accuracy: 98.88% | Dataset-specific performance | Metaheuristic TL optimization with opposition learning |
Model | Accuracy | Precision | Recall | F1-Score |
---|---|---|---|---|
MobileNet | 0.937587 | 0.937604 | 0.937587 | 0.937594 |
Xception | 0.889043 | 0.889281 | 0.889043 | 0.888823 |
DenseNet | 0.884882 | 0.885869 | 0.884882 | 0.884459 |
CNN | 0.610264 | 0.610927 | 0.610264 | 0.610538 |
EfficientNet | 0.540915 | 0.292589 | 0.540915 | 0.379761 |
VGG16 | 0.540915 | 0.292589 | 0.540915 | 0.379761 |
ResNet50 | 0.540915 | 0.292589 | 0.540915 | 0.379761 |
Class | Precision | Recall | F1-Score |
---|---|---|---|
Benign | 0.94 | 0.95 | 0.95 |
Malignant | 0.94 | 0.93 | 0.94 |
Overall | 0.94 | 0.94 | 0.94 |
Accuracy | 0.9417 |
Class | Precision | Recall | F1-Score |
---|---|---|---|
Benign | 0.98 | 0.98 | 0.98 |
Malignant | 0.98 | 0.98 | 0.98 |
Overall | 0.98 | 0.98 | 0.98 |
Accuracy | 0.9806 |
Model | Accuracy | Precision | Recall | F1-Score |
---|---|---|---|---|
MobileNet | 0.935361 | 0.903614 | 0.892857 | 0.898204 |
DenseNet | 0.927757 | 0.922078 | 0.845238 | 0.881988 |
Xception | 0.904943 | 0.904110 | 0.785714 | 0.840764 |
VGG16 | 0.904943 | 0.855422 | 0.845238 | 0.850299 |
ResNet50 | 0.794677 | 0.941176 | 0.380952 | 0.542373 |
CNN | 0.711027 | 0.633333 | 0.226190 | 0.333333 |
EfficientNet | 0.680608 | 0.340000 | 0.500000 | 0.680000 |
Class | Precision | Recall | F1-Score |
---|---|---|---|
Benign | 0.95 | 0.95 | 0.95 |
Malignant | 0.91 | 0.91 | 0.91 |
Overall | 0.93 | 0.93 | 0.93 |
Accuracy | 0.9422 |
Class | Precision | Recall | F1-Score |
---|---|---|---|
Benign | 0.96 | 0.97 | 0.97 |
Malignant | 0.94 | 0.92 | 0.93 |
Overall | 0.95 | 0.94 | 0.95 |
Accuracy | 0.9544 |
Model | Accuracy | Precision | Recall | F1-Score |
---|---|---|---|---|
MobileNet | 0.9195 | 0.92 | 0.92 | 0.92 |
DenseNet | 0.8591 | 0.86 | 0.85 | 0.86 |
Xception | 0.8541 | 0.85 | 0.85 | 0.85 |
VGG16 | 0.7522 | 0.75 | 0.75 | 0.75 |
ResNet50 | 0.6138 | 0.62 | 0.58 | 0.56 |
CNN | 0.5849 | 0.59 | 0.58 | 0.59 |
EfficientNet | 0.5635 | 0.32 | 0.56 | 0.41 |
Class | Precision | Recall | F1-Score |
---|---|---|---|
Benign | 0.98 | 0.99 | 0.99 |
Malignant | 0.98 | 0.97 | 0.97 |
Overall | 0.98 | 0.98 | 0.98 |
Accuracy | 0.9896 |
Class | Precision | Recall | F1-Score |
---|---|---|---|
Benign | 0.99 | 0.99 | 0.98 |
Malignant | 0.97 | 0.97 | 0.97 |
Overall | 0.98 | 0.98 | 0.98 |
Accuracy | 0.9870 |
Model | Dataset | Accuracy | Precision | Recall | F1-Score |
---|---|---|---|---|---|
Optimized MobileNet FLA | MIAS | 0.9806 | 0.9800 | 0.9800 | 0.9800 |
Optimized MobileNet DOA | MIAS | 0.9417 | 0.9400 | 0.9400 | 0.9400 |
MobileNet | MIAS | 0.9376 | 0.9376 | 0.9376 | 0.9376 |
Optimized MobileNet FLA | Breast Ultrasound | 0.9544 | 0.9500 | 0.9500 | 0.9500 |
Optimized MobileNet DOA | Breast Ultrasound | 0.9422 | 0.9400 | 0.9400 | 0.9400 |
MobileNet | Breast Ultrasound | 0.9354 | 0.9036 | 0.8929 | 0.8982 |
Optimized MobileNet DOA | Fusion | 0.9896 | 0.9800 | 0.9800 | 0.9800 |
Optimized MobileNet FLA | Fusion | 0.9870 | 0.9800 | 0.9800 | 0.9800 |
MobileNet | Fusion | 0.9195 | 0.9800 | 0.9200 | 0.9200 |
References | Model/Method | Dataset | Results |
---|---|---|---|
Teoh et al., 2024 [34] | Ensemble (AlexNet, GoogLeNet, VGG16, ResNet-50), Transfer Learning | Mammograms | 93.05% (microcalcifications), 88.59% (normal) |
Sarwar et al., 2025 [36] | Few-Shot Learning + Transfer Learning (Relation Network) | Mammograms (Few-shot) | Accuracy: 95.6%, AUC: 0.970 |
Raghavan et al., 2024 [37] | Attention-Guided Grad-CAM+En semble CNNs | Infrared Thermograms (DMR) | Accuracy: 98.04%, AUC: 0.97 |
Alhsnony et al., 2024 [42] | Custom Simplified CNN | MIAS, DDSM, Fusion | Accuracy: 94.23%(MIAS), 95.53% (DDSM), 96.67% (Fusion) |
Houssein et al., 2022 [49] | ResNet50 + Improved Marine Predators Algorithm (IMPA) | MIAS, CBIS-DDSM | Accuracy: up to 98.88% |
Naas et al., 2025 [51] | DeepLabV3++ Autoencoder + Grad-CAM (XAI) | Breast Ultrasound | Accuracy: 97.4%, Dice: 0.981 |
Proposed Method | Optimized MobileNet + FLA | MIAS Dataset | Precision/Recall/F1: 0.98. Accuracy: 98.06% |
Proposed Method | Optimized MobileNet + FLA | Breast Ultrasound Images | Accuracy:95.44%, Precision/Recall/F1: 0.95 |
Proposed Method | Optimized MobileNet + DOA | Fusion Dataset | Accuracy: 98.96% Precision/Recall/F1: 0.98 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Mustafa, M.A.; Erdem, O.A.; Söğüt, E. Hybrid Optimization and Explainable Deep Learning for Breast Cancer Detection. Appl. Sci. 2025, 15, 8448. https://doi.org/10.3390/app15158448
Mustafa MA, Erdem OA, Söğüt E. Hybrid Optimization and Explainable Deep Learning for Breast Cancer Detection. Applied Sciences. 2025; 15(15):8448. https://doi.org/10.3390/app15158448
Chicago/Turabian StyleMustafa, Maral A., Osman Ayhan Erdem, and Esra Söğüt. 2025. "Hybrid Optimization and Explainable Deep Learning for Breast Cancer Detection" Applied Sciences 15, no. 15: 8448. https://doi.org/10.3390/app15158448
APA StyleMustafa, M. A., Erdem, O. A., & Söğüt, E. (2025). Hybrid Optimization and Explainable Deep Learning for Breast Cancer Detection. Applied Sciences, 15(15), 8448. https://doi.org/10.3390/app15158448