Next Article in Journal
Multi-Scale Feature Fusion GANomaly with Dilated Neighborhood Attention for Oil and Gas Pipeline Sound Anomaly Detection
Next Article in Special Issue
Early Heart Attack Detection Using Hybrid Deep Learning Techniques
Previous Article in Journal
Assessing the Impact of Prior Coding and Artificial Intelligence Learning on Non-Computing Majors’ Perception of AI in a University Context
Previous Article in Special Issue
A Maximum Value for the Kullback–Leibler Divergence between Quantized Distributions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

AI-Based Breast Cancer Detection System: Deep Learning and Machine Learning Approaches for Ultrasound Image Analysis

Department of Computer Science and Engineering, College of Engineering, Qatar University, Doha 2713, Qatar
*
Authors to whom correspondence should be addressed.
Information 2025, 16(4), 278; https://doi.org/10.3390/info16040278
Submission received: 24 February 2025 / Revised: 18 March 2025 / Accepted: 20 March 2025 / Published: 30 March 2025

Abstract

:
This work addresses the critical need for the early detection of breast cancer, a significant health concern worldwide. Using a combination of advanced deep learning and machine learning techniques, we offer a comprehensive solution to enhance breast cancer detection accuracy. By leveraging state-of-the-art convolutional neural networks (CNNs) like GoogLeNet, AlexNet, and ResNet18, alongside traditional classifiers such as k-nearest neighbors (KNN) and support vector machine (SVM), we ensure robust prediction capabilities. Our preprocessing methods significantly improve input data quality, leading to promising detection accuracies. For instance, ResNet-18 achieved impressive results, outperforming other models. Furthermore, our integration of these algorithms into a user-friendly MATLAB R2024b application ensures easy access for medical professionals, facilitating timely diagnosis and treatment. This work represents a vital step towards more effective breast cancer diagnosis, underscoring the importance of early intervention for improved patient outcomes.

Graphical Abstract

1. Introduction

Breast cancer is one of the most common forms of cancer worldwide, with millions of new cases diagnosed each year. It remains the second leading cause of cancer-related death among women globally, accounting for a significant burden on healthcare systems [1]. The early detection of breast cancer is crucial, as it greatly improves the chances of successful treatment and long-term survival. The progression of breast cancer typically follows a series of stages, from localized tumors to metastatic cancer, and early identification of these stages is key to determining appropriate treatment options. Despite advances in medical imaging, the accurate detection and classification of breast cancer remains challenging, especially when tumors are small or present in dense breast tissue, making it difficult for traditional imaging techniques, such as mammography, to detect them effectively [2]. Ultrasound imaging has become a vital tool in breast cancer diagnosis, offering advantages like real-time imaging, no radiation exposure, and effectiveness in imaging dense breast tissue [3]. However, the accuracy of ultrasound-based diagnoses depends heavily on the skill and experience of the clinician interpreting the images, which can lead to variability in diagnosis.
In this context, artificial intelligence (AI) and machine learning (ML) approaches offer promising solutions to improve the accuracy and efficiency of breast cancer detection. These technologies have demonstrated the potential to automate the analysis of ultrasound images, reduce diagnostic errors, and support clinicians in making timely and informed decisions. Our proposed solution builds on this potential by integrating deep learning and machine learning techniques to provide a reliable, accessible, and user-friendly tool for early breast cancer detection. Specifically, our solution aims to streamline the diagnostic process by introducing a computer-aided diagnostic (CAD) system that offers ease of access for specialists via a user interface to diagnose ultrasound breast images. The motivation behind our endeavor lies in the imperative to enhance healthcare outcomes for individuals at risk of breast cancer, as timely diagnosis can significantly impact treatment options and prognosis.
By integrating state-of-the-art convolutional neural networks (CNNs) such as GoogLeNet, AlexNet, and ResNet18 for deep learning-based prediction, complemented by traditional machine learning classifiers like k-nearest neighbors (KNN) and support vector machine (SVM), we aim to provide medical professionals with robust tools for early detection. To construct our solution, we collected baseline ultrasound images of breasts from 600 female patients aged from 25 to 75 years in 2018 [4]. Preprocessing steps are integral to our approach, as they enhance the quality of input data before feeding them into the models. Our final prediction models, implemented using deep and machine learning hybrid approaches, are seamlessly integrated into a MATLAB application, which is deployed as a standalone application supported on any Windows OS for ease of use by medical practitioners. The evaluation of our prediction models demonstrates promising results, with high accuracy achieved using both deep learning and machine learning approaches. These findings underscore the potential of our comprehensive solution to significantly improve early detection rates and ultimately enhance patient outcomes in the battle against breast cancer.

2. Related Work

In this section, we explore previous research efforts in the field of breast cancer detection, focusing on the methodologies and techniques employed for early diagnosis and classification. We analyze the advancements made in medical imaging using preprocessing steps, machine and deep learning algorithms, computational approaches, and results. In [5], the researchers fine-tuned the hyperspectral imaging (HSI) dataset for detecting breast cancer. This dataset is structured as image cubes, where each pixel contains detailed spectral information across numerous narrow spectral bands. Moreover, they utilized a hyperspectral camera to capture images for their dataset, which compromises both near-infrared and visible ranges. For preprocessing, the Lee filter was employed to address speckle noise in the hyperspectral breast cancer detection dataset. This filter is a local adaptive filter that replaces each pixel value with a weighted average of neighboring pixels. The classification method used in the study is the fuzzy c-means clustering technique. It is a variation of the conventional k-means clustering method that allows for the soft assignment of data points to clusters, where each data point has a degree of membership in each cluster rather than a binary assignment. This method was applied to segment cancerous regions in breast tissue samples based on their spectral characteristics. In the end, they achieved promising results with a sensitivity of 98.23% and 95.42%, specificity of 94.7% and 92.09%, and accuracy of 97.03% and 94.29% on two different samples. In [6], the authors demonstrated that deep learning and CNNs can overshadow the features of the region of interest with global features from the entire image. To address this issue, they introduced BCHisto-Net, a method designed to capture both local and global features focusing on the region of interest (ROI). Their solution comprises two main steps. The first step involves color normalization, where histopathological images undergo color normalization using the Reinhard method [7]. This process aligns color distributions across images, ensuring consistent representation for accurate analysis using BCHisto-Net. The second step introduces BCHisto-Net, which utilizes two branches: one for global features and another for local features [8]. Global features are derived from a color-normalized image, while local features begin with Otsu’s thresholding method to identify regions of interest (ROIs), followed by Gaussian smoothing for improved feature extraction [8]. Within this second step, the BCHisto-Net architecture incorporates parallel global and local feature extractors, each employing an inception module for feature extraction [9]. The inception module utilizes convolution operations with varying receptive fields, concatenated along the channel axis to effectively capture low-level features. The global branch processes the color-normalized image, while the local branch employs Otsu’s thresholding for ROI detection and Gaussian smoothing. Subsequent feature extraction involves five convolution blocks with dropout layers to prevent overfitting. The global and local features are aggregated and passed through fully connected layers for classification. Parameter optimization involves BCE loss for each branch independently, ensuring different weights for distinct feature extraction, culminating in a holistic training approach. BCHisto-Net achieved high accuracy rates of 95% and 89% on the KMC and BreakHis datasets, respectively, demonstrating its effectiveness in classifying breast histopathological images at 100× magnification. In the third related work [10], recent strides in computational breast cancer diagnosis are highlighted using innovative applications of neural networks. A notable contribution in this domain involves the development of a feed-forward backpropagation neural network (FNN) specifically tailored for analyzing Hematoxylin and Eosin (H&E)-stained histopathological images. This method adeptly employs digital image processing techniques to extract eight distinct features from a substantial dataset comprising over 2600 breast biopsy images, with a primary focus on discerning cell nuclei characteristics. Within this framework, the FNN architecture is structured with a single hidden layer housing 20 neurons, alongside 4 output neurons tasked with distinguishing between benign and malignant tumors. Furthermore, it effectively categorizes malignant tumors into three specific types: Type1, Type2, and Type3. Notably, the study reports impressive diagnostic accuracy, with a precision of 96.07% in discriminating tumor types and an overall classification accuracy of 95.80% during testing phases. These findings underscore the potential of neural networks as adjunct tools for pathologists, enhancing diagnostic precision and pinpointing cases necessitating further investigation. This research aligns with broader trends in medical image processing, where machine learning models play a pivotal role in refining diagnostic workflows in clinical settings. The noteworthy high accuracy rates and the nuanced approach to the multi-classification of malignant tumors highlight the distinct contributions of this study within the realm of automated breast cancer detection systems. In the fourth related work [11], the researchers delve into the domain of medical image analysis, with a specific focus on the classification of breast cancer utilizing deep convolutional neural networks (CNNs). Their study centers on the adaptation of the DenseNet model, recognized for its proficiency in handling intricate image processing tasks, to meet the demands of breast cancer diagnosis. Using this adaptation, they achieve a notable 96% accuracy rate in multi-class classification, surpassing human expert performance. This success underscores the transformative potential of deep learning algorithms in augmenting diagnostic capabilities, particularly in critical areas such as cancer detection, where accuracy is paramount. Furthermore, the findings emphasize the significance of data availability in refining and bolstering the efficacy of deep learning models, suggesting that, with access to larger datasets, further performance enhancements are attainable. Ultimately, the research reaffirms the utility of CNNs in medical image analysis and highlights their potential to revolutionize diagnostic paradigms, offering unprecedented accuracy and efficiency in identifying and classifying complex medical conditions like breast cancer.

3. Proposed Solution

3.1. Preprocessing

In the preprocessing pipeline for our dataset, we begin by reading the dataset, which comprises 437 benign, 210 malignant, and 133 normal RGB images. Following by applying data augmentation techniques, we expanded our dataset from 647 to 17,480 images, augmenting the original images to enrich our dataset. Each image was then converted form unsigned 8-bit integers (unit8 format) to double format to simplify the processing and reduce computational complexity. Moving forward, we implemented a series of image enhancement steps. Firstly, each image underwent a median filter applied separately to each one of them in order to reduce noise. Next, we applied a high-pass filter to enhance image edges and fine details. This involves subtracting the original image from a Gaussian-filtered version of the image with a standard deviation (std) of 20, resulting in edges of the image; later, we added this high-pass filter to the original image, resulting in sharper image. After these enhancements, we resized the images to a standard size of 224 × 224 pixels. Finally, as a last step before classification, the images were converted to unsigned 8-bit integers (unit8 format) to ensure compatibility with subsequent processing steps. These preprocessing steps collectively aim to improve image quality, enhance features, and prepare the dataset for effective classification analysis.

3.2. Deep Learning Approach

Our innovative solution to breast cancer detection integrates cutting-edge deep learning techniques with traditional machine learning methodologies to create a robust and accurate diagnostic tool. In our project, we strategically employed three cutting-edge convolutional neural networks (CNNs): GoogLeNet [9,12], ResNet18 [13], and AlexNet [14]. These CNNs were meticulously selected for their track record of delivering exceptional accuracy and efficiency in handling complex image data. GoogLeNet, with its impressive 144 layers and an input image size of 224 × 224 × 3, stands out for its hallmark inception modules that efficiently capture multi-scale features, making it ideal for scenarios with constrained computational resources but demanding deep networks. To harness GoogLeNet’s power effectively, we replaced its original fully connected layer and classification layer with a new fully connected and classification layers tailored for our medical images. Additionally, we set the weight learn rate factor and bias learn rate factor to 10, ensuring optimal adaptation to our dataset. ResNet18, boasting 71 layers and sharing the same input image size as GoogLeNet, is renowned for its deep structure and ingenious use of residual connections, effectively tackling the vanishing gradient problem. Following the same strategy as with GoogLeNet, we replaced its fully connected and classification layers with new counterparts, maintaining consistency with the weight learn rate factor and bias learn rate. AlexNet, comprising 25 layers and an input image size of 227 × 227 × 3, underwent a similar transformation. We replaced its last three layers—fully connected, SoftMax, and classification—with new ones, each endowed with a weight learn rate factor and bias learn rate of 20, providing enhanced adaptability to our dataset. For the training regimen, we meticulously selected optimization algorithms tailored to each network’s characteristics. GoogLeNet and ResNet18 benefited from the Adam optimizer [15], while AlexNet utilized RMSProp [16]. The mini-batch size was set to 10 to balance computational efficiency and training stability, ensuring that each iteration processed a manageable subset of the data. The maximum number of training epochs was capped at 50, as this struck a balance between training efficiency and convergence. Shuffling the training data at the start of each epoch maintained robustness and ensured unbiased model training. Moreover, setting the learning rate to 0.0001 fine-tuned model performance supported with learning rate drop factor of 0.5 after each 10 epochs to avoid overfitting, optimizing prediction accuracy across all networks. Post-training, we leveraged the confusion matrix to meticulously assess model performance, computing accuracy, prediction, recall, and F1 score percentages. This comprehensive evaluation methodology provided invaluable insights into each model’s effectiveness, guiding iterative improvements, and ensuring robust performance in real-world scenarios.

3.3. Machine Learning Approch

In the machine learning approach, the activations function is used as a feature extractor. This function processes the dataset alongside a trained CNN model and a specific layer from which features are to be extracted. In our case, we used the ResNet-18 CNN model and selected the 2D average pooling layer, referred to as ‘pool5’ for feature extraction. This layer was chosen because it provided the highest accuracy percentage. The function’s output consists of rows, where each row represents the extracted features of a single image. Each image is represented by 512 columns, corresponding to the number of features per image. Moving on with the utilized classifiers, starting with K-Nearest Neighbors (KNN) which is the widely used classification techniques for medical images. KNN operates on the principle of similarity, where it assigns a class label to a new data point based on the majority class label of its nearest neighbors in the training [17]. Expanding on this, KNN considers the distance between data points using various distance metrics. One such metric is the distance weight of the squared inverse, where the inverse of distances is weighted, giving more importance to closer neighbors. This metric is effective in situations where nearby points have higher relevance for classification [18]. Additionally, KNN can utilize the cosine distance metric, which measures the cosine of the angle between two vectors. This metric is particularly useful for text data or high-dimensional sparse data where the magnitude of vectors is not as critical as their orientation [19]. The cosine distance metric is ideal for analyzing ultrasound images in breast cancer detection. In these images, where features are often sparse and high-dimensional, the angle between vectors is more crucial than their magnitudes. This metric focuses on relative orientations, aiding in the accurate identification of cancerous patterns based on feature alignments within the image. In summary, KNN with the incorporation of distance weight of squared inverse and cosine distance metrics offers a flexible approach to breast cancer classification tasks, leveraging different distance calculations based on the characteristics of the data. SVM Linear is a machine learning algorithm used for classification tasks, particularly effective when dealing with linearly separable datasets. It works by finding the optimal hyperplane that best separates data points of different classes in a high-dimensional space [20]. This hyperplane maximizes the margin, which is the distance between the closest data points from different classes, leading to the better generalization and robustness of the classifier. SVM Linear achieves this separation by using a linear kernel function, which calculates the dot product of feature vectors in the input space. The decision boundary is then determined based on the sign of this dot product, assigning data points to different classes accordingly. One of SVM Linear’s strengths is its efficiency in handling large-scale datasets with a linear separation boundary. It is computationally efficient and memory-friendly compared to non-linear SVM kernels, making it suitable for applications with a large number of features and data points.

3.4. User Interface

We implemented our project with a user-friendly interface by developing an application utilizing MATLAB. We ensured the integration of our highest accuracy models into the final version of the Breast Cancer Detection Application that was developed in collaboration between all authors. For our deep learning model, we have chosen ResNet-18 CNN, while for the machine learning approach, we opted for the KNN algorithm with 1 neighbor due to its superior accuracy percentage. In our application, we ensured that it does not require MATLAB to be operational. Instead, we managed to deploy the application as an executable file (.exe) that can run on any device supporting the Windows operating system. This deployment implementation simplifies the application’s usability and removes the limitation of MATLAB installation. Within our application, users have the option to select any ultrasound image of a breast cancer case. Once the user has chosen the image, it will be displayed on the application user interface. Subsequently, the user can preprocess the image for either the deep learning or machine learning approach. Upon pressing the preprocessing button, the preprocessed image will be displayed on the user interface. The final step involves pressing the button designated for classifying the image, whether using the deep learning or machine learning approach. Upon pressing the classify button, the classification result will be displayed in the corresponding text field, indicating whether the image is normal or abnormal, and the abnormal image can be classified further as benign or malignant. With these steps, we have developed a reliable application that can be utilized by any health institution.
The proposed CAD system is designed to assist in breast cancer classification through deep and machine learning techniques. Figure 1 presents the high-level architecture of the system. In this process, a doctor selects an ultrasound image, which is then preprocessed to enhance its suitability for analysis using learning models. The image is subsequently classified as either normal, abnormal, benign, or malignant based on the trained model. To support classification, the system utilizes a dataset containing breast ultrasound images categorized into three primary classes: normal, benign, and malignant. Figure 2 illustrates sample images from each of these categories, highlighting the visual differences that the models are trained to detect. A user-friendly MATLAB-based interface was developed to deploy the models and streamline the diagnostic process. Figure 3a shows an example of a benign case being classified using the application, while Figure 3b displays a malignant case. The UI allows users to select and preprocess images, then classify them using either deep learning or machine learning, with the results displayed clearly within the application.

4. Simulation Results

For our breast cancer detection project, we leveraged the “Breast Ultrasound Images Dataset”, released in 2018 [4]. This dataset is specifically designed for advanced medical image analysis tasks, such as semantic segmentation and classification, and is a crucial resource for enhancing the accuracy of breast cancer diagnosis. The dataset consists of 780 ultrasound images, annotated at the pixel level to support semantic segmentation tasks. These annotations include 780 objects across three primary categories: benign, malignant, and normal. We have encountered all of these annotated categories to study pathological findings. We have noticed that the dataset contained some duplicated images in which it would affect the model’s accuracy, so all of the duplicated images were deleted. To ensure a rigorous evaluation of our models, we divided the dataset into training, validation, and testing sets. Approximately 60% of the images were allocated for training, 10% for validation, and the remaining 30% were allocated for testing. This distribution was designed to provide a substantial training dataset and adequate resources for effective model validation and testing. Our project’s primary challenge was to develop robust models capable of performing classification tasks accurately. We employed state-of-the-art convolutional neural networks (CNNs) such as GoogLeNet, AlexNet, and ResNet18 for training our models. These deep learning models are critical for identifying complex patterns indicative of benign and malignant conditions. The utilized hardware specifications used in implementing this project are as follows:: AMD EPYC 74F3 24-Core Processor, NVIDIA A10-24Q GPU, and 440 GB RAM. The results of our study are highly promising, indicating that the integration of classification techniques considerably enhances the accuracy of breast cancer detection.
Table 1 shows the highest performance metrics which underscore the effectiveness of our approach, combining detailed image analysis with powerful machine learning and deep learning algorithms.
Additionally, the inclusion of these models in the MATLAB application provides a user-friendly tool for medical professionals, facilitating early diagnosis and improving the likelihood of successful patient outcomes.

5. Discussion

Our study investigates the effectiveness of both machine learning and deep learning techniques in detecting breast cancer from ultrasound images. We employed Support Vector Machine (SVM) and K-Nearest Neighbors (KNN) as representative machine learning methods, alongside deep learning models such as GoogLeNet, AlexNet, and ResNet18. Our results indicate that both SVM and KNN achieved high accuracy in distinguishing between all four classes, which are normal, abnormal, benign, and malignant, with SVM scoring 99.72% accuracy for benign and malignant classes and 97.65% for distinguishing between normal and abnormal classes. These findings align with previous research, showcasing SVM’s strength in classification tasks, especially when using a linear kernel, which we found to be the most effective for our dataset.
Table 2 presents the performance of SVM classifiers with different kernel integrations for the four classes: benign, malignant, normal, and abnormal.
Apart from the Support Vector Machine (SVM) model, another commonly used classification method is the K-Nearest Neighbors (KNN) algorithm. The choice of the parameter k, representing the number of neighbors considered, greatly influences KNN’s performance. In our research, we assessed how different values of k affected the KNN algorithm’s performance, varying from 1 to 5. Typically, smaller k values make the model more sensitive to data noise and outliers, possibly causing overfitting; however, in our proposed solution, the k value was 1, since the dataset split was clear to predict by the trained models. Hence, selecting the optimal k value is crucial for achieving optimal results. After extensive evaluation, we found that setting the value k to 1 resulted in the highest accuracy of 99.72% for benign and malignant, while 98.63% for normal and abnormal. This suggests that considering only the closest neighbor for classification yields the most accurate results for our dataset. Furthermore, other performance metrics such as precision, recall, and F1 score also showed favorable outcomes for k = 1, with all metrics exceeding 98%. Our results echo previous research utilizing machine learning algorithms for similar tasks, affirming the efficacy of KNN with a small k value for detecting breast cancer from ultrasound images. By refining parameter selection, we aim to improve the accuracy and reliability of our model, contributing to advancements in early breast cancer detection techniques.
Table 3 shows the performance of KNN classifiers with different selections of the number of neighbors for classifying the four classes: benign, malignant, normal, and abnormal.
In parallel with our exploration of machine learning techniques, we investigated the capabilities of deep learning models, focusing on evaluating the efficacy of GoogLeNet, ResNet18, and AlexNet for breast cancer detection from ultrasound images. Notably, ResNet-18 demonstrated the highest accuracy of 99.72% for benign and malignant classes and 99.66% for normal and abnormal classes, indicating its proficiency in efficiently capturing multi-scale features. While AlexNet achieved slightly lower accuracies at 98.27% for benign and malignant classes and 98.69% for normal and abnormal classes, however, AlexNet’s performance remained robust across precision, recall, and F1 scores.
Table 4 summarizes the classification performance of CNN models across the four classes: benign, malignant, normal, and abnormal.
These results underscore the resilience of deep learning techniques in accurately classifying ultrasound images for breast cancer detection. By leveraging complex neural architectures and effectively learning intricate patterns from medical data, these models show promise for enhancing diagnostic capabilities in healthcare settings. However, further exploration and validation across diverse datasets are imperative to establish their generalizability and scalability in real-world clinical applications. Nonetheless, the demonstrated effectiveness of deep learning models in our study highlights their potential to significantly impact early breast cancer detection and improve patient outcomes.
Table 5 provides a comprehensive summary of related work studies in comparison to the proposed solution.
Our study demonstrates the superior performance of deep learning models in breast cancer detection from ultrasound images compared to existing research. In contrast to the first related work utilizing hyperspectral imaging [5], where accuracies ranged around 95–98%, our deep learning models consistently exceeded 99% accuracy specifically for GoogLeNet and Resnet 18 CNNs, showcasing substantial advancements in diagnostic capabilities. Similarly, when compared to the second related work focusing on histopathological images [6], our research achieved remarkable accuracies exceeding 99% across various machine learning and deep learning models, surpassing the accuracies achieved by BCHisto-Net [6]. Additionally, the third related work employs a feed-forward backpropagation neural network (FNN) for analyzing histopathological images [10], achieving accuracies of 96.07% in distinguishing tumor types and an overall classification accuracy of 95.80%; on the contrary, our machine learning approach achieved 99.72% for benign and malignant classes and 98.63% for normal and abnormal classes, while in the deep learning approach, we achieved 99.72% for benign and malignant classes and 99.66% for normal and abnormal classes. Despite the differences in methodologies and imaging modalities, all studies underscored the efficacy of deep and machine learning techniques in improving breast cancer detection, signaling significant progress in diagnostic capabilities across varied image types.

6. Conclusions

In conclusion, this project demonstrates a substantial advancement in the field of medical diagnostics, specifically in the early detection of breast cancer. By effectively combining cutting-edge deep learning models such as GoogLeNet, AlexNet, and ResNet18 with traditional machine learning classifiers like KNN and SVM, we have developed a robust diagnostic tool that enhances the accuracy of breast cancer detection. Our preprocessing techniques ensure the quality of the input data, which is crucial for achieving high detection accuracy. The notable performance of ResNet-18 highlights the effectiveness of our chosen methodologies. Additionally, the user-friendly MATLAB application we developed simplifies the use of our advanced models in clinical field, enabling medical professionals to make timely and accurate diagnoses, which are essential for optimal patient care. The promising results obtained from the evaluation of our models reaffirm the potential of integrated deep and machine learning solutions to significantly improve early detection rates, thereby having a profound impact on patient outcomes and advancing the fight against breast cancer. This project not only contributes to technological advancements in healthcare, but also underscores the critical importance of early cancer detection and the potential to save lives through innovation and accessibility.

Author Contributions

Conceptualization, A.M., A.A. and U.Q.; Methodology, A.M., A.A. and U.Q.; Software, A.M., A.A. and U.Q.; Validation, A.M., A.A. and U.Q.; Formal analysis, A.M., A.A. and U.Q.; Investigation, A.M., A.A. and U.Q.; Resources, A.M., A.A. and U.Q.; Data curation, A.M., A.A. and U.Q.; Writing—original draft, A.M., A.A. and U.Q.; Writing—review & editing, A.M., A.A. and U.Q.; Visualization, A.M., A.A. and U.Q.; Supervision, A.M., A.A. and U.Q.; Project administration, A.M., A.A. and U.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are openly available in the dataset published by Al-Dhabyani et al. (2020) [4] and can be accessed at https://doi.org/10.1016/j.dib.2019.104863.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ferlay, J.; Ervik, M.; Lam, F.; Colombet, M.; Mery, L.; Piñeros, M.; Znaor, A.; Soerjomataram, I.; Bray, F. Global Cancer Observatory: Cancer Today; International Agency for Research on Cancer: Lyon, France, 2020. [Google Scholar]
  2. Boyd, N.F.; Guo, H.; Martin, L.J.; Sun, L.; Stone, J.; Fishell, E.; Jong, R.A.; Hislop, G.; Chiarelli, A.; Minkin, S.; et al. Mammographic density and the risk and detection of breast cancer. N. Engl. J. Med. 2007, 356, 227–236. [Google Scholar] [CrossRef] [PubMed]
  3. American Cancer Society. Breast Ultrasound. 2021. Available online: https://www.cancer.org/cancer/breast-cancer/screening-tests-and-early-detection/breast-ultrasound.html (accessed on 5 March 2025).
  4. Al-Dhabyani, W.; Gomaa, M.; Khaled, H.; Fahmy, A. Dataset of breast ultrasound images. Data Brief 2020, 28, 104863. [Google Scholar] [CrossRef] [PubMed]
  5. Youssef, A.; Moa, B.; El-Sharkawy, Y.H. A Novel Visible and Near-Infrared Hyperspectral Imaging Platform for Automated Breast-Cancer Detection. Photodiagn. Photodyn. Ther. 2024, 46, 104048. Available online: https://www.sciencedirect.com/science/article/pii/S1572100024000875 (accessed on 5 March 2025). [CrossRef] [PubMed]
  6. Rashmi, R.; Prasad, K.; Udupa, C.B.K. BCHisto-Net: Breast histopathological image classification by global and local feature aggregation. Artif. Intell. Med. 2021, 121, 102191. Available online: https://www.sciencedirect.com/science/article/pii/S0933365721001846#s0025 (accessed on 5 March 2025).
  7. Reinhard, E.; Adhikhmin, M.; Gooch, B.; Shirley, P. Color transfer between images. IEEE Comput. Graph. Appl. 2001, 21, 34–41. [Google Scholar] [CrossRef]
  8. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  9. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  10. Singh, S.; Gupta, P.R.; Sharma, M.K. Breast cancer detection and classification of histopathological images. Int. J. Eng. Sci. Technol. 2010, 3, 4228. [Google Scholar]
  11. Nawaz, M.A.A.; Hassan, T. Multi-Class Breast Cancer Classification using Deep Learning Convolutional Neural Network. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 316–332. [Google Scholar] [CrossRef]
  12. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  13. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  14. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the 26th Annual Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 25. [Google Scholar] [CrossRef]
  15. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  16. Graves, A. Generating Sequences with Recurrent Neural Networks. arXiv 2013, arXiv:1308.0850. [Google Scholar]
  17. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  18. Li, S.; Jain, A.K. Handbook of Face Recognition; Springer Science & Business Media: New York, NY, USA, 2007. [Google Scholar]
  19. Manning, C.D.; Raghavan, P.; Schütze, H. Introduction to Information Retrieval; Cambridge University Press: Cambridge, UK, 2008. [Google Scholar]
  20. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar]
Figure 1. Represents the high-level architecture of the system.
Figure 1. Represents the high-level architecture of the system.
Information 16 00278 g001
Figure 2. Shows sample images from the three classes: normal, benign, and malignant.
Figure 2. Shows sample images from the three classes: normal, benign, and malignant.
Information 16 00278 g002
Figure 3. User interface: (a) displays benign case example from user application; (b) displays malignant case example from user application.
Figure 3. User interface: (a) displays benign case example from user application; (b) displays malignant case example from user application.
Information 16 00278 g003aInformation 16 00278 g003b
Table 1. Highest model performances.
Table 1. Highest model performances.
Model Performance Benign/Malignant
ModelAccuracyPrecisionRecallF1_Score
KNN99.72%99.74%99.63%99.68%
SVM 99.72%99.74%99.63%99.68%
ResNet-1899.72%99.74%99.63%99.68%
Model Performance Normal/Abnormal
ModelAccuracyPrecisionRecallF1_Score
KNN 98.63%96.60%98.78%97.65%
SVM 97.65%95.21%96.74%95.95%
ResNet-1899.66%99.27%99.53%99.40%
Table 2. SVM performance.
Table 2. SVM performance.
SVM Performance for Different Kernels Benign/Malignant
AccuracyPrecisionRecallF1_Score
Gaussian98.55%98.95%97.75%98.32%
Polynomial99.72%99.74%99.63%99.68%
Linear99.72%99.74%99.63%99.68%
SVM Performance for Different Kernels Normal/Abnormal
AccuracyPrecisionRecallF1_Score
Gaussian97.31%95.13%95.48%95.30%
Polynomial66.00%64.37%75.12%61.15%
Linear97.65%95.21%96.74%95.95%
Table 3. KNN performance.
Table 3. KNN performance.
KNN Accuracy for Different Values of K Benign/Malignant
KAccuracy
199.72%
299.72%
399.72%
499.72%
599.72%
KNN Accuracy for Different Values of K Normal/Abnormal
KAccuracy
198.63%
298.63%
398.11%
498.05%
597.82%
Table 4. CNN performance.
Table 4. CNN performance.
CNN Deep Learning Models Benign/Malignant
CNNAccuracyPrecisionRecallF1_Score
GoogLeNet99.65%99.58%99.63%99.60%
Resnet1899.72%99.74%99.63%99.68%
AlexNet98.27%97.80%98.27%98.03%
CNN Deep Learning Models Normal/Abnormal
CNNAccuracyPrecisionRecallF1_Score
GoogLeNet99.60%99.62%98.97%99.29%
Resnet1899.66%99.27%99.53%99.40%
AlexNet98.69%98.27%98.75%98.50%
Table 5. Proposed solution vs. related work results.
Table 5. Proposed solution vs. related work results.
Paper NumberDataset TypeAccuracy
Proposed solution (Resnet 18)Ultrasound Images99.72% for benign/malignant
99.66% for normal/abnormal
Paper [5]Hyperspectral Imaging (HSI) dataset (near-infrared and visible)97.03% and 94.29% (on two different samples)
Paper [6]KMC and BreakHis (histopathological images)95% (KMC dataset) and 89% (BreakHis dataset)
Paper [10]H&E-stained histopathological images (over 2600 breast biopsies)95.80% (overall classification) and 96.07% (tumor-type level)
Paper [11]DenseNet-based approach on breast histopathological images96% (multi-class classification)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Moursi, A.; Aboumadi, A.; Qidwai, U. AI-Based Breast Cancer Detection System: Deep Learning and Machine Learning Approaches for Ultrasound Image Analysis. Information 2025, 16, 278. https://doi.org/10.3390/info16040278

AMA Style

Moursi A, Aboumadi A, Qidwai U. AI-Based Breast Cancer Detection System: Deep Learning and Machine Learning Approaches for Ultrasound Image Analysis. Information. 2025; 16(4):278. https://doi.org/10.3390/info16040278

Chicago/Turabian Style

Moursi, Amro, Abdulrahman Aboumadi, and Uvais Qidwai. 2025. "AI-Based Breast Cancer Detection System: Deep Learning and Machine Learning Approaches for Ultrasound Image Analysis" Information 16, no. 4: 278. https://doi.org/10.3390/info16040278

APA Style

Moursi, A., Aboumadi, A., & Qidwai, U. (2025). AI-Based Breast Cancer Detection System: Deep Learning and Machine Learning Approaches for Ultrasound Image Analysis. Information, 16(4), 278. https://doi.org/10.3390/info16040278

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop