From previous research, several studies have been performed to develop a deep CNN architectures for the classification of both natural and medical images. A few deep-learning-based methods for classifying breast cancer masses in mammography pictures have been proposed. In [
9], authors created a deep-belief-network (DBN)-based approach for determining whether mammography images are normal or abnormal. A discrete wavelet transform was used to extract image features, specifically the gray level co-occurrence matrix features from the HL and LL wavelet sub-bands. The authors showed that a deep belief network (DBN) CAD system could enable automated hierarchical feature extraction to offer more flexibility for intricate design patterns. Yet, there are several obstacles to overcome in this approach, including the demand for large and varied datasets for efficient performance, the need for extensive computer resources during training, and possible issues understanding the learned features. Moreover, to successfully build a DBN-based CAD system, it is imperative to strike a balance between model complexity and resource requirements. The results on the MIAS dataset included an accuracy of 91.5%, a specificity of 72.4%, and a sensitivity of 94.1%. The authors in [
10] proposed a CNN approach for automatically classifying breast cancer from mammography and ultrasound images. The method has five adjustable convolutional blocks, each composed of four convolutional layers, with a single fully connected layer, which serves as the classifier. The method uses a few customizable parameters to automatically extract key features from images. The benefit of using deep Convolutional Neural Networks (CNNs) and a multi-modal approach is that they allow for an automated diagnosis and thorough analysis. Nonetheless, some difficulties can occur due to the CNN model’s interpretability, its reliance on large and high-quality data, the requirement of strong validation, and moral issues surrounding healthcare automation. The authors performed many simulations on mammography datasets (DDSM, MIAS, and INbreast) and ultrasound datasets (BUS-1 and BUS-2) and found that their evaluation metrics were better than the current best practices. Furthermore, data augmentation allowed for less overfitting. Using the DDSM, MIAS, and INbreast datasets, their CNN algorithms obtained accuracies of 90.68%, 96.55%, and 91.28%, respectively. Additional accuracies of 100% and 89.73% were achieved on the BUS-1 and BUS-2 datasets, respectively. In [
11], the authors began by removing noise, then adding a logarithmic spatial transform to improve the images, and finally, deleting the oblique and pectoral muscles and backdrop. Then, they used a fractional Fourier transform to obtain the coefficients of the time–frequency spectrum, which were then reduced using the PCA technique. In conclusion, the following performance results were obtained using the classifiers (k-nearest neighbors and SVM): in the case of SVM, the sensitivity was 92.22%, the specificity was 92.10%, and the accuracy was 92.16%. In addition, in [
12], the researchers suggested the use of a multi-scale all-Convolutional Neural Network (MA-CNN) for the classification of mammography images. To keep the connections between close pixels, instead of pooling, a long stride convolution was used. The possible advantages include the multi-scale approach and the particular emphasis on mammography characteristics. This notwithstanding, there are some limitations to this method, such as its dependence on varied and high-quality data, its interpretability, the necessity of a thorough validation, possible computing demands, and its practical application in a clinical environment. The obtained sensitivity, accuracy, and AUC on the MIAS dataset were 96%, 96.47%, and 99%, respectively.
In [
13], a YOLO-based CAD system utilizing deep learning was introduced to detect and classify masses related to breast cancer. The methodology comprises four sequential steps: After completing preprocessing, the model uses a deep convolutional network to extract the features. Mass detection is then performed, and a fully connected neural network is used for mass classification. The Digital Database for Screening Mammography (DDSM) and a pre-trained model from the ImageNet dataset were utilized, with weights assigned accordingly. Finally, the model was fine-tuned. The YOLO technique was employed in computer-aided design (CAD) to detect objects in real-time precisely. This enables the efficient processing of large CAD files. Its versatility and effectiveness make it possible to identify a wide range of object classes and integrate them into current systems. When it comes to localizing small objects, the YOLO method may not be as accurate as slower options. In CAD settings with limited resources, the large and complicated models and the need for many training data can be a problem. The effectiveness of YOLO depends on the needs of the particular CAD application, and the tradeoff between speed and accuracy is a factor. With the help of two different datasets from the DDSM database, consisting of an original amount of 600 photos and their augmented set of 2400 images, the researchers evaluated their system’s performance. They achieved an impressive overall accuracy of 97% and an area under the curve (AUC) of 96.45%. The researchers in [
14] presented a deep-belief-network (DBN)-based CAD system for identifying breast cancer. This approach involves the extraction of regions of interest (ROIs) utilizing two distinct methodologies. The initial approach selects four randomly chosen regions of interest (ROIs) from an identified mass, with each ROI having dimensions of 32 by 32 px. The second technique makes use of every ROI that has been found. This technique employs morphological processes and adaptive thresholding to detect masses with an accuracy of 86%. Due to its limitations in detecting masses in dense regions, many forms of breast tissue pose difficulties in identification and diagnosis. After the extraction of the ROIs, this technique provides 347 statistical parameter settings to identify the optimal one. Most of the time, hyperparameter tuning is performed by hand because the search area is large and it can be costly to test each setup. They then used deep belief networks (DBNs) to classify the images. The classification technique achieved an accuracy of 90.86% for malignant tumors and 92.86% for benign tumors. Additionally, the AUCs for the total mass method and the ROI mass method were 93.54% and 86.56%, respectively. A technique for identifying and classifying breast cancer using mammogram images was presented in [
15]. The researchers used a Convolutional Neural Network to classify mammogram images after several preprocessing procedures to adjust the CNN classifier’s parameters. Using the MIAS dataset, they achieved a percentage of 82.71% for the accuracy. The authors in [
16] developed a classification system for mammogram images called CNN improvement for breast cancer classification (CNNI-BCC). The CNNI-BCC model classifies the images of breasts into three classes: malignant, benign, and normal masses. They had an accuracy rate of 90.50% and a specificity of 90.71%. Although evolutionary algorithms have been employed to optimize CNN parameter values, their use has not received much research. Today, in research, metaheuristic algorithms are employed for designing deep learning architecture. For example, the authors in [
17] used both fuzzy logic modeling and a better quantum-behaved Particle Swarm Optimization method. To evaluate the impact of particular fuzzy variables on surface degradation, they conducted ball-on-disk tests. They improved the fuzzy model by making the fuzzy variables’ membership functions more optimal to increase the prediction accuracy. In [
18], the researchers developed a novel method to detect breast cancer in mammogram images, leveraging feature extraction and reduction techniques. The authors utilized various pre-trained Convolutional Neural Network (CNN) models to extract the features. These features were then combined, and the most-useful ones were chosen based on mutual information. The selected features were then classified using various machine learning algorithms, including neural networks (NNs), k-nearest neighbors (kNN), random forest (RF), and support vector machine (SVM). The proposed algorithm was evaluated on different datasets, including the newly introduced RSNA dataset, MIAS, and DDSM. The authors in [
19] conducted research to assess the effectiveness of genetic algorithms in the context of neural network classifiers for categorizing land cover in multispectral remote sensing data. A genetic algorithm analysis was performed in a hybrid environment with backpropagation, but the network properties and how they affect categorization were not specifically thought about.
Lorenzo et al. utilized Particle Swarm Optimization (PSO) to select the deepening parameters, as described in their study [
20]. The algorithm is based on natural behavior and imitates flocks of birds or fish. It was initially proposed by Eberhart and Kennedy in their publication [
21]. Based on the experiment results using the LeNet-4 network, it was shown that PSO can significantly improve the accuracy of classification on the MNIST dataset. However, their approach and methodology were not suitable for the CIFAR-10 datasets.
To be more precise, the large number of hyperparameters is an obstacle to achieving better results, despite the positive findings reached by CNN architectures in the detection of breast cancer. As a result, optimizing the hyperparameters for CNN design is crucial to enhancing CNN performance. This study developed an enhanced Convolutional Neural Network (CNN) structure for classifying mammography datasets. The architecture was refined using the Particle Swarm Optimization (PSO) approach to determine new hyperparameters. This might potentially be advantageous for healthcare professionals in the diagnosis of breast cancer.