Enhancing Brain Tumor Detection Through Custom Convolutional Neural Networks and Interpretability-Driven Analysis
Abstract
:1. Introduction
- Development of a custom CNN model specifically tailored for brain tumor detection using MRI scans.
- Comparative analysis with State-of-the-Art pretrained models, including ResNetV2, DenseNet201, and VGG16.
- Integration of oversampling techniques and class weighting to handle class imbalance.
- Enhancement of model interpretability using gradient-based attribution methods and saliency maps.
- Comprehensive performance evaluation using multiple metrics, including accuracy, precision, recall, and F1-score.
2. Literature Review
2.1. Conflicts and Gaps
2.2. Proposed Approach
3. Methodology
3.1. Data Collection
3.2. Data Description
3.3. Data Preprocessing
3.4. Model Architecture and Training Hyperparameters
3.4.1. Custom CNN Architecture
- Convolutional layers: These layers extract features such as edges, textures, and patterns from the MRI scans, using 3 × 3 kernels and ReLU activation.
- Max-pooling layers: These layers reduce the spatial dimensions of the feature maps, thus decreasing computational complexity while retaining important information.
- Fully connected layers: The first fully connected layer has 512 neurons with ReLU activation and a dropout rate of 0.5 to prevent overfitting. The second fully connected layer has 4 neurons (corresponding to the 4 classes in the dataset) and uses Softmax activation to output class probabilities.
- Dropout regularization: A dropout rate of 0.5 was applied to prevent overfitting, ensuring that the model generalizes well to unseen data.
3.4.2. Convolutional Layer Operation
3.4.3. Sigmoid Activation
3.4.4. Fine-Tuning Pretrained Models
- Layer modification: The original dense layer, which outputs 1000 classes, was replaced with a new dense layer configured to output 4 classes. This adjustment directly aligns the model with the four tumor categories specific to our dataset.
- Transfer-learning strategy: This is used to optimize the model for our task while leveraging the following powerful feature-extraction capabilities of these pretrained models:
- ⚬
- Layer freezing: Initially, we froze many of the lower layers of the pretrained models. This approach retained the learned weights from the ImageNet training, allowing the model to use these robust feature representations while preventing modifications during the early training stages.
- ⚬
- Unfreezing layers: After training the model with frozen lower layers for a few epochs, we gradually unfroze the higher layers. This allowed the model to fine-tune and adapt more specific features that are particularly relevant to our MRI dataset of brain tumors.
- Learning-rate adjustment: During the fine-tuning process, we adjusted the learning rates to ensure a smooth transition from general ImageNet tasks to our specific classification problem.
- ⚬
- Learning rate for new layers: A lower learning rate was set for the newly added output layers to allow for a more gradual adjustment to the specific features of our dataset, thereby reducing the risk of drastic changes that could hinder performance.
- ⚬
- Initial learning rate: The learning rate was set to 0.0001 for the new layers, while the learning rate for the frozen layers remained lower to retain the learned representations without disruption.
- Training strategy: The models underwent training with the following strategies.
- ⚬
- Early stopping: To prevent overfitting and ensure optimal performance, we employed early stopping. This technique monitored the validation loss during training and halted the process when performance ceased to improve.
- ⚬
- Batch size: A batch size of 32 was maintained throughout the training process to ensure efficient learning and convergence.
3.4.5. Training Hyperparameters
- Learning rate: The learning rate is a crucial hyperparameter that determines the step size at each iteration while moving toward a minimum of the loss function.
- ⚬
- Custom CNN: For the custom CNN model, an initial learning rate of 1 × 10−4 (0.0001) was selected based on preliminary experiments, ensuring that the model could learn effectively without making drastic updates to the weights.
- ⚬
- Fine-tuned models: For the fine-tuned pretrained models, a lower learning rate of 1 × 10−5 (0.00001) was set for the newly added layers. This adjustment allowed the model to transition smoothly from general ImageNet tasks to the specific classification challenges posed by our MRI dataset.
- Batch size: The batch size determines the number of training examples utilized in one iteration to update model weights.
- ⚬
- A consistent batch size of 32 was chosen for all models. This value strikes a balance between stable gradient estimation and efficient training time, allowing for effective convergence while managing memory usage.
- Optimizer: The optimizer is responsible for updating the model’s weights based on the gradients computed during backpropagation.
- ⚬
- The Adam optimizer was selected for all models due to its adaptive learning-rate capabilities. Adam is known for its efficiency and effectiveness in various deep-learning tasks, particularly with large datasets.
- Dropout rate: Dropout is a regularization technique used to prevent overfitting by randomly deactivating a fraction of neurons during training.
- ⚬
- A dropout rate of 0.5 was applied to the fully connected layers of the custom CNN. This rate helps ensure that the model does not rely too heavily on any single neuron, thereby enhancing generalization to unseen data.
- Loss function: The choice of loss function impacts how well the model learns from the training data.
- ⚬
- The categorical cross-entropy loss function was employed for both the custom CNN and fine-tuned models, as this loss function is particularly suited for multi-class classification tasks. It quantifies the difference between the true label and the predicted probabilities, guiding the optimization process effectively.
- Early stopping: Early stopping is used to avoid overfitting and ensure that the model generalizes well to new data.
- ⚬
- Early stopping was implemented, which monitored the validation loss during training. Training was halted when the validation loss stopped improving for a predetermined number of epochs (patience), preventing the model from continuing to learn patterns that may not generalize well.
3.4.6. Statistical Validation of Model Performance Using ANOVA
3.5. Model Selection
3.6. Handling Overfitting and Generalization
3.7. Model Training
3.8. Model Evaluation
3.9. Computational Resource Requirements
4. Class-Imbalance Remediation and Interpretability
4.1. Class-Imbalance Remediation
- Oversampling techniques: Our primary approach to mitigating class imbalance involved advanced data-augmentation techniques, such as rotation, translation, scaling, and flipping. By generating additional samples for the underrepresented tumor types, we aimed to balance the class distribution and provide the model with sufficient examples to learn the subtle features characteristic of each category [27].
- Class weighting: In addition to oversampling, we implemented class weighting as a complementary strategy. Assigning higher weights to minority classes during training ensured that the model paid equal attention to all classes, regardless of their representation in the dataset. This approach helped prevent bias toward the majority class and improved the model’s ability to accurately identify critical, underrepresented tumor types [28]. While undersampling was considered as an alternative approach, we opted against it to avoid reducing the overall dataset size and potentially discarding valuable information. Our chosen method directly addressed class imbalance through oversampling, leading to more reliable results and conclusions.
- Generalization: These class imbalance-mitigation techniques were essential for maximizing the model’s performance and generalization in accurately classifying tumor types. By ensuring a balanced representation of all classes, we contribute to a deeper understanding of human biology and demonstrate our commitment to scientific rigor and valid findings [29].
- Effectiveness and potential biases: The combination of oversampling and class-weighting techniques proved highly effective in addressing the class-imbalance issue in our dataset. After applying these strategies, we observed a significant improvement in the model’s performance, with an increase in accuracy from 93.27% to 94.51% and a decrease in test loss from 0.4532 to 0.1400. However, it is important to acknowledge potential biases that may be introduced by these techniques. Oversampling methods, such as advanced data augmentation, can potentially generate synthetic samples that do not accurately represent the true distribution of the minority class, leading to overfitting or the introduction of artifacts.
4.2. Qualitative Evaluations and Error Analysis
- Common misclassifications: Our analysis revealed recurring misclassifications in certain tumor types, such as gliomas and meningiomas, where similar structural features posed challenges for accurate classification. Additionally, pituitary tumors were occasionally misclassified, possibly due to their small size and subtle appearance in MRI scans. Understanding these common misclassifications helps us identify specific features or patterns that the model struggles to capture. This knowledge informs potential adjustments to the model architecture or feature-extraction methods to enhance classification accuracy [30].
- Challenging Cases: Some cases presented unique challenges for accurate classification, particularly tumors with atypical morphologies or rare histological subtypes. Aggressive glioma subtypes, like glioblastomas, often posed difficulties due to their heterogeneous appearance and rapid growth patterns. Similarly, cases involving multiple or recurrent tumors were challenging to classify accurately. By examining these challenging cases, we gain insights into the model’s limitations and areas for improvement. Expanding the training dataset to include a broader range of tumor variations or exploring ensemble methods may help address these challenges effectively [30].
- Error patterns and limitations: Our analysis also identified broader error patterns and limitations in the model’s performance. For instance, the model showed a tendency to misclassify tumors located in specific brain regions, suggesting potential biases or limitations in spatial information processing. Additionally, the model’s performance degraded when processing low-quality or artifact-ridden MRI scans, highlighting the importance of robust preprocessing techniques. Identifying these error patterns and limitations guides future research efforts to enhance model performance. Exploring advanced attention mechanisms or developing dedicated modules for handling low-quality data could address these challenges effectively [30].
4.3. Interpretability and Visualization
- Feature interpretation: We utilized gradient-based attribution methods like Integrated Gradients and Guided Backpropagation [31]. These methods helped us visualize the features the model learned from the brain scan images and how they contribute to the final classification (tumor type or healthy).
- Saliency maps: We also generated saliency maps to pinpoint the specific regions within the brain scans that most significantly influence the model’s output [31]. This helps us understand which parts of the image hold the most weight for the model’s decision-making process.
4.4. GUI Design
4.5. Methodological Choices
5. Results
5.1. Performance Metrics of Trained CNN Models
5.2. Performance Comparison With and Without ImageNet Weights
- With ImageNet weights: The models were initialized with weights pretrained on the ImageNet dataset.
- Without ImageNet weights: The models were randomly initialized and trained from scratch on our brain tumor dataset.
5.3. Training Dynamics and Learning Curves
5.3.1. Training Before Sampling
5.3.2. Training After Sampling
5.3.3. Learning-Curve Analysis
5.4. Overfitting Metrics
5.5. ROC Curve and Confusion Matrix
5.6. Class-Imbalance Mitigation
5.7. Gradient-Based Attribution Methods
5.8. User Interface and Model Metrics
5.9. Statistical Analysis
6. Discussion
6.1. Model Performance and Convergence
6.2. Addressing Class Imbalance
6.3. Enhancing Interpretability
6.4. Clinical Implications
6.5. Contribution to Brain Tumor Detection
6.6. Comparison with Related Models
6.7. Exploration of Transformer-Based Architectures
6.8. Security Threats and Countermeasures
- Differential privacy: This technique ensures that individual patient data points do not significantly influence the model, reducing the likelihood of privacy breaches from model outputs.
- Secure multi-party computation: By distributing the computation across multiple parties, the risk of an adversary gaining access to sensitive medical data is minimized.
6.9. Energy-Efficient Long-Term Health-Monitoring Systems
6.10. AI-Empowered IoT Security for Smart Cities
6.11. Ethical Considerations
6.12. Limitations and Future Work
7. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Khazaei, Z.; Goodarzi, E.; Borhaninejad, V.; Iranmanesh, F.; Mirshekarpour, H.; Mirzaei, B.; Naemi, H.; Bechashk, S.M.; Darvishi, I.; Ershad Sarabi, R.; et al. The association between incidence and mortality of brain cancer and human development index (HDI): An ecological study. BMC Public Health 2020, 20, 1696. [Google Scholar] [CrossRef] [PubMed]
- Bernstock, J.D.; Gary, S.E.; Klinger, N.; Valdes, P.A.; Ibn Essayed, W.; Olsen, H.E.; Chagoya, G.; Elsayed, G.; Yamashita, D.; Schuss, P.; et al. Standard clinical approaches and emerging modalities for glioblastoma imaging. Neuro-Oncol. Adv. 2022, 4, vdac080. [Google Scholar] [CrossRef] [PubMed]
- Sabeghi, P.; Zarand, P.; Zargham, S.; Golestany, B.; Shariat, A.; Chang, M.; Yang, E.; Rajagopalan, P.; Phung, D.C.; Gholamrezanezhad, A. Advances in Neuro-Oncological Imaging: An Update on Diagnostic Approach to Brain Tumors. Cancers 2024, 16, 576. [Google Scholar] [CrossRef] [PubMed]
- Wu, J.; Li, C.; Gensheimer, M.; Padda, S.; Kato, F.; Shirato, H.; Wei, Y.; Schönlieb, C.-B.; Price, S.J.; Jaffray, D.; et al. Radiological tumour classification across imaging modality and histology. Nat. Mach. Intell. 2021, 3, 787–798. [Google Scholar] [CrossRef]
- Orr, B.A. Pathology, diagnostics, and classification of medulloblastoma. Brain Pathol. 2020, 30, 664–678. [Google Scholar] [CrossRef]
- ZainEldin, H.; Gamel, S.A.; El-Kenawy, E.M.; Alharbi, A.H.; Khafaga, D.S.; Ibrahim, A.; Talaat, F.M. Brain Tumor Detection and Classification Using Deep Learning and Sine-Cosine Fitness Grey Wolf Optimization. Bioengineering 2022, 10, 18. [Google Scholar] [CrossRef]
- Saraswat, B.K.; Vaibhav, V.; Pal, P.; Singh, A.K.; Tiwari, N. Brain Tumor Detection. IJRASET 2023, 11, 5634–5640. [Google Scholar] [CrossRef]
- Rajeev, S.K.; Rajasekaran, M.P.; Ramaraj, K.; Vishnuvarthanan, G.; Arunprasath, T.; Muneeswaran, V. A Hybrid CNN-LSTM Network For Brain Tumor Classification Using Transfer Learning. In Proceedings of the 2023 9th International Conference on Smart Computing and Communications (ICSCC), Kochi, Kerala, India, 17–19 August 2023; pp. 77–82. [Google Scholar] [CrossRef]
- Aakanksha, M. Brain Tumor Detection using Deep Learning. Int. J. Res. Appl. Sci. Eng. Technol. 2023, 11, 490–493. [Google Scholar] [CrossRef]
- Singh, A. Review of Brain Tumor Detection from MRI Images. In Proceedings of the 2016 3rd International Conference on Computing for Sustainable Global Development (INDIACom), New Delhi, India, 16–18 March 2016; pp. 3997–4000. [Google Scholar]
- Tambe, U.Y.; Shanthini, A. Brain Tumor Detection & Classification into Different Categories using Deep Learning Model. In Proceedings of the 2023 International Conference on Advanced Computing Technologies and Applications (ICACTA), Mumbai, India, 6–7 October 2023. [Google Scholar]
- Badjie, B.; Deniz Ülker, E. A Deep Transfer Learning Based Architecture for Brain Tumor Classification Using MR Images. Inf. Technol. Control 2022, 51, 332–344. [Google Scholar] [CrossRef]
- Banu, R. Brain Tumour Detection and Classification Using U-Net Deep Neural Network. Int. J. Creat. Res. Thoughts (IJCRT) 2022, 10, 816–820. [Google Scholar]
- Liu, Z.; Mao, H.; Wu, C.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11966–11976. [Google Scholar] [CrossRef]
- Dai, Z.; Liu, H.; Le, Q.V.; Tan, M. CoAtNet: Marrying Convolution and Attention for All Data Sizes. arXiv 2021, arXiv:2106.04803. [Google Scholar]
- Tajane, K.; Rathkanthiwar, V.; Chava, G.; Dhavale, S.; Chawda, G.; Pitale, R. EffiConvRes: An Efficient Convolutional Neural Network with Residual Connections and Depthwise Convolutions. In Proceedings of the 2023 7th International Conference on Computing, Communication, Control And Automation (ICCUBEA 2023), Pune, India, 18–19 August 2023. [Google Scholar] [CrossRef]
- Todi, A.; Narula, N.; Sharma, M.; Gupta, U. ConvNext: A Contemporary Architecture for Convolutional Neural Networks for Image Classification. In Proceedings of the 3rd International Conference on Innovative Sustainable Computational Technologies, Graphic Era Deemed to Be University, Dehradun, India, 8–9 September 2023; pp. 1–6. [Google Scholar] [CrossRef]
- Priyadarshini, P.; Kanungo, P.; Kar, T. Multigrade brain tumor classification in MRI images using Fine tuned efficientnet. e-Prime 2024, 8, 100498. [Google Scholar] [CrossRef]
- Kassu, J.S. Research Design and Methodology; Abu-Taieh, E., El Mouatasim, A., Al Hadid, I.H., Eds.; Cyberspace; IntechOpen: Rijeka, Croatia, 2019; Chapter 3. [Google Scholar]
- Kadam, A. Brain Tumor Classification using Deep Learning Algorithms. Int. J. Res. Appl. Sci. Eng. Technol. 2021, 9, 417–426. [Google Scholar] [CrossRef]
- Razzaq, M.; Clément, F.; Yvinec, R. An overview of deep learning applications in precocious puberty and thyroid dysfunction. Front. Endocrinol. 2022, 13, 959546. [Google Scholar] [CrossRef]
- Alzubaidi, L.; Zhang, J.; Al-Amidie, M.; Farhan, L.; Fadhel, M.A.; Duan, Y.; Santamaría, J.; Al-Dujaili, A.; Al-Shamma, O.; Humaidi, A.J. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef]
- Aliferis, C.; Simon, G. Overfitting, Underfitting and General Model Overconfidence and Under-Performance Pitfalls and Best Practices in Machine Learning and AI; Springer: Cham, Switzerland, 2024. [Google Scholar] [CrossRef]
- Salehin, I.; Kang, D. A Review on Dropout Regularization Approaches for Deep Neural Networks within the Scholarly Domain. Electronics 2023, 12, 3106. [Google Scholar] [CrossRef]
- Wang, X.; Yan, L.; Zhang, Q. Research on the Application of Gradient Descent Algorithm in Machine Learning. In Proceedings of the 2021 International Conference on Computer Network, Electronic and Automation (ICCNEA), Xi’an, China, 24–26 September 2021; pp. 11–15. [Google Scholar] [CrossRef]
- Matsuyama, E.; Nishiki, M.; Takahashi, N.; Watanabe, H. Using Cross Entropy as a Performance Metric for Quantifying Uncertainty in DNN Image Classifiers: An Application to Classification of Lung Cancer on CT Images. J. Biomed. Sci. Eng. 2024, 17, 1–12. [Google Scholar] [CrossRef]
- Gnip, P.; Vokorokos, L.; Drotár, P. Selective oversampling approach for strongly imbalanced data. PeerJ Comput. Sci. 2021, 7, e604. [Google Scholar] [CrossRef]
- Araf, I.; Idri, A.; Chairi, I. Cost-sensitive learning for imbalanced medical data: A review. Artif. Intell. Rev. 2024, 57, 80. [Google Scholar] [CrossRef]
- Johnson, J.L.; Adkins, D.; Chauvin, S. A Review of the Quality Indicators of Rigor in Qualitative Research. Am. J. Pharm. Educ. 2020, 84, 7120–7146. [Google Scholar] [CrossRef]
- Althubaiti, A. Information bias in health research: Definition, pitfalls, and adjustment methods. J. Multidiscip. Healthc. 2016, 9, 211–217. [Google Scholar] [CrossRef] [PubMed]
- Rajbahadur, G.K.; Wang, S.; Oliva, G.A.; Kamei, Y.; Hassan, A.E. The Impact of Feature Importance Methods on the Interpretation of Defect Classifiers. TSE 2022, 48, 2245–2261. [Google Scholar] [CrossRef]
- Deepak, S.; Ameer, P.M. Retrieval of brain MRI with tumor using contrastive loss based similarity on GoogLeNet encodings. Comput. Biol. Med. 2020, 125, 103993. [Google Scholar] [CrossRef] [PubMed]
- Çinar, A.; Yildirim, M. Detection of tumors on brain MRI images using the hybrid convolutional neural network architecture. Med. Hypotheses 2020, 139, 109684. [Google Scholar] [CrossRef] [PubMed]
- Khare, N.; Devan, P.; Chowdhary, C.; Bhattacharya, S.; Singh, G.; Singh, S.; Yoon, B. SMO-DNN: Spider Monkey Optimization and Deep Neural Network Hybrid Classifier Model for Intrusion Detection. Electronics 2020, 9, 692. [Google Scholar] [CrossRef]
- Kharrat, A.; Gasmi, K.; Ben Messaoud, M.; Benamrane, N.; Abid, M. Medical Image Classification Using an Optimal Feature Extraction Algorithm and a Supervised Classifier Technique. Int. J. Softw. Sci. Comput. Intell. 2011, 3, 19–33. [Google Scholar] [CrossRef]
- Díaz-Pernas, F.J.; Martínez-Zarzuela, M.; Antón-Rodríguez, M.; González-Ortega, D. A Deep Learning Approach for Brain Tumor Classification and Segmentation Using a Multiscale Convolutional Neural Network. Healthcare 2021, 9, 153. [Google Scholar] [CrossRef]
- Hemanth, D.J.; Anitha, J.; Naaji, A.; Geman, O.; Popescu, D.E.; Hoang Son, L. A Modified Deep Convolutional Neural Network for Abnormal Brain Image Classification. IEEE Access 2019, 7, 4275–4283. [Google Scholar] [CrossRef]
- Paul, J.S.; Plassard, A.; Landman, B.; Fabbri, D. Deep Learning for Brain Tumor Classification. In Proceedings of the Medical Imaging 2017: Biomedical Applications in Molecular, Structural, and Functional Imaging, Orlando, FL, USA, 11–16 February 2017; Volume 10137, p. 1013710. [Google Scholar] [CrossRef]
- Tian, Z.; Cui, L.; Liang, J.; Yu, S. A Comprehensive Survey on Poisoning Attacks and Countermeasures in Machine Learning. ACM Comput. Surv. 2023, 55, 1–35. [Google Scholar] [CrossRef]
- Wu, B.; Wei, S.; Zhu, M.; Zheng, M.; Zhu, Z.; Zhang, M.; Chen, H.; Yuan, D.; Liu, L.; Liu, Q. Defenses in Adversarial Machine Learning: A Survey. arXiv 2023, arXiv:2312.08890. [Google Scholar]
- Zhou, S.; Liu, C.; Ye, D.; Zhu, T.; Zhou, W.; Yu, P.S. Adversarial Attacks and Defenses in Deep Learning: From a Perspective of Cybersecurity. ACM Comput. Surv. 2023, 55, 1–39. [Google Scholar] [CrossRef]
- Wang, F.; Wang, X.; Ban, X.J. Data poisoning attacks in intelligent transportation systems: A survey. Transp. Res. Part C Emerg. Technol. 2024, 165, 104750. [Google Scholar] [CrossRef]
- Khalid, N.; Qayyum, A.; Bilal, M.; Al-Fuqaha, A.; Qadir, J. Privacy-preserving artificial intelligence in healthcare: Techniques and applications. Comput. Biol. Med. 2023, 158, 106848. [Google Scholar] [CrossRef] [PubMed]
- Yu, H.; Li, N.; Zhao, N. How Far Are We from Achieving Self-Powered Flexible Health Monitoring Systems: An Energy Perspective. Adv. Energy Mater. 2021, 11, 2058. [Google Scholar] [CrossRef]
Source | Methodology | Main Findings | Limitations |
---|---|---|---|
[7] | Deep learning-based approach combining CNNs and RNNs to analyze multimodal MRI data, with data augmentation. | Improved sensitivity and specificity in brain tumor detection through a deep learning-based approach combining CNNs and RNNs. | Limited availability of labeled data addressed through data augmentation. |
[8] | Utilization of deep learning for medical image analysis, preprocessing MRI images, and classification with a hybrid CNN-LSTM model. | Outperformed existing models in brain tumor classification with a high validation accuracy. | The proposed method relies on pretrained models like AlexNet for feature extraction, potentially limiting its ability to adapt to new and diverse datasets without further fine-tuning. |
[9] | Development of a deep-learning system using convolutional neural networks for brain tumor detection from MRI scans. | Accurate detection of brain tumors from MRI scans using deep learning-based system. | Limited availability of diverse and representative datasets for training may constrain the generalizability of the model’s predictions. |
[10] | Utilization of deep-learning models (ResNet50, ConvNeXt, and custom CNN) for brain tumor detection from MRI scans. | Deep-learning models offer efficient tumor detection on MRI images for clinicians. | Limited availability of diverse and large-scale datasets for training and testing deep-learning models, which may affect the generalizability and robustness of the developed brain tumor-detection system. |
[11] | Use of CNN for brain tumor detection and classification, development of a deep-learning model for tumor categorization. | Deep-learning model accurately classifies brain tumors into different categories with high accuracy. | Complexity of MRI images, limited classification into 4 tumor types, generalizability not discussed. |
[12] | Utilization of transfer-learning model (AlexNet’s CNN) for brain tumor detection and classification in MR images. | Transfer learning with AlexNet’s CNN improves brain tumor detection and classification in MR images. | Not mentioned. |
[13] | Development of DL model based on U-Net CNN for classifying different brain tumor types. | DL model based on U-Net CNN classifies different brain tumor types with high accuracy. | Limited availability of diverse and large-scale datasets for training and testing the U-Net model, which may affect the generalizability and robustness of the developed brain tumor detection-and-classification system. |
Aspect | Key Findings | Methodologies | Gaps/Limitations |
---|---|---|---|
Deep-learning approaches | Many studies demonstrate the effectiveness of CNNs in analyzing medical images for brain tumor detection, leading to high accuracy rates. | Utilization of CNN architectures to process MRI scans and classify brain tumors into various subtypes. Data preprocessing, including normalization and augmentation, is used to enhance model performance. Transfer-learning techniques to improve model accuracy by transferring knowledge from pretrained models. | Lack of interpretability in deep-learning models. Limited generalizability due to focus on specific datasets. |
Multimodal imaging | Integration of MRI, CT, and PET data can enhance tumor boundary delineation and diagnostic accuracy. | Utilization of multimodal datasets combining information from different imaging techniques. Fusion techniques, such as feature concatenation or attention mechanisms, to integrate information from multiple modalities. Training of models using deep-learning or traditional machine-learning algorithms. | Challenges in integrating and harmonizing data from disparate sources. Standardization of imaging protocols and data preprocessing techniques is crucial. |
Addressing class imbalance | Strategies like oversampling, class weighting, and specialized loss functions mitigate the negative impact of class imbalance on model performance. | Employing techniques such as oversampling to generate synthetic samples for minority classes. Applying class weighting during model training to give higher importance to minority classes. Designing specialized loss functions to penalize misclassifications of minority classes more heavily. | While these strategies improve model performance, they may not fully address the underlying imbalance in the dataset. Further research is needed to explore novel approaches for handling class imbalance effectively. |
Interpretability of models | Techniques for visualizing and interpreting model predictions, such as gradient-based attribution methods and saliency maps, improve model transparency. | Visualization techniques to highlight influential regions in input images and visualize the features learned by the model. Employing gradient-based attribution methods like Integrated Gradients and Guided Backpropagation to identify influential pixels in the input image. Generating saliency maps to visualize areas of the image that contribute most to the model’s outputs. | While these techniques provide valuable insights into model predictions, they may not always capture the complex decision-making process of deep-learning models. Further research is needed to develop more interpretable models and visualization techniques tailored to medical-imaging tasks. |
Challenges | Description |
---|---|
Limited generalization | Studies often focus on specific datasets, making findings less applicable. |
Class imbalance | Some tumor types are underrepresented, leading to biased predictions. |
Interpretability | Deep-learning models are hard to interpret, making it challenging for clinicians to understand predictions. |
Goal | Description |
---|---|
Developing a custom CNN model | Conceive and train a bespoke CNN architecture tailored to brain tumor detection. Leverage domain-specific insights and architectural innovations to enhance sensitivity, specificity, and resilience. |
Addressing class imbalance | Employ strategies like oversampling minority classes, class weighting, or specialized loss functions to alleviate the effects of class imbalance and ensure equitable performance across all tumor categories. |
Enhancing interpretability | Explore techniques for visualizing and interpreting the features learned by the CNN model, including gradient-based attribution methodologies, activation maximization, and saliency maps. These techniques provide valuable insights into the model’s decision-making rationale and facilitate collaboration between clinicians and AI systems. |
Glioma | Meningioma | Pituitary | No Tumor | |
---|---|---|---|---|
Training | 826 | 822 | 827 | 395 |
Testing | 100 | 115 | 74 | 105 |
Layer Type | Output Shape | Kernel Size | Activation Function | Number of Parameters |
---|---|---|---|---|
Input Layer | (128, 128, 3) | - | - | 0 |
Convolutional Layer 1 | (128, 128, 32) | 3 × 3 | ReLU | 896 |
Max-Pooling Layer 1 | (64, 64, 32) | 2 × 2 | - | 0 |
Convolutional Layer 2 | (64, 64, 64) | 3 × 3 | ReLU | 18,496 |
Max-Pooling Layer 2 | (32, 32, 64) | 2 × 2 | - | 0 |
Convolutional Layer 3 | (32, 32, 128) | 3 × 3 | ReLU | 73,856 |
Max-Pooling Layer 3 | (16, 16, 128) | 2 × 2 | - | 0 |
Flatten Layer | (32,768) | - | - | 0 |
Fully Connected Layer 1 | (512) | - | ReLU | 16,777,472 |
Dropout Layer | (512) | - | - | 0 |
Fully Connected Layer 2 | (4) | - | Softmax | 2052 |
Architecture | Number of Layers | Kernel Sizes | Activation Function |
---|---|---|---|
Proposed CNN | 3 convolutional + 2 fully connected | 3 × 3 | ReLU |
ResNetV2 | 50 (including residual blocks) | 3 × 3 | ReLU |
VGG16 | 16 (13 convolutional + 3 fully connected) | 3 × 3 | ReLU |
DenseNet201 | 201 (including dense blocks) | 3 × 3 | ReLU |
Model | Total Parameters (Without Sigmoid) | Total Parameters (with Sigmoid) |
---|---|---|
Proposed CNN | 819,290,760 | 83,840 |
ResNetV2 | ~25.6 million | ~25.6 million |
VGG16 | 138,357,544 | 123,651,176 |
DenseNet201 | ~20 million | ~20 million |
Model | L2 Regularization |
---|---|
Proposed CNN | 6.154 × 10−7 |
ResNetV2 | 4.185 × 10−10 |
VGG16 | 1.171 × 10−9 |
DenseNet201 | 7.678 × 10−10 |
Component | Training Specifications | Inference Specifications |
---|---|---|
GPU | NVIDIA GeForce GTX 1080 Ti (12 GB VRAM) | NVIDIA GeForce GTX 1650 (4 GB VRAM) |
CPU | Intel Core i7-8700K (3.7 GHz, 6 cores) | Intel Core i5-9300H (2.4 GHz, 4 cores) |
RAM | Corsair 32 GB DDR4 | Corsair 16 GB DDR4 |
Model | Accuracy | Loss |
---|---|---|
Proposed CNN | 0.9521 | 0.3386 |
ResNetV2 | 0.9269 | 0.5018 |
VGG16 | 0.8685 | 0.5602 |
DenseNet201 | 0.6497 | 1.2507 |
Model | Initialization | Accuracy (%) | Loss |
---|---|---|---|
ResNetV2 | With ImageNet Weights | 93.45 | 0.352 |
ResNetV2 | Without ImageNet | 88.27 | 0.522 |
VGG16 | With ImageNet Weights | 90.62 | 0.405 |
VGG16 | Without ImageNet | 85.48 | 0.590 |
DenseNet201 | With ImageNet Weights | 92.38 | 0.376 |
DenseNet201 | Without ImageNet | 87.31 | 0.533 |
Proposed CNN | N/A | 94.51 | 0.140 |
Metric | Oversampled Dataset | Non-Oversampled Dataset |
---|---|---|
Training accuracy (%) | 94.17 | 95.21% |
Validation accuracy (%) | 94.19 | 86.42% |
Test accuracy (%) | 94.51 | 87.00% |
Metric | Original Value | Improvement | New Value (with Improvement) |
---|---|---|---|
Accuracy | 95.215 | −0.705% | 94.51 |
Loss | 0.3386 | −0.1986 | 0.140 |
Tumor Type | Precision | Recall | F1-Score |
---|---|---|---|
Glioma tumors | 95% | 92% | 93% |
Meningioma tumor | 93% | 88% | 90% |
No tumor | 94% | 99% | 96% |
Pituitary tumor | 97% | 98% | 97% |
Weighted average | 94% | 94% | 94% |
Metric | Value |
---|---|
F-statistic | 15.34 |
p-value | <0.001 |
Comparison | Significant Difference | p-Value |
---|---|---|
Proposed CNN vs. VGG16 | Yes | <0.01 |
Proposed CNN vs. DenseNet201 | Yes | <0.001 |
ResNetV2 vs. DenseNet201 | Yes | <0.01 |
Proposed CNN vs. ResNetV2 | No | N/A |
Metric | EfficientNetV2S | Custom CNN Model |
---|---|---|
Accuracy | 98.48% | 94.51% |
Precision | 98.5% | 94% |
Recall | 98% | 94% |
F1-score | 98% | 94% |
Interpretability tools | Grad-CAM | Saliency Maps, Gradient Attribution |
Source | Classified Method | Accuracy | Additional Information |
---|---|---|---|
[32] | Siamese Neural Network (GoogLeNet) | 97.64% | The Siamese Neural Network achieves a commendable accuracy of 97.64%. However, its performance is slightly lower than the proposed method. While Siamese networks are effective for tasks like image similarity and verification, their suitability for brain tumor detection may vary depending on the dataset and task requirements. |
[33] | Hybrid CNN (Resnet50) | 97.20% | The Hybrid CNN, utilizing ResNet50 architecture, achieves an accuracy of 97.20%, which is slightly lower than both the proposed method and the Siamese Neural Network. Hybrid CNN architectures often combine features from multiple CNN architectures to improve performance. However, their complexity may pose challenges in interpretation and implementation. |
[34] | Optimal DNN and Spider-Monkey Optimization | 99.30% | Preethi and Aishwarya’s method, employing an Optimal DNN with Spider-Monkey Optimization, achieves the highest accuracy of 99.30%. While the accuracy is impressive, the complexity of the optimization technique and the interpretability of the model may be limiting factors for practical applications. |
[35] | Wavelet Transform and Support Vector Machine | 98.14% | Kharrat et al. achieved an accuracy of 98.14% using Wavelet Transform combined with Support Vector Machine (SVM). While SVMs are known for their effectiveness in classification tasks, the reliance on feature engineering and the interpretability of the model may be challenging compared to deep-learning approaches. |
[36] | Multiscale CNN | 97.30% | Diaz-Pernas et al.’s Multiscale CNN achieved an accuracy of 97.30%, demonstrating robust performance in brain tumor detection. Multiscale CNN architectures leverage features at multiple resolutions, offering a comprehensive representation of the input data. However, they may require more computational resources during training and inference. |
[37] | Modified Deep CNN | 96.40% | Hemanth et al. achieved an accuracy of 96.40% with a Modified Deep CNN, demonstrating competitive performance in tumor detection. Modifications to standard CNN architectures can improve their effectiveness for specific tasks. However, the degree of modification and its impact on model interpretability should be carefully considered. |
[38] | Convolutional Neural Network | 91.43% | Paul et al.’s CNN achieved an accuracy of 91.43%, which is relatively lower compared to other methods. The lower accuracy may be attributed to various factors such as dataset characteristics, model architecture, or training methodology. |
[12] | Deep transfer learning (AlexNet) | 99.62% | While Badjie and Ülker achieved impressive results using AlexNet for two-class classification, our study expands upon this by classifying four classes, making the problem more complex. Transfer learning remains highly effective, but the challenges increase as more tumor types are introduced. |
Proposed method | Custom CNN | 94.51% | The proposed method utilizes a custom CNN architecture tailored specifically for brain tumor detection. This approach allows for better capturing of features relevant to tumor classification, leading to a high accuracy of 94.51%. The custom CNN architecture offers flexibility and adaptability to the dataset and clinical requirements, potentially making it more suitable for real-world applications compared to standardized models. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Dewage, K.A.K.W.; Hasan, R.; Rehman, B.; Mahmood, S. Enhancing Brain Tumor Detection Through Custom Convolutional Neural Networks and Interpretability-Driven Analysis. Information 2024, 15, 653. https://doi.org/10.3390/info15100653
Dewage KAKW, Hasan R, Rehman B, Mahmood S. Enhancing Brain Tumor Detection Through Custom Convolutional Neural Networks and Interpretability-Driven Analysis. Information. 2024; 15(10):653. https://doi.org/10.3390/info15100653
Chicago/Turabian StyleDewage, Kavinda Ashan Kulasinghe Wasalamuni, Raza Hasan, Bacha Rehman, and Salman Mahmood. 2024. "Enhancing Brain Tumor Detection Through Custom Convolutional Neural Networks and Interpretability-Driven Analysis" Information 15, no. 10: 653. https://doi.org/10.3390/info15100653
APA StyleDewage, K. A. K. W., Hasan, R., Rehman, B., & Mahmood, S. (2024). Enhancing Brain Tumor Detection Through Custom Convolutional Neural Networks and Interpretability-Driven Analysis. Information, 15(10), 653. https://doi.org/10.3390/info15100653