1. Introduction
Skin cancer, the most prevalent malignancy worldwide, originates from the uncontrolled growth of skin cells in the outermost layer of the body [
1,
2]. This disease has become alarmingly common, with approximately one in three cancer diagnoses being skin cancer and over 3.5 million cases reported each year in the United States [
3,
4]. Skin cancers can be broadly categorized into melanoma and nonmelanoma skin cancers (NMSCs), with the latter divided into basal cell carcinoma (BCC) and squamous cell carcinoma (SCC) [
5]. While BCC generally exhibits a lower tendency for metastasis, SCC represents a significant health risk, contributing substantially to NMSC-related metastasis and mortality [
6,
7]. SCC, in particular, poses a greater clinical challenge owing to its potential for aggressive spread and distinct histological classifications, ranging from well-differentiated (Grade I) to poorly differentiated (Grades II and III) and undifferentiated or invasive stages (Grade IV) [
8].
Primary SCC diagnosis often involves dermoscopic examination and tissue biopsy, followed by Mohs micrographic surgery (MMS), where tissue samples are microscopically analyzed to assess for residual cancer cells [
9,
10]. Effective margin assessment in SCC treatment is essential to ensure complete removal of tumor cells, which is crucial for reducing recurrence [
11]. However, manual microscopic examinations demand substantial expertise and are heavily dependent on the judgment of trained pathologists [
12,
13]. The variability in diagnostic accuracy due to subjective interpretation has led to a demand for computational assistance through artificial intelligence (AI) and deep learning models for histopathological image analysis [
14].
Histopathological image quality is critical in AI-based diagnostic accuracy [
15] and is affected by multiple factors, including microscope quality, staining techniques, reagent reliability, skill level of laboratory personnel, and environmental conditions during diagnostic processes [
16,
17]. In resource-limited settings, the lack of high-quality equipment and trained personnel often results in low-quality histopathological images [
18,
19].
Figure 1 presents a comparative example.
Figure 1A, from a well-resourced laboratory in India, demonstrates the clarity and detail that can aid in precise diagnosis [
20]. In contrast,
Figure 1B from a low-resource laboratory in Ethiopia showed limitations in image quality owing to substandard equipment and resources, which can impair accurate diagnosis and treatment planning [
21].
Over the past decade, convolutional neural networks (CNNs) are the leading AI models for histopathological image analysis [
22]. However, their performance is highly sensitive to image quality, leading to decreased accuracy when analyzing low-quality images [
23,
24,
25]. Vision transformers (ViTs), an emerging architecture in deep learning, have recently gained popularity because of their ability to outperform CNNs in various image analysis tasks, particularly in suboptimal image quality situations [
26,
27,
28]. ViTs can capture long-range dependencies and structural details more effectively, making them promising candidates for improving diagnostic accuracy, even with low-quality images [
29,
30].
In this study, we proposed a ViT-based model specifically designed for histopathological images with compromised quality, focusing on the classification of SCC margins. By examining the efficacy of ViTs in this context, we aimed to demonstrate their potential to enhance diagnostic accuracy and clinical decision-making in settings with limited resources.
2. Materials and Methods
2.1. Dataset and Preprocessing
Tissue samples were sourced from the Jimma University Medical Center, a leading healthcare facility in Ethiopia [
21]. The dataset can be accessed publicly at (
https://osf.io/3ma4p/) (accessed on 28 August 2024) and provides an important resource for researchers and clinicians. Tissue samples were obtained from patients diagnosed with skin-related conditions, specifically SCC. Hematoxylin and eosin (H&E) staining, a widely used method for the histological examination of tissue samples, was consistently applied across all samples. This standard staining method aids in highlighting cellular structures, allowing for easier visualization of abnormalities, such as cancerous cells and tissue margins. The dataset comprises histopathological slides obtained from 50 patients, offering a diverse representation of SCC grades. The distribution of patients included 17 with well-differentiated SCC, 15 with moderately differentiated SCC, and 18 with invasive SCC, reflecting varied degrees of cancer progression. This variation in differentiation stages is critical for training machine learning models that need to differentiate between varying levels of tumor aggressiveness. Additionally, the dataset included 345 normal tissue images that were classified as margin negative and represented healthy tissues. In contrast, 483 images containing tumor cells were designated as margin positive, indicating the presence of cancer at the tissue margins. The source dataset is particularly valuable for studying SCC margin classification, which is critical for ensuring complete tumor resection during surgery. Margin classification is crucial for determining whether cancer cells remain at surgical margins, which can influence treatment decisions and patient outcomes. By providing these images, along with their margin labels, the dataset supports the development of AI models aimed at automating the detection of tumor cells in low-quality or resource-limited settings, where manual analysis by pathologists may be challenging. Example images of normal and tumor tissue samples are shown in
Figure 2.
The original resolution of the SCC margin cell images in this study was 2048 × 1536 pixels, which is a high level of detail typically required for histopathological analyses. However, to make the images more manageable and suitable for deep learning models, all images were resized to 224 × 224 pixels. This reduction in resolution is a common practice in machine learning to reduce the computational overhead while preserving key tissue features for analysis. A size of 224 × 224 pixels is often preferred for CNNs and other machine learning models because it offers a balance between processing efficiency and maintaining the necessary details for identifying abnormalities, such as cancerous cells. Furthermore, image resizing facilitates the generation of smaller, localized portions of the original images, enabling the model to focus on specific tissue regions and increase its ability to detect tumors even in lower-quality images. The patch-based approach is particularly useful in histopathology because tumors or abnormal cells may appear differently depending on their location within the tissue.
To further improve model performance and prevent overfitting, where the model becomes overly tuned to the training data, various data augmentation techniques were applied. These methods include flipping, scaling, and image rotation. Flipping the images horizontally or vertically mimicked the natural variations in tissue orientation during sample preparation and imaging. Scaling adjusts the size of the images, simulating cases in which the tumor or tissue features may appear larger or smaller, which is common in clinical practice. Rotation allows the model to learn from images at different angles, thereby improving its robustness to variations in slide presentation. These augmentations artificially increase the size and diversity of the training dataset, allowing the model to generalize better to unseen data, which is particularly critical when working with small datasets or images from resource-limited settings, where data variability can be high. By creating multiple versions of each image, the model learns more generalized patterns, preventing it from memorizing the specific features of the training images, thus improving its ability to accurately classify new images.
To assess the generalization capabilities of the model and further reduce the risk of overfitting, 5-fold cross-validation was used. This technique divides a dataset into five equally sized subsets or folds. The model was trained on four folds and tested on the remaining folds; this process was repeated for each fold. This ensured that every sample in the dataset was used for both training and testing, thereby providing a more robust evaluation of model performance. In this study, 90% of the dataset was used for training, allowing the model to learn from a large portion of the data, whereas the remaining 10% was used for testing. Cross-validation provides a more reliable estimate of the performance of the model on unseen data, ensuring that it does not overfit any particular subset of the data. In this work, we used stratified cross-validation to address the class imbalance in the dataset. Stratified cross-validation ensures that each fold of the dataset maintains the same proportion of classes as the original dataset, which is especially important in cases of imbalanced data. By preserving the distribution of classes across all folds, we ensured that the model is trained and evaluated on representative samples of both the majority (margin positive) and minority (margin negative) classes. This approach helps to mitigate the risk of biased performance metrics and allows the model to learn more effectively from the underrepresented class, ultimately leading to a more robust and fair evaluation of its performance.
2.2. Best Model and Parameters Selection
Preliminary experiments were conducted to identify the most suitable CNN and ViT models for the task, along with optimization of their key parameters. The initial selection included three CNN architectures, ResNet50, EfficientNetB2, and InceptionV3, and three ViT models, ViT-B16, ViT-B32, and ViT-L32. These models were evaluated using a range of learning rates (0.0001, 0.001, 0.01, and 0.1) and five optimizers: Adam, Adadelta, Adamax, Adagrad, and Stochastic Gradient Descent (SGD). The aim of this study was to explore how these factors influenced the ability of the model to classify SCC margin images effectively and to choose the best optimized parameters and models. In these experiments, the models’ pretrained ImageNet weights were retained, and only the final output layer was replaced and fine-tuned for the specific task. A fixed number of 50 epochs was selected for training to allow sufficient learning while maintaining computational efficiency.
The results of these preliminary experiments, which guided the selection of the best-performing models, are shown in
Figure 3. In these early tests, the performance differences between the models and optimizers were assessed based on the classification accuracy and area under the receiver operator curve (AUC) values. The outcomes revealed valuable insights into the behavior of the models under different configurations, helping to refine the choice of architecture and optimization strategy for subsequent experiments. From the preliminary experiments, we identified the best-performing model for this study. The ViT model, ViT-B16, was selected as the optimal ViT, whereas the InceptionV3 model was chosen as the best CNN model. These models were selected based on their superior performance in terms of classification accuracy and efficiency when applied to the dataset used in this study. In addition to selecting the models, optimal training parameters were determined. Among the various optimizers tested, the Adam optimizer yielded the best results for both the ViT and CNN models. The ability of Adam to adaptively adjust the learning rate during training makes it particularly effective for fine-tuning models on a specific histopathology dataset. Moreover, a learning rate of 0.0001 emerged as the most effective starting point for both models because it balanced the speed of convergence with stable training and minimized the risk of overfitting. Therefore, based on the results of these preliminary experiments, the study proceeded with the ViT-B16 and InceptionV3 models using the Adam optimizer with a starting learning rate of 0.0001 as the final selected parameter for further training and evaluation.
2.3. Proposed Model
In this study, a ViT-based transfer learning approach was used to classify low-quality SCC images as either positive (present tumor) or negative (no tumor). The core idea behind transfer learning is to leverage a model pretrained on a large, generalized dataset, such as ImageNet, and fine-tune it for a specific task, in this case, the classification of histopathological images. Pretraining on ImageNet allows the model to learn the general features of images, such as edges, textures, and basic shapes, which can then be adapted for more specialized applications, such as SCC image classification. This approach is particularly beneficial when working with limited data as it reduces the need for extensive training from scratch and low-quality images and adds to the previous knowledge acquired from large datasets such as ImageNet.
The transfer learning model in this study is based on ViT, a deep learning architecture that has gained attention owing to its success in handling image classification tasks. The ViT model, originally pretrained on the ImageNet dataset, was customized for SCC image classification by modifying the architecture after the Multi-Layer Perceptron (MLP) head. Specifically, after the pretrained ViT layers, a flattening layer was added to convert the 2D features into a 1D vector, followed by a batch normalization layer to stabilize learning and prevent overfitting. In the final step, dense layers were added to allow the model to learn complex relationships within the data, followed by a SoftMax output layer to classify the images into two categories: margin negative (normal) and margin positive (tumor present). The architecture and design of the model, including additional layers, are shown in
Figure 4.
2.4. Implementation
The best model was trained at a learning rate of 0.0001 and was selected based on preliminary experiments to ensure stable convergence. The training was performed on a high-performance computing (HPC) cluster with NVIDIA GPUs (NVIDIA Corporation, Santa Clara, CA, USA), enabling the efficient processing of the computationally intensive ViT model and a large dataset of SCC images. The training was performed over 200 epochs, a duration found to be effective for capturing the intricate details of histopathological structures without overfitting. The model weights were optimized using the Adam optimizer, which was selected for its adaptability and efficiency in deep learning tasks. A batch size of 64 was used to balance the memory efficiency and training speed. Additionally, L2 regularization was applied in the form of weight decay, which aided in preventing the model from learning overly complex patterns that could lead to overfitting.
2.5. Performance Measures
In this study, several performance metrics were employed to evaluate the ability of the model to classify SCC margin images accurately. The values were assessed using five-fold cross-validation with a 95% confidence interval. These metrics included accuracy, precision, recall, F1 score, and AUC. Each metric provides a unique perspective on model performance, offering insights into how well the model balances between identifying true positives and minimizing false positives. For
TP, true positive;
TN, true negative;
FP, false positive;
FN, false negative, the metrics are given in Equations (1)–(4) below.
3. Results
To evaluate the proposed ViT-based approach for low-quality histopathological images in classifying SCC cases, we ran experiments involving different epochs, employing addition and removal layers, and compared them with the best-performing CNN models. Consequently, the proposed ViT-based transfer learning approach performed better than the CNNs and various additional layer combinations.
Table 1 presents the results of measuring the performance of the proposed method in terms of accuracy, AUC, F1 score, precision, and recall for various epoch numbers, including 50, 100, 150, 200, and 250 epochs. Among all epoch numbers utilized, the best model provided the highest performance results when running for 200 epoch, providing the most optimized performance. The proposed ViT-based SCC images classification provided the highest accuracy of 0.928 ± 0.027, AUC of 0.927 ± 0.028, F1 score of 0.926 ± 0.029, precision of 0.922 ± 0.028, and recall of 0.928 ± 0.027 running for 200 epochs.
Figure 5 shows the learning curve (the
Y-axis represents accuracy, and the
X-axis represents the number of epochs), confusion matrix, and ROC curve of the proposed model. From the figure, it can be deduced that the learning regimen was smooth, making the proposed model optimal for classifying SCC image data. The training, validation, and test accuracies were 0.998 ± 0.002, 0.937 ± 0.010, and 0.928 ± 0.027, respectively. The training, validation, and test losses were 0.346 ± 0.015, 0.389 ± 0.008, and 0.390 ± 0.002, respectively. The confusion matrix shows a small number of misclassified images, proving the effectiveness of the proposed model in classifying SCC images that were never observed during training. The ROC curve showed an almost perfect result, demonstrating the capability of the proposed method for low-quality histopathological image classification.
The proposed ViT-based method was compared with CNN-based methods to evaluate the effectiveness of the ViT models in classifying low-quality histopathology images relative to CNNs. To do this, the three best-performing CNN models chosen from preliminary studies were selected, trained, and evaluated on the same data distribution as that of the ViT-based model by keeping all parameters the same as the parameters chosen from the preliminary study and were all optimized parameters for both the ViT and CNN models. The results of these comparisons are presented in
Table 2. Based on these results, the highest performing CNN-based approach for low-quality histopathology image classification on the SCC dataset was InceptionV3 with the highest accuracy of 0.860 ± 0.049, AUC of 0.837 ± 0.029, F1 score of 0.854 ± 0.044, precision of 0.858 ± 0.057, and recall of 0.854 ± 0.039 running for 200 epochs. When compared to the ViT-based model that has the highest accuracy of 0.928 ± 0.027, AUC of 0.927 ± 0.028, F1 score of 0.926 ± 0.029, precision of 0.922 ± 0.028, and recall of 0.928 ± 0.027 running for 200 epochs, the ViT-based approach performed better than that of the CNN-based approach.
A comparison of the CNN-based approach was also performed using visual machine learning measures, including the learning curve, confusion matrix, and ROC curve.
Figure 6 shows the learning curve, confusion matrix, and ROC curve of the InceptionV3-based model (the highest performing CNN-based model) for classifying SCC images. Compared to the ViT-based approach depicted in
Figure 5, the ViT-based approach provided better results when evaluated visually.
Furthermore, an ablation study involving the use of different layer configurations was conducted to evaluate whether this affected the performance of the SCC image classification. To this end, we designed a ViT-based approach with three configurations, including the first configuration with pretrained ViT with only the output layer attached; configuration two with a pretrained ViT model with a flattening layer, batch normalization, dense layer, batch normalization, and output layer; and the third configuration with a pretrained ViT model with a flattening layer, batch normalization, two dense layers, batch normalization, and output layer, which were used for comparison with the proposed model. As shown in
Table 3, the proposed model outperformed all three configurations, although all three configurations provided better results than the best-performing CNN model, reinforcing the fact that ViT-based models outperformed CNN-based models in the classification of SCC images, irrespective of how the ViT-based approach was configured.
Figure 7 shows the differences in performance (in terms of accuracy (
Figure 7A) and AUC (
Figure 7B)) between the various ViT-based configurations employed for the ablation study, reinforcing the fact that the proposed combination is the most optimized and effective architecture of the ViT-based approach for low-quality histopathological image classification.
4. Discussion
These findings highlight the potential of ViT-based models for addressing critical diagnostic challenges in histopathological analysis. Traditional CNN-based models are the benchmark for image classification tasks but are notably sensitive to image quality, limiting their efficacy in resource-constrained settings. By leveraging the global attention mechanism of ViTs, this study achieved a significant leap in the diagnostic accuracy of SCC margin classification, particularly with low-quality images.
A standout aspect of the ViT model is its ability to capture long-range dependencies and structural features, which are often blurred or degraded in suboptimal images. This capability is crucial for SCC margin classification, in which the fine details of tissue morphology are vital for accurate diagnosis. Additionally, the proposed architecture, which incorporates flattening, batch normalization, and dense layers, demonstrates the importance of tailored configurations for specific tasks. The ablation study reinforced the idea that even if modifications in the ViT architecture were made, they did not result in substantial performance variations. This indicates that the ViT-based approach is highly effective for low-quality histopathological images.
In addition to these technical achievements, this study has profound implications for clinical practice in under-resourced environments. By reducing the reliance on high-quality imaging equipment and pathologist expertise, the proposed model bridges the gap between technological capabilities and clinical needs, enabling more equitable access to effective cancer diagnostics. Moreover, the results emphasized the potential of AI to enhance diagnostic precision while reducing variability among pathologists. This aligns with the broader goals of precision medicine, in which personalized and accurate treatment decisions are paramount. Therefore, collaboration between healthcare providers and policymakers is essential for scaling AI-driven diagnostics in global health systems. In general, this work not only contributes to the body of research on transformer models in medical imaging but also provides a foundation for future studies aimed at addressing the challenges of diagnostic accuracy in low-quality imaging conditions.
However, the computational demands of ViTs pose a barrier to their widespread adoption, particularly in low-resource settings lacking high-performance hardware. For instance, compared to EfficientNetB2, ViT is computationally expensive having 80 M trainable parameters (against 5 M for EfficientNetB2), requiring longer training time of 5 h over 50 epochs (against 3 h for EfficentNetB2), and having GFLOPs of 17 (against 2 for EfficientNetB2). Future studies should explore lightweight adaptations of ViT models or hybrid architectures that balance computational efficiency with diagnostic accuracy. Expanding the dataset to include diverse SCC subtypes and integrating real-world clinical workflows will further validate the utility of this model. The evaluation focused on a single SCC dataset, which restricted the generalizability of the results to other histopathological image types and cancers. In addition, although the model successfully classified low-quality images, further investigation is required to assess its performance under varying clinical imaging conditions. The black-box nature of transformers also presents a challenge for clinical adoption because their decision-making processes are non-interpretable. This limitation could impact the clinicians’ ability to trust and utilize these models in practice.
Future studies should focus on expanding the dataset to include diverse types of histopathological images to assess the generalizability of the model across various cancers and imaging settings. Further research on interpretability methods for transformers would also enhance clinical transparency and trust in these models. Hybrid architectures that integrate the strengths of CNNs and transformers can offer another pathway for improving the performance and interpretability of complex visual recognition tasks in medical imaging. Furthermore, the issue of class imbalance in the dataset must be resolved by considering patient-level data split for cross-validation and considering more evaluation approaches such as using precision–recall curves and others.
5. Conclusions
This study established the efficacy of a ViT-based model for classifying SCC margins in low-quality histopathological images, thus outperforming CNN-based approaches in terms of accuracy, precision, and robustness. These results underscore the transformative potential of ViTs in addressing the diagnostic challenges inherent in resource-constrained settings where suboptimal imaging conditions often prevail. Our findings have several important implications. The proposed ViT model enhances diagnostic accuracy and democratizes access to reliable cancer diagnostics. By mitigating the dependency on high-quality imaging and specialized expertise, this approach has the potential to improve surgical outcomes and reduce global cancer recurrence rates. However, this study has some limitations. The computational requirements of ViTs remain a significant hurdle, particularly in low-resource environments where access to high-performance hardware is limited. Although diverse, the dataset used in this study may not capture the full spectrum of SCC presentations observed in clinical practice. Expanding the dataset and validating the model across various patient populations and imaging conditions will strengthen its clinical applicability. Future studies should focus on the development of lightweight and hardware-efficient ViT variants for real-time deployment. Furthermore, integrating this model into existing clinical workflows and evaluating its impact on patient outcomes is critical. The exploration of hybrid AI models that combine the strengths of ViTs and CNNs could also provide valuable insights. Ultimately, this study paves the way for advancing AI-driven histopathological analysis, with implications extending beyond SCC to other cancers and diseases and offering a scalable solution to global healthcare challenges.
Author Contributions
Conceptualization, S.-y.P., G.A. and S.-w.C.; methodology, S.-y.P., G.A. and S.-w.C.; software, S.-y.P., G.A. and S.-w.C.; validation, S.-y.P., G.A. and S.-w.C.; formal analysis, S.-y.P., G.A. and S.-w.C.; investigation, S.-y.P., G.A. and S.-w.C.; resources, S.-w.C.; data curation, S.-y.P., G.A., B.D.W. and S.-w.C.; writing—original draft preparation, S.-y.P., G.A., S.-D.Y. and S.-w.C.; writing—review and editing, S.-y.P., G.A., B.D.W., K.C.J., S.-D.Y. and S.-w.C.; visualization, S.-y.P., G.A. and S.-w.C.; supervision, S.-w.C.; project administration, S.-w.C.; funding acquisition, S.-w.C. All authors have read and agreed to the published version of the manuscript.
Funding
This research was supported by the Basic Science Research Program of the National Research Foundation of Korea (NRF), funded by the Ministry of Education (RS-2023-00240521).
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
In this study, histopathological images obtained from the Jimma University Medical Center were used (data are publicly available and can be accessed at
https://osf.io/3ma4p/ (accessed on 28 August 2024)).
Conflicts of Interest
The authors declare no conflicts of interest. The funders had no influence on the study design, collection, analyses, interpretation of the data, writing of the manuscript, or decision to publish the results.
References
- Hasan, N.; Nadaf, A.; Imran, M.; Jiba, U.; Sheikh, A.; Almalki, W.H.; Almujri, S.S.; Mohammed, Y.H.; Kesharwani, P.; Ahmad, F.J. Skin Cancer: Understanding the Journey of Transformation from Conventional to Advanced Treatment Approaches. Mol. Cancer 2023, 22, 168. [Google Scholar] [CrossRef]
- Khan, N.H.; Mir, M.; Qian, L.; Baloch, M.; Khan, M.F.A.; Rehman, A.-U.; Ngowi, E.E.; Wu, D.-D.; Ji, X.-Y. Skin Cancer Biology and Barriers to Treatment: Recent Applications of Polymeric Micro/Nanostructures. J. Adv. Res. 2022, 36, 223–247. [Google Scholar] [CrossRef]
- Paulson, K.G.; Gupta, D.; Kim, T.S.; Veatch, J.R.; Byrd, D.R.; Bhatia, S.; Wojcik, K.; Chapuis, A.G.; Thompson, J.A.; Madeleine, M.M.; et al. Age-Specific Incidence of Melanoma in the United States. JAMA Dermatol. 2020, 156, 57–64. [Google Scholar] [CrossRef]
- Linos, E.; Chren, M.; Cenzer, I.S.; Covinsky, K.E. Skin Cancer in U.S. Elderly Adults: Does Life Expectancy Play a Role in Treatment Decisions? J. Am. Geriatr. Soc. 2016, 64, 1610–1615. [Google Scholar] [CrossRef]
- Kohnehshahri, M.K.; Sarkesh, A.; Khosroshahi, L.M.; HajiEsmailPoor, Z.; Aghebati-Maleki, A.; Yousefi, M.; Aghebati-Maleki, L. Current Status of Skin Cancers with a Focus on Immunology and Immunotherapy. Cancer Cell Int. 2023, 23, 174. [Google Scholar] [CrossRef]
- Dettrick, A.; Foden, N.; Hogan, D.; Azer, M.; Blazak, J.; Atwell, D.; Buddle, N.; Min, M.; Livingston, R.; Banney, L.; et al. The Hidden Australian Skin Cancer Epidemic, High-Risk Cutaneous Squamous Cell Carcinoma: A Narrative Review. Pathology 2024, 56, 619–632. [Google Scholar] [CrossRef]
- Roky, A.H.; Islam, M.M.; Ahasan, A.M.F.; Mostaq, M.S.; Mahmud, M.Z.; Amin, M.N.; Mahmud, M.A. Overview of Skin Cancer Types and Prevalence Rates across Continents. Cancer Pathog. Ther. 2024, 2, E01–E36. [Google Scholar] [CrossRef]
- Filippini, D.M.; Carosi, F.; Querzoli, G.; Fermi, M.; Ricciotti, I.; Molteni, G.; Presutti, L.; Foschini, M.P.; Locati, L.D. Rare Head and Neck Cancers and Pathological Diagnosis Challenges: A Comprehensive Literature Review. Diagnostics 2024, 14, 2365. [Google Scholar] [CrossRef]
- Zürcher, S.; Martignoni, Z.; Hunger, R.E.; Benzaquen, M.; Seyed Jafari, S.M. Mohs Micrographic Surgery for Cutaneous Squamous Cell Carcinoma. Cancers 2024, 16, 2394. [Google Scholar] [CrossRef]
- Verdaguer-Faja, J.; Toll, A.; Boada, A.; Guerra-Amor, Á.; Ferrándiz-Pulido, C.; Jaka, A. Management of Cutaneous Squamous Cell Carcinoma of the Scalp: The Role of Imaging and Therapeutic Approaches. Cancers 2024, 16, 664. [Google Scholar] [CrossRef]
- Levy, J.J.; Davis, M.J.; Chacko, R.S.; Davis, M.J.; Fu, L.J.; Goel, T.; Pamal, A.; Nafi, I.; Angirekula, A.; Suvarna, A.; et al. Intraoperative Margin Assessment for Basal Cell Carcinoma with Deep Learning and Histologic Tumor Mapping to Surgical Site. npj Precis. Oncol. 2024, 8, 2. [Google Scholar] [CrossRef]
- Hanna, M.G.; Ardon, O.; Reuter, V.E.; Sirintrapun, S.J.; England, C.; Klimstra, D.S.; Hameed, M.R. Integrating Digital Pathology into Clinical Practice. Mod. Pathol. 2022, 35, 152–164. [Google Scholar] [CrossRef] [PubMed]
- Gurcan, M.N.; Boucheron, L.E.; Can, A.; Madabhushi, A.; Rajpoot, N.M.; Yener, B. Histopathological Image Analysis: A Review. IEEE Rev. Biomed. Eng. 2009, 2, 147–171. [Google Scholar] [CrossRef]
- Luan, H.; Yang, K.; Hu, T.; Hu, J.; Liu, S.; Li, R.; He, J.; Yan, R.; Guo, X.; Qian, N.; et al. Review of Deep Learning-Based Pathological Image Classification: From Task-Specific Models to Foundation Models. Futur. Gener. Comput. Syst. 2025, 164, 107578. [Google Scholar] [CrossRef]
- Baxi, V.; Edwards, R.; Montalto, M.; Saha, S. Digital Pathology and Artificial Intelligence in Translational Medicine and Clinical Practice. Mod. Pathol. 2022, 35, 23–32. [Google Scholar] [CrossRef]
- Meyerholz, D.K.; Beck, A.P. Principles and Approaches for Reproducible Scoring of Tissue Stains in Research. Lab. Investig. 2018, 98, 844–855. [Google Scholar] [CrossRef]
- Kreiss, L.; Jiang, S.; Li, X.; Xu, S.; Zhou, K.C.; Lee, K.C.; Mühlberg, A.; Kim, K.; Chaware, A.; Ando, M.; et al. Digital Staining in Optical Microscopy Using Deep Learning—A Review. PhotoniX 2023, 4, 34. [Google Scholar] [CrossRef]
- Bai, B.; Yang, X.; Li, Y.; Zhang, Y.; Pillar, N.; Ozcan, A. Deep Learning-Enabled Virtual Histological Staining of Biological Samples. Light Sci. Appl. 2023, 12, 57. [Google Scholar] [CrossRef]
- Haghighat, M.; Browning, L.; Sirinukunwattana, K.; Malacrino, S.; Alham, N.K.; Colling, R.; Cui, Y.; Rakha, E.; Hamdy, F.C.; Verrill, C.; et al. Automated Quality Assessment of Large Digitised Histology Cohorts by Artificial Intelligence. Sci. Rep. 2022, 12, 5002. [Google Scholar] [CrossRef]
- Rahman, T.Y.; Mahanta, L.B.; Chakraborty, C.; Das, A.K.; Sarma, J.D. Textural Pattern Classification for Oral Squamous Cell Carcinoma. J. Microsc. 2018, 269, 85–93. [Google Scholar] [CrossRef]
- Wako, B.D.; Dese, K.; Ulfata, R.E.; Nigatu, T.A.; Turunbedu, S.K.; Kwa, T. Squamous Cell Carcinoma of Skin Cancer Margin Classification From Digital Histopathology Images Using Deep Learning. Cancer Control 2022, 29, 107327482211325. [Google Scholar] [CrossRef]
- Ahmed, A.A.; Abouzid, M.; Kaczmarek, E. Deep Learning Approaches in Histopathology. Cancers 2022, 14, 5264. [Google Scholar] [CrossRef]
- Jeong, H.K.; Park, C.; Jiang, S.W.; Nicholas, M.; Chen, S.; Henao, R.; Kheterpal, M. Image Quality Assessment Using Convolutional Neural Network in Clinical Skin Images. JID Innov. 2024, 4, 100285. [Google Scholar] [CrossRef]
- Tang, S.; Jing, C.; Jiang, Y.; Yang, K.; Huang, Z.; Wu, H.; Cui, C.; Shi, S.; Ye, X.; Tian, H.; et al. The Effect of Image Resolution on Convolutional Neural Networks in Breast Ultrasound. Heliyon 2023, 9, e19253. [Google Scholar] [CrossRef] [PubMed]
- Hu, C.; Sapkota, B.B.; Thomasson, J.A.; Bagavathiannan, M.V. Influence of Image Quality and Light Consistency on the Performance of Convolutional Neural Networks for Weed Mapping. Remote Sens. 2021, 13, 2140. [Google Scholar] [CrossRef]
- Atabansi, C.C.; Nie, J.; Liu, H.; Song, Q.; Yan, L.; Zhou, X. A Survey of Transformer Applications for Histopathological Image Analysis: New Developments and Future Directions. Biomed. Eng. Online 2023, 22, 96. [Google Scholar] [CrossRef]
- Maurício, J.; Domingues, I.; Bernardino, J. Comparing Vision Transformers and Convolutional Neural Networks for Image Classification: A Literature Review. Appl. Sci. 2023, 13, 5521. [Google Scholar] [CrossRef]
- Ayana, G.; Barki, H.; Choe, S. Pathological Insights: Enhanced Vision Transformers for the Early Detection of Colorectal Cancer. Cancers 2024, 16, 1441. [Google Scholar] [CrossRef]
- Raghu, M.; Unterthiner, T.; Kornblith, S.; Zhang, C.; Dosovitskiy, A. Do Vision Transformers See Like Convolutional Neural Networks? Adv. Neural Inf. Process. Syst. 2021, 34, 12116–12128. [Google Scholar]
- Ayana, G.; Lee, E.; Choe, S. Vision Transformers for Breast Cancer Human Epidermal Growth Factor Receptor 2 Expression Staging without Immunohistochemical Staining. Am. J. Pathol. 2023, 194, 402–414. [Google Scholar] [CrossRef]
- Komura, D.; Ishikawa, S. Machine Learning Methods for Histopathological Image Analysis. Comput. Struct. Biotechnol. J. 2018, 16, 34–42. [Google Scholar] [CrossRef] [PubMed]
- Mezei, T.; Kolcsár, M.; Joó, A.; Gurzu, S. Image Analysis in Histopathology and Cytopathology: From Early Days to Current Perspectives. J. Imaging 2024, 10, 252. [Google Scholar] [CrossRef] [PubMed]
- Hoque, M.Z.; Keskinarkaus, A.; Nyberg, P.; Seppänen, T. Stain Normalization Methods for Histopathology Image Analysis: A Comprehensive Review and Experimental Comparison. Inf. Fusion 2024, 102, 101997. [Google Scholar] [CrossRef]
- McGenity, C.; Clarke, E.L.; Jennings, C.; Matthews, G.; Cartlidge, C.; Freduah-Agyemang, H.; Stocken, D.D.; Treanor, D. Artificial Intelligence in Digital Pathology: A Systematic Review and Meta-Analysis of Diagnostic Test Accuracy. npj Digit. Med. 2024, 7, 114. [Google Scholar] [CrossRef]
- Madusanka, N.; Jayalath, P.; Fernando, D.; Yasakethu, L.; Lee, B.-I. Impact of H&E Stain Normalization on Deep Learning Models in Cancer Image Classification: Performance, Complexity, and Trade-Offs. Cancers 2023, 15, 4144. [Google Scholar] [CrossRef]
- Asaf, M.Z.; Rao, B.; Akram, M.U.; Khawaja, S.G.; Khan, S.; Truong, T.M.; Sekhon, P.; Khan, I.J.; Abbasi, M.S. Dual Contrastive Learning Based Image-to-Image Translation of Unstained Skin Tissue into Virtually Stained H&E Images. Sci. Rep. 2024, 14, 2335. [Google Scholar] [CrossRef]
- Cong, C.; Liu, S.; Di Ieva, A.; Pagnucco, M.; Berkovsky, S.; Song, Y. Colour Adaptive Generative Networks for Stain Normalisation of Histopathology Images. Med. Image Anal. 2022, 82, 102580. [Google Scholar] [CrossRef]
- Shavlokhova, V.; Sandhu, S.; Flechtenmacher, C.; Koveshazi, I.; Neumeier, F.; Padrón-Laso, V.; Jonke, Ž.; Saravi, B.; Vollmer, M.; Vollmer, A.; et al. Deep Learning on Oral Squamous Cell Carcinoma Ex Vivo Fluorescent Confocal Microscopy Data: A Feasibility Study. J. Clin. Med. 2021, 10, 5326. [Google Scholar] [CrossRef]
- Li, H.; Zhang, Y.; Chen, P.; Shui, Z.; Zhu, C.; Yang, L. Rethinking Transformer for Long Contextual Histopathology Whole Slide Image Analysis. arXiv 2024, arXiv:2410.14195. [Google Scholar]
- Pezoulas, V.C.; Zaridis, D.I.; Mylona, E.; Androutsos, C.; Apostolidis, K.; Tachos, N.S.; Fotiadis, D.I. Synthetic Data Generation Methods in Healthcare: A Review on Open-Source Tools and Methods. Comput. Struct. Biotechnol. J. 2024, 23, 2892–2910. [Google Scholar] [CrossRef]
- Islam, T.; Hafiz, M.S.; Jim, J.R.; Kabir, M.M.; Mridha, M.F. A Systematic Review of Deep Learning Data Augmentation in Medical Imaging: Recent Advances and Future Research Directions. Healthc. Anal. 2024, 5, 100340. [Google Scholar] [CrossRef]
- Rahat, F.; Hossain, M.S.; Ahmed, M.R.; Jha, S.K.; Ewetz, R. Data Augmentation for Image Classification Using Generative AI. arXiv 2024, arXiv:2409.00547. [Google Scholar]
- Salvi, M.; Branciforti, F.; Molinari, F.; Meiburger, K.M. Generative Models for Color Normalization in Digital Pathology and Dermatology: Advancing the Learning Paradigm. Expert Syst. Appl. 2024, 245, 123105. [Google Scholar] [CrossRef]
- Albalawi, E.; Thakur, A.; Ramakrishna, M.T.; Khan, S.B.; SankaraNarayanan, S.; Almarri, B.; Hadi, T.H. Oral Squamous Cell Carcinoma Detection Using EfficientNet on Histopathological Images. Front. Med. 2024, 10, 1349336. [Google Scholar] [CrossRef]
- Hetz, M.J.; Bucher, T.-C.; Brinker, T.J. Multi-Domain Stain Normalization for Digital Pathology: A Cycle-Consistent Adversarial Network for Whole Slide Images. Med. Image Anal. 2024, 94, 103149. [Google Scholar] [CrossRef]
- Ochi, M.; Komura, D.; Onoyama, T.; Shinbo, K.; Endo, H.; Odaka, H.; Kakiuchi, M.; Katoh, H.; Ushiku, T.; Ishikawa, S. Registered Multi-Device/Staining Histology Image Dataset for Domain-Agnostic Machine Learning Models. Sci. Data 2024, 11, 330. [Google Scholar] [CrossRef]
- Mudeng, V.; Farid, M.N.; Ayana, G.; Choe, S. Domain and Histopathology Adaptations–Based Classification for Malignancy Grading System. Am. J. Pathol. 2023, 193, 2080–2098. [Google Scholar] [CrossRef]
- Ayana, G.; Dese, K.; Abagaro, A.M.; Jeong, K.C.; Yoon, S.-D.; Choe, S. Multistage Transfer Learning for Medical Images. Artif. Intell. Rev. 2024, 57, 232. [Google Scholar] [CrossRef]
- Xu, S.; Xiang, S.; Meng, F.; Wu, Q. Enhancing Unsupervised Domain Adaptation for Person Re-Identification with the Minimal Transfer Cost Framework. Comput. Mater. Contin. 2024, 80, 4197–4218. [Google Scholar] [CrossRef]
- Guichemerre, A.; Belharbi, S.; Mayet, T.; Murtaza, S.; Shamsolmoali, P.; McCaffrey, L.; Granger, E. Source-Free Domain Adaptation of Weakly-Supervised Object Localization Models for Histology. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 17–18 June 2024. [Google Scholar]
- Roy, K.; Banik, D.; Bhattacharjee, D.; Nasipuri, M. Patch-Based System for Classification of Breast Histology Images Using Deep Learning. Comput. Med. Imaging Graph. 2019, 71, 90–103. [Google Scholar] [CrossRef]
- Park, Y.; Kim, M.; Ashraf, M.; Ko, Y.S.; Yi, M.Y. MixPatch: A New Method for Training Histopathology Image Classifiers. Diagnostics 2022, 12, 1493. [Google Scholar] [CrossRef] [PubMed]
- Ciga, O.; Xu, T.; Nofech-Mozes, S.; Noy, S.; Lu, F.-I.; Martel, A.L. Overcoming the Limitations of Patch-Based Learning to Detect Cancer in Whole Slide Images. Sci. Rep. 2021, 11, 8894. [Google Scholar] [CrossRef] [PubMed]
- Kaczmarzyk, J.R.; Gupta, R.; Kurc, T.M.; Abousamra, S.; Saltz, J.H.; Koo, P.K. ChampKit: A Framework for Rapid Evaluation of Deep Neural Networks for Patch-Based Histopathology Classification. Comput. Methods Programs Biomed. 2023, 239, 107631. [Google Scholar] [CrossRef] [PubMed]
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).