Deep Learning Approaches for the Classification of Keloid Images in the Context of Malignant and Benign Skin Disorders
Abstract
:1. Introduction
Previous Work
- i.
- To train various CNN models to classify keloid images as benign lesions and not as malignant lesions (and for this, we use a large variety of skin lesion images, from keloids to melanoma, basal cell carcinoma, squamous cell carcinoma, seborrheic keratosis, nevus, etc., which could be either malignant or benign);
- ii.
- To train the CNN models to classify images of various skin disorders in three separate classes: malignant, benign, or keloid;
- iii.
- The identification and classification of keloid lesions.
- Using non-dermatoscopic (clinical) images.
2. Materials and Methods
2.1. Data
- The first dataset was taken from [34] and was obtained using a smartphone-based application created to help doctors and medical students collect clinical photos of skin lesions, as well as clinical information about the patient. Each dataset sample contains up to 26 attributes: a clinical diagnosis; a picture of the lesion; the patient’s age; the location of the lesion; if the lesion itches, bleeds, or has bled, hurts, has recently risen, has altered its pattern, is elevated, etc. The dataset contains 2298 images taken from 1373 patients between 2018 and 2019, which are available in Portable Network Graphics (PNGs) in raw format (i.e., just as they were taken). Each image in this dataset focused on one of the following six common skin disorders (also summarised in Table 2): melanoma, basal cell carcinoma, squamous cell carcinoma (which are considered malignant), actinic keratosis, seborrheic keratosis, and nevus (which are considered benign). We emphasise that 58.4% of the skin lesions and 100% of the skin cancers in this dataset are histopathology-confirmed (and this is about the same percentage as for the ISIC [43] dataset). For more information on the data, see [34].
- The second dataset is taken from [44] and consists of different non-dermatoscopic keloid images; we chose 274 keloid images and cropped them to ensure that these images are consistent with the (zoomed-in) images in the first dataset. Figure 2 shows a sample of the (keloid) images in this second dataset [44].
2.2. Deep Learning Approaches
2.2.1. Data Preprocessing
2.2.2. Model Details
- VGG16: VGG16 was proposed in [47] by the Visual Geometry Group (VGG) at the University of Oxford. It consists of 13 convolutional layers (with filter sizes), 5 max-pooling layers, and 3 dense (i.e., fully connected) layers including the output layer. The rectified linear unit (ReLU) activation function is used for each convolutional and dense layer except for the output layer, which uses the “softmax” activation function. Each convolutional layer use a convolutional filter with a stride of 1 and same padding, which makes the resulting feature maps retain the same spatial dimensions as the input. The convolutional layers are stacked on each other, with the number of input filters doubled after each max-pooling layer (with a stride of 2). The depth of the convolutional layers is also increased monotonically. Figure 4 and Table 3 show a summary of the VGG16 architecture.
- DenseNet121: This model, which was proposed in [49], is a feedforward network that connects all layers to all other layers in a feedforward fashion. It comprises 121 layers and consists of densely connected convolutional layers within dense blocks, promoting feature reuse and gradient flow. The model also consists of transition layers which help to control the growth of complexity between blocks. Figure 5 and Table 4 show a summary of the DenseNet121 architecture.
- InceptionV3: InceptionV3 is an extension of GoogleNet (a model that has been shown to demonstrate strong classification performance in some biological applications [68,69]). InceptionV3 uses the inception model to minimise the number of parameters required to be trained, thus lowering the computational cost by concatenating numerous convolutional filters of various sizes into a new filter. Figure 6 and Table 5 summarise the InceptionV3 architecture.
- MobileNet: The MobileNet model uses depthwise separable convolutions (a form of factorised convolutions that factorise a classical convolution into a depthwise convolution and a convolution called a pointwise convolution) to reduce computational cost and model size. In one step of these convolutions, the depthwise convolution applies a single filter per input channel, and the pointwise convolution applies a convolution to combine the outputs of the depthwise convolution. For more details on MobileNet, see [50]. Table 6 summarises the MobileNet architecture.
- EfficientNetB0: EfficientNet is a family of CNNs that proposed a new scaling method after carefully identifying that better accuracy could be achieved when the depth, width, and resolution of the network are carefully balanced. The proposed scaling approach uses a simple compound coefficient to uniformly scale the depth, width, and resolution to reduce the computational cost and the model size. EfficientNet consists of 8 models in the range B0–B7, where the B1–B7 models are scaled increasingly from the baseline model using different compound coefficients. The baseline model EfficientNetB0 is based on mobile inverted bottleneck convolution (MBConv) blocks [70]. In this study, we focus only on the baseline model (EfficientNetB0) due to limited computational resources. For more details on EfficientNet, see [51]. Table 7 summarises the EfficientNetB0 architecture.
- Xception: The Xception model uses depthwise separable convolutions like MobileNets and the Inception models described above. However, this model is based on the hypothesis that cross-channels correlations and spatial correlations mappings in the feature maps of CNNs can be decoupled entirely. This is a stronger assumption than that of the Inception models, and hence the name “Xception” that derived from the phrase “Extreme Inception”. For more details on the Xception model, see [52]. Table 8 summarises the Xception architecture.
- InceptionRNV2: This is the second verion of the InceptionResNet model. It is largely similar to the Inception model already described above (see InceptionV3). However, in InceptionResNet (i.e., a residual version of the Inception model), a cheaper Inception block is used preceded by a filter expansion layer which scales up the dimension of filter bank, hence cushioning the effect of the the dimensionality reduction caused by the Inception block. Another distinction between the InceptionResNet models and the vanilla Inception models is that batch normalisation is applied only to the standard layers not the summations, hereby increasing the overall number of Inception blocks. In this study, we used the second version of the InceptionResNet model. For more details on the InceptionResNet model, see [53]. Table 9 summarises the InceptionResNetV2 architecture.
- EfficientNetV2-L: This is an extension of the EfficientNet models described above (see 5) with faster training time, because it uses about half the number of parameters used in EfficientNet. In this new family of models, the scaling approach introduced in [51] was combined with a training-aware neural architecture search (NAS) to optimise the training speed and number of parameters. Unlike the original EfficientNet, which uses depthwise separable convolutions, the fused-MBConv block fuses the initial pointwise and depthwise convolutions, reducing the computational cost. As opposed to the vanilla EfficientNet, EfficienNetV2 utilises both MBConv and the fused-MBConv [71] in the early layers. This new family comprises three variants (i.e., small, medium, and large) based on their size and performance. In this study, we made use of the largest variant (i.e., EfficientNetV2-L). For more details on EfficientNetV2, see [54]. Table 10 summarises the EfficientNetV2-L architecture.
- NASNet-L: This model was designed using a proposed search space called NAS (that enables transferability) to find optimal network architectures. It utilises reinforcement learning to explore a preset search space of architectures while optimising performance and efficiency and also using a regularisation technique called “ScheduledDropPath”. It has different variants, NASNet-A, B, and C, tailored for different use cases (where A is the most accurate and was designed to deliver high performance, while B and C provide a trade-off between efficiency and accuracy, with B being more accurate than C). The model also includes large and small versions depending on the resources available. The large model (i.e., NASNet-L) is particularly effective for high-performance, while the small model (i.e., NASNetMobile) is particularly effective for resource-constrained environments like mobile devices. In this study, we used NASNet-L. For more details on NASNet models, see [55]. Table 11 summarises the NASNet-L architecture.
- Freezing the base model (i.e., training only the weights and biases of the added model).
- Fine-tuning the base model by unfreezing all layers of the base model (i.e., training the weights and biases of the base model alongside the added model).
- Training the model on the original train data after splitting it into train, validation, and test datasets.
- Training the model on an oversampled train dataset.
- Training the model on an augmented train dataset.
2.2.3. Evaluation Metrics
3. Results
3.1. Binary Classification: Benign vs. Malignant Lesions
3.2. Identifying Keloids as Keloids and Not Only as Benign vs. Malignant Skin Disorders
- (a)
- Original training data: In Table 18, we present the performance of the 10 models considered in this study (on test data) when they were trained for 50 epochs on the original training dataset validated on the original validation dataset (while fine-tuning all layers of the base models). We see here that VGG16 equally outperformed the rest of the model on all metrics, followed by Xception and DensNet121, respectively.
- (b)
- Oversampled training data: In Table 19, we present the performance of the trained model on oversampled data. Here, we see that the performance of each of the models improved in comparison to Table 18. Here again, VGG16 performed better than the rest of the models on all metrics followed by Xception (with a slightly lower AUC that DenseNet121) and DenseNet121.
- (c)
- Augmented training data: In an attempt to further improve the result of the models, instead of random oversampling, we applied data augmentation (such as rotations, flips, zooms, etc.) to increase the training dataset to introduce variability and to help the model generalise better. In Table 20, we see an increase in the performance of all the models as expected in comparison to Table 18 and Table 19. We see again that VGG16 outperformed the rest of the models on all metrics followed by EfficientNetV2L and InceptionV3, which both had lower AUC ROC scores compared to DenseNet121, which had the second best AUC ROC score, while its other performances ranked below InceptionV3.We emphasise here that out of the pretrained CNN models used, MobileNet model was the worst-performing model on the test data considered in this study.
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.8276 | 0.8301 | 0.8276 | 0.8276 | 0.9376 |
MobileNet | 0.7471 | 0.7534 | 0.7471 | 0.7480 | 0.8914 |
DenseNet121 | 0.7931 | 0.7942 | 0.7931 | 0.7922 | 0.9305 |
InceptionV3 | 0.7816 | 0.7841 | 0.7816 | 0.7823 | 0.9095 |
EfficientNetB0 | 0.7624 | 0.7663 | 0.7624 | 0.7632 | 0.8925 |
Xception | 0.8199 | 0.8245 | 0.8199 | 0.8202 | 0.9309 |
InceptionRNV2 | 0.7854 | 0.7854 | 0.7854 | 0.7854 | 0.9091 |
EfficientNetV2L | 0.7778 | 0.7843 | 0.7778 | 0.7743 | 0.9111 |
NASNetLarge | 0.6743 | 0.7266 | 0.6743 | 0.6685 | 0.8665 |
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.8506 | 0.8530 | 0.8506 | 0.8502 | 0.9427 |
MobileNet | 0.7625 | 0.7650 | 0.7625 | 0.7634 | 0.9032 |
DenseNet121 | 0.8084 | 0.8101 | 0.8084 | 0.8091 | 0.9295 |
InceptionV3 | 0.8008 | 0.8056 | 0.8008 | 0.8015 | 0.9205 |
EfficientNetB0 | 0.7969 | 0.7989 | 0.7969 | 0.7969 | 0.9148 |
Xception | 0.8199 | 0.8239 | 0.8199 | 0.8205 | 0.9282 |
InceptionRNV2 | 0.8007 | 0.8036 | 0.8008 | 0.8008 | 0.9215 |
EfficientNetV2L | 0.7969 | 0.7983 | 0.7969 | 0.7973 | 0.9179 |
NASNetLarge | 0.7050 | 0.7222 | 0.7050 | 0.7061 | 0.8728 |
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.8774 | 0.8813 | 0.8774 | 0.8778 | 0.9519 |
MobileNet | 0.8046 | 0.8045 | 0.8046 | 0.8046 | 0.9240 |
DenseNet121 | 0.8391 | 0.8399 | 0.8391 | 0.8394 | 0.9452 |
InceptionV3 | 0.8467 | 0.8474 | 0.8467 | 0.8470 | 0.9403 |
EfficientNetB0 | 0.8429 | 0.8422 | 0.8429 | 0.8424 | 0.9406 |
Xception | 0.8046 | 0.8072 | 0.8046 | 0.8046 | 0.9334 |
InceptionRNV2 | 0.8199 | 0.8222 | 0.8199 | 0.8206 | 0.9359 |
EfficientNetV2L | 0.8467 | 0.8502 | 0.8467 | 0.8470 | 0.9322 |
NASNetLarge | 0.8391 | 0.8410 | 0.8391 | 0.8394 | 0.9365 |
3.3. New Test Data: Clinical Images
3.4. Keloid vs. Similar-Looking Malignant Lesions
4. Discussion and Research Limitation
4.1. Discussion
4.2. Research Limitations
- The image dataset used in this study contained non-dermatoscopic (clinical) images of keloids, some malignant skin cancers/lesions, and other benign lesions (including melanomas, basal cell carcinomas, squamous cell carcinomas, seborrheic keratosis, nevus, and actinic keratosis); hence, it may not perform well when tested on other skin lesions not present in training data or dermatoscopic images of skin lesions present in training data.
- In addition, as we mentioned in Section 2.1, not all data we used were pathologically validated (especially the non-cancerous lesions), which could have impacted the classification results we obtained.
- In Remark 1, we highlighted that the models performed poorly on images taken at a long distance from the skin lesion, as the models were trained on images focused on the skin lesions. Also, zoomed-in images of the same original pictures led to blurred images and possible misclassification.
- Lastly, as previously mentioned in Remark 1, the trained data were relatively small in number and might not contain all possible anatomical regions where such lesions can occur (e.g., the eye or close to the eye; see Figure 1). Note that these aspects can impact the performance of the model, as exemplified in Figure 10, where the algorithm classified the lesion as “malignant” (see Figure 10a) when the lower part of the iris and the sclera is obvious in the zoom-in, while classifying the lesion as “benign“ (see Figure 10b) when only the lower part of the sclera is obvious in the zoom-in.
5. Conclusions
6. Comparison with Published Literature
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Appendix A. Details of the Performance Metrics
- Accuracy, which measures the proportion of correct predictions made by the model defined as
- Precision, which measures the ratio of predictions among all positive predictions ( and ), is given by
- Recall (also known as sensitivity or TP rate), which measures the proportion of correctly predicted positive cases (TP) out of all actual positive cases (TP + FN), is given by
- A metric that combines the precision and recall is the . It provides a good evaluation of model performance and is defined below as
- We can also evaluate the performance of each model using the Area Under the Receiver Operating Characteristic Curve (AUC-ROC). The AUC-ROC measures the ability of a classifier to differentiate between the positive and negative classes across different threshold values, while the ROC curve itself is a plot of the FP rate (x axis) against the TP rate (i.e., the recall on the y axis).
- The confusion matrix is another (visual) metric that describes the performance of a classification model by comparing the actual values with the predicted values.
Appendix B. Hardware and Software Specification
References
- Behera, D.; Debata, I.; Mohapatra, D.; Agarwal, A. Appearances can be deceptive: An intriguing case of a keloidal mass. Indian J. Dermatol. 2022, 67, 629. [Google Scholar] [CrossRef] [PubMed]
- Grant-Kels, J.M.; Bason, E.T.; Grin, C.M. The misdiagnosis of malignant melanoma. J. Am. Acad. Dermatol. 1999, 40, 539–548. [Google Scholar] [CrossRef] [PubMed]
- Hwang, S.-M.; Pan, H.-C.; Hwang, M.-K.; Kim, M.-W.; Lee, J.-S. Malignant skin tumor misdiagnosed as a benign skin lesion. Arch. Craniofac. Surg. 2016, 17, 86–89. [Google Scholar] [CrossRef]
- Jia, J.; Wang, M.; Song, L.; Feng, Y. A melanotic malignant melanoma presenting as a keloid. A case report. Medicine 2017, 96, e9047. [Google Scholar] [CrossRef] [PubMed]
- Nicholas, R.S.; Stodell, M. An important case. BMJ Case Rep. 2014, 2014, bcr2014203600. [Google Scholar] [CrossRef]
- Sondermann, W.; Zimmer, L.; Schadendorf, D.; Roesch, A.; Klode, J.; Dissemond, J. Initial misdiagnosis of melanoma located on the foot is associated with poorer prognosis. Medicine 2016, 95, e4332. [Google Scholar] [CrossRef]
- Ucak, M. A rare case of misdiagnosis: Recurrence of dermatofibrosarcoma protuberans that was treated surgically as keloid. Med. Arch. 2018, 72, 74–75. [Google Scholar] [CrossRef]
- Wang, Y.; Yang, F.; Wu, X.L.; Liu, F.; Jin, R.; Gu, C.; Ni, T.; Wang, X.; Yang, Q.; Luo, X.; et al. On the diagnosis and treatment of cutaneous and soft tissue tumours misdiagnosed as scars: Lessons from four cases. Int. Wound J. 2019, 16, 793–799. [Google Scholar] [CrossRef]
- Newcomer, J.B.; Durbin, A.; Wilson, C. Cutaneous Metastasis of Lung Adenocarcinoma: Initially Mimicking and Misdiagnosed as Keloids. Cureus 2022, 14, e27285. [Google Scholar] [CrossRef]
- Niessen, F.B.; Spauwen, P.H.; Schalkwijk, J.; Kon, M. On the nature of hypertrophic scars and keloids: A review. Plast. Reconstr. Surg. 1999, 104, 1435–1458. [Google Scholar] [CrossRef]
- Vincent, A.S.; Phan, T.T.; Mukhopadhyay, A.; Lim, H.Y.; Halliwell, B.; Wong, K.P. Human skin keloid fibroblasts display bioenergetics of cancer cells. J. Investig. Dermatol. 2008, 128, 702–709. [Google Scholar] [CrossRef] [PubMed]
- Gauglitz, G.G.; Korting, H.C.; Pavicic, T.; Ruzicka, T.; Jeschke, M.G. Hypertrophic scarring and keloids: Pathomechanisms and current and emerging treatment strategies. Mol. Med. 2011, 17, 113–125. [Google Scholar] [CrossRef] [PubMed]
- Satish, L.; Lyons-Weiler, J.; Hebda, P.A.; Wells, A. Gene expression patterns in isolated keloid fibroblasts. Wound Repair Regen. 2006, 14, 463–470. [Google Scholar] [CrossRef]
- Eftimie, R.; Rollin, G.; Adebayo, O.; Urcun, S.; Bordas, S.P.A. Modelling keloids dynamics: A brief review and new mathematical perspectives. Bull. Math. Biol. 2023, 85, 117. [Google Scholar] [CrossRef]
- Bostanci, S.; Akay, B.N.; Alizada, M.; Okçu, A.H. Cutaneous leiomyosarcoma misdiagnosed as a keloid: A case report with dermatoscopy. Turkderm.-Turk. Arch. Dermatol. Venereol. 2023, 57, 154–156. [Google Scholar] [CrossRef]
- Portugal, E.H.; Alves, J.C.R.R.; Da Fonseca, R.P.L.; De Souza Andrade, J.; De Melo Alemida, A.C.; Araujo, I.C.; Pereira, N.A.; Da Silva, R.L.F. Dermatofibrosarcoma protuberans misdiagnosed as keloid and treated with triamcinolone acetonid. Rev. Bras. Cir. Plast. 2016, 31, 82–87. [Google Scholar] [CrossRef]
- Tiong, W.; Basiron, N. Challenging Diagnosis of a Rare Case of Spontaneous Keloid Scar. J. Med. Cases 2014, 5, 466–469. [Google Scholar] [CrossRef]
- Lallas, A.; Paschou, E.; Manoli, S.M.; Papageorgiou, C.; Spyridis, I.; Liopyris, K.; Bobos, M.; Moutsoudis, A.; Lazaridou, E.; Apalla, Z. Dermatoscopy of melanoma according to type, anatomic site and stage. Ital. J. Dermatol. Venereol. 2021, 156, 274–288. [Google Scholar] [CrossRef]
- Puig, S.; Cecilia, N.; Malvehy, J. Dermoscopic criteria and basal cell carcinoma. G. Ital. Dermatol. Venereol. 2012, 147, 135–140. [Google Scholar]
- Sgouros, D.; Theofili, M.; Zafeiropoulou, T.; Lallas, A.; Apalla, Z.; Zaras, A.; Liopyris, K.; Pappa, G.; Polychronaki, E.; Kousta, F.; et al. Dermoscopy of Actinic Keratosis: Is There a True Differentiation between Non-Pigmented and Pigmented Lesions? J. Clin. Med. 2023, 12, 1063. [Google Scholar] [CrossRef]
- Fee, J.A.; McGrady, F.P.; Hart, N.D. Dermoscopy use in primary care: A qualitative study with general practitioners. BMC Primary Care 2022, 23, 47. [Google Scholar] [CrossRef] [PubMed]
- Forsea, A.M.; Tschandl, P.; Zalaudek, I.; Del Marmol, V.; Soyer, H.P.; EurodermoscopyWorking Group; Argenziano, G.; Geller, A.C. The impact of dermoscopy on melanoma detection in the practice of dermatologists in Europe: Results of a pan-European survey. J. Eur. Acad. Dermatol. Venereol. 2017, 31, 1148–1156. [Google Scholar] [CrossRef]
- Skvara, H.; Teban, L.; Fiebiger, M.; Binder, M.; Kittler, H. Limitations of dermoscopy in the recognition of melanoma. Arch. Dermatol. 2005, 141, 155–160. [Google Scholar] [CrossRef] [PubMed]
- Chollet, F. Deep Learning with Python; Manning Publications: Shelter Island, NY, USA, 2018. [Google Scholar]
- Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef] [PubMed]
- Jo, T.; Nho, K.; Saykin, A. Deep Learning in Alzheimer’s Disease: Diagnostic Classification and Prognostic Prediction Using Neuroimaging Data. Front. Aging Neurosci. 2019, 11, 220. [Google Scholar] [CrossRef]
- Pacheco, A.G.C.; Krohling, R.A. The impact of patient clinical information on automated skin cancer detection. Comput. Biol. Med. 2020, 116, 103545. [Google Scholar] [CrossRef]
- Frid-Adar, M.; Diamant, I.; Klang, E.; Amitai, M.; Goldberger, J.; Greenspan, H. GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing 2018, 321, 321–331. [Google Scholar] [CrossRef]
- Thirumalaisamy, S.; Thangavilou, K.; Rajadurai, H.; Saidani, O.; Alturki, N.; Mathivanan, S.K.; Jayagopal, P.; Gochhait, S. Breast Cancer Classification Using Synthesized Deep Learning Model with Metaheuristic Optimization Algorithm. Diagnostics 2023, 13, 2925. [Google Scholar] [CrossRef]
- Houfani, D.; Slatnia, S.; Remadna, I.; Zerhouni, N.; Saouli, H. Breast cancer classification using machine learning techniques: A comparative study. Med. Technol. J. 2020, 4, 535–544. [Google Scholar] [CrossRef]
- Ghosh, P.; Azam, S.; Quadir, R.; Karim, A.; Shamrat, F.M.J.M.; Bhowmik, S.K.; Jonkman, M.; Hasib, K.M.; Ahmed, K. SkinNet-16: A deep learning approach to identify benign and malignant skin lesions. Front. Oncol. 2022, 12, 535–544. [Google Scholar] [CrossRef]
- Kim, J.; Oh, I.; Lee, Y.N.; Kim, J.; Lee, H.J. Predicting the severity of postoperative scars using artificial intelligence based on images and clinical data. Sci. Rep. 2023, 13, 13448. [Google Scholar] [CrossRef]
- Li, S.; Wang, H.; Xiao, Y.; Zhang, M.; Yu, N.; Zeng, A.; Wang, X. A Workflow for Computer-Aided Evaluation of Keloid Based on Laser Speckle Contrast Imaging and Deep Learning. J. Pers. Med. 2022, 12, 981. [Google Scholar] [CrossRef] [PubMed]
- Pacheco, A.G.C.; Lima, G.R.; Salomão, A.S.; Krohling, B.; Biral, I.P.; Angelo, G.G.; Alves, F.C.R.; Esgario, J.G.M.; Simora, A.C.; Castro, P.B.C.; et al. PAD-UFES-20: A skin lesion dataset composed of patient data and clinical images collected from smartphones. Mendeley Data 2020, V1. [Google Scholar] [CrossRef] [PubMed]
- Kaggle. Skin Cancer: Malignant vs. Benign. Available online: https://www.kaggle.com/datasets/fanconic/skin-cancermalignant-vs-benign (accessed on 1 November 2021).
- Skin Cancer MNIST: HAM10000. Available online: https://www.kaggle.com/datasets/kmader/skin-cancer-mnist-ham10000?fbclid=IwAR0ZHT4XR1ZBC8fHdPtVkQRlOF0XQIyJNSiG1rohE4rYb7K8vkKx9eVpSjg (accessed on 5 June 2022).
- Lu, J. Convolutional neural network and vision transformer for image classification. Appl. Comput. Eng. 2023, 5, 104–108. [Google Scholar] [CrossRef]
- Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in Vision: A Survey. Assoc. Comput. Mach. 2022, 54, 10s. [Google Scholar] [CrossRef]
- Steiner, A.; Kolesnikov, A.; Zhai, X.; Wightman, R.; Uszkoreit, J.; Beyer, L. How to train your ViT? Data, Augmentation, and Regularization in Vision Transformers. arXiv 2022, arXiv:2106.10270. [Google Scholar]
- Acar, E.M.; Kilitci, A.; Kaya, Z.; Abbas, O.L.; Kemeriz, F. Basal cell carcinoma presenting as an excoriated cicatricial plaque: A case report. Dermatol. Sin. 2018, 36, 211–213. [Google Scholar] [CrossRef]
- Lewis, J. Keloidal basal cell carcinoma. Am. J. Dermatopathol. 2007, 29, 485. [Google Scholar] [CrossRef]
- Mizuno, H.; Uysal, A.C.; Koike, S.; Hyakusoku, H. Squamous cell carcinoma of the auricle arising from keloid after radium needle therapy. J. Craniofac. Surg. 2006, 17, 360–362. [Google Scholar] [CrossRef]
- Gutman, D.; Codella, N.C.F.; Celebi, E.; Helba, B.; Marchetti, M.; Mishra, N.; Halpern, A. Skin lesion analysis toward melanoma detection: A challenge at the International Symposium on Biomedical Imaging (ISBI) 2016, hosted by the International Skin Imaging Collaboration (ISIC). arXiv 2016, arXiv:1605.01397. [Google Scholar]
- Kaggle. Keloid Scars vs Hypertrophic Scars. Available online: https://www.kaggle.com/datasets/quratulainislam/keloid-vs-scar (accessed on 16 February 2024).
- Cimpean, I.; Theate, I.; Vanhooteghem, O. Seborrheic keratosis evolution into squamous cell carcinoma: A truly modified sun-related tumor? A case report and review of the literature. Dermatol. Rep. 2019, 11, 7999. [Google Scholar] [CrossRef]
- Mentzel, T.; Wiesner, T.; Cerroni, L.; Hantschke, M.; Kutzner, H.; Rütten, A.; Häberle, M.; Bisceglia, M.; Chibon, F.; Coindre, J.M. Malignant dermatofibroma: Clinicopathological, immunohistochemical, and molecular analysis of seven cases. Mod. Pathol. 2013, 26, 256–267. [Google Scholar] [CrossRef] [PubMed]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
- Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2020, arXiv:1905.11946. [Google Scholar]
- Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. arXiv 2017, arXiv:1610.02357. [Google Scholar]
- Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. arXiv 2016, arXiv:1602.07261. [Google Scholar] [CrossRef]
- Tan, M.; Le, Q.V. EfficientNetV2: Smaller Models and Faster Training. arXiv 2021, arXiv:2104.00298. [Google Scholar]
- Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning Transferable Architectures for Scalable Image Recognition. arXiv 2018, arXiv:1707.07012. [Google Scholar]
- Tejada, M.; Espinoza, R.; Hinojosa, L. Use of Xception Architecture for the Classification of Skin Lesions. Int. J. Comput. Vis. 2024, 12, 145–160. [Google Scholar]
- Kurniawan, V.A.T.; Niswary, E.C.; Chandranegara, D.R. Brain Tumor Classification Using InceptionResNetV2 and Transfer Learning. J. Comput. Biol. Med. 2024, 18, 98–110. [Google Scholar]
- Liu, D.; Wang, W.; Wu, X.; Yang, J. EfficientNetV2 Model for Breast Cancer Histopathological Image Classification. IEEE Int. Conf. Electr. Eng. Comput. Sci. 2022, 7, 384–387. [Google Scholar]
- Rahman, M.A.; Bazgir, E.; Hossain, S.M.S.; Maniruzzaman, M. Skin Cancer Classification Using NASNet. Int. J. Sci. Res. Arch. 2024, 11, 23–36. [Google Scholar]
- Ismail, M.; Nisa, F.; Ahmad, R. Utilising VGG16 of Convolutional Neural Network for Brain Tumor and Alzheimer’s Classification. J. Adv. Med. Sci. 2024, 10, 123–135. [Google Scholar]
- Costa, M.; Sousa, J.; Almeida, P. Classification of X-ray Images for Detection of Childhood Pneumonia using MobileNet. IEEE Trans. Med. Imaging 2020, 39, 1123–1135. [Google Scholar]
- Mulya, R.F.; Utami, E.; Ariatmanto, D. Classification of Acute Lymphoblastic Leukemia based on White Blood Cell Images using InceptionV3 Model. J. RESTI (Rekayasa Sist. Dan Teknol. Inf.) 2023, 7, 5182. [Google Scholar] [CrossRef]
- Zerouaoui, H.; Idri, A. Classifying Breast Cytological Images using Deep Learning Architectures. Proc. Int. Conf. Data Sci. Artif. Intell. 2022, 557–564. [Google Scholar] [CrossRef]
- Sait, N.A.; Kathirvelan, J. Detection and classification of progressive supranuclear palsy from MRI images using deep learning. J. Appl. Res. Technol. 2024, 22, 2228. [Google Scholar] [CrossRef]
- Feller, L.; Khammissa, R.A.G.; Kramer, B.; Altini, M.; Lemmer, J. Basal cell carcinoma, squamous cell carcinoma and melanoma of the head and face. Head Face Med. 2016, 12, 11. [Google Scholar] [CrossRef]
- Udriștoiu, A.L.; Stanca, A.E.; Ghenea, A.E.; Vasile, C.M.; Popescu, M.; Udriștoiu, Ș.C.; Iacob, A.V.; Castravete, S.; Gruionu, L.G.; Gruionu, G. Skin diseases classification using deep learning methods. Curr. Health Sci. J. 2020, 46, 136–140. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
- Shin, H.C.; Roth, H.R.; Gao, M.; Lu, L. Deep Convolutional Neural Networks for Computer-Aided Detection: CNN Architectures, Dataset Characteristics and Transfer Learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef]
- Kumar, A.; Kim, J.; Lyndon, D.; Fulham, M.; Feng, D. An Ensemble of Fine-Tuned Convolutional Neural Networks for Medical Image Classification. IEEE J. Biomed. Health Inform. 2017, 21, 31–40. [Google Scholar] [CrossRef] [PubMed]
- Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
- Gupta, S.; Tan, M. EfficientNet-EdgeTPU: Creating Accelerator-Optimized Neural Networks with AutoML. 2019. Available online: https://research.google/blog/efficientnet-edgetpu-creating-accelerator-optimized-neural-networks-with-automl/ (accessed on 7 January 2025).
- Goder, M.; Kornhaber, R.; Bordoni, D.; Winkler, E.; Haik, J.; Tessone, A. Cutaneous basal cell carcinoma arising within a keloid scar: A case report. OncoTargets Ther. 2016, 9, 4793–4796. [Google Scholar]
- Majumder, A.; Srivastava, S.; Ranjan, P. Squamous cell carcinoma arising in a keloid scar. Med. J. Armed Forces India 2019, 75, 222–224. [Google Scholar] [CrossRef]
- Ogawa, R.; Yoshitatsu, S.; Yoshida, K.; Miyashita, T. Is radiation therapy for keloids acceptable? The risk of radiation-induced carcinogenesis. Plast. Reconstr. Surg. 2009, 124, 1196–1201. [Google Scholar] [CrossRef]
- Kaggle. Skin Cancer ISIC. Available online: https://www.kaggle.com/datasets/nodoubttome/skin-cancer9-classesisic/data (accessed on 8 February 2024).
- International Skin Imaging Collaboration (ISIC). ISIC Gallery. Available online: https://gallery.isic-archive.com (accessed on 25 October 2024).
- Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 180161. [Google Scholar] [CrossRef]
- Metzcar, J.; Jutzeler, C.R.; Macklin, P.; Köhn-Luque, A.; Brüningk, S.C. A review of mechanistic learning in mathematical oncology. Front. Immunol. 2024, 15, 1363144. [Google Scholar] [CrossRef]
- Brutti, F.; La Rosa, F.; Lazzeri, L.; Benvenuti, C.; Bagnoni, G.; Massi, D.; Laurino, M. Artificial Intelligence Algorithms for Benign vs. Malignant Dermoscopic Skin Lesion Image Classification. Bioengineering 2023, 10, 11322. [Google Scholar] [CrossRef]
- Bechelli, S.; Delhommelle, J. Machine learning and deep learning algorithms for skin cancer classification from dermoscopic images. Bioengineering 2022, 9, 97. [Google Scholar] [CrossRef]
- Udrea, A.; Mitra, G.; Costea, D.; Noels, E.; Wakkee, M.; Siegel, D.M.; de Carvalho, T.M.; Nijsten, T. Accuracy of a smartphone application for triage of skin lesions based on machine learning algorithms. J. Eur. Acad. Dermatol. Venereol. 2020, 34, 648–655. [Google Scholar] [CrossRef]
- Codella, N.; Rotemberg, V.; Tschandl, P.; Celebi, M.E.; Dusza, S.; Gutman, D.; Helba, B.; Kalloo, A.; Liopyris, K.; Marchetti, M.; et al. Skin lesion analysis toward melanoma detection 2018: A challenge hosted by the international skin imaging collaboration (ISIC). arXiv 2019, arXiv:1902.03368. [Google Scholar]
- Codella, N.C.; Gutman, D.; Celebi, M.E.; Helba, B.; Marchetti, M.A.; Dusza, S.W.; Kalloo, A.; Liopyris, K.; Mishra, N.; Kittler, H.; et al. Skin lesion analysis toward melanoma detection: A challenge at the 2017 international symposium on biomedical imaging (ISBI), hosted by the international skin imaging collaboration (ISIC). In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; pp. 168–172. [Google Scholar]
Study | Dataset | Algorithm | Category | Results | Limitation |
---|---|---|---|---|---|
[28] | Liver lesion images (CT scans) | GAN + CNN | Benign vs. Malignant Liver Lesions | Improved CNN accuracy with GAN augmentation (AUC-ROC = 0.93, recall = 0.857, specificity = 0.924) | GAN-generated images may not fully capture real variations; limited dataset size |
[27] | Non-dermatoscopic skin lesion images + clinical data | Several CNN classifiers | Skin Cancer Detection | Accuracy = 0.788 ± 0.025, precision = 0.8 ± 0.028, recall = 0.788 ± 0.025 | Dependency on patient metadata; limited generalisation ability |
[29] | Mammogram images tested on the MIAS and DDSM | ResNet101 + Metaheuristic optimisation | Breast Cancer Classification | DDSM (Accuracy = 0.986, recall 0.987) MIAS (accuracy = 0.992, recall = 0.979) | High computational cost; possible overfitting |
[31] | HAM10000 + ISIC dataset | proposed CNN-based SkinNet-16 | Benign vs. Malignant Skin Lesions | Accuracy ≈ 0.992, recall ≈ 0.98 | Limited dataset; only binary classification |
[32] | Post-operative scar images + clinical data | Proposed CBAM + ResNet50 | Scar Severity Prediction | Accuracy = 0.733, AUC-ROC = 0.912, recall = 0.733 | Variability in scar assessment; need for real-world validation |
[33] | Keloid images (Laser Speckle Contrast Imaging) | Proposed a cascaded vision transformer architecture | Keloid Severity Evaluation | Average accuracy = | Limited dataset; requires clinical validation |
Diagnostic Class | Skin Disease | Nbr. of Images | Total Nbr. Images |
---|---|---|---|
Melanoma | 52 | ||
Malignant | Basal Cell Carcinoma | 845 | 1089 |
Squamous Cell Carcinoma | 192 | ||
Seborrheic Keratosis | 235 | ||
Nevus | 244 | 1209 | |
Benign | Actinic Keratosis | 730 | |
Total Images | 2298 |
Layers | Output Size | VGG16 |
---|---|---|
Convolution | conv | |
Pooling | max pool | |
Convolution | conv | |
Pooling | max pool | |
Convolution | conv | |
Pooling | max pool | |
Fully Connected | 4096 | Fully connected layer |
Fully Connected | 4096 | Fully connected layer |
Output | 1000 (classes) | Output layer with softmax activation |
Layers | Output Size | DenseNet121 |
---|---|---|
Convolution | conv | |
Pooling | max pool | |
Dense Block | 6 layers | |
Transition Layer | conv, avg pool | |
Dense Block | 12 layers | |
Transition Layer | conv, avg pool | |
Dense Block | 24 layers | |
Transition Layer | conv, avg pool | |
Dense Block | 16 layers | |
Pooling | avg pool | |
Fully Connected | 1024 | Fully connected layer |
Output | 1000 (classes) | Output layer with softmax activation |
Layers | Output Size | InceptionV3 |
---|---|---|
Convolution | conv | |
Convolution | conv | |
Convolution | conv | |
Pooling | max pool | |
Convolution | conv | |
uses the Convolution | conv | |
Inception Module | Inception A | |
Inception Module | Inception B | |
Inception Module | Inception C | |
Pooling | avg pool | |
Fully Connected | 2048 | Fully connected layer |
Output | 1000 (classes) | Output layer with softmax activation |
Layers | Output Size | MobileNet |
---|---|---|
Convolution | conv | |
Depthwise Separable Conv | depthwise, pointwise | |
Pooling | max pool | |
Depthwise Separable Conv | depthwise, pointwise | |
Depthwise Separable Conv | depthwise, pointwise | |
Pooling | max pool | |
Depthwise Separable Conv | depthwise, pointwise | |
Depthwise Separable Conv | depthwise, pointwise | |
Pooling | avg pool | |
Fully Connected | 1024 | Fully connected layer |
Output | 1000 (classes) | Output layer with softmax activation |
Layers | Output Size | EfficientNet-B0 |
---|---|---|
Convolution | conv, 32 filters, stride 2 | |
MBConv1 | depthwise, 16 filters | |
MBConv6 | depthwise, 24 filters, stride 2 | |
MBConv6 | depthwise, 40 filters, stride 2 | |
MBConv6 | depthwise, 80 filters, stride 2 | |
MBConv6 | depthwise, 112 filters | |
MBConv6 | depthwise, 192 filters, stride 2 | |
MBConv6 | depthwise, 320 filters | |
Convolution | conv, 1280 filters | |
Pooling | Global average pooling | |
Fully Connected | 1280 | Fully connected layer |
Output | 1000 (classes) | Output layer with softmax activation |
Layers | Output Size | Xception |
---|---|---|
Convolution | conv, 32 filters, stride 2 | |
Convolution | conv, 64 filters | |
Entry Flow | 3 × SeparableConv layers, 128 filters, stride 2 | |
Entry Flow | 3 × SeparableConv layers, 256 filters, stride 2 | |
Entry Flow | 3 × SeparableConv layers, 728 filters, stride 2 | |
Middle Flow | 8 × (3 × SeparableConv layers, 728 filters) | |
Exit Flow | 3 × SeparableConv layers, 1024 filters, stride 2 | |
Exit Flow | 3 × SeparableConv layers, 2048 filters | |
Pooling | Global average pooling | |
Fully Connected | 2048 | Fully connected layer |
Output | 1000 (classes) | Output layer with softmax activation |
Layers | Output Size | InceptionResNetV2 |
---|---|---|
Convolution | conv, 32 filters, stride 2 | |
Convolution | conv, 32 filters | |
Convolution | conv, 64 filters, stride 2 | |
Inception-ResNet-A | 5 × Inception-ResNet-A modules | |
Reduction-A | Transition layer (pooling, conv) | |
Inception-ResNet-B | 10 × Inception-ResNet-B modules | |
Reduction-B | Transition layer (pooling, conv) | |
Inception-ResNet-C | 5 × Inception-ResNet-C modules | |
Convolution | conv, 1536 filters | |
Pooling | Global average pooling | |
Fully Connected | 1536 | Fully connected layer |
Output | 1000 (classes) | Output layer with softmax activation |
Layers | Output Size | EfficientNetV2-L |
---|---|---|
Convolution | conv, 32 filters, stride 2 | |
MBConv1 | depthwise, 32 filters | |
MBConv4 | depthwise, 64 filters, stride 2 | |
Fused-MBConv4 | fused conv, 128 filters, stride 2 | |
Fused-MBConv6 | fused conv, 256 filters, stride 2 | |
MBConv6 | depthwise, 512 filters, stride 2 | |
MBConv6 | depthwise, 1280 filters | |
Convolution | conv, 1280 filters | |
Pooling | Global average pooling | |
Fully Connected | 1280 | Fully connected layer |
Output | 1000 (classes) | Output layer with softmax activation |
Layers | Output Size | NASNetLarge |
---|---|---|
Convolution | conv, 96 filters, stride 2 | |
Normal Cell | 5 × Normal cells, 168 filters | |
Reduction Cell | Transition layer (pooling, conv) | |
Normal Cell | 5 × Normal cells, 336 filters | |
Reduction Cell | Transition layer (pooling, conv) | |
Normal Cell | 5 × Normal cells, 672 filters | |
Reduction Cell | Transition layer (pooling, conv) | |
Normal Cell | 5 × Normal cells, 1344 filters | |
Reduction Cell | Transition layer (pooling, conv) | |
Normal Cell | 5 × Normal cells, 4032 filters | |
Pooling | Global average pooling | |
Fully Connected | 4032 | Fully connected layer |
Output | 1000 (classes) | Output layer with softmax activation |
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.7893 | 0.7642 | 0.7297 | 0.7465 | 0.8809 |
MobileNet | 0.8008 | 0.7287 | 0.8468 | 0.7833 | 0.8776 |
DenseNet121 | 0.7969 | 0.7589 | 0.7658 | 0.7623 | 0.8841 |
InceptionV3 | 0.7356 | 0.6721 | 0.7387 | 0.7039 | 0.8123 |
EfficientNetB0 | 0.5747 | 0.0 | 0.0 | 0.0 | 0.5608 |
Xception | 0.7635 | 0.7333 | 0.6937 | 0.7130 | 0.8280 |
InceptionRNV2 | 0.7165 | 0.6761 | 0.6396 | 0.6574 | 0.8131 |
EfficientNetV2L | 0.5754 | 0.0 | 0.0 | 0.0 | 0.5943 |
NASNetLarge | 0.6743 | 0.625 | 0.5856 | 0.6047 | 0.7370 |
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.8467 | 0.8142 | 0.8288 | 0.8214 | 0.9274 |
MobileNet | 0.7816 | 0.7455 | 0.7387 | 0.7421 | 0.8633 |
DenseNet121 | 0.8238 | 0.8495 | 0.7117 | 0.7745 | 0.9039 |
InceptionV3 | 0.8122 | 0.7719 | 0.7930 | 0.7822 | 0.8880 |
EfficientNetB0 | 0.8008 | 0.7706 | 0.7568 | 0.7636 | 0.8793 |
Xception | 0.8161 | 0.8316 | 0.7117 | 0.7670 | 0.8958 |
InceptionRNV2 | 0.8161 | 0.7838 | 0.7838 | 0.7838 | 0.8793 |
EfficientNetV2L | 0.8314 | 0.7815 | 0.8378 | 0.8097 | 0.8869 |
NASNetLarge | 0.7203 | 0.6173 | 0.9009 | 0.7326 | 0.8501 |
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.7969 | 0.7900 | 0.7117 | 0.7488 | 0.8787 |
MobileNet | 0.7893 | 0.7154 | 0.8378 | 0.7718 | 0.8674 |
DenseNet121 | 0.7893 | 0.7188 | 0.8288 | 0.7699 | 0.8742 |
InceptionV3 | 0.7203 | 0.6301 | 0.8288 | 0.7160 | 0.8001 |
EfficientNetB0 | 0.5785 | 0.5785 | 0.2162 | 0.3038 | 0.5615 |
Xception | 0.7548 | 0.7156 | 0.7027 | 0.7091 | 0.8031 |
InceptionRNV2 | 0.7471 | 0.7064 | 0.6937 | 0.7000 | 0.8250 |
EfficientNetV2L | 0.5402 | 0.4764 | 0.8198 | 0.6026 | 0.6011 |
NASNetLarge | 0.6935 | 0.6449 | 0.6216 | 0.6330 | 0.7815 |
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.8237 | 0.7982 | 0.7839 | 0.7909 | 0.9255 |
MobileNet | 0.7931 | 0.7822 | 0.7117 | 0.7453 | 0.8668 |
DenseNet121 | 0.8084 | 0.8020 | 0.7297 | 0.7642 | 0.8917 |
InceptionV3 | 0.7854 | 0.7391 | 0.7658 | 0.7522 | 0.8641 |
EfficientNetB0 | 0.8391 | 0.8224 | 0.7928 | 0.8073 | 0.8978 |
Xception | 0.8352 | 0.8269 | 0.7748 | 0.8000 | 0.9031 |
InceptionRNV2 | 0.8276 | 0.8000 | 0.7928 | 0.7964 | 0.8889 |
EfficientNetV2L | 0.8199 | 0.7963 | 0.7748 | 0.7854 | 0.8591 |
NASNetLarge | 0.7050 | 0.5966 | 0.9459 | 0.7317 | 0.8762 |
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.8084 | 0.8081 | 0.7207 | 0.7619 | 0.8808 |
MobileNet | 0.8276 | 0.8367 | 0.7387 | 0.7847 | 0.8965 |
DenseNet121 | 0.8084 | 0.7798 | 0.7658 | 0.7727 | 0.8962 |
InceptionV3 | 0.7165 | 0.6529 | 0.7117 | 0.6810 | 0.7974 |
EfficientNetB0 | 0.5057 | 0.4505 | 0.7387 | 0.5597 | 0.5575 |
Xception | 0.7356 | 0.6875 | 0.6937 | 0.6906 | 0.8186 |
InceptionRNV2 | 0.7356 | 0.7100 | 0.6396 | 0.6730 | 0.8191 |
EfficientNetV2L | 0.5556 | 0.4868 | 0.8288 | 0.6133 | 0.6845 |
NASNetLarge | 0.7011 | 0.6774 | 0.5676 | 0.6176 | 0.7694 |
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.8506 | 0.8600 | 0.7748 | 0.8152 | 0.9205 |
MobileNet | 0.7854 | 0.7570 | 0.7297 | 0.7431 | 0.8728 |
DenseNet121 | 0.8467 | 0.8318 | 0.8018 | 0.8165 | 0.9101 |
InceptionV3 | 0.8352 | 0.8469 | 0.7477 | 0.7943 | 0.9168 |
EfficientNetB0 | 0.8199 | 0.7909 | 0.7838 | 0.7873 | 0.9020 |
Xception | 0.8276 | 0.8113 | 0.7748 | 0.7926 | 0.9052 |
InceptionRNV2 | 0.8200 | 0.8137 | 0.7477 | 0.7793 | 0.8976 |
EfficientNetV2L | 0.8506 | 0.8461 | 0.7928 | 0.8186 | 0.9197 |
NASNetLarge | 0.8391 | 0.8556 | 0.7477 | 0.7981 | 0.9026 |
Model | Accuracy | Precision | Recall | AUC | |
---|---|---|---|---|---|
VGG16 | 0.8647 | 0.8674 | 0.8647 | 0.8378 | 0.8982 |
MobileNet | 0.8346 | 0.8045 | 0.8346 | 0.8100 | 0.8727 |
DenseNet121 | 0.9098 | 0.9110 | 0.9098 | 0.8972 | 0.8872 |
InceptionV3 | 0.8797 | 0.8956 | 0.8797 | 0.8559 | 0.8848 |
EfficientNetB0 | 0.8421 | 0.8172 | 0.8421 | 0.8133 | 0.8867 |
Xception | 0.8346 | 0.8045 | 0.8346 | 0.8100 | 0.8836 |
InceptionRNV2 | 0.8647 | 0.8521 | 0.8647 | 0.8447 | 0.8976 |
EfficientNetV2L | 0.8571 | 0.8343 | 0.8571 | 0.8344 | 0.8942 |
NASNetLarge | 0.8571 | 0.8551 | 0.8571 | 0.8297 | 0.8819 |
Authors | Dataset | Category | Accuracy | Precision | Recall |
---|---|---|---|---|---|
Brutti et al. [79] | Dermatoscopic image dataset from [77,82,83] | Benign vs. Malignant | |||
Bechelli and Delhommelle [80] | Dermatoscopic image dataset from [35,77] | Benign vs. Malignant | 0.88 | 0.93 | 0.83 |
Pacheco and Krohling [27] | Non-dermatoscopic image dataset from [34] | Multiple (six) skin lesion class | 0.788 ± 0.025 | 0.8 ± 0.028 | 0.79 ± 0.027 |
Udrea et al. [81] | Non-dermatoscopic image dataset (private) | Unspecified | unavailable | unavailable | 0.951 |
Benign vs. Malignant; | 0.851 | 0.860 | 0.775 | ||
Ours | Non-dermatoscopic image dataset from [34] and [44] | Benign vs. Malignant vs. Keloid; | 0.877 | 0.881 | 0.877 |
Keloid vs. BCC vs. SCC | 0.91 | 0.911 | 0.91 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Adebayo, O.E.; Chatelain, B.; Trucu, D.; Eftimie, R. Deep Learning Approaches for the Classification of Keloid Images in the Context of Malignant and Benign Skin Disorders. Diagnostics 2025, 15, 710. https://doi.org/10.3390/diagnostics15060710
Adebayo OE, Chatelain B, Trucu D, Eftimie R. Deep Learning Approaches for the Classification of Keloid Images in the Context of Malignant and Benign Skin Disorders. Diagnostics. 2025; 15(6):710. https://doi.org/10.3390/diagnostics15060710
Chicago/Turabian StyleAdebayo, Olusegun Ekundayo, Brice Chatelain, Dumitru Trucu, and Raluca Eftimie. 2025. "Deep Learning Approaches for the Classification of Keloid Images in the Context of Malignant and Benign Skin Disorders" Diagnostics 15, no. 6: 710. https://doi.org/10.3390/diagnostics15060710
APA StyleAdebayo, O. E., Chatelain, B., Trucu, D., & Eftimie, R. (2025). Deep Learning Approaches for the Classification of Keloid Images in the Context of Malignant and Benign Skin Disorders. Diagnostics, 15(6), 710. https://doi.org/10.3390/diagnostics15060710