Next Article in Journal
Acoustic Emission in Bone Biomechanics: A Comprehensive Review of Mechanical Properties and Predictive Damage Modeling
Next Article in Special Issue
The Influence of Ambient Factors on Emotional Wellbeing of Older Adults: A Review
Previous Article in Journal
Flake-Shaped Nickel Hydroxide Supported on Carbon Cloth as an Electrochemical Sensor for Efficient Detection of Phosphate Under Neutral Conditions
Previous Article in Special Issue
Bird Species Detection Net: Bird Species Detection Based on the Extraction of Local Details and Global Information Using a Dual-Feature Mixer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Advanced Deep Learning Models for Melanoma Diagnosis in Computer-Aided Skin Cancer Detection

1
Department of Software Engineering & AI, Media Design School, Auckland 1010, New Zealand
2
School of Engineering, Computer, and Mathematical Sciences, Auckland University of Technology, Auckland 1010, New Zealand
3
Division of Intelligent Future Technologies, Mälardalen University, 721 23 Västerås, Sweden
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(3), 594; https://doi.org/10.3390/s25030594
Submission received: 25 November 2024 / Revised: 5 January 2025 / Accepted: 8 January 2025 / Published: 21 January 2025

Abstract

:
The most deadly type of skin cancer is melanoma. A visual examination does not provide an accurate diagnosis of melanoma during its early to middle stages. Therefore, an automated model could be developed that assists with early skin cancer detection. It is possible to limit the severity of melanoma by detecting it early and treating it promptly. This study aims to develop efficient approaches for various phases of melanoma computer-aided diagnosis (CAD), such as preprocessing, segmentation, and classification. The first step of the CAD pipeline includes the proposed hybrid method, which uses morphological operations and context aggregation-based deep neural networks to remove hairlines and improve poor contrast in dermoscopic skin cancer images. An image segmentation network based on deep learning is then used to extract lesion regions for detailed analysis and calculate the optimized classification features. Lastly, a deep neural network is used to distinguish melanoma from benign lesions. The proposed approaches use a benchmark dataset named International Skin Imaging Collaboration (ISIC) 2020. In this work, two forms of evaluations are performed with the classification model. The first experiment involves the incorporation of the results from the preprocessing and segmentation stages into the classification model. The second experiment involves the evaluation of the classifier without employing these stages i.e., using raw images. From the study results, it can be concluded that a classification model using segmented and cleaned images contributes more to achieving an accurate classification rate of 93.40% with a 1.3 s test time on a single image.

1. Introduction

Melanocytes proliferate, and fibrous tissues overgrow in the deeper layers of the skin, thereby causing skin cancer. In the outermost layer of the skin, cancerous cells are visible. Several factors contribute to the development of skin cancer, including fair complexion, overexposure to sun rays, sunburn, genetic history, and weak immunity [1,2]. In many cases, skin cancer is indistinguishable from a normal mole due to its shape, size, and color. As a result, it is challenging to spot with the naked eye. There are main five types of skin cancer [3], as shown in Figure 1. It is crucial to be able to distinguish melanoma (MEL) from other types of skin cancer such as basal cell carcinoma (BCC), benign keratosis-like (BKL) lesions, congenital melanocytic nevus (CMN), and benign (BEN). Since MEL is an aggressive form of skin cancer, it is less common than the other types. The other four types of cancer are considered benign because they do not invade other parts of the body. These cancers do not spread to local structures or distant parts of the body. The growth of benign tumors is usually slow, and their borders are clearly defined. The majority of benign tumors are not dangerous. Hence, the primary objective of this study was to detect melanoma and distinguish it from benign cancer.
Image-assessment-based diagnosis is crucial in diagnosing abnormalities, including skin cancer, all over the body. The American Academy of Dermatologists Association (AADA) [4] estimates that one out of every five Americans develops skin cancer at some point in their lives. A total of 197,700 new cases of melanoma occurred in the U.S. in 2022, including 97,920 noninvasive (in situ) cases and 99,780 invasive cases [5,6]. The International Agency for Research on Cancer (IARC) also pointed out in its latest study [7] that there will be an increase in cutaneous melanoma cases by more than 50% from 2020 to 2040. In Australia and New Zealand, the incidences observed for melanoma are 42 cases per 100,000 men and 31 cases per 100,000 women, followed by Western Europe (19 cases in both sexes), Northern America (18 cases in men and 14 in women), and Northern Europe (17 cases in men and 18 cases in women) [7]. The incidence of melanoma is low in most countries in Africa and Asia, where it is commonly less than 1 per 100,000. By 2040, scientists estimated that nearly 100,000 melanoma deaths worldwide will have occurred due to global population changes, and more than 500,000 new cases of melanoma are expected every year.
The early detection of melanoma is imperative for ensuring high survival rates. If diagnosed early, the disease can be successfully treated. A CAD system can diagnose a wide range of disorders quickly, efficiently, and consistently. In addition to offering precise and cost-effective disease detection and protection, CAD also offers advanced tumor disease detection and protection. A variety of imaging technologies are used to assess disorders of the human organ system, including magnetic resonance imaging (MRI), positron emission tomography (PET), and X-rays. For skin lesion analysis and prognosis, dermatoscopes are used to visually examine the infected skin area with a magnified view.
In the current clinical setting, however, there are a few limitations that complicate diagnosis. A dermatologist, for instance, examines a skin lesion visually based on various characteristics such as its size, border, shape, and color, called the ABCDE criteria or seven-point checklist [8]. There are, however, inaccuracies due to complex lesion patterns and variations in interpretation. A second issue is that skin cancer diagnosis is an expensive procedure, as the average treatment costs in the USA for existing patients are approximately USD 44.9 million, and for newly diagnosed patients, it is approximately USD 932.5 million [9]. In addition, the accuracy of skin lesion diagnostics is lower among dermatologists with little expertise. Physicians use a variety of subjective, time-consuming, and error-prone methods to evaluate and analyze lesion images. Because skin lesions are so complex, the images are difficult to understand. Analyzing images for skin lesions, evaluating them, and being aware of them requires unambiguous identification of lesion pixels. Thus, computer-aided diagnostic and prediction systems for skin cancer detection have made significant advances by utilizing deep learning approaches in computer vision. These computer-aided models are affordable and fast to analyze initial skin lesions.
The purpose of this study was to develop a variety of deep learning models for the different phases of a computer-aided melanoma diagnosis system, which utilizes lesion image sites to correctly diagnose the early stages of cancer. The system under consideration relies on preprocessing, segmentation, and classification operations. The major novel contributions made to each phase such as noise removal, image contrast enhancement, segmentation, and classification of the melanoma CAD system in this research are listed as follows:
  • Reconstruction of the entire input image using a custom-designed intensity-adjustment-based interpolation technique after eliminating hairlines (Section 3.2.1).
  • Application of a context aggregation deep learning network to improve image contrast, as well as rescaling and standardization of images to fit the training models during the learning procedure (Section 3.2.2).
  • Design of a novel deep neural network incorporating atrous dilated convolutions to extract lesion segmentation maps with high spatial resolution (Section 3.3).
  • Design of a novel deep convolutional neural network (N-DCNN) to calculate the high- to low-level features of a lesion to distinguish benign and malignant lesions (Section 3.4).
  • Performance analysis of the classification model utilizing unprocessed (raw images) and preprocessed (enhanced images) datasets.

2. Related Studies

Several categorization algorithms and strategies have been devised to enhance the accuracy of the melanoma diagnosis process. This study focused on deep learning models for different stages of skin cancer in CAD systems to develop an optimized approach in contrast to a variety of other existing image classification models. In the case of skin cancer problems, the presence of hairlines is a major challenge because such kind of noisy data produce poor feature sets, which can further lead to poor image segmentation and classification results. The hairline removal method is a two-phase process: First, the localization and occlusion of hair pixels are performed using morphological operations [10,11], edge operators [12], and Gaussian derivatives [13]. Second, to repair the color and texture of occluded areas, bi-linear interpolation and histogram equalization methods [14] are used. There are old and limited studies [15,16] that contributed to the employment of hairline removal preprocessing methods to clean dermoscopic skin cancer images. The major problem not addressed in the literature is the detection and painting of very thin and complex hairlines in the skin samples used in the preprocessing phase of melanoma CAD models. Furthermore, the original color contrast and texture information of the image that has been lost after the hair pixel’s occlusion are not restored by the existing methods.
For the segmentation stage of a melanoma CAD model, several methods have been presented, such as machine-learning-based unsupervised [17], and optimization [18] methods. Due to the ineffectiveness of such approaches in producing accurate lesion segmentation masks, recent studies have utilized a variety of deep learning frameworks to extract information about lesions. CNNs, such as fully convolutional neural networks (FCNs) [19] and deconvolution networks [20], as extensions of FCNs, SegNet [21], and UNet [22], have been proposed by researchers as potential segmentation methods. These networks have some limitations, such as pooling layers that reduce resolution and discard essential image information. To perform semantic segmentation, class maps must be precisely known, and ‘where’ information must be preserved. For some problem domains, encoder–decoder networks (SegNet and UNet) maintain output image quality efficiently; however, their complexity and long execution times can cause them to be heavy. The concept of atrous convolutions, introduced by Chen [23], enables the direct control of the resolution to preserve the feature map information computed in the deep convolutional layers at the expense of long computation time due to their large networks. Therefore, there is a high demand for an optimized and less complex architecture that can accurately extract lesion information.
Further, for the last stage of the melanoma CAD model, several image-based skin classification systems have been developed using traditional machine learning techniques, such as K-nearest neighbors (KNN) [24,25,26], support vector machine (SVM) [26,27,28], artificial neural network (ANN) [29,30,31], and deep learning approaches [32]. Review articles on these traditional classifiers can be found [33,34]. The reviews point out that traditional machine-learning techniques are feature-dependent. The wrong selection of features can lead to poor classification results. Several deep learning techniques have been proposed for automated classification tasks [35,36,37]. For example, in 2012, Krizhevsky et al. [38] proposed AlexNet to classify 1.2 million images in the ImageNet containing 1000 different classes. A few attempts were also made by integrating traditional machine learning algorithms and deep convolutional neural networks (CNN) to improve the performance on classification problems. Ulzii et al. [39] proposed a technique that employs a pretrained AlexNet and the SVM classifier to classify 3753 skin lesion images. They explained the advantages of AlexNet in terms of its few training parameters and validity. Mahbod et al. [40] presented three deep learning networks, AlexNet, VGG16, and ResNet18, as feature generators. They trained the SVM classifier with feature sets from three neural networks and fused their results to make final predictions to detect three categories of skin cancer. They concluded in their study that fusing multiple features from various networks leads to a better classification rate. However, these approaches still require the additional step of crafting the feature sets employed in machine learning classifiers.
Some recent works have used optimized deep learning frameworks to enhance the melanoma classification performance. Guoda et al. [41] considered preprocessing and classification as key steps in achieving high performance. To reduce the inconsistencies among images, the entire intensity of the image was improved during the preprocessing stage. Then, a crossbred deep learning model was developed to predict cancer, and a comparison was made with other pretrained networks such as ResNet50, InceptionV3, and ResNet-Inception. Another study presented in [42] proposed a three-step superpixel method, which was further enhanced by mapping it through a new activation function. The deep feature extraction was performed via a pretrained ResNet50 deep learning model, and then features were optimized using a modified grasshopper method. However, that study did not consider the removal of hairlines and the impact of using segmented images instead of original images.
The literature suggests there is an increasing need for a high-performing skin cancer diagnosis system. To achieve high diagnosis accuracy, this research aimed to develop an automated classification system to determine malignant skin cancer. The above-discussed studies on skin cancer classification have utilized well-known deep learning architectures for classification purposes. However, deep learning frameworks for classification and segmentation still need to be enhanced and optimized as well as made more efficient. Moreover, to the best of our knowledge, the impact of employing results from the preprocessing stages on the final classification outcome was not examined in existing studies. This study also considered a complex and challenging skin cancer dataset from the ISIC 2020 challenge, which consists of several samples from dark-skinned people, the older population, and highly noisy conditions.

3. Materials and Methods

This research classified skin cancer into two types, benign and malignant, where benign further includes keratosis, basal cell carcinoma, and nevus lesions. Deep learning approaches were used for different phases of the melanoma CAD system for the classification of skin cancer. In Figure 2, a systematic methodology is presented that accepts the skin cancer image dataset, and its learning is performed in two ways. Step 1 performs classification of the raw images with a few basic preprocessing operations, and step 2 involves a series of phases such as preprocessing, segmentation, and classification. During the preprocessing phase, hairlines are obliterated using a hairline removal method, followed by a context aggregation deep learning model to regain the contrast and smoothness lost during hair detection and repatriation. In the segmentation operation, lesions are separated from the background using a neural network based on dilated atrous convolutions. In the last stage, melanoma and benign tumors are classified using an N-DCNN model. As a final step, performance metrics were used to compare and analyze the classification results of steps 1 and 2.

3.1. Dataset Preparation

In this study, a dermoscopic image dataset from ISIC 2020 [43,44] was used for training and testing the network. The ISIC is a well-known, open-access repository of dermatology images that presents a variety of challenging and complex skin cancer samples. There are 33,126 lesion samples in the original ISIC 2020 dataset collected from 2056 patients. The lesions were collected from men and women. Skin cancer is more common in men than women. In the given dataset, 15,981 lesions are from women, and 17,145 belong to men. In the dataset, 16 samples were collected for each patient to observe changes over time. Because skin cancer develops over time, its ABCDE (asymmetry, border, color, diameter, and evolution) characteristics change over time. Several of the lesions likely came from the same patient who developed symptoms of cancer, which is why their mole changed in shape, color, and size. For some patients, no changes were observed at the time of the next scan; thus, such images formed duplicates [45]. This may result in biases and overfitting in the network. To prevent these issues, data normalization was performed to eliminate duplicates. After removing duplicates, the dataset consisted of 12,882 skin samples, 7093 of which were BEN samples, and 5789 of which were MEL samples. There are also some challenging samples in the present dataset, including samples from people with dark skin and samples from older individuals. The dataset contains highly unbalanced classes, which may result in low accuracy of the model, for example, with significantly fewer MEL samples than BEN samples. Hence, synthetic samples were generated for classes with fewer images using random oversampling. Of the samples, 70% were used for training, 10% for validation, and 20% for testing.

3.2. Preprocessing

3.2.1. Hairlines Removal Method

Various external and environmental factors can cause noise artifacts in skin cancer images. The main concern is the presence of hairlines and poor contrast, which compromise the model’s ability to diagnose. The proposed hairlines removal approach first transforms the color space from RGB to grayscale, and then hairlines are detected through morphological closing and subtraction operations.
Close img = ( I ( x , y ) SE ) SE
Diff img = Close img I
The closing operation is performed between the original image I and the structuring element S E by applying dilation (⊕), followed by an erosion operation (⊖), as given in Equation (1). An image-closing operation removes the irrelevant parts of an image. In addition, it eliminates small holes, fills gaps between boundary lines, and smooths contours from the outside. To detect differences between two images and level uneven parts, a subtraction is performed between an image obtained after closing C l o s e i m g and an original image I, as presented in Equation (2). After subtraction, the resultant image contains the pixels that differ between the two. A binarization process using T = 25 as a threshold is used to produce hair masks and refine hair pixels. During this comparison, background pixels are set to ‘0’, and foreground pixels, including those associated with hair, are set to ‘1’. To repair the boundaries and bridge the gaps in the resultant image, a dilation operation is applied.
Dilation img = Diff img > T
Dilation img = Dilation img SE
New img = I , if Dilation img = = 0 o t h e r w i s e , I new , with new intensity remapping
Furthermore, the pixels in a dilation image are copied into a new image based on a comparison: if a pixel value is ‘0’, then it is considered the background of the new image. The pixels having non-zero values are compared against the threshold to check if it is a hairy pixel, and, if detected, it is replaced with the intensity of the neighboring pixels. The above-given operations are explained in Equations (3)–(5). The image output produced after each operation of the hairline removal method is displayed in Figure 3.

3.2.2. Contrast Enhancement Model

During the previous step of hair removal, the contrast and resolution of the skin images are significantly affected because the output image’s pixel was reconstructed, so it is necessary to apply a postprocessing method to enhance image quality. A deep-learning-based multiscale context aggregation network (MCAN) is used to approximate the contextual characteristics of skin cancer images. Detailed information about how hairlines are occluded and how MCAN improves medical image contrast can be found in our other research paper [46]. To achieve the highest accuracy, MCAN must first be trained on a set of images with high-frequency details. As a result of training, the network can process degraded images directly without the need for conventional processing. The network aggregates multiscale information of images and preserves local and global information at the deepest level of the network.
An illustration of the MCAN model’s architecture is shown in Figure 4. In this network, original medical images are extracted and divided into patches of 256 × 256 each. The MRABs (multiscale residual aggregation blocks) are the building blocks of the network. There are eight context modules in the MCAN, called MRAB, which consist of layers such as the input, 2D convolution, batch normalization, leaky ReLU activation, and regression layers. Multiscale contextual information is gathered by these context modules and passed on to the next MRAB module.
An image patch with a different dilation rate is applied with 32 convolutions in the MRAB. Each convolution layer C o n v i …, C o n v N has its own dilation rate set at 2 , 4 , 6 , 8 , 16 , 32 , 64 , and 128, where N = 8 . To aggregate contextual information, the network uses dilation convolutions systematically without deteriorating the image’s quality. To ensure the normalization and identity branches are adjusted and strengthened, two more custom layers are created, adaptive normalization μ and λ . An initial training phase was conducted on original skin cancer images to estimate contextual information about the cancer images. Then, the model was applied to occluded hair images to restore their contrast.

3.3. Lesion Segmentation Model

Segmentation of lesions is a challenging task in CAD since it involves identifying the precise boundary of the lesions. For clinicians to better understand lesion patterns, an atrous convolutional neural network (ACNN) is used to segment infected areas from skin photographs automatically. The proposed segmentation deep learning model utilizes Chen et al.’s [23,47] original concept of atrous rates in the convolution operation. By incorporating atrous rates into the ACNN architecture as shown in Figure 5, a wide view of fields is achieved. The structure of ACNN is designed into 5 blocks having 16 feature extraction layers with different dilation factors. There are two primary steps in ACNN: pixel classification, or segmentation, and feature extraction. To enhance the filter’s view, the network uses multiple scaling rates combined with atrous convolutions. Segmentation models use enhanced input images through preprocessing methods such as hair removal and context aggregation, which are converted to a specified size and then processed by various sub-blocks to compute features.
In the network, the convolutional layer performs the feature calculations. An image I x , y , k is converted into a feature map F M ( x , y , k ) by sliding a small matrix called a kernel or filter over it and transforming the pixels’ values as follows:
FM ( x , y , k ) = i = 1 m h j = 1 m w k = 1 m c I x + i , y + j , k K i , j , k r = 1
The height, width, and number of channels of an input image I [ x , y , k ] can be defined as m [ h ] , m [ w ] , and m [ c ] . The channels in an input image I [ x , y , k ] and kernel K [ i , j , k ] need to match. Each convolution layer was constructed using atrous convolutions with varying dilation rates. The filter can assimilate a wider context by utilizing a larger field of view (the space of the input vector). The methodology provides an efficient way of balancing correct localization with context absorption without requiring extra computations. Rather than using general convolutions, atrous convolutions are used, which helps to extract more contextual information. The convolutional operation, as described in (6) is a standard operation that uses a dilation rate of 1.
In contrast, the following equation describes the atrous convolution operation, if the dilation rate is greater than 1.
FM ( x , y , k ) i = 1 m h j = 1 m w k = 1 m c I x + i , y + j , k . r K i , j , k r 2
The dilation factor increasedsby ‘2’ with each successive two-dimensional convolution layer. Figure 6 shows how translating a filter’s input image by dilation rates 1, 2, and 4 affects its view on the input image.
This network uses leakyReLU to multiply all negative values by a fixed scalar and transform them into positive ones. To normalize the data, a batch normalization layer is used in the network. The hyperparameters used for the training of the network are shown in Table 1.

3.4. Classification Model

Melanoma and benign skin cancers were classified using a novel deep convolutional neural network (N-DCNN) architecture. The N-DCNN for melanoma detection is a carefully crafted network that extracts low-level and high-level skin details by arranging many layers in an orderly fashion. The N-DCNN architecture consists of 11 blocks, as shown in Figure 7.
A deep network is constructed by repeating a few blocks such as 2, 4, and 2 times. A three-channel RGB image is accepted by the network’s first layer, which has dimensions of 128 × 128 . Then, it applies a convolutional operation that slides 8 kernels of size 3 × 3 over the image. The multiple convolutional layers in several blocks calculate a rich feature set from skin cancer images, which is passed to the next layer, max-pooling. The pooling operation reduces the size of the feature maps by taking the maximum values. In each block, the convolution layer is followed by batch normalization and the leakyReLU activation layer. Normalizing the feature maps from the previous layers is accomplished through batch normalization. To prevent overfitting, it regulates the network’s learning process. In this case, leakyReLU is used since it has a slight slope for negative values instead of a zero slope for standard ReLU. A negative value is converted into a positive value using the leakyReLU function by multiplying it by a scalar value s = 0.3 :
l e a k y R e L U = x × s , x < 0 x , x 0
By constructing a deep network of repeating blocks, edge, color, and pattern information about the lesions can be obtained. To interpret the probability of falling into a certain category, the softmax function calculates a confidence score. It is less complex and lightweight than other state-of-the-art networks because it generates fewer learnable parameters (3.3 M) and kernels (3.1 K). A backpropagation algorithm is used to optimize the weights by reducing the loss-concerning gradients. To make small changes in the direction of optimization, a stochastic gradient optimizer updates the network’s weights and biases.
θ i + 1 = θ i α L ( θ i )
Iterations are represented by i, learning parameters (0.001) are represented by α > 0 , parameter vectors are described by θ , and the loss function’s gradient is defined by L ( θ [ i ] ) . At each iteration of the algorithm, the gradient is evaluated, and the parameters are updated over a small set of batches. Often, networks become stuck in local minima due to large weight values. The gradient descent algorithm was modified to reduce oscillations using a momentum term γ :
θ i + 1 = θ i α L ( θ i ) + γ ( θ i + θ i + 1 )
The error is calculated using the cross-entropy loss function as follows:
L o s s = 1 N i = 1 K j = 1 N w i T i j l o g ( P i j )
A network determines the weights w by taking into account the number of observations N and the number of classes K. The hyperparameters utilized by the N-DCNN for training are given in Table 2.

4. Results

The N-DCNN was evaluated on a large dataset, ISIC 2020, that is composed of dermoscopic lesion images for the classification of skin lesions. A N-DCNN classification network was trained and evaluated in two ways, as explained in Figure 2: the first using raw image datasets after applying general preprocessing operations such as data normalization and augmentation, and the second using enhanced and segmented images. Network training was conducted on GeForce GTX 1080 Ti hardware with a computation capacity of ‘7.5’. The original dataset contained many duplicates that were not suitable for network training. Thus, to eliminate duplicate samples in a dataset, the algorithm given in Figure 8 was designed. In the case of identical images, when the correlation coefficient was greater than 0.99, one copy of each image was discarded.
The distribution of data after removing duplicates is described in Table 3. It can be seen that the adopted datasets were highly skewed among the two skin cancer types i.e., MEL and BEN. In the network training, data augmentation was applied using three common operations: rotation from 300 to + 300 , scaling by 0.8 in the X direction, and translating by 5 and + 5 in the Y direction. The training set was only subjected to these operations, whereas the validation and test sets were not altered, and their original data distribution was used.
To compare the model’s performance with other state-of-art-methods, several metrics [48,49] were considered, such as precision (PRE), recall (REC), accuracy (ACC), specificity (SPE), F1 score, and learnable parameters.
A C C = T P + T N T P + F P + T N + F N
P R E = T P T P + F P
R E C = T P T P + F N
S P E = T N T N + F P
F 1 - Score = 2 T P 2 T P + F P + F N
The terms TP, FP, TN, and FN represent true positive, false positive, true negative, and false negative, respectively. The ACC measures the percentage of correctly identified samples to the total number of predictions. PRE and REC are significant metrics when assessing the performance of a model, since PRE measures all positive predicted rates, and REC determines the actual positive ratio from all positively identified samples. A metric called SPE measures the model’s ability to identify the TN of each class. A harmonic mean can be calculated by considering FP and FN when computing the F1 score. A value near one indicates the perfect PRE and REC.
The total number of parameters of the N-DCNN was calculated as the sum of all layer’s parameters including the convolutional layer, fully connected layer, and batch normalization layer. The pooling layer parameters were not added as the pooling layer does not contribute parameters. The parameters for each layer were calculated using different formulas.
  • For the convolution layer, the parameters were calculated as
    P a r a m e t e r s = ( K e r n e l _ w i d t h × K e r n e l _ h e i g h t × n o . _ o f _ k e r n e l s _ i n _ t h e _ p r e v i o u s _ l a y e r + 1 ) × n o . _ o f _ k e r n e l s
  • At the fully connected layer, the parameters are calculated as
    P a r a m e t e r s = ( c u r r e n t _ l a y e r _ n e u r o n s , c × p r e v i o u s _ l a y e r _ n e u r o n s , p ) + 1 × c
  • For the batch normalization layer,
    P a r a m e t e r s = 2 × n u m b e r _ o f _ f e a t u r e s
The ISIC validation dataset contained a different proportion of samples from both classes, and hyperparameters were optimized to improve performance on the validation dataset. The trained model with fine-tuned parameters was used to evaluate the test set, which was unseen by the network. The performance of the N-DCNN on the ISIC validation set is shown in Figure 9a by plotting its performance between loss and accuracy over the number of epochs. With an increasing number of iterations per epoch, the accuracy of the network gradually increased. The confusion matrix given in Figure 9b demonstrates that the model achieved TPs of 1293, TNs of 1050, FNs of 108, and FPs of 126 on the ISIC test dataset. On the ISIC datasets, the area under the curve (AUC) for the MEL class was 0.9642, which illustrates the trade-off between sensitivity and specificity achieved by the model. Figure 9c shows the true positive vs. false positive curves.
As shown in Table 4, the N-DCNN performed better than the other state-of-the-art approaches in skin lesion classification. Based on metrics such as ACC, PRE, REC, SPE, and F1 score, the most impressive results are highlighted in bold. For the ISIC 2020 dataset, N-DCNN achieved higher values for all metrics than the other methods. The winners of the ISIC 2020 challenge were determined primarily by their AUC. There were quite a few duplicates in the dataset, and, to the best of our knowledge, no study has tried to eliminate duplicates. In this analysis, the N-DCNN model achieved the highest AUC score among the considered methods, as shown in Table 5.
Furthermore, in Table 6, a comparison between the proposed N-DCNN and some efficient baseline CNN models such as ResNet18, Inceptionv3, and Xception is presented. The comparison was performed between these popular networks on the same dataset. In Table 6, it is evident that the proposed model performed better on the ISIC test set than the other networks in terms of ACC, PRE, and RECm with the fewest learnable parameters, making it lightweight and simple.
The preprocessed and segmented images were used to further evaluate the performance of the classification network. The IA-HR algorithm was applied first to remove hairlines, and contrast was increased using MCACNN before DilatedSkinNet segmentation was used to extract the accurate ROIs from the lesion. With the preprocessed training, validation, and test samples of the ISIC dataset, the N-DCNN model was again trained, validated, and evaluated. On the preprocessed data, the N-DCNN model was trained with the same tuned hyperparameters. Table 7 and Figure 10 show the performance of the N-DCNN on the raw and preprocessed data.

5. Discussion

Skin cancer prediction as malignant or benign is accomplished using dermoscopic images of the skin. Dermoscopy is the safest and most time-effective [57] mode of capturing these lesions. There is a great need for computerized diagnostic systems to determine types of skin cancer as the number of cases is continuously growing all over the world [58]. Clinical analysis traditionally analyzes the lesions according to the ABCDE [59] criteria, which is time-consuming, and diagnosis is difficult. Moreover, the lesions’ high inter- and intra-class similarities, morphological complexities, and variations have further motivated the development of an automated classification framework. Various lesion image features such as textures, geometrical aspects, color, and shape characteristics are used to determine skin conditions [41]. The development of an efficient model that can capture the in-depth feature details of a lesion image for accurate classification is required. Thus, in the proposed work, a novel approach named N-DCNN, inspired by convolutional neural networks, was designed for dermoscopic skin cancer classification. The classification model was developed via the careful organization of deep learning layers; further, it was trained and tested on the publicly available ISIC 2020 dataset. The dermoscopic image dataset consists of duplicate images; thus, to overcome bias issues, the duplicate removal method was designed based on a correlation index. Further, a data augmentation method was applied to balance the classes and to avoid overfitting. The presence of hairlines, low contrast, and unwanted background area were considered in this investigation regarding whether they impact classification accuracy. From the statistical results presented in Table 7, it can be seen that the preprocessing operations including the removal of hairlines, increasing contrast, and segmenting lesions significantly improved the performance of the N-DCNN classification network. For each stage of preprocessing, novel approaches were developed such as a hair removal method Section 3.2.1 based on morphological operations, and a contrast enhancement model Section 3.2.2 to regain the lost texture contrast of the lesion during the hair removal process. After this, ACNN, a segmentation deep learning model Section 3.3 based on the atrous dilation concept, is applied to remove unwanted pixels in the background. Through the computation of rich features at multiple layers, thN-DCNN efficiently analyzes very complex lesion patterns. The model gained an ACC of 90.92% using raw data and 93.40% using preprocessed data. Furthermore, the inference time for raw images was 76 s. On the other hand, the preprocessed images required only 17 s for the same test set. The low inference time can be attributed to the fact that the network takes less time to compute lesion features from an image that contains the ROI only rather than processing the whole image. For melanoma classification, the test time on a single image using N-DCNN was quite low at 1.8 s on raw and 1.3 s on preprocessed data. It was found that N-DCNN could improve the accuracy and reduce the architectural and computational complexity. Table 4 demonstrates the classification performance of the proposed N-DCNN model with state-of-the-art classifiers. The method by S. Nasiri in [50] achieved an average accuracy of 75%. M.R.Hassan [51] performed a comparative analysis of various deep learning approaches and listed VGG16 as the best model that achieved 93.18%, but the number of parameters was higher. Brinker [52] designed a CNN model on 12,378 open-source dermoscopic images and obtained a precision rate of 74.10%. The study compared the results of 100 images with 157 dermatologists, and deep learning models outperformed the experts. Another study by Kwasigroch [53] achieved a low accuracy rate of 77%. When compared with these studies, the proposed N-DCNN model showed promising accuracy of 91% on raw images and 93.4% on preprocessed images. Further, these studies have not considered the issues of hairlines, low contrast, and segmenting ROIs. Moreover, the studies [54,55,56] mentioned in Table 5 did not remove the duplicates from the ISIC 2020 dataset. N-DCNN performed well despite the irrelevant visual indicators and imaging artifacts when evaluated on a large and enhanced dermoscopic image dataset.

6. Conclusions

In this work, a deep convolutional neural network was designed for melanoma vs. benign classification. An automatic system for detecting melanoma would play a significant role in the early diagnosis of this global health problem. Skin cancer occurrence is quickly increasing and influencing various communities globally. Many have lost their lives because of slow testing processes and the lack of facilities because this cancer was not diagnosed in the early stages. This work proposed a novel N-DCNN model for the binary classification of melanoma and benign, which was designed by organizing different deep learning layers and by hypertuning network parameters. Resizing, oversampling, and augmentation operations were applied to the data before the training process. The experimental results showed that the N-DCNN model outperformed CNN models such as ResNet18, InceptionV3, Xception, and other related methods [50,51,52,53]. The model also achieved the highest AUC compared with the related methods [54,55,56]. As a diagnostic tool, the N-DCNN model demonstrates a high true positive rate and is reliable in predicting the correct lesion category. In addition, other major contributions of this work are the development of an image duplicate elimination method, a hairline removal algorithm, image contrast enhancement, and a segmentation model. The proposed N-DCNN was evaluated using preprocessed datasets by removing hairlines as well as applying image enhancement and segmentation approaches to improve its performance. A significant improvement in N-DCNN performance was obtained with preprocessed data, achieving 93.4% accuracy, illustrating the importance of the preprocessing and segmentation operations.

Author Contributions

Conceptualization: R.K. and H.G.; methodology: R.K. and H.G.; software: R.K.; validation: R.K. and H.G.; data collection: R.K.; writing—original draft preparation: R.K.; writing—review and editing: H.G. and M.L.; supervision: H.G. and M.L.; proofreading and revision: H.G. and M.L. All authors have read and agreed to the published version of this manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data were obtained from a publicly available benchmark dataset: https://www.isic.org/.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
N-DCNNNovel deep convolutional neural network
ISICInternational skin imaging collaboration
ReLURectified linear unit
BNBatch normalization
MCANMultiscale context aggregation network
ACNNAtrous convolutional neural network

References

  1. Massey University of New Zealand. Environmental Health Indicators New Zealand. Available online: https://www.ehinz.ac.nz/indicators/uv-exposure/melanoma/ (accessed on 10 July 2020).
  2. Autier, P.; for Epimel, D.; Eortc Melanoma Cooperative Group, J.F. Influence of sun exposures during childhood and during adulthood on melanoma risk. Int. J. Cancer 1998, 77, 533–537. [Google Scholar] [CrossRef]
  3. Schadendorf, D.; Fisher, D.E.; Garbe, C.; Gershenwald, J.E.; Grob, J.J.; Halpern, A.; Herlyn, M.; Marchetti, M.A.; McArthur, G.; Ribas, A.; et al. Melanoma. Nat. Rev. Dis. Prim. 2015, 1, 1–20. [Google Scholar] [CrossRef] [PubMed]
  4. Stern, R.S. Prevalence of a history of skin cancer in 2007: Results of an incidence-based model. Arch. Dermatol. 2010, 146, 279–282. [Google Scholar] [CrossRef]
  5. Cokkinides, V.; Albano, J.; Samuels, A.; Ward, M.; Thum, J. American Cancer Society: Cancer Facts and Figures; American Cancer Society: Atlanta, GA, USA, 2005. [Google Scholar]
  6. Siegel, R.L.; Miller, K.D.; Jemal, A. Cancer statistics, 2022. CA Cancer J. Clin. 2022, 72, 7–33. [Google Scholar] [CrossRef] [PubMed]
  7. Arnold, M.; Singh, D.; Laversanne, M.; Vignat, J.; Vaccarella, S.; Meheus, F.; Cust, A.E.; de Vries, E.; Whiteman, D.C.; Bray, F. Global burden of cutaneous melanoma in 2020 and projections to 2040. JAMA Dermatol. 2022, 158, 495–503. [Google Scholar] [CrossRef]
  8. Di Leo, G.; Paolillo, A.; Sommella, P.; Fabbrocini, G. Automatic diagnosis of melanoma: A software system based on the 7-point check-list. In Proceedings of the 43rd Hawaii International Conference on System Sciences, Honolulu, HI, USA, 5–8 January 2010. [Google Scholar]
  9. Guy, G.P., Jr.; Ekwueme, D.U.; Tangka, F.K.; Richardson, L.C. Melanoma treatment costs: A systematic review of the literature, 1990–2011. Am. J. Prev. Med. 2012, 43, 537–545. [Google Scholar] [CrossRef]
  10. Salido, J.A.; Ruiz, C., Jr. Hair artifact removal and skin lesion segmentation of dermoscopy images. Asian J. Pharm. Clin. Res. 2018, 11, 36–39. [Google Scholar] [CrossRef]
  11. Zaqout, I.S. An efficient block-based algorithm for hair removal in dermoscopic images. Comput. Opt. 2017, 41, 521–527. [Google Scholar] [CrossRef]
  12. Kasmi, R.; Mokrani, K.; Rader, R.; Cole, J.; Stoecker, W. Biologically inspired skin lesion segmentation using a geodesic active contour technique. Ski. Res. Technol. 2016, 22, 208–222. [Google Scholar] [CrossRef] [PubMed]
  13. Abbas, Q.; Garcia, I.F.; Emre Celebi, M.; Ahmad, W. A feature-preserving hair removal algorithm for dermoscopy images. Ski. Res. Technol. 2013, 19, e27–e36. [Google Scholar] [CrossRef]
  14. Garcia-Arroyo, J.L.; Garcia-Zapirain, B. Segmentation of skin lesions in dermoscopy images using fuzzy classification of pixels and histogram thresholding. Comput. Methods Programs Biomed. 2019, 168, 11–19. [Google Scholar] [CrossRef] [PubMed]
  15. Zghal, N.S.; Derbel, N. Melanoma skin cancer detection based on image processing. Curr. Med. Imaging 2020, 16, 50–58. [Google Scholar] [CrossRef]
  16. Amin, J.; Sharif, A.; Gul, N.; Anjum, M.A.; Nisar, M.W.; Azam, F.; Bukhari, S.A.C. Integrated design of deep features fusion for localization and classification of skin cancer. Pattern Recognit. Lett. 2020, 131, 63–70. [Google Scholar] [CrossRef]
  17. Garg, S.; Jindal, B. Skin lesion segmentation using k-mean and optimized fire fly algorithm. Multimed. Tools Appl. 2021, 80, 7397–7410. [Google Scholar] [CrossRef]
  18. Khan, M.A.; Sharif, M.; Akram, T.; Damaševičius, R.; Maskeliūnas, R. Skin lesion segmentation and multiclass classification using deep learning features and improved moth flame optimization. Diagnostics 2021, 11, 811. [Google Scholar] [CrossRef]
  19. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  20. Hong, S.; Noh, H.; Han, B. Decoupled deep neural network for semi-supervised semantic segmentation. arXiv 2015, arXiv:1506.04924. [Google Scholar]
  21. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  22. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  23. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  24. Sáez, A.; Sánchez-Monedero, J.; Gutiérrez, P.A.; Hervás-Martínez, C. Machine learning methods for binary and multiclass classification of melanoma thickness from dermoscopic images. IEEE Trans. Med. Imaging 2015, 35, 1036–1045. [Google Scholar] [CrossRef] [PubMed]
  25. Gautam, D.; Ahmed, M.; Meena, Y.K.; Ul Haq, A. Machine learning—Based diagnosis of melanoma using macro images. Int. J. Numer. Methods Biomed. Eng. 2018, 34, e2953. [Google Scholar] [CrossRef]
  26. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  27. Maglogiannis, I.; Pavlopoulos, S.; Koutsouris, D. An integrated computer supported acquisition, handling, and characterization system for pigmented skin lesions in dermatological images. IEEE Trans. Inf. Technol. Biomed. 2005, 9, 86–98. [Google Scholar] [CrossRef] [PubMed]
  28. Jaworek-Korjakowska, J.; Kłeczek, P. Automatic classification of specific melanocytic lesions using artificial intelligence. BioMed Res. Int. 2016, 2016, 8934242. [Google Scholar] [CrossRef] [PubMed]
  29. Fujisawa, Y.; Otomo, Y.; Ogata, Y.; Nakamura, Y.; Fujita, R.; Ishitsuka, Y.; Watanabe, R.; Okiyama, N.; Ohara, K.; Fujimoto, M. Deep-learning-based, computer-aided classifier developed with a small dataset of clinical images surpasses board-certified dermatologists in skin tumour diagnosis. Br. J. Dermatol. 2019, 180, 373–381. [Google Scholar] [CrossRef] [PubMed]
  30. Masood, A.; Ali Al-Jumaily, A. Computer aided diagnostic support system for skin cancer: A review of techniques and algorithms. Int. J. Biomed. Imaging 2013, 2013, 323268. [Google Scholar] [CrossRef] [PubMed]
  31. Brinker, T.J.; Hekler, A.; Utikal, J.S.; Grabe, N.; Schadendorf, D.; Klode, J.; Berking, C.; Steeb, T.; Enk, A.H.; Von Kalle, C. Skin cancer classification using convolutional neural networks: Systematic review. J. Med. Internet Res. 2018, 20, e11936. [Google Scholar] [CrossRef] [PubMed]
  32. Premaladha, J.; Ravichandran, K. Novel approaches for diagnosing melanoma skin lesions through supervised and deep learning algorithms. J. Med. Syst. 2016, 40, 96. [Google Scholar] [CrossRef] [PubMed]
  33. Popescu, D.; El-Khatib, M.; El-Khatib, H.; Ichim, L. New Trends in Melanoma Detection Using Neural Networks: A Systematic Review. Sensors 2022, 22, 496. [Google Scholar] [CrossRef]
  34. Jones, O.; Matin, R.; van der Schaar, M.; Bhayankaram, K.P.; Ranmuthu, C.; Islam, M.; Behiyat, D.; Boscott, R.; Calanzani, N.; Emery, J.; et al. Artificial intelligence and machine learning algorithms for early detection of skin cancer in community and primary care settings: A systematic review. Lancet Digit. Health 2022, 4, e466–e476. [Google Scholar] [CrossRef]
  35. Jeyakumar, J.P.; Jude, A.; Priya Henry, A.G.; Hemanth, J. Comparative Analysis of Melanoma Classification Using Deep Learning Techniques on Dermoscopy Images. Electronics 2022, 11, 2918. [Google Scholar] [CrossRef]
  36. Alam, T.M.; Shaukat, K.; Khan, W.A.; Hameed, I.A.; Almuqren, L.A.; Raza, M.A.; Aslam, M.; Luo, S. An Efficient Deep Learning-Based Skin Cancer Classifier for an Imbalanced Dataset. Diagnostics 2022, 12, 2115. [Google Scholar] [CrossRef] [PubMed]
  37. Bakator, M.; Radosav, D. Deep learning and medical diagnosis: A review of literature. Multimodal Technol. Interact. 2018, 2, 47. [Google Scholar] [CrossRef]
  38. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  39. Dorj, U.O.; Lee, K.K.; Choi, J.Y.; Lee, M. The skin cancer classification using deep convolutional neural network. Multimed. Tools Appl. 2018, 77, 9909–9924. [Google Scholar] [CrossRef]
  40. Mahbod, A.; Schaefer, G.; Wang, C.; Ecker, R.; Ellinge, I. Skin lesion classification using hybrid deep neural networks. In Proceedings of the ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 1229–1233. [Google Scholar]
  41. Gouda, W.; Sama, N.U.; Al-Waakid, G.; Humayun, M.; Jhanjhi, N.Z. Detection of skin cancer based on skin lesion images using deep learning. Healthcare 2022, 10, 1183. [Google Scholar] [CrossRef] [PubMed]
  42. Afza, F.; Sharif, M.; Mittal, M.; Khan, M.A.; Hemanth, D.J. A hierarchical three-step superpixels and deep learning framework for skin lesion classification. Methods 2022, 202, 88–102. [Google Scholar] [CrossRef] [PubMed]
  43. Sloan Kettering Cancer Center. The International Skin Imaging Collaboration. 2019. Available online: https://www.isic-archive.com/#!/topWithHeader/wideContentTop/main/ (accessed on 13 February 2019).
  44. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 180161. [Google Scholar] [CrossRef] [PubMed]
  45. Rotemberg, V.; Kurtansky, N.; Betz-Stablein, B.; Caffery, L.; Chousakos, E.; Codella, N.; Combalia, M.; Dusza, S.; Guitera, P.; Gutman, D.; et al. A patient-centric dataset of images and metadata for identifying melanomas using clinical context. Sci. Data 2021, 8, 34. [Google Scholar] [CrossRef] [PubMed]
  46. Kaur, R.; GholamHosseini, H.; Sinha, R. Hairlines removal and low contrast enhancement of melanoma skin images using convolutional neural network with aggregation of contextual information. Biomed. Signal Process. Control 2022, 76, 103653. [Google Scholar] [CrossRef]
  47. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  48. Powers, D.M. Evaluation: From precision, recall and F-measure to ROC, informedness, markedness and correlation. arXiv 2020, arXiv:2010.16061. [Google Scholar]
  49. Tharwat, A. Classification assessment methods. Appl. Comput. Inf. 2021, 17, 168–192. [Google Scholar] [CrossRef]
  50. Nasiri, S.; Helsper, J.; Jung, M.; Fathi, M. DePicT Melanoma Deep-CLASS: A deep convolutional neural networks approach to classify skin lesion images. BMC Bioinform. 2020, 21, 84. [Google Scholar] [CrossRef]
  51. Hasan, M.R.; Fatemi, M.I.; Monirujjaman Khan, M.; Kaur, M.; Zaguia, A. Comparative Analysis of Skin Cancer (Benign vs. Malignant) Detection Using Convolutional Neural Networks. J. Healthc. Eng. 2021, 2021, 5895156. [Google Scholar] [CrossRef] [PubMed]
  52. Brinker, T.J.; Hekler, A.; Enk, A.H.; Klode, J.; Hauschild, A.; Berking, C.; Schilling, B.; Haferkamp, S.; Schadendorf, D.; Holland-Letz, T.; et al. Deep learning outperformed 136 of 157 dermatologists in a head-to-head dermoscopic melanoma image classification task. Eur. J. Cancer 2019, 113, 47–54. [Google Scholar] [CrossRef] [PubMed]
  53. Kwasigroch, A.; Grochowski, M.; Mikołajczyk, A. Neural architecture search for skin lesion classification. IEEE Access 2020, 8, 9061–9071. [Google Scholar] [CrossRef]
  54. Karki, S.; Kulkarni, P.; Stranieri, A. Melanoma classification using EfficientNets and Ensemble of models with different input resolution. In Proceedings of the 2021 Australasian Computer Science Week Multiconference, Dunedin, New Zealand, 1–5 February 2021; pp. 1–5. [Google Scholar]
  55. Ha, Q.; Liu, B.; Liu, F. Identifying melanoma images using efficientnet ensemble: Winning solution to the siim-isic melanoma classification challenge. arXiv 2020, arXiv:2010.05351. [Google Scholar]
  56. O’Brien, M.; Bukowski, J.; Hager, G.; Pezeshk, A.; Unberath, M. Evaluating neural network robustness for melanoma classification using mutual information. In Proceedings of the Medical Imaging 2022: Image Processing, San Diego, CA, USA, 20–24 February 2022; Volume 12032, pp. 173–177. [Google Scholar]
  57. Bajwa, M.N.; Muta, K.; Malik, M.I.; Siddiqui, S.A.; Braun, S.A.; Homey, B.; Dengel, A.; Ahmed, S. Computer-aided diagnosis of skin diseases using deep neural networks. Appl. Sci. 2020, 10, 2488. [Google Scholar] [CrossRef]
  58. Dildar, M.; Akram, S.; Irfan, M.; Khan, H.U.; Ramzan, M.; Mahmood, A.R.; Alsaiari, S.A.; Saeed, A.H.M.; Alraddadi, M.O.; Mahnashi, M.H. Skin cancer detection: A review using deep learning techniques. Int. J. Environ. Res. Public Health 2021, 18, 5479. [Google Scholar] [CrossRef] [PubMed]
  59. Rigel, D.S.; Carucci, J.A. Malignant melanoma: Prevention, early detection, and treatment in the 21st century. CA Cancer J. Clin. 2000, 50, 215–236. [Google Scholar] [CrossRef]
Figure 1. Images of different skin cancer types: (a) MEL, (b) BCC, (c) BKL, (d) CMN, (e) BEN.
Figure 1. Images of different skin cancer types: (a) MEL, (b) BCC, (c) BKL, (d) CMN, (e) BEN.
Sensors 25 00594 g001
Figure 2. Overall design of the proposed system.
Figure 2. Overall design of the proposed system.
Sensors 25 00594 g002
Figure 3. (a) Input image; (b) grayscale; (c) dilation, erosion, and closing operations; (d) hair mask; (e) noise-free image.
Figure 3. (a) Input image; (b) grayscale; (c) dilation, erosion, and closing operations; (d) hair mask; (e) noise-free image.
Sensors 25 00594 g003
Figure 4. Overview of the MCAN model.
Figure 4. Overview of the MCAN model.
Sensors 25 00594 g004
Figure 5. The architecture of the segmentation model (ACNN).
Figure 5. The architecture of the segmentation model (ACNN).
Sensors 25 00594 g005
Figure 6. Atrous convolution operations by selecting the blue pixel using the dilation rates. (a) rate = 1, (b) rate = 2, (c) rate = 4.
Figure 6. Atrous convolution operations by selecting the blue pixel using the dilation rates. (a) rate = 1, (b) rate = 2, (c) rate = 4.
Sensors 25 00594 g006
Figure 7. The architecture of a classification network, N-DCNN.
Figure 7. The architecture of a classification network, N-DCNN.
Sensors 25 00594 g007
Figure 8. Duplicate image elimination method.
Figure 8. Duplicate image elimination method.
Sensors 25 00594 g008
Figure 9. Performance of classification network: (a) ACC vs. loss, (b) confusion matrix, (c) sensitivity vs. specificity curves.
Figure 9. Performance of classification network: (a) ACC vs. loss, (b) confusion matrix, (c) sensitivity vs. specificity curves.
Sensors 25 00594 g009
Figure 10. Performance illustration of N-DCNN model in terms of accuracy and loss with preprocessed and raw images.
Figure 10. Performance illustration of N-DCNN model in terms of accuracy and loss with preprocessed and raw images.
Sensors 25 00594 g010
Table 1. The training parameters of the ACNN.
Table 1. The training parameters of the ACNN.
ParameterValues
Input image size 192 × 256 × 3
Batch size16
Learning parameter, α 0.01
L2 regularization0.005
Momentum, γ 0.9
Epochs30
Loss function E ( θ ) weighted cross-entropy
OptimiserSGDM
Table 2. Hyperparameters for the training the N-DCNN.
Table 2. Hyperparameters for the training the N-DCNN.
ParameterValues
Learning algorithmSGDM
Learning rate0.001
Mini-batch size32
Epochs100
Activation functionLeaky ReLU
Data augmentationRandom oversampling, rotation,
translation, and scaling
Momentum0.99
Regularization0.0005
Table 3. Distribution of ISIC data among training, validation, and test sets.
Table 3. Distribution of ISIC data among training, validation, and test sets.
ClassesTraining SamplesAugmented TrainingValidation SamplesTest SamplesTotal Samples
MEL4052496557911585789
BEN4965496570914197093
Total901799301288257712,882
Table 4. Comparison of N-DCNN with other state-of-the-art methods.
Table 4. Comparison of N-DCNN with other state-of-the-art methods.
StudiesDatasetNo. of
Images
ACC%PRE%REC%SPE%F1 Score%Parameters
(Millions)
S. Nasiri [50]ISIC134675.0077.0073.0078.0075.00
M.R.Hasan [51]ISIC659493.18134.2 M
T.J Brinker [52]ISIC12,37874.1087.5060.00>23 M
Kwasigroch, A. [53]ISIC13,60077.007.18 M
ProposedISIC12,88290.9291.1292.2989.2991.703.32 M
Table 5. N-DCNN performance compared to other related methods based on AUC.
Table 5. N-DCNN performance compared to other related methods based on AUC.
StudiesDatasetDuplicates RemovedModelAUC
S. Karki [54]2020NOEnsemble Nets94.11
Q. Ha [55]2020NOEnsemble Efficient-Net94.90
M. O’Brien [56]2020NODNN59.1
Proposed2020YesN-DCNN96.42
Table 6. Comparison of proposed N-DCNN with baseline CNNs on the ISIC datasets.
Table 6. Comparison of proposed N-DCNN with baseline CNNs on the ISIC datasets.
ApproachACCPRERECSPEF1 ScoreParametersTest Time
ResNet1885.3284.8688.5481.5486.6611 M171 s
Inceptionv390.5092.8290.5290.4891.6624 M260 s
Xception82.780.079.881.982.122.8 M202 s
N-DCNN90.9291.1292.2989.2991.703.32 M76 s
Table 7. Performance comparison of N-DCNN on the preprocessed data.
Table 7. Performance comparison of N-DCNN on the preprocessed data.
DataACCPRERECSPEF1 ScoreExecution Time (s)Per Image Test Time (s)
Raw data90.9292.1192.2989.2991.70761.8
Preprocessed data93.4093.4594.5192.0893.98171.3
Note: Execution time was for the complete test dataset. The best values are marked in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kaur, R.; GholamHosseini, H.; Lindén, M. Advanced Deep Learning Models for Melanoma Diagnosis in Computer-Aided Skin Cancer Detection. Sensors 2025, 25, 594. https://doi.org/10.3390/s25030594

AMA Style

Kaur R, GholamHosseini H, Lindén M. Advanced Deep Learning Models for Melanoma Diagnosis in Computer-Aided Skin Cancer Detection. Sensors. 2025; 25(3):594. https://doi.org/10.3390/s25030594

Chicago/Turabian Style

Kaur, Ranpreet, Hamid GholamHosseini, and Maria Lindén. 2025. "Advanced Deep Learning Models for Melanoma Diagnosis in Computer-Aided Skin Cancer Detection" Sensors 25, no. 3: 594. https://doi.org/10.3390/s25030594

APA Style

Kaur, R., GholamHosseini, H., & Lindén, M. (2025). Advanced Deep Learning Models for Melanoma Diagnosis in Computer-Aided Skin Cancer Detection. Sensors, 25(3), 594. https://doi.org/10.3390/s25030594

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop