Previous Article in Journal
Computed Tomography Angiography Identified High-Risk Coronary Plaques: From Diagnosis to Prognosis and Future Management
Previous Article in Special Issue
Prediction of the Cause of Glaucoma Disease Identified by Glaucoma Optical Coherence Tomography Test in Relation to Diabetes and Hypertension at a National Hospital in Seoul: A Retrospective Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assessing the Efficacy of the Spectrum-Aided Vision Enhancer (SAVE) to Detect Acral Lentiginous Melanoma, Melanoma In Situ, Nodular Melanoma, and Superficial Spreading Melanoma

1
Department of Dermatology, Dalin Tzu Chi General Hospital, No. 2, Min-Sheng Rd., Dalin Town, Chiayi 62247, Taiwan
2
Institute of Medicine, School of Medicine, College of Medicine, National Yang Ming Chiao Tung University, No. 155, Sec. 2, Li-Nong Street, Beitou District, Taipei 112304, Taiwan
3
Department of Surgery, Division of Plastic and Reconstructive Surgery, Taichung Veterans General Hospital, 1650 Taiwan Boulevard Sect. 4, Taichung 407219, Taiwan
4
Department of Mechanical Engineering, National Chung Cheng University, 168, University Rd., Min Hsiung, Chia Yi 62102, Taiwan
5
Department of Dermatology, Chung Shan Medical University Hospital, No. 110, Sec. 1, Jianguo N. Rd., South Dist., Taichung City 40201, Taiwan
6
Institute of Medicine, School of Medicine, Chung Shan Medical University, No. 110, Sec. 1, Jianguo N. Rd., South Dist., Taichung City 40201, Taiwan
7
Department of Surgery, Division of General Surgery, Kaohsiung Armed Forces General Hospital, 2, Zhongzheng 1st. Rd., Lingya District, Kaohsiung 80284, Taiwan
8
Department of Technology Development, Hitspectra Intelligent Technology Co., Ltd., Kaohsiung 80661, Taiwan
*
Authors to whom correspondence should be addressed.
Diagnostics 2024, 14(15), 1672; https://doi.org/10.3390/diagnostics14151672 (registering DOI)
Submission received: 22 May 2024 / Revised: 18 July 2024 / Accepted: 26 July 2024 / Published: 1 August 2024

Abstract

:
Skin cancer is the predominant form of cancer worldwide, including 75% of all cancer cases. This study aims to evaluate the effectiveness of the spectrum-aided visual enhancer (SAVE) in detecting skin cancer. This paper presents the development of a novel algorithm for snapshot hyperspectral conversion, capable of converting RGB images into hyperspectral images (HSI). The integration of band selection with HSI has facilitated the identification of a set of narrow band images (NBI) from the RGB images. This study utilizes various iterations of the You Only Look Once (YOLO) machine learning (ML) framework to assess the precision, recall, and mean average precision in the detection of skin cancer. YOLO is commonly preferred in medical diagnostics due to its real-time processing speed and accuracy, which are essential for delivering effective and efficient patient care. The precision, recall, and mean average precision (mAP) of the SAVE images show a notable enhancement in comparison to the RGB images. This work has the potential to greatly enhance the efficiency of skin cancer detection, as well as improve early detection rates and diagnostic accuracy. Consequently, it may lead to a reduction in both morbidity and mortality rates.

1. Introduction

Skin cancer is the prevailing form of cancer in people and its occurrence is increasing significantly [1]. Skin cancers are classified based on the specific type of cell they originate from and their clinical characteristics [2]. Acral lentiginous melanoma is an uncommon kind of melanoma that primarily develops on the hands, feet, and nail beds [3]. While it is the most commonly identified subtype of the disease in many areas, especially in Latin America, Africa, and Asia, it is still relatively unexplored due to its distinct genetic causes [4]. Melanoma in situ refers to melanomas that are confined to the epidermis [5]. Despite being a rare type of skin cancer, its annual growth rate has climbed by 9.5%, making it one of the fastest-growing malignancies [6]. There have been no reported deaths directly caused by instances of melanoma in situ. However, people with this condition are at a significantly increased risk of getting another type of malignant melanoma (MM) in the future [7]. Cutaneous melanoma refers to a diverse group of malignant melanocytic growths that vary in terms of their epidemiology, morphology, growth patterns, genetics, and ability to spread to other parts of the body [8]. Nodular melanoma and superficial spreading melanoma are two prominent kinds of cutaneous melanoma. Nodular melanoma is a subtype of melanoma that is distinguished by its fast development rate. It is estimated that nodular melanoma invades the skin at a pace of approximately 0.5 mm per month [9].
Early diagnosis is the crucial determinant for the effective treatment of skin cancer [10]. A skin biopsy is a method of acquiring a sample from a suspected site of skin cancer, which is a difficult, time-consuming, and unpleasant procedure. Computer-aided diagnosis (CAD) offers a non-invasive, expedited, and real-time method for diagnosing skin cancer. Kousis et al. devised 11 advanced deep learning techniques to achieve a precise identification of skin cancer using a mobile application. Among these methods, the one utilizing DenseNet169 yielded the highest performance, with an accuracy of 92.25% and a recall rate of 93.59% [11]. Mehr et al. employed the inception-ResNet-v2 CNN deep learning architecture to identify skin cancer in lesion images, achieving an average accuracy of 89.3% [12]. Agrahari et al. developed a very efficient system for detecting skin cancer in many classes. They utilized a pre-trained MobileNet model on the HAM10000 ISIC dataset. The system demonstrated a level of performance that was comparable to that of a dermatological expert [13]. However, the majority of research primarily concentrates on RGB imaging, which utilizes only three distinct bands for picture processing. Nevertheless, the utilization of hyperspectral imaging (HSI) can greatly augment the quantity of data that can be acquired from each pixel, resulting in a greater level of precision.
The HSI technique is composed of numerous bands that have the ability to capture very detailed spectral responses that enable HSI to effectively identify even the most minor alterations in histopathological images [14]. An HSI image consists of pixels, each of which has a complete spectrum that makes it a highly effective tool for defining and studying biological substances [15]. In recent years, this new imaging modality has been utilized in medical applications [16,17,18,19]. Tsai et al. utilized HSI in conjunction with single shot multibox detector (SSD) architecture to detect esophageal cancer. Their study showed a significant improvement of over 5% in detection accuracy when HSI was employed instead of white-light imaging (WLI) [20]. HSI has not been fully employed in the field of skin cancer diagnosis within the bio-medical industry. The majority of studies primarily focus on detecting the most common types of skin cancer, such as basal cell carcinoma (BCC), squamous cell carcinoma (SCC), and seborrheic keratosis (SK), while paying less attention to the less common forms of skin cancer [21].
Narrow band imaging (NBI) is a specific sort of hyperspectral imaging (HSI) that captures images within a small range of wavelengths. This focused technique improves particular characteristics of an image, making it extremely efficient for medical diagnostics and other applications where precision and differentiation are essential [22]. This technology is based on the utilization of two short wavelength light beams, one at 415 nm (blue) and the other at 540 nm (green), to enhance optical images [23]. Longer wavelengths of light have a greater ability to enter tissues because of their scattering and absorption qualities. Specifically, 415 nm improves the visibility of superficial mucosal vascular forms, whereas 540 nm increases submucosal IPCLs [24]. This results in a noticeable difference in the picture, with the superficial vessels appearing in brown and the submucosal vessels appearing in cyan. This enhances the distinction between the blood vessels and the surrounding mucosa. [25]. Many studies have proven that NBI is better than white-light imaging (WLI) in terms of accuracy, sensitivity, and specificity [26,27,28,29,30].
Therefore, in this study, a novel method known as the spectrum-aided vision enhancer (SAVE) has been developed which has the ability to convert any WLI image into an HSI image and has been combined with band selection to select particular narrow bands. These SAVE images and WLI images of four different classes of skin cancers including acral lentiginous melanoma, melanoma in situ, nodular melanoma, and superficial spreading melanoma were trained using multiple YOLO architectures and the results were compared in terms of sensitivity, precision, F1-score, and accuracy.

2. Materials and Methods

2.1. Dataset

A total of 878 images was used in this study for the analysis of four different skin cancer classes including acral lentiginous melanoma with a total of 342 images, superficial spreading melanoma with a total of 253 images, nodular melanoma with a total of 100 images, and melanoma in situ with 183 images. The dataset was divided into training, testing, and validation in the ratio of 7:2:1, respectively, as shown in Table 1. These images were obtained from our partner hospital as well as from the International Skin Imaging Collaboration (ISIC), with equal contributions. Labelimg (version 1.8.6) software was used for the annotation and marking of the images. Image augmentation techniques such as clockwise and counter-clockwise rotation of 90 degrees and shear of 10 degrees in both vertical and horizontal orientations were also performed. Images were resized to 640 × 640 pixels to fit the YOLO architectures. The early disable function was also disabled to make the architectures run for 600 epochs fully, the batch size was set to 16, and value curves of the training and validation sets were determined after 600 epochs of training the loss function.
ALM is represented by 231 close-up images, 12 dermoscopy images, and 102 clinical images. NM is represented by 61 close-up images, 16 dermoscopy images, and 26 clinical images. MIS is represented by 21 close-up images, 20 dermoscopy images, and 136 clinical images. SSM is represented by 35 close-up images, 34 dermoscopy images, and 184 clinical images. The distribution of images for each diagnosis group exhibits a notable disparity, which has the potential to induce bias in the model’s performance. The model is trained using a balanced strategy to mitigate the impact of any single modality on the outcomes, despite the variance in the quantity of images between modalities. This entails the utilization of normalization and augmentation strategies to alleviate the influence of features peculiar to each modality. The design of our model incorporates a wide range of images from multiple modalities, which improves its capacity to generalize across diverse input sources. The use of diverse imaging techniques in the model’s training data is advantageous rather than restrictive, as it equips the model to effectively handle real-world situations. The performance indicators, including as precision, recall, and mean average precision (mAP), are obtained through thorough validation and testing across all modalities. This guarantees that the model’s performance remains consistent and dependable, regardless of the type of image. Our methodology primarily emphasizes finding the fundamental illness patterns rather than surface-level picture attributes. The combination of the SAVE algorithm and the YOLO framework enhances the emphasis on pathological traits rather than artifacts peculiar to a certain modality (for a more detailed description of the dataset please see Supplementary Materials).

2.2. YOLO Architectures

The choice of YOLO v5 for this experiment was made because of prior research findings that have shown its higher detection speed when compared with alternative architectures, such as Densenet or SSD. This supremacy is ascribed to its capacity to improve real-time performance. The YOLO v5 model includes a total of three modules: the neck, backbone, and head terminals. Convolutional Neural Networks (CNNs) often consist of architectural components such as focus, CONV-BN-Leaky ReLU (CBL), cross-stage partial (CSP), and spatial pyramid pooling (SPP) models [31]. Focus possesses the ability to combine multiple high-resolution images, divide the input image into smaller segments, reduce the required CUDA 11.8 memory and layers, enhance the rate of both forward and backward propagation, and extract image features. The provided image, which has a fixed size of 640 × 640 pixels, is split into four separate pictures, each measuring 320 × 320 pixels. After being partitioned, these images undergo CON-CAT, where they are combined with slices and layers of convolution kernels. This process produces an image with dimensions of 320 × 320 pixels. This partitioning strategy is utilized to optimize the efficacy of the training process. The SPP pooling layer is specifically designed to overcome the limitations of input size while preserving the integrity of the image. In this context, the term “neck” denotes a collection of layers that integrate characteristics from an image, resulting in the development of feature pyramid networks (FPN) and path aggregation networks (PAN) [32]. The system comprises the CBL, Upsample, CSP2_X, and several other models. YOLO v5 preserves the CSP1_X architecture from YOLO v4’s CSPDarknet-53 version and integrates the CSP2_X architecture to reduce the model’s dimensions and capture a more extensive range of image details. The GIoU Loss is a loss function utilized to quantify the difference between predicted and actual bounding boxes in the head. The loss function of YOLO v5 comprises three distinct types of losses: classification losses, confidence losses, and bounding box regression losses [33]. The loss function quantifies the discrepancy between the observed and predicted results of the model. The YOLO v5 model’s loss function is defined as follows:
L G I O U = i = 0 S 2 j = 0 B I i j o b j 1 I O U + A c U A c
λ n o o b j i = 0 S 2 j = 0 B I i j n o o b j c ^ i j l o g C i j + 1 C i j l o g 1 C i j
L c l a s s = i = 0 S 2 I i j n o o b j
c classes P ^ i j c l o g P i j c + 1 P ^ i j c l o g 1 P i j c
In January 2023, Ultralytics (Los Angeles, CA, USA) confirmed the newest member of the YOLO family, YOLO v8 [34]. The YOLO v8 concept does not rely on anchors. This suggests that instead of making a prediction about the distance between an item and a known anchor box, it directly calculates the center of the object. Anchor-free detection reduces the need for predicting bounding boxes, hence accelerating Non-Maximum Suppression (NMS), a complex post-processing step that filters potential detections after inference. The fundamental unit was modified, with C2f substituting C3, and the original 6 × 6 convolution of the stem being substituted with a 3 × 3. YOLO v8 utilizes the PAN-FPN approach in the neck section to enhance the fusion and exploitation of feature layer information at various scales [35,36]. YOLO v8 utilizes two methods for increasing resolution and many C2f modules in combination with the ultimate decoupled head structure to create the neck module. It also incorporates the concept of CSP from YOLO v5 [37]. YOLO v8 includes the concept of separating the head in YOLOx and applies it to the latter half of the neck. The integration of confidence and regression boxes results in a heightened level of accuracy.
Wang et al. released YOLO v9, the most recent version of the YOLO family of real-time object identification models, in February 2024 [38]. This marks a notable progression in the field of object detection, with the goal of surpassing methods based on convolution and transformers. This model incorporates programmable gradient information (PGI) and the Generalized Efficient Layer Aggregation Network (GELAN) to improve accuracy [39]. This type is specifically intended for object detection jobs and is renowned for its exceptional efficiency in real-time detection, offering enhanced accuracy and speed [40]. YOLO v9 has been evaluated against its predecessor, YOLO v8, in multiple tests, showcasing its higher accuracy while keeping the same training efficiency. Although YOLO v9 has a longer training period, it provides a favorable balance between accuracy and efficiency, making it a potential option for applications that need accurate object detection. The performance of the model in reliably detecting objects has been thoroughly analyzed in terms of precision, recall, F1-score, and loss functions.

2.3. Spectrum-Aided Vision Enhancer

This study devised a VIS-HSI transformation mechanism called SAVE (developed by Hitspectra Intelligent Technology Co., Ltd., Kaohsiung City, Taiwan), which can convert an RGB image taken by an electronic camera into an HSI image, as illustrated in Figure 1. Firstly, it is essential to determine the relationship among an RGB picture and the spectrometer’s readings for different colors. The specified colors for calibration are identified as the X-Rite Classic. The tool consists of 24 squares displaying a diverse range of color samples commonly found in natural surroundings. The samples include the colors red, green, blue, cyan, magenta, and yellow, as well as six different hues of gray. X-Rite has gained popularity as a preferred option for color calibration in recent years. The main objective of the camera was to capture photographs that accurately represented the colors of the Macbeth Color Checker, which acted as the reference point. The patch image, which contained 24 colors, was transformed to the CIE 1931 XYZ color space. The digital camera recorded an image that was subsequently stored in JPEG, using the standard RGB (sRGB) color scheme. The first stage entailed converting the R, G, and B values (ranging from 0 to 255) within the sRGB color space into a more limited range that spans from 0 to 1. Later on, the Gamma function was employed to transform the normalized sRGB characteristics into calibrated RGB values. The RGB values were converted to the CIE 1931 color space using a transformation matrix. This matrix depicts the quantitative relationship among the wavelengths in the visible spectrum (VIS) and the perceived color in the natural world. However, the images captured with a digital camera might be affected by various causes such as nonlinear response, dark current, improper color separation, or color shift. Thus, Equation (5), which represents a matrix with varying values, was utilized. Equation (6) was utilized to compute the calibrated values of X, Y, and Z, known as XYZcorrect, as part of the error correction procedure. (the equations for modifying the 24-color patch picture and 24-color reflectance spectrum information to transform them into the XYZ color space are explained in Supplementary Materials Section S1).
C = X Y Z S p e c t r u m × p i n v ( V )
X Y Z C o r r c n t = C × [ V ]
The brightness ratio was determined by utilizing the Y value of the XYZ color gamut space, as this parameter exhibits a direct correlation with brightness. The data obtained from the reflectance spectrum were converted into the XYZ value (XYZSpectrum), which was then standardized to suit the XYZ color gamut space. The correction coefficient matrix C was derived through the application of multiple regression analysis, as described in Equation (3). The transformation matrix (M) was derived using the reflectance spectrum data (Rspectrum). Principle component analysis (PCA) was conducted on the Rspectrum dataset in order to obtain 6 significant principal components (PCs) and eigenvectors. These components were able to account for 99.64% of the information. The average root mean square error (RMSE) between the XYZcorrect and XYZSpectrum models for the 24 target colors was 0.19, indicating a minor difference. Next, a multiple regression analysis (MRA) which is a method of statistical analysis that can be employed to examine the correlation between one dependent variable and multiple independent variables was conducted on M and correlated with the six significant PCs. These PCs were utilized in the multivariate regression analysis of XYZcorrect. The variable Vcolor was deliberately selected due to its comprehensive coverage of all permutations of the X, Y, and Z values. The calibration of the camera is a crucial component of the SAVE algorithm. Following this, the color exhibited a striking resemblance to the color acquired by the spectrum analyzer, rendering the distinction challenging to perceive. Prior to camera calibration, the mean chromatic aberration of all 24 color blocks was 10.76. After calibrating, the mean chromatic aberration decreased to a minimum of 0.63. The analog spectrum (SSpectrum) was computed from XYZcorrect using Equation (4). Next, the SSpectrum was assessed in comparison to the Rspectrum. The RMSE for the 24 color blocks was found to be 0.056, while the mean color variation between the analog spectrum and reflectance spectrum produced from the spectrometer was determined to be 0.75. This observation suggests that the colors derived from the reflectance spectrum closely resembled the colors associated with the observed values. Hence, employing the aforementioned methodology, it is possible to convert an RGB image into an HSI image.
The spectrometer employed for this study was the Ocean Optics QE65000, which produces a reflectance spectrum of a 24-color patch. The brightness ratio was calculated by using the value for Y of the XYZ color gamut space, as this variable has a direct association with brightness. The reflectance spectrum data were transferred into the XYZ value (XYZSpectrum), and subsequently adjusted to conform to the XYZ color gamut space. The correction coefficient matrix C was calculated by applying MRA, as outlined in Equation (7). The transformation matrix (M) was obtained by utilizing the reflectance spectrum data (Rspectrum). PCA was conducted on the Rspectrum dataset to identify the 6 most relevant principal components (PCs) and their corresponding eigenvectors. These components were capable of explaining 99.64% of the data being analyzed. The RMSE comparing the XYZcorrect and XYZSpectrum estimates for the 24 target colors had an average value of 0.19, suggesting a negligible disparity. Subsequently, an MRA was conducted on variable M and its correlation was examined with the six significant principal components. The PCs were employed in the multivariate regression analysis of XYZcorrect. The variable Vcolor was intentionally chosen since it encompasses every possible combination of the X, Y, and Z values. The calibration of the camera is an essential element of the SAVE algorithm. After calibrating the camera, the color closely matched the color obtained by the spectrum analyzer, making it difficult to distinguish between them. Before calibrating the camera, the mean chromatic aberration of all 24 color blocks was 10.76. After calibrating the camera, the chromatic aberration was reduced to a minimum value of 0.63. Equation (8) was used to calculate the analog spectrum (SSpectrum) from XYZcorrect. Subsequently, the SSpectrum was evaluated in relation to the Rspectrum. The RMSE for the 24 color blocks was calculated to be 0.056. Additionally, the average color discrepancy within the SSpectrum and Rspectrum generated by the spectrometer was measured to be 0.75. This finding indicates that the colors obtained from the reflectance spectrum closely matched the colors linked to the observed values. Therefore, by utilizing the previously indicated approach, it is feasible to transform an RGB image into an HSI image.
M = S c o r e × p i n v ( V C o l o r )
[ S S p e c t r u m ] 380 ~ 780   n m = E V M [ V C o l o r ]

2.4. Band Selection

Although the SAVE algorithm has the ability to convert the RGB images into HSI images, suitable narrow-bands must be selected from the spectrum of visible bands to enhance the diagnostic accuracy. For this, in this study, the narrow bands used in the Olympus endoscope were simulated, calculated, and replicated. The simulated NBI image from the SAVE conversion algorithm must be similar to the real NBI image captured by Olympus endoscope. Also, for this calibration, the standard 24-color checker was used. The NBI image simulated from the HSI conversion algorithm was compared with the real NBI image captured by the Olympus (Tokyo, Japan) endoscope. The CIEDE 2000 color difference between each of the 24 color blocks was measured and minimized. After the correction, the average color difference of the 24 color blocks was found to be only 2.79, which is negligible. Although the peak absorption wavelengths of hemoglobin are 415 nm and 540 nm, the real NBI image taken by the Olympus endoscope includes not only green and blue colors, but also various shades of brown, which correspond to a wavelength of 650 nm. Figure 2 displays the contrast of the wavelength between SAVE, Olympus WLI, and Olympus NBI. Thus, it can be inferred that there is a nuanced image post-processing technique that enhances the realism of NBI videos. Therefore, in this study, in addition to the wavelengths of 415 nm and 540 nm, three other regions at 600 nm, 700 nm, and 780 nm also exhibit a light spectrum. The entirety of the process is shown in Figure 3.

2.5. Evaluation Indices

The diagnostic performance of the YOLO architectures was assessed in this study using multiple factors, including precision, recall, specificity, accuracy, and F1 score. Accuracy, in the context of machine learning, is a quantitative metric that evaluates the proportion of accurate predictions made by the model compared to the total number of predictions. The calculation involves evaluating the proportion of accurate predictions to the overall number of predictions [41]. A higher degree of precision indicates a model that exhibits superior performance in terms of overall accuracy. Precision is a numerical measure that determines the proportion of accurate positive predictions made by a model out of the overall amount of positive predictions. The calculation entails the division of the total amount of true positives by the sum of precise positives and false positives [42]. Precision is a metric that quantifies the accuracy of positive predictions by measuring the percentage of predicted positive instances that are truly correct. Recall, also known as sensitivity, is a quantitative metric that evaluates the proportion of accurate positive predictions made by a model compared to the overall amount of actual positive occurrences. The calculation requires dividing the total amount of true positives by the sum of true positives and false negatives. The concept of recall emphasizes the model’s ability to correctly identify positive instances from the overall amount of actual positives [43]. The F1-score is a quantitative metric that quantifies the harmonic average of precision and recall. The purpose of this is to achieve a fair assessment between these two metrics [44]. The calculation is obtained by multiplying the values of precision and recall and then multiplying the resulting product by 2. The product is then divided by the total number of precision and recall values. The F1-score is a useful metric for evaluating the effectiveness of the architecture in classification tasks as it considers both false positives and false negatives. Mean Average Precision (mAP) is a crucial metric employed in object detection tasks to evaluate the precision of models in detecting and classifying objects in images [45]. The calculation involves determining the average precision (AP) for all recall values ranging from 0 to 1. mAP, which stands for mean Average Precision, is a commonly used metric in the field of computer vision.
A c c u r a c y = T P + T N T P + T N + F P + F N × 100
P r e c i s i o n = T P T P + F P × 100
R e c a l l = T P T P + F N × 100
F 1 s c o r e = 2 × P × R P + R × 100
TP = True positive; TN = True negative; FN = False negative; FP = False positive; P = Precision; and R = Recall.

3. Results

3.1. SAVE Performance Evaluation

The data generated through simulation were utilized for the initial evaluation of the system’s performance. The characterization involved measuring their spectrum emissions, while also taking into account the sensitivities provided by the camera manufacturers. The spectral curves of the 24-color Macbeth Color Checker chart were utilized for both the training and validation sets of the samples. Table 2 displays the RMSEs and standard deviation (SD) values of the XYZ color space coordinates before and after calibration for a total of 24 colors. RMSE measures the average difference between anticipated and actual values. A lower RMSE number indicates a higher level of accuracy in the calibration process. The XYZ values denote color coordinates that are crucial for precise image representation. SD quantifies the dispersion of these numbers, where a smaller SD indicates more consistency and reliability of the data. Table 2 illustrates the efficacy of the calibration method in diminishing mistakes and enhancing the accuracy of color depiction in photographs. Calibrating the SAVE is essential to ensure the dependability of its ability to convert RGB photos into HSI-NBI for precise skin cancer detection.
The calibration of the camera is a crucial component of the SAVE algorithm. Figure 4 displays the color disparity outcomes prior to and following calibration. Following the calibration of the camera, the color exhibited a striking resemblance to the color acquired by the spectrum analyzer, rendering the distinction challenging to perceive. Prior to camera calibration, the mean chromatic aberration of all 24 color blocks was 10.76. After calibrating the camera, the average chromatic aberration decreased to a minimum of 0.63.
In Figure 5, the reflectance values of the six primary colors inside the 24-color block are illustrated. These colors are blue (13), red (15), green (14), yellow (16), magenta (17), and cyan (18). Based on the analysis of the 24 color blocks, it was noted that the red block exhibited the most significant disparity between the simulated and actual reflectance values, particularly across the longer wavelength range of 600 to 780 nm. One of the limitations of the study is considered to be this element. All of the remaining 23 color blocks exhibited RMSEs below 0.1, with the color black demonstrating the lowest RMSE of 0.015. The RMSE was merely 0.056, suggesting that the majority of the color could be replicated with precision.
RMSE values can be graphically and mathematically represented by calculating the disparity between simulated and measured colors. The representation of a color can be denoted as LAB, where L represents lightness, channel A, and channel B, respectively. The numerical definition of any color can be achieved by manipulating the values of L, A, and B. The L, A, and B values of the simulated and computed colors are depicted in Figure 6. The average color disparity was about 0.75, suggesting that the replicated color was visually precise.

3.2. Model Performance Evaluation

In this study examining the efficacy of the SAVE algorithm using different YOLO machine learning models and the Roboflow 3.0 model, a comprehensive analysis was performed focusing on metrics like precision, recall, mAP, and F1-score are shown in Table 3 and Table 4. The YOLO v8 model emerged as the most effective, outperforming other models such as YOLO v9, YOLO v5, YOLO-NAS, and Roboflow 3.0 in various aspects of object detection. This superior performance is highlighted by its higher mAP score, which reflects better overall prediction accuracy across multiple threshold settings. The superior precision of the YOLO v8-SAVE model, achieving above 90%, contrasts with a lower recall of 71%, indicating that while the model is highly accurate in detecting true positives, it falls short in identifying all actual positives. This is further underscored by comparing it to the YOLO v8-RGB model, which balanced precision and recall more evenly at 84.3% and 75.5%, respectively. This discrepancy suggests that the YOLO v8 model, when configured for the SAVE algorithm, prioritizes accuracy over coverage, potentially overlooking some true positive cases in favor of maintaining high confidence in the cases it does detect. To address this imbalance, particularly the lower recall, adjustments such as modifying the decision threshold, enhancing feature selection, or enriching the training data with more diverse examples could be beneficial. These changes might help the model better identify subtle features indicative of positive cases, thereby increasing the recall without sacrificing precision significantly. YOLO v5, while second in overall performance, demonstrated an interesting pattern where the recall was higher relative to its precision. This suggests that YOLO v5 may be better suited in scenarios where missing a true positive is more detrimental than misclassifying a negative as a positive. The high mAP of 81.9% for the YOLO v5-SAVE model indicates robustness in its predictive capability across various thresholds, although it trails behind YOLO v8 in precision and overall performance. The YOLO v9 model had a similar result in comparison with the YOLO v5 model in terms of precision and mAP. However, both YOLO v5 and YOLO v9 had a better recall and F1-score. The YOLO-NAS model showed comparatively poor results in precision, recall, and mAP, highlighting possible limitations in its architecture or training data that could be impeding its performance. This might require a reevaluation of the model’s structural parameters or an enhancement of its learning algorithm to better capture the necessary features for accurate diagnosis.
In terms of precision, acral lentiginous melanoma and nodular melanoma, in general, showed higher precision in the SAVE model in both YOLO v5 and YOLO v8 models while melanoma in situ and superficial spreading melanoma had inconsistent results, with a decrease of almost 23% and 7% in the YOLO v5-SAVE model and an 18% and 1% increase in YOLO v8 respectively. In terms of recall across the two ML architectures, the RGB model had either better or similar results compared to SAVE except in two occasions where SAVE outperformed the WLI model. YOLO v9 had a different pattern of results with acral lentiginous melanoma being the lowest in terms of precision with 65.7% in the RGB model, which was improved 29% in the SAVE model, and melanoma in situ had a higher precision of 94.9% in the RGB model which decreased to 79.8% in the SAVE model. However, similar to the other models, the recall values when compared with the precision were less, although the average mAP50 in the YOLO v9 model was found to be higher. This finding showcased the model’s exceptional level of accuracy in learning lesion features. The accuracy rates of both the RGB model and the SAVE model appear to be similar due to the relatively smaller number of images used. Nevertheless, this study demonstrates that the SAVE-based conversion algorithms, which possess the ability to transform RGB images into HSI images, have the potential to accurately classify and detect skin cancers. While the majority of the findings are comparable, this study has notably enhanced the accuracy rate as shown in Figure 7.

4. Discussion

Given these varied performances, the choice of model can be strategic depending on the specific requirements of the medical diagnostic task. For instance, in settings where missing a diagnosis could lead to severe outcomes, a model with higher recall like YOLO v5 might be preferable. Conversely, in situations where precision is paramount to avoid unnecessary treatments, YOLO v8 would be more appropriate. Thus, selecting the optimal model involves a trade-off between these metrics, tailored to the needs of the medical application. This analysis not only underscores the strengths and weaknesses of each model but also offers a pathway to refine these tools for enhanced diagnostic accuracy in medical imaging. One key limitation of this study is the potential lack of generalizability due to the specific dataset used, which might not adequately represent the diverse range of skin types and cancer stages. Additionally, the study primarily focused on quantitative metrics such as precision and recall, potentially overlooking qualitative aspects like model interpretability and clinical usability. The models were evaluated in controlled experimental settings, which may not fully capture their performance in real-world clinical environments. Lastly, the computational demands of these advanced models might limit their practical application in resource-constrained settings, necessitating further optimization for broader deployment. To overcome the limitations identified in this study, future research will prioritize diversifying the dataset to include a wider range of skin types, conditions, and stages of cancer to enhance the models’ generalizability. Incorporating qualitative assessments such as user feedback from clinicians can help refine the models for better usability and interpretability in real-world settings. Additionally, developing lightweight versions of these models or leveraging newer, more efficient machine learning frameworks could reduce computational demands, making the technology more accessible in resource-limited environments. Collaboration with clinical partners for pilot testing can also provide critical insights into practical challenges and user requirements.
The future scope of this study in evaluating the SAVE algorithm through various YOLO models and the Roboflow 3.0 framework in skin cancer detection is expansive and promising. Given the preliminary findings, future research can explore several avenues to enhance the accuracy and applicability of these models. Firstly, refining the SAVE algorithm with advanced machine learning techniques such as deep learning ensemble methods could further improve the precision and recall metrics, potentially leading to better diagnostic outcomes. Exploring the integration of additional imaging modalities alongside hyperspectral imaging could also provide richer data for model training and validation, increasing the robustness of the diagnostic system. Moreover, the application of the models in real-world clinical settings as part of routine skin cancer screenings could provide valuable feedback on their operational effectiveness and user-friendliness. This practical deployment can help identify unforeseen challenges and user requirements, leading to iterative model improvements. Additionally, expanding the dataset to include a broader spectrum of skin types, conditions, and stages of cancer will be crucial for ensuring the models’ generalizability and sensitivity across diverse patient populations. Another significant area for future research is the development of real-time diagnostic tools using these models. Leveraging the real-time processing capabilities of YOLO architectures can facilitate the development of mobile or handheld devices for immediate skin cancer screening and diagnosis, making effective use of the models’ ability to provide instant results. These efforts combined could vastly improve the early detection and treatment of skin cancer, ultimately saving lives and reducing healthcare costs.

5. Conclusions

This study has effectively demonstrated the superior performance of the YOLO v8 model integrated with the SAVE algorithm for detecting skin cancer, surpassing other models like YOLO v5, YOLO-NAS, YOLO v9, and Roboflow 3.0 in terms of precision, recall, mAP, and F1-score. Despite its impressive precision, YOLO v8 showed a comparatively lower recall, suggesting a need for adjustments in model threshold and training strategies to better balance detection accuracy with the ability to identify all relevant cases. YOLO v5 also showed promising results, particularly in recall, indicating its potential utility in scenarios where higher sensitivity is required. The study highlights the importance of tailored model selection based on specific diagnostic needs and sets the stage for the further refinement and real-world application of these advanced machine learning tools in the medical field. Future research should focus on enhancing model generalizability, exploring additional imaging modalities, and expanding the deployment of these technologies in clinical settings.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics14151672/s1, Figure S1. Training and Loss curves of WLI and SAVE model in YOLO v5. Figure S2. F1-Confidence Curve of WLI and SAVE in YOLO v5. Figure S3. Training and Loss curves of WLI and SAVE models in YOLO v8. Figure S4. F1-Confidence Curve of WLI and SAVE in YOLO v8. Figure S5. Training and Loss curves of WLI and SAVE models in YOLO v9. Figure S6. F1-Confidence Curve of WLI and SAVE in YOLO v9. Figure S7. Comparison between WLI and SAVE when using YOLO v5 in terms of %. Figure S8. Comparison between WLI and SAVE when using YOLO v5 in terms of quantitative numbers. Figure S9. Comparison between WLI and SAVE when using YOLO v8 in terms of %. Figure S10. Comparison between WLI and SAVE when using YOLO v8 in terms of quantitative numbers. Figure S11. Training and Loss curves of WLI and SAVE model in YOLO v9. Figure S12. Prediction results of in the RGB model. Figure S13. Prediction results in SAVE.

Author Contributions

Conceptualization, T.-L.L., S.-C.H., Y.-P.H., and H.-C.W.; methodology, A.M., H.-C.W., Y.-P.H., K.N., and S.-C.H.; software, K.N., R.K., H.-C.W., and A.M.; validation, S.-C.H., R.K., T.-L.L., Y.-P.H., and H.-C.W.; formal analysis, C.-T.L., S.-C.H., R.K., Y.-P.H., T.-L.L., and H.-C.W.; investigation, S.-C.H., K.N., and A.M.; resources, S.-C.H., T.-L.L., and H.-C.W.; data curation, A.M., R.K., H.-C.W., Y.-P.H., and C.-T.L.; writing—original draft preparation, A.M. and C.-T.L.; writing—review and editing, A.M. and H.-C.W.; supervision, Y.-P.H. and H.-C.W.; project administration, H.-C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Science and Technology Council, the Republic of China under the grants NSTC 112-2221-E-194-036 and 112-2222-E-194-002. This work was financially/partially supported by the Dalin Tzu Chi Hospital, Buddhist Tzu Chi Medical Foundation-National Chung Cheng University Joint Research Program (DTCRD113-C-01) and the Kaohsiung Armed Forces General Hospital Research Program KAFGH_D_113024 in Taiwan.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Dalin Tzu Chi General Hospital (B11302024).

Informed Consent Statement

Written informed consent was waived in this study because of the retrospective, anonymized nature of the study design.

Data Availability Statement

The data presented in this study are available from the corresponding author (H.-C.W.) upon reasonable request.

Conflicts of Interest

Hsiang-Chen Wang was employed by the company Hitspectra Intelligent Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Linares, M.A.; Zakaria, A.; Nizran, P. Skin cancer. Prim. Care Clin. Off. Pract. 2015, 42, 645–659. [Google Scholar] [CrossRef] [PubMed]
  2. Simoes, M.F.; Sousa, J.S.; Pais, A.C. Skin cancer and new treatment perspectives: A review. Cancer Lett. 2015, 357, 8–42. [Google Scholar] [CrossRef] [PubMed]
  3. Goydos, J.S.; Shoen, S.L. Acral lentiginous melanoma. Melanoma 2016, 321–329. [Google Scholar]
  4. Basurto-Lozada, P.; Molina-Aguilar, C.; Castaneda-Garcia, C.; Vázquez-Cruz, M.E.; Garcia-Salinas, O.I.; Álvarez-Cano, A.; Martínez-Said, H.; Roldán-Marín, R.; Adams, D.J.; Possik, P.A. Acral lentiginous melanoma: Basic facts, biological characteristics and research perspectives of an understudied disease. Pigment Cell Melanoma Res. 2021, 34, 59–71. [Google Scholar] [CrossRef] [PubMed]
  5. Balch, C.M.; Gershenwald, J.E.; Soong, S.-J.; Thompson, J.F.; Atkins, M.B.; Byrd, D.R.; Buzaid, A.C.; Cochran, A.J.; Coit, D.G.; Ding, S. Final version of 2009 AJCC melanoma staging and classification. J. Clin. Oncol. 2009, 27, 6199. [Google Scholar] [CrossRef] [PubMed]
  6. National Cancer Institute Surveillance, Epidemiology, and End Results (SEER) Program. Cancer Statistics, SEER Data & Software, Registry Operations. 2018. Available online: https://seer.cancer.gov/archive/csr/1975_2018/ (accessed on 27 July 2024).
  7. Higgins, H.W., II; Lee, K.C.; Galan, A.; Leffell, D.J. Melanoma in situ: Part, I. Epidemiology, screening, and clinical features. J. Am. Acad. Dermatol. 2015, 73, 181–190. [Google Scholar] [CrossRef] [PubMed]
  8. Corneli, P.; Zalaudek, I.; Magaton Rizzi, G.; di Meo, N. Improving the early diagnosis of early nodular melanoma: Can we do better? Expert Rev. Anticancer Ther. 2018, 18, 1007–1012. [Google Scholar] [CrossRef]
  9. Liu, W.; Dowling, J.P.; Murray, W.K.; McArthur, G.A.; Thompson, J.F.; Wolfe, R.; Kelly, J.W. Rate of growth in melanomas: Characteristics and associations of rapidly growing melanomas. Arch. Dermatol. 2006, 142, 1551–1558. [Google Scholar] [CrossRef] [PubMed]
  10. Dildar, M.; Akram, S.; Irfan, M.; Khan, H.U.; Ramzan, M.; Mahmood, A.R.; Alsaiari, S.A.; Saeed, A.H.M.; Alraddadi, M.O.; Mahnashi, M.H. Skin cancer detection: A review using deep learning techniques. Int. J. Environ. Res. Public Health 2021, 18, 5479. [Google Scholar] [CrossRef]
  11. Kousis, I.; Perikos, I.; Hatzilygeroudis, I.; Virvou, M. Deep learning methods for accurate skin cancer recognition and mobile application. Electronics 2022, 11, 1294. [Google Scholar] [CrossRef]
  12. Mehr, R.A.; Ameri, A. Skin cancer detection based on deep learning. J. Biomed. Phys. Eng. 2022, 12, 559. [Google Scholar]
  13. Agrahari, P.; Agrawal, A.; Subhashini, N. Skin cancer detection using deep learning. In Proceedings of the Futuristic Communication and Network Technologies: Select Proceedings of VICFCNT 2020, Chennai, India, 6–7 November 2020; pp. 179–190. [Google Scholar]
  14. Lu, B.; Dao, P.D.; Liu, J.; He, Y.; Shang, J. Recent advances of hyperspectral imaging technology and applications in agriculture. Remote Sens. 2020, 12, 2659. [Google Scholar] [CrossRef]
  15. El Masry, G. Principles of hyperspectral imaging technology. In Hyperspectral Imaging for Food Quality Analysis and Control; Sun, D.-W., Ed.; Elsevier: Dublin, Ireland, 2010; pp. 3–43. [Google Scholar]
  16. Fei, B. Hyperspectral imaging in medical applications. In Data Handling in Science and Technology; Elsevier: Dublin, Ireland, 2019; Volume 32, pp. 523–565. [Google Scholar]
  17. Tsai, T.-J.; Mukundan, A.; Chi, Y.-S.; Tsao, Y.-M.; Wang, Y.-K.; Chen, T.-H.; Wu, I.-C.; Huang, C.-W.; Wang, H.-C. Intelligent Identification of Early Esophageal Cancer by Band-Selective Hyperspectral Imaging. Cancers 2022, 14, 4292. [Google Scholar] [CrossRef]
  18. Mukundan, A.; Tsao, Y.-M.; Wang, H.-C. Early esophageal detection using hyperspectral engineering and convolutional neural network. In Proceedings of the Optics in Health Care and Biomedical Optics XIII, Beijing, China, 14–16 October 2023; pp. 12–19. [Google Scholar]
  19. Yang, K.-Y.; Mukundan, A.; Tsao, Y.-M.; Shi, X.-H.; Huang, C.-W.; Wang, H.-C. Assessment of hyperspectral imaging and CycleGAN-simulated narrowband techniques to detect early esophageal cancer. Sci. Rep. 2023, 13, 20502. [Google Scholar] [CrossRef] [PubMed]
  20. Tsai, C.-L.; Mukundan, A.; Chung, C.-S.; Chen, Y.-H.; Wang, Y.-K.; Chen, T.-H.; Tseng, Y.-S.; Huang, C.-W.; Wu, I.-C.; Wang, H.-C. Hyperspectral imaging combined with artificial intelligence in the early detection of esophageal cancer. Cancers 2021, 13, 4593. [Google Scholar] [CrossRef] [PubMed]
  21. Huang, H.-Y.; Hsiao, Y.-P.; Mukundan, A.; Tsao, Y.-M.; Chang, W.-Y.; Wang, H.-C. Classification of Skin Cancer Using Novel Hyperspectral Imaging Engineering via YOLOv5. J. Clin. Med. 2023, 12, 1134. [Google Scholar] [CrossRef]
  22. Sano, Y.; Tanaka, S.; Kudo, S.E.; Saito, S.; Matsuda, T.; Wada, Y.; Fujii, T.; Ikematsu, H.; Uraoka, T.; Kobayashi, N. Narrow-band imaging (NBI) magnifying endoscopic classification of colorectal tumors proposed by the Japan NBI Expert Team. Dig. Endosc. 2016, 28, 526–533. [Google Scholar] [CrossRef] [PubMed]
  23. Zheng, C.; Lv, Y.; Zhong, Q.; Wang, R.; Jiang, Q. Narrow band imaging diagnosis of bladder cancer: Systematic review and meta-analysis. BJU Int. 2012, 110, E680–E687. [Google Scholar] [CrossRef]
  24. Gono, K.; Obi, T.; Yamaguchi, M.; Ohyama, N.; Machida, H.; Sano, Y.; Yoshida, S.; Hamamoto, Y.; Endo, T. Appearance of enhanced tissue features in narrow-band endoscopic imaging. J. Biomed. Opt. 2004, 9, 568–577. [Google Scholar] [CrossRef]
  25. van Schaik, J.E.; Halmos, G.B.; Witjes, M.J.; Plaat, B.E. An overview of the current clinical status of optical imaging in head and neck cancer with a focus on Narrow Band imaging and fluorescence optical imaging. Oral Oncol. 2021, 121, 105504. [Google Scholar] [CrossRef]
  26. Kim, D.H.; Lee, M.H.; Lee, S.; Kim, S.W.; Hwang, S.H. Comparison of Narrowband Imaging and White-Light Endoscopy for Diagnosis and Screening of Nasopharyngeal Cancer. Otolaryngol. Head Neck Surg. 2022, 166, 795–801. [Google Scholar] [CrossRef] [PubMed]
  27. Ezoe, Y.; Muto, M.; Horimatsu, T.; Minashi, K.; Yano, T.; Sano, Y.; Chiba, T.; Ohtsu, A. Magnifying narrow-band imaging versus magnifying white-light imaging for the differential diagnosis of gastric small depressive lesions: A prospective study. Gastrointest. Endosc. 2010, 71, 477–484. [Google Scholar] [CrossRef] [PubMed]
  28. Ezoe, Y.; Muto, M.; Uedo, N.; Doyama, H.; Yao, K.; Oda, I.; Kaneko, K.; Kawahara, Y.; Yokoi, C.; Sugiura, Y. Magnifying narrowband imaging is more accurate than conventional white-light imaging in diagnosis of gastric mucosal cancer. Gastroenterology 2011, 141, 2017–2025.e3. [Google Scholar] [CrossRef] [PubMed]
  29. Park, J.-S.; Seo, D.-W.; Song, T.J.; Park, D.H.; Lee, S.S.; Lee, S.K.; Kim, M.-H. Usefulness of white-light imaging–guided narrow-band imaging for the differential diagnosis of small ampullary lesions. Gastrointest. Endosc. 2015, 82, 94–101. [Google Scholar] [CrossRef] [PubMed]
  30. Du Le, V.; Wang, Q.; Gould, T.; Ramella-Roman, J.C.; Pfefer, T.J. Vascular contrast in narrow-band and white light imaging. Appl. Opt. 2014, 53, 4061–4071. [Google Scholar] [CrossRef] [PubMed]
  31. Ren, J.; Wang, Z.; Zhang, Y.; Liao, L. YOLOv5-R: Lightweight real-time detection based on improved YOLOv5. J. Electron. Imaging 2022, 31, 033033. [Google Scholar] [CrossRef]
  32. Liu, H.; Sun, F.; Gu, J.; Deng, L. Sf-yolov5: A lightweight small object detection algorithm based on improved feature fusion mode. Sensors 2022, 22, 5817. [Google Scholar] [CrossRef] [PubMed]
  33. Wang, J.; Xiao, T.; Gu, Q.; Chen, Q. YOLOv5_CSL_F: YOLOv5’s loss improvement and attention mechanism application for remote sensing image object detection. In Proceedings of the 2021 International Conference on Wireless Communications and Smart Grid (ICWCSG), Hangzhou, China, 13–15 August 2021; pp. 197–203. [Google Scholar]
  34. Hussain, M. YOLO-v1 to YOLO-v8, the rise of YOLO and its complementary nature toward digital manufacturing and industrial defect detection. Machines 2023, 11, 677. [Google Scholar] [CrossRef]
  35. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  36. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 10–17 June 2018; pp. 8759–8768. [Google Scholar]
  37. Wang, C.-Y.; Liao, H.-Y.M.; Wu, Y.-H.; Chen, P.-Y.; Hsieh, J.-W.; Yeh, I.-H. CSPNet: A new backbone that can enhance learning capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 390–391. [Google Scholar]
  38. Wang, C.-Y.; Yeh, I.-H.; Liao, H.-Y.M. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar]
  39. Chien, C.-T.; Ju, R.-Y.; Chou, K.-Y.; Chiang, J.-S. YOLOv9 for Fracture Detection in Pediatric Wrist Trauma X-ray Images. arXiv 2024, arXiv:2403.11249. [Google Scholar] [CrossRef]
  40. Aziz, F.; Saputri, D.U.E. Efficient Skin Lesion Detection using YOLOv9 Network. J. Med. Inform. Technol. 2024, 2, 11–15. [Google Scholar] [CrossRef]
  41. Arisholm, E.; Briand, L.C.; Johannessen, E.B. A systematic and comprehensive investigation of methods to build and evaluate fault prediction models. J. Syst. Softw. 2010, 83, 2–17. [Google Scholar] [CrossRef]
  42. Gray, D.; Bowes, D.; Davey, N.; Sun, Y.; Christianson, B. Further thoughts on precision. In Proceedings of the 15th Annual Conference on Evaluation & Assessment in Software Engineering (EASE 2011), Durham, UK, 11–12 April 2011; pp. 129–133. [Google Scholar]
  43. Saito, T.; Rehmsmeier, M. The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets. PLoS ONE 2015, 10, e0118432. [Google Scholar] [CrossRef] [PubMed]
  44. Lipton, Z.C.; Elkan, C.; Narayanaswamy, B. Thresholding classifiers to maximize F1 score. arXiv 2014, arXiv:1402.1892. [Google Scholar]
  45. Henderson, P.; Ferrari, V. End-to-end training of object class detectors for mean average precision. In Proceedings of the Computer Vision–ACCV 2016: 13th Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; pp. 198–213, (revised 2017). [Google Scholar]
Figure 1. SAVE flowchart.
Figure 1. SAVE flowchart.
Diagnostics 14 01672 g001
Figure 2. Comparison of wavelengths of Olympus WLI, Olympus NBI, and SAVE.
Figure 2. Comparison of wavelengths of Olympus WLI, Olympus NBI, and SAVE.
Diagnostics 14 01672 g002
Figure 3. Flowchart of the procedure for band selection.
Figure 3. Flowchart of the procedure for band selection.
Diagnostics 14 01672 g003
Figure 4. Color difference before and after camera calibration.
Figure 4. Color difference before and after camera calibration.
Diagnostics 14 01672 g004
Figure 5. RMSEs between analog and measured spectra of each color block.
Figure 5. RMSEs between analog and measured spectra of each color block.
Diagnostics 14 01672 g005
Figure 6. LAB values of the simulated and observed colors.
Figure 6. LAB values of the simulated and observed colors.
Diagnostics 14 01672 g006
Figure 7. Different classes of skin cancer used in this study. (a) and (e) represent acral lentiginous melanoma in RGB and SAVE, respectively; (b) and (f) represent melanoma in situ in RGB and SAVE, respectively; (c) and (g) represent nodular melanoma in RGB and SAVE, respectively; and (d) and (h) represent superficial spreading melanoma in RGB and SAVE, respectively.
Figure 7. Different classes of skin cancer used in this study. (a) and (e) represent acral lentiginous melanoma in RGB and SAVE, respectively; (b) and (f) represent melanoma in situ in RGB and SAVE, respectively; (c) and (g) represent nodular melanoma in RGB and SAVE, respectively; and (d) and (h) represent superficial spreading melanoma in RGB and SAVE, respectively.
Diagnostics 14 01672 g007
Table 1. Dataset used in this study.
Table 1. Dataset used in this study.
ClassificationTrainingValidationTesting
Acral Lentignious Melanoma2396934
Melanoma in Situ1283718
Nodular Melanoma702010
Superficial Spreading Melanoma1785025
Total61517687
Table 2. RMSEs of the XYZ values before and after calibration.
Table 2. RMSEs of the XYZ values before and after calibration.
S. NoBefore CalibrationAfter CalibrationRMSESD
XYZXYZ
110.969.924.6311.149.874.260.240.30
238.7435.8018.6538.5735.9418.660.130.08
316.6219.0724.1316.4818.7924.110.180.17
410.3312.864.6210.1613.034.850.190.19
524.0523.8731.5524.1624.0731.600.130.08
630.1242.1532.4030.1042.1732.420.020.002
738.1030.244.2838.0430.374.220.090.04
811.7011.4725.9011.6411.3725.910.070.02
929.0119.919.6229.2019.789.600.130.08
108.266.499.638.066.499.860.180.17
1134.1544.068.4434.1544.028.530.060.01
1247.9944.556.0548.0544.346.170.150.11
136.825.7921.076.905.9121.000.090.04
1414.5523.557.2214.5823.517.120.070.02
1521.0812.253.5721.0112.283.650.060.01
1658.4060.697.5458.3860.797.420.090.04
1728.9819.5420.6728.9419.5220.660.020.002
1812.8119.0128.5412.8419.1028.560.050.01
1982.1288.5467.2082.3188.7367.510.240.30
2054.7458.9245.5254.2858.4044.750.601.89
2133.0835.7327.2433.2635.8227.540.210.23
2218.1819.6214.9418.8620.3115.620.682.43
239.1310.018.138.569.267.210.763.04
242.873.192.393.103.352.680.230.27
Average0.190.39
Table 3. Evaluation of different ML models with respect to the two imaging modalities.
Table 3. Evaluation of different ML models with respect to the two imaging modalities.
ModelImage ModalityPrecisionRecallmApF1-Score
YOLO v5RGB0.8040.7160.7970.751
SAVE0.7990.8290.8190.810
YOLO v8RGB0.8430.7550.8070.795
SAVE0.9040.7100.8010.794
YOLO v9RGB0.8060.6050.7370.65
SAVE0.7830.6660.7750.71
YOLO-NASRGB0.7330.5410.6590.623
SAVE0.7310.6650.6900.69
Roboflow 3.0RGB0.7190.6430.6600.675
SAVE0.7810.6130.6800.68
Table 4. Precision, recall, mAP of the four different classes of skin cancer studies in YOLO v5 and YOLO v8.
Table 4. Precision, recall, mAP of the four different classes of skin cancer studies in YOLO v5 and YOLO v8.
ArchitectureModelSkin Cancer TypesPrecisionRecallmAP50mAP 50–95
YOLO v-5RGBAcral Lentiginous Melanoma0.8650.8750.9660.6
Melanoma in Situ0.780.5450.5970.323
Nodular Melanoma0.7320.90.9320.679
Superficial Spreading Melanoma0.7840.5420.6120.33
SAVEAcral Lentiginous Melanoma0.9190.8120.9030.551
Melanoma in Situ0.5510.4550.4810.223
Nodular Melanoma0.840.80.8420.6
Superficial Spreading Melanoma0.6690.6250.5980.299
YOLO v-8RGBAcral Lentiginous Melanoma0.9380.950.9650.562
Melanoma in Situ0.6850.5450.6080.294
Nodular Melanoma0.8670.90.9320.62
Superficial Spreading Melanoma0.8820.6230.7250.419
SAVEAcral Lentiginous Melanoma0.9460.8120.9570.55
Melanoma in Situ0.8590.5550.70.286
Nodular Melanoma0.8510.90.8910.631
Superficial Spreading Melanoma0.8750.5820.6390.338
YOLO v-9RGBAcral Lentiginous Melanoma0.6570.8460.8140.508
Melanoma in Situ0.9490.6670.8260.46
Nodular Melanoma0.8820.2310.5330.33
Superficial Spreading Melanoma0.7370.6760.7760.472
SAVEAcral Lentiginous Melanoma0.84910.9950.4
Melanoma in Situ0.7980.6070.7490.45
Nodular Melanoma0.7550.5710.7650.416
Superficial Spreading Melanoma0.730.4840.5790.31
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, T.-L.; Lu, C.-T.; Karmakar, R.; Nampalley, K.; Mukundan, A.; Hsiao, Y.-P.; Hsieh, S.-C.; Wang, H.-C. Assessing the Efficacy of the Spectrum-Aided Vision Enhancer (SAVE) to Detect Acral Lentiginous Melanoma, Melanoma In Situ, Nodular Melanoma, and Superficial Spreading Melanoma. Diagnostics 2024, 14, 1672. https://doi.org/10.3390/diagnostics14151672

AMA Style

Lin T-L, Lu C-T, Karmakar R, Nampalley K, Mukundan A, Hsiao Y-P, Hsieh S-C, Wang H-C. Assessing the Efficacy of the Spectrum-Aided Vision Enhancer (SAVE) to Detect Acral Lentiginous Melanoma, Melanoma In Situ, Nodular Melanoma, and Superficial Spreading Melanoma. Diagnostics. 2024; 14(15):1672. https://doi.org/10.3390/diagnostics14151672

Chicago/Turabian Style

Lin, Teng-Li, Chun-Te Lu, Riya Karmakar, Kalpana Nampalley, Arvind Mukundan, Yu-Ping Hsiao, Shang-Chin Hsieh, and Hsiang-Chen Wang. 2024. "Assessing the Efficacy of the Spectrum-Aided Vision Enhancer (SAVE) to Detect Acral Lentiginous Melanoma, Melanoma In Situ, Nodular Melanoma, and Superficial Spreading Melanoma" Diagnostics 14, no. 15: 1672. https://doi.org/10.3390/diagnostics14151672

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop