Next Article in Journal
ORPP—An Ontology for Skill-Based Robotic Process Planning in Agile Manufacturing
Previous Article in Journal
Electrothermal Averaged Model of a Half-Bridge DC–DC Converter Containing a Power Module
Previous Article in Special Issue
Predicting Economic Trends and Stock Market Prices with Deep Learning and Advanced Machine Learning Techniques
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Feature Extraction and Selection Method to Diagnose Burn Depth from Burn Images

by
Xizhe Zhang
1,
Qi Zhang
1,*,
Peixian Li
2,
Jie You
3,
Jingzhang Sun
4 and
Jianhang Zhou
5
1
Faculty of Data Science, City University of Macau, Macau SAR, China
2
School of Information Engineering, Yancheng Teachers University, Yancheng 224002, China
3
Shenzhen Intellirocks Tech Co., Ltd., Shenzhen 518000, China
4
School of Cyberspace Security, Hainan University, Haikou 570228, China
5
Department of Intelligent Media, Institute of Scientific and Industrial Research (SANKEN), Osaka University, Suita 565-0871, Japan
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(18), 3665; https://doi.org/10.3390/electronics13183665
Submission received: 10 August 2024 / Revised: 9 September 2024 / Accepted: 12 September 2024 / Published: 14 September 2024

Abstract

:
Burn wound depth is a significant determinant of patient treatment. Typically, the evaluation of burn depth relies heavily on the clinical experience of doctors. Even experienced surgeons may not achieve high accuracy and speed in diagnosing burn depth. Thus, intelligent burn depth classification is useful and valuable. Here, an intelligent classification method for burn depth based on machine learning techniques is proposed. In particular, this method involves extracting color, texture, and depth features from images, and sequentially cascading these features. Then, an iterative selection method based on random forest feature importance measure is applied. The selected features are input into the random forest classifier to evaluate this proposed method using the standard burn dataset. This method classifies burn images, achieving an accuracy of 91.76% when classified into two categories and 80.74% when classified into three categories. The comprehensive experimental results indicate that this proposed method is capable of learning effective features from limited data samples and identifying burn depth effectively.

1. Introduction

Burns are a global public health issue, with an estimated 11 million people requiring medical care for burns each year, according to the World Health Organization (WHO). Additionally, 180,000 people die from burns annually, with the vast majority occurring in low- and middle-income countries [1].
Damage from burn affects all cell proteins, causing denaturation. Based on clinical observations of skin anatomy, burns are commonly classified into four categories based on depth: superficial, superficial dermal, deep dermal, and full-thickness burns [2,3].
  • Superficial burn (first-degree burn [2])
Superficial burn only affects the skin epidermis, resulting in redness without blisters, dry surface, and pain upon touch, often accompany by edema.
2.
Superficial dermal burn
Superficial dermal burn penetrates into the papillary or superficial layer of the dermis, showing redness and moistness. Applying pressure to the red area may cause it to turn white and then back to red upon release, often accompanied by moderate edema and blister formation within minutes of injury.
3.
Deep dermal burn
Deep dermal burn extends down to the reticular layer of the dermis or deeper, with colors appearing red or waxy white. When pressure is applied, the red area continues to turn white, but upon pressure release, capillary refill may disappear or be slow. Blisters are usually not present, and the exposed surface of the wound is moist. Edema in the burn area is typically pronounced.
4.
Full-thickness burn
Full-thickness burn penetrates can affect the subcutaneous tissue layer. The skin affected by the burn is dry and tough, with colors possibly ranging from brown to red. A notable feature of a full-thickness burn is the absence of pain and pronounced edema in the burn area [4].
For superficial burn, the wounds can heal on their own without leaving scars. Thus, the analysis of it is not performed in this work. Superficial dermal burns can also heal on their own, but improper treatment may cause skin discoloration, which can gradually recover over time. Surgical treatment, grafting, or reconstructive therapy are necessary for deep dermal burn and full-thickness burn. Otherwise, the wound could worsen and become infected [5]. Deep dermal burn and full-thickness burn can be classified together because they both require surgical intervention and are unable to self-heal. Superficial burn and superficial dermal burn do not require surgery and can heal on their own [6].
Therefore, burns are categorized into superficial dermal burn, deep dermal burn, and full-thickness burn in this paper. Figure 1 illustrates these three types of burn injuries from left to right: superficial dermal burn, deep dermal burn, and full-thickness burn. In addition, deep dermal burns are combined with full-thickness burns into one category designated non-self-healing burns, while superficial skin burns are referred to as self-healing burns.
Currently, the initial clinical assessment of burn areas is often challenging because medical doctors must consider various factors during diagnosis, including the cause of the injury, the extent of the lesions, and tissue penetration [4]. The diagnostic accuracy of clinical assessment through visual and tactile examinations ranges between 50% and 80% [7]. In the worst-case scenario, some medical experts may have a difference of nearly 30% in diagnostic accuracy [7]. The high cost and workload of experienced doctors pose challenges during treatment, particularly in regions with lower economic levels, where ensuring the quality of burn patient care is difficult [8]. Considering these issues, a tool that automatically assesses the depth of burns is crucial for assisting clinicians in determining the severity of burns.
The schematic diagram of the method proposed in this work is shown in Figure 2. Firstly, the color histograms of the RGB, HSV, and L*A*B* color models are extracted from burn images as their color features. Texture features are then extracted by applying a Gabor filter to the images. Deep features are extracted in a modified ResNet-50 network named Burn-ResNet51. These three features are concatenated and arranged, and a feature selection method based on random forest (RF) is applied to reduce the dimensionality of the features in the training set, filtering out less important features. Subsequently, the corresponding features from the test set are selected. Finally, classification is performed using six classifiers, such as random forest. In summary, this study contributes in the following ways:
  • This proposed method can extract multiple types of features from burn images, which can represent the characteristics of images from multiple perspectives. Thus, the differences between different types of burn images can be captured more accurately and comprehensively.
  • Simple modifications are made to the ResNet-50 network, creating a new network named Burn-ResNet51, which reduces the dimensionality of features extracted in the max pooling layer to an optimal value, enabling better expression of deep features.
  • A random-forest-based feature extraction method is used before classification, which mitigates the negative impact of redundant features.
  • To ensure that each feature contributes to the model’s performance enhancement, multiple ablation experiments are conducted, thereby verifying the rationality of this proposed model and helping guide subsequent improvements to the model.
  • This proposed method is validated on the Burns BIP_US dataset and achieves the best performance compared to other common burn depth classification methods [9].
The rest of this paper is organized as follows. Section 2 introduces the related literature reviews or surveys of the methods involved in this study. In Section 3, four main approaches used in burn depth diagnosis are presented. Then, some experiments are conducted in Section 4 to prove the effectiveness of this proposed method. Section 5 illustrates possible future research directions. Finally, Section 6 offers the conclusion.

2. Related Work

Imaging techniques were used to assess burn depth in the past. Laser Doppler imaging (LDI) is one of the more effective traditional assessment methods. But the limitation of this approach lies in its accuracy of less than 80% within 24–48 h after the patient’s injury. Moreover, LDI also suffers from high costs and difficult calibration [10].
With the advancement of technology, machine learning models are increasingly being used for the classification of burn depth. For example, Badea et al. [11] used color as the primary feature to distinguish between different burn depths. However, the performance of this method was only comparable to the average performance reported in the literature by specialized doctors. Similarly, Ref. [12] leveraged color features to employ multidimensional scaling (MDS) to determine certain physical features, which were then transformed into mathematical features. Through classifiers and feature selection methods, the accuracy in binary classification reached 79.73%.
Since a single-color feature provides limited information, Ref. [13] also considered texture features in addition to color features. Feature extraction was performed using discrete wavelet transform (DWT) followed by principal component analysis (PCA) to reduce feature dimensionality. Texture features were then extracted from the decomposed images using gray-level co-occurrence matrix (GLCM). Similarly, the researchers in Ref. [14] also studied color and texture features. Their method transfers the images from the RGB space to the Cr space (one of YCbCr transformation) to separate the skin region from others, while contrast and uniformity were selected as texture features. This method performed well when classifying sub-images of burn pictures into two categories using SVM. Yadav et al. [9] used a binary classification method to classify a burn image dataset. The experiment combined traditional feature extraction methods, including the HOG feature extraction method, with an SVM classifier, achieving a classification accuracy of 82.43%.
In addition to color and texture features, depth features can also assist clinical doctors in identifying the depth of burns. For example, Rostami et al. [15] applied a deep learning method using the pre-trained deep convolutional neural network AlexNet on the dataset of burn images for fine-tuning and classification. The accuracy was 77.8% when classified into three categories and 82.43% when classified into two categories. Furthermore, the authors of [16] used a dataset of multiple images of body parts; their proposed method used three pre-trained and fine-tuned CNN models, ResNet-50, VGG-16, and VGG-19, for burn image classification. Ultimately, ResNet-50 attained an accuracy of 83.52% in the classification of burn depth into two categories, and surpassed the performance of other models. When the features extracted from the fully connected layer of ResNet-50 were combined with an SVM classifier, the accuracy increased to 84.85%. Moreover, Ref. [17] also used a Simultaneously Evolved SVM (SE-SVM) to classify deep features extracted from four different deep convolutional networks and applied them to diagnose COVID-19 from X-ray images. These studies demonstrate that pre-trained deep convolutional neural networks can extract effective features and the features can be used in the traditional classifier, thereby improving classification accuracy to a certain extent.
ResNet-50 not only demonstrates excellent performance in diagnosing burn depth but also shows promising results in the field of brain tumor classification. Kumar et al. [18] proposed a method that utilizes transfer learning with residual networks and initializes the ResNet-50 model with image weights using the RNGAP (global average pooling) model. This approach enables the model to perform three-level brain tumor image classification tasks. A similar approach can be seen in this study, Abubakar et al. [19] extracted deep features in fully connected layers of ResNet50 and VGG16. Then, combining these features with SVM also leads to good classification of burn depth. These studies [15,16,18,19] have shown that utilizing the ResNet-50 model can alleviate the gradient vanishing problem. Ref. [18] demonstrated that the global average pooling layer can effectively reduce overfitting, thereby improving classification performance. However, the network models used in these studies were either classified after the fully connected softmax layer [15,18], or only features extracted from the final fully connected layer were used for classification [16,19,20,21]. This suggests that features can also be extracted from the convolutional layers, pooling layers, or fully connected layers of the network and classified using other classifiers to achieve transfer learning. Although the classification part can utilize artificial neural networks composed of fully connected layers, softmax layers, and output layers, its efficiency still cannot surpass certain specific classifiers in some cases. Accordingly, studies [22,23,24] indicated that features extracted from layers other than the fully connected layer exhibit good generalization capabilities. Ref. [25] also mentioned that features extracted from the fully connected layer have inferior portability. Therefore, features are considered to be extracted from the global average pooling layer of the modified ResNet-50 network to obtain depth features, which may enhance the generalization capabilities of the features and partially alleviate overfitting.
Compared to the methods mentioned above, the authors in [26,27,28] attempted to enhance intelligent diagnostic performance by utilizing multiple features. The authors of [28] selected color features, texture features, and deep features extracted by stacked sparse autoencoders, combined the three extracted features, and input them into multiple classifiers. Ultimately, the accuracy of dividing burn injury images into two categories and three categories was 85.86% and 76.87%. Although, this method is effective, there may be redundant information among the extracted features. However, in Ref. [26], the authors reduced the dimension of the features by PCA after extracting the deep features. Then, they used the SVM classifier for face classification; the classification accuracy on face dataset reached 97.19%. Similarly, in Ref. [27], the authors used the univariate measurement and Poisson distribution feature selection approach after extracting and fusing deep features. Finally, the selected features were input into the Multi-class Support Vector Machine (MC-SVM) for classification of skin diseases. The experimental results show that this method achieved higher performance visually and in terms of enhanced quantitative evaluation with enhanced accuracy.
Therefore, directly combining multiple features is not the desired choice. The studies mentioned above proved that proper feature selection methods can improve classification performance. Additionally, there may still be room for improvement in the extracted deep features. Taking these points into consideration, the aim is to design a novel method for extracting depth features. A range of diverse features are then integrated and feature selection techniques are employed to identify useful ones, thereby enhancing the performance of the method. In Table 1, a summary table of the work related to burn depth diagnosis covered in this section is presented to facilitate comparison with this proposed approach. It should be noted that the advantages and disadvantages listed in the table are compared to those of other research in Table 1.

3. Method

In this section, the methods for extracting color features, texture features, and depth features from burn images specifically are discussed, as well as the method of feature selection after feature extraction to reduce dimensionality for burn depth classification.

3.1. Color Feature

Typically, medical experts often analyze the depth of burns based on color characteristics such as pink-white, beige, or dark brown in the burn area, or its texture features [29]. Different depths of burns exhibit different color appearances, with superficial dermal burn appearing as red and possibly moist, deep burns showing a mixture of red or waxy white, and full-thickness burns presenting skin colors of brown or red [3]. Therefore, theoretically, color can serve as an important basis for classifying burn depth in burn images.
The color histogram method is utilized to extract color features from images. Color histogram is widely used in various image retrieval tasks [30]. Essentially, this method calculates the number of occurrences of color intensities in the entire image or specific regions of interest. A histogram is generated for each color channel of the color model. Then, the histogram is quantized, as each entry in the histogram can be represented by a range of intensity values, and this process is known as binning. Further details on the color histogram can be explored in Refs. [31,32].
RGB (red, green, blue), HSV (hue, saturation, value), and L*A*B* (lightness, a* chromaticity layer, b* chromaticity layer) are three commonly used color spaces [33]. Generally, the RGB color space is easy to understand and use but may not be suitable for describing color attributes such as brightness and saturation. The HSV color space is used for tasks such as color recognition and segmentation in image processing [33]. It provides a more intuitive and natural way to describe color attributes, which makes it suitable for color-related operations in image processing. The L*A*B* color space is closer to the human visual system, offering more precise color description and control and performing well in image segmentation tasks [34]. Thus, different color spaces can express varying image information.
Here, images are considered to be converted into these three color spaces. Next, color histograms are used to count the occurrences of each color channel in each space and then merge them into color features. Figure 3a shows a superficial dermal burn image, while Figure 3b represents the RGB color space histogram as shown below. The histogram in Figure 3b is generated using the Color Thresholder tool with the bin of the histograms for each channel set to 256 [35].
In the RGB color space of a color image, the color histogram is represented by the k-level intensity values for the R, G, and B channels [28]. Specifically, the color histogram function for the R channel is defined as follows:
H R = H R 1 ,   H R 2 ,   . . .   H R k
H R represents all pixel values of the R channel’s k-level intensity value [31]. Similarly, in the HSV color space, the color histogram is represented by the k-level values for the H, S, and V channels. In the L*A*B* color space, the color histogram is represented by k-level values for the L*, a*, and b* channels. These color histograms can be concatenated to form a vector-representing image. Ultimately, the color features of each image can be defined as follows:
C o l o r i = [   H R 1 ,   . . .   H R k ,   H G 1 ,   . . .   H G k ,   H B 1 ,   . . .   H B k , H H 1 ,   . . .   H H k ,   H S 1 ,   . . .   H S k ,   H V 1 ,   . . .   H V k , H L 1 ,   . . .   H L k ,   H a 1 ,   . . .   H a k ,   H b 1 ,   . . .   H b k   ]
C o l o r i represents the color feature of the ith sample. H R k , H G k , H B k are the k-level vectors of the R, G and B channels, respectively, H H k , H S k , H V k , H L k , H a k and H b k are produced based on a similar approach.

3.2. Texture Feature

Skin with varying depths of burns exhibits different texture characteristics. In addition to assessing the color of the burned area, medical experts also consider texture features when making diagnoses, such as determining whether the affected area is moist [12].
The Gabor filter has garnered significant attention and has become a prominent method for extracting texture features due to its ability to optimally localize in the spatial domain and the frequency domain. Gabor filter feature extractor comprises a filter bank with filters of varying frequencies and orientations. These feature extractors can transform images from their original space into a feature space. Therefore, the texture features of each image are represented by unique features in the feature space [36].
A key advantage of the Gabor filter is its capacity to capture image details at various frequencies (scales) and orientations [37], offering the best spatial and frequency resolution combination [38]. Currently, the Gabor filter is extensively utilized in diverse fields, including diabetes detection [39], Alzheimer’s disease monitoring [37], and face recognition [40]. Therefore, the Gabor filter is chosen for extracting texture from burn images. For each pixel I(i,j) in each image i, the kernel of the Gabor filter can be defined using the following formula:
G a b o r ( i , j ) = e x p ( i c o s θ + j s i n θ ) 2 + η 2 ( i s i n θ + j c o s θ ) 2 2 σ 2 · c o s ( 2 π i c o s θ + j s i n θ λ + ξ )
The wavelength λ and direction θ serve as the primary parameters of Gabor filter. λ denotes the filter’s periodicity in the spatial domain, while θ dictates the filter’s orientation for detecting edges and textures in the image. ξ signifies the offset, σ denotes the standard deviation, and η represents the ratio of dimension to diameter. Drawing from insights glean from prior studies [28], eight directions and five wavelengths are employed in the used Gabor filter. The eight directions include the following:
θ = {   0 ° ,   22.5 ° ,   . . . ,   157.5 °   }  
These five wavelengths can be expressed as follows:
λ = {   2 k 2   | k = 1 ,   2 ,   ,   5   }
A kernel size of 39 × 39 is utilized in this work. Subsequently, the Gabor kernel filters each burn image to obtain texture features.

3.3. Depth Feature

Inspired by the success of deep learning applications, deep learning is used to extract features and classify burn depth. Unlike traditional manual methods, deep learning can automatically extract useful information, rendering it more adaptable to diverse scenarios. Typically, the shallower layers of a deep network capture basic features like texture and edge information, while deeper layers construct intricate feature hierarchies that exceed human visual perception [41].
Deep Convolutional Neural Networks (CNNs) possess the ability to autonomously learn hierarchical features from low to high levels. Theoretically, networks with more layers can tackle increasingly complex tasks, albeit accompanied by challenges, such as a decrease in saturation or accuracy, or vanishing and exploding gradients. ResNet-50 alleviates both problems by using deep residual pretrained architecture [42]. The ResNet-50 architecture is trained on the ImageNet dataset, which comprises approximately 1.5 million natural scene images. It reduces the computing cost by using a smaller dataset for training and achieves a better accuracy [43].
The ResNet-50 network can effectively extract a large number of complex features from images, and its advantages have been verified in numerous studies [15,16,42]. Furthermore, some studies have demonstrated that features extracted from layers other than the fully connected layer possess good generalization ability [25,44]. Therefore, inspired by these results, an attempt is made to extract features from the global average pooling layer.
Here, burn images are subjected to simple transformations (simple transformation includes random rotation of the images by 0°–180°, random scaling by 0 to 10 times, and the size of the images being adjusted to 224 × 224 × 3). Then, these images are validated ten times by five basic CNN models.
Burn images are categorized into three classes: superficial dermal burn, deep dermal burn, and full-thickness burn. Additionally, these burn images can also be presented as two classes, i.e., self-healing burn (superficial dermal burn) and non-self-healing burn (deep dermal burn and full-thickness burn). The mean classification accuracies are then presented for fairly comparing. Table 2 shows the accuracies of using different CNN models for classifying two or three burn classes.
Considering both the classification performance and the model type, as seen in Table 2, DenseNet-19 achieves the highest accuracy for burns with 2 classes, with ResNet-50 closely following by a margin of about 1%. However, ResNet-50 outperforms the others, showcasing a superiority of around 1.7% over DenseNet-19 when the burn depth consisting of 3 classes. Interestingly, despite having the most extensive network layers, ResNet-101 did not exhibit the best performance in this scenario. These basic CNN networks contain several layers for classification with the same function (including the full connection layer, SoftMax layer, and classification layer). Therefore, the reason for the varying accuracy rates may be the different features of the burn images that can be extracted. To improve the network, two perspectives are considered. One is to change the internal structure of the network in order to extract more comprehensive feature information from images or using alternative classification methods. In the following part, this proposed method for extracting deep features is introduced from the perspective of changing the internal network structure. As for the other perspective, detailed explanations are provided in the experimental section.
To ensure a more comprehensive extraction of depth features, a novel CNN model, named Burn-ResNet51, is introduced to classify burn images based on burn depth. Burn-ResNet51 employs transfer learning and adjustment of parameters of the original network to extract profound features essential for accurate classification. In particular, ResNet-50, as the backbone of the proposed network, comprises 49 convolutional layers. It uses the residual structure to effectively mitigates issues such as gradient vanishing and explosion [45]. However, extracting features directly from the final fully connected layer may lead to excessively large feature dimensions, resulting in reduced classification performance and an increased risk of overfitting.
To enhance performance, Burn-ResNet51 has modified the ResNet-50 architecture. When dividing burn depth into two or three categories, the structure of Burn-ResNet51 remains unchanged, while the parameters of some layers are adjusted. The following section elaborates on how the structure of Burn-ResNet51 is derived.
Initially, the last three layers of the pre-trained ResNet-50 architecture (the fully connected layer, softmax layer, and classification layer) are adapted to suit this task. Specifically, the original fully connected layer in the pre-trained network is replaced with a new fully connected layer that categorizes burn images into three or two classes. Subsequently, three additional layers (‘Conv’, ‘Batch_Normaliz’, and ‘Activation_Relu’) are inserted between the ‘activation_49_relu’ and ‘avg_pool’ layers of the ResNet-50 architecture, as shown in Figure 4b. The parameters of ‘Conv’ and ‘Batch_Normaliz’ are fine-tuned to automatically extract image features. After these adjustments, the ‘avg_pool’ layer is followed by the last three newly added layers (‘full_connected’, ‘softmax’, and ‘ClassificationLayer’), as previously mentioned [44]. Figure 4 provides a visual representation of both the basic network structure and the modified network; the red box and red arrow in Figure 4a need to be modified, and the blue boxes in Figure 4b were modified.
Figure 4a illustrates the original ResNet-50 architecture, while Figure 4b showcases the modified Burn-ResNet51 architecture after incorporating new layers. In this section, regarding the three newly introduced layers, the number of filters in the convolutional layer and the ε in the normalization layer are primarily fine-tuned. The rationale behind adjusting these parameters is elaborated on below.
Studies [46,47] have demonstrated that the kernel size and the number of filters in the convolutional layer significantly impact classification accuracy. Therefore, the kernel size of the new added convolutional layers in Burn-ResNet51 is set to 3 × 3, which is the same as the kernel size of the new convolutional layer in Ref. [44] with a similar network structure to the one in this work. Meanwhile, for the number of filters, [48] is taken into account, which suggests that the feature maps extracted from the final convolutional layer of a CNN can be transformed into feature vectors. This statement can be interpreted as follows: each convolutional filter contributes to a specific component of the feature vectors, with the dimensionality of the extracted feature vectors being directly linked to the number of convolutional filters utilized, and the number of feature maps dictating the total count of extracted feature vectors.
Especially in this work, where the dataset is relatively small and the feature dimensions of depth features are not very large, the dimensionality of features and the corresponding number of filters appear particularly important. As a result, the influence of four different filter quantities (200, 300, 400) on classification accuracy is explored.
Moreover, according to the definition of batch normalization in Ref. [49], a mini-batch Ɓ of size m is taken into account, which signifies the presence of m activation values within the mini-batch. Ɓ is represented as follows:
Ɓ =   x 1 ,   x 2 ,   . . .   x m  
If i is any value between 1 and m, the x ^ i is obtained by the normalization of xi, which can be expressed as follows:
x ^ i = x i μ Ɓ σ Ɓ 2 + ε
where μ Ɓ and σ Ɓ represent the mean and variance of   x 1 to x m in each mini-batch, respectively. ε is a constant added to the mini-batch variance for numerical stability.
To make sure that the transformation inserted in the network can represent the identity transform, y i is used to represent the linear change in x ^ i . The Batch Normalizing Transform is expressed as in the following formula:
y i = B N γ , β x i = γ x ^ i + β
To sum up, within this Batch Normalizing Transform algorithm, ε serves as a constant added to the small batch variance to uphold numerical stability [49]. Consequently, the focus is on exploring the impact of various ε values (1 × 10³, 1 × 10−4, 1 × 10−5) on classification accuracy during parameter selection.
With the above parameters set, the burn images are simply transformed (simple transformation includes random rotation of the images by 0°–180°, random scaling by 0 to 10 times, and the size of the images being adjusted to 224 × 224 × 3); then, the new images are inputted into the burn-Resnet51 network. Subsequently, features are extracted from the ‘avg_pool’ layer as the depth features of the burn image.

3.4. Feature Selection

The initial feature extraction process generates a large number of features. To enhance model performance through the selection of more effective features, a random-forest-based approach is employed to assess feature importance and facilitate selection. This method is commonly utilized in variable selection scenarios involving dense features [50,51,52,53]. Random forest stands out as an efficient technique requiring minimal parameter configuration. It excels in handling multi-class classification and multivariate regression tasks and has found successful applications across various computer vision domains [52], including face recognition [53], action recognition [50], and facial expression recognition [51].
Random forest comprises an ensemble of decision trees built on multiple bootstrap samples. Bootstrapping is a random sampling method with replacement from the training dataset. Additionally, a subset of features is randomly chosen from the entire sample to construct each node of the tree [54].
A key aspect of the random forest feature selection method is the utilization of Out-Of-Bag (OOB) error estimates. The Out-Of-Bag set represents a sample not utilized during tree training. Initially, random forest estimates the OOB error ( e r r ( X j ) ) for each feature, followed by randomly substituting the feature values with those from the OOB sample set. Subsequently, the OOB error is recalculated for the altered feature value ( e r r ( X o o b j ) ) [52]. The permutation importance measure, a widely adopted feature importance metric, is leveraged. Feature importance ( V I ) is determined by the increase in average error when a feature ( X j ) randomly alters one of its values from the OOB set ( X o o b j ). The number of trees in the forest is expressed as nb_tree, and feature importance ( V I ) is as follows [54]:
V I ( X j ) = 1 n b _ t r e e j ( | e r r ( X j ) e r r ( X o o b j ) | )
The random forest method is employed for feature selection, and the specific operational process is illustrated in Figure 5. Initially, a random forest model with m decision trees is constructed with a specified training set. Subsequently, the feature importance measure ( V I ) for the selected feature set is computed in each iteration using the random forest model prepared based on the Out-Of-Bag (OOB) set selected during training. The features are then arranged in descending order of V I , and the top n features are selected. Then, the corresponding n features are selected out from the test set. Ultimately, various classifiers, e.g., decision tree, KNN, etc., are utilized for classification, and the classification outcome is determined. Further details regarding the specific parameter configurations involved in the feature selection process will be discussed in subsequent experiments.
From this section, a 900-dimensional color feature vector is extracted for the three color spaces using color histograms. Subsequently, a 40-dimensional texture feature vector is extracted using Gabor filters, along with a 300-dimensional depth feature vector obtained from the Burn-ResNet51 model. Therefore, a total of 1240-dimensional feature vector is entered into the feature selector. The next section presents the fine-tuning and selection of some of the remaining parameters.

4. Experiment

This section covers the experimental aspects. First, the source and content of the dataset used in this experiment are introduced. Next, the process of selecting the optimal number of filters in the convolutional layer and the ε value in the normalization layer during deep feature extraction are discussed. The method of selecting the optimal number of selected features and the number of trees in nb_tree during feature selection are elaborated based on experiments. Finally, different combinations of three features are classified by random forest classifier. At the same time, some other common classifiers are also used for comparison and evaluation.
To evaluate the performance of the proposed method, accuracy, precision, recall, and F1 score are used as metrics to evaluate the classification performance, and are defined as follows. TP, FP, TN and FN represent true positive, false positive, true negative and false negative errors, respectively [28].
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 s c o r e = 2 T P 2 T P + F P + F N
During experimentation with diverse classifiers, fine-tuning is conducted to achieve optimal outcomes. Furthermore, each group of experiments is randomly repeated 20 times with the various aforementioned settings; then, their mean classification accuracies are reported for fair comparison. The experiments are performed on a PC with Intel(R) i7-4770 (2.90 GHz) CPU, NVIDIA GeForce RTX 2060 and 16GB RAM. The depth feature extraction, feature selection, and classification parts of the program are available at https://github.com/windgoddy/zxz/tree/main (accessed on 22 August 2024).

4.1. Dataset Setting

In this experiment, the Burns BIP_US dataset from [9] is chosen, which comprises 94 standard burn images. Following the dataset configuration in previous studies [28], twenty burn images are used for training, including 9 superficial dermal burns, 5 deep dermal burns, and 6 full-thickness burns. The remaining 74 images are designated the test set, consisting of 33 superficial dermal burns, 20 deep dermal burns, and 21 full-thickness burns. Table 3 provides a visual representation of the dataset’s configuration.

4.2. Depth Feature Parameter Setting

In order to select the optimal number of filters (NumFilters) and the ε value, a total of 12 combinations of NumFilters and ε are considered. Then, different combinations of these parameters are used to implement the proposed method.
In this section, the following strategy is used to find the optimal depth feature parameters for this proposed method. First, each parameter combination is used to extract deep features, which are then combined with pre-prepared color and texture features. Subsequently, feature selection is performed on the training dataset using parameters that will be discussed in the next section as optimal, and the corresponding features for the test dataset are selected. Finally, classification is carried out using a random forest classifier, with the classification results evaluated based on accuracy. Note that, in addition to the two parameters, other parts of the proposed Burn-ResNet51 model for extracting depth features have been adjusted and parameterized according to the image category. Table 4 shows the 12 combinations of NumFilters and ε, along with the corresponding classification results.
From this table, it can be observed that when ε is 0.0001 and NumFilters is 300, the highest accuracy of 91.76% is achieved when classifying burn images into two categories. Similarly, when ε is 0.001 and NumFilters is also 300, the highest accuracy of 80.74% is obtained when classifying burn images into three categories. This indicates that combining 300 extracted features with different values of ε can lead to optimal classification results.
Furthermore, it is evident from these combinations that these changes in the number of filters or ε value have a varying degree of impact on the accuracy of the classification. It follows that selecting appropriate values for NumFilters and ε is crucial for the extraction of deep features. Therefore, 0.0001 is used for ε and 300 for NumFilters with two classes of burn depth. Thus, 0.001 is used for ε and 300 for NumFilters with three classes of burn depth.

4.3. Feature Selection Parameter Settings

The random forest feature selection method is sensitive to the number of selected features (NumFeatures) and the number of trees (NumTrees) in the forest model [54]. These two parameters are used to adjust the performance of the model and improve the learning ability. Therefore, the following strategy is employed to determine the optimal parameters in this section. First, the depth features are extracted using the parameter combinations selected in the previous section, and then they are combined with color and texture features.
Subsequently, the NumTrees in the forest is fixed at 50 when the burn image is divided into two classes and at 600 when divided into three classes. At the same time, NumFeatures for candidate selection is defined as (200, 300… 1000, 1248 (no feature selection)). Different combinations of NumTrees and NumFeatures selected from the candidate range are used to implement this proposed method. Finally, the random forest classifier is used for classification, and the results for different NumFeatures are shown in Table 5.
Similarly, according to the NumFeatures determined by the above experiment, when the burn image is divided into two classes, NumFeatures is fixed to 700, and when the burn image is divided into three classes, NumFeatures is fixed to 500. The number of candidate trees is also defined as (10, 50, 100, 200, 300… 700). The proposed method is performed by using different combinations of NumFeatures and the parameter NumTrees selected from the candidate range. Finally, the random forest classifier is used for classification, and the classification results for different NumTrees are shown in Table 6.
From the two tables above, it can be observed that for binary image classification, the random forest feature selection method achieves optimal performance when selecting 700 features and using 50 trees. The optimal performance is achieved with 500 features and 600 trees when classifying burn images into three categories. Therefore, 700 is used for NumFeatures and 50 for NumTrees with two classes of burn depth. Furthermore, 500 is used for NumFeatures and 600 for NumTrees with three classes of burn depth.
These results show that the selected NumFeatures and the NumTrees in the forest have a certain impact on classification performance. Especially in the case of multiple classification, where parameter variations affect performance, whereas the impact is less significant for binary classification. It is observed that, in the case of binary classification, changes in the NumFeatures and NumTrees parameters do not have a significant impact on the experimental results. However, these parameters are still used. This is not only due to its improvement in accuracy by about 1%, but also its reduction in feature dimension and computational cost.
It should be noted that the two features selected in this section will only be used in the experiment where all three features are employed (the experiment in Section 4.4 using a random forest as a classifier). For other classification experiments, the parameters described in this section are not used. Instead, a similar approach is used to determine the specific parameters applicable to different scenarios.

4.4. Experiments after Combining the Three Features

After determining the necessary parameters, in order to more accurately evaluate the performance of the proposed method and further explore the relationship between these features and classification performance, several experiments are designed to investigate these issues. Following the classification process, Accuracy, Precision, Recall, and F1-score are applied as evaluation metrics for classification performance. The subsequent part provides detailed descriptions of these experiments and analyzes the experimental results.

4.4.1. Experimental Result

First, three different sets of features are extracted and concatenated. Then, feature selection is performed on the training dataset, and the corresponding features for the test dataset are selected. A random forest classifier was used to classify the depth of burn images. In addition, the classification results are compared with those of other common classifiers: support vector machine optimized with particle swarm optimization (PSO-SVM) [55], decision tree, k-nearest neighbors (KNN) [2], long short-term memory (LSTM), and 1-dimension convolutional neural network (1D-CNN).
In order to strike a balance between model complexity, generalization capability, and computational efficiency, the key parameter values of each classifier were tuned within the specified range according to experience in such a way that the highest average of recognition accuracy was obtained. For the random forest classifier, the number of trees is chosen as the main parameter to be adjusted. The process begins with 50 trees, gradually increasing the number. The number of trees is then adjusted based on classification performance for different parameters until a satisfactory result is achieved. Ultimately, the number of trees in the random forest classifier is fixed at 300. Similar adjustments are made for other classifiers to attain better outcomes.
The final classification results for dividing burn depth into two categories using three features are presented in Table 7, while the results for dividing burn depth into three categories are shown in Table 7. To visually compare the relationship between the results of each classifier, the contents of Table 7 and Table 8 are plotted with bar graphs (as shown in Figure 6 and Figure 7).
From Table 7, it can be seen that compared to other classifiers, the random forest classifier performs better in classifying the concatenation of three features into two categories (Accuracy 91.76%, Precision 91.62%, Recall 91.98%, F1-score 91.70%). Specifically, the detection accuracy of self-healing burns using random forest is 94.09%, while the detection accuracy of non-self-healing burns is 89.88%. From Table 8, the random forest classifier outperforms other classifiers in classifying the concatenation of three features into three categories (Accuracy 80.74%, Precision 77.77%, Recall 77.99%, F1-score 77.64%). In particular, the accuracy of superficial dermal burn using random forest is 96.21%, that of deep dermal burn is 67.04%, and that of full-thickness burn is 70.71%.
To further evaluate the performance of this proposed method, the images of the original dataset are rearranged evenly, and 20, 30, 40, 50, and 60 images are set for training, while the remaining images are reserved for testing. The accuracy rates obtained by random forest classification after dividing the data into three classes and two classes are shown in Figure 8. It can be seen from the chart that by changing the size of the training set, the performance of the model gradually improves as the amount of data increases, which indicates that this proposed model has good generalization ability.

4.4.2. Analysis of Experimental Results

Three effective burn depth classification methods selected from the relevant literature are compared with this proposed method, as shown in Table 9. Meanwhile, to provide a more intuitive comparison of the advantages of this proposed method, a bar chart (Figure 9) is plotted based on the results of dividing the burn images into two categories, as shown in Table 9.
Table 9 shows a detailed comparison of the best performance of these methods and that of this proposed method, and it can be seen that the results have different degrees of improvement in various aspects. Especially when classifying burn images into three categories, the accuracy of this proposed method is obviously higher than that of other methods.
To assess whether the proposed method has statistical significance, a paired t-test with two tails [56] is conducted on the accuracy obtained with different classifiers. As shown in Table 10, the p-values of all these classifiers are all lower than 0.05, indicating that the proposed method has statistically significant differences in terms of accuracy compared to the other approaches mentioned. Although the accuracy of using PSO-SVM to classify images into two classes is 90.95%, which is close to the best classification result, its p-value is approximately 0.02. Statistically, the performance of the random forest classifier is better than that of PSO-SVM. Therefore, it can be concluded that the proposed method outperforms the other methods presented in terms of performance.
Following many experiments, 12 representative burn images are selected from the classification results of each classifier as a group, resulting in five groups for the five classifiers, as shown in Figure 10. The images framed in red boxes are images that cannot be correctly classified in either the two-class or three-class cases. The images framed in blue boxes can only be correctly divided into two categories. The images framed in green boxes can be correctly classified in both cases. To make it easier to explain, there are circular markers in the bottom right corner of some images, and circles of the same color indicate the same image.
It can be observed from Figure 10 that the red circles mark images which belong to deep dermal burn; these cannot be correctly classified by any classifier. The yellow circles mark images that can be correctly classified by the PSO-SVM classifier and decision tree classifier. The green circles mark images that can be correctly classified by the random forest classifier, LSTM classifier, and 1D-CNN classifier. The image marked with the blue circle can only be correctly classified by the KNN classifier. Furthermore, it is found that the random forest classifier and 1D-CNN classifier may achieve better classification results for images with blisters.

4.5. Ablation Study

Combining the findings from Table 7 and Table 8, it is evident that the random forest classifier demonstrates the most favorable performance, followed by PSO-SVM. These results indicate that the proposed approach is effective in classifying burn depth in burn images. However, it is not conclusively established that all three types of features contribute positively to classification. Therefore, a series of ablation experiments are conducted, involving the isolation of each feature for individual classification using six classifiers. Subsequently, pairwise combinations of these features are tested with the random forest classifier to verify that each feature contributes to the classification.
Table 11 and Table 12 present the results of categorizing depth features into two classes and three classes using classifiers after extracting them individually. It is evident that PSO-SVM achieves the best performance, with an accuracy of 90.54% for two classes and 76.08% for three classes. Random forest, 1D-CNN, and LSTM show comparable performance, indicating that the depth features extracted using Burn-ResNet51 network possess good discriminative ability to aid classifiers in distinguishing between different categories. The satisfactory performance of these classifiers may also suggest a certain level of universality in the depth features. Thus, the extraction of depth features using Burn-ResNet51 network can be considered effective and provides a solid foundation for further research. However, from another perspective, when categorized into two classes, the individual depth features already achieve an accuracy of 89.19% when classified with the random forest, indicating that the influence of color and texture features on performance is limited.
When categorizing color features into two classes and three classes using classifiers after extracting them individually, it can be clearly seen from Table 13 and Table 14 that individual color features are less effective than depth features in conveying comprehensive information about burn images. This may be due to the fact that color features are more susceptible to noise, illumination changes, and other issues. When the images are categorized into three classes, random forest and decision tree classifiers exhibit superior classification performance compared to other classifiers. However, PSO-SVM still performs better in the two-class scenario.
Table 15 and Table 16 present the results of categorizing texture features into two classes and three classes using classifiers after extracting them individually. It is evident that the classification performance of texture features alone is not ideal. However, valuable information can still be derived. For instance, in the two-class scenario, random forest, PSO-SVM, LSTM, and 1D-CNN classifiers are more sensitive to the second class of images, which may partially compensate for the lower sensitivity of color features to the second class. This observation may also indicate that texture features can extract image information from a different perspective than color features. However, the classification results of texture features are not as good in the three-class scenario.
After conducting experiments separately on each feature, it can be concluded that depth features can more accurately represent burn image information compared to color and texture features. However, it is still not very effective to prove that all three features have a positive effect on classification. Therefore, the three types of features are combined pairwise for analysis and verification.
Table 17 and Table 18 display the results of combining color features with texture features, color features with depth features, and texture features with depth features, and classifying them using the random forest classifier.
To provide a more intuitive comparison of the contributions made by different features in classification, all the classification results obtained using random forest in this section are compiled into a bar chart (as shown in Figure 11). It can be observed that despite the inferior classification performance when classifying texture features individually, they can still play a role when combined with other features, especially in the three-class scenario, as demonstrated in the previous experiments.
From all the experiments above, it can be seen that the proposed method is effective and feasible for classifying burn images of different depths, demonstrating that each feature contributes to the classification.

5. Discussion

A method is proposed to extract multiple features from burn images and perform feature selection. In particular, Burn-ResNet51 network is used for depth feature extraction and the random forest feature selection method is used for feature selection, enhancing the classification performance. Finally, utilizing random forest to classify burn images, the results show that the accuracy of binary classification is 91.76%, and the accuracy of three-class classification is 80.74%.
When extracting color features, color histograms are extracted from the RGB, HSV, and L*A*B* color spaces of the images as their respective color features. The advantage of this approach lies in that each color space has its unique color representation method, which can provide more color information. At the same time, color histograms can display the distribution of various colors in the image, which is helpful for image identification and classification. Currently, most studies consider the commonly used color models, namely RGB, HSV, and L*A*B* [20,25,31]. However, in some research, other color models also perform well. For example, HSI performs well in image enhancement [57], and YCbCr performs well in skin color segmentation [58]. Since the color of burned skin also contains some skin color characteristics, the YCbCr color space may also perform well. Therefore, it may be possible to extract more accurate color features by comparing various color spaces in different color extraction methods (color histograms, color correlograms, color co-occurrence matrices) [59].
Similarly, the Gabor filter used for extracting texture features possesses the drawback of non-orthogonality, resulting in redundant features at different scales. Most of the time, a single-filter resolution may not effectively capture the spatial structure in natural textures [60]. In order to compare with the Gabor filter, another geometric feature descriptor [61,62] is selected, namely the local binary pattern (LBP) [28], as a method for extracting features. However, after replacing the Gabor filter with LBP, three types of features were selected, achieving an accuracy rate of 90.28% for dividing burn depth into two categories and 79.84% for dividing it into three categories. Both of these experimental results are approximately 1% lower than those achieved by this proposed method. When only features extracted by LBP were selected, the accuracy rate for dividing burn depth into two categories was 48.92%, while the accuracy rate for dividing it into three categories was 35.37%. The experimental results are all lower than those of the original method; therefore, the Gabor filter is still chosen to extract features.
With the continuous development of deep learning, network structures are becoming increasingly intricate. Different tasks correspond to their own unique network structures and parameters, greatly enhancing the upper limit of deep learning. Therefore, future research should comprehensively consider features from fully connected layers, convolutional layers and various 1D-CNN architectures, and fine-tune parameters to optimize image classification.
In order to further evaluate the performance of the method, the total computing time is recorded (as shown in Table 19), which is the time taken to train the model on the burn injury images in the training set and then classify the images once. The total computing time includes training time and testing time, with records also kept for the computing time without feature selection. When the burn depth is divided into two classes, the computing time is 76.91 s. For three classes, the total computing time is 77.82 s. It was observed that when images are categorized into three categories, the computing time for the feature selection process increases significantly due to the excessive number of trees. However, at the same time, the corresponding accuracy also improves significantly (as shown in Table 6). Moreover, a random forest is used as a feature selection method, which is based on random sampling and is sensitive to noise or outliers, potentially affecting the accuracy of feature selection. Therefore, selecting an efficient and stable feature selection method is also an important direction for future research.
From the dataset selected in this work, the number of burn images is limited, and there is a lack of complex burn skin images (e.g., burn skin with skin diseases) for comparison analysis. This insufficiency could undermine the performance of the proposed approach when dealing with different depth regions of complex burn images or burn images with background interference. Therefore, it is necessary to collect more diverse burn samples to further validate the effectiveness of the proposed method [28].
In addition, this method is based on the analysis of burn images, which makes it difficult to completely avoid the influence of factors such as brightness, illumination angle, sharpness, background information. While the classification performance of this proposed method has been greatly improved compared with other relatively simple methods, it is still difficult to distinguish the correct category of the images similar to those marked by blue and red circles in Figure 10, which have a large color difference from other images in the same class. Therefore, perhaps combining the method proposed in this study with the incomplete multi-view clustering (IMC) method that restores views with missing information [63] or the method from Ref. [64] that solves the double incomplete learning problem of multi-view multi-label classification with missing views and missing labels would be very helpful for the classification of burn depth in burn images.

6. Conclusions

A novel method is introduced to assist medical experts in determining the depth of burn injuries through automated evaluation of burn images. This method utilizes color histograms extracted from three different color models as color features, and employs Gabor filters to extract texture features. Subsequently, features are extracted from the global average pooling layer of an enhanced Burn-ResNet51 network (a derived network of the ResNet-50) as depth features. These three types of features are then combined and selected using a feature selection method based on random forest. Finally, a classifier is used to classify the selected features.
A comprehensive series of experiments is conducted to highlight the critical role of deep learning in feature extraction and to validate the effectiveness of this proposed method. Compared with other existing burn depth classification techniques, this proposed method exhibits superior performance. In the future, more burn injury samples can be collected for testing, and it is also an interesting research direction to choose appropriate feature extraction methods and feature selection methods based on the characteristics of burn injury images.

Author Contributions

Conceptualization, X.Z.; methodology, X.Z. and Q.Z.; software, X.Z., Q.Z. and J.Y.; validation, P.L., J.S. and J.Z.; formal analysis, J.Y., J.S. and J.Z.; investigation, X.Z. and P.L.; data curation, Q.Z. and J.S.; writing—original draft preparation, X.Z.; writing—review and editing, Q.Z.; visualization, P.L. and J.Z.; supervision, Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Hainan Province Science and Technology Talents Innovation Project (KJRC2023D30).

Data Availability Statement

Data are available at [http://personal.us.es/rboloix/Burns_BIP_US_database.zip (accessed on 25 January 2024)].

Conflicts of Interest

The author Jie You was employed by the company Shenzhen Intellirocks Tech Co., Ltd. The remaining authors declare that this research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.

References

  1. Johnson, C. Management of burns. Surgery 2018, 36, 435–440. [Google Scholar] [CrossRef]
  2. Acha, B.; Serrano, C.; Fondón, I.; Gómez-Cía, T. Burn depth analysis using multidimensional scaling applied to psychophysical experiment data. IEEE Trans. Med. Imaging 2013, 32, 1111–1120. [Google Scholar] [CrossRef]
  3. Noorbakhsh, S.I.; Bonar, E.M.; Polinski, R.; Amin, M.S. Educational case: Burn injury—Pathophysiology, classification, and treatment. Acad. Pathol. 2021, 8, 23742895211057239. [Google Scholar] [CrossRef]
  4. Markiewicz-Gospodarek, A.; Kozioł, M.; Tobiasz, M.; Baj, J.; Radzikowska-Büchner, E.; Przekora, A. Burn wound healing: Clinical complications, medical care, treatment, and dressing types: The current state of knowledge for clinical practice. Int. J. Environ. Res. Public Health 2022, 19, 1338. [Google Scholar] [CrossRef]
  5. Chrapusta, A.; Pabiańczyk, R.; Nessler, M.; Nessler, K.; Cieślik, K. Aquacel Ag w leczeniu oparzeń IIb i IIb/III stopnia ręki u dorosłych–doświadczenia własne. In Forum Zakażeń; Evereth Publishing: Warszawa, Poland, 2012; pp. 71–76. [Google Scholar]
  6. Jaskille, A.D.; Shupp, J.W.; Jordan, M.H.; Jeng, J.C. Critical review of burn depth assessment techniques: Part I. Historical review. J. Burn Care Res. 2009, 30, 937–947. [Google Scholar] [CrossRef] [PubMed]
  7. Lee, S.; Lukan, J.; Boyko, T.; Zelenova, K.; Makled, B.; Parsey, C.; Norfleet, J.; De, S. A deep learning model for burn depth classification using ultrasound imaging. J. Mech. Behav. Biomed. Mater. 2022, 125, 104930. [Google Scholar] [CrossRef] [PubMed]
  8. Yakupu, A.; Zhang, J.; Dong, W.; Song, F.; Dong, J.; Lu, S. The epidemiological characteristic and trends of burns globally. BMC Public Health 2022, 22, 1596. [Google Scholar] [CrossRef] [PubMed]
  9. Yadav, D.; Sharma, A.; Singh, M.; Goyal, A. Feature extraction based machine learning for human burn diagnosis from burn images. IEEE J. Transl. Eng. Health Med. 2019, 7, 1–7. [Google Scholar] [CrossRef]
  10. Hoeksema, H.; Van de Sijpe, K.; Tondu, T.; Hamdi, M.; Van Landuyt, K.; Blondeel, P.; Monstrey, S. Accuracy of early burn depth assessment by laser Doppler imaging on different days post burn. Burns 2009, 35, 36–45. [Google Scholar] [CrossRef]
  11. Badea, M.-S.; Vertan, C.; Florea, C.; Florea, L.; Bădoiu, S. Automatic burn area identification in color images. In Proceedings of the 2016 International Conference on Communications (COMM), Bucharest, Romania, 9–10 June 2016; pp. 65–68. [Google Scholar]
  12. Serrano, C.; Boloix-Tortosa, R.; Gómez-Cía, T.; Acha, B. Features identification for automatic burn classification. Burns 2015, 41, 1883–1890. [Google Scholar] [CrossRef]
  13. Kuan, P.; Chua, S.; Safawi, E.; Wang, H.; Tiong, W. A comparative study of the classification of skin burn depth in human. J. Telecommun. Electron. Comput. Eng. JTEC 2017, 9, 15–23. [Google Scholar]
  14. Wantanajittikul, K.; Auephanwiriyakul, S.; Theera-Umpon, N.; Koanantakool, T. Automatic segmentation and degree identification in burn color images. In Proceedings of the 4th 2011 Biomedical Engineering International Conference, Chiang Mai, Thailand, 29–31 January 2012; pp. 169–173. [Google Scholar]
  15. Rostami, B.; Niezgoda, J.; Gopalakrishnan, S.; Yu, Z. Multiclass burn wound image classification using deep convolutional neural networks. arXiv 2021, arXiv:2103.01361. [Google Scholar]
  16. Chauhan, J.; Goyal, P. Deep learning based fully automatic efficient burn severity estimators for better burn diagnosis. In Proceedings of the 2020 International Joint Conference on Neural Networks (IJCNN), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar]
  17. Nalepa, J.; Bosowski, P.; Dudzik, W.; Kawulok, M. Fusing deep learning with support vector machines to detect COVID-19 in X-ray images. In Proceedings of the Asian Conference on Intelligent Information and Database Systems, Ho Chi Minh City, Vietnam, 28–30 November 2022; pp. 340–353. [Google Scholar]
  18. Kumar, R.L.; Kakarla, J.; Isunuri, B.V.; Singh, M. Multi-class brain tumor classification using residual network and global average pooling. Multimed. Tools Appl. 2021, 80, 13429–13438. [Google Scholar] [CrossRef]
  19. Abubakar, A.; Ugail, H.; Smith, K.M.; Bukar, A.M.; Elmahmudi, A. Burns depth assessment using deep learning features. J. Med. Biol. Eng. 2020, 40, 923–933. [Google Scholar] [CrossRef]
  20. Abubakar, A.; Ugail, H.; Bukar, A.M. Can machine learning be used to discriminate between burns and pressure ulcer? In Proceedings of the Intelligent Systems and Applications: Proceedings of the 2019 Intelligent Systems Conference (IntelliSys), London, UK, 5–6 September 2019; Volume 2, pp. 870–880. [Google Scholar]
  21. Abubakar, A.; Ugail, H.; Bukar, A.M. Noninvasive assessment and classification of human skin burns using images of Caucasian and African patients. J. Electron. Imaging 2020, 29, 041002. [Google Scholar] [CrossRef]
  22. Cimpoi, M.; Maji, S.; Vedaldi, A. Deep filter banks for texture recognition and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3828–3836. [Google Scholar]
  23. Cimpoi, M.; Maji, S.; Kokkinos, I.; Mohamed, S.; Vedaldi, A. Describing textures in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3606–3613. [Google Scholar]
  24. Napoletano, P. Hand-crafted vs. learned descriptors for color texture classification. In Proceedings of the Computational Color Imaging: 6th International Workshop, CCIW 2017, Milan, Italy, 29–31 March 2017; pp. 259–271. [Google Scholar]
  25. Barburiceanu, S.; Meza, S.; Orza, B.; Malutan, R.; Terebes, R. Convolutional neural networks for texture feature extraction. Applications to leaf disease classification in precision agriculture. IEEE Access 2021, 9, 160085–160103. [Google Scholar] [CrossRef]
  26. Lu, X.; Duan, X.; Mao, X.; Li, Y.; Zhang, X. Feature extraction and fusion using deep convolutional neural networks for face detection. Math. Probl. Eng. 2017, 2017, 1376726. [Google Scholar] [CrossRef]
  27. Maqsood, S.; Damaševičius, R. Multiclass skin lesion localization and classification using deep learning based features fusion and selection framework for smart healthcare. Neural Netw. 2023, 160, 238–258. [Google Scholar] [CrossRef]
  28. Zhang, B.; Zhou, J. Multi-feature representation for burn depth classification via burn images. Artif. Intell. Med. 2021, 118, 102128. [Google Scholar] [CrossRef]
  29. Acha, B.a.; Serrano, C.; Acha, J.I.; Roa, L.M. Segmentation and classification of burn images by color and texture information. J. Biomed. Opt. 2005, 10, 034014. [Google Scholar] [CrossRef]
  30. Xue, B.; Wanjun, L. Research of image retrieval based on color. In Proceedings of the 2009 International Forum on Computer Science-Technology and Applications, Chongqing, China, 25–27 December 2009; pp. 283–286. [Google Scholar]
  31. Rachmawati, E.; Khodra, M.L.; Supriana, I. Histogram based color pattern identification of multiclass fruit using feature selection. In Proceedings of the 2015 International Conference on Electrical Engineering and Informatics (ICEEI), Denpasar, Indonesia, 10–11 August 2015; pp. 43–48. [Google Scholar]
  32. Nazir, A.; Ashraf, R.; Hamdani, T.; Ali, N. Content based image retrieval system by using HSV color histogram, discrete wavelet transform and edge histogram descriptor. In Proceedings of the 2018 International Conference on Computing, Mathematics and Engineering Technologies (iCoMET), Sukkur, Pakistan, 3–4 March 2018; pp. 1–6. [Google Scholar]
  33. Bora, D.J.; Gupta, A.K.; Khan, F.A. Comparing the performance of L* A* B* and HSV color spaces with respect to color image segmentation. arXiv 2015, arXiv:1506.01472. [Google Scholar]
  34. Bora, D.J.; Gupta, A.K. A novel approach towards clustering based image segmentation. arXiv 2015, arXiv:1506.01710. [Google Scholar]
  35. Color Thresholder. Available online: https://www.mathworks.com/help/images/ref/colorthresholder-app.html?s_tid=srchtitle_site_search_1_Color%20Thresholder (accessed on 3 June 2024).
  36. Li, W.; Mao, K.; Zhang, H.; Chai, T. Selection of gabor filters for improved texture feature extraction. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, China, 26–29 September 2010; pp. 361–364. [Google Scholar]
  37. Keserwani, P.; Pammi, V.C.; Prakash, O.; Khare, A.; Jeon, M. Classification of Alzheimer disease using Gabor texture feature of hippocampus region. Int. J. Image Graph. Signal Process. 2016, 8, 13–20. [Google Scholar] [CrossRef]
  38. Bianconi, F.; Fernández, A. Evaluation of the effects of Gabor filter parameters on texture classification. Pattern Recognit. 2007, 40, 3325–3335. [Google Scholar] [CrossRef]
  39. Ting, S.; Zhang, B. Diabetes mellitus detection based on facial block texture features using the Gabor filter. In Proceedings of the 2014 IEEE 17th International Conference on Computational Science and Engineering, Chengdu, China, 19–21 December 2014; pp. 1–6. [Google Scholar]
  40. Yu, L.; He, Z.; Cao, Q. Gabor texture representation method for face recognition using the Gamma and generalized Gaussian models. Image Vis. Comput. 2010, 28, 177–187. [Google Scholar] [CrossRef]
  41. Li, S.; Song, W.; Fang, L.; Chen, Y.; Ghamisi, P.; Benediktsson, J.A. Deep learning for hyperspectral image classification: An overview. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6690–6709. [Google Scholar] [CrossRef]
  42. ElGhany, S.A.; Ibraheem, M.R.; Alruwaili, M.; Elmogy, M. Diagnosis of Various Skin Cancer Lesions Based on Fine-Tuned ResNet50 Deep Network. Comput. Mater. Contin. 2021, 68, 117–135. [Google Scholar] [CrossRef]
  43. Çınar, A.; Yıldırım, M.; Eroğlu, Y. Classification of pneumonia cell images using improved ResNet50 model. Trait. Signal 2021, 38, 165–173. [Google Scholar] [CrossRef]
  44. Elpeltagy, M.; Sallam, H. Automatic prediction of COVID− 19 from chest images using modified ResNet50. Multimed. Tools Appl. 2021, 80, 26451–26463. [Google Scholar] [CrossRef]
  45. Zhu, J.; Sun, L.; Wang, Y.; Subramani, S.; Peng, D.; Nicolas, S.C. A ResNet based multiscale feature extraction for classifying multi-variate medical time series. KSII Trans. Internet Inf. Syst. TIIS 2022, 16, 1431–1445. [Google Scholar]
  46. Ahmed, W.S. The impact of filter size and number of filters on classification accuracy in CNN. In Proceedings of the 2020 International Conference on Computer Science and Software Engineering (CSASE), Duhok, Iraq, 16–18 April 2020; pp. 88–93. [Google Scholar]
  47. Agrawal, A.; Mittal, N. Using CNN for facial expression recognition: A study of the effects of kernel size and number of filters on accuracy. Vis. Comput. 2020, 36, 405–412. [Google Scholar] [CrossRef]
  48. Passalis, N.; Tefas, A. Learning bag-of-features pooling for deep convolutional neural networks. In Proceedings of the Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5755–5763. [Google Scholar]
  49. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  50. Fanelli, G.; Yao, A.; Noel, P.-L.; Gall, J.; Van Gool, L. Hough forest-based facial expression recognition from video sequences. In Proceedings of the Trends and Topics in Computer Vision: ECCV 2010 Workshops, Heraklion, Crete, Greece, 10–11 September 2010; Revised Selected Papers, Part I 11; Springer: Berlin/Heidelberg, Germany, 2012; pp. 195–206. [Google Scholar]
  51. Yao, A.; Gall, J.; Van Gool, L. A hough transform-based voting framework for action recognition. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2061–2068. [Google Scholar]
  52. Gharsalli, S.; Emile, B.; Laurent, H.; Desquesnes, X.; Vivet, D. Random forest-based feature selection for emotion recognition. In Proceedings of the 2015 International Conference on Image Processing Theory, Tools and Applications (IPTA), Orleans, France, 10–13 November 2015; pp. 268–272. [Google Scholar]
  53. Zhao, X.; Kim, T.-K.; Luo, W. Unified face analysis by iterative multi-output random forests. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1765–1772. [Google Scholar]
  54. Gharsalli, S.; Emile, B.; Laurent, H.; Desquesnes, X. Feature selection for emotion recognition based on random forest. In Proceedings of the International Conference on Computer Vision Theory and Applications, Rome, Italy, 27–29 February 2016; pp. 610–617. [Google Scholar]
  55. Zhang, Y.; Xie, X.; Cheng, T. Application of PSO and SVM in image classification. In Proceedings of the 2010 3rd International Conference on Computer Science and Information Technology, Chengdu, China, 9–11 July 2010; pp. 629–631. [Google Scholar]
  56. Ruxton, G.D. The unequal variance t-test is an underused alternative to Student’s t-test and the Mann–Whitney U test. Behav. Ecol. 2006, 17, 688–690. [Google Scholar] [CrossRef]
  57. Asmare, M.H.; Asirvadam, V.S.; Iznita, L. Color space selection for color image enhancement applications. In Proceedings of the 2009 International Conference on Signal Acquisition and Processing, Kuala Lumpur, Malaysia, 3–5 April 2009; pp. 208–212. [Google Scholar]
  58. Kaur, A.; Kranthi, B. Comparison between YCbCr color space and CIELab color space for skin color segmentation. Int. J. Appl. Inf. Syst. 2012, 3, 30–33. [Google Scholar]
  59. Patel, J.M.; Gamit, N.C. A review on feature extraction techniques in content based image retrieval. In Proceedings of the 2016 International Conference on Wireless Communications, Signal Processing and Networking (WiSPNET), Chennai, India, 23–25 March 2016; pp. 2259–2263. [Google Scholar]
  60. Humeau-Heurtier, A. Texture feature extraction methods: A survey. IEEE Access 2019, 7, 8975–9000. [Google Scholar] [CrossRef]
  61. Rangayyan, R.M. Biomedical Image Analysis; CRC Press: Boca Raton, FL, USA, 2004. [Google Scholar]
  62. Farag, A.; Ali, A.; Graham, J.; Farag, A.; Elshazly, S.; Falk, R. Evaluation of geometric feature descriptors for detection and classification of lung nodules in low dose CT scans of the chest. In Proceedings of the 2011 IEEE International Symposium on Biomedical Imaging: From Nano to Macro, Chicago, IL, USA, 30 March–2 April 2011; pp. 169–172. [Google Scholar]
  63. Liu, C.; Wen, J.; Wu, Z.; Luo, X.; Huang, C.; Xu, Y. Information recovery-driven deep incomplete multiview clustering network. IEEE Trans. Neural Netw. Learn. Syst. 2023. early access. [Google Scholar] [CrossRef]
  64. Wen, J.; Liu, C.; Deng, S.; Liu, Y.; Fei, L.; Yan, K.; Xu, Y. Deep double incomplete multi-view multi-label learning with incomplete labels and missing views. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 11396–11408. [Google Scholar] [CrossRef]
Figure 1. These are typical images of three different burn depths. The images from left to right are (a) superficial dermal burn; (b) deep dermal burn; (c) full-thickness burn.
Figure 1. These are typical images of three different burn depths. The images from left to right are (a) superficial dermal burn; (b) deep dermal burn; (c) full-thickness burn.
Electronics 13 03665 g001
Figure 2. The schematic diagram of this proposed method.
Figure 2. The schematic diagram of this proposed method.
Electronics 13 03665 g002
Figure 3. (a) The typical sample of burn image; (b) the corresponding histogram of RGB color space of burn image (a).
Figure 3. (a) The typical sample of burn image; (b) the corresponding histogram of RGB color space of burn image (a).
Electronics 13 03665 g003
Figure 4. The architectures of (a) the original ResNet-50 model and (b) the Burn-ResNet51 model used in the proposed method.
Figure 4. The architectures of (a) the original ResNet-50 model and (b) the Burn-ResNet51 model used in the proposed method.
Electronics 13 03665 g004
Figure 5. Feature selection architecture based on the random forest.
Figure 5. Feature selection architecture based on the random forest.
Electronics 13 03665 g005
Figure 6. Comparison of the results of detecting two categories of burn depths with random forest, PSO-SVM, decision tree, KNN, LSTM, and 1D-CNN.
Figure 6. Comparison of the results of detecting two categories of burn depths with random forest, PSO-SVM, decision tree, KNN, LSTM, and 1D-CNN.
Electronics 13 03665 g006
Figure 7. Comparison of the results of detecting three categories of burn depths with random forest, PSO-SVM, decision tree, KNN, LSTM, and 1D-CNN.
Figure 7. Comparison of the results of detecting three categories of burn depths with random forest, PSO-SVM, decision tree, KNN, LSTM, and 1D-CNN.
Electronics 13 03665 g007
Figure 8. The accuracies of dividing the images into two categories and three categories after setting the number of training images from 20 to 60.
Figure 8. The accuracies of dividing the images into two categories and three categories after setting the number of training images from 20 to 60.
Electronics 13 03665 g008
Figure 9. Comparison of burn depth classified into two categories using the method in Refs. [9,15,28] and our proposed method.
Figure 9. Comparison of burn depth classified into two categories using the method in Refs. [9,15,28] and our proposed method.
Electronics 13 03665 g009
Figure 10. Twelve typical burn images from each classifier’s classification results are selected as a group, and each group of images is further divided into three subgroups based on whether they were correctly classified. The images of these five groups of classification results belong to these classifiers as follows: (a) the random forest classifier; (b) the PSO-SVM classifier; (c) the decision tree classifier; (d) the KNN classifier; (e) the LSTM classifier; (f) the 1D-CNN classifier. Some pictures have circular marks in the lower right corner, and the same colored circular marks indicate that they are the same picture.
Figure 10. Twelve typical burn images from each classifier’s classification results are selected as a group, and each group of images is further divided into three subgroups based on whether they were correctly classified. The images of these five groups of classification results belong to these classifiers as follows: (a) the random forest classifier; (b) the PSO-SVM classifier; (c) the decision tree classifier; (d) the KNN classifier; (e) the LSTM classifier; (f) the 1D-CNN classifier. Some pictures have circular marks in the lower right corner, and the same colored circular marks indicate that they are the same picture.
Electronics 13 03665 g010
Figure 11. Comparison of the accuracies using random forest for classification after selecting different features individually or different feature combinations.
Figure 11. Comparison of the accuracies using random forest for classification after selecting different features individually or different feature combinations.
Electronics 13 03665 g011
Table 1. Summary table of research related to burn depth diagnosis.
Table 1. Summary table of research related to burn depth diagnosis.
StudiesAdvantagesDisadvantagesPublished
Year
Hoeksema et al. [10]The accuracy of LDI is higher in burn depth assessment after the fifth day of burn injury.The assessment accuracy in the early stage of burn is lower.
High cost.
Difficult to calibrate.
2009
Serrano et al. [12]The accuracy of distinguishing burns that can heal themselves or not is higher. Vulnerable to factors affecting the color of images.2016
Kuan et al. [13]Focus on both color and texture
Low cost and velocity.
Low accuracy.2017
Wantanajittikul et al. [14]Focus on both color and texture.The accuracy is low when burns are classified into three categories.2012
Yadav et al. [9]The accuracy of distinguishing burns that can heal themselves or not is higher.
Accurately classify burns with blisters.
The accuracy is low when burns are classified into three categories.
Focus only on color.
2019
Rostami et al. [15]High accuracy.Require more computational resources.
Focus only on deep features.
2021
Zhang et al. [28]Focus on multiple features of burn images.
High accuracy.
Require more computational resources.2021
Abubakar et al. [19]High accuracy.Focus only on deep features.
Require more computational resources.
Program runs slower.
2020
Table 2. The results of using different CNN models for classifying two or three burn classes.
Table 2. The results of using different CNN models for classifying two or three burn classes.
CNN ModelAccuracy (Two Categories) (%)Accuracy (Three Categories) (%)
ResNet-1882.8468.47
ResNet-5085.6073.38
ResNet-10182.7770.95
VGG-1680.4171.37
DenseNet-1986.4971.62
Table 3. A statistical table of the sample quantities for different types of burns in the training and test sets.
Table 3. A statistical table of the sample quantities for different types of burns in the training and test sets.
Burn Depth TypeTraining SetTesting Set
Three CategoriesTwo CategoriesThree CategoriesTwo CategoriesThree CategoriesTwo Categories
Superficial dermal burnCapable of self-healing993333
Deep dermal burnIncapable of self-healing 5112041
Full-thickness burn621
Total2074
Table 4. Classification accuracy with different combinations of NumFilters and ɛ.
Table 4. Classification accuracy with different combinations of NumFilters and ɛ.
ɛNumFiltersTwo CategoriesThree Categories
Accuracy (%)Accuracy (%)
0.00120087.7778.42
0.000189.6776.49
0.0000187.7774.26
0.00130088.3880.74
0.000191.7675.68
0.0000190.6880.55
0.00140090.8278.85
0.000189.3277.10
0.0000185.1465.43
Table 5. Classification accuracy under different number of features.
Table 5. Classification accuracy under different number of features.
NumFeaturesTwo CategoriesThree Categories
Accuracy (%)Accuracy (%)
20091.2178.24
30090.2779.73
40090.2780.14
50090.8880.74
60091.1580.54
70091.7677.03
80091.0176.49
90090.7476.95
100090.8177.03
1248
(no feature selection)
90.8878.39
Table 6. Classification accuracy under different number of trees.
Table 6. Classification accuracy under different number of trees.
NumTreesTwo CategoriesThree Categories
Accuracy (%)Accuracy (%)
1090.3564.87
5091.7674.32
10091.1574.32
20090.5878.38
30090.5579.73
40090.8879.73
50091.2879.73
60091.3580.74
70090.8879.19
Table 7. The classification results of different classifiers when three features are selected and for categorizing burn images into two burn classes.
Table 7. The classification results of different classifiers when three features are selected and for categorizing burn images into two burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II (%)
Random Forest91.7691.62/91.98/91.7094.09/89.88
PSO-SVM90.9590.78/91.03/90.8791.81/90.24
Decision Tree83.4582.73/81.32/80.9778.89/83.76
KNN85.8886.76/86.56/85.7390.26/82.87
LSTM87.3088.18/87.90/87.2193.48/82.32
1D-CNN90.5490.74/91.17/90.5396.97/85.37
Table 8. The classification results of different classifiers when three features are selected and for categorizing burn images into three burn classes.
Table 8. The classification results of different classifiers when three features are selected and for categorizing burn images into three burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II/III (%)
Random Forest80.7477.77/77.99/77.6496.21/67.04/70.71
PSO-SVM76.6276.70/75.86/73.3798.70/45.74/84.14
Decision Tree67.7763.21/64.85/65.2982.91/59.04/50.57
KNN69.8968.28/67.07/66.4381.57/56.58/64.05
LSTM70.2769.05/70.46/65.3292.70/37.11/80.57
1D-CNN74.5171.09/72.82/70.3095.3049.44/73.71
Table 9. When selecting three features, a comparison is carried out between the results of classifying burn images into different categories and the results of Refs. [9,15,28].
Table 9. When selecting three features, a comparison is carried out between the results of classifying burn images into different categories and the results of Refs. [9,15,28].
StudiesAccuracy (%)Precision/Recall/F1-Score (%)Category I/II/III (%)
Three CategoriesThe proposed method80.7477.77/77.99/77.6496.21/67.04/70.71
Rostami et al. [15]77.8\\
Zhang et al. [28]76.8775.28/76.83/75.2193.35/55.92/78.23
Two CategoriesThe proposed method91.7691.62/91.98/91.7094.09/89.88
Rostami et al. [15]90.590.6/87.9/89.22\
Zhang et al. [28]85.8686.79/87.72/86.1493.48/84.74
Yadav et al. [9]82.4382/88/85\
Table 10. The p-values calculated for the accuracies obtained by the different classifiers.
Table 10. The p-values calculated for the accuracies obtained by the different classifiers.
PSO-SVMDecision TreeKNNLSTM1D-CNN
Three categories9.58 × 10−52.45 × 10−125.92 × 10−143.89 × 10−163.89 × 10−10
Two categories0.0222570.0001321.87 × 10−144.44 × 10−187.33 × 10−26
Table 11. The classification results of different classifiers when depth features are selected and for categorizing burn images into two burn classes.
Table 11. The classification results of different classifiers when depth features are selected and for categorizing burn images into two burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II (%)
Random Forest88.9288.04/88.18/88.0988.09/88.27
PSO-SVM90.5490.37/90.58/90.4690.91/90.24
Decision Tree83.4584.62/83.60/83.1784.60/82.61
KNN85.5486.05/85.58/85.2784.61/86.55
LSTM86.0886.95/86.18/85.8787.12/85.24
1D-CNN89.4889.94/89.73/89.1693.17/86.29
Table 12. The classification results of different classifiers when depth features are selected and for categorizing burn images into three burn classes.
Table 12. The classification results of different classifiers when depth features are selected and for categorizing burn images into three burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II/III (%)
Random Forest75.9573.20/75.46/73.6088.18/59.63/78.57
PSO-SVM76.0873.11/75.11/73.4790/58.89/76.43
Decision Tree66.2863.54/66.98/63.9180.75/45.88/72.50
KNN68.9269.47/65.91/62.1685.33/59.31/51.08
LSTM75.8774.25/75.23/72.0487.27/55.56/82.86
1D-CNN75.1473.40/74.42/72.2989.09/57.04/77.14
Table 13. The classification results of different classifiers when color features are selected and for categorizing burn images into two burn classes.
Table 13. The classification results of different classifiers when color features are selected and for categorizing burn images into two burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II (%)
Random Forest69.1270.04/69.98/69.1177.88/62.07
PSO-SVM71.6972.94/71.66/70.5468.79/74.54
Decision Tree68.3869.18/68.75/67.8070.74/66.77
KNN68.1870.80/69.80/67.7579.47/60.11
LSTM72.8473.92/71.98/71.4164.09/79.88
1D-CNN71.3571.63/71.31/71.0872.33/70.29
Table 14. The classification results of different classifiers when color features are selected and for categorizing burn images into three burn classes.
Table 14. The classification results of different classifiers when color features are selected and for categorizing burn images into three burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II/III (%)
Random Forest62.0358.22/57.62/56.7886.97/39.81/46.07
PSO-SVM57.1654.87/51.76/50.7983.63/35.93/35.71
Decision Tree63.2462.67/61.36/60.3871.48/58.28/54.32
KNN55.6855.74/54.3/53.4578.30/38.48/45.81
LSTM56.0353.74/54.89/50.7083.03/25.93/55.71
1D-CNN52.6550.18/51.34/47.4373.93/26.48/52.43
Table 15. The classification results of different classifiers when texture features are selected and for categorizing burn images into two burn classes.
Table 15. The classification results of different classifiers when texture features are selected and for categorizing burn images into two burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II (%)
Random Forest63.7270.21/59.97/56.2925.30/94.63
PSO-SVM61.2865.02/57.54/53.5122.88/92.20
Decision Tree61.6862.12/61.27/60.0756.66/65.88
KNN60.1461.19/60.20/59.1556.06/64.34
LSTM48.5846.03/46.06/44.0022.73/69.39
1D-CNN54.8752.68/51.94/49.3224.85/79.02
Table 16. The classification results of different classifiers when texture features are selected and for categorizing burn images into three burn classes.
Table 16. The classification results of different classifiers when texture features are selected and for categorizing burn images into three burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II/III (%)
Random Forest36.6248.42/42.08/36.8222.88/33.07/66.29
PSO-SVM40.5446.56/42.92/39.0129.79/51.96/45.71
Decision Tree43.2441.71/40.57/43.7155.14/45.04/19.14
KNN47.146.90/44.26/43.1758.31/46.22/28.27
LSTM33.8436.59/34.01/31.1720.97/44.48/39.57
1D-CNN33.0336.79/37.19/33.1132.21/25.37/53.57
Table 17. The classification results of random forest classifier when features are combined in pairs and for categorizing burn images into two burn classes.
Table 17. The classification results of random forest classifier when features are combined in pairs and for categorizing burn images into two burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II (%)
Color + Texture69.3269.29/69.51/69.2271.21/67.80
Color + Depth90.4190.26/90.54/90.3391.81/89.27
Texture + Depth89.1988.76/88.91/88.8188.79/89.02
Table 18. The classification results of random forest classifier when features are combined in pairs and for categorizing burn images into three burn classes.
Table 18. The classification results of random forest classifier when features are combined in pairs and for categorizing burn images into three burn classes.
Feature CombinationAccuracy (%)Precision/Recall/F1-Score (%)Category I/II/III (%)
Color + Texture63.0458.68/57.83/57.1486.67/41.11/45.71
Color + Depth80.2777.11/77.66/77.0496.36/65.19/71.43
Texture + Depth78.5876.05/77.73/75.3796.97/54.44/81.79
Table 19. Summary table of several kinds of computing time of the proposed method.
Table 19. Summary table of several kinds of computing time of the proposed method.
Two CategoriesThree Categories
Training time (s)73.9470.28
Testing time (s)2.977.54
Computing time (s)76.9177.82
Computing time without feature selection (s)76.9973.25
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Zhang, Q.; Li, P.; You, J.; Sun, J.; Zhou, J. Multi-Feature Extraction and Selection Method to Diagnose Burn Depth from Burn Images. Electronics 2024, 13, 3665. https://doi.org/10.3390/electronics13183665

AMA Style

Zhang X, Zhang Q, Li P, You J, Sun J, Zhou J. Multi-Feature Extraction and Selection Method to Diagnose Burn Depth from Burn Images. Electronics. 2024; 13(18):3665. https://doi.org/10.3390/electronics13183665

Chicago/Turabian Style

Zhang, Xizhe, Qi Zhang, Peixian Li, Jie You, Jingzhang Sun, and Jianhang Zhou. 2024. "Multi-Feature Extraction and Selection Method to Diagnose Burn Depth from Burn Images" Electronics 13, no. 18: 3665. https://doi.org/10.3390/electronics13183665

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop