Next Article in Journal
Investigating Contrastive Pair Learning’s Frontiers in Supervised, Semisupervised, and Self-Supervised Learning
Next Article in Special Issue
Celiac Disease Deep Learning Image Classification Using Convolutional Neural Networks
Previous Article in Journal
AIDA (Artificial Intelligence Dystocia Algorithm) in Prolonged Dystocic Labor: Focus on Asynclitism Degree
Previous Article in Special Issue
Automated Lung Cancer Diagnosis Applying Butterworth Filtering, Bi-Level Feature Extraction, and Sparce Convolutional Neural Network to Luna 16 CT Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gastric Cancer Image Classification: A Comparative Analysis and Feature Fusion Strategies

Department of Mathematics and Computer Science, University of Cagliari, Via Ospedale 72, 09124 Cagliari, Italy
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2024, 10(8), 195; https://doi.org/10.3390/jimaging10080195
Submission received: 3 July 2024 / Revised: 1 August 2024 / Accepted: 7 August 2024 / Published: 10 August 2024

Abstract

:
Gastric cancer is the fifth most common and fourth deadliest cancer worldwide, with a bleak 5-year survival rate of about 20%. Despite significant research into its pathobiology, prognostic predictability remains insufficient due to pathologists’ heavy workloads and the potential for diagnostic errors. Consequently, there is a pressing need for automated and precise histopathological diagnostic tools. This study leverages Machine Learning and Deep Learning techniques to classify histopathological images into healthy and cancerous categories. By utilizing both handcrafted and deep features and shallow learning classifiers on the GasHisSDB dataset, we conduct a comparative analysis to identify the most effective combinations of features and classifiers for differentiating normal from abnormal histopathological images without employing fine-tuning strategies. Our methodology achieves an accuracy of 95% with the SVM classifier, underscoring the effectiveness of feature fusion strategies. Additionally, cross-magnification experiments produced promising results with accuracies close to 80% and 90% when testing the models on unseen testing images with different resolutions.

1. Introduction

Gastric cancer is the fifth most prevalent cancer globally and the fourth leading cause of cancer-related deaths, with a global 5-year survival rate hovering around 20%. Despite significant research into the disease’s pathobiology, predicting its progression remains difficult, contributing to the persistently low survival rates. Furthermore, medical diagnostics’ intricate and time-consuming nature can lead to missing critical details during microscopic examinations, potentially resulting in misdiagnoses [1,2].
While recent advancements in computer technology, especially in Machine Learning (ML) and Deep Learning (DL), have enabled notable progress [3,4,5,6], there remain significant challenges [6,7]. Existing models often require extensive fine-tuning and customization to perform well in specific medical imaging tasks, which may not always be feasible in practical scenarios [3,6]. Additionally, the transferability and general applicability of features derived from pre-trained models to the medical domain are not well understood [6,8,9]. There is also a lack of comprehensive studies comparing the effectiveness of handcrafted (HC) features versus deep features across different classifiers in the context of gastric cancer histopathological images [4,6,9].
Creating computational tools that can automatically and accurately perform histopathological diagnoses is essential to addressing these challenges. In this study, we contribute to the field of gastric cancer pathological image classification by utilizing the GasHisSDB dataset. It has been explicitly designed to evaluate the effectiveness of shallow learning classifiers using both HC features and deep features derived from pre-trained Convolutional Neural Network (CNN) architectures and to investigate the extent to which these general features, without any specific fine-tuning strategy, can be effectively used in the medical context of gastric cancer classification.
The rationale for this approach is two-fold. First, by employing these deep features without any specific optimization or fine-tuning, we aim to assess their inherent potential, transferability, and general applicability of these features to medical imaging tasks. This investigation is important for understanding the robustness and effectiveness of general-purpose features in specialized domains, particularly when resources for fine-tuning or designing custom models may be limited.
Second, establishing a baseline performance using general features allows us to benchmark the potential gains that can be achieved through more specific adaptations in future work. This study serves as a foundational step, providing insights into the inherent capabilities of pre-trained models in the medical imaging domain, which can inform and guide subsequent efforts in fine-tuning and custom model design.
Moreover, we investigate feature fusion techniques, exploring the combination of both HC and deep features. This exploration allows us to examine how integrating different types of features can lead to improved classification performance. By understanding the synergies between these feature sets, we can identify optimal strategies for enhancing the robustness and accuracy of pathology image classification.
Additionally, we conducted a cross-magnification experiment to assess how varying image resolutions affect the performance of classification models in histopathological image analyses. This experiment is particularly important as it addresses the practical challenges faced in histopathology, where images may be captured at different magnifications [10,11]. High-magnification images provide detailed cellular structures, while low-magnification images offer a broader tissue architecture perspective [10,11]. By analyzing how different resolutions affect classification accuracy, we contribute valuable knowledge that can inform future practices in the field, ensuring that classifiers are effective and adaptable to the diverse conditions under which pathology images are obtained.
To sum up, the contributions of this work are the following:
  • We proposed a comparative analysis of various HC and deep features across four different ML classifiers to identify the most stable and high-performing feature–classifier pairs for classifying gastric cancer histopathological images and distinguishing between normal and abnormal cells;
  • We explored and analyzed various feature fusion techniques to determine their effectiveness in enhancing classification accuracy in the task at hand;
  • We conducted a cross-magnification experiment to evaluate the impact of different image resolutions on classification performance, providing insights into the efficacy of utilizing multiple magnifications in pathology image analyses;
  • Since different magnifications highlight unique tissue features, we conducted a cross-magnification experiment to assess the impact of varying image resolutions on classification performance, providing insights on the use of different magnifications in this field;
  • We thoroughly evaluated the GasHisSDB dataset and compared our results with the state-of-the-art techniques.
The rest of this manuscript is organized as follows. Section 2 reviews the existing literature to contextualize our research within the field. Section 3 outlines the dataset and methodologies employed in this study. Section 4 presents the findings of our experiments, highlighting the performance of different feature categories, various classifiers, and feature combinations. Section 5 offers an in-depth analysis of our results, comparing them with previous studies and exploring their implications. Finally, the Section 6 summarizes our contributions and suggests directions for future research.

2. Related Work

The early detection and accurate diagnosis of gastric cancer (GC) is crucial, as patients with early-stage gastric cancer (EGC) have a much higher 5-year survival rate of 70–90% compared to only 10–30% for advanced gastric cancer (AGC) [9]. However, the accuracy of standard white-light endoscopy for detecting EGC is limited to 70–80%, heavily relying on the expertise of the endoscopist [9]. In recent years, researchers have increasingly explored the use of Computer Vision (CV) and DL techniques to assist in detecting and classifying gastric cancer from endoscopic and pathological images [12].
One of the first studies in this area was by Hirasawa et al., who developed a novel CNN for detecting and recognizing gastric cancer in video images [3]. Similarly, Yoon et al. developed an optimized CNN model for EGC detection and prediction [7]. Beyond endoscopic image analyses, researchers have also explored the use of CV techniques for gastric cancer classification using pathological images. For instance, Zhao et al. conducted a systematic review on the application of CNNs for identifying gastric cancer [4]. They found that a total of 27 studies had used CNN-based models for gastric cancer detection, classification, segmentation, and margin delineation from various medical imaging modalities, including endoscopy and pathology.
The reported accuracy of the CNN-based systems ranged from 77.3% to 98.7%, demonstrating the strong potential of these techniques for assisting clinicians in the diagnosis of gastric cancer [4]. One notable study in this domain was by Xie et al., who developed an optimized GoogleNet model for the diagnosis of gastric cancer pathological images [5]. Their improved model, which combined the strengths of two network structures, achieved a sensitivity of 97.61% and a specificity of 99.47% in recognizing gastric cancer pathological sections [5].
In this context, Hu et al. proposed a comprehensive dataset, named GasHisSDB, with 245,196 sub-sized gastric histopathology images labeled as normal or gastric cancer, which were derived from 600 whole slide images (WSIs) [2]. It was introduced to overcome the limitations of the existing datasets, particularly their small sample sizes [2,13]. Several follow-up studies have used the GasHisSDB dataset, starting from its proposal, where Hu et al. evaluated the performance of various ML and DL models [8], while several authors proposed optimized approaches to accomplish this task. For instance, Yong et al. proposed an ensemble DL approach based on EfficientNetB0, EfficientNetB1, DenseNet-121, DenseNet169, and MobileNet [6] whereas Li et al. introduced a lightweight gated fully fused network (LGFFN) with a gated hybrid input (GHI) module. The LGFFN-GHI comprises two main components: feature extraction and classification modules. The feature extraction module uses a cross-attention mechanism to fuse features from different scales. The classification module then takes the fused features and outputs the final classification prediction [14]. In addition, Fu et al. proposed MCLNet, a multidimensional CNN based on ShuffleNet. It extracts the correlation features between pixels in an image by one-dimensional convolution to achieve pixel-level and patch-level feature interaction.
Overall, the reviewed studies demonstrate significant progress in applying CV and DL techniques for gastric cancer classification from endoscopic and pathological images. However, despite these advances, there remain notable gaps and challenges. Many existing studies rely heavily on specific fine-tuning and optimization strategies tailored to particular datasets or clinical settings, which may limit the generalizability and transferability of their findings to broader contexts.
In this context, our study aims to advance the classification of gastric cancer using histopathologic images by addressing these gaps. Our primary objective is to propose a robust system that does not rely on ad hoc adjustments or fine-tuning. By leveraging features from non-optimized yet generic methods, we explore the feasibility of offering a generalizable solution that performs consistently across various magnifications and datasets. This approach is crucial for developing automated diagnostic tools that can be widely applicable and effective in diverse clinical environments.
Our study builds on the foundational work in this field by providing a comprehensive comparative analysis of both HC and deep features across multiple classifiers. We investigate the inherent potential and transferability of general features extracted from pre-trained CNNs without specific optimization. This analysis is important for understanding the robustness and effectiveness of general-purpose features in specialized domains, particularly when resources for fine-tuning or designing custom models may be limited.
Additionally, we explore feature fusion techniques to assess how integrating different types of features can enhance classification performance. By understanding the synergies between HC and deep features, we aim to identify optimal strategies for improving the robustness and accuracy of pathology image classification. Furthermore, we conduct cross-magnification experiments to evaluate the impact of different image resolutions on classification performance, addressing practical challenges faced in histopathology where images may be captured at varying magnifications.
In summary, our study contributes to the field by providing a detailed comparative analysis of feature extraction methods, exploring feature fusion strategies, and evaluating the effects of image magnification on classification accuracy. These efforts aim to develop a more generalizable and effective approach for automated gastric cancer diagnoses, advancing the application of ML and DL techniques in gastric cancer classification and addressing key limitations in current research that can influence the performance of the systems, such as the size and diversity of the training datasets, the specific CNN architectures employed, and the clinical context in which they are deployed.

3. Materials and Methods

This section provides details of the components used in our study. We begin with an overview of the dataset in Section 3.1, detailing its composition and relevance to our research objectives. Following this, in Section 3.2, we present both feature extraction methods employed, HC and deep, whereas in Section 3.3, we describe the classification methods applied. In addition, we discuss the performance evaluation measures in Section 3.4. Finally, Section 3.5 outlines the experimental setup and implementation details.

3.1. Dataset

The GasHisSDB dataset is a publicly available gastric histopathology image dataset [2]. It contains a total of 245,196 sub-sized gastric histopathology images, which were derived from 600 WSIs, stained with H&E, of 2048 × 2048 pixels. The images were scanned using a NewUsbCamera and digitized at 20 × magnification. Two experienced pathologists from Liaoning Cancer Hospital and Institute provided the labels, classifying the images as either normal or abnormal (gastric cancer). A normal image is characterized by the absence of cancerous regions, reflecting typical microscopic cell observations. In contrast, an image is labeled as abnormal when approximately 50% of its area is occupied by cancerous regions [2]. The dataset is divided into three image sub-databases, each of them containing images with specific resolutions: 160 × 160 (S-A), 120 × 120 (S-B), and 80 × 80 (S-C) pixels. The distribution of the dataset images is summarized in Table 1, while Figure 1 shows two image samples.

3.2. Feature Extraction Methods

Features derived from images include a wide array of descriptors designed to capture morphological, pixel-level, and textural information, denoted as handcrafted features. As noted by [15], HC features can be broadly categorized into three main groups: invariant moments, texture features, and color features. To them, we have added a set of deep features, i.e., features obtained by the activations of off-the-shelf CNNs. In the following, we present a brief summary of each category along with the specific descriptors utilized.

3.2.1. Invariant Moments

An image moment is a weighted average of pixel intensities in an image used to extract specific properties. Moments are crucial in image analyses and pattern recognition, helping to characterize segmented objects [16]. This study employs three distinct types of moments: Zernike, Legendre, and Chebyshev. A concise overview of these moment types follows.
Chebyshev Moments (CH) constitute a class of discrete orthogonal moments [17], based on Chebyshev polynomials [18] with the maximum possible leading coefficient constrained by an absolute value of 1 within the interval [ 1 , 1 ] . This study used the first- and second-order moments, denoted as CH_1 and CH_2, respectively. Both moments were calculated to the fifth order. They are defined as
T p q = x = 0 M 1 y = 0 N 1 T p ( x ) T q ( y ) f ( x , y )
where T p ( x ) and T q ( y ) are the Chebyshev polynomials of order p and q, respectively, and f ( x , y ) is the image function.
Second-order Legendre Moments (LM) are a type of continuous orthogonal moment that can be used for image analyses. LM capture information about the shape and orientation of an image. They are calculated using the second-order Legendre polynomials, which are orthogonal over the interval [ 1 , 1 ] [19,20]; they can capture and represent objects’ shape and spatial characteristics within an image. In our analysis, we extracted the LM of order 5. The LM are defined as
L p q = 2 p + 1 2 2 q + 1 2 x = 1 1 y = 1 1 P p ( x ) P q ( y ) f ( x , y )
where P p ( x ) and P q ( y ) are the Legendre polynomials of order p and q, respectively, and f ( x , y ) is the image function.
Zernike Moments (ZM) are a type of continuous orthogonal moment that is defined over the unit circle [20]. They are calculated using Zernike polynomials, which form an orthogonal basis [21]. In this study, we extracted the ZM of order 5 with a repetition of 5. The ZM are defined as
Z n m = n + 1 π x = 1 1 y = 1 1 V n m * ( x , y ) f ( x , y )
where V n m ( x , y ) are the Zernike polynomials, V n m * denotes the complex conjugate, n is the order, m is the repetition, and f ( x , y ) is the image function.
The specific forms of the Chebyshev, Legendre, and Zernike polynomials are provided below:
  • Chebyshev polynomial  T p ( x ) of order p:
    T p ( x ) = cos ( p · arccos ( x ) )
  • Legendre polynomial  P p ( x ) of order p:
    P p ( x ) = 1 2 p k = 0 p ( 1 ) k p k 2 p 2 k p x p 2 k
  • Zernike polynomial  V n m ( x , y ) of order n with repetition m:
    V n m ( x , y ) = R n m ( ρ ) · e j m θ
    where ρ = x 2 + y 2 and θ = arctan ( y / x ) , and R n m ( ρ ) is the radial polynomial defined as
    R n m ( ρ ) = s = 0 ( n | m | ) / 2 ( 1 ) s ( n s ) ! s ! ( ( n + | m | ) / 2 s ) ! ( ( n | m | ) / 2 s ) ! ρ n 2 s

3.2.2. Texture Features

Texture serves as a visual feature indicative of homogeneity within an image. It reveals the organization and arrangement of surface structures exhibiting gradual or periodic variations. Rather than relying on individual pixel characteristics, a texture analysis requires statistical calculations over regions encompassing multiple pixels [22]. The texture is characterized by the gray-level distribution of pixels and their surrounding spatial neighbors, encapsulating local texture information. Additionally, global texture information is determined by the extent of repetition of this local texture information. For the sake of this work, we have considered two widely employed methods, now described.
Rotation-Invariant Haralick (HAR) Features: Thirteen HAR features were extracted from the Gray-Level Co-occurrence Matrix (GLCM) and then transformed into rotation-invariant features [23]. To achieve rotation invariance, four variations of the GLCM were calculated, each with a distance parameter d = 1 and angular orientations θ = [ 0 , 45 , 90 , 135 ] .
The Gray-Level Co-occurrence Matrix (GLCM) is defined as
P ( i , j , d , θ ) = x = 1 N y = 1 N 1 if I ( x , y ) = i and I ( x + d cos θ , y + d sin θ ) = j 0 otherwise
From the GLCM, we extracted the first 13 HAR features. They are defined as follows:
Angular Second Moment ( ASM ) = i = 0 N 1 j = 0 N 1 P ( i , j ) 2
Contrast = n = 0 N 1 n 2 i = 0 N 1 j = 0 N 1 P ( i , j ) , | i j | = n
Correlation = i = 0 N 1 j = 0 N 1 ( i · j · P ( i , j ) ) μ x · μ y σ x · σ y
where
μ x = i = 0 N 1 i j = 0 N 1 P ( i , j ) μ y = j = 0 N 1 j i = 0 N 1 P ( i , j ) σ x = i = 0 N 1 ( i μ x ) 2 j = 0 N 1 P ( i , j ) σ y = j = 0 N 1 ( j μ y ) 2 i = 0 N 1 P ( i , j )
Variance = i = 0 N 1 j = 0 N 1 ( i μ ) 2 P ( i , j )
where
μ = i = 0 N 1 j = 0 N 1 i · P ( i , j )
Inverse Difference Moment ( IDM ) = i = 0 N 1 j = 0 N 1 P ( i , j ) 1 + ( i j ) 2
Sum Average = i = 2 2 N i · P x + y ( i )
where
P x + y ( k ) = i = 0 N 1 j = 0 N 1 P ( i , j ) , i + j = k
Sum Variance = i = 2 2 N ( i Sum Average ) 2 · P x + y ( i )
Sum Entropy = i = 2 2 N P x + y ( i ) log P x + y ( i )
Entropy = i = 0 N 1 j = 0 N 1 P ( i , j ) log P ( i , j )
Difference Variance = Variance of P x y ( k )
where
P x y ( k ) = i = 0 N 1 j = 0 N 1 P ( i , j ) , | i j | = k
Difference Entropy = i = 0 N 1 P x y ( i ) log P x y ( i )
Information Measures of Correlation 1 = H X Y H X Y 1 max { H X , H Y }
where
H X Y = i = 0 N 1 j = 0 N 1 P ( i , j ) log P ( i , j ) H X Y 1 = i = 0 N 1 j = 0 N 1 P ( i , j ) log ( P x ( i ) P y ( j ) ) H X = i = 0 N 1 P x ( i ) log P x ( i ) H Y = j = 0 N 1 P y ( j ) log P y ( j )
Information Measures of Correlation 2 = 1 exp 2 ( H X Y 2 H X Y )
where
H X Y 2 = i = 0 N 1 j = 0 N 1 P x ( i ) P y ( j ) log ( P x ( i ) P y ( j ) )
Local Binary Pattern (LBP) is a powerful method for capturing the texture and patterns in an image, as described in [24]. In this study, we computed the histogram of the LBP and transformed it into a rotation-invariant form [25]. This histogram was then extracted and used as the feature vector. The LBP map was generated within a neighborhood defined by a radius of r = 1 and eight neighbors ( n = 8 ) .
The LBP operator assigns a binary code to each pixel by thresholding its neighborhood with the center pixel value. The LBP code for a pixel ( x c , y c ) is given by
L B P P , R = p = 0 P 1 s ( i p i c ) 2 p
where s ( x ) is the sign function
s ( x ) = 1 if x 0 0 otherwise

3.2.3. Color Features

Histograms are the most widely employed method for characterizing the color properties of images since they effectively represent the global color distribution within an image, indicating their proportion. The descriptors that can be extracted from the histogram are invariant to image rotation, translation, and scaling changes. However, they have a significant limitation in that they cannot describe the local distribution of colors, the spatial location of each color, or specific objects within the image [26]. In this study, these descriptors were calculated from images that underwent a conversion to grayscale, streamlining the process of the analysis and computation.
Histogram (Hist) Features: From the histogram, which characterizes the overall color distribution within the image, we derived seven statistical descriptors: the mean, standard deviation, smoothness, skewness, kurtosis, uniformity, and entropy.
For a histogram h ( i ) with N bins, the statistical descriptors are defined as follows:
Mean = 1 N i = 0 N 1 i · h ( i )
Standard Deviation = 1 N i = 0 N 1 ( i μ ) 2 · h ( i )
Smoothness = 1 1 1 + Variance
Skewness = 1 N σ 3 i = 0 N 1 ( i μ ) 3 · h ( i )
Kurtosis = 1 N σ 4 i = 0 N 1 ( i μ ) 4 · h ( i ) 3
Uniformity = i = 0 N 1 h ( i ) 2
Entropy = i = 0 N 1 h ( i ) log h ( i )
Autocorrelogram (AC): The AC captures the spatial correlation of colors within an image. It is a restricted version of the more general color correlogram, considering only the spatial correlation between pixels of the same color [27]. Specifically, the color autocorrelogram calculates the probability that a pixel of a given color will be found at a certain distance, d, away from another pixel of the same color. Our research considered four discrete distances: d = 1 , 2 , 3 , 4 . The four resulting probability vectors are concatenated to form a comprehensive feature vector.
The autocorrelogram for color k at distance d is defined as
AC k ( d ) = 1 N k i = 1 N k j N d ( i ) δ ( C ( i ) , k ) δ ( C ( j ) , k )
where N k is the number of pixels of color k, N d ( i ) is the set of pixels at distance d from pixel i, and δ is the Kronecker delta function.
Haar-like (Haar) Features: The key idea behind these features is to calculate the difference in the sum of pixel intensities across rectangular regions in an image. This allows detecting edges, lines, and center-surround features that indicate the presence of an object [28]. Haar-like features can be calculated using integral images to speed up the process. The integral image at a location ( x , y ) is defined as
I I ( x , y ) = x x y y I ( x , y )
Using the integral image, the sum of pixel intensities within a rectangular region can be computed efficiently, allowing for the calculation of Haar-like features.

3.2.4. Deep Features

With deep features, we refer to the characteristics of an image derived from the CNN activations since they have proven to be a potent strategy for enhancing the predictive power of classifiers [29]. These deep features were extracted from off-the-shelf CNN architectures pre-trained on the well-known natural image dataset ImageNet [30].
Specifically, depending on the architecture, deep features were extracted from one of the following layers: (i) the penultimate layer, (ii) the final fully connected layer, or (iii) the last pooling layer. This approach ensures extracting features encapsulating the network’s learned global knowledge. Notably, the fine-tuning strategy for the classification phase was not employed to maintain the generalization ability of the networks [31,32]. Detailed specifications regarding the selected layers for feature extraction, input dimensions, and the count of trainable parameters for each CNN model are outlined in Table 2, while a brief explanation of the CNNs employed is now provided.
AlexNet consists of a sequence of convolutional and max-pooling layers, culminating in three fully connected layers [33]. With only five convolutional layers, it represents the most shallow architecture used in this study.
DarkNet builds upon the established principles of inception and batch normalization. This study employs two specific versions of DarkNet, incorporating 19 [34] and 53 [35] convolutional layers. These configurations form the foundational network for the You Only Look Once object detection method.
DenseNet, proposed by Huang et al. [36], addresses the typical CNN characteristic of having layers equal to the number of connections. Specifically, the number of connections is L ( L + 1 ) / 2 , where L denotes the number of layers. Each layer’s input comprises the output from all preceding layers, which then serves as the input for the subsequent layer.
EfficientNet stands out for its uniform and efficient scaling of network width, depth, and resolution through compound scaling. Proposed by Tan et al. [37], this study employs the EfficientNetB0 version.
Inception-v3 uses the inception layer concept by incorporating factorized, smaller, and asymmetric convolutions [38]. Inception models are notable for their multi-branch architectures, combining filters of various sizes integrated through concatenation within each branch.
Inception-ResNet-v2 merges the strengths of ResNet and Inception architectures [39]. The Inception-ResNet block combines variously sized convolutional filters with residual connections, featuring four max-pooling layers and 160 convolutional layers.
ResNet refers to a family of deep architectures that use residual learning [40]. These architectures integrate skip-connections or recurrent units to link blocks of convolutional and pooling layers, with each block followed by batch normalization [43]. This study employs three ResNet variants, ResNet-18, ResNet-50, and ResNet-101, with the numbers indicating the respective network depths.
VGG comprises a series of convolutional layers followed by max-pooling, which enhances its deep representation capabilities [41]. This study uses VGG19, featuring 19 layers.
XceptionNet extends the Inception architecture by employing depth-wise separable convolutions to improve efficiency and reduce parameter count. This approach aims to capture complex feature dependencies by focusing on cross-channel correlations [42].

3.3. Classification Methods

After feature extraction, HC and deep features served as inputs for four classical ML algorithms to classify GasHisSDB. Here is a brief overview of these classifiers.
Decision Tree (DT) is a hierarchical data structure used for prediction. Each internal node represents a feature, with branches denoting possible feature values and leaves representing different categories. The algorithm optimizes this structure by pruning nodes that minimally contribute to category separation, thereby merging instances at higher levels. Classification is achieved by tracing the path from the root to a leaf node [44].
k-Nearest Neighbor (kNN): The kNN classifier categorizes observations by considering the classes of the k training examples nearest to the observation in question. This method employs a local strategy for classification, leveraging the proximity of neighboring instances to determine the class [45].
Support Vector Machine (SVM): SVM differentiates categories by mapping examples to opposite sides of a decision boundary. The one-vs.-rest approach is employed for multiclass problems, training individual classifiers to distinguish each class from all others [46].
Random Forest (RF): This algorithm aggregates predictions from multiple Decision Trees, each constructed from random subsets of features and examples. By fostering diversity among the trees, this ensemble method enhances model robustness, improving resilience against data imbalance and mitigating overfitting. The use of 100 trees specifically enhances the random forest’s predictive accuracy [47].

3.4. Performance Evaluation Measures

In evaluating the performance of a binary classifier on a dataset, each instance is classified as either negative or positive based on the classifier’s predictions. The result of this classification, when compared to the actual target value, determines the following performance measures:
  • True Negatives (TNs): instances correctly predicted as negative.
  • False Positives (FPs): instances incorrectly predicted as positive.
  • False Negatives (FNs): instances incorrectly predicted as negative.
  • True Positives (TPs): instances correctly predicted as positive.
As detailed below, we assess the classifier’s performance using several measures specifically defined for binary classification tasks.
  • Accuracy (A): It is the ratio of correct predictions to the total number of predictions:
    A = T P + T N T P + F N + F P + T N
  • Precision (P) is the ratio of TPs to the sum of TPs and FPs, indicating the classifier’s efficiency in predicting positive instances:
    P = T P T P + F P
  • Recall (R), also known as sensitivity, is the ratio of TPs to the sum of TPs and FNs:
    R = T P T P + F N
  • Specificity (S) is the ratio of TNs to the sum of TNs and FPs:
    S = T N T N + F P
  • F1-score (F1) is the harmonic mean of P and R, considering both FPs and FNs:
    F 1 = 2 · P · R P + R
  • Matthews Correlation Coefficient (MCC) is a comprehensive measure that considers all elements of the confusion matrix (TP, TN, FP, FN). Ranging from 1 to + 1 , it provides a high score only when the classifier performs well in both the positive and negative classes:
    M C C = T P · T N F P · F N ( T P + F P ) ( T P + F N ) ( T N + F P ) ( T N + F N )
  • Balanced accuracy (BACC) is defined as the mean of specificity and sensitivity:
    B A C C = S + R 2

3.5. Experimental Setup

The experiments were performed on a workstation with an Intel(R) Core(TM) i9-8950HK @ 2.90 GHz CPU, 32 GB of RAM, and an NVIDIA GTX1050 Ti GPU with 4 GB of memory. MATLAB R2021b was used for all implementations and experimental evaluations.
This study deliberately did not use image augmentations to concentrate on extracting pure features from the original images. Moreover, we used Euclidean distance as a distance measure for kNN with k = 1 ; note that with k = 1 , no voting strategy is required. In addition, the SVM kernel function uses a linear kernel, and the number of DTs composing the RF has been set to 100.
In addition, a 5-fold cross-validation (CVal) approach was employed for the testing strategy. This method ensures statistical reliability by repeatedly training and testing the same dataset. Specifically, the dataset is divided into 80% for training and 20% for testing at each iteration.

4. Experimental Results

In this section, we detail the comprehensive experimental analysis conducted to evaluate the performance of various feature extraction and classification techniques on the GasHisSDB dataset. Section 4.1 presents the results obtained with HC features, whereas Section 4.2 explores the use of CNN as feature extractors. This is succeeded by Section 4.3, where we discuss the outcomes of combining HC and deep features to enhance classification accuracy. To ensure robustness across different magnifications, Section 4.4 evaluates the consistency and reliability of our methods when applied to images at various magnification levels. Finally, Section 4.5 provides a critical analysis of our results in the context of existing research. Finally, please note that for the sake of comparison, we have reported only the results obtained with the two best-performing classifiers, i.e., RF and SVM. We report kNN and DT results in Appendix A.

4.1. HC Feature Performance

The outcomes obtained by RF and SVM trained with HC features are presented in Table 3 and Table 4, while Table A1 and Table A2 report the results obtained with DT and kNN, respectively.
The performance obtained with the SVM (Table 3) shows that the CH_1 and CH_2 features provide the best accuracy (75.92% and 72.50%, respectively) and balanced accuracy (73.45% and 67.66%, respectively), showing their potential for discriminating between normal and abnormal tissues. Interestingly, while the Hist features achieve high precision (96.29%), it suffers from very low recall (9.00%), leading to a much lower balanced accuracy (54.23%).
As for the RF (Table 4), when trained with the LBP features, it achieves the highest accuracy (79.57%), precision (80.74%), and F1 (83.77%). The LBP feature’s strong performance across most metrics suggests its effectiveness in capturing essential patterns in histopathological images. Contrastively, the Haar feature again demonstrates a significantly lower accuracy (62.48%) and balanced accuracy (53.97%), indicating its relative ineffectiveness in this context.
Finally, the LBP again emerges as the top performer with both DT (Table A1) and kNN (Table A2). DT with LBP obtained an accuracy of 71.22% and a BACC of 69.87%. This indicates that despite the DT classifier’s simplicity, LBP features can still capture discriminative information effectively. Haar features, however, perform poorly with a balanced accuracy of 53.37%. Instead, kNN with LBP obtained an accuracy of 69.51% and an F1 of 74.58%. Conversely, the Haar feature shows poor performance with an accuracy of 42.17% and a balanced accuracy of 48.53%, reaffirming its limitations for this task.
The consistency across classifiers underlines LBP features’ robustness, even without top-notch performance.

4.2. Deep Feature Performance

The results of SVM and RF classifiers trained using deep features are summarized in Table 5 and Table 6. In Table A3 and Table A4, the results obtained with DT and kNN are reported, respectively.
As shown in Table 5, SVM trained with DenseNet-201 features achieves the highest accuracy (86.02%) and balanced accuracy (84.23%), followed closely by DarkNet-53 and EfficientNetB0.
Even with RF (Table 6), DenseNet-201 achieves the highest accuracy (91.93%) and balanced accuracy (91.33%), indicating its superior feature extraction capability. Other deep features, such as those from DarkNet-53 and ResNet-101, also perform exceptionally well with RF.
In addition, the features extracted from DenseNet-201 also excel with DT and kNN. As shown in Table 6, DT achieves the highest accuracy (84.92%) and balanced accuracy (84.20%). In contrast, kNN (Table A4) shows superior performance with DenseNet-201 and DarkNet-53, which gained accuracies of 88.21% and 88.25%, respectively, and balanced accuracies of 87.23% and 87.22%. These results further confirm that even simpler classifiers can benefit significantly from the rich feature representations pre-trained CNNs provide.
This consistent top performance across different classifiers highlights DenseNet-201’s strong feature extraction capabilities for histopathological images, even without any fine-tuning strategy.

4.3. Feature Fusion Performance

Despite the significant results achieved from the previous classification, further efforts were made to enhance performance. An additional experiment explored the potential of integrating the representative power of both HC and deep features into a feature fusion strategy. More specifically, this experiment focused on LBP, DenseNet-201, and EfficientNetB0, which resulted in the best HC and the best two deep features, respectively. They were evaluated with all possible combinations using the best three classifiers from the previous stage: DT, SVM, and RF. This integration aimed to leverage their combined strengths for improved performance. The results are shown in Table 7.
For the fusion of LBP and DenseNet-201 features, SVM emerged as the most effective classifier with an accuracy of 94.41% and F1 of 95.40%. This performance indicates that SVM, when trained with this fusion of features, is highly reliable. RF also performed robustly with an accuracy of 92.16%, showing strong capability but slightly lagging behind SVM.
In the case of combining LBP with EfficientNetB0 features, SVM again demonstrated superior performance with an accuracy of 94.05%, F1 of 95.11%, MCC of 87.53, and balanced accuracy of 93.68%. This reiterates SVM’s effectiveness across different feature combinations. RF showed a notable drop in performance compared to the previous fusion strategy, suggesting that this combination might not be as effective for RF.
The combination of DenseNet-201 and EfficientNetB0 features led to SVM achieving the highest measures overall, with an accuracy of 94.89% and F1-score of 95.78%. This indicates that the deep features from these two CNNs complement each other well, providing rich information for the classifier. RF and DT also performed better with this fusion strategy than HC features, highlighting the benefit of purely deep features.
Finally, the most complex feature fusion strategy, combining LBP with both DenseNet-201 and EfficientNetB0, resulted in the highest overall performance for SVM, with an accuracy of 95.03%, and F1 of 95.90%. This indicates that the incorporation of both HC and multiple deep features provides a comprehensive feature set that enhances classification performance. RF also showed its best performance with this fusion strategy, suggesting that adding more feature types helps improve its robustness and generalization.

4.4. Cross-Magnification Performance

Table 8 and Table 9 detail the performance measures of two cross-magnification experiments conducted using different classifiers and feature fusion strategies on the GasHisSDB dataset. The experiments involved training on a 160 × 160 image sub-database and testing on smaller dimensions ( 120 × 120 and 80 × 80 , respectively).
  • Results of testing on S-B: The classifiers were evaluated on the 120 × 120 test set in the first experiment. Combining LBP and DenseNet-201 as features yielded varied results across different classifiers. The RF classifier outperformed others, achieving an accuracy of 89.04%, an F1 of 90.76%, and a balanced accuracy of 89.76%. The SVM also demonstrated strong performance, particularly with a precision of 97.42%, though it lagged in recall compared to RF.
When integrating LBP with EfficientNetB0, the performance metrics showed a slight decline, especially noticeable in the DT classifier, which recorded an accuracy of 85.17%. The RF continued to maintain relatively high performance, albeit slightly lower than with DenseNet-201.
The fusion of DenseNet-201 and EfficientNetB0 features displayed a notable improvement in classifier performance. RF again led the results with an accuracy of 89.55%, an F1-score of 91.30%, and a balanced accuracy of 89.90%. The DT classifier also performed well under this strategy, achieving high precision and recall rates.
Combining all three feature sets (LBP, DenseNet-201, and EfficientNetB0) resulted in marginal improvements across the board. RF achieved the highest accuracy at 89.56%, while the SVM exhibited the highest precision at 98.45%. This comprehensive feature fusion strategy enhanced the robustness and consistency of the classifiers’ performance, particularly evident in the balanced accuracy and MCC scores.
  • Results of testing on S-C: The second experiment, with testing on the 80 × 80 sub-database, illustrated a greater challenge for the classifiers, reflected in the generally lower performance values. The LBP + DenseNet-201 combination showed that RF remained the most reliable classifier with an accuracy of 78.89% and an F1-score of 79.73%. While demonstrating high precision at 96.18%, the SVM struggled with recall and balanced accuracy, indicating a possible reliance on the 160 × 160 pixels’ image data.
In the LBP + EfficientNetB0 strategy, all classifiers showed decreased performance, with the SVM particularly underperforming in terms of recall and F1. RF again stood out, albeit with lower scores than the previous experiment.
The fusion of DenseNet-201 and EfficientNetB0 improved the values slightly, with RF achieving an accuracy of 79.73% and a balanced accuracy of 81.69%. This strategy illustrated a more balanced performance across the classifiers, with DT and SVM showing moderate improvements in precision and recall.
Lastly, the combination of all three feature sets in this second experiment underscored RF as the most robust classifier with an accuracy of 78.44% and an F1 of 79.09%. The SVM showed better balanced accuracy compared to previous setups, though it still struggled with recall.

4.5. Comparison with the State of the Art

Table 10 showcases a comparative analysis of the performance of our work against previous state-of-the-art studies on the GasHisSDB dataset.
In Hu et al.’s work [2], two models, VGG16 and ResNet50, were tested with a 40/40/20 split. VGG16 achieved accuracies of 96.12%, 96.47%, and 95.90% across S-C, S-B, and S-A, respectively. Similarly, ResNet50 showed comparable performance with 96.09%, 95.94%, and 96.09% in the same sub-databases.
In the study of [48], an InceptionV3 model trained from scratch using a 40/20/40 split achieved a remarkable 98.83% accuracy in the S-A sub-database. Furthermore, combining InceptionV3 and ResNet50 through feature concatenation yielded a very close accuracy of 98.80%.
Li et al. [14] used a local–global feature fuse network (LGFFN) with a 60/20/20 split, achieving an accuracy of 96.81% in the S-A. This approach leverages the strengths of local and global features to improve classification performance. On the other hand, [49] employed MCLNet based on ShuffleNetV2 with an 80/-/20 split, reporting high accuracies of 96.28%, 97.95%, and 97.85% across S-C, S-B, and S-A, respectively.
The ensemble method adopted in [6] with a 40/20/40 split exhibited outstanding results, with accuracies of 97.72%, 98.68%, and 99.20% in S-C, S-B, and S-A, respectively. This ensemble approach amalgamates the strengths of multiple models, thereby achieving superior performance and robustness in classification tasks.
The main differences between the state of the art and our work are that our evaluation followed a 5-fold Cval protocol; we used only HC features and features extracted from pre-trained, off-the-shelf CNNs to evaluate the extent to which non-specialized and non-tuned features can accomplish the binary classification task faced in this study. We tested on S-B and S-C by using only models trained on S-A to investigate the influence of the image resolution size in this scenario.
As can be seen, we reported two models: SVM with feature fusion and RF with feature fusion. The SVM model achieved accuracies of 60.31%, 85.82%, and 95.03% in S-C, S-B, and S-A, respectively. Similarly, the RF model showed accuracies of 78.44%, 89.56%, and 92.26% in the same categories. Although lower in S-B and S-C due to the training/testing strategy employed, compared to previous studies, these results highlight the potential of feature fusion techniques in improving classification performance, even without the need for a complex fine-tuning strategy that can be time-consuming and, above all, require a high amount of labeled data that can be complex in medical scenarios [50]. The 5-fold CV method ensures a more robust evaluation by repeatedly training and testing on different subsets of the data, thus providing a reliable estimate of the models’ performance.

5. Discussion

This section analyzes the key aspects of our study. Specifically, Section 5.1 examines the relative performance and merits of HC versus deep features, whereas Section 5.2 discusses the outcomes of combining HC and deep features, analyzing how this fusion impacts the overall classification performance. Next, Section 5.3 evaluates the robustness and adaptability of our classification models across different magnifications of the considered dataset. Finally, Section 5.4 addresses the constraints and potential weaknesses of our study.

5.1. On the HC vs. Deep Feature Comparison

The comparative analysis presented in Section 4.1 and Section 4.2 reveals that, on the one hand, LBP consistently performs well among HC features, demonstrating robustness and reliability across different classifiers, even without exceptional performance. On the contrary, Haar features generally perform poorly, suggesting that they are less suitable for this task.
On the other hand, deep features extracted from pre-trained CNNs, especially DenseNet-201 and DarkNet-53, consistently outperform HC features. This underscores the advantage of using Deep Learning models for feature extraction in complex tasks such as histopathological image classification, even without fine-tuning strategies.
In addition, the random forest classifier has shown strong performance with both HC and deep features, indicating its versatility and effectiveness in handling various feature types. More precisely, RF with features extracted from DenseNet-201 demonstrated their reliability for the task.
In summary, the detailed performance evaluation across various feature–classifier combinations provides valuable insights into the strengths and weaknesses of different approaches. The consistent superiority of deep features, particularly those from DenseNet-201 and EfficientNetB0, suggests a clear direction in this domain, emphasizing the integration of advanced Deep Learning techniques for enhanced classification accuracy, effectiveness, and robustness. This is the main reason that motivated us to pick them along with LBP among the HC features to investigate feature fusion strategies.

5.2. On the Feature Fusion Performance

Across all feature fusion strategies, SVM consistently outperformed both DT and RF, demonstrating its superior ability to handle the diverse and complex feature sets derived from combining HC and deep features. SVM’s consistently high performance across various combinations suggests that it is highly adaptable to different types of features.
While generally strong, RF showed variability in performance depending on the feature combination, indicating that it might be more sensitive to the quality and type of features used. DT, on the other hand, consistently lagged behind SVM and RF, pointing to its relatively lower ability to utilize complex feature sets effectively.
The combination of LBP, DenseNet-201, and EfficientNetB0 features, mainly when used with SVM, provides the most reliable and high-performing strategy for classifying histopathological images. This fusion strategy leverages the strengths of both HC and deep features, resulting in a robust classification framework.

5.3. On the Cross-Magnification Performance

Overall, the experiments reveal that the RF classifier consistently outperforms DT and SVM across various feature fusion strategies and test set dimensions. The combination of DenseNet-201 and EfficientNetB0 generally provides the most reliable feature set, enhancing classifier performance and demonstrating the feasibility of using features provided by HC methods or pre-trained CNNs in histopathological image classification.
The comprehensive analysis demonstrates the efficacy of feature fusion and the importance of choosing robust classifiers to achieve high accuracy and reliability in medical image classification tasks.

5.4. Limitations

While comprehensive in its approach to evaluating the performance of shallow learning classifiers on histopathological image classification, this study presents several limitations that must be acknowledged.
First, reliance on pre-trained CNNs for feature extraction without any fine-tuning specific to the dataset at hand may limit the potential performance of the classifiers. Fine-tuning these networks could potentially yield features more tailored to the specific characteristics of the histopathological images, thereby improving classification accuracy.
Second, the experiments were conducted using a single dataset, GasHisSDB, which might limit the generalizability of the findings. The performance measures observed might vary significantly when applied to other histopathological datasets with different image characteristics, variations in staining procedures, or differing disease profiles. A broader validation across multiple datasets would provide more robust evidence of the classifiers’ effectiveness.
Third, this study employed a specific image resolution sub-database ( 160 × 160 for training and testing, 120 × 120 and 80 × 80 for testing). The impact of image resolution on classifier performance was not extensively explored, and it is possible that different resolutions could influence the feature extraction and classification processes.
Additionally, this study did not incorporate image augmentation techniques commonly used in image classification tasks to improve model generalization by artificially increasing the size and variability of the training dataset. The absence of augmentation may result in overfitting, particularly given the limited data available for training.
In summary, while this work provides valuable insights into the use of shallow learning classifiers for histopathological image classification with features supplied by non-fine-tuned methods, the limitations discussed here suggest avenues for further research to enhance and validate the findings.

6. Conclusions

This study comprehensively evaluates shallow learning classifiers for histopathological image classification using HC and deep features.
The comparative analysis of HC versus deep features demonstrates the clear superiority of deep features, particularly those extracted from pre-trained CNNs such as DenseNet-201 and EfficientNetB0. These features consistently outperform HC features, highlighting the advanced feature extraction capability of DL models in complex image classification tasks. Among HC features, LBP shows robust performance, while Haar features are less effective.
Our exploration of feature fusion techniques shows that combining features can significantly enhance classification performance. The SVM classifier, in particular, excels in handling diverse and complex feature sets, outperforming both DT and RF classifiers across various combinations. The fusion of LBP, DenseNet-201, and EfficientNetB0 features emerges as the most reliable strategy, leveraging the strengths of both HC and deep features.
In addition, our cross-magnification experiments underscore the robustness of RF classifiers, which consistently perform well across different image resolutions. The combination of features from DenseNet-201 and EfficientNetB0 proves effective in maintaining high classification accuracy, demonstrating the feasibility of utilizing features from pre-trained CNNs and HC methods.
While this study advances our understanding of shallow learning classifiers in histopathological image classification, our results open the field for several future works. For instance, with feature importance and explainability techniques, we plan to investigate the most effective features for the classifiers’ final prediction to simplify the workflow with feature selection strategies. Moreover, we aim to integrate further DL methods like Vision Transformer and leverage them as feature extractors in this context.

Author Contributions

Conceptualization, A.L. and C.D.R.; methodology, M.U., A.L. and C.D.R.; investigation, M.U., A.L. and C.D.R.; software, M.U. and A.L.; writing—original draft, M.U. and A.L.; writing—review and editing, M.U., A.L. and C.D.R.; supervision, A.L. and C.D.R. All authors have read and agreed to the published version of the manuscript.

Funding

We acknowledge financial support under the National Recovery and Resilience Plan (NRRP), Mission 4 Component 2 Investment 1.5—Call for tender No. 3277 published on 30 December 2021 by the Italian Ministry of University and Research (MUR) funded by the European Union—NextGenerationEU. Project Code ECS0000038—Project Title eINS Ecosystem of Innovation for Next Generation Sardinia—CUP F53C22000430001-Grant Assignment Decree No. 1056 adopted on 23 June 2022 by the Italian Ministry of University and Research (MUR) and by the project DEMON “Detect and Evaluate Manipulation of ONline information” funded by MIUR under the PRIN 2022 grant 2022BAXSPY (CUP F53D23004270006, NextGenerationEU).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the material used and developed for this work is available at the following GitHub repository: https://github.com/MurkoZawa/HistopathologyClassification (accessed on 2 July 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this document:
MLMachine Learning
DLDeep Learning
HCHandcrafted
CNNConvolutional Neural Network
GCGastric Cancer
EGCEarly-Stage Gastric Cancer
AGCAdvanced Gastric Cancer
CVComputer Vision
WSIWhole Slide Image
LGFFNLightweight Gated Fully Fused Network
GHIGated Hybrid Input
CHChebyshev Moment
LMSecond-Order Legendre Moment
ZMZernike Moment
HARRotation-Invariant Haralick
LBPLocal Binary Pattern
HistHistogram
ACAutocorrelogram
HaarHaar-Like
DTDecision Tree
kNNk-Nearest Neighbor
SVMSupport Vector Machine
RFRandom Forest
TNTrue Negative
FPFalse Positive
FNFalse Negative
TPTrue Positive
AAccuracy
PPrecision
RRecall
SSpecificity
FF1-Score
MCCMatthews Correlation Coefficient
BACCBalanced Accuracy
CvalCross-Validation

Appendix A. Further Results

Table A1. Performance obtained with DT trained with HC features. Values are shown in terms of %.
Table A1. Performance obtained with DT trained with HC features. Values are shown in terms of %.
DescriptorAPRSF1MCCBACC
AC62.7869.3968.9753.2669.1822.2061.12
Haar59.7062.5683.3123.4371.468.3353.37
Hist68.7874.4973.7161.2274.1034.8467.46
HAR68.7874.8572.9962.3273.9135.1067.66
LBP71.2276.2376.2663.4776.2539.7469.87
CH_171.1176.1576.1763.3576.1639.5269.76
CH_271.0576.1276.0763.3576.0939.4169.71
LM71.3276.1676.6463.1676.4039.8769.90
ZM58.0665.7464.2448.5764.9812.7356.40

Appendix B

Table A2. Performance obtained with kNN trained with HC features. Values are shown in terms of %.
Table A2. Performance obtained with kNN trained with HC features. Values are shown in terms of %.
DescriptorAPRSF1MCCBACC
AC57.3470.0451.6666.0659.4617.4258.86
Haar42.1756.9918.4078.6727.82−3.6148.53
Hist64.6471.9768.1559.2470.0127.0763.70
HAR61.0668.5366.0553.4167.2619.2959.73
LBP69.5175.3273.8662.8274.5836.5068.34
CH_166.1672.4171.2858.2971.8429.4564.78
CH_265.4672.0970.1458.2971.1028.2464.21
LM66.3272.5771.3858.5571.9729.8164.97
ZM57.4065.0364.1946.9764.6011.1255.58
Table A3. Performance obtained with DT trained with deep features. Values are shown in terms of %.
Table A3. Performance obtained with DT trained with deep features. Values are shown in terms of %.
DescriptorAPRSF1MCCBACC
AlexNet75.0079.4779.1968.5779.3347.7273.88
DarkNet-1978.2582.0482.0472.4282.0454.4677.23
DarkNet-5381.6484.7884.9576.5784.8661.5580.76
DenseNet-20184.9287.5187.6080.8087.5668.4284.20
EfficientNet B078.7982.6282.2973.4182.4655.6477.85
Inception-v374.5479.2178.6068.3078.9046.8173.45
Inception-ResNet-v273.5278.0278.3566.1078.1844.4972.22
ResNet-1876.7381.2080.1371.5080.6651.4675.82
ResNet-5081.3084.6284.4776.4284.5560.8780.45
ResNet-10180.1383.6783.4874.9783.5858.4279.23
VGG1976.4080.9779.7971.2080.3750.8075.49
Xception79.3883.3082.4974.5982.8956.9478.54
Table A4. Performance obtained with kNN trained with deep features. Values are shown in terms of %.
Table A4. Performance obtained with kNN trained with deep features. Values are shown in terms of %.
DescriptorAPRSF1MCCBACC
AlexNet80.8583.8084.7774.8284.2859.7879.80
DarkNet-1983.9986.4787.2079.0586.8466.4083.13
DarkNet-5388.2588.8992.1182.3290.4875.2587.22
DenseNet-20188.2189.0491.8482.6390.4275.1687.23
EfficientNet B087.5389.0190.6082.8289.8073.7986.71
Inception-ResNet-v276.1379.8980.9868.6980.4349.8574.83
Inception-v380.8883.6585.0474.4884.3459.8079.76
ResNet-10187.8988.5791.8781.7990.1974.4886.83
ResNet-1884.2185.8988.4777.6887.1666.7383.07
ResNet-5086.9788.2091.0980.4189.6273.1985.75
VGG1983.3585.9486.9477.9586.4465.8882.45
Xception85.1586.7489.4478.9588.0770.1784.20

References

  1. Ilic, M.; Ilic, I. Epidemiology of stomach cancer. World J. Gastroenterol. 2022, 28, 1187. [Google Scholar] [CrossRef] [PubMed]
  2. Hu, W.; Li, C.; Li, X.; Rahaman, M.M.; Ma, J.; Zhang, Y.; Chen, H.; Liu, W.; Sun, C.; Yao, Y.; et al. GasHisSDB: A new gastric histopathology image dataset for computer aided diagnosis of gastric cancer. Comput. Biol. Med. 2022, 142, 105207. [Google Scholar] [CrossRef] [PubMed]
  3. Hirasawa, T.; Aoyama, K.; Tanimoto, T.; Ishihara, S.; Shichijo, S.; Ozawa, T.; Ohnishi, T.; Fujishiro, M.; Matsuo, K.; Fujisaki, J.; et al. Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images. Gastric Cancer 2018, 21, 653–660. [Google Scholar] [CrossRef] [PubMed]
  4. Zhao, Y.; Hu, B.; Wang, Y.; Yin, X.; Jiang, Y.; Zhu, X. Identification of gastric cancer with convolutional neural networks: A systematic review. Multim. Tools Appl. 2022, 81, 11717–11736. [Google Scholar] [CrossRef] [PubMed]
  5. Xie, K.; Peng, J. Deep learning-based gastric cancer diagnosis and clinical management. J. Radiat. Res. Appl. Sci. 2023, 16, 100602. [Google Scholar] [CrossRef]
  6. Yong, M.P.; Hum, Y.C.; Lai, K.W.; Lee, Y.L.; Goh, C.H.; Yap, W.S.; Tee, Y.K. Histopathological gastric cancer detection on GasHisSDB dataset using deep ensemble learning. Diagnostics 2023, 13, 1793. [Google Scholar] [CrossRef] [PubMed]
  7. Yoon, H.J.; Kim, S.; Kim, J.H.; Keum, J.S.; Oh, S.I.; Jo, J.; Chun, J.; Youn, Y.H.; Park, H.; Kwon, I.G.; et al. A lesion-based convolutional neural network improves endoscopic detection and depth prediction of early gastric cancer. J. Clin. Med. 2019, 8, 1310. [Google Scholar] [CrossRef] [PubMed]
  8. Hu, W.; Chen, H.; Liu, W.; Li, X.; Sun, H.; Huang, X.; Grzegorzek, M.; Li, C. A comparative study of gastric histopathology sub-size image classification: From linear regression to visual transformer. Front. Med. 2022, 9, 1072109. [Google Scholar] [CrossRef] [PubMed]
  9. Zhang, K.; Wang, H.; Cheng, Y.; Liu, H.; Gong, Q.; Zeng, Q.; Zhang, T.; Wei, G.; Wei, Z.; Chen, D. Early gastric cancer detection and lesion segmentation based on deep learning and gastroscopic images. Sci. Rep. 2024, 14, 7847. [Google Scholar]
  10. Marini, N.; Otálora, S.; Podareanu, D.; van Rijthoven, M.; van der Laak, J.; Ciompi, F.; Müller, H.; Atzori, M. Multi_Scale_Tools: A Python Library to Exploit Multi-Scale Whole Slide Images. Front. Comput. Sci. 2021, 3, 684521. [Google Scholar] [CrossRef]
  11. Ashtaiwi, A. Optimal Histopathological Magnification Factors for Deep Learning-Based Breast Cancer Prediction. Appl. Syst. Innov. 2022, 5, 87. [Google Scholar] [CrossRef]
  12. Cao, R.; Tang, L.; Fang, M.; Zhong, L.; Wang, S.; Gong, L.; Li, J.; Dong, D.; Tian, J. Artificial intelligence in gastric cancer: Applications and challenges. Gastroenterol. Rep. 2022, 10, 64. [Google Scholar] [CrossRef]
  13. Hu, W.; Li, C.; Rahaman, M.M.; Chen, H.; Liu, W.; Yao, Y.; Sun, H.; Grzegorzek, M.; Li, X. EBHI: A new Enteroscope Biopsy Histopathological H&E Image Dataset for image classification evaluation. Phys. Medica 2023, 107, 102534. [Google Scholar]
  14. Li, S.; Liu, W. LGFFN-GHI: A Local-Global Feature Fuse Network for Gastric Histopathological Image Classification. J. Comput. Commun. 2022, 10, 91–106. [Google Scholar] [CrossRef]
  15. Putzu, L.; Loddo, A.; Ruberto, C.D. Invariant Moments, Textural and Deep Features for Diagnostic MR and CT Image Retrieval. In Proceedings of the 19th International Conference of Computer Analysis of Images and Patterns, CAIP 2021, Virtual, 28–30 September 2021; Proceedings, Part I; Lecture Notes in Computer Science. Tsapatsoulis, N., Panayides, A., Theocharides, T., Lanitis, A., Pattichis, C.S., Vento, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2021; Volume 13052, pp. 287–297. [Google Scholar] [CrossRef]
  16. Ruberto, C.D.; Loddo, A.; Putzu, L. On The Potential of Image Moments for Medical Diagnosis. J. Imaging 2023, 9, 70. [Google Scholar] [CrossRef]
  17. Mukundan, R.; Ong, S.H.; Lee, P.A. Image analysis by Tchebichef moments. IEEE Trans. Image Process. 2001, 10, 1357–1364. [Google Scholar] [CrossRef] [PubMed]
  18. Ruberto, C.D.; Putzu, L.; Rodriguez, G. Fast and accurate computation of orthogonal moments for texture analysis. Pattern Recognit. 2018, 83, 498–510. [Google Scholar] [CrossRef]
  19. Teh, C.; Chin, R.T. On Image Analysis by the Methods of Moments. IEEE Trans. Pattern Anal. Mach. Intell. 1988, 10, 496–513. [Google Scholar] [CrossRef]
  20. Teague, M.R. Image analysis via the general theory of moments. J. Opt. Soc. Am. 1980, 70, 920–930. [Google Scholar] [CrossRef]
  21. Wee, C.; Raveendran, P. On the computational aspects of Zernike moments. Image Vis. Comput. 2007, 25, 967–980. [Google Scholar] [CrossRef]
  22. Mirjalili, F.; Hardeberg, J.Y. On the Quantification of Visual Texture Complexity. J. Imaging 2022, 8, 248. [Google Scholar] [CrossRef] [PubMed]
  23. Putzu, L.; Ruberto, C.D. Rotation Invariant Co-occurrence Matrix Features. In Proceedings of the 19th International Conference of Image Analysis and Processing, ICIAP 2017, Catania, Italy, 11–15 September 2017; Proceedings, Part I; Lecture Notes in Computer Science. Battiato, S., Gallo, G., Schettini, R., Stanco, F., Eds.; Springer: Berlin/Heidelberg, Germany, 2017; Volume 10484, pp. 391–401. [Google Scholar] [CrossRef]
  24. He, D.C.; Wang, L. Texture unit, texture spectrum, and texture analysis. IEEE Trans. Geosci. Remote. Sens. 1990, 28, 509–512. [Google Scholar]
  25. Ojala, T.; Pietikäinen, M.; Mäenpää, T. Multiresolution Gray-Scale and Rotation Invariant Texture Classification with Local Binary Patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  26. Van de Weijer, J.; Schmid, C. Coloring Local Feature Extraction. In Proceedings of the 9th European Conference on Computer Vision, ECCV 2006, Graz, Austria, 7–13 May 2006; Proceedings, Part II; Lecture Notes in Computer Science. Leonardis, A., Bischof, H., Pinz, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; Volume 3952, pp. 334–348. [Google Scholar] [CrossRef]
  27. Huang, J.; Kumar, R.; Mitra, M.; Zhu, W.; Zabih, R. Image Indexing Using Color Correlograms. In Proceedings of the 1997 IEEE Conference on Computer Vision and Pattern Recognition (CVPR ’97), San Juan, Puerto Rico, 17–19 June 1997; IEEE Computer Society: Piscataway, NJ, USA, 1997; pp. 762–768. [Google Scholar] [CrossRef]
  28. Viola, P.A.; Jones, M.J. Rapid Object Detection using a Boosted Cascade of Simple Features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, (CVPR 2001), Kauai, HI, USA, 8–14 December 2001; IEEE Computer Society: Piscataway, NJ, USA, 2001; pp. 511–518. [Google Scholar] [CrossRef]
  29. Bodapati, J.D.; Veeranjaneyulu, N. Feature Extraction and Classification UsingDeep Convolutional Neural Networks. J. Cyber Secur. Mobil. 2019, 8, 261–276. [Google Scholar] [CrossRef]
  30. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, (CVPR 2009), Miami, FL, USA, 20–25 June 2009; IEEE Computer Society: Piscataway, NJ, USA, 2009; pp. 248–255. [Google Scholar] [CrossRef]
  31. Putzu, L.; Piras, L.; Giacinto, G. Convolutional neural networks for relevance feedback in content based image retrieval. Multim. Tools Appl. 2020, 79, 26995–27021. [Google Scholar] [CrossRef]
  32. Wang, H.; Wu, X.; Huang, Z.; Xing, E.P. High-Frequency Component Helps Explain the Generalization of Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; IEEE Computer Vision Foundation: Piscataway, NJ, USA, 2020; pp. 8681–8691. [Google Scholar] [CrossRef]
  33. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  34. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Piscataway, NJ, USA, 2017; pp. 6517–6525. [Google Scholar] [CrossRef]
  35. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  36. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Piscataway, NJ, USA, 2017; pp. 2261–2269. [Google Scholar] [CrossRef]
  37. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA, 9–15 June 2019; Chaudhuri, K., Salakhutdinov, R., Eds.; PMLR: New York, NY, USA, 2019; Volume 97, pp. 6105–6114. [Google Scholar]
  38. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; IEEE Computer Society: Piscataway, NJ, USA, 2016; pp. 2818–2826. [Google Scholar] [CrossRef]
  39. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Singh, S., Markovitch, S., Eds.; AAAI Press: Washington, DC, USA, 2017; pp. 4278–4284. [Google Scholar] [CrossRef]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; IEEE Computer Society: Piscataway, NJ, USA, 2016; pp. 770–778. [Google Scholar] [CrossRef]
  41. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015; Conference Track Proceedings. Bengio, Y., LeCun, Y., Eds.; ACM: New York, NY, USA, 2015. [Google Scholar]
  42. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Piscataway, NJ, USA, 2017; pp. 1800–1807. [Google Scholar] [CrossRef]
  43. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 6–11 July 2015; Bach, F.R., Blei, D.M., Eds.; JMLR Workshop and Conference Proceedings. JMLR: Cambridge, MA, USA, 2015; Volume 37, pp. 448–456. [Google Scholar]
  44. Quinlan, J.R. Learning efficient classification procedures and their application to chess end games. In Machine Learning; Springer: Berlin/Heidelberg, Germany, 1983; pp. 463–482. [Google Scholar]
  45. Cover, T.M.; Hart, P.E. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  46. Lin, Y.; Lv, F.; Zhu, S.; Yang, M.; Cour, T.; Yu, K.; Cao, L.; Huang, T.S. Large-scale image classification: Fast feature extraction and SVM training. In Proceedings of the The 24th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; IEEE Computer Society: Piscataway, NJ, USA, 2011; pp. 1689–1696. [Google Scholar]
  47. Breiman, L. Random Forests. Mach. Learn. 2001, 4, 5–32. [Google Scholar] [CrossRef]
  48. Springenberg, M.; Frommholz, A.; Wenzel, M.; Weicken, E.; Ma, J.; Strodthoff, N. From modern CNNs to vision transformers: Assessing the performance, robustness, and classification strategies of deep learning models in histopathology. Med. Image Anal. 2023, 87, 102809. [Google Scholar] [CrossRef]
  49. Fu, X.; Liu, S.; Li, C.; Sun, J. MCLNet: An multidimensional convolutional lightweight network for gastric histopathology image classification. Biomed. Signal Process. Control. 2023, 80, 104319. [Google Scholar] [CrossRef]
  50. Song, Y.; Wang, T.; Cai, P.; Mondal, S.K.; Sahoo, J.P. A comprehensive survey of few-shot learning: Evolution, applications, challenges, and opportunities. ACM Comput. Surv. 2023, 55, 1–40. [Google Scholar] [CrossRef]
Figure 1. Sample images from the GasHisSDB dataset, acquired with the H&E staining method. The hematoxylin is alkaline, and stains cell nuclei a purplish blue, and eosin is acidic and stains the extracellular matrix and cytoplasm pink, with other structures taking on different shades, hues, and combinations of these colors [2].
Figure 1. Sample images from the GasHisSDB dataset, acquired with the H&E staining method. The hematoxylin is alkaline, and stains cell nuclei a purplish blue, and eosin is acidic and stains the extracellular matrix and cytoplasm pink, with other structures taking on different shades, hues, and combinations of these colors [2].
Jimaging 10 00195 g001
Table 1. Description of GasHisSDB with details on its subdivision and number of images per class.
Table 1. Description of GasHisSDB with details on its subdivision and number of images per class.
Sub-DatabaseSizeAbnormalNormal
S-A 160 × 160 pixels13,12420,160
S-B 120 × 120 pixels24,80140,460
S-C 80 × 80 pixels59,15187,500
Total 97,076148,120
Table 2. Employed CNN details including reference paper, number of trainable parameters in millions, input shape, feature extraction layer, and related feature vector size.
Table 2. Employed CNN details including reference paper, number of trainable parameters in millions, input shape, feature extraction layer, and related feature vector size.
CNNParameters (M)Input ShapeFeature Layer# of Features
AlexNet [33]60 224 × 224 Pen. FC4096
DarkNet-19 [34]20.8 224 × 224 Conv191000
DarkNet-53 [35]20.8 224 × 224 Conv531000
DenseNet-201 [36]25.6 224 × 224 Avg. Pool1920
EfficientNetB0 [37]5.3 224 × 224 Avg. Pool1280
Inception-v3 [38]21.8 299 × 299 Last FC1000
Inception-ResNet-v2 [39]55 299 × 299 Avg. Pool1536
ResNet-18 [40]11.7 224 × 224 Pool5512
ResNet-50 [40]26 224 × 224 Avg. Pool1024
ResNet-101 [40]44.6 224 × 224 Pool51024
VGG19 [41]144 224 × 224 Pen. FC4096
XceptionNet [42]22.9 299 × 299 Avg. Pool2048
Table 3. Performance obtained with SVM trained with HC features. Values are shown in terms of %.
Table 3. Performance obtained with SVM trained with HC features. Values are shown in terms of %.
DescriptorAPRSF1MCCBACC
AC67.3069.1882.9943.2075.4528.7163.09
Haar62.1862.3894.5912.3875.1812.4553.49
Hist44.6796.299.0099.4716.4717.9154.23
HAR62.0773.6458.2168.0065.0225.6463.10
LBP62.6470.0067.0655.8568.5022.6961.46
CH_175.9277.3785.1461.7581.0748.6173.45
CH_272.5071.5790.5544.7679.9540.7867.66
LM73.3272.6589.7348.1180.2942.6168.92
ZM63.8469.7871.0852.7270.4323.9361.90
Table 4. Performance obtained with RF trained with HC features. Values are shown in terms of %.
Table 4. Performance obtained with RF trained with HC features. Values are shown in terms of %.
DescriptorAPRSF1MCCBACC
AC71.7672.8984.9751.4778.4739.0968.22
Haar62.4862.6594.2213.7175.2613.6153.97
Hist73.2677.2479.1964.1578.2043.6671.67
HAR76.7878.6984.5564.8481.5250.6374.69
LBP79.5780.7487.0368.1183.7756.6177.57
CH_178.1179.9985.1767.2882.5053.5676.22
CH_278.0779.8485.3466.9082.5053.4376.12
LM78.2580.2085.0967.7382.5853.8776.41
ZM65.1968.1679.8442.7073.5424.2661.27
Table 5. Performance obtained with SVM trained with deep features. Values are shown in terms of %.
Table 5. Performance obtained with SVM trained with deep features. Values are shown in terms of %.
DescriptorAPRSF1MCCBACC
AlexNet72.9777.4976.5665.8177.0242.8071.18
DarkNet-1977.8682.2080.8271.9581.5053.2876.39
DarkNet-5382.8385.8484.9976.8885.4163.6580.94
DenseNet-20186.0289.0186.6181.8487.7969.9184.23
EfficientNet B082.8485.7685.3576.9685.5563.4881.15
Inception-v373.6078.8874.7367.5176.7545.8671.12
Inception-ResNet-v269.8775.0371.6562.7873.2935.6867.21
ResNet-1877.7882.1780.8271.8681.4953.1376.34
ResNet-5082.7786.1184.3477.8785.2263.4381.11
ResNet-10182.5285.7684.7176.8185.2363.0080.76
VGG1979.6083.9481.7873.7082.8457.6777.74
XceptionNet82.2485.7584.2276.2284.9762.5380.22
Table 6. Performance obtained with RF trained with deep features. Values are shown in terms of %.
Table 6. Performance obtained with RF trained with deep features. Values are shown in terms of %.
DescriptorAPRSF1MCCBACC
AlexNet84.0285.5588.5777.0387.0366.2982.80
DarkNet-1988.3088.6892.4981.8790.5475.3387.18
DarkNet-5390.3090.7293.5585.3092.1179.5889.42
DenseNet-20191.9392.6194.2088.4693.4083.0591.33
EfficientNet B089.8989.9693.7783.9291.8378.7188.85
Inception-v385.5285.6491.4276.4688.4469.3983.94
Inception-ResNet-v283.2584.1089.2174.1086.5864.5581.65
ResNet-1886.9987.3291.8779.5089.5372.5485.68
ResNet-5089.9290.1293.6384.2391.8478.7788.93
ResNet-10189.5989.7693.4883.6291.5878.0788.55
VGG1985.9886.6190.9278.4088.7170.4184.66
Xception88.5889.0892.4982.5990.7575.9487.54
Table 7. Performance measures of different classifiers trained with a feature fusion strategy. The classifiers used are DT, SVM, and RF. The strategies compared include combinations of HC and deep features: LBP + DenseNet-201, LBP + EfficientNetB0, DenseNet-201 + EfficientNetB0, and LBP + DenseNet-201 + EfficientNetB0. Values are shown in terms of %.
Table 7. Performance measures of different classifiers trained with a feature fusion strategy. The classifiers used are DT, SVM, and RF. The strategies compared include combinations of HC and deep features: LBP + DenseNet-201, LBP + EfficientNetB0, DenseNet-201 + EfficientNetB0, and LBP + DenseNet-201 + EfficientNetB0. Values are shown in terms of %.
StrategyClassifierAPRSF1MCCBACC
LBP+
DenseNet-201
DT88.2189.0491.8482.6390.4275.1687.23
SVM94.4195.1495.6692.5095.4088.2994.08
RF92.1692.8394.3588.8093.5883.5391.57
LBP+
EfficientNetB0
DT87.5589.0190.6382.8289.8173.8286.72
SVM94.0594.7895.4491.9295.1187.5393.68
RF89.6590.1993.0384.4691.5978.2188.74
DenseNet-201 +
EfficientNetB0
DT90.3091.0593.1385.9492.0879.5989.54
SVM94.8995.7695.8193.4995.7889.3194.65
RF91.8392.3194.3787.9293.3382.8291.15
LBP +
DenseNet-201 +
EfficientNetB0
DT90.3191.0793.1385.9892.0979.6389.56
SVM95.0395.8695.9393.6495.9089.5994.79
RF92.2692.6794.7288.5093.6883.7491.61
Table 8. Performance measures of the first cross-magnification experiment. Different classifiers were trained on the 160 × 160 split of the GasHisSDB with a feature fusion strategy and tested on the 120 × 120 split. The classifiers used were DT, SVM, and RF. Values are shown in terms of %.
Table 8. Performance measures of the first cross-magnification experiment. Different classifiers were trained on the 160 × 160 split of the GasHisSDB with a feature fusion strategy and tested on the 120 × 120 split. The classifiers used were DT, SVM, and RF. Values are shown in terms of %.
StrategyClassifierAPRSF1MCCBACC
LBP +
DenseNet-201
DT86.6990.5687.6885.0889.0972.1086.38
SVM86.4197.4280.2096.5387.9874.5188.37
RF89.0495.1286.7892.7490.7677.8789.76
LBP +
EfficientNetB0
DT85.1788.6687.2581.7987.9568.7184.52
SVM85.0296.8178.4295.7986.6572.0487.10
RF87.3890.2089.3684.1589.7873.3086.76
DenseNet-201 +
EfficientNetB0
DT88.4391.6589.5086.6990.5675.6688.09
SVM85.8898.4078.5097.9287.3374.1988.21
RF89.5594.3688.4391.3791.3078.5189.90
LBP +
DenseNet-201 +
EfficientNetB0
DT88.4291.6489.5086.6790.5575.6588.08
SVM85.8298.4578.3697.9887.2674.1288.17
RF89.5694.7987.9992.1291.2678.6790.05
Table 9. Performance measures of the second cross-magnification experiment. Different classifiers were trained on the 160 × 160 split of the GasHisSDB with a feature fusion strategy and tested on the 80 × 80 split. The classifiers used were DT, SVM, and RF. Values are shown in terms of %.
Table 9. Performance measures of the second cross-magnification experiment. Different classifiers were trained on the 160 × 160 split of the GasHisSDB with a feature fusion strategy and tested on the 80 × 80 split. The classifiers used were DT, SVM, and RF. Values are shown in terms of %.
StrategyClassifierAPRSF1MCCBACC
LBP +
DenseNet-201
DT77.0586.2373.2382.7079.2054.8977.97
SVM68.9296.1849.8997.0765.7049.8373.48
RF78.8993.2969.6292.6079.7361.4181.11
LBP +
EfficientNetB0
DT63.5889.1944.3492.0559.2339.0968.20
SVM54.3696.0424.5398.5039.0731.4461.51
RF71.3892.2756.7992.9670.3150.6374.88
DenseNet-201 +
EfficientNetB0
DT74.7088.6966.0287.5575.6952.8976.78
SVM59.9696.8134.0298.3450.3439.0066.18
RF79.7392.8471.5491.8480.8162.3981.69
LBP +
DenseNet-201 +
EfficientNetB0
DT74.7088.6966.0287.5575.6952.8976.78
SVM60.3196.7234.6698.2651.0339.3966.46
RF78.4493.8868.3393.4179.0961.1080.87
Table 10. Performance comparison of our work and the previous state-of-the-art works on the GasHisSDB dataset. * indicates that the proposed approach was trained on S-A and directly tested on S-B and S-C without fine-tuning.
Table 10. Performance comparison of our work and the previous state-of-the-art works on the GasHisSDB dataset. * indicates that the proposed approach was trained on S-A and directly tested on S-B and S-C without fine-tuning.
WorkSplit (%)Model DetailsA (%)
S-CS-BS-A
[2]40/40/20VGG1696.1296.4795.90
40/40/20ResNet5096.0995.9496.09
[48]40/20/40InceptionV3 trained from scratch--98.83
40/20/40InceptionV3 + ResNet50 (feature concatenation)--98.80
[14]60/20/20LGFFN--96.81
[49]80/-/20MCLNet based on ShuffleNetV296.2897.9597.85
[6]40/20/40Ensemble97.7298.6899.20
Ours5-fold CValSVM with feature fusion60.31 *85.82 *95.03
Ours5-fold CValRF with feature fusion78.44 *89.56 *92.26
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Loddo, A.; Usai, M.; Di Ruberto, C. Gastric Cancer Image Classification: A Comparative Analysis and Feature Fusion Strategies. J. Imaging 2024, 10, 195. https://doi.org/10.3390/jimaging10080195

AMA Style

Loddo A, Usai M, Di Ruberto C. Gastric Cancer Image Classification: A Comparative Analysis and Feature Fusion Strategies. Journal of Imaging. 2024; 10(8):195. https://doi.org/10.3390/jimaging10080195

Chicago/Turabian Style

Loddo, Andrea, Marco Usai, and Cecilia Di Ruberto. 2024. "Gastric Cancer Image Classification: A Comparative Analysis and Feature Fusion Strategies" Journal of Imaging 10, no. 8: 195. https://doi.org/10.3390/jimaging10080195

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop