Next Article in Journal
The Impact of Histologic Portal T-Cell Density on the Clinical Outcomes in Hepatic Graft-versus-Host Disease and Autoimmune Liver Diseases
Previous Article in Journal
Performance Evaluation of Open Channel Buhlmann Fecal Calprotectin Turbo Assay on Abbott Alinity C Analyzer
Previous Article in Special Issue
Brain and Ventricle Volume Alterations in Idiopathic Normal Pressure Hydrocephalus Determined by Artificial Intelligence-Based MRI Volumetry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Gastric Cancer Detection with Ensemble Learning on Digital Pathology: Use Case of Gastric Cancer on GasHisSDB Dataset

1
College of Engineering, Northeastern University, Boston, MA 02115, USA
2
Khoury College of Computer Sciences, Northeastern University, Boston, MA 02115, USA
3
MaineHealth Institute for Research, Scarborough, ME 04074, USA
4
Maine Medical Center, Portland, ME 04102, USA
5
The Roux Institute, Department of Bioengineering, College of Engineering at Northeastern University, Boston, MA 02115, USA
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(16), 1746; https://doi.org/10.3390/diagnostics14161746 (registering DOI)
Submission received: 12 July 2024 / Revised: 1 August 2024 / Accepted: 2 August 2024 / Published: 12 August 2024

Abstract

:

Simple Summary

Gastric cancer is a major worldwide health concern, underscoring the importance of early detection to enhance patient outcomes. Traditional histological analysis, while considered the gold standard, is labour intensive and manual. Deep learning (DL) is a potential approach, but existing models fail to extract all of the visual data required for successful categorization. This work overcomes these constraints by using ensemble models that mix different deep-learning architectures to improve classification performance for stomach cancer diagnosis. Using the Gastric Histopathology Sub-Size Images Database, the ensemble models obtained an average accuracy of more than 99% at various resolutions. ResNet50, VGGNet, and ResNet34 performed better than EfficientNet and VitNet, with the ensemble model continuously delivering higher accuracy. These findings show that ensemble models may accurately detect important characteristics from smaller picture patches, allowing pathologists to diagnose stomach cancer early and increasing patient survival rates.

Abstract

Gastric cancer has become a serious worldwide health concern, emphasizing the crucial importance of early diagnosis measures to improve patient outcomes. While traditional histological image analysis is regarded as the clinical gold standard, it is labour intensive and manual. In recognition of this problem, there has been a rise in interest in the use of computer-aided diagnostic tools to help pathologists with their diagnostic efforts. In particular, deep learning (DL) has emerged as a promising solution in this sector. However, current DL models are still restricted in their ability to extract extensive visual characteristics for correct categorization. To address this limitation, this study proposes the use of ensemble models, which incorporate the capabilities of several deep-learning architectures and use aggregate knowledge of many models to improve classification performance, allowing for more accurate and efficient gastric cancer detection. To determine how well these proposed models performed, this study compared them with other works, all of which were based on the Gastric Histopathology Sub-Size Images Database, a publicly available dataset for gastric cancer. This research demonstrates that the ensemble models achieved a high detection accuracy across all sub-databases, with an average accuracy exceeding 99%. Specifically, ResNet50, VGGNet, and ResNet34 performed better than EfficientNet and VitNet. For the 80 × 80-pixel sub-database, ResNet34 exhibited an accuracy of approximately 93%, VGGNet achieved 94%, and the ensemble model excelled with 99%. In the 120 × 120-pixel sub-database, the ensemble model showed 99% accuracy, VGGNet 97%, and ResNet50 approximately 97%. For the 160 × 160-pixel sub-database, the ensemble model again achieved 99% accuracy, VGGNet 98%, ResNet50 98%, and EfficientNet 92%, highlighting the ensemble model’s superior performance across all resolutions. Overall, the ensemble model consistently provided an accuracy of 99% across the three sub-pixel categories. These findings show that ensemble models may successfully detect critical characteristics from smaller patches and achieve high performance. The findings will help pathologists diagnose gastric cancer using histopathological images, leading to earlier identification and higher patient survival rates.

1. Introduction

The GI tract, spanning 25 feet from the oral cavity to the anus, transports ingested substances. The digestive process begins with the oesophagus and continues through the stomach and small intestines, extracting important nutrients. Waste is then eliminated through the colon and rectum [1,2]. Tumours in these organs often result from aberrant cell growth caused by DNA changes [3]. Mutations can be caused by a variety of reasons, including health conditions, genetics, or lifestyle. The uncontrolled proliferation of malignant cells in the gastrointestinal system is caused by genetic, environmental, and lifestyle factors that interact. Common gastrointestinal (GI) cancers include oesophageal cancer, colorectal cancer, gastric cancer, bile duct cancer, anal cancer, colon cancer, gallbladder cancer, pancreatic cancer, gastrointestinal stromal tumours, liver cancer, rectal cancer, gastric cancer, and small intestine cancer. In 2020, gastric cancer was one of the top three most common cancers in 19 nations, with around 1.1 million cases reported (720,000 men, 370,000 females). Early identification of GI cancer aids in cancer treatment and reduces health-related complications.
Traditional approaches to the identification of cancer include the estimation of body fat percentage and its subsequent correlation with cancer. Other methods include identifying common microbes associated with cancer in food. Another way is the use of Indocyanine Green (ICG) in gastrointestinal surgery, which is gaining popularity, particularly for lymph node diagnosis and operative field imaging [4,5].
However, pathologists must physically assess tissue samples, which is a tough, time-consuming, and subjective procedure. Moreover, different pathologists may provide different results, making the analysis susceptible to errors. The accuracy of histopathological analysis is heavily dependent on the pathologists’ experience and knowledge, making the manual process susceptible to mistakes such as incorrect detection and diagnosis. Furthermore, a scarcity of pathologists creates significant delays in examining patient cases, potentially leading to late cancer discovery [6,7].
Various computer-aided detection (CAD) strategies have been investigated for the diagnosis of gastric cancer utilizing histopathological imaging. For more than 30 years, researchers have studied computer-aided diagnosis in gastroenterology, creating datasets from endoscopic images using various methodologies. The most widely researched issue is the identification of aberrant pathological signs in a specific location of the GI tract, notably polyps. There has also been research on the detection and categorization of disorders throughout the GI system, which has included clinical findings, anatomical markers, and therapies [8]. Figure 1 shows the steps of extraction of the histopathological image from slides.

2. Materials and Methods

2.1. Literature Reviews

Various computer-aided detection (CAD) strategies have been investigated for the diagnosis of gastric cancer utilizing histopathological imaging. For more than 30 years, researchers have studied computer-aided diagnosis in gastroenterology, creating datasets from endoscopic images using various methodologies. The most widely researched issue is the identification of aberrant pathological signs in a specific location of the GI tract, notably polyps. There has also been research on the detection and categorization of disorders throughout the GI system, which has included clinical findings, anatomical markers, and therapies [8]. During the first 20 years of development, image processing required the extraction of features using various approaches before their categorization using statistical methods [9]. These characteristics are divided into three categories: spatial, frequency, and high-level descriptors. Spatial characteristics are retrieved using pixel-based and histogram techniques, whilst frequency information is collected using Fourier and wavelet transform algorithms. High-level characteristics are retrieved with edge and region-based methods. Statistical machine learning approaches were frequently employed to categorize these characteristics.
Later, machine learning (ML) was to become commonly used in CAD to diagnose gastric cancer by extracting handmade elements such as colour, texture, and form. For this purpose, support vector machines (SVM), random forests, and Adaboost are among the most frequently employed machine learning classifiers. In recent research, deep learning has been utilized to automate the feature selection process. Several studies have shown that deep convolutional neural networks (CNN) excel at tasks such as recognizing and segmenting histopathological images related to cancer, metastasis, and genetic mutation analysis. Some investigations have indicated that these networks perform similarly to human pathologists [10,11].
Deep learning techniques have advanced significantly over the past decade, notably with the CNN architecture. This design allows for the extraction and categorization of spatial as well as high-level features, making it a major area of study for academia. Several solutions have been proposed, including hybrid approaches based on CNN attributes, transfer learning, the development of novel CNN models, and research into other deep learning networks [12].
Gastroenterology CAD research has a long history and covers a wide range of topics. For studies involving the classification of the KvasirV2 and HayperKvasir datasets, which were employed in the studies conducted [9,12], Melaku et al. (2019) utilized VGGNet and InceptionV3 to classify the Hyper KVASIR dataset with 98% accuracy using SVM. M Hmoud et al. (2020) evaluated GoogLeNet, ResNet-50, and AlexNet on the KVASIR dataset, with AlexNet outperforming the others and achieving 97% accuracy. Yogapriya et al. (2021) used VGG16, ResNet-18, and GoogLeNet on the KVASIR v2 dataset, with VGG16 leading the way with an accuracy of about 96.33%. Furthermore, Zenebe et al. (2022) proposed a unique deep convolutional neural network (CNN) with a spatial attention mechanism for categorizing gastrointestinal (GI) illnesses [13,14]. When evaluated on a dataset of 12,147 GI images, the model demonstrated an impressive accuracy of 92.84%. This study emphasizes the importance of using pre-trained models in the correct diagnosis of gastrointestinal disorders, showcasing numerous methodologies, and achieving significant advances in this sector.
The study by Weiming Hua, Chen Lia, et al., [15] centres on enhancing deep learning models for cancer detection. This research utilizes various deep learning architectures and training methodologies to boost model accuracy, achieving notable results in cancer detection. Ensemble learning, which integrates multiple models to enhance performance, has seen increasing popularity in the field of medical imaging. Previous studies, including those by Hua et al. and others, have demonstrated that ensemble methods often surpass the performance of individual models. However, the specific combination of models and their application to particular datasets, like gastric histopathology images, have not been extensively investigated. Furthermore, the referenced paper does not deeply address data preprocessing techniques, cross-validation methodologies, the selection of the best epochs, or detailed performance metrics. These areas present opportunities for further exploration to enhance the robustness and generalizability of deep learning models in medical image analysis.
Several studies on medical image analysis have investigated the possibility of deep learning models for reliable diagnosis. Varun Sapra, Luxmi Sapra, and Akashdeep Bhardwaj, et al. [16] have proposed strategies that include pruning, quantization, and knowledge distillation to minimize ensemble model size and computing costs while maintaining accuracy. While their research produced important theoretical insights, it required a thorough technical review of real datasets.
To assess how well the proposed models worked, this research evaluated them using the recently available Gastric Histopathology Sub-Size Image Database (GasHisSDB) [17]. The main contributions of this study are the development of effective deep ensemble learning models for detecting gastric cancer that outperform current other models included in the research on the GasHisSDB dataset and the development of their ability to successfully identify gastric histology images with lower resolution, perhaps resulting in a reduction of the digital scanners, data storage, and computer servers needed for histopathology activities. This may increase the chance of the early detection of gastric cancer and enhance the rates of patient survival [18,19].

2.2. Dataset Description

The Gastric Histopathology Sub-Size Image Database (GasHisSDB), with a link to the data available in the Supplementary Material, consists of 600 pictures of stomach cancer pathology obtained from a patient’s pathological slides (see Figure 1) obtained from a specific section of their gastrointestinal tract, each measuring 2048 × 2048 pixels. The images were obtained by scanning with a new USB camera at a magnification of 20. Four pathologists from Longhua Hospital at the Shanghai University of Traditional Chinese Medicine then provided tissue-specific labelling. In conjunction with five Northeastern University biomedical experts, the photos were cropped into 245,196 sub-sized gastric cancer pathology images. Two qualified pathologists from Liaoning Cancer Hospital and Institute calibrated these pictures. The dataset was then classed as aberrant or normal, with photographs reduced to three different sizes (160 × 160, 120 × 120, and 80 × 80) for each group. The dataset of gastrointestinal images is separated into three sizes: 80 × 80, 120 × 120, and 160 × 160. Each size category is further classified into two groups: abnormal and normal (see Figure 2b–d). The 80 × 80 size category has 81,390 photos, with 34,350 classed as abnormal and 47,040 labelled as normal. With regard to the 120 × 120 size category, there are 65,260 photos altogether, with 24,800 rated as abnormal and 40,460 as normal. Finally, in the 160 × 160 size group, there are 33,272 photos, with 13,117 classified as abnormal and 20155 as normal (as described in see Figure 3). In this project, the dataset of each class was divided into a five-fold cross validation.
Figure 1. Process of extraction of the histopathology image.
Figure 1. Process of extraction of the histopathology image.
Diagnostics 14 01746 g001

2.3. Methodology Overview

In this work, CNN architectures utilizing transfer learning and ensemble approaches were demonstrated to detect gastric cancer patches. The process is divided into four major steps: (1) creating the dataset by eliminating empty patches (see Figure 2a) and augmenting, (2) tailoring pre-trained networks or base models, (3) choosing the most effective base models to form ensemble models, and (4) assessing and presenting the models using different metrics and the class activation map.
To enhance the model performance, data preprocessing was conducted to create a more balanced dataset by removing non-informative empty patches. This was undertaken because the presence of these empty patches would bias the training process and therefore jeopardize the model performance. After empty patches are removed, data augmentation was employed to increase the dataset size for training.

2.4. Empty Patch Removal Process

Empty patches are defined as those where more than half of the pixels have an RGB intensity value greater than 230 across all channels. The following is an overview of the empty patch removal procedure, including the proportion of empty patches deleted. In the 120 × 120 resolution dataset, 15.92% of patches were removed from the abnormal subclass, while 45.01% were removed from the normal subclass. For the 160 × 160 resolution dataset, 14.77% of patches were removed from the abnormal subclass, and 44.23% were removed from the normal subclass. In the 80 × 80 resolution dataset, 17.45% of patches were removed from the abnormal subclass and 45.54% were removed from the normal subclass. After eliminating the empty patches, the remaining patches were utilized to train the model. This preprocessing phase guaranteed that the dataset contained only relevant images, allowing the model to extract detailed visual characteristics for precise categorization. Figure 4 illustrates the count differences between empty patches, which are mostly backgrounds, and non-empty patches, which represent tissues in the histopathological images.

2.5. Pretrained Networks as Base Models

Convolutional neural networks (CNNs) have played an important role in numerous applications since deep learning’s inception, this is due to continual advances in strength, efficiency, and adaptability. CNNs are an excellent illustration of this breakthrough, as they are particularly built for computer vision problems and use convolutional layers inspired by natural visual processes. Multiple CNN structures have evolved throughout time, each improving accuracy, speed, and overall performance, and are usually compared with the ImageNet project, a massive visual database that promotes advancements in computer vision. Historically, training CNNs from scratch required significant computer resources and time. Transfer learning (TL) provides a practical shortcut by exploiting prior information from trained models to accelerate optimization and perhaps improve classification accuracy. TL entails the transfer of weights from pre-trained models, using insights acquired from varied datasets, and increasing the speed of training processes to improve model accuracy, particularly in complicated architectures [20,21].

2.5.1. ResNet34 Architecture

ResNet34 is a member of the residual networks (ResNet) family, which was introduced by He et al. in 2015. ResNet34 uses residual learning to address the issue of vanishing gradients, which is common in deep neural networks. This design is made up of 34 layers, with shortcut connections that allow gradients to flow straight across the network. These residual connections enable the training of very deep networks by overcoming the degradation issue. ResNet34 strikes a balance between depth and computational efficiency, beating shallower networks while maintaining manageable computational costs.
ResNet34 is a variation of ResNet, a CNN architecture created by Microsoft Research. ResNet34 includes 34 layers and uses residual connections to overcome the vanishing gradient problem, making training more efficient. ResNet models are popular due to their efficacy in a variety of applications.
The design starts with an input layer that processes a two-dimensional picture, followed by a 7 × 7 convolutional layer with 64 filters and a stride of 2, which includes padding to retain spatial dimensions (see Figure 5). A max pooling layer with a stride of two decreases the spatial dimensions. The network is divided into multiple stages of residual blocks. The first stage consists of three residual blocks, each with two 3 × 3 convolutional layers with 64 filters and shortcut connections. The second stage consists of four residual blocks, each with two 3 × 3 convolutional layers and 128 filters. The first block down-samples the feature maps with a stride of two, and the shortcut connection adjusts dimensions. The third stage consists of six residual blocks, each with two 3 × 3 convolutional layers and 256 filters. The first block downscales with a stride of 2 and the shortcut connection adjusts appropriately. The last stage is made up of three residual blocks, each with two 3 × 3 convolutional layers and 512 filters. The first block down-samples with a stride of two, and the shortcut connection does the same. An average pooling layer then reduces spatial dimensions to 1 × 1, and the pooled feature maps are fed via a fully connected layer with 1000 output units, which correspond with the number of classes in the ImageNet dataset.

2.5.2. ResNet50 Architecture

ResNet50 is a 50-layer version of the ResNet architecture, deeper than ResNet34. This depth can improve performance on certain tasks, but it also necessitates more computational resources for training.
ResNet50’s architecture handles two-dimensional input data using a succession of layers geared for effective feature extraction. It starts with zero-padding, then adds a convolutional layer, batch normalization, ReLU activation, and max-pooling to minimize spatial dimensions (see Figure 6). The network is organized into five phases, each of which contains convolutional and identity blocks. Stage 1 prepares the data, whereas Stages 2–5 contain convolutional blocks for down-sampling and identity blocks for deep feature extraction. Following these steps, an average pooling layer reduces the spatial dimensions to one by one, followed by a flattening layer and a fully linked layer for classification. The inclusion of identity blocks allows for the direct addition of input to output, which alleviates the vanishing gradient problem and enables deeper network training. This design facilitates the effective learning of complicated characteristics, hence improving image recognition performance.

2.5.3. VGGNet16 Architecture

The Visual Geometry Group (VGG) at the University of Oxford developed VGGNet16, a deep CNN architecture. It is well known for being both simple and successful in picture classification applications. VGGNet16 is made up of 16 layers, the first 13 of which are convolutional, followed by three fully connected layers. Each convolutional layer has a 3 × 3 kernel, and max-pooling layers are used after a sequence of convolutional layers to minimize spatial dimensions. Although VGGNet16 has a large number of parameters, which makes it computationally costly, it achieves good accuracy on benchmark datasets thanks to its deep architecture and consistent layer design.
VGGNet, created by the Visual Geometry Group at the University of Oxford, is renowned for its simplicity. It is mostly composed of layered convolutional layers with 3 × 3 kernels and max-pooling layers. VGGNet includes several variations, including VGG16 and VGG19, which differ in the number of layers.
The VGGNet16 architecture (see Figure 7) runs an input picture through a sequence of convolutional layers, each followed by ReLU activation. It starts with two convolutional layers of 64 filters each, followed by max-pooling (see Figure 7). This pattern is repeated with increasing numbers of filters (128, 256, and eventually 512). To minimize spatial dimensions, max-pooling is applied after each convolutional layer step. Three fully connected layers follow the convolutional layers, with the last layer serving as a softmax layer for 1000-class classification. This deep design, which makes regular use of modest 3 × 3 filters, enables good feature extraction while keeping reasonable computing cost, making VGGNet16 a strong model for image recognition applications.

2.5.4. EfficientNet Architecture

EfficientNet is a class of CNN models developed by Google AI that outperforms earlier models using fewer parameters and FLOPs. They employ a novel scaling technique to improve network depth, breadth, and resolution for improved resource management.
The EfficientNet design starts with a 3 × 3 convolutional layer, then numerous mobile inverted bottleneck (MBConv) blocks with different kernel sizes and expansion factors (as shown in Figure 8). The design employs a mix of 3 × 3 and 5 × 5 MBConv blocks to efficiently record spatial characteristics at various sizes. Each MBConv block is composed of a depth wise convolution followed by a pointwise convolution, allowing the model to capture complicated patterns while being computationally efficient. The network gradually raises the number of filters and the resolution of feature maps using these MBConv blocks, resulting in a highly detailed and rich feature map. This architecture strikes a compromise between depth, breadth, and resolution, producing an efficient and strong model for picture categorization tasks.

2.5.5. VitNet Architecture

The vision transformer is a novel CNN architecture created by Google Research that employs self-attention techniques seen in transformer topologies. Rather than employing convolutional layers like typical CNNs, ViT employs a transformer encoder, which allows it to recognize relationships across large distances in pictures. This method has shown outstanding results in several computer vision tasks, particularly when trained on large datasets.
The vision transformer (ViT) architecture (as shown in Figure 9) divides an input picture into fixed-size, non-overlapping patches, which are then flattened into a 1D vector (as shown in Figure 9). The vectors are then projected linearly onto a lower-dimensional space. Position embeddings are applied to these vectors in order to keep the positioning information of the patches inside the original picture. The generated sequence is routed through a transformer encoder, which consists of many layers of multi-head self-attention mechanisms and feed-forward neural networks. This encoder captures both the global environment and the interactions between patches. The output of the transformer encoder is then passed into a multi-layer perceptron (MLP) head, which computes the final predicted probability for each class. This design enables ViT to harness the benefits of transformers in collecting long-range relationships and global picture context.

2.5.6. Ensemble Architecture

Convolutional neural networks (CNNs) and transfer learning have significantly increased neural network performance, but there is still potential for improvement. This article proposes the use of ensemble methods to improve the effectiveness of the three models that have been pre-trained. Ensemble learning, a machine learning and statistics-based approach combines the skills of many algorithms to extract relevant insights from data [19]. During this analysis, stacking was determined to be the most appropriate method. This requires training various ML algorithms on the information before merging them to create a composite algorithm capable of successfully combining their input.
In this research, ensemble model architecture was used using ResNet34 and VGGNet16 as basis models. Figure 10 shows the proposed ensemble, which comprises both ResNet34 and VGGNet16. Initially, each model was trained independently to determine its unique performance. We then selected the best-performing epochs for each model based on validation accuracy. The best-performing epochs were then used to build the ensemble model. To improve the robustness and generalizability of the ensemble techniques, cross-validation was used. During each fold of the cross-validation approach, the research utilized the best weights from the previous folds to train the ensemble. For example, if cross-validation was undertaken on fold 1, the best epochs from folds 2, 3, 4, and 5 of ResNet34 and VGGNet16 were used to establish the ensemble model. The ensemble model architecture was created to leverage the complementary characteristics of ResNet34 and VGGNet16. ResNet34, with its residual connections, successfully mitigates the vanishing gradient problem, allowing for deeper network training. In contrast, VGGNet16, noted for its simplicity and constant layer architecture, excels in capturing fine-grained characteristics. By integrating both models, the ensemble makes use of VGGNet16’s comprehensive feature extraction and ResNet34’s depth-wise learning capabilities. Training the ensemble model entailed freezing the early layers of both base models to preserve their pre-trained feature extraction capabilities while fine-tuning the subsequent layers to fit them to the unique dataset. This hybrid strategy enabled the ensemble model to outperform individual models, as indicated by the improved accuracy and resilience across many validation criteria. In summary, the proposed ensemble model, which includes ResNet34 and VGGNet16, showed considerable performance increases. The strategic use of cross-validation and the incorporation of complementary model architectures demonstrate the effectiveness of ensemble techniques in deep learning applications.

2.6. Interpretability of the Ensemble Model

Understanding how ensemble models make conclusions is critical for gaining acceptance and confidence in medical diagnoses. Our work makes use of class activation mapping (CAM) tools to display and comprehend the ensemble models’ decision-making processes. CAM identifies the key areas in histopathological pictures that have the most impact on the model’s predictions, offering insights into the characteristics examined by the models. During the training phase, we divided the dataset into a subset of validation pictures in order to determine which areas the models focused on when discriminating between normal and diseased tissues. Our cooperating pathologists then inspected the produced maps to confirm that the highlighted locations matched significant histological findings. This technique not only confirmed the model’s focus areas, but also helped to better grasp its decision-making framework.
Furthermore, we used ensemble learning with a stacking technique, in which each of the base models (ResNet34, VGGNet16, and so on) were trained separately. These models’ outputs were integrated using a meta-classifier, which pooled their predictions. This multi-model decision-making technique improves resilience and minimizes the danger of incorrect classifications by combining the capabilities of many architectures. By adding these interpretability methodologies, we may offer a clear picture of the model’s underlying workings, increasing the trustworthiness and dependability of diagnostic results.

2.7. Experimental Setting

The data were divided into training and validation sets. Each network was trained for 20 epochs using 5-fold cross-validation to create the model. The weights from the epoch with the best validation accuracy were chosen as the final representations for each model. Various metrics were then employed to assess accuracy, followed by many objective assessment factors to determine overall performance. The dataset was rigorously separated into training and validation sets to ensure a thorough evaluation of the model’s performance. The training approach used 5-fold cross-validation across 20 epochs for each network. This strategy provides a thorough evaluation by cycling through many train–test divides, lowering the danger of overfitting and guaranteeing that the model generalizes well to new data. For example, using an 80 × 80 dataset, we split it into five equal sections. During the initial fold, the model was trained on the first four components and verified on the fifth. In the second fold, the model was trained on parts two, three, four, and five, with validation on the first component. The third fold featured training on parts three, four, five, and one, followed by validation on part two. This method was repeated for the remaining two folds, ensuring that each portion was only used as a validation set once. After running the 20 epochs for each fold, we chose the weights from the epoch with the best validation accuracy to represent the final model for that fold. This strategy ensured that the model parameters that were picked functioned best on data that were not known, hence increasing the model’s dependability. Various metrics were then employed to assess accuracy, followed by many objective assessment factors to determine overall performance [22].

3. Results

The performance evaluation criteria used include accuracy, sensitivity, specificity, Jaccard index, and area under the curve (AUC). Positive samples are those that include abnormal or malignant patches, whereas negative samples contain normal or healthy patches. The phrases true positive (TP), false positive (FP), true negative (TN), and false negative (FN) are used to describe the various prediction results.
1. Accuracy: Accuracy is the ratio of properly identified samples to the total number of samples. This is computed as follows:
A c c u r a c y = ( T P + T N ) ( T P + T N + F P + F N )
2. Sensitivity (recall): Sensitivity, also known as recall, is the proportion of real positive samples that the model properly identifies. This is provided as follows:
S e n s i t i v i t y = ( T P ) ( T P + F N )
3. Specificity refers to the fraction of real negative samples properly detected by the model. This is computed as follows:
S p e c i f i c i t y = ( T N ) ( T N + F P )
4. The Jaccard index, commonly known as the intersection over union (IoU), assesses the similarity between expected and observed positive samples. This is provided as follows:
J a c c a r d   i n d e x = ( T P ) ( T P + F P + F N )
5. Area under the curve (AUC): The area under the receiver operating characteristic (ROC) curve, or AUC, relates the true positive rate (sensitivity) to the false positive rate (1-specificity). A higher AUC implies improved model performance.
By examining these measures, a thorough picture of the model’s performance can be acquired, particularly when discriminating between aberrant (positive) and normal (negative) data.
When analysing the success of machine learning models, performance measures must be considered. These metrics provide values that indicate the overall performance of a statistical or machine learning technique. In classification tasks, performance measures evaluate the model’s capacity to accurately categorize data points as well as its consistency in producing the right classifications. Classification accuracy and F1 score are both measures of classification task accuracy, whereas AUC reflects a model’s overall ability to forecast accurately. The study’s findings, which were obtained by examining these performance metrics, are shown in the Table 1, Table 2 and Table 3 below.

4. Discussion

In this work, the proposed model was trained on datasets with varied picture sizes, such as 80 × 80 and 120 × 120 and 160 × 160. Findings highlights the critical significance of ongoing innovation and exploration in increasing medical machine learning and hence improving healthcare practices. Notably, the findings show that the top five ensemble models had high detection accuracy across all sub-databases. The overall ensemble model demonstrated the highest accuracy, (see Table 1, Table 2 and Table 3) surpassing the performance of VGGNet, ResNet34, and ResNet50, which also outperformed VitNet and EfficientNet. The only exception was in the 160 × 160 sub-database, where EfficientNet achieved an accuracy of 92%. By combining multidisciplinary techniques and technology breakthroughs, one can pave the way towards a future in which the early and precise identification of gastrointestinal disorders is not only achievable but also common practice in protecting human health and wellbeing.
This deliberate method was used to assist the model build a solid knowledge of multiple picture dimensions, making it more flexible to a variety of real-world scenarios. Nonetheless, it is worth noting that the proposed model’s knowledge may have been expanded much more with greater computing resources. Greater computational capability might have enabled a more in-depth examination of the dataset’s intricacies, as well as the discovery of insights beyond existing capabilities.
These advanced methodologies offer various structures and improvement tactics that may increase the efficacy of the model. Time restrictions precluded the incorporation of these algorithms into the current system; nonetheless, their use holds great promise for improving cancer detection techniques.
This research prioritizes in innovation and placing people at the heart of the research. The objective is to make research applications as user-friendly as possible while also emphasizing how they might benefit medical practitioners. This research strives to integrate the most recent research work with practical applications, avoiding plagiarism and crafting a research story that is honest and true to the research’s commitment to expanding scientific understanding.

Experimental Setting

In the future, the focus will be on two major goals: developing an easy-to-use web app for cancer detection and expanding current algorithms to include new forms of cancer. The primary goal of this research is to develop a model that will work for all forms of cancer, not just specific forms. This will be accomplished by utilizing many data sources and innovative algorithms in order to obtain a thorough understanding of cancer, hence assisting healthcare professionals with diagnosis.
By integrating multidisciplinary methodologies and capitalizing on technological developments, we hope to contribute towards a future in which the early and precise identification of GI problems is not only possible, but is also the standard for protecting human health and wellbeing.

5. Conclusions

Detecting and diagnosing gastrointestinal (GI) illnesses is critical for human health, yet it can be challenging owing to limited medical competence and expensive expenses. The use of machine learning, particularly deep learning techniques, has the potential to increase the speed and accuracy of GI illness identification. This research study investigated the efficiency of ensemble approaches using five pre-trained models on the Gastric Histopathology Sub-size Image Database (GasHisSDB), which comprises a diverse set of pictures in various pixel sizes and categories. Considerable boost in prediction accuracy can be seen when utilizing ensemble learning, which combines the predictive skills of several models, as opposed to using individual models and in contrast with previous studies that used comparable datasets. This demonstrates the potential of ensemble approaches to improve the capabilities of medical machine learning systems, resulting in more effective and precise diagnoses. The current approach is based on transfer learning, a technique that improves model learning by leveraging knowledge from previously trained models. Moreover, we employed an ensemble strategy to improve performance by merging various classifiers. Following a rigorous review, the current approach revealed a good accuracy for the test dataset, exceeding current evaluation techniques. This demonstrates how deep learning may help alleviate the pressure on healthcare systems while also improving human health outcomes [22,23,24,25]. In this study, advanced deep ensemble learning models were created that used transfer learning from multiple pre-trained networks, including VitNet, EfficientNet, VGGNet, ResNet34, and ResNet50, to improve stomach cancer diagnosis. The study found that using base models in ensemble learning resulted in high identification accuracy (97.57% to 99.72%) for histopathology pictures with resolutions ranging from 80 × 80 pixels to 160 × 160 pixels. The experimental results demonstrate that ensemble models may extract key information even from smaller picture patches while retaining good performance. This improvement implies the possibility of adopting digital scanners with lower specifications, as well as reduced data storage and computing needs for histopathology operations. This, in turn, might increase the speed of stomach cancer identification and perhaps increase survival rates. Continued work in these areas is intended to push the boundaries of medical image processing and enhance clinical results.
However, it is critical to recognize the limits of the current research. The usage of a restricted dataset emphasizes the need to have access to larger, higher-quality datasets to enhance and confirm the approaches. Furthermore, computational restrictions may have influenced the scope of the results. Future studies might seek to introduce new preprocessing methods and to optimize algorithms to boost performance. Furthermore, the field of medical image retrieval offers several chances for continuous research, including the ability to use multiple deep learning approaches and models for complete examination.

Supplementary Materials

The GasHisSDB dataset is openly available at this Link: https://paperswithcode.com/dataset/gashissdb (Access Date: 11 March 2024).

Author Contributions

Conceptualization, S.M.A.A.-H., M.D. and S.A.; Methodology, M.D., G.R.M., S.M.A.A.-H. and S.A.; Software, G.R.M., I.H.A., M.D., S.M.A.A.-H. and S.A.; Validation, I.H.A., G.R.M., S.M.A.A.-H., A.B., B.A., R.C., S.T.R. and S.A.; Formal analysis, S.A.; Investigation, S.M.A.A.-H., A.B., B.A., R.C., S.T.R. and S.A.; Writing—original draft, G.R.M.; Writing—review & editing, S.M.A.A.-H. and S.A.; Visualization, S.A.; Supervision, S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available in this article.

Acknowledgments

We would like to express our deepest gratitude to the Roux Institute, the IEAI and the Alfond Foundation for their invaluable support and contributions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rai, H.M. Cancer detection and segmentation using machine learning and deep learning techniques: A review. Multimed. Tools Appl. 2024, 83, 27001–27035. [Google Scholar] [CrossRef]
  2. Kuntz, S.; Krieghoff-Henning, E.; Kather, J.N.; Jutzi, T.; Höhn, J.; Kiehl, L.; Hekler, A.; Alwers, E.; von Kalle, C.; Fröhling, S.; et al. Gastrointestinal cancer classification and prognostication from histology using deep learning: Systematic review. Eur. J. Cancer 2021, 155, 200–215. [Google Scholar] [CrossRef] [PubMed]
  3. Suzuki, H.; Yoshitaka, T.; Yoshio, T.; Tada, T. Artificial intelligence for cancer detection of the upper gastrointestinal tract. Dis. Esophagus 2021, 33, 254–262. [Google Scholar] [CrossRef] [PubMed]
  4. Chao, W.L.; Manickavasagan, H.; Krishna, S.G. Application of Artificial Intelligence in the Detection and Differentiation of Colon Polyps: A Technical Review for Physicians. Diagnostics 2019, 9, 99. [Google Scholar] [CrossRef]
  5. Li, B.; Meng, M.Q.H. Automatic polyp detection for wireless capsule endoscopy images. Expert. Syst. Appl. 2012, 39, 10952–10958. [Google Scholar] [CrossRef]
  6. Guo, L.; Gong, H.; Wang, Q.; Zhang, Q.; Tong, H.; Li, J.; Lei, X.; Xiao, X.; Li, C.; Jiang, J.; et al. Detection of multiple lesions of gastrointestinal tract for endoscopy using artificial intelligence model: A pilot study. Surg. Endosc. 2021, 35, 6532–6538. [Google Scholar] [CrossRef] [PubMed]
  7. Charfi, S.; El Ansari, M.; Balasingham, I. Computer-aided diagnosis system for ulcer detection in wireless capsule endoscopy images. IET Image Process. 2019, 13, 1023–1030. [Google Scholar] [CrossRef]
  8. Mirza, O.M.; Alsobhi, A.; Hasanin, T.; Ishak, M.K.; Karim, F.K.; Mostafa, S.M. Computer Aided Diagnosis for Gastrointestinal Cancer Classification Using Hybrid Rice Optimization With Deep Learning. IEEE Access 2023, 11, 76321–76329. [Google Scholar] [CrossRef]
  9. Liedlgruber, M.; Uhl, A. Computer-aided decision support systems for endoscopy in the gastrointestinal tract: A review. IEEE Rev. Biomed. Eng. 2011, 4, 73–88. [Google Scholar] [CrossRef]
  10. Sivari, E.; Bostanci, E.; Guzel, M.S.; Acici, K.; Asuroglu, T.; Ercelebi Ayyildiz, T. A New Approach for Gastrointestinal Tract Findings Detection and Classification: Deep Learning-Based Hybrid Stacking Ensemble Models. Diagnostics 2023, 13, 720. [Google Scholar] [CrossRef]
  11. Gunasekaran, H.; Ramalakshmi, K.; Swaminathan, D.K.; J, A.; Mazzara, M. GIT-Net: An Ensemble Deep Learning-Based GI Tract Classification of Endoscopic Images. Bioengineering 2023, 10, 809. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  12. Naz, J.; Sharif, M.; Yasmin, M.; Raza, M.; Khan, M.A. Detection and Classification of Gastrointestinal Diseases using Machine Learning. Curr. Med. Imaging Former. Curr. Med. Imaging Rev. 2020, 17, 479–490. [Google Scholar] [CrossRef]
  13. Wang, X.; Qian, H.; Ciaccio, E.J.; Lewis, S.K.; Bhagat, G.; Green, P.H.; Xu, S.; Huang, L.; Gao, R.; Liu, Y. Celiac disease diagnosis from videocapsule endoscopy images with residual learning and deep feature extraction. Comput. Methods Programs Biomed. 2020, 187, 105236. [Google Scholar] [CrossRef]
  14. Renna, F.; Martins, M.; Neto, A.; Cunha, A.; Libânio, D.; Dinis-Ribeiro, M.; Coimbra, M. Artificial Intelligence for Upper Gastrointestinal Endoscopy: A Roadmap from Technology Development to Clinical Practice. Diagnostics 2022, 12, 1278. [Google Scholar] [CrossRef] [PubMed]
  15. Hu, W.; Li, C.; Li, X.; Rahaman, M.M.; Ma, J.; Zhang, Y.; Chen, H.; Liu, W.; Sun, C.; Yao, Y.; et al. GasHisSDB: A new gastric histopathology image dataset for computer aided diagnosis of gastric cancer. Comput. Biol. Med. 2022, 142, 105207. [Google Scholar] [CrossRef]
  16. Sapra, V.; Sapra, L.; Bhardwaj, A.; Bharany, S.; Saxena, A.; Karim, F.K.; Ghorashi, S.; Mohamed, A.W. Integrated approach using deep neural network and CBR for detecting severity of coronary artery disease. Alex. Eng. J. 2023, 68, 709–720. [Google Scholar] [CrossRef]
  17. Yong, M.P.; Hum, Y.C.; Lai, K.W.; Lee, Y.L.; Goh, C.H.; Yap, W.S.; Tee, Y.K. Histopathological Gastric Cancer Detection on GasHisSDB Dataset Using Deep Ensemble Learning. Diagnostics 2023, 13, 1793. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  18. Su, Q.; Wang, F.; Chen, D.; Chen, G.; Li, C.; Wei, L. Deep convolutional neural networks with ensemble learning and transfer learning for automated detection of gastrointestinal diseases. Comput. Biol. Med. 2022, 150, 106054. [Google Scholar] [CrossRef] [PubMed]
  19. Almanifi, O.R.A.; Razman, M.A.M.; Khairuddin, I.M.; Abdullah, M.A.; Majeed, A.P.P.A. Automated Gastrointestinal Tract Classification Via Deep Learning and The Ensemble Method. In Proceedings of the 2021 21st International Conference on Control, Automation and Systems (ICCAS), Jeju, Republic of Korea, 12–15 October 2021; pp. 602–606. [Google Scholar] [CrossRef]
  20. Haile, M.B.; Salau, A.O.; Enyew, B.; Belay, A.J.; Jin, Z. Detection and classification of gastrointestinal disease using convolutional neural network and SVM. Cogent Eng. 2022, 9, 2084878. [Google Scholar] [CrossRef]
  21. Billah, M.; Waheed, S. Gastrointestinal polyp detection in endoscopic images using an improved feature extraction method. Biomed. Eng. Lett. 2018, 8, 69–75. [Google Scholar] [CrossRef]
  22. Abin, D.; Mudavadkar, G.; Minase, R.; Handoo, M.; Kumawat, R. MRI Image Enhancement for Brain Tumor Detection using Hybridization of Contrast Enhancement Techniques. In Proceedings of the 2022 International Conference on Knowledge Engineering and Communication Systems (ICKES), Chickballapur, India, 28–29 December 2022; pp. 1–5. [Google Scholar] [CrossRef]
  23. Ghanzouri, I.; Amal, S.; Ho, V.; Safarnejad, L.; Cabot, J.; Brown-Johnson, C.G.; Leeper, N.; Asch, S.; Shah, N.H.; Ross, E.G. Performance and Usability Testing of an Automated Tool for Detection of Peripheral Artery Disease Using Electronic Health Records. Sci. Rep. 2022, 12, 13364. [Google Scholar] [CrossRef]
  24. Singh, A.; Wan, M.; Harrison, L.; Breggia, A.; Christman, R.; Winslow, R.L.; Amal, S. Visualizing Decisions and Analytics of Artificial Intelligence Based Cancer Diagnosis and Grading of Specimen Digitized Biopsy: Case Study for Prostate Cancer. In Proceedings of the Companion Proceedings of the 28th International Conference on Intelligent User Interfaces, Sydney, Australia, 27–31 March 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 166–170. [Google Scholar]
  25. Yasmin, F.; Hassan, M.M.; Hasan, M.; Zaman, S.; Bairagi, A.K.; El-Shafai, W.; Fouad, H.; Chun, Y.C. GastroNet: Gastrointestinal Polyp and Abnormal Feature Detection and Classification with Deep Learning Approach. IEEE Access 2023, 11, 97605–97624. [Google Scholar] [CrossRef]
Figure 2. Examples of histopathological gastric images. (a) The pre-processing step and (bd) examples of different image sizes.
Figure 2. Examples of histopathological gastric images. (a) The pre-processing step and (bd) examples of different image sizes.
Diagnostics 14 01746 g002
Figure 3. Summaries of the total number of images for every subclass in an experiment setup.
Figure 3. Summaries of the total number of images for every subclass in an experiment setup.
Diagnostics 14 01746 g003
Figure 4. Summarises the total number of the empty and non-empty patch images in the dataset.
Figure 4. Summarises the total number of the empty and non-empty patch images in the dataset.
Diagnostics 14 01746 g004
Figure 5. ResNet34 model architecture.
Figure 5. ResNet34 model architecture.
Diagnostics 14 01746 g005
Figure 6. ResNet50 model architecture.
Figure 6. ResNet50 model architecture.
Diagnostics 14 01746 g006
Figure 7. VGGNet16 model architecture.
Figure 7. VGGNet16 model architecture.
Diagnostics 14 01746 g007
Figure 8. EfficientNet model architecture.
Figure 8. EfficientNet model architecture.
Diagnostics 14 01746 g008
Figure 9. VITNet model architecture.
Figure 9. VITNet model architecture.
Diagnostics 14 01746 g009
Figure 10. Ensemble model detailed architecture.
Figure 10. Ensemble model detailed architecture.
Diagnostics 14 01746 g010
Table 1. The effectiveness of the several deep learning models was assessed using an 80-pixel sub-database, displayed above.
Table 1. The effectiveness of the several deep learning models was assessed using an 80-pixel sub-database, displayed above.
ModelFoldTrain AccuracyTrain LossVal AccuracyVal LossJaccard IndexAUCSpecificitySensitivity
Resnet34198.75080.033193.61310.21750.74470.97190.96040.8494
297.81920.060493.72060.20920.74720.97170.96380.8427
398.85460.032993.81110.22100.73840.97600.96990.8204
497.77730.059693.82260.17240.74660.97770.96250.8491
598.17670.048793.79100.23990.76400.97540.95180.8900
ResNet50198.48180.040993.74080.19360.87630.89790.96820.8277
296.24190.095894.24390.18800.88600.90690.97030.8435
398.44100.041393.49710.20240.87420.90420.95770.8508
497.06540.076992.76880.20160.86010.90710.94310.8711
593.54110.167393.84590.18690.87720.90640.96480.8480
VitNet183.62430.363183.75910.35550.69450.71320.93470.4916
282.98350.381684.28360.34900.69690.70760.94930.4660
378.49740.475578.86560.45600.59460.50540.99880.0121
482.56120.389884.28410.35050.69860.70150.94880.4542
578.32170.523177.41050.53530.56740.510
VggNet198.73710.036693.75910.25790.87630.97460.97470.8053
296.26930.096994.33410.17850.88680.97570.97380.8353
398.56830.041793.92200.21200.87630.97510.97200.8178
494.32240.151993.76810.16210.87560.97330.96850.8245
598.36360.048793.79100.23290.87680.97240.96930.8302
EfficientNet174.94740.503575.40570.51420.60940.74780.71720.7784
282.44350.386683.04740.36850.68290.67260.94210.4030
381.03590.413982.37090.38550.67040.63460.94630.3229
482.33790.387984.22310.35310.69750.67250.94190.4030
582.65710.380583.52100.36510.69440.68160.94010.4232
Ensemble198.35870.043699.34300.02520.98670.99040.99570.9850
298.51870.041099.24210.02110.98390.98750.99620.9787
398.60010.038499.20560.02370.98360.98980.99360.9861
497.92340.062797.81970.06510.95620.96420.98860.9398
598.56420.039099.08690.02210.98230.98660.99430.9789
The most interesting results are shown in bold.
Table 2. The effectiveness of the several deep learning models was assessed using a 120-pixel sub-database, displayed above.
Table 2. The effectiveness of the several deep learning models was assessed using a 120-pixel sub-database, displayed above.
ModelFoldTrain AccuracyTrain LossVal AccuracyVal LossJaccard IndexAUCSpecificitySensitivity
Resnet34198.86270.032796.81670.10970.83470.98790.98120.9075
299.28470.018797.02970.13270.84440.98150.98840.8884
398.91450.033696.93640.10250.83920.99070.99010.8762
499.16400.023496.44750.12080.81750.98790.98360.8778
599.29570.021496.68930.10740.81490.98970.98770.8642
ResNet50199.55520.012596.72570.11200.93160.95180.97560.9281
299.24490.021396.98560.11560.93720.92540.99510.8557
399.72290.008997.09300.11980.93880.94930.98330.9154
499.50070.014296.81660.12530.93390.93790.98530.8905
599.50420.014797.07480.09950.93790.94060.98600.8951
VitNet186.69630.310987.47150.28750.75180.73950.94880.5301
284.69480.349586.18260.31850.72540.71230.94700.4775
382.28170.448681.75310.42870.63100.510
485.74470.329887.91230.29290.75810.76070.94640.5750
584.20840.346586.32650.30850.72310.67670.95820.3951
VggNet176.61850.476475.61620.50530.61520.76570.77760.7538
288.37400.275889.17680.26110.78890.74760.95920.5361
388.13510.282288.33880.27320.77560.73720.96050.5139
486.65760.314088.28260.28130.76730.69790.95910.4368
588.37020.274588.41980.27810.77790.74600.95940.5325
EfficientNet199.19490.026596.81670.13720.93400.98780.98670.8818
299.89780.005496.94160.17160.93630.98860.98730.8884
398.43390.046497.18240.11300.94030.98830.99010.8897
496.45440.101196.03220.10980.91760.98600.97570.8905
599.83660.006196.96140.14100.93380.99020.983630.9005
Ensemble197.25810.072897.61250.06460.95010.95930.98530.9332
297.79730.062298.17380.05330.96270.96470.99130.9381
397.78940.060397.87560.06180.95310.96360.98740.9399
497.84570.062597.57780.06850.94860.95490.98760.9223
598.97350.031699.42500.02260.98720.98950.99650.9826
The most interesting results are shown in bold.
Table 3. The effectiveness of the several deep learning models was assessed using a 160-pixel sub-database, displayed above.
Table 3. The effectiveness of the several deep learning models was assessed using a 160-pixel sub-database, displayed above.
ModelFoldTrain AccuracyTrain LossVal AccuracyVal LossJaccard IndexAUCSpecificitySensitivity
Resnet34199.63110.013297.98070.06480.89490.99610.98870.9398
299.54960.014898.17760.06560.90220.99130.99220.9341
399.73370.008398.10380.05350.89280.99760.99120.9308
497.11840.089697.20050.08210.86650.99310.98060.9361
599.54370.021398.41870.07380.91050.99290.98620.9739
ResNet50199.36600.028898.43960.05790.96780.97390.99040.9573
299.58420.017098.31430.05920.96280.97190.98940.9544
399.82630.006698.10380.07760.95990.96840.98740.9494
499.95380.001897.93480.07430.95740.96200.99030.9338
599.77180.011298.37080.06250.96480.98320.98390.9826
VitNet182.26140.452681.68880.45320.62950.510
284.49190.342686.28700.31410.73070.69800.95550.4405
381.92120.463883.02480.45270.65430.510
484.37060.350284.67180.33930.69770.68410.94980.4184
582.04830.395186.39190.30690.72740.63370.97760.2898
VggNet177.32460.478476.58260.48200.62910.77320.76430.7822
290.08750.229693.07020.18380.86070.79240.96040.6244
391.06230.215592.80180.19110.85720.81390.96410.6636
491.00690.217893.09250.16550.85890.81680.96280.6709
591.44760.207792.01460.19450.84440.81860.96590.6712
EfficientNet199.97690.001498.21010.11210.96230.99460.99150.9398
299.68820.009897.94980.07950.95840.99440.99520.9088
398.87730.034797.96830.08960.95790.99210.99340.9122
499.27380.024497.20050.09400.94640.99120.98340.9243
599.66920.009198.13120.08140.95860.99480.98960.9391
Ensemble198.47850.041298.80670.03130.97510.98290.99100.9749
299.08770.026898.90660.03250.97530.97750.99550.9594
399.12030.029899.18730.02600.98300.98660.99450.9787
499.48130.015099.72460.00790.99420.99640.99770.9952
598.91650.031999.32910.02450.98360.99010.99480.9855
The most stunning results are shown in bold.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mudavadkar, G.R.; Deng, M.; Al-Heejawi, S.M.A.; Arora, I.H.; Breggia, A.; Ahmad, B.; Christman, R.; Ryan, S.T.; Amal, S. Gastric Cancer Detection with Ensemble Learning on Digital Pathology: Use Case of Gastric Cancer on GasHisSDB Dataset. Diagnostics 2024, 14, 1746. https://doi.org/10.3390/diagnostics14161746

AMA Style

Mudavadkar GR, Deng M, Al-Heejawi SMA, Arora IH, Breggia A, Ahmad B, Christman R, Ryan ST, Amal S. Gastric Cancer Detection with Ensemble Learning on Digital Pathology: Use Case of Gastric Cancer on GasHisSDB Dataset. Diagnostics. 2024; 14(16):1746. https://doi.org/10.3390/diagnostics14161746

Chicago/Turabian Style

Mudavadkar, Govind Rajesh, Mo Deng, Salah Mohammed Awad Al-Heejawi, Isha Hemant Arora, Anne Breggia, Bilal Ahmad, Robert Christman, Stephen T. Ryan, and Saeed Amal. 2024. "Gastric Cancer Detection with Ensemble Learning on Digital Pathology: Use Case of Gastric Cancer on GasHisSDB Dataset" Diagnostics 14, no. 16: 1746. https://doi.org/10.3390/diagnostics14161746

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop