Next Article in Journal
Potential Role of the Microbiome in Acne: A Comprehensive Review
Next Article in Special Issue
Artificial Intelligence Prediction Model for the Cost and Mortality of Renal Replacement Therapy in Aged and Super-Aged Populations in Taiwan
Previous Article in Journal
Metabolic and Nutritional Characteristics of Long-Stay Critically Ill Patients
Previous Article in Special Issue
Shared Decision-Making in Chronic Patients with Polypharmacy: An Interventional Study for Assessing Medication Appropriateness
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Artificial Intelligence-Based Classification of Multiple Gastrointestinal Diseases Using Endoscopy Videos for Clinical Diagnosis

Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro 1-gil, Jung-gu, Seoul 04620, Korea
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2019, 8(7), 986; https://doi.org/10.3390/jcm8070986
Submission received: 31 May 2019 / Revised: 2 July 2019 / Accepted: 5 July 2019 / Published: 7 July 2019
(This article belongs to the Special Issue The Future of Artificial Intelligence in Clinical Medicine)

Abstract

:
Various techniques using artificial intelligence (AI) have resulted in a significant contribution to field of medical image and video-based diagnoses, such as radiology, pathology, and endoscopy, including the classification of gastrointestinal (GI) diseases. Most previous studies on the classification of GI diseases use only spatial features, which demonstrate low performance in the classification of multiple GI diseases. Although there are a few previous studies using temporal features based on a three-dimensional convolutional neural network, only a specific part of the GI tract was involved with the limited number of classes. To overcome these problems, we propose a comprehensive AI-based framework for the classification of multiple GI diseases by using endoscopic videos, which can simultaneously extract both spatial and temporal features to achieve better classification performance. Two different residual networks and a long short-term memory model are integrated in a cascaded mode to extract spatial and temporal features, respectively. Experiments were conducted on a combined dataset consisting of one of the largest endoscopic videos with 52,471 frames. The results demonstrate the effectiveness of the proposed classification framework for multi-GI diseases. The experimental results of the proposed model (97.057% area under the curve) demonstrate superior performance over the state-of-the-art methods and indicate its potential for clinical applications.

1. Introduction

Different types of gastrointestinal (GI) diseases, such as colorectal cancer and tumor, are the leading cause of death in the USA [1]. According to the American Cancer Society, approximately 76,940 people lost their lives in 2016 owing to different types of cancers in the GI tract [1]. The effective diagnosis of such GI diseases is a tedious and time-consuming task. Most of the small GI lesions remain imperceptible during the early stages, which ultimately evolves into a fatal ailment. Therefore, it is essential to develop computerized approaches that can assist the physicians in effective diagnosis and treatment. Therefore, substantial efforts were focused over the past few decades to develop artificial intelligence (AI)-based computer-aided diagnosis (CAD) tools and applications in various medical fields [2,3,4]. These fields include the detection of brain tumor [5], classification of different types of skin cancers, diagnosis in radiation oncology, diabetic retinopathy, histologic classification of gastric biopsy, and endoscopy [6,7,8,9,10,11,12,13,14,15].
In the field of endoscopy, the recent AI-based CAD tools utilize the strength of deep learning (a set of advanced machine learning algorithms) for the analysis of various types of endoscopic scans. In general, deep learning algorithms are used to extract the optimal representations of training data. A training algorithm optimizes the learnable parameters of the deep learning model [16]. Based on the optimal features extracted from the training dataset, the CAD tool can analyze the newly acquired clinical images prospectively. Artificial neural networks (ANNs) are the key component of such deep learning-based image analysis tools that logically emulates the structure and activity of the brain neurons on a computer. Various types of ANNs were proposed, including convolutional neural networks (CNNs), in the field of image recognition [16,17]. However, all the supervised ANN-based image recognition methods require a training dataset, which is used to estimate the optimal network parameters for sufficient training. In the field of medical image analysis, similar training of an ANN model is performed, which is called supervised learning (SL). In SL, the available training dataset consists of both input images and appropriate output information. The primary portion of ANN model is the stack of multiple layers, which comprise of learnable filters with different values of size and depth. These layers extract the complex features from the available training dataset by using different learning algorithms. Finally, the ANN network learns from these features without using other handcrafted features [16,17]. After performing sufficient training, an ANN-based diagnostic framework demonstrates the best performance in various clinical applications.
This research primarily focuses on the analysis of different deep learning models used in the classification of GI diseases. We analyze in depth the performance of the most recent CNN models considering the following perspectives: (1) the importance of spatial and temporal features in the classification of GI diseases; (2) feature selection from different layers within a CNN network; (3) combining the CNN and long short-term memory (LSTM); and (4) analyzing the effects of the temporal features by considering different number of successive frames. Thus, we proposed a cascaded deep feature-based framework by combining the deep residual network (ResNet) and LSTM to obtain the best classification accuracy. Finally, we provide a novel spatiotemporal features-based pretrained model for the classification of multiple GI diseases, which is our primary contribution. We have also ensured that our pretrained model and the video indices of the experimental endoscopic videos are publicly available for other researchers [18].
The rest of this paper is organized as follows: The related studies on endoscopy for the detection and classification of different GI diseases are provided in Section 2, and a brief summary of our contribution to this research is explained in Section 3. In Section 4, a comprehensive description of the proposed classification framework for multiple GI diseases is presented. In Section 5, we illustrate the experimental setup and performance analysis of the proposed method to validate its performance and efficiency over the previous deep learning and handcrafted features-based methods. A discussion on certain important issues relevant to this paper is presented in Section 6. Finally, Section 7 draws a conclusion of our research work.

2. Related Works

In recent years, the strength of deep learning-based algorithms has been utilized in the field of endoscopy, including capsule endoscopy (CE), esophagogastroduodenoscopy (EGD), and colonoscopy [6,7,8,9,10,11,12,13,14,15]. To facilitate the physicians with the effective diagnosis of different GI lesions, several CNN-based CAD tools have been proposed in the literature. These CAD tools are capable of detecting and classifying even small lesions in the GI tract, which often remain imperceptible to the human visual system. Before the advent of deep learning methods, many previous studies have focused on the handcrafted feature-based methods, which mainly consider texture and color information.
Most of the previous studies have been carried out to perform the detection and classification of different type of GI polyps in the field of CE. Generally, these methods followed a common approach of the feature extracting and then classification to detect and classify the GI polyps. In [19], Karargyris et al. proposed a geometric and texture features based method for the detection of small bowel polyps and ulcers in CE. Log Gabor filters and the SUSAN edge detector was used to preprocess the images and, finally, the geometric features were extracted to detect the polyp and ulcer region. Li et al. [20] utilized the advantages of a discrete wavelet transform and uniform local binary pattern (LBP) with a support vector machine (SVM) to classify the normal and abnormal tissues. In this feature extraction approach, wavelet transform combines the capability of multiresolution analysis and uniform LBP to provide robustness to illumination changes, which results in better performance.
Similarly, another texture features-based automatic tumor recognition framework was proposed in [6] for wireless CE images. In this framework, a similar integrated approach was adopted based on LBP and discrete wavelet transform to extract the texture features of the scale and rotation invariants. Finally, the selected features were classified by using an SVM. Yuan et al. [21] proposed an integrated polyps detection algorithm by combing the Bag of Features (BoF) method with the saliency map. In the first step, the BoF method characterizes the local features by using a scale-invariant feature transform (SIFT) feature vectors with k-means clustering. Then saliency features were obtained by generating saliency map histogram. Finally, both BoF and saliency features were fed into the SVM to perform classification. Later, Yuan et al. [22] extended this approach with the addition of LBP, uniform LBP (ULBP), complete LBP (CLBP), and histogram of oriented gradients (HoG) features along with SIFT features for capturing more discriminative texture information. Finally, these features were classified by using SVM and Fisher’s linear discriminant analysis (FLDA) classifiers by considering different combinations of local features. The combination of SIFT and CLBP features with SVM classifier resulted in top classification accuracy.
Seguí et al. presented a deep CNN system for small intestine motility characterization [7]. This CNN-based method exploited the general representation of six different intestinal motility events by extracting deep features, which resulted in superior classification performance when compared to the other handcrafted features-based methods. Another CNN-based CAD tool was presented in [15] to quantitatively analyze the celiac disease in a fully automated approach by using CE videos. This proposed method utilized the advantages of a well-known CNN model (i.e., GoogLeNet) to distinguish between the normal and abnormal (i.e., diagnosed with celiac disease) patients. Thus, the effective characterization of the celiac disease resulted in better diagnosis and treatment when compared to the manual analysis of CE videos. In [12], a multistage deep CNN-based framework for hookworm (i.e., intestinal parasite) detection was proposed using CE images. Two different CNN networks, named as edge extraction network and hookworm classification network, were unified, which simultaneously characterized the visual and tubular patterns of hookworms.
In the field of EGD, a deep learning-based CAD tool was proposed for the diagnosis of Helicobacter pylori (H. pylori) infection [9]. In this proposed framework, two-stage CNN models were used. In the first stage, a 22-layers deep CNN was fine-tuned for the classification (i.e., positive or negative) of H. pylori infection. Then, in the second stage, another CNN was used to further classify the dataset (EGD images) according to eight different anatomical locations. Similarly, Takiyama et al. proposed another CNN-based classification model to categorize the anatomical location of the human GI tract [8]. This technique could categorize the EGD images into four major anatomical locations (i.e., larynx, esophagus, stomach, and duodenum) and three subcategories for the stomach images (upper, middle, and lower regions). A pretrained CNN architecture, named as GoogLeNet, was used for this classification problem, which demonstrated high classification performance. In a recent study by Hirasawa et al. [13], a fully automated diagnostic tool for gastric cancer was proposed by utilizing the detecting capability of deep CNN-based architectures. A single-shot multibox detector (SSD) architecture was used to detect early and advanced stages of gastric cancer from EGD images. The proposed method demonstrated substantial detection capability even for small lesions when compared to the conventional methods. The results of this study illustrated its practical usability in clinical practice for better diagnosis and treatment. However, it demonstrated certain limitations as only high-quality EGD images could be used from the same type of endoscope and endoscopic video system.
Generally, the various deep learning-based methods demonstrate either the problem of over-fitting or under-fitting owing to the utilization of a large number of network parameters and the limited amount of data available in the training dataset. This problem degrades the system performance in a real-world scenario. A similar problem also occurs in the domain of medical image analysis owing to the unavailability of a sufficiently large training dataset. To address this issue, a transfer learning mechanism is often adopted in this domain. In the field of colonoscopy, Zhang et al. [10] used this approach for automatic detection and classification of colorectal polyps. A novel transfer learning approach was applied to train the two different CNN models for the source domain (i.e., nonmedical dataset) and then fine-tuning was performed for the target domain (i.e., medical dataset). Their method performed the polyp detection and classification tasks in two different stages. In the first stage, an image of interest (i.e., polyp image) was selected by using the CNN-based polyp detection model. In the second stage, another CNN model was further used to categorize the detected polyp image into either a hyperplastic polyp or an adenomatous colorectal polyp. The results of this study demonstrated that the CNN-based diagnoses achieved a higher accuracy and recall rate than endoscopist diagnoses. However, their method is not applicable for real-time colonoscopy image analysis owing to the use of multistage CNN models. Another study by Byrne et al. [14], presented a single deep CNN-based real-time colorectal polyp classification framework using the colonoscopy video images. In this study, a simple CNN model was trained to classify each input frame into one of four different categories, i.e., hyperplastic polyp, adenomatous polyp, no polyp, or unsuitable. The end-to-end processing time of this CNN model was 50 ms per frame, resulting in its applicability for the real-time classification of polyps. In another study [11], an offline and online three-dimensional (3D) deep CNN framework was proposed for automatic polyp detection. Two different 3D-CNNs, named as offline 3D-CNN and online 3D-CNN, were simultaneously used to exploit the more general representation of features for the task of effective polyp detection. In this complete framework, the offline 3D-CNN effectively reduced the number of false positives, whereas the online 3D-CNN was used to further improve the polyp detection. The experimental results showed that the 3D fully convolutional network was capable of learning more representative spatiotemporal features from colonoscopy videos in comparison with the handcrafted or two-dimensional (2D) CNN features-based methods.
Endoscopy is a direct imaging modality, which captures the internal structure of the human GI tract in the form of videos rather than a still image. Therefore, it is possible to extract both spatial and temporal information from endoscopic data to enhance the diagnostic capability of different deep CNN-based CAD tools. Most of the previous studies considered only the spatial information for classification and detection of different GI diseases without considering the temporal information. The loss of temporal information affects the overall performance of the CAD tools. In addition, the maximum number of classes to be managed in the previous studies were also limited to eight [9], which only considered limited GI diseases, such as a tumor or cancer.
To address these issues from previous researches, we considered 37 different categories in our proposed work, which included both normal and diseased cases related to different parts of the human GI tract. We proposed a novel two-stage deep learning-based framework to enhance the classification performance of different GI diseases by considering both spatial and temporal information. Two different models named as ResNet and LSTM were trained separately to extract the spatial and temporal features, respectively. In Table 1, the strengths and weaknesses of previous studies and our proposed method are summarized.

3. Contribution

This is the first approach towards the classification of multiple GI diseases that includes 37 different categories related to normal and diseased cases while considering different parts of the human GI tract. The major contributions of this study can be summarized in the following five ways when compared to the previous methods.
(1)
To the best of our knowledge, this is the first approach to develop a comprehensive deep learning-based framework for the classification of multiple GI diseases by considering deep spatiotemporal features. In contrast, most of the previous studies [6,7,8,9,10,11,12,13,14,15] considered the limited number of classes that are related to a specific type of GI portion.
(2)
We proposed a novel cascaded ResNet and LSTM-based framework in the medical domain to learn both spatial and temporal features for the different type of GI diseases. When compared to the previous methods based on handcrafted features and simple 2D-CNNs, our method can manage the large intraclass and low interclass variations among multiple classes more effectively.
(3)
We deeply analyzed the performance of our proposed method by selecting the multilevel spatial features for LSTM from the different layers of the ResNet network. Furthermore, the performance of multilevel spatial features was also analyzed by applying principal component analysis (PCA).
(4)
We compared the performance of the various state-of-the-art CNN models and different handcrafted feature-based approaches. Our analysis was more detailed, in contrast to previous studies [8,9], which provided only a limited performance analysis for a small number of classes related to a specific GI part.
(5)
Finally, we have ensure that our trained model and video indices of experimental endoscopic videos are publicly available through [18]; therefore, other researchers can evaluate and compare its performance.

4. Proposed Method

This section presents our proposed method for the classification of multiple GI diseases, including the CNN architecture for the extraction of spatial features, LSTM-based network for the extraction of temporal features, and finally, the classification portion comprises of fully connected (FC) layers.

4.1. Overview of the Proposed Approach

The conventional image or video classification framework is comprised of two main stages, known as the feature extraction stage and the classification stage. There are also certain other preprocessing steps such as image resizing or batch normalization (BN) to adjust the dataset according to the network compatibility. A brief flowchart of our method for the classification of multiple GI diseases based on deep spatiotemporal features is illustrated in Figure 1. In the first preprocessing step, the size of each endoscopic video frame was adjusted to 224   ×   224   ×   3 (according to the input layer size of the CNN model). In the next steps, we used a cascaded CNN and LSTM-based deep network to extract the spatial and temporal features, respectively, by using the resized sequence of frames. Using the CNN model, a sequence of spatial feature vectors was extracted, which was subsequently inputted to the LSTM for the extraction of temporal features. The final output of the LSTM comprises of a single feature vector that contains both the spatial and temporal information for each given sequence of frames. In the last step, the classification of the extracted spatiotemporal feature vector was performed by categorizing the given video sequence into one of 37 different categories (i.e., 37 different categories presenting the normal and diseased cases related to the human GI tract).

4.2. Structure of Our Proposed Model

Our proposed classification framework consists of a cascaded CNN and LSTM-based deep networks with the capability to classify the video data based on spatiotemporal features. The primary advantage of our network is its capability to categorize a variable length sequence of n successive images (i.e., I 1 , I 2 , I 3 , , I n ) with significant performance gain. For example, the use of more successive images results in better classification performance. In addition, our cascaded deep learning model demonstrated high performance in comparison with only CNN-based models. That is because the CNN models only extract the spatial information by processing each input image independently rather than considering both spatial and temporal features in the case of a video dataset. Owing to the loss of temporal information in a CNN model, the overall classification performance is deteriorated. To overcome the limitation of previous spatial features-based methods in the medical domain, our study included a spatial variant of a recurrent neural network (RNN) named as LSTM along with the conventional CNN model to enhance the classification performance. The overall structure of our proposed classification framework is shown in Figure 2. The complete framework is comprised of three different stages, i.e., spatial features extraction, temporal features extraction, and finally, the classification stage. In each stage, a specific set of deep learning procedures was applied to the given input sequence of endoscopic frames. Thus, the final class label was predicted for the input sequence using 37 different categories of different GI diseases. The detailed explanation of each stage is presented in the subsequent sections.

4.2.1. Spatial Features Extraction using a Convolutional Neural Network

The first stage of our proposed classification framework included a deep CNN model named ResNet18 [23], which was used for spatial features extraction from each input frame. The primary reasons for selecting ResNet18 [23] was the high classification accuracy and the optimal number of learnable parameters when compared to the other state-of-the-art deep CNN models [16,24,25,26,27]. In a later section the experimental results quantitatively illustrate the significance of our selected ResNet18 model when compared to the other models.
The complete structure of the extraction model for spatial features is illustrated in Figure 2. The entire network consists of multiple residual units, which can be considered as the basic building block. These residual units are categorized into two different types based on the type of shortcut connectivity (i.e., 1 × 1 convolutional-mapping-based shortcut connectivity and identity-mapping-based shortcut connectivity) [23]. The shortcut connectivity in an identity-mapping-based residual unit maintains the depth of previous feature map without any modification whereas the shortcut connectivity in the 1 × 1 convolutional-mapping-based residual unit increases the depth of the previous feature map by applying the 1 × 1 convolution. Moreover, in each residual unit, there are two convolutional layers with a filter size of 3 × 3 in sequential order. These filters contain the learnable parameters, which are optimized during the training procedure. ResNet18 consists of a total of eight residual units, including five identity mapping-based residual units and three 1 × 1 convolutional mapping-based residual units, as shown in Figure 3. The use of more identity mapping-based residual units results in better performance in terms of computational complexity and training time. In addition, both types of residual units result in smoother information propagation in both forward and backward directions [28].
The layer-wise structural details are further explained in Table 2, which demonstrates the flow of information processing by the different layers of ResNet18 in a sequential order. In general, the convolutional and FC layers are the main components of a conventional CNN model, which are used for features extraction and classification, respectively. There are also certain other layers without including the learnable parameters, such as a rectified linear unit (ReLU) layer, softmax, max pooling, average pooling, and a classification layer. Our selected ResNet18 model primarily contains a total of eighteen layers in which there are seventeen convolutional layers and one FC layer. These layers encompass the learnable parameters (i.e., filter coefficients and biases), which are optimized through the training procedure. Each convolutional layer is followed by the BN layer (it normalizes the feature map of each channel) and then a ReLU layer (it performs a threshold operation).
The first convolutional layer (i.e., Conv1) of our selected model generates an output feature map X 1 of size 112 × 112 × 64 by applying 64 different filters of size 7   ×   7   ×   3 over the given input image X . After Conv1, the next max pooling layer further processes the output feature map X 1 by applying a filter of 3 × 3 pixels and generates a down-sampled feature map X 2 of size 56 × 56 × 64 . This output feature map X 2 is passed through the first identity mapping-based residual unit that applies the two convolution filters (Conv2-1 and Conv2-2) in sequential order and generates an intermediate feature map as f ( X 2 , W 2 ) . Finally, the output feature map X 3 of size 56 × 56 × 64 is generated by adding X 2 and f ( X 2 , W 2 ) . The second identity mapping-based residual unit also performs a similar operation and converts the feature map X 3 to a new feature map X 4 . The next 1 × 1 convolutional mapping-based residual unit further processes the feature map X 4 by applying the two convolution filters (Conv4-1 and Conv4-2) in sequential order and generates the first intermediate output feature map as f ( X 4 , W 4 ) . Meanwhile, a 1 × 1 convolution filter (Conv4-3) converts the feature map X 4 to the second intermediate output feature map as h ( X 4 , W 4 ) . Finally, the output feature map X 5 is obtained by adding both intermediate feature maps f ( X 4 , W 4 ) and h ( X 4 , W 4 ) .
Similarly, all the successive residual units process the output feature map of the previous residual unit in the same way by using a different number of filters with different sizes and stride values as listed in Table 2. Finally, the optimal feature vector x of size 1 × 1 × 512 is obtained after applying the average pooling layer with filter size 7 × 7 pixels over the last output feature map X 10 (i.e., the output of the last convolutional layer). In this way, a set of n feature vectors { x 1 , x 2 , x 3 , x n } are obtained by processing all the successive images ( I 1 , I 2 , I 3 , I n ). These extracted feature vectors are further used as the input to the LSTM network for temporal feature extraction. The remaining three layers (i.e., FC, softmax, and classification layer) only participate in the training procedure. Therefore, after completing the training process, the output feature vector is selected after the average pooling layer for further temporal feature extraction and classification rather than the final classification layer.

4.2.2. Temporal Features Extraction by Long Short-term Memory Model

In the second stage, LSTM, a variant of the RNN model [29], was used to exploit the temporal information from the set of n features vectors that were extracted in the first stage by using ResNet18. The structure of LSTM consists of n LSTM cells [30]. Figure 2 (Stage 2) illustrates the flow of n features vectors ( x 1 , x 2 , x 3 , x n ) through the multiple LSTM cells. In the figure, h n and c n denote the output (also known as the hidden state) and cell state at time step n , respectively. The hidden state, h n , contains the output of the LSTM cell for the time step n and the cell state c n holds the information learned from all the previous time steps (i.e., 1   to   n 1 ). The first LSTM cell (at time step n = 1 ) uses the initial state of the network ( h 0 , c 0 ) and the input feature vector x 1 to compute the first output h 1 and the updated cell state c 1 . At time step n (where n 1 ), the LSTM cell uses the current state of the network ( h n 1 , c n 1 ) and the input feature vector x n to calculate the output h n and the updated cell state c n . Thus, the temporal information is exploited in the LSTM stage by using all the spatial feature vectors.
The basic structure of a standard LSTM cell is shown in Figure 4, which illustrates the flow of data at time step n . In general, four components, named as input gate ( i n ) , forget gate ( f n ) , cell candidate ( g n ) , and output gate ( o n ) , are responsible for controlling the state information at time step n . The i n controls the level of the cell state update, whereas the f n controls the level of the cell state reset. The g n adds the information to the cell state and finally, the o n controls the level of the cell state added to the hidden state. Based on these components, the complete structure of the cell is divided into three gates, named as forget, input, and output gates, as highlighted in Figure 4.
Furthermore, the three different type of learnable parameters, termed as input weights, W   =   [ W i n , W f n , W g n , W o n ] T , recurrent weights, R   =   [ R i n , R f n , R g n , R o n ] T , and bias, b   =   [ b i n , b f n , b g n , b o n ] T , are included in the LSTM cell, which are responsible for learning the temporal information after performing sufficient training. These learnable parameters ( W , R , b ) and cell components ( i n , f n , g n , o n ) are used to calculate the cell state c n and output h n at time step n . The following mathematical computations are performed to determine the state information and cell components:
c n   =   f n × c n 1 + g n × i n
h n   =   o n × t a n h ( c n )
i n   =   σ ( W i n x n + R i n h n 1 + b i n )
f n   =   σ ( W f n x n + R f n h n 1 + b f n )
g n   =   t a n h ( W g n x n + R g n h n 1 + b g n )
o n   =   σ ( W o n x n + R o n h n 1 + b o n )
where t a n h is the hyperbolic tangent function, which is calculated as t a n h ( x ) = ( e x e x ) / ( e x + e x ) . It is used as a state activation function. The function σ is the sigmoid function, which is calculated as σ ( x )   =   ( 1 + e x ) 1 to compute the gate activation function.
In the first stage, ResNet18 processed the sequence of n successive images (i.e., I 1 , I 2 , I 3 , I n ) in a sequential order to extract the spatial features. Then, the LSTM model processed all the spatial feature vectors (a set of n feature vectors { x 1 , x 2 , x 3 , x n } ) in a parallel fashion in the second stage. Therefore, the feature accumulation block, as shown in Figure 2, is used to accumulate all the spatial feature vectors (obtained from ResNet18 in the first stage) before inputting it to the LSTM model in the second stage. The layer-wise structural details of our proposed LSTM model are listed in Table 3. The final output of the LSTM model contains both the spatial and temporal information, which is followed by the stack of FC layers to perform the final classification.

4.2.3. Classification

In the final classification stage, the output h n of the LSTM cell at the last time step n is selected as the final output feature vector rather than using all the outputs (i.e., h 1 , h 2 , h 3 , , h n ). Then, a stack consisting of FC, softmax, and classification layers is used to perform the final classification as shown in Figure 2. The output of the last LSTM cell is followed by a FC layer where the number of nodes is equal to the number of classes. The primary purpose of the FC layer is to determine the larger patterns by combining all the spatiotemporal features learned by the previous layers across the images. It multiplies the input feature vector obtained from the last LSTM cell by a weight matrix W and then adds a bias vector b . The final output obtained after this FC layer is presented as y   =   W · h n + b . The next softmax layer converts the output y in terms of probability by applying the softmax function [31]. Finally, the classification layer considers the output from the softmax layer and assigns each input to one of the 37 different categories by using the cross-entropy loss function [31]. In conclusion, the final class label is assigned to the given sequence of n successive images by exploiting both the spatial and temporal information.

5. Experimental Setup and Performance Analysis

In this section, we analyze the performance of our proposed ResNet18 and LSTM-based classification framework. We provide the details of the selected endoscopy dataset, experimental configurations, various performance analysis metrics used to evaluate the quantitative performance, observations, and analysis of the results as well as the comparison with other methods.

5.1. Dataset

To evaluate the performance of the proposed multiple GI diseases classification framework, we selected an open access endoscopic videos dataset from Gastrolab [32] and the KVASIR dataset [33]. The datasets contain various endoscopic videos related to different parts of the human GI tract, including both normal and diseased cases. The details of each individual video (including the information about normal and diseased cases as well as the anatomical district) are included as the video name. Based on the available information, the complete dataset was categorized into 37 different classes including both normal and diseased cases related to different parts of the human GI tract. These different classes include the multiple anatomical locations (i.e., esophagus, stomach, small intestine, large intestine, and rectum) of the human GI tract as shown in Figure 5.
Furthermore, the details of multiple subcategories of each anatomical district and their corresponding number of classes with types of diseases and the number of training and testing sequences are listed in Table 4. The entire dataset contains a total of 77 video files including 52,471 frames. In the preprocessing part, all these frames were resized into fixed dimensions with the spatial size of 224 × 224 ; subsequently, they were converted into a standard bitmap file format. We performed the two-fold cross-validation by randomly dividing the entire dataset as 50% for training and the remaining 50% for testing. That is, in all the performance comparisons, the numbers of training data are the same as those of the testing data as shown in Table 4.
In the first stage, an online data augmentation [34] process (including random translation and in-plain rotation) was used to solve the class imbalance problem [35] caused by the different number of training samples in each class. The data augmentation process was performed only for the training dataset in the first stage (i.e., spatial features extraction using ResNet18), and was not performed for the testing dataset.
The visual representation of our selected dataset for each class is shown in Figure 6. In this diagram, each individual image presents a specific class from the total of 37 different classes (i.e., C1, C2, C3, …, C37). The primary challenge in our selected dataset was the high intra-class variance caused by the different types of lesion structures and texture properties within the same class as depicted in Figure 7. Furthermore, different viewing conditions and dynamic structural changes during the endoscopy procedure may also increase the intra-class variance. To solve this problem, a high level of abstraction was required to present the common characteristics of such types of datasets with high intra-class variance. In addition, a sufficient amount of training dataset related to a particular domain can also enhance the overall performance of the CAD systems. This type of dataset aids in analyzing the performance of our proposed framework in a challenging scenario.

5.2. Experimental Setup and Training

The proposed framework was implemented with MATLAB R2018b (MathWorks, Inc., Natick, MA, USA) [36] on a Windows 10 operating system. The deep learning library named as deep learning toolbox was included in MATLAB for the implementation of various CNN models [37]. Any people who purchase MATLAB R2018b [36] can use this library with the licenses based on the credits to the authors of the CNN models. All the experiments were performed on a desktop computer with a 3.50 GHz Intel® (Santa Clara, CA, USA) Core-i7-3770K central processing unit (CPU) [38], 16 GB random access memory (RAM), and an NVIDIA (Santa Clara, CA, USA) GeForce GTX 1070 graphics card [39]. The use of the graphics card provides the parallel processing capability for both the training and the testing phase.
As explained in Section 4, our proposed method combined two types of image features for classification of multiple GI diseases, i.e., the spatial features extracted by a deep CNN model in the first stage, and then the temporal features that were extracted by using the LSTM model in the second stage. Both the networks were trained separately by using the stochastic gradient descent [40] optimizer method, which is generally used for optimal training of CNNs. It is a more efficient back propagation algorithm for learning the discriminative linear classifiers by using a convex loss function. Its primary goal is to optimize the learnable parameters of the model (i.e., filter weights and biases) by considering the derivative of the loss function. In addition, we initialized the parameters of the first stage CNN model by using a pretrained ResNet18 model, which was successfully trained on the ImageNet dataset [41]. This scheme was widely used in previous studies to initialize the network parameters to make the network training process easier and time effective. In the case of the LSTM model, the initial weights were randomly initialized by using a Gaussian distribution with zero mean and 0.001 standard deviation, and the biases were initialized to zero. In Table 5, the parameters of the training procedure used in our experiments are listed.
The performance of our proposed method was evaluated by performing the cascaded training of our ResNet18 and LSTM-based classification framework. In the first stage, we performed the training of ResNet18 by using the training dataset (as listed in Table 4). Figure 8 shows the progress of training loss and accuracy according to the different number of epochs for both folds of cross-validations. The training loss approaches zero after a certain number of epochs, and the training accuracy approaches 100%, which illustrate that our selected model is sufficiently trained. In addition, after performing several training experiments for different CNN models, we determined that the fine-tuning of a pretrained model results in faster convergence rather than training from scratch. In other words, we used the ResNet18 model which was pretrained with the ImageNet dataset [41]. Then, we performed the fine-tuning of this model with our training dataset of Table 4. Therefore, we selected a pretrained model of ResNet18 for spatial feature extraction in the first stage. Moreover, the average accuracy of our selected ResNet18 based on the spatial features was higher than other deep CNN models. Thus, both the ResNet18 and LSTM models were interconnected in a cascaded fashion, and separate trainings were performed for both networks. The second stage training process was started after completing the training for the ResNet18 model.
In the second stage, the output feature vectors (extracted from the trained ResNet18 model in the first stage using the training dataset) were used to train our proposed LSTM model. In this stage, each training sample comprised of a set of n feature vectors (extracted from n successive frames in the first stage) instead of a single feature vector. Thus, an intermediate features-based dataset was generated from the extracted feature vectors, which was further used for temporal feature extraction. In our experiment, a total of fifteen (i.e., n   =   15 ) successive frames were used to generate a set of fifteen feature vectors for each training sample. Figure 9 shows the progress of training loss and accuracy for both folds of cross-validations. The training loss approaches to zero after a certain number of iterations in the first epoch and the training accuracy approaches 100%, which shows the optimal convergence of the second stage (LSTM) of our model. In Figure 9, it can also be observed that the convergence of LSTM is faster and smoother when compared to ResNet18 (in the first stage). The primary reason for this result is the use of an intermediate dataset (i.e., a set of discriminative spatial feature vectors) for temporal feature extraction rather than using the successive frames.

5.3. Evaluation of the Performance by Proposed Method

5.3.1. Performance Analysis Metric

We employed average accuracy, F1 score, mean average prevision (mAP), and mean average recall (mAR) [42] to quantitatively evaluate the performance of our proposed ResNet18 and LSTM-based classification model. Based on these four parameters, we evaluated the overall performance of the model by calculating the average value for all the classes. These four metrics are defined as:
A c c u r a c y   =   1 K k   =   1 K T P k + T N k T P k + T N k + F P k + F N k
F 1. S c o r e   =   2 × m A P × m A R m A P + m A R
m A P   =   1 K k   =   1 K T P k T P k + F P k
m A R   =   1 K k   =   1 K T P k T P k + T N k
where T P k , F P k , T N k , and F N k denote the number of true positives, false positives, true negatives, and false negatives, respectively, for each class k . The value of T P k presents the number of correctly classified images from class k , F P k shows the number of images that are misclassified as belonging to class k . T N k indicates the number of images correctly classified that do not belong to class k and F N k denotes the number of misclassified images that actually belong to class k . Here K denotes the total number of classes, which is equal to 37 in our research.

5.3.2. Testing of the Proposed Method

The length of successive frames performs an important role in the system performance. The small number of successive frames results in low temporal information, whereas the long sequence length increases the processing time and the effects of noise. Therefore, we performed the training of our LSTM model for thirty different number of frames (i.e., n   =   1 , 2 , 3 , , 30 ). Then, the testing performance was evaluated for each step size. Figure 10 shows the average performance results according to different number of frames. In Figure 10, the green square box indicates the maximum average performance whereas the red square box illustrates the maximum performance with respect to different performance metrics (i.e., accuracy, F1 score, mAP, and mAR). Finally, based on the overall maximum average performance, we determined that the best accuracy could be obtained when the numbers of frame was 15 ( n   =   15 ).
As the next experiment, we performed a layer-wise performance comparison between our method (ResNet18 + LSTM) and only a ResNet18 model by selecting the features from the different parts of the network. Moreover, this additional experiment was also used to investigate the more discriminative features at certain intermediate layers that could result in better performance. For this experiment, the output feature vectors were extracted from five different layers (i.e., Conv6-2, Conv7-2, Conv8-2, Conv9-2, Avg. pooling, as listed in Table 2) of ResNet18 with the feature map size of 14 × 14 × 256 (50,176), 14 × 14 × 256 (50,176), 7 × 7 × 512 (25,088), 7 × 7 × 512 (25,088), and 1 × 1 × 512 (512), respectively. In the case of our method, the classification performance for each layer was obtained by further extracting the temporal information from the LSTM model using these features. The layer-wise features from ResNet18 model were classified using a k-nearest neighbor (KNN) classifier, which is widely used for pattern classification [43]. The complete layer-wise performances of our method and ResNet18 are listed in Table 6. Based on the overall performance, we concluded that the deeper features result in better classification performance in the case of our method and the ResNet18 model. However, the layer-wise performance of our method was still higher than the conventional ResNet18.
Moreover, our method also showed a high average accuracy of 90.48% and mAP of 91.29% when a still image (i.e., n   =   1 ) was used, which are higher values when compared to other CNN-based methods (an accuracy of 89.95% and mAP of 90.72% in the case of conventional ResNet18).
The extracted features from the last average pooling layer of ResNet18 were further analyzed by applying PCA [44] technique as a post processing step. The main objective of this analysis was to explore the discriminative nature of the features (i.e., to check if our selected features were distinctive or redundant). For this purpose, all the extracted features of dimension 1 × 512 from the last average pooling layer were projected to the eigenspace by applying the PCA. This eigenspace presented all the input feature vectors in a new coordinate system in a more distinctive way. The dimensions of these newly obtained features are selected based on the maximum variance (i.e., greater than 99%) of the projected data on all the possible axes. The eigenvalue corresponding to each feature vector was used to select a feature vector. In the case of our dataset, a new set of feature vectors (with the feature dimension 1 × 136 ) was obtained by selecting a total of 136 eigenvectors with the highest eigenvalues. In our proposed model, this new set of feature vectors were further used as inputs to the LSTM model to explore the temporal information and then the final classification performance was obtained as listed in Table 7. In addition, the PCA feature-based performance was evaluated for ResNet18 by using the KNN classifier, which is also presented in Table 7. According to these final classification results, we concluded that the PCA-based features reduced the performance in both cases (i.e., our proposed model and ResNet18), whereas the original high dimension features resulted in better performance. Finally, it can be concluded that our extracted features (from the last average pooling layer) were already diverse, and the performance of our method was still high in comparison with conventional ResNet18 after applying the PCA.
Figure 11 illustrates the more comprehensive classification performance of our model in terms of the confusion matrix. It can be observed from these results that only a few classes (i.e., C16, C31, C33, C34) showed a low classification performance owing to the high inter class similarities in terms of lesion textures or GI organ structures. However, the overall performance of our proposed method was significantly high for a dataset with several classes.

5.3.3. Comparisons with Previous Methods

The performance of our proposed ResNet18 and LSTM-based methods were compared with the various state-of-the-art deep CNN-based CAD tools that are used in the endoscopy domain [8,12,14,15]. To ensure a fair comparison, the performances of all the existing baseline methods were evaluated with our selected dataset using the same training and testing data of two-fold cross-validation. In a recent study related to endoscopy, two different CNN models—GoogLeNet [8,12,15] and InceptionV3 [14]—were primarily used in the diagnosis of various type of GI diseases. Therefore, the performance of these two models were evaluated in comparison with our proposed method. The experimental results showed that our method outperformed these two deep CNN models [8,12,14,15] with significant performance gain as listed in Table 8.
Further, we also compared the performance of our method with the recent CNN models [16,23,24,25] used in image classification domains other than endoscopy. The main objective of these comparisons was to estimate the performance of the existing state-of-the-art CNN models in the endoscopy image analysis domain. The complete experimental results for all the selected baseline methods are listed in Table 8. These results confirm that our proposed ResNet18 and LSTM-based method shows the highest performance in the endoscopy image analysis domain for the classification of multiple GI diseases.
The discriminative ability of our proposed method, in contrast with other baseline methods, can also be observed through the receiver operating characteristics (ROC) curve (an effective measure used to evaluate the diagnostic ability of a model). It is created by plotting the true positive rate (known as the probability of detection) against the false positive rate (known as the probability of false alarm) at various threshold settings. From Figure 12, it can be observed that our proposed method also shows the highest value for the area under the curve (AUC) with a value of 97.057% in comparison with all the other selected baseline methods (i.e., SqueezeNet: 82.131%, AlexNet: 87.328%, GoogLeNet: 91.097%, VGG19: 92.039%, VGG16: 93.060%, InceptionV3: 95.000%, ResNet50: 95.924%, and ResNet18: 95.705%). All these ROC curves are presented by the average values obtained from two-fold cross-validations. The figure on the left side provides an enlarged view to illustrate the performance difference more clearly.
The complete parametric and structural details of our proposed model and the other selected models are listed in Table 9. The AUC performance of ResNet18 is comparable with the second-best model named as ResNet50, as shown in Figure 12; however, the training parameters of ResNet18 are significantly less than half of that of ResNet50, as listed in Table 9. Therefore, we adopted the ResNet18 architecture as the backbone model to extract the spatial features, which are further used as inputs to the LSTM model to exploit the temporal information. In our proposed framework, the total learnable parameters were approximately 13.17M (including both ResNet18 and LSTM), which were still significantly lower than the second-best model (i.e., ResNet50) as shown in Table 9.
Furthermore, a sensitivity analysis was performed to evaluate the robustness of our method and other CNN models. A Monte Carlo simulation step [27] was performed to analyze this sensitivity performance. In this simulation setup, the performance of each individual CNN model was evaluated in an iterative way by randomly selecting 20% of the complete testing dataset as a subset of the testing dataset. A total of 200 iterations were performed for both folds of cross-validations. Finally, the average performance (i.e., average accuracy, F1 score, mAP, and mAR) as well as standard deviation were obtained for each model. The overall sensitivity performance of our method and all the selected models are illustrated in Figure 13. It can be observed in Figure 13a–d that the overall sensitivity performance of our proposed method is higher while considering average accuracy, F1 score, mAP, and mAR when compared to all the existing baseline models.
A t-test performance analysis [46] was further performed to illustrate the significance of the performance difference between our method and ResNet18. The reason why the t-test performance analysis was performed only against ResNet18 is because ResNet18 shows the second-best accuracy as shown in Table 8. In general, this performance analysis is often used to illustrate the performance difference between two systems or algorithms in a more discriminative way. It is based on a null hypothesis ( H ), which assumes that there is no performance difference (i.e., H   =   0 ) between two models. Then, a rejection score (p-value) is calculated to check the validity of the null hypothesis based on the performance of the two models (in this case, our method and the second-best model). Figure 14 illustrates the t-test performance (for the values of mean ( μ ), standard deviation ( ρ ), and p-value) for our method and the second-best model. These results were calculated for all the performance measures. The obtained rejection scores (p-values) in case of the average accuracy, F1 score, mAP, and mAR were 1.51 × 10 43 , 6.87 × 10 20 , 4.67 × 10 10 , and 1.03 × 10 33 , respectively. All these p-values are less than 0.01, which indicate that the null hypothesis is rejected (i.e., H 0 ) at a 99% confidence score for all the performance metrics. Based on these results, it can be concluded that there is a significant performance difference between our method and the second-best method. Furthermore, the higher mean ( μ ) performance of our method indicates its superiority over the second-best baseline model.
We also performed Cohen’s d [47] analysis, by which the size of the difference between the two groups were demonstrated using the effect size [48]. Cohen’s d analysis is widely used for analyzing the difference between two measured values. Generally, Cohen’s d is classified as small at approximately 0.2–0.3, as medium at approximately 0.5, and as large at greater than or equal to 0.8. For example, if the calculated Cohen’s d is closer to 0.2–0.3 than 0.5 and 0.8, we can say that the difference between measured values has a small effect size. If the calculated Cohen’s d is closer to 0.8 than 0.2–0.3 and 0.5, we can say that the difference between measured values has a large effect size. The calculated Cohen’s d values for the performance of the two models (our method and the second-best model) were approximately 1.57 (closer to 0.8), 0.96 (closer to 0.8), 0.64 (closer to 0.8), and 1.33 (closer to 0.8) for average accuracy, F1 score, mAP, and mAR, respectively. Consequently, we concluded that the difference in the performances between our method and the second-best model has a large effect while considering the average accuracy, F1 score, mAP, and mAR.
In this section, we present the performances of various handcrafted feature-based methods that were also compared with our proposed CNN and LSTM-based classification framework for further comparison. In this comparison, three known handcrafted feature extraction methods, named as LBP [49], histogram of oriented gradients (HoG) [50], and multilevel LBP (MLBP) [51], were considered. Then, the extracted features from each method were classified by using four different classifiers: adaptive boosting (AdaBoostM2) [52], multiclass SVM (multi-SVM) [53], random forest (RF) [54], and KNN. All these handcrafted feature-based methods exploit the low-level features (i.e., edge or corner information). We evaluated the performance of 12 different handcrafted feature-based classification methods for our selected dataset to obtain a fair comparison. The detailed results for all these classification methods are listed in Table 10.
Among all these handcrafted feature extraction and classification methods, HoG + RF (i.e., HoG feature extraction method followed by the RF classifier) demonstrated superior performance. Hence, the HoG feature extraction method exploited the more discriminative low-level features in comparison with the other two methods. Furthermore, the RF classifier considers a tree structure to determine the classification decision, which resulted in a better performance and controlled the over-fitting problem. However, there is a significant performance difference between our method and the best handcrafted feature-based method (HoG + RF). Our proposed method outperformed all the handcrafted feature-based methods.

6. Discussion

Our proposed deep CNN and LSTM-based classification framework shows the best performance with a high AUC of 97.057%. This remarkable performance of our proposed system increases its usability in the diagnosis of several GI diseases by automatically detecting different types of GI lesions or abnormalities, such as polyps, ulcers, or cancers from endoscopic videos. Our AI-based CAD system can assist the physicians in an effective diagnosis and treatment of many complex GI diseases. Furthermore, the classification of the endoscopic videos can, itself, be beneficial in retrieving the previously stored videos related to the current situation of a patient. Thus, the past cases can provide a path toward correct diagnostic decision. Therefore, we can also utilize our proposed classification framework for efficient endoscopic video frame retrieval by using the predicted class labels. The overall block diagram for our class prediction-based retrieval system is shown in Figure 15. In this retrieval section, the first step is to predict the actual class for the given query (i.e., successive endoscopic video frames). To predict the actual class label, a probability score corresponding to each class label is obtained for the given query by using our proposed classification framework. Based on the highest probability score, the corresponding class label is chosen as the actual class label. In the second step, the relevant cases related to input query frames are explored only within the predicted class based on feature matching. In this feature matching stage, the extracted spatiotemporal feature vector from the input query frames is matched one by one with the feature database of that predicted class by calculating the Euclidean distance. Based on the minimum distance, the frame index (i.e., name or ID information) is selected. Finally, the relevant frame is retrieved from the database by using the frame index information obtained in previous stage.
A few correctly retrieved examples are illustrated in Figure 16 by using our class prediction-based retrieval system. It can be observed that the retrieved endoscopic frames have high intra-class variance with varying illumination and contrast. However, our proposed system still outperforms with 100% retrieval performance for all the selected cases. Moreover, the classification performance for these selected example cases is also 100%, which can be observed in Figure 11 (confusion matrix performance for each class). Further, Figure 17 shows the probability score corresponding to each input query. It can be observed that the highest probability score is obtained for the actual predicted class, which shows that the proposed classification model is capable of extracting the discriminative features for the given query. In conclusion, this significant performance gain (in both classification and retrieval sections) shows that our method can be robust to the high intra-class variance of a dataset.
There are a few classes in our selected dataset that show the low retrieval performance, as shown in Figure 18. The primary reason for this performance degradation is the anatomical structural overlapping and identical shape of different GI lesions among different classes. Figure 18a shows a few incorrectly retrieved results as C30 (i.e., tuber adenoma in sigmoid colon) and C32 (i.e., ulcerative colitis in rectosigmoid part of large intestine) are retrieved for an input query of C16 (i.e., severe Crohn’s disease in terminal ileum of small intestine). It can be observed from Figure 18a that the lesion characteristics among these three classes (i.e., C16, C30, and C32) show a resemblance that may cause the incorrect retrieval. Similarly, certain other incorrect retrieval cases were obtained for an input query of C31 (i.e., polypoid cancer in sigmoid colon), C33 (i.e., severe Crohn’s disease in the rectum), and C34 (i.e., adenocarcinoma in the rectum) owing to identical lesion characteristics, as shown in Figure 18b–d. Moreover, Figure 19 shows the probability score corresponding to each input query in which significantly higher probability scores can be observed corresponding to multiple predicted class labels. These multiple higher scores show the structural or lesion similarities among multiple classes, which can result in classification errors. However, the retrieval performance in these cases can be enhanced by exploring the input query in multiple classes, which can be selected based on a multiple probability scores that is greater than a certain threshold.

7. Conclusions

In this paper, a novel CNN and LSTM-based classification framework was proposed for the classification of multiple GI diseases using endoscopic videos. Moreover, our proposed classification framework is further utilized to design a class prediction-based endoscopic video retrieval system. The proposed spatiotemporal features-based method is capable of encoding more discriminative representations of multiple endoscopy scans when compared to the features learned only from spatial information. Therefore, both spatial and temporal information results in better classification and retrieval performance. The performance of the proposed method was evaluated thoroughly using a publicly available dataset from GastroLab as well as the KVASIR database. Moreover, the same dataset and experimental protocol was adopted for the various state-of-the-art methods to make a fair comparison. The proposed method achieved 97.057% area under the curve as the best results, together with an average accuracy of 92.57%, F1 score of 93.41%, mAP of 94.58%, and mAR of 92.28. In addition, the obtained t-test rejection scores (p-values) of our proposed and second-best method are less than 0.01 ( 1.51 × 10 43 , 6.87 × 10 20 , 4.67 × 10 10 , and in the case of the average accuracy, F1 score, mAP, and mAR, respectively), which indicate that the null hypothesis is rejected (i.e., H 0 ) at a 99% confidence score for all the performance metrics. After performing a detailed analysis, we observed that our method consistently achieved high classification performance in comparison with various state-of-the-art deep CNN and handcrafted features-based methods of LBP, HoG, and MLBP. The classification and retrieval performance of the proposed system reveals its applicability to clinical diagnosis, treatment, education, and research. We also ensured that our trained model is publicly available to aid other researchers in performance comparisons.
As a future work, we are planning to increase the dataset by considering more than 37 classes. In addition, we are planning to perform the real-time detection of small lesions using an endoscopic video. We also plan to improve the overall classification performance by combing multiple deep CNN models.

Author Contributions

M.O. and K.R.P. designed the overall system. In addition, they wrote and revised the paper. M.A., J.C., and T.M. helped to design the comparative analysis and experiments.

Acknowledgments

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2018R1D1A1B07041921), by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (NRF-2019R1F1A1041123), and by the Bio & Medical Technology Development Program of the NRF funded by the Korean government, MSIT (NRF-2016M3A9E1915855).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Siegel, R.L.; Miller, K.D.; Jemal, A. Cancer Statistics. Ca-Cancer J. Clin. 2016, 66, 7–30. [Google Scholar] [CrossRef] [PubMed]
  2. Vuong, Q.-H.; Ho, M.-T.; Vuong, T.-T.; La, V.-P.; Ho, M.-T.; Nghiem, K.-C.P.; Tran, B.X.; Giang, H.-H.; Giang, T.-V.; Latkin, C.; et al. Artificial intelligence vs. natural stupidity: Evaluating AI readiness for the vietnamese medical information system. J. Clin. Med. 2019, 8, 168. [Google Scholar] [CrossRef] [PubMed]
  3. Tran, B.X.; Vu, G.T.; Ha, G.H.; Vuong, Q.-H.; Ho, M.-T.; Vuong, T.-T.; La, V.-P.; Ho, M.-T.; Nghiem, K.-C.P.; Nguyen, H.L.T.; et al. Global evolution of research in artificial intelligence in health and medicine: A bibliometric study. J. Clin. Med. 2019, 8, 360. [Google Scholar] [CrossRef] [PubMed]
  4. Owais, M.; Arsalan, M.; Choi, J.; Park, K.R. Effective diagnosis and treatment through content-based medical image retrieval (CBMIR) by using artificial intelligence. J. Clin. Med. 2019, 8, 462. [Google Scholar] [CrossRef] [PubMed]
  5. Cheng, C.-H.; Liu, W.-X. Identifying degenerative brain disease using rough set classifier based on wavelet packet method. J. Clin. Med. 2018, 7, 124. [Google Scholar] [CrossRef] [PubMed]
  6. Li, B.; Meng, M.Q.-H. Tumor recognition in wireless capsule endoscopy images using textural features and SVM-based feature selection. IEEE Trans. Inf. Technol. Biomed. 2012, 16, 323–329. [Google Scholar] [PubMed]
  7. Seguí, S.; Drozdzal, M.; Pascual, G.; Radeva, P.; Malagelada, C.; Azpiroz, F.; Vitrià, J. Generic feature learning for wireless capsule endoscopy analysis. Comput. Biol. Med. 2016, 79, 163–172. [Google Scholar] [CrossRef] [Green Version]
  8. Takiyama, H.; Ozawa, T.; Ishihara, S.; Fujishiro, M.; Shichijo, S.; Nomura, S.; Miura, M.; Tada, T. Automatic anatomical classification of esophagogastroduodenoscopy images using deep convolutional neural networks. Sci. Rep. 2018, 8, 1–8. [Google Scholar] [CrossRef]
  9. Shichijo, S.; Nomura, S.; Aoyama, K.; Nishikawa, Y.; Miura, M.; Shinagawa, T.; Takiyama, H.; Tanimoto, T.; Ishihara, S.; Matsuo, K.; et al. Application of convolutional neural networks in the diagnosis of Helicobacter pylori infection based on endoscopic images. EBioMedicine 2017, 25, 106–111. [Google Scholar] [CrossRef]
  10. Zhang, R.; Zheng, Y.; Mak, T.W.C.; Yu, R.; Wong, S.H.; Lau, J.Y.W.; Poon, C.C.Y. Automatic detection and classification of colorectal polyps by transferring low-level CNN features from nonmedical domain. IEEE J. Biomed. Health Inf. 2017, 21, 41–47. [Google Scholar] [CrossRef]
  11. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.A. Integrating online and offline three-dimensional deep learning for automated polyp detection in colonoscopy videos. IEEE J. Biomed. Health Inf. 2017, 21, 65–75. [Google Scholar] [CrossRef] [PubMed]
  12. He, J.-Y.; Wu, X.; Jiang, Y.-G.; Peng, Q.; Jain, R. Hookworm detection in wireless capsule endoscopy images with deep learning. IEEE Trans. Image Process. 2018, 27, 2379–2392. [Google Scholar] [CrossRef] [PubMed]
  13. Hirasawa, T.; Aoyama, K.; Tanimoto, T.; Ishihara, S.; Shichijo, S.; Ozawa, T.; Ohnishi, T.; Fujishiro, M.; Matsuo, K.; Fujisaki, J.; et al. Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images. Gastric Cancer 2018, 21, 653–660. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Byrne, M.F.; Chapados, N.; Soudan, F.; Oertel, C.; Perez, M.L.; Kelly, R.; Iqbal, N.; Chandelier, F.; Rex, D.K. Real-time differentiation of adenomatous and hyperplastic diminutive colorectal polyps during analysis of unaltered videos of standard colonoscopy using a deep learning model. Gut 2019, 68, 94–100. [Google Scholar] [CrossRef] [PubMed]
  15. Zhou, T.; Han, G.; Li, B.N.; Lin, Z.; Ciaccio, E.J.; Green, P.H.; Qin, J. Quantitative analysis of patients with celiac disease by video capsule endoscopy: A deep learning method. Comput. Biol. Med. 2017, 85, 1–6. [Google Scholar] [CrossRef]
  16. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  17. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  18. Dongguk, CNN and LSTM Models for the Classification of Multiple Gastrointestinal (GI) Diseases, and Video Indices of Experimental Endoscopic Videos. Available online: http://dm.dgu.edu/link.html (accessed on 15 June 2019).
  19. Karargyris, A.; Bourbakis, N. Detection of small bowel polyps and ulcers in wireless capsule endoscopy videos. IEEE Trans. Biomed. Eng. 2011, 58, 2777–2786. [Google Scholar] [CrossRef]
  20. Li, B.; Meng, M.Q.-H. Automatic polyp detection for wireless capsule endoscopy images. Expert Syst. Appl. 2012, 39, 10952–10958. [Google Scholar] [CrossRef]
  21. Yuan, Y.; Meng, M.Q.-H. Polyp classification based on bag of features and saliency in wireless capsule endoscopy. In Proceedings of the IEEE International Conference on Robotics and Automation, Hong Kong, China, 31 May–7 June 2014; pp. 3930–3935. [Google Scholar]
  22. Yuan, Y.; Li, B.; Meng, M.Q.-H. Improved bag of feature for automatic polyp detection in wireless capsule endoscopy images. IEEE Trans. Autom. Sci. Eng. 2016, 13, 529–535. [Google Scholar] [CrossRef]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  24. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360v4. [Google Scholar]
  25. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar]
  26. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  27. Raychaudhuri, S. Introduction to Monte Carlo simulation. In Proceedings of the IEEE Winter Simulation Conference, Miami, FL, USA, 7–10 December 2008; pp. 91–100. [Google Scholar]
  28. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 630–645. [Google Scholar]
  29. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  30. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  31. Heaton, J. Artificial Intelligence for Humans; Deep Learning and Neural Networks; Heaton Research, Inc.: St. Louis, MO, USA, 2015; Volume 3. [Google Scholar]
  32. Gastrolab—The Gastrointestinal Site. Available online: http://www.gastrolab.net/ni.htm (accessed on 1 February 2019).
  33. Pogorelov, K.; Randel, K.R.; Griwodz, C.; Eskeland, S.L.; de Lange, T.; Johansen, D.; Spampinato, C.; Dang-Nguyen, D.-T.; Lux, M.; Schmidt, P.T.; et al. KVASIR: A multi-class image dataset for computer aided gastrointestinal disease detection. In Proceedings of the 8th ACM Multimedia Systems Conference, Taipei, Taiwan, 20–23 June 2017; pp. 164–169. [Google Scholar]
  34. Peng, X.; Tang, Z.; Yang, F.; Feris, R.S.; Metaxas, D. Jointly optimize data augmentation and network training: Adversarial data augmentation in human pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2226–2234. [Google Scholar]
  35. Wong, S.C.; Gatt, A.; Stamatescu, V.; McDonnell, M.D. Understanding data augmentation for classification: When to warp? In Proceedings of the IEEE International Conference on Digital Image Computing: Techniques and Applications, Gold Coast, Australia, 30 November–2 December 2016; pp. 1–6. [Google Scholar]
  36. MATLAB R2018b. Available online: https://www.mathworks.com/products/matlab.html (accessed on 1 February 2019).
  37. Deep Learning Toolbox. Available online: https://in.mathworks.com/products/deep-learning.html (accessed on 1 July 2019).
  38. Intel® Core i7-3770K Processor. Available online: https://ark.intel.com/content/www/us/en/ark/products/65523/intel-core-i7-3770k-processor-8m-cache-up-to-3-90-ghz.html (accessed on 1 February 2019).
  39. GeForce GTX 1070. Available online: https://www.geforce.com/hardware/desktop-gpus/geforce-gtx-1070/specifications (accessed on 1 February 2019).
  40. Bottou, L. Stochastic gradient descent tricks. In Neural Networks: Tricks of the Trade; Springer: Berlin, Germany, 2012; pp. 421–436. [Google Scholar]
  41. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  42. Hossin, M.; Sulaiman, M.N. A review on evaluation metrics for data classification evaluations. Int. J. Data Min. Knowl. Manag. Process. 2015, 5, 1–11. [Google Scholar]
  43. Cover, T.M.; Hart, P.E. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  44. Ilin, A.; Raiko, T. Practical approaches to principal component analysis in the presence of missing values. J. Mach. Learn. Res. 2010, 11, 1957–2000. [Google Scholar]
  45. Precision and Recall. Available online: https://en.wikipedia.org/wiki/Precision_and_recall (accessed on 21 May 2019).
  46. Student’s T-test. Available online: https://en.wikipedia.org/wiki/Student%27s_t-test (accessed on 20 March 2019).
  47. Cohen, J. A power primer. Psychol. Bull. 1992, 112, 155–159. [Google Scholar] [CrossRef] [PubMed]
  48. Nakagawa, S.; Cuthill, I.C. Effect size, confidence interval and statistical significance: A practical guide for biologists. Biol. Rev. 2007, 82, 591–605. [Google Scholar] [CrossRef] [PubMed]
  49. Subrahmanyam, M.; Maheshwari, R.P.; Balasubramanian, R. Local maximum edge binary patterns: A new descriptor for image retrieval and object tracking. Signal Process. 2012, 92, 1467–1479. [Google Scholar] [CrossRef]
  50. Velmurugan, K.; Baboo, S.S. Image retrieval using Harris corners and histogram of oriented gradients. Int. J. Comput. Appl. 2011, 24, 6–10. [Google Scholar]
  51. Nguyen, D.T.; Pham, T.D.; Baek, N.R.; Park, K.R. Combining deep and handcrafted image features for presentation attack detection in face recognition systems using visible-light camera sensors. Sensors 2018, 18, 699. [Google Scholar] [CrossRef]
  52. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
  53. Hsu, C.-W.; Lin, C.-J. A comparison of methods for multiclass support vector machines. IEEE Trans. Neural Netw. 2002, 13, 415–425. [Google Scholar] [PubMed]
  54. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
Figure 1. Overall flow diagram of the proposed classification framework.
Figure 1. Overall flow diagram of the proposed classification framework.
Jcm 08 00986 g001
Figure 2. Overview of the proposed cascaded convolutional neural network and long short-term memory (LSTM)-based deep architecture for the classification of multiple gastrointestinal (GI) diseases.
Figure 2. Overview of the proposed cascaded convolutional neural network and long short-term memory (LSTM)-based deep architecture for the classification of multiple gastrointestinal (GI) diseases.
Jcm 08 00986 g002
Figure 3. Residual block of ResNet18 with (a) 1 × 1 convolutional-mapping-based residual unit and (b) identity-mapping-based residual unit.
Figure 3. Residual block of ResNet18 with (a) 1 × 1 convolutional-mapping-based residual unit and (b) identity-mapping-based residual unit.
Jcm 08 00986 g003
Figure 4. Internal connectivity of a standard LSTM cell.
Figure 4. Internal connectivity of a standard LSTM cell.
Jcm 08 00986 g004
Figure 5. Different anatomical districts of the human GI tract.
Figure 5. Different anatomical districts of the human GI tract.
Jcm 08 00986 g005
Figure 6. Examples from each class of the 37 different categories (i.e., C1 to C37) including both normal and diseased cases.
Figure 6. Examples from each class of the 37 different categories (i.e., C1 to C37) including both normal and diseased cases.
Jcm 08 00986 g006
Figure 7. Selected sample images for illustrating the high intra-class variance: (a) C18; (b) C22; (c) C34; (d) C32; (e) C13; and (f) C30.
Figure 7. Selected sample images for illustrating the high intra-class variance: (a) C18; (b) C22; (c) C34; (d) C32; (e) C13; and (f) C30.
Jcm 08 00986 g007
Figure 8. Training loss and accuracy plots during the first stage (i.e., spatial features extraction by ResNet18): (a) 1st fold cross-validation; and (b) 2nd fold cross-validation.
Figure 8. Training loss and accuracy plots during the first stage (i.e., spatial features extraction by ResNet18): (a) 1st fold cross-validation; and (b) 2nd fold cross-validation.
Jcm 08 00986 g008
Figure 9. Training loss and accuracy plots during the second stage (i.e., temporal features extraction using LSTM): (a) 1st fold cross-validation; and (b) 2nd fold cross-validation.
Figure 9. Training loss and accuracy plots during the second stage (i.e., temporal features extraction using LSTM): (a) 1st fold cross-validation; and (b) 2nd fold cross-validation.
Jcm 08 00986 g009
Figure 10. Classification performance of our framework according to the number of frames ( n ) for LSTM.
Figure 10. Classification performance of our framework according to the number of frames ( n ) for LSTM.
Jcm 08 00986 g010
Figure 11. Confusion matrix of the proposed method. The entry in the i t h row and j t h column corresponds to the percentage of samples from class i that were classified as class j . Precision and recall are calculated as “TPk/ (TPk + FPk) “and “TPk/ (TPk + FNk)” [45], respectively.
Figure 11. Confusion matrix of the proposed method. The entry in the i t h row and j t h column corresponds to the percentage of samples from class i that were classified as class j . Precision and recall are calculated as “TPk/ (TPk + FPk) “and “TPk/ (TPk + FNk)” [45], respectively.
Jcm 08 00986 g011
Figure 12. Receiver operating characteristic curves of our proposed method and other baseline models with the area under the curve (AUC).
Figure 12. Receiver operating characteristic curves of our proposed method and other baseline models with the area under the curve (AUC).
Jcm 08 00986 g012
Figure 13. Sensitivity analysis plot of our method and various baseline models in terms of (a) average accuracy; (b) average F1 score; (c) mAP; and (d) mAR.
Figure 13. Sensitivity analysis plot of our method and various baseline models in terms of (a) average accuracy; (b) average F1 score; (c) mAP; and (d) mAR.
Jcm 08 00986 g013
Figure 14. t-test performance of our method and the second-best model in terms of (a) average accuracy; (b) average F1 score; (c) mAP; and (d) mAR.
Figure 14. t-test performance of our method and the second-best model in terms of (a) average accuracy; (b) average F1 score; (c) mAP; and (d) mAR.
Jcm 08 00986 g014
Figure 15. Class prediction-based retrieval system by using our proposed classification framework.
Figure 15. Class prediction-based retrieval system by using our proposed classification framework.
Jcm 08 00986 g015
Figure 16. Examples of the correctly retrieved frames by our proposed method: (a) C6; (b) C19; (c) C24; and (d) C37.
Figure 16. Examples of the correctly retrieved frames by our proposed method: (a) C6; (b) C19; (c) C24; and (d) C37.
Jcm 08 00986 g016aJcm 08 00986 g016b
Figure 17. Examples of the correctly classified frames by our proposed method with probability score graph: (a) C6; (b) C19; (c) C24; and (d) C37.
Figure 17. Examples of the correctly classified frames by our proposed method with probability score graph: (a) C6; (b) C19; (c) C24; and (d) C37.
Jcm 08 00986 g017aJcm 08 00986 g017b
Figure 18. Examples of the incorrectly retrieved frames by our proposed method: (a) C16; (b) C31; (c) C33; and (d) C34.
Figure 18. Examples of the incorrectly retrieved frames by our proposed method: (a) C16; (b) C31; (c) C33; and (d) C34.
Jcm 08 00986 g018aJcm 08 00986 g018b
Figure 19. Examples of the incorrectly classified frames by our proposed method with probability score graph: (a) C16; (b) C31; (c) C33; and (d) C34.
Figure 19. Examples of the incorrectly classified frames by our proposed method with probability score graph: (a) C16; (b) C31; (c) C33; and (d) C34.
Jcm 08 00986 g019aJcm 08 00986 g019b
Table 1. Comparison of our proposed and existing methods for endoscopy disease classification.
Table 1. Comparison of our proposed and existing methods for endoscopy disease classification.
Endoscopy TypeMethodPurposeNo. of ClassesStrengthWeakness
CELog Gabor filter, SUSAN edge detection and SVM [19]Small bowel polyps and ulcers detection2Computationally efficientLimited dataset and number of classes
Low detection performance
CETexture features (ULBP, wavelet) + SVM [20]Polyp detection in GI tract2Robust to illumination change and scale invariantLimited dataset and number of classes
CETexture features (LBP, wavelet) + SVM [6]Tumor recognition in the digestive tract 2Invariant to illumination change
Extract multiresolution features
Limited dataset and number of classes
CETexture features (SIFT, Saliency) + SVM [21]Polyp classification2Extract scale invariant featuresLimited dataset and number of classes
CETexture features (SIFT, HoG, LBP, CLBP, ULBP) + SVM, FLDA [22]Polyp Detection2Extract scale invariant features
High classification performance
Limited dataset and number of classes
CECNN [7]Small intestine movement characterization 6High classification performanceLimited number of classes
CECNN [15]Celiac disease classification2High sensitivity and specificityLimited dataset and number of classes
CECNN [12]Hookworm detection2Edge extraction network results in better performance Limited number of classes
EGDCNN [9]H. pylori infection detection 9Comparable performance of second CNN with the clinical diagnosis reference standardCAD performance should be enhanced.
A limited number of classes
EGDCNN [8]Anatomical classification of GI images 6High classification performance
Computationally efficient
Limited number of classes
Only used for anatomical classification
EGDCNN-based SSD detector [13]Gastric cancer detection2High sensitivity
Computationally efficient
Overall low positive prediction value
Limited dataset and number of classes
ColonoscopyCNN [10]Colorectal polyp detection and classification 3High detection performanceLimited dataset and number of classes
Low classification performance
ColonoscopyCNN [14]Real-time colorectal polyp type analysis4High accuracy and sensitivityLimited number of classes
Low specificity
ColonoscopyOnline and offline 3D-CNN [11]Detection of colorectal polyps 2Computationally efficientCAD performance should be enhanced.
EGD, Colonoscopy, Sigmoidoscopy,
Rectoscopy
CNN (ResNet) + LSTM
(Proposed)
Classification of multiple GI diseases 37Computationally efficient
High classification performance
Cascaded training of CNN and LSTM requires more time
Table 2. Layer-wise configuration details of deep ResnNet18 model in our study.
Table 2. Layer-wise configuration details of deep ResnNet18 model in our study.
Layer NameFeature Map
Size
FiltersKernel SizeStride#PaddingTotal Learnable Parameters
Image input layer 224 × 224   × 3 n/an/an/an/an/a
Conv1 112 × 112 × 64 64 7 × 7 × 3 239600
Max pooling 56 × 56 × 64 1 3 × 3 21
Conv2-1–Conv2-2
(Identity Mapping)
56 × 56 × 64
56 × 56 × 64
64
64
3 × 3 × 64
3 × 3 × 64
1
1
1
1
74,112
Conv3-1–Conv3-2
(Identity Mapping)
56 × 56 × 64
56 × 56 × 64
64
64
3 × 3 × 64
3 × 3 × 64
1
1
1
1
74,112
Conv4-1–Conv4-2
(1 × 1 Convolutional Mapping)
28 × 28 × 128
28 × 28 × 128
28 × 28 × 128
128
128
128
3 × 3 × 64
3 × 3 × 128
1 × 1 × 64
2
1
2
1
1
0
230,528
Conv5-1–Conv5-2
(Identity Mapping)
28 × 28 × 128
28 × 28 × 128
128
128
3 × 3 × 128
3 × 3 × 128
1
1
1
1
295,680
Conv6-1–Conv6-3
(1 × 1 Convolutional Mapping)
14 × 14 × 256
14 × 14 × 256
14 × 14 × 256
256
256
256
3 × 3 × 128
3 × 3 × 256
1 × 1 × 128
2
1
2
1
1
0
919,808
Conv7-1–Conv7-2
(Identity Mapping)
14 × 14 × 256
14 × 14 × 256
256
256
3 × 3 × 256
3 × 3 × 256
1
1
1
1
1,181,184
Conv8-1–Conv8-3
(1 × 1 Convolutional Mapping)
7 × 7 × 512
7 × 7 × 512
7 × 7 × 512
512
512
512
3 × 3 × 256
3 × 3 × 512
1 × 1 × 256
2
1
2
1
1
0
3,674,624
Conv9-1–Conv9-2
(Identity Mapping)
7 × 7 × 512
7 × 7 × 512
512
512
3 × 3 × 512
3 × 3 × 512
1
1
1
1
4,721,664
Avg pooling 1 × 1 × 512 1 7 × 7 70
FC layer37 18,981
Softmax37
Classification layer37
Total number of learnable parameters: 11,200,293
Table 3. Layer-wise configuration details of long short-term memory (LSTM) model in our study.
Table 3. Layer-wise configuration details of long short-term memory (LSTM) model in our study.
Layer NameFeature Map Size Total Learnable
Sequence input layer n × 1 × 1 × 512
LSTM6001,951,200
Dropout600
FC layer3722,237
Softmax37
Classification layer37
Total learnable parameters: 1,973,437
Table 4. Details of multiple subcategories of each anatomical district and their corresponding classes.
Table 4. Details of multiple subcategories of each anatomical district and their corresponding classes.
Gastrointestinal TractClass Name
(Normal/Disease Cases)
Training Set (Frames)Testing Set (Frames)Total
Anatomical DistrictSubcategory
EsophagusLarynxC1: Normal387387774
Upper partC2: Normal6256251250
C3: Esophageal candidiasis419419838
C4: Esophageal papillomatosis272272544
Lower part (z-line)C5: Normal250250500
StomachCardiaC6: Hiatal hernia6486481296
FundusC7: Atrophic gastritis241241482
C8: Atrophic and xanthoma gastritis255254509
BodyC9: Benign hyperplastic polyps107010702140
C10: Adenocarcinoma (Cancer)9559551910
PylorusC11: Normal127512752550
Small IntestineDuodenumC12: Normal423423846
C13: Ulcer134513452690
C14: Papilla vateri7027021404
Terminal IleumC15: Crohn’s disease8408401680
IleocecalC16: Severe Crohn’s disease278278556
Ileocecal valveC17: Crohn’s disease8388381676
Large IntestineCaecumC18: Adenocarcinoma (Cancer)130113012602
C19: Melanosis coli342342684
C20: Caecal angiectasia403404807
C21: Appendix aperture6946941388
Ascending/
Transverse/Descending Colon
C22: Adenocarcinoma (Cancer)129312932586
C23: Melanosis coli6036041207
C24: Other types of polyps250250500
C25: Dyed resection margins250250500
C26: Dyed lifted polyps250250500
C27: Melanosis coli and tuber adenoma243243486
C28: Inflammatory polyposis382382764
C29: Normal5005001000
Sigmoid ColonC30: Tuber adenoma221222124424
C31: Polypoid cancer282282564
RectosigmoidC32: Ulcerative colitis207120714142
Rectum C33: Severe Crohn’s disease 107410742148
C34: Adenocarcinoma (Cancer)136213622724
C35: Tuber adenoma106910692138
C36: Normal420420840
C37: A focal radiation injury411411822
Table 5. Parameters of the stochastic gradient descent method for the training of both ResNet18 and LSTM models in our experiments.
Table 5. Parameters of the stochastic gradient descent method for the training of both ResNet18 and LSTM models in our experiments.
ModelNumber of Training EpochsInitial Learning RateMomentumL2-RegularizationLearning Rate Drop FactorMini-Batch Size
ResNet1880.0010.90.00010.110
LSTM100.00010.90.00010.150
Table 6. Performance comparison of our method using ResNet18 + LSTM with the conventional ResNet18 model based on feature extraction from different layers (unit: %).
Table 6. Performance comparison of our method using ResNet18 + LSTM with the conventional ResNet18 model based on feature extraction from different layers (unit: %).
Layer NameFeature Dim.ResNet18 [23]Proposed
Accuracy
± Std
F1 score
± Std
mAP
± Std
mAR
± Std
Accuracy
± Std
F1 score
± Std
mAP
± Std
mAR
± Std
Conv6-250,17675.86 ± 4.0378.62 ± 1.2881.64 ± 0.3575.85 ± 2.6987.15 ± 1.0287.61 ± 0.0488.85 ± 0.8186.40 ± 0.85
Conv7-250,17677.13 ± 3.6179.61 ± 0.7382.42 ± 0.7677.02 ± 2.0288.02 ± 2.7888.94 ± 1.1891.20 ± 0.1286.81 ± 2.36
Conv8-225,08884.39 ± 1.5484.75 ± 0.6985.92 ± 0.2083.62 ± 1.1589.07 ± 0.1089.96 ± 0.8891.24 ± 0.8688.72 ± 0.91
Conv9-225,08887.10 ± 0.7087.57 ± 0.4788.19 ± 0.1786.97 ± 1.0989.39 ± 1.1089.70 ± 1.6990.24 ± 1.6189.18 ± 1.76
Avg. pooling51289.95 ± 1.2690.35 ± 1.7490.72 ± 1.1789.99 ± 2.2992.57 ± 0.6693.41 ± 0.1294.58 ± 0.3792.28 ± 0.58
Table 7. Performance comparisons of our method (ResNet18 + LSTM) with the conventional ResNet18 with and without PCA (unit: %).
Table 7. Performance comparisons of our method (ResNet18 + LSTM) with the conventional ResNet18 with and without PCA (unit: %).
MethodResNe18 [23]Proposed
Accuracy
± Std
F1 score
± Std
mAP
± Std
mAR
± Std
Accuracy
± Std
F1 score
± Std
mAP
± Std
mAR
± Std
With PCA
(No. of eigenvectors = 136)
88.50 ± 1.0190.16 ± 0.1691.85 ± 0.1188.52 ± 0.2090.01 ± 0.1791.82 ± 0.3794.22 ± 0.4089.54 ± 0.33
Without PCA89.95 ± 1.2690.35 ± 1.7490.72 ± 1.1789.99 ± 2.2992.57 ± 0.6693.41 ± 0.1294.58 ± 0.3792.28 ± 0.58
Table 8. Comparative classification performance of proposed method and different baseline CNN models (unit: %).
Table 8. Comparative classification performance of proposed method and different baseline CNN models (unit: %).
MethodsAccuracyF1 ScoremAPmAR
Fold
1
Fold
2
Avg. ± StdFold
1
Fold
2
Avg. ± StdFold
1
Fold
2
Avg. ± StdFold
1
Fold
2
Avg. ± Std
SqueezeNet [24]78.6977.0077.84 ± 1.1977.5375.9576.74 ± 1.1278.3875.1676.77 ± 2.2776.7076.7676.73 ± 0.04
AlexNet [16]79.1980.9780.08 ± 1.2680.3180.6680.49 ± 0.2480.5580.8580.70 ± 0.2180.0880.4780.28 ± 0.27
GoogLeNet [8,12,15,17]83.3685.8284.59 ± 1.7484.9985.2985.14 ± 0.2184.6785.9285.29 ± 0.8985.3284.6684.99 ± 0.47
VGG19 [25]84.8185.4985.15 ± 0.4884.5786.0285.29 ± 1.0385.4886.2785.88 ± 0.5683.6785.7784.72 ± 1.48
VGG16 [25]83.8887.5785.72 ± 2.6184.8486.7785.80 ± 1.3785.2087.2886.24 ± 1.4784.4886.2685.37 ± 1.26
InceptionV3 [14,26]87.2388.6187.92 ± 0.9887.8089.1088.45 ± 0.9286.5089.2487.87 ± 1.9389.1488.9689.05 ± 0.13
ResNet50 [23]88.9490.1789.55 ± 0.8790.1391.0690.60 ± 0.6689.5991.8290.70 ± 1.5890.6890.3290.50 ± 0.26
ResNet18 [23]90.8489.0689.95 ± 1.2691.5889.1390.35 ± 1.7491.5589.8990.72 ± 1.1791.6288.3789.99 ± 2.29
Proposed92.1093.0392.57 ± 0.66 93.4993.3393.41 ± 0.1294.3294.8494.58 ± 0.3792.6891.8792.28 ± 0.58
Table 9. Parametric and structural comparisons of different deep CNN models with our proposed model.
Table 9. Parametric and structural comparisons of different deep CNN models with our proposed model.
CNN ModelsSize (MB)No. of Conv. LayersNo. of FC LayersNo. of LSTM LayersNetwork DepthParameters (Millions)Image Input Size
SqueezeNet [24]4.6 MB18 181.24227-by-227
AlexNet [16]227 MB53 861227-by-227
GoogLeNet [8,12,15,17]27 MB211 227.0224-by-224
VGG19 [25]535 MB163 19144224-by-224
VGG16 [25]515 MB133 16138224-by-224
InceptionV3 [14,26]89 MB471 4823.9299-by-299
ResNet50 [23]96 MB491 5025.6224-by-224
ResNet18 [23]44 MB171 1811.7224-by-224
Proposed48 MB17111913.17224-by-224
Table 10. Comparison of classification performance of the proposed method with different handcrafted feature-based methods (unit: %).
Table 10. Comparison of classification performance of the proposed method with different handcrafted feature-based methods (unit: %).
MethodClassifiersAccuracyF1 ScoremAPmAR
Fold
1
Fold
2
Avg. ± StdFold
1
Fold
2
Avg. ± StdFold
1
Fold
2
Avg. ± StdFold
1
Fold
2
Avg. ± Std
LBP
[49]
AdaBoostM236.9034.5735.74 ± 1.6528.8526.5527.70 ± 1.6336.9034.5735.74 ± 1.6523.6821.5522.61 ± 1.51
Multi-SVM45.5342.1543.84 ± 2.3943.3441.3542.35 ± 1.4144.0541.9442.99 ± 1.4942.6640.7741.72 ± 1.34
RF57.3756.8457.10 ± 0.3753.4054.3153.85 ± 0.6454.5355.0654.79 ± 0.3752.3153.5852.95 ± 0.90
KNN49.6851.2450.46 ± 1.1046.2848.4447.36 ± 1.5345.7347.9946.86 ± 1.5946.8448.9047.87 ± 1.46
HoG [50]AdaBoostM240.2838.4139.35 ± 1.3333.0432.6832.86 ± 0.2540.2838.4139.35 ± 1.3328.0028.4428.22 ± 0.31
Multi-SVM47.9651.7349.84 ± 2.6751.9555.6653.80 ± 2.6368.1366.6467.39 ± 1.0541.9747.7944.88 ± 4.11
RF60.1062.7261.41 ± 1.8561.7364.6663.19 ± 2.0768.0369.2968.66 ± 0.8956.4960.6158.55 ± 2.91
KNN50.1456.2653.20 ± 4.3352.2257.1354.68 ± 3.4757.3759.4558.41 ± 1.4747.9354.9851.45 ± 4.99
MLBP
[51]
AdaBoostM246.4241.6244.02 ± 3.4040.0434.8537.45 ± 3.6746.4241.6244.02 ± 3.4035.2029.9832.59 ± 3.69
Multi-SVM56.1854.7655.47 ± 1.0053.7252.4953.10 ± 0.8755.7053.8154.75 ± 1.3351.8751.2351.55 ± 0.45
RF61.5661.2461.40 ± 0.2256.9858.1657.57 ± 0.8458.4159.7559.08 ± 0.9555.6256.6556.13 ± 0.73
KNN54.3856.4355.40 ± 1.4550.9053.4952.20 ± 1.8350.9253.2152.06 ± 1.6150.8853.7852.33 ± 2.05
Proposed92.1093.0392.57 ± 0.66 93.4993.3393.41 ± 0.1294.3294.8494.58 ± 0.3792.6891.8792.28 ± 0.58

Share and Cite

MDPI and ACS Style

Owais, M.; Arsalan, M.; Choi, J.; Mahmood, T.; Park, K.R. Artificial Intelligence-Based Classification of Multiple Gastrointestinal Diseases Using Endoscopy Videos for Clinical Diagnosis. J. Clin. Med. 2019, 8, 986. https://doi.org/10.3390/jcm8070986

AMA Style

Owais M, Arsalan M, Choi J, Mahmood T, Park KR. Artificial Intelligence-Based Classification of Multiple Gastrointestinal Diseases Using Endoscopy Videos for Clinical Diagnosis. Journal of Clinical Medicine. 2019; 8(7):986. https://doi.org/10.3390/jcm8070986

Chicago/Turabian Style

Owais, Muhammad, Muhammad Arsalan, Jiho Choi, Tahir Mahmood, and Kang Ryoung Park. 2019. "Artificial Intelligence-Based Classification of Multiple Gastrointestinal Diseases Using Endoscopy Videos for Clinical Diagnosis" Journal of Clinical Medicine 8, no. 7: 986. https://doi.org/10.3390/jcm8070986

APA Style

Owais, M., Arsalan, M., Choi, J., Mahmood, T., & Park, K. R. (2019). Artificial Intelligence-Based Classification of Multiple Gastrointestinal Diseases Using Endoscopy Videos for Clinical Diagnosis. Journal of Clinical Medicine, 8(7), 986. https://doi.org/10.3390/jcm8070986

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop