Next Article in Journal
Nonischemic Super-Responders in Fusion CRT Pacing with Normal Atrioventricular Conduction
Previous Article in Journal
Multi-Scale Hybrid Network for Polyp Detection in Wireless Capsule Endoscopy and Colonoscopy Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep-Learning-Based Real-Time and Automatic Target-to-Background Ratio Calculation in Fluorescence Endoscopy for Cancer Detection and Localization

1
Department of Mechanical Engineering, University of Washington, Seattle, WA 98195, USA
2
Division of Gastroenterology, Department of Internal Medicine, University of Michigan, Ann Arbor, MI 48109, USA
3
Department of Mechanical Engineering, University of Michigan, Ann Arbor, MI 48109, USA
4
Department of Biomedical Engineering, University of Michigan, Ann Arbor, MI 48109, USA
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(9), 2031; https://doi.org/10.3390/diagnostics12092031
Submission received: 16 July 2022 / Revised: 15 August 2022 / Accepted: 15 August 2022 / Published: 23 August 2022
(This article belongs to the Section Medical Imaging and Theranostics)

Abstract

:
Esophageal adenocarcinoma (EAC) is a deadly cancer that is rising rapidly in incidence. The early detection of EAC with curative intervention greatly improves the prognoses of patients. A scanning fiber endoscope (SFE) using fluorescence-labeled peptides that bind rapidly to epidermal growth factor receptors showed a promising performance for early EAC detection. Target-to-background (T/B) ratios were calculated to quantify the fluorescence images for neoplasia lesion classification. This T/B calculation is generally based on lesion segmentation with the Chan–Vese algorithm, which may require hyperparameter adjustment when segmenting frames with different brightness and contrasts, which impedes automation to real-time video. Deep learning models are more robust to these changes, while accurate pixel-level segmentation ground truth is challenging to establish in the medical field. Since within our dataset the ground truth contained only a frame-level diagnosis, we proposed a computer-aided diagnosis (CAD) system to calculate the T/B ratio in real time. A two-step process using convolutional neural networks (CNNs) was developed to achieve automatic suspicious frame selection and lesion segmentation for T/B calculation. In the segmentation model training for Step 2, the lesion labels were generated with a manually tuned Chan–Vese algorithm using the labeled and predicted suspicious frames from Step 1. In Step 1, we designed and trained deep CNNs to select suspicious frames using a diverse and representative set of 3427 SFE images collected from 25 patient videos from two clinical trials. We tested the models on 1039 images from 10 different SFE patient videos and achieved a sensitivity of 96.4%, a specificity of 96.6%, a precision of 95.5%, and an area under the receiver operating characteristic curve of 0.989. In Step 2, 1006 frames containing suspicious lesions were used for training for fluorescence target segmentation. The segmentation models were tested on two clinical datasets with 100 SFE frames each and achieved mean intersection-over-union values of 0.89 and 0.88, respectively. The T/B ratio calculations based on our segmentation results were similar to the manually tuned Chan–Vese algorithm, which were 1.71 ± 0.22 and 1.72 ± 0.28, respectively, with a p-value of 0.872. With the graphic processing unit (GPU), the proposed two-step CAD system achieved 50 fps for frame selection and 15 fps for segmentation and T/B calculation, which showed that the frame rejection in Step 1 improved the diagnostic efficiency. This CAD system with T/B ratio as the real-time indicator is designed to guide biopsies and surgeries and to serve as a reliable second observer to localize and outline suspicious lesions highlighted by fluorescence probes topically applied in organs where cancer originates in the epithelia.

1. Introduction

Esophageal cancer is the eighth most common cancer in the world. With a poor 5-year survival rate of less than 20%, this cancer ranks sixth in mortality worldwide [1]. Esophageal adenocarcinoma (EAC), the prevalent subtype of esophageal cancer in developed countries, is rising in incidence due to increased risk factors such as gastroesophageal reflux disease and obesity [2,3]. Barrett’s esophagus (BE), a known precursor for EAC, transforms via low-grade dysplasia (LGD) and high-grade dysplasia (HGD) before progressing to EAC [4]. Due to the lack of an efficient early detection method for Barrett’s neoplasia, EAC is usually diagnosed at an advanced stage with poor patient prognosis.
Standard endoscopy screening with white light illumination and random, four-quadrant biopsy is recommended for cancer surveillance in BE patients but is labor-intensive and limited by sampling error [5]. Novel endoscopic imaging technologies with molecular agents targeting cancer-specific biomarkers have been developed for early EAC detection [6]. In our previous studies, a scanning fiber endoscope (SFE) with fluorescence-labeled peptides has demonstrated feasibility to visualize the expressions of EGFR and ErbB2 for Barrett’s neoplasia detection [7,8]. To quantify the fluorescence images, target-to-background (T/B) ratios were calculated. These T/B ratios measured the ratios between the average intensities of high-contrast fluorescence targets and the neighboring background regions [7]. This quantitative ratio can be more important for more widespread use of fluorescence molecular probes during cancer surveillance and screening without an expert endoscopist. In this case, a computer-aided diagnosis (CAD) system with real-time and automatic T/B ratio calculation can serve as a reliable trained observer to localize and outline suspicious lesions for biopsy guidance and cancer detection. This real-time T/B calculation is limited by the speed of target segmentation. Currently, the Chan–Vese algorithm is applied to segment the fluorescence targets, with speed limited to only 2–3 frames/s [9]. Since the Chan–Vese approach evolves the level set iteratively to minimize the energy function, the algorithm is not easily computed in parallel processing, which impedes real-time computation. In addition, hyperparameters need to be adjusted for in the Chan–Vese algorithm when the contrast and brightness of the images change, which makes automatic T/B calculation more difficult. Moreover, to achieve this automatic CAD system, an automatic frame selection is needed to reject frames with diverse image artifacts, including (1) saturation, (2) dye pooling, (3) air bubbles, and (4) instruments (mother endoscope), as well as to select frames with suspicious lesions for further segmentation and T/B quantification, which has been performed manually in our previous studies [7,8].
Deep learning, a machine-learning subfield, has been successfully applied to many areas of science and technology, including computer vision [10,11], speech recognition [12,13], games [14], and bioinformatics [15], and has been expanded to the field of medical image analysis. Recently, there have been significant developments in computer-aided detection and diagnosis using deep-learning techniques in various medical-imaging modalities, e.g., MRI, computed tomography, and ultrasounds [16,17]. The main contribution of deep learning in medical image analysis is mainly around two topics: (1) detection and classification, and (2) localization and segmentation.

1.1. Detection and Classification

Convolutional neural networks (CNNs) have become a state-of-the-art tool for medical image classification and have been widely used for cancer classification. Litjens et al. used a CNN to detect cancer areas from H-and-E-stained whole slides for prostate and breast cancer detection. All the slides containing prostate cancer and metastases of breast cancer could be identified automatically, while 30–40% of the slides containing benign and normal tissue could be excluded without any human intervention [18]. Ardila et al. proposed end-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography and achieved a state-of-the-art performance, with a 94.4% area under the curve (AUC) on 6716 National Lung Cancer Screening Trial cases [19]. Kooi et al. showed that a CNN model trained on a large dataset of mammographic lesions (45,000 images) outperformed a state-of-the art CAD system trained using manually designed features (respective AUC: 92.9% vs. 90.6%) [20].

Localization and Segmentation

Another contribution of deep learning to the medical-imaging field is the potential to improve the speed and accuracy of cancer localization and segmentation in clinical settings. A CNN-based algorithm was applied for the automatic localization and segmentation of rectal cancers on multiparametric MR imaging, with segmentation results comparable to those manually labeled by experts and a Dice similarity of 0.70 [21]. A hybrid network that fused DenseNet and UNet [22] together achieved a competitive performance for liver and tumor segmentation from 3D CT volumes. Features from both inter- and intraslices were extracted and jointed together for better 3D segmentation [23]. A two-pathway CNN that extracted both local and global contexts was applied to segment brain tumors from MRI scans and showed improved accuracy and speed over other algorithms reported on the 2013 BRATS [24].
There has also been an increasing number of studies that have applied deep learning to advance endoscopic imaging [25,26,27,28]. With the development of massive parallel architecture, or graphic processing units (GPUs), deep-learning techniques have demonstrated promising results in real-time endoscopic cancer surveillance [25,26]. Wang et al. developed a deep-learning algorithm by training it with 5545 colonoscopy images from 1290 patients for polyp detection and localization, with a processing speed of 25 frames per second. The algorithm achieved 94.38% sensitivity and 95.92% specificity validated on 27,113 colonoscopy images [25]. Although the segmentation of fluorescence images remains challenging due to the “patchy” appearance of the fluorescence targets, nonuniform intensities in both targets and backgrounds, and losses in contrast and resolution due to the scattering of light from biological tissue, there have been deep-learning models showing promising results for nucleus segmentation in fluorescence microscopy [29,30]. The robustness and real-time capability of deep learning allow real-time and automatic T/B calculation, while the training of deep-learning models generally requires a dataset with pixel-level lesion labels as ground truth [29].
In this paper, we present the development and validation of our deep-learning-based pipeline for automatic and real-time T/B calculation with frame-level diagnosis as ground truth only, which can be embedded in a CAD system for cancer surveillance and screening using fluorescence endoscopy with molecular agents. The proposed pipeline has two steps, with real-time and automatic suspicious frame selection in Step 1 and segmentation and T/B quantification in Step 2. The use of the T/B ratio preserves the clinically validated quantitative neoplastic lesion analysis for classification, unlike pure “black box” approaches to cancer detection and classification [31]. In Step 1, the frame-based diagnostic ground truth is used for training, and predicted suspicious frames in test videos and originally labelled frames can be combined as a larger segmentation dataset for Step 2. The labels of the segmentation training set are generated by a manually tuned Chan–Vese algorithm. The labor is reduced for labelling more nondiagnostic frames by looking into the predicted suspicious frames only.
The speed of Step 1 in our pipeline reaches 50 fps, and Step 2 is 15 fps. The efficient frame selection in Step 1 further improves the real-time capability of the T/B ratio diagnosis compared with the segmentation models in Step 2 only. To the best of our knowledge, there is no deep-learning-based CAD system for wide-field fluorescence endoscopy to provide (1) real-time suspicious frame selection or (2) real-time and automatic quantification using T/B ratios. In the experiment, we validate each step of the system in two clinical studies using fluorescence-labeled peptides for Barrett’s neoplasia detection.

2. Materials and Methods

The whole CAD system could be divided into steps: (1) a suspicious frame selection model and (2) fluorescence target segmentation. The segmented targets were then used to generate a 30-pixel-wide background via morphological operations. T/B ratios were calculated from the mean intensities of each region. The pipeline is shown in Figure 1.
We trained different CNN architectures as backbones to extract features for the following tasks: (1) classification and (2) segmentation. These CNN architectures usually consisted of the following layers: convolutional layers, batch normalization layers, activation layers, pooling layers, and skip connection layers. All these layers were assembled to form different backbone architectures for feature extraction. In this paper, two different backbone architectures, MobileNetV2 [32] and Xception [33], were used.

2.1. Suspicious Frame Selection Model

Network architectures with Xception or MobileNetV2 backbones, followed by a global average pooling layer and 2 fully connected layers, were applied for suspicious frame selection. The whole architecture for frame selection is shown in Figure 2.

2.2. Fluorescence Target Segmentation Model

Architectures for semantic segmentation usually consist of two parts: an encoder and a decoder. The encoder compresses the input image into smaller vectors with context information. On the other hand, the decoder expands the extracted information and reconstructs an output with the size of the original input image to obtain a pixel-level classification. In this study, 2 different architectures, UNet and BiSeNet, were evaluated for fluorescence target segmentation.
UNet [22] consists of a symmetric encoder and decoder. The encoder extracts feature vectors from the input images. The decoder combines spatial information from different levels with up-convoluted context information. In this study, Xception and MobileNetV2 were used as backbones to extract features for the encoder. Figure 3 shows an example of the UNet architecture with the MobileNetV2 backbone.
BiSeNet33 contains two pathways, a spatial path with a small stride to preserve the spatial information and to generate high-resolution features and a context path as the encoder to obtain a sufficient receptive field and to extract context information for segmentation. Xception and MobileNetV2 were used as backbones to extract features within the context path. An attention refinement module was introduced to refine the features of the last two stages of the context path. A feature fusion module was introduced to combine features from the two pathways efficiently. The overall architecture of BiSeNet is shown in Figure 4.

2.3. Regularization

Different training techniques were applied to decrease overfitting and to improve the model performance without sacrificing the training cost. In this study, data augmentation was applied to increase the variability of the training images. We used 3 different types of image transformations for augmentation: (1) geometric distortion, (2) brightness and contrast adjustment, and (3) image blur and noise. For geometric distortion, we used random scaling, shifting, flipping, and rotating in the full range of 0–90 degrees. For blur and noise, we used random motion blur, Gaussian blur, and Gaussian noise. All the augmentation methods mentioned above were pixel-wise adjustments, which retained the original pixel information. “Early stopping” was applied to avoid overfitting [35]. Models with the maximum validation accuracy were saved. In addition, we used weights that were pretrained on the ImageNet data corpus as initialization for our backbone architectures. Natural images in the ImageNet dataset share fundamental features that can be transferred to medical images [36].

3. Experiments

3.1. Datasets and Preprocessing

Two datasets from two clinical studies were used for training and testing the deep-learning models presented in this work:
  • The first set contained a total of 15 videos from an in vivo Multiplex imaging study (NCT03589443) performed in patients with Barrett’s neoplasia. The fluorescence-labeled peptides of QRH*-Cy5 and KSP*-IRDye800 targeting EGFR and ErbB2, respectively, were topically applied in the lower esophagus. Videos were recorded using SFE with a resolution of 720 × 720 at 30 Hz. Separate channels were used to record fluorescence images from QRH*-Cy5 (red) and KSP*-IRDye800 (green), as well as reflectance (blue). The combined duration of 15 videos was approximately 96 min (~173,000 frames) [7].
  • A separate set contained a total of 20 videos from an in vivo Dimer imaging study (NCT03852576) performed in patients with Barrett’s neoplasia. Only IRDye800 was used to label dimer peptide QRH*-KSP*-E3-IRDye800, targeting both EGFR and ErbB2, and was again topically applied. The NIR fluorescence projected to the green channel was recorded with reflectance as the blue channel. The combined duration of 20 videos was approximately 88 min (~158,000 frames) [8].
All the images were preprocessed to have the same format before being passed to the deep-learning models. First, the individual frames were rescaled to a fixed size of 480 × 480 pixels to reduce the number of operations and to increase the model processing speeds. Then, the values of the pixels in each frame were normalized by dividing by 255. For the suspicious region detection task, fluorescence was projected to the green channel with reflectance as the blue channel. Each image from the Multiplex dataset was converted to two separate images. For the fluorescence target segmentation task, the fluorescence channels were extracted and converted to greyscale.

3.2. Model Development

We trained and evaluated the suspicious frame selection model and the fluorescence target segmentation model separately.

3.2.1. Frame Selection Model

The preprocessed images were passed to the frame selection model and were classified into two classes: (1) frames containing bright, suspicious lesions and (2) frames containing a normal esophagus or frames with different artifacts, including dye pooling, air bubbles, saturation, or an instrument (mother endoscope). The datasets from the two clinical studies were divided into the following categories.
Training and validation: 3427 images were extracted from 15 videos of the Multiplex study and 10 videos from the Dimer study. There were 1475 images with suspicious bright regions and 1952 images with different artifacts. The 3427 images were then split into training (83%) and validation (17%) groups under random and stratified sampling.
Testing: 1039 images were extracted from the remaining 10 unused videos from the Dimer study, which contained 443 images with suspicious bright regions and 596 images with artifacts.
We augmented the dataset by using different combinations of image transformations during training. We adjusted the batch size and optimized the initial learning rate for each backbone architecture individually. We employed the same learning rate schedule (decreasing by 0.8 every 10 epochs) for each backbone. The networks were trained with the ADAM optimizer [37] to minimize the binary cross-entropy loss.
In terms of evaluation, area under the curve (AUC) values of the ROC curve were computed and compared for different combinations of backbone and augmentation techniques over the test set, which measured the model performances irrespective of the chosen classification threshold. The accuracy, sensitivity, specificity, and precision were calculated for the model with the highest AUC. A gradient-weighted class activation map (grad-CAM) was applied to visualize the implicit attention of the trained CNN model. The discriminative features of the objects were highlighted by the grad-CAM, which provided a direct interpretation of the network when making its classification [38].

3.2.2. Fluorescence Target Segmentation Model

Different model architectures of BiSeNet or UNet with backbones of MobileNetV2 or Xception were trained and evaluated for fluorescence target segmentation. Only frames with suspicious regions were selected from the two clinical studies for the model training and evaluation.
  • Training and validation: 1006 images from 11 videos from the Multiplex study were randomly separated into training (910) and validation (96) groups;
  • Testing A: 100 images were extracted from 4 different videos from the Multiplex study;
  • Testing B: 100 images were extracted from the Dimer study.
All the images were paired with a pixel-level annotated segmentation map, which was used as ground truth for segmentation. The annotated segmentation maps were generated using the Chan–Vese algorithm. We applied the same training protocol for training the frame selection model. Different combinations of image augmentation were evaluated. The batch size and optimized initial learning rate were adjusted for each model and backbone individually. The ADAM optimizer was used to minimize the sparse, categorical cross-entropy. The performance was measured in terms of accuracy and pixel intersection-over-union (IOU) and averaged across the 2 classes [39]:
I O U A , B = A B A B ,
where A is the ground truth and B is the segmented fluorescence target.
Herein, we used a Windows computer with an Intel(R) Core(TM) i7-1065G7 CPU, a Toshiba NVMe KIOXIA 512 GB SSD, and an NVIDIA RTX 2080 SUPER GPU with a CUDA 10.0 for training and evaluation. All the experiments were implemented using Tensorflow [40] software library, an open-source deep-learning library.

4. Results

4.1. Suspicious Frame Selection

The suspicious frame selection results are shown in Table 1. Networks initialized from weights pretrained on the ImageNet data corpus demonstrated better performances compared to those starting from random weights. Different combinations of image augmentation techniques were evaluated. Networks with geometric distortions surpassed those without geometric distortions by a significant margin (one-tailed paired t-test, p-value). Neither (1) adjusting the brightness and contrast of the image nor (2) adding noise or blur to the image during training showed an improvement in model performance. Networks using the Xception backbone performed slightly better than those with the MobileNetV2 backbone (one-tailed paired t-test, p-value of 0.012).
At a sensitivity level (true positive rate) of 96.4%, the best model using the Xception backbone with geometric distortion, brightness, and contrast adjustment augmentation during training (highlighted in bold in Table 1) had a specificity of 96.6%, a precision of 95.5%, and an overall accuracy of 96.8%. The confusion matrix and the ROC of the test data are shown in Figure 5. As can be seen, there was a reasonably large range of high sensitivities, high specificities, and high precisions (all greater than 90%).
A grad-CAM was applied to visualize the discriminative features detected by the network. As can be seen in Figure 6a–d, fluorescence-highlighted lesions from different locations with different brightness and contrasts were detected and used as features for classification. Figure 6e–g demonstrated that the trained model correctly distinguished artifacts such as air bubbles, pooled dye, and the mother endoscope and classified these frames as noise.

4.2. Fluorescence Target Segmentation

A total of 1006 images were used to train and validate the model for the fluorescence target segmentation task. The test set results, as presented in Table 2 and Table 3, showed the following:
(1)
Networks with geometric augmentation during training surpassed those without geometric augmentation by a significant margin (one-tailed paired t-test, p-value), which was consistent with our findings for the frame selection model;
(2)
Neither (a) adjusting the brightness and contrast of the images nor (b) adding noise or blur to the images during training showed an improvement in performance;
(3)
UNet outperformed BiSeNet in segmentation accuracy (one-tailed paired t-test, p-value);
(4)
Using pretrained ImageNet weights as initialization did not improve the performance;
(5)
BiSeNet models using the Xception backbone outperformed BiSeNet models using the MobileNetV2 backbone regarding to accuracy and mIOU (one-tailed paired t-test, p-value).
The UNet model with the MobileNetV2 backbone, geometric distortions, and image blur and noise augmentation performed better when segmenting the fluorescence images (highlighted in red in Table 3) and achieved mIOU values of 0.89 and 0.88 on the Multiplex and Dimer test datasets, respectively. Figure 7 shows examples of fluorescence target segmentation results tested on the Multiplex and Dimer datasets.
To better evaluate the accuracy of the whole pipeline for T/B calculation, we compared the T/B ratios calculated using the deep-learning pipeline with those using the Chan–Vese algorithm. A total of 58 frames with unique, suspicious bright regions were selected from 31 patient videos from the Dimer study. The average T/B ratios from the segmentation results of our best model for deep learning and the Chan–Vese algorithm were 1.71 ± 0.22 and 1.72 ± 0.28, respectively, with a p-value of 0.872 calculated using a two-tailed paired t-test. There was no significant difference when comparing the deep-learning T/B ratios with the Chan–Vese T/B ratios, which meant our segmentation deep-learning model could reach the accuracy of the labels we used for training.

4.3. Speed

We achieved average frame-processing times of 20 ms for the frame selection task and 42 ms (UNet with MobileNetV2) for the segmentation task with an NVIDIA RTX 2080S external GPU. In comparison to the ground truth Chen–Vese method, all the deep-learning methods increased in speed by ≥10 × . The total processing time of the CAD system for each frame, including frame selection, fluorescence target segmentation, and T/B calculation, was around 70 ms.

5. Discussion

In this study, a deep-learning CAD pipeline was developed to achieve real-time and automatic T/B calculation. This research was an extension of our prior work, where we have shown that the T/B ratio can be used as a quantitative and accurate indicator for Barrett’s neoplasia detection and localization [7,8]. When combined with fluorescence endoscopy (SFE) and fluorescence molecular probes, this deep-learning CAD pipeline could serve as a reliable observer to localize and outline suspicious lesions for guiding biopsies, which has the potential to greatly improve routine surveillance for BE patients. Our study highlighted the development of artificial intelligence in fluorescence endoscopy undergoing clinical trials, specifically SFE for early EAC detection in a limited subject population at a tertiary referral hospital.
The reported pipeline consisted of two deep-learning models: (1) suspicious frame selection and (2) fluorescence target segmentation. Frames with suspicious bright regions were automatically selected, while frames with normal squamous cells or artifacts, including dye pooling, air bubbles, saturation, etc., were rejected in the frame selection model. A CNN model with an Xception backbone for frame selection achieved a uniform and consistent high performance with a sensitivity of 96.4%, a specificity of 96.6%, a precision of 95.5%, an overall accuracy 96.8%, and an AUC of 0.989 (Table 1). This suggests that the model likely offered sufficient sensitivity for selecting suspicious frames for further T/B calculation. To better unmask the “black box” of deep learning and to better understand how the models made their decisions, a grad-CAM was applied. The features highlighted by the grad-CAMs showed consistency with those selected by experienced human observers. As can be seen in Figure 6e,f, frames with air bubbles and dye pooling in the GE junction were rejected for further quantification since these features could diminish the robustness of the T/B ratio in esophageal neoplasia detection. On the contrary, frames with patchy lesions lightened by fluorescence probes were selected. The models learned these features for classification by themselves.
In an attempt to segment the fluorescence target, UNet and BiSeNet architectures with Xception and MobileNetV2 backbones were evaluated on two different datasets from two clinical studies. Different combinations of augmentation methods were tested to improve the model performance. As shown in Table 3, the best model, UNet with the MobileNetV2 backbone with added geometric distortions and image blur augmentation during training, achieved mIOU values of 0.89 and 0.88, respectively, for the Multiplex and Dimer studies. The T/B ratio calculated from the segmentation results had 1% difference compared with that calculated from the manually tuned Chan–Vese method, which generated training labels for our segmentation learning. The segmentation models trained only using images from one clinical study showed comparable performances to the images from the other clinical study. This finding demonstrated that the models could be generally applied to studies using SFE with different molecular probes for cancer detection in the esophagus, and most likely in the oral cavity [41], bladder [42], and colon [43,44], where topically applied fluorescence probes demonstrate early detection in the epithelia.
In our training dataset, we only had frame-based diagnostic labels without pixel-level lesion masks from pathologists. To leverage the classification labels thoroughly and embed them into the loss function, the suspicious frame selection model was separated. In the supervised training of the segmentation model, the labels were generated with a manually fine-tuned Chan–Vese algorithm for different frames. Continued improvements would be fusing the two-step model into one model for both Barrett’s neoplasia detection and localization when segmentation labels from pathologists are available. In this case, the segmentation models can be trained directly on the video frames. Architectures such as Mask R-CNN [45] and YOLACT [46] could be applied to localize each object in the image, including EAC and HGD lesions, dye pooling, and forceps. Instance segmentation masks could also be generated for each neoplasia lesion for T/B calculation. However, this task would require a relatively large volume of patient data to train the models, which could eventually be fulfilled when the technique is utilized in a multisite clinical trial for BE surveillance. Obtaining the pixel-level object masks as data annotations is labor-intensive, and label noise is a common feature of medical image datasets. The major sources of label noise include interobserver variability, human annotator error, and errors in computer-generated labels [47,48]. In addition to changing the network architectures for noise-labeling specifically [49,50], a large number of studies have kept the original network and only modified the loss functions [51,52], learning to reweight training data based on their noise levels [53], exploit data and label consistency [54], and optimize training procedures [55,56]. In a future study with obtained segmentation labels, further explorations should be conducted to train with noisy labels.
Regarding the speed of the deep-learning models, we achieved averages of ~20 ms for suspicious frame selection and ~42 ms for fluorescence target segmentation on images of 480 × 480 pixels using a single-threaded PC with an NVIDIA RTX 2080S external GPU. This is comparable to studies where deep learning has been investigated with endoscopic imaging for real-time cancer detection. By using a multithreaded processing system, Wang et al. could process at least 25 frames per second (fps) to localize polyps during clinical colonoscopies [25]. Byrne et al. developed an AI model to differentiate diminutive adenomas from hyperplastic polyps, with a frame-processing time of 50 ms [57]. Our CAD system could achieve 50 fps for frame selection. Once the frame was selected for T/B calculation, additional time was needed for fluorescence target segmentation. The whole pipeline, including both frame selection and target segmentation, dropped the speed to ~15 fps. A real-time assistant software should operate at a speed of at least 25 fps according to PAL and NTSC standards, where video encodings are standardized to 25 or 30 fps [58]. A multithreaded system could be implemented to accelerate the processing speed. Each thread processed one image at a time. To further accelerate the speed, the input image could be down-sampled. The models with input images of 224 × 224 pixels could achieve an average of ~16 ms for suspicious frame selection and ~26 ms for fluorescence target segmentation, which decreased the processing time by 40%. We believe that future advances in computer hardware and deep-learning architecture can allow for the use of even larger input images while preserving real-time capabilities.
Several training techniques were investigated to improve the model performance. Weights pretrained on the ImageNet data corpus were used as initialization for the backbone architectures, which improved the performance for the frame selection models but not for the fluorescence target models. The input images for the fluorescence target models contained only a fluorescence channel, while the input images for the frame selection models contained both fluorescence and reflectance channels. Weights transferred from ImageNet could be applied to the reflectance channel but not to the fluorescence channel. Different augmentation techniques were tested, including (1) geometric distortions (2) brightness and contrast adjustments, and (3) image blur and noise. Looking at Table 1 and Table 2, there was no common combination of augmentations that was optimal for both datasets and tasks. Only models with geometric distortion augmentation during training demonstrated significantly improved performances in both suspicious frame selection and fluorescence target segmentation for both datasets. These geometric distortions overcame the positional and dimensional biases. On the contrary, (a) adjusting the brightness and contrast of the image and (b) adding noise or blur to the image during training showed negligible improvement in performance. However, we did not perform a search for augmentation hyperparameters. Further optimizations may further improve model performance.
Single-channel fluorescence images containing informative molecular expressions that were endoscopically invisible with conventional reflectance imaging were used in our segmentation. However, full-color images do contain important supplementary structural information on tissues and tumors that can contribute to computer-aided diagnosis and tumor segmentation. In our future work, full-color and single fluorescence channels are coregistered during imaging with a multimodal scanning fiber endoscope (mmSFE) [59] and generate four-channel 30 Hz videos with the same field of view of a 70–100 degree cone angle, which produces varying spatial resolutions across the depth of focus of 3 to 50 mm (15 microns to submillimeters, respectively) [60]. Due to the extremely high sensitivity of photomultiplier light detection on each of the four channels, the total laser power can be kept below 5 mW to reduce laser safety requirements for the clinical staff.
Adapting state-of-the-art segmentation networks from three-channel to four-channel input is straightforward by changing the convolution kernel size, but we could not directly take advantage of transfer learning from large, natural datasets, which increase the performance when the target dataset is small [61]. Applying transfer learning from an RGB dataset to a multichannel dataset is not trivial in the sense that additional channels cannot benefit from the same training as the RGB channels if no large, annotated dataset with the same additional channels exists [62]. In addition, target lesions from different modalities can have different characteristics that require feature fusion for more accurate segmentation. Working with multichannel data has been studied, and different ways of merging features have been proposed for natural images [63,64]. Segmentation using multimodality consisting of fusing multiple pieces of information to improve segmentation in medical fields is also under exploration, where fusing multiple modalities is the key challenge. According to the fusion strategies, network architectures can be categorized into input-level (feature level) fusion networks, layer-level (classifier level) fusion networks, and decision-level fusion networks [65]. Input-level fusion networks are adopted by most existing multimodal medical image segmentation networks. In this case, multimodality images are directly integrated in the input space to learn a unified feature representation, which contains the intrinsic, multimodal representation of the data [66,67]. It is used to support the learning of a traditional segmentation network. In a layer-level fusion network, images of each modality are used to learn separate feature sets of their own. These single-modality features are concatenated and fed to the decision layer to obtain the final segmentation results [68,69,70]. Like layer-level fusion, decision-level fusion segmentation networks use each modality image as the input of the single segmentation network, while in this case the final outputs of the individual networks are integrated to obtain the segmentation result. The single network can better exploit the unique information of the corresponding modality independently, which benefits when multimodal images have little direct complementary information in their original image spaces due to different image acquisition techniques [65]. Most of the fusion strategies in this category are based on averaging and majority voting [71,72]. Decision-level fusion was investigated to achieve a better performance on brain tumor segmentation among different fusion strategies because it could learn complex and complementary feature information from different modalities compared to input-level fusion networks [73]. To take advantage of different fusion schemes, some recent works have adopted three schemes simultaneously on the same dataset [66,74].
Despite the promising performance of our CAD system with real-time and automatic T/B ratio calculation, there were several limitations in this study. One limitation was that 4466 frames from only 35 patient videos were applied for training and testing the models, which limited the robustness of the models. However, we did not directly classify each frame as Barrett’s neoplasia or not. Instead, suspicious frames were selected for further T/B calculation, which did not require a relatively large volume of data containing neoplastic lesions with various appearances. For the segmentation task, the single foreground object class (fluorescence) and the high tolerance of the segmentation performance diminished the effect of the small dataset. In addition, we applied different techniques to further minimize the effect of the small dataset, e.g., image augmentation and ImageNet pretrained weights as initialization. Another limitation was that images used for model development were from patients with advanced BE or a history of neoplasia. A paucity of normal esophagus tissue limited the generalization of our model to a larger population. In addition, race and ethnicity could potentially introduce biases to machine-learning models [75]. Future studies should require larger and more-balanced datasets to assess model robustness against different medical centers, patient populations, and patient races and ethnicities.

6. Conclusions

We reported the successful development of a deep-learning-based CAD system for automatic and real-time T/B calculation. This accelerated CAD system based on T/B quantification provided similar high accuracy and faster processing speed (15 fps) compared with clinical studies using manual post-processing protocols without the use of pixel-level segmentation ground truth for training. Thus, this demonstration of a potentially more consistent analysis should be useful in clinical settings in three ways. First, it may serve as a reliable second observer to guide biopsies with the T/B ratio as a real-time indicator of lesion extent. Second, the system may be suitable for routine screening or surveillance using unsedated transnasal or capsule endoscopy, especially in under-served or under-developed areas where there is a lack of experienced clinicians. Third, it can be integrated into surgical procedure to provide real-time feedback during treatment. Further clinical studies should be conducted to determine if Barrett’s neoplasia detection rates may be increased and made more consistent with the assistance of a real-time CAD system.

Author Contributions

Conceptualization, Y.J., J.C., C.G., T.D.W. and E.J.S.; methodology, Y.J. and C.G.; software, Y.J.; validation, Y.J., J.C. and C.G.; formal analysis, Y.J. and C.G.; investigation, Y.J. and J.C.; resources, T.D.W. and E.J.S.; data curation, Y.J. and J.C.; writing—original draft preparation, Y.J.; writing—review and editing, Y.J., C.G., T.D.W. and E.J.S.; visualization, Y.J.; project administration, E.J.S.; funding acquisition, T.D.W. and E.J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the NIH National Cancer Institute (NCI) and initially part of the Barrett’s Esophagus Translational Research Network (BETRNet) (grant numbers U54 CA163059 and R44 CA261318-01A1).

Institutional Review Board Statement

The clinical data were obtained from clinical trial studies conducted in accordance with the Declaration of Helsinki, approved by the Institutional Review Board of the University of Michigan (NCT03589443, NCT03852576), and registered at ClinicalTrials.gov (http://www.clinicaltrials.gov/) for studies involving humans, accessed 22 August 2022.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. All data are not publicly available due to human testing restrictions.

Conflicts of Interest

All authors declare no conflict of interest other than patents assigned to and managed by the University of Washington (Y.J., C.G., and E.J.S.) and the University of Michigan (J.C. and T.D.W.). In addition, E.J.S. is a consultant at VerAvanti Inc., an SBIR recipient of NIH NCI funding. VerAvanti had no role in the data analysis, writing, or editing of the manuscript.

References

  1. Bray, F.; Ferlay, J.; Soerjomataram, I.; Siegel, R.L.; Torre, L.A.; Jemal, A. Torre, and Ahmedin Jemal. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Melina, A.; Rutherford, M.J.; Bardot, A.; Ferlay, J.; Andersson, T.M.L.; Myklebust, T.Å.; Tervonen, H.; Grad, V.T.; Ransom, D.; Shack, L.; et al. Progress in cancer survival, mortality, and incidence in seven high-income countries 1995–2014 ICBP SURVMARK-2: A population-based study. Lancet Oncol. 2019, 20, 1493–1505. [Google Scholar]
  3. Vingeliene, S.; Chan, D.S.M.; Vieira, A.R.; Polemiti, E.; Stevens, C.; Abar, L.; Rosenblatt, D.N.; Greenwood, D.C.; Norat, T. An update of the WCRF/AICR systematic literature review and meta-analysis on dietary and anthropometric factors and esophageal cancer risk. Ann. Oncol. 2017, 28, 2409–2419. [Google Scholar] [CrossRef]
  4. Bhat, S.; Coleman, H.G.; Yousef, F.; Johnston, B.T.; McManus, D.T.; Gavin, A.T.; Murray, L.J. Risk of malignant progression in Barrett’s esophagus patients: Results from a large population-based study. J. Natl. Cancer Inst. 2011, 103, 1049–1057. [Google Scholar] [CrossRef] [Green Version]
  5. Shaheen, N.J.; Falk, G.W.; Iyer, P.G.; Gerson, L.B. ACG clinical guideline: Diagnosis and management of Barrett’s esophagus. Off. J. Am. Coll. Gastroenterol. ACG 2016, 111, 30–50. [Google Scholar] [CrossRef]
  6. Goetz, M.; Wang, T.D. Molecular imaging in gastrointestinal endoscopy. Gastroenterology 2010, 138, 828–833. [Google Scholar] [CrossRef] [Green Version]
  7. Chen, J.; Jiang, Y.; Chang, T.-S.; Joshi, B.; Zhou, J.; Rubenstein, J.H.; Wamsteker, E.J.; Kwon, R.S.; Appelman, H.; Beer, D.G.; et al. Multiplexed endoscopic imaging of Barrett’s neoplasia using targeted fluorescent heptapeptides in a phase 1 proof-of-concept study. Gut 2021, 70, 1010–1013. [Google Scholar] [CrossRef] [PubMed]
  8. Chen, J.; Jiang, Y.; Chang, T.-S.; Rubenstein, J.H.; Kwon, R.S.; Wamsteker, E.-J.; Prabhu, A.; Zhao, L.; Appelman, H.D.; Owens, S.R.; et al. Detection of Barrett’s neoplasia with a near-infrared fluorescent heterodimeric peptide. Endoscopy 2022. [Google Scholar] [CrossRef]
  9. Jiang, Y.; Gong, Y.; Rubenstein, J.H.; Wang, T.D.; Seibel, E.J. Toward real-time quantification of fluorescence molecular probes using target/background ratio for guiding biopsy and endoscopic therapy of esophageal neoplasia. J. Med. Imaging 2017, 4, 024502. [Google Scholar] [CrossRef] [Green Version]
  10. Christian, S.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  11. Joseph, R.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  12. Ilya, S.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  13. Kyunghyun, C.; van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing EMNLP, Doha, Qatar, 25–29 October 2014. [Google Scholar]
  14. Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; van den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of Go with deep neural networks and tree search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef]
  15. Seonwoo, M.; Lee, B.; Yoon, S. Deep learning in bioinformatics. Brief. Bioinform. 2017, 18, 851–869. [Google Scholar]
  16. Milletari, F.; Ahmadi, S.-A.; Kroll, C.; Plate, A.; Rozanski, V.; Maiostre, J.; Levin, J.; Dietrich, O.; Ertl-Wagner, B.; Bötzel, K.; et al. Hough-CNN: Deep learning for segmentation of deep brain regions in MRI and ultrasound. Comput. Vis. Image Underst. 2017, 164, 92–102. [Google Scholar] [CrossRef] [Green Version]
  17. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Litjens, G.; Sánchez, C.I.; Timofeeva, N.; Hermsen, M.; Nagtegaal, I.; Kovacs, I.; Hulsbergen-Van de Kaa, C.; Bult, P.; Van Ginneken, B.; Van Der Laak, J. Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis. Sci. Rep. 2016, 6, 26286. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Ardila, D.; Kiraly, A.P.; Bharadwaj, S.; Choi, B.; Reicher, J.J.; Peng, L.; Tse, D.; Etemadi, M.; Ye, W.; Corrado, G.; et al. End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nat. Med. 2019, 25, 954–961. [Google Scholar] [CrossRef] [PubMed]
  20. Kooi, T.; Litjens, G.; van Ginneken, B.; Gubern-Mérida, A.; Sánchez, C.I.; Mann, R.; den Heeten, A.; Karssemeijer, N. Large scale deep learning for computer aided detection of mammographic lesions. Med. Image Anal. 2017, 35, 303–312. [Google Scholar] [CrossRef] [PubMed]
  21. Trebeschi, S.; Van Griethuysen, J.J.M.; Lambregts, D.; Lahaye, M.J.; Parmar, C.; Bakers, F.C.H.; Peters, N.H.G.M.; Beets-Tan, R.G.H.; Aerts, H.J.W.L. Deep learning for fully-automated localization and segmentation of rectal cancer on multiparametric MR. Sci. Rep. 2017, 7, 5301. [Google Scholar] [CrossRef]
  22. Olaf, R.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  23. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.; Heng, P.-A. H-DenseUNet: Hybrid densely connected UNet for liver and tumor segmentation from CT volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef] [Green Version]
  24. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.-M.; Larochelle, H. Brain tumor segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef] [Green Version]
  25. Wang, P.; Xiao, X.; Brown, J.R.G.; Berzin, T.M.; Tu, M.; Xiong, F.; Hu, X.; Liu, P.; Song, Y.; Zhang, D.; et al. Development and validation of a deep-learning algorithm for the detection of polyps during colonoscopy. Nat. Biomed. Eng. 2018, 2, 741–748. [Google Scholar] [CrossRef]
  26. Urban, G.; Tripathi, P.; Alkayali, T.; Mittal, M.; Jalali, F.; Karnes, W.; Baldi, P. Deep learning localizes and identifies polyps in real time with 96% accuracy in screening colonoscopy. Gastroenterology 2018, 155, 1069–1078. [Google Scholar] [CrossRef] [PubMed]
  27. Ali, S.; Zhou, F.; Braden, B.; Bailey, A.; Yang, S.; Cheng, G.; Zhang, P.; Li, X.; Kayser, M.; Soberanis-Mukul, R.D.; et al. An objective comparison of detection and segmentation algorithms for artefacts in clinical endoscopy. Sci. Rep. 2020, 10, 2748. [Google Scholar] [CrossRef] [PubMed]
  28. Izadyyazdanabadi, M.; Belykh, E.; Mooney, M.A.; Eschbacher, J.M.; Nakaji, P.; Yang, Y.; Preul, M.C. Prospects for theranostics in neurosurgical imaging: Empowering confocal laser endomicroscopy diagnostics via deep learning. Front. Oncol. 2018, 8, 240. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Caicedo, J.C.; Roth, J.; Goodman, A.; Becker, T.; Karhohs, K.W.; Broisin, M.; Molnar, C.; McQuin, C.; Singh, S.; Theis, F.J.; et al. Evaluation of deep learning strategies for nucleus segmentation in fluorescence images. Cytom. Part A 2019, 95, 952–965. [Google Scholar] [CrossRef] [Green Version]
  30. Moen, E.; Bannon, D.; Kudo, T.; Graf, W.; Covert, M.; Van Valen, D. Deep learning for cellular image analysis. Nat. Methods 2019, 16, 1233–1246. [Google Scholar] [CrossRef]
  31. van der Sommen, F.; de Groof, J.; Struyvenberg, M.; van der Putten, J.; Boers, T.; Fockens, K.; Schoon, E.J.; Curvers, W.; De With, P.; Mori, Y.; et al. Machine learning in GI endoscopy: Practical guidance in how to interpret a novel field. Gut 2020, 69, 2035–2045. [Google Scholar] [CrossRef]
  32. Mark, S.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  33. François, C. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  34. Yu, C.; Wang, J.; Peng, C.; Gao, C.; Yu, G.; Sang, N. Bisenet: Bilateral segmentation network for real-time semantic segmentation. In Proceedings of the European Conference on Computer Vision ECCV, Munich, Germany, 8–14 September 2018; pp. 325–341. [Google Scholar]
  35. Lutz, P. Early stopping-but when? In Neural Networks: Tricks of the Trade; Springer: Berlin/Heidelberg, Germany, 1998; pp. 55–69. [Google Scholar]
  36. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012. [Google Scholar]
  37. P, K.D.; Ba, J.L. Adam: A method for stochastic gradient descent. In Proceedings of the ICLR: International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
  38. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning deep features for discriminative localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2921–2929. [Google Scholar]
  39. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  40. Martín, A.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. {TensorFlow}: A system for {Large-Scale} machine learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation OSDI 16, Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  41. Jasper, V.; de Wit, J.G.; jan Voskuil, F.; Witjes, M.J.H. Improving oral cavity cancer diagnosis and treatment with fluorescence molecular imaging. Oral Dis. 2021, 27, 21–26. [Google Scholar]
  42. Pan, Y.; Volkmer, J.-P.; Mach, K.E.; Rouse, R.V.; Liu, J.-J.; Sahoo, D.; Chang, T.C.; Metzner, T.J.; Kang, L.; van de Rijn, M.; et al. Endoscopic molecular imaging of human bladder cancer using a CD47 antibody. Sci. Transl. Med. 2014, 6, ra148–ra260. [Google Scholar] [CrossRef]
  43. Kelly, K.; Alencar, H.; Funovics, M.; Mahmood, U.; Weissleder, R. Detection of invasive colon cancer using a novel, targeted, library-derived fluorescent peptide. Cancer Res. 2004, 64, 6247–6251. [Google Scholar] [CrossRef] [Green Version]
  44. Hsiung, P.-L.; Hardy, J.; Friedland, S.; Soetikno, R.; Du, C.B.; Wu, A.; Sahbaie, P.; Crawford, J.M.; Lowe, A.W.; Contag, C.; et al. Detection of colonic dysplasia in vivo using a targeted heptapeptide and confocal microendoscopy. Nat. Med. 2008, 14, 454–458. [Google Scholar] [CrossRef]
  45. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  46. Daniel, B.; Zhou, C.; Xiao, F.; Lee, Y.J. Yolact: Real-time instance segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 9157–9166. [Google Scholar]
  47. Vorontsov, E.; Cerny, M.; Régnier, P.; Di Jorio, L.; Pal, C.J.; Lapointe, R.; Vandenbroucke-Menu, F.; Turcotte, S.; Kadoury, S.; Tang, A. Deep learning for automated segmentation of liver lesions at ct in patients with colorectal cancer liver metastases. Radiol. Artif. Intell. 2019, 1, 180014. [Google Scholar] [CrossRef] [PubMed]
  48. Karimi, D.; Dou, H.; Warfield, S.K.; Gholipour, A. Deep learning with noisy labels: Exploring techniques and remedies in medical image analysis. Med. Image Anal. 2020, 65, 101759. [Google Scholar] [CrossRef] [PubMed]
  49. Thekumparampil, K.K.; Khetan, A.; Lin, Z.; Oh, S. Robustness of conditional gans to noisy labels. In Proceedings of the 31th International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  50. Bekker, A.J.; Goldberger, J. Training deep neural-networks based on unreliable labels. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP, Shanghai, China, 20–25 March 2016; pp. 2682–2686. [Google Scholar]
  51. Ghosh, A.; Kumar, H.; Sastry, P.S. Robust loss functions under label noise for deep neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar] [CrossRef]
  52. Giorgio, P.; Rozza, A.; Menon, A.K.; Nock, R.; Qu, L. Making deep neural networks robust to label noise: A loss correction approach. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1944–1952. [Google Scholar]
  53. Shen, Y.; Sanghavi, S. Learning with bad training data via iterative trimmed loss minimization. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 5739–5748. [Google Scholar]
  54. Lee, K.; Yun, S.; Lee, K.; Lee, H.; Li, B.; Shin, J. Robust inference via generative classifiers for handling noisy labels. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 3763–3772. [Google Scholar]
  55. Gong, C.; Bin, K.; Seibel, E.J.; Wang, X.; Yin, Y.; Song, Q. Synergistic Network Learning and Label Correction for Noise-Robust Image Classification. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP, Singapore, 23–27 May 2022; pp. 4253–4257. [Google Scholar]
  56. Yu, X.; Han, B.; Yao, J.; Niu, G.; Tsang, I.; Sugiyama, M. How does disagreement help generalization against label corruption? In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 7164–7173. [Google Scholar]
  57. Byrne, M.F.; Chapados, N.; Soudan, F.; Oertel, C.; Pérez, M.L.; Kelly, R.; Iqbal, N.; Chandelier, F.; Rex, D.K. Real-time differentiation of adenomatous and hyperplastic diminutive colorectal polyps during analysis of unaltered videos of standard colonoscopy using a deep learning model. Gut 2019, 68, 94–100. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  58. Bernal, J.; Tajkbaksh, N.; Sanchez, F.J.; Matuszewski, B.J.; Chen, H.; Yu, L.; Angermann, Q.; Romain, O.; Rustad, B.; Balasingham, I.; et al. Comparative validation of polyp detection methods in video colonoscopy: Results from the MICCAI 2015 endoscopic vision challenge. IEEE Trans. Med. Imaging 2017, 36, 1231–1249. [Google Scholar] [CrossRef] [PubMed]
  59. Lee, C.M.; Engelbrecht, C.J.; Soper, T.D.; Helmchen, F.; Seibel, E.J. Scanning fiber endoscopy with highly flexible, 1 mm catheterscopes for wide-field, full-color imaging. J. Biophotonics 2010, 3, 385–407. [Google Scholar] [CrossRef] [Green Version]
  60. Miller, S.J.; Lee, C.M.; Joshi, B.P.; Gaustad, A.; Seibel, E.J.; Wang, T.D. Special section on endomicroscopy technologies and biomedical applications: Targeted detection of murine colonic dysplasia in vivo with flexible multispectral scanning fiber endoscopy. J. Biomed. Opt. 2012, 17, 021103. [Google Scholar] [CrossRef]
  61. Deepak, S.; van der Putten, P.; Plaat, A. On the impact of data set size in transfer learning using deep neural networks. In Proceedings of the International Symposium on Intelligent Data Analysis, Stockholm, Sweden, 13–15 October 2016; Springer: Cham, Switzerland, 2016; pp. 50–60. [Google Scholar]
  62. de La Comble, A.; Prepin, K. Efficient transfer learning for multi-channel convolutional neural networks. In Proceedings of the 2021 17th International Conference on Machine Vision and Applications MVA, Aichi, Japan, 25–27 July 2021; pp. 1–6. [Google Scholar]
  63. Ophoff, T.; Van Beeck, K.; Goedemé, T. Exploring RGB+ Depth fusion for real-time object detection. Sensors 2019, 19, 866. [Google Scholar] [CrossRef] [Green Version]
  64. Choe, G.; Kim, S.-H.; Im, S.; Lee, J.-Y.; Narasimhan, S.G.; Kweon, I.S. RANUS: RGB and NIR urban scene dataset for deep scene parsing. IEEE Robot. Autom. Lett. 2018, 3, 1808–1815. [Google Scholar] [CrossRef]
  65. Zhou, T.; Ruan, S.; Canu, S. A review: Deep learning for medical image segmentation using multi-modality fusion. Array 2019, 3, 100004. [Google Scholar] [CrossRef]
  66. Guo, Z.; Li, X.; Huang, H.; Guo, N.; Li, Q. Deep learning-based image segmentation on multimodal medical imaging. IEEE Trans. Radiat. Plasma Med. Sci. 2019, 3, 162–169. [Google Scholar] [CrossRef]
  67. Nabil, I.; Rahman, M.S. MultiResUNet: Rethinking the U-Net architecture for multimodal biomedical image segmentation. Neural Netw. 2020, 121, 74–87. [Google Scholar]
  68. Jose, D.; Gopinath, K.; Yuan, J.; Lombaert, H.; Desrosiers, C.; Ayed, I.B. HyperDense-Net: A hyper-densely connected CNN for multi-modal image segmentation. IEEE Trans. Med. Imaging 2018, 38, 1116–1126. [Google Scholar]
  69. Jose, D.; Desrosiers, C.; Ayed, I.B. IVD-Net: Intervertebral disc localization and segmentation in MRI with a multi-modal UNet. In Proceedings of the International Workshop and Challenge on Computational Methods and Clinical Applications for Spine Imaging, Granada, Spain, 16 September 2018; Springer: Cham, Switzerland, 2018; pp. 130–143. [Google Scholar]
  70. Chen, L.; Wu, Y.; Souza, A.M.D.; Abidin, A.Z.; Wismüller, A.; Xu, C. MRI tumor segmentation with densely connected 3D CNN. In Proceedings of the Medical Imaging 2018: Image Processing, Houston, TX, USA, 11–13 February 2018; Volume 10574, pp. 357–364. [Google Scholar]
  71. Kamnitsas, K.; Bai, W.; Ferrante, E.; McDonagh, S.; Sinclair, M.; Pawlowski, N.; Rajchl, M.; Lee, M.; Kainz, B.; Rueckert, D.; et al. Ensembles of multiple models and architectures for robust brain tumour segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Quebec City, QC, Canada, 14 September 2017; Springer: Cham, Switerland, 2017; pp. 450–462. [Google Scholar]
  72. Rokach, L. Ensemble-based classifiers. Artif. Intell. Rev. 2010, 33, 1–39. [Google Scholar] [CrossRef]
  73. Aygün, M.; Şahin, Y.H.; Ünal, G. Multi modal convolutional neural networks for brain tumor segmentation. arXiv 2018, arXiv:1809.06191. [Google Scholar]
  74. Zhe, G.; Li, X.; Huang, H.; Guo, N.; Li, Q. Medical image segmentation based on multi-modal convolutional neural network: Study on image fusion schemes. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging ISBI 2018, Washington, DC, USA, 4–7 April 2018; pp. 903–907. [Google Scholar]
  75. Gianfrancesco, M.A.; Tamang, S.; Yazdany, J.; Schmajuk, G. Potential biases in machine learning algorithms using electronic health record data. JAMA Intern. Med. 2018, 178, 1544–1547. [Google Scholar] [CrossRef]
Figure 1. Overall CAD pipeline for automatic and real-time T/B calculation with two steps: (1) suspicious frame selection model and (2) fluorescence target segmentation model.
Figure 1. Overall CAD pipeline for automatic and real-time T/B calculation with two steps: (1) suspicious frame selection model and (2) fluorescence target segmentation model.
Diagnostics 12 02031 g001
Figure 2. Suspicious frame selection model architecture with MobileNetV2 backbone.
Figure 2. Suspicious frame selection model architecture with MobileNetV2 backbone.
Diagnostics 12 02031 g002
Figure 3. UNet architecture with MobileNetV2 backbone for fluorescence target segmentation.
Figure 3. UNet architecture with MobileNetV2 backbone for fluorescence target segmentation.
Diagnostics 12 02031 g003
Figure 4. An overview of the bilateral segmentation network consisting of (a) network architecture, (b) components of the attention refinement module (ARM), and (c) components of the feature fusion module (FFM) [34].
Figure 4. An overview of the bilateral segmentation network consisting of (a) network architecture, (b) components of the attention refinement module (ARM), and (c) components of the feature fusion module (FFM) [34].
Diagnostics 12 02031 g004
Figure 5. The confusion matrix and ROC curve for the best model (Xception backbone with geometric distortion, brightness, and contrast adjustment augmentation during training) evaluated over test data.
Figure 5. The confusion matrix and ROC curve for the best model (Xception backbone with geometric distortion, brightness, and contrast adjustment augmentation during training) evaluated over test data.
Diagnostics 12 02031 g005
Figure 6. Examples of suspicious frame selection model visualized using grad-CAM. (ad) Lesions from different locations with different brightness and contrasts were detected and used as features for classification. The heatmaps generated by grad-CAM overlaid well with the lesions. (e) Air bubbles were detected as artifacts and highlighted by grad-CAM. (f) Dye pooling in gastroesophageal (GE) junction was detected and classified as noise. (g) Image of the inside of the mother endoscope working channel was classified as artifact frame in spite of the debris in the top left corner appearing as a suspicious lesion. The model used features from other regions (highlighted by grad-CAM) for classification.
Figure 6. Examples of suspicious frame selection model visualized using grad-CAM. (ad) Lesions from different locations with different brightness and contrasts were detected and used as features for classification. The heatmaps generated by grad-CAM overlaid well with the lesions. (e) Air bubbles were detected as artifacts and highlighted by grad-CAM. (f) Dye pooling in gastroesophageal (GE) junction was detected and classified as noise. (g) Image of the inside of the mother endoscope working channel was classified as artifact frame in spite of the debris in the top left corner appearing as a suspicious lesion. The model used features from other regions (highlighted by grad-CAM) for classification.
Diagnostics 12 02031 g006
Figure 7. Representatives of segmentation results for BiSeNet and UNet models with MobileNetV2 and Xception backbones.
Figure 7. Representatives of segmentation results for BiSeNet and UNet models with MobileNetV2 and Xception backbones.
Diagnostics 12 02031 g007
Table 1. Summary of suspicious frame selection results for different architectures (MobileNetV2 and Xception) and different combinations of augmentation.
Table 1. Summary of suspicious frame selection results for different architectures (MobileNetV2 and Xception) and different combinations of augmentation.
BackboneImageNetRotation, Shift, Scale, FlipBrightness, ContrastBlur, Gaussian NoiseAUC
MobileNetV2 0.926
0.976
0.973
0.97
0.969
0.887
0.953
0.926
0.915
Xception0.989
0.989
0.986
0.981
0.95
0.948
0.966
0.928
Table 2. Summary of fluorescence target segmentation results for BiSeNet model (mIOU: mean IOU over two classes).
Table 2. Summary of fluorescence target segmentation results for BiSeNet model (mIOU: mean IOU over two classes).
Model + BackboneImageNetRotation, Shift, Scale, FlipBrightness, ContrastBlur, Gaussian NoiseMultiplexDimer
AccuracymIOUAccuracymIOU
BiSeNet + MobileNetV20.9770.8680.9590.844
0.9770.8690.9590.849
0.9760.8640.9620.858
0.9730.8480.9580.846
0.9720.8390.9540.824
0.970.8280.9520.824
0.9760.8640.9560.832
0.9670.8180.9520.818
0.9770.8710.9570.841
0.9770.870.9590.846
0.9780.8740.9640.865
0.9770.8720.9620.858
0.9680.8190.9480.805
0.9680.8260.9550.831
0.9710.8360.9520.817
0.9690.8280.9540.829
BiSeNet + Xception0.9780.8730.9640.864
0.9780.8740.9620.859
0.9780.8780.9660.873
0.980.8860.9650.868
0.9730.8410.9530.821
0.9740.850.9540.829
0.9750.8560.9570.837
0.9750.8590.9550.831
0.9780.8730.9630.861
0.9780.8750.9620.859
0.980.8850.9640.865
0.9790.880.9650.87
0.970.8270.9560.83
0.9710.8330.9510.815
0.9720.840.9560.834
0.9720.8450.9560.833
Table 3. Summary of fluorescence target segmentation results for UNet model.
Table 3. Summary of fluorescence target segmentation results for UNet model.
Model + BackboneImageNetRotation, Shift, Scale, FlipBrightness, ContrastBlur, Gaussian NoiseMultiplexDimer
AccuracymIOUAccuracymIOU
UNet
+
MobileNetV2
0.980.8890.9620.857
0.9820.8970.9620.862
0.9810.8910.9680.88
0.9790.8810.9630.863
0.9740.8470.9550.829
0.9730.8450.9550.83
0.9760.8650.9590.845
0.970.8290.9520.816
0.9790.880.9630.865
0.9790.8770.9630.86
0.9790.8820.9670.877
0.9790.8810.9640.865
0.9720.8390.9490.81
0.970.8340.9620.857
0.9720.840.9580.844
0.960.7970.9570.839
UNet
+
Xception
0.9820.8970.9660.872
0.9790.8830.9650.87
0.9810.8920.9660.875
0.9780.8770.9640.867
0.9670.8230.9540.83
0.9720.840.9570.838
0.9760.8620.9570.838
0.9720.8450.9560.835
0.9810.8920.9640.866
0.9810.8930.9660.875
0.9810.8950.9660.875
0.9810.8910.9660.875
0.9690.8310.960.852
0.9710.8410.9580.844
0.9730.8510.960.85
0.9720.8420.9570.837
Note: mIOU is the mean IOU over two classes.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jiang, Y.; Chen, J.; Gong, C.; Wang, T.D.; Seibel, E.J. Deep-Learning-Based Real-Time and Automatic Target-to-Background Ratio Calculation in Fluorescence Endoscopy for Cancer Detection and Localization. Diagnostics 2022, 12, 2031. https://doi.org/10.3390/diagnostics12092031

AMA Style

Jiang Y, Chen J, Gong C, Wang TD, Seibel EJ. Deep-Learning-Based Real-Time and Automatic Target-to-Background Ratio Calculation in Fluorescence Endoscopy for Cancer Detection and Localization. Diagnostics. 2022; 12(9):2031. https://doi.org/10.3390/diagnostics12092031

Chicago/Turabian Style

Jiang, Yang, Jing Chen, Chen Gong, Thomas D. Wang, and Eric J. Seibel. 2022. "Deep-Learning-Based Real-Time and Automatic Target-to-Background Ratio Calculation in Fluorescence Endoscopy for Cancer Detection and Localization" Diagnostics 12, no. 9: 2031. https://doi.org/10.3390/diagnostics12092031

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop