Next Article in Journal
Artificial Intelligence in Orthopedic Radiography Analysis: A Narrative Review
Previous Article in Journal
Coronary Artery Bypass Graft Surgery Brings Better Benefits to Heart Failure Hospitalization for Patients with Severe Coronary Artery Disease and Reduced Ejection Fraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Using Endobronchial-Ultrasound-Guided Transbronchial Needle Aspiration Image to Improve the Overall Diagnostic Yield of Sampling Mediastinal Lymphadenopathy

1
Graduate Institute of Biomedical Engineering, National Taiwan University of Science and Technology, Taipei 106335, Taiwan
2
Graduate Institute of Applied Science and Technology, National Taiwan University of Science and Technology, Taipei 106335, Taiwan
3
Department of Pathology, Tri-Service General Hospital, Taipei 11490, Taiwan
4
Institute of Pathology and Parasitology, National Defense Medical Center, Taipei 11490, Taiwan
5
Department of Thoracic Surgery, Tri-Service General Hospital, National Defense Medical Center, Taipei 11490, Taiwan
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(9), 2234; https://doi.org/10.3390/diagnostics12092234
Submission received: 7 July 2022 / Revised: 23 August 2022 / Accepted: 13 September 2022 / Published: 16 September 2022
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Lung cancer is the biggest cause of cancer-related death worldwide. An accurate nodal staging is critical for the determination of treatment strategy for lung cancer patients. Endobronchial-ultrasound-guided transbronchial needle aspiration (EBUS-TBNA) has revolutionized the field of pulmonology and is considered to be extremely sensitive, specific, and secure for lung cancer staging through rapid on-site evaluation (ROSE), but manual visual inspection on the entire slide of EBUS smears is challenging, time consuming, and worse, subjective, on a large interobserver scale. To satisfy ROSE’s needs, a rapid, automated, and accurate diagnosis system using EBUS-TBNA whole-slide images (WSIs) is highly desired to improve diagnosis accuracy and speed, minimize workload and labor costs, and ensure reproducibility. We present a fast, efficient, and fully automatic deep-convolutional-neural-network-based system for advanced lung cancer staging on gigapixel EBUS-TBNA cytological WSIs. Each WSI was converted into a patch-based hierarchical structure and examined by the proposed deep convolutional neural network, generating the segmentation of metastatic lesions in EBUS-TBNA WSIs. To the best of the authors’ knowledge, this is the first research on fully automated enlarged mediastinal lymph node analysis using EBUS-TBNA cytological WSIs. We evaluated the robustness of the proposed framework on a dataset of 122 WSIs, and the proposed method achieved a high precision of 93.4%, sensitivity of 89.8%, DSC of 82.2%, and IoU of 83.2% for the first experiment (37.7% training and 62.3% testing) and a high precision of 91.8 ± 1.2, sensitivity of 96.3 ± 0.8, DSC of 94.0 ± 1.0, and IoU of 88.7 ± 1.8 for the second experiment using a three-fold cross-validation, respectively. Furthermore, the proposed method significantly outperformed the three state-of-the-art baseline models, including U-Net, SegNet, and FCN, in terms of precision, sensitivity, DSC, and Jaccard index, based on Fisher’s least significant difference (LSD) test ( p < 0.001 ). For a computational time comparison on a WSI, the proposed method was 2.5 times faster than U-Net, 2.3 times faster than SegNet, and 3.4 times faster than FCN, using a single GeForce GTX 1080 Ti, respectively. With its high precision and sensitivity, the proposed method demonstrated that it manifested the potential to reduce the workload of pathologists in their routine clinical practice.

1. Introduction

In normal physical examinations, one of the most difficult malignancies to identify at an early stage is lung cancer [1], which is one of the most often diagnosed diseases. As the number of lung cancer patients rises, the health care systems in both industrialized and developing nations are being put under immense strain [2]. Lung cancer is mainly divided into non-small-cell lung cancer (NSCLC) and small-cell lung carcinoma. About 80% to 85% of lung cancers cases are NSCLC. The subtypes of NSCLC are adenocarcinoma, squamous-cell carcinoma, and some other subtypes [3]. Although there have been new developments in the diagnosis, classification, and treatment of lung cancer, the overall survival rate is still poor [4]. At the time of diagnosis, the majority of the cases have distant metastasis [5]. Because of their advanced clinical stage, the great majority of patients with NSCLC do not need surgical resection [6]. Mediastinal lymphadenopathy (ML) presents a diagnostic challenge. ML may be caused by infections, granulomatous disease, reactive hyperplasia, and metastatic tumor [7]. In patients with NSCLC, determining mediastinal and hilar metastases may help guide therapy decisions, provide prognostic information, and help with patient care [8]. ML nodes need to be sampled to reach a diagnostic conclusion. Currently, there are several evaluation methods, such as positron emission tomography/computed tomography (PET/CT), endobronchial ultrasound (EBUS)-guided transbronchial needle aspiration (TBNA), B-mode morphological ultrasound, and elastography, that can be used to classify the nodal stage before surgery [9]. Although PET/CT is the main instrument for the preoperative examination of lung cancer patients, EBUS-TBNA has superior test performance and PET/CT cannot be regarded as an alternative method to be used to replace EBUS-TBNA for tissue sampling [10].
EBUS-TBNA is a new clinical technology and currently the preferred method of assessing advanced lung cancer with mediastinal lymphadenopathy. Real-time ultrasound guidance gives good diagnostic value during mediastinal lymph node collection with EBUS-TBNA, a minimally invasive method. Some of its benefits include being cost-effective, easy to use, and safe [11]. Because of this, EBUS-TBNA is considered a better option for mediastinal lymph node sampling than traditional mediastinoscopy. When combined with EBUS-TBNA, EBUS-TBNA may provide clinical information in 19% of patients, negating the need for further invasive testing. Direct smears from needle aspiration and brushing specimens allow the EBUS-TBNA to be used for on-site assessment [11,12,13,14,15]. The EBUS-TBNA treatment, which requires only little sedation and may be performed as an outpatient procedure, can be used to get a sample from the hilar lymph nodes [16]. There have been several prospective protocol-based studies [17] that have shown that EBUS-TBNA is 95 percent accurate in detecting and staging lung cancer. Microscopic pathological examination is usually the gold standard for diagnosing different types of cancer [18,19]. When dealing with suspected metastatic mediastinal lesions, rapid on-site assessment (ROSE) seems to be especially helpful in improving the EBUS-diagnostic TBNA’s yield [20]. The preliminary diagnosis provided by ROSE may minimize the number of invasive procedures (such as mediastinoscopy) required in the future [21]. When used in clinical settings, the ROSE of EBUS-TBNA needs the collaboration of many medical specialists for analysis and diagnosis, which is not always achievable owing to a lack of medical personnel. Furthermore, if malignant cells are missed during the manual screening procedure, the patient runs the risk of undergoing unneeded surgery as a result of their condition worsening.
It is now possible to convert glass slides into whole-slide images (WSIs), which enables the examination of pathological images using computer-based algorithms [22]. WSIs are high-resolution images with huge file sizes on the order of 10 gigapixels, making it difficult for pathologists to manually examine all the information on the histopathology slide due to the large quantity of information on each slide. Thus, a qualified cancer diagnosis necessitates peer review and consensus, a requirement that can be costly to meet in hospitals and small cancer centers with a scarcity of skilled pathologists. Deep learning algorithms have been frequently used in pathological image analysis applications in recent years, demonstrating their strength in representation learning. Automating the pathological image analysis aids pathologist in making accurate diagnoses in a short amount of time. Deep learning has previously shown promise in helping pathologists diagnose, classify, and segment cancer [23,24,25]. Examples include Rijthoven et al. [26], who proposed a segmentation model that combined context and details by utilizing many branches of encoder–decoder convolution neural networks for breast and lung cancer segmentation in WSIs. Wang et al. [27] proposed a fast and fully automatic hierarchical deep learning approach that utilized the coarse-to-fine strategy for bone marrow nucleated differential count on WSIs. Dov et al. [28] proposed a weakly supervised two-stage deep learning model for detecting thyroid cancer on WSIs. Tang et al. [29] proposed a DeFusionNet that can efficiently and accurately detect defocus blur. Masud et al. [30] utilized pretrained deep learning models for the detection of breast cancer in ultrasound images.
To the best of the authors’ knowledge, this is the first research on automatic segmentation of enlarged mediastinal lymph nodes metastasis in EBUS-TBNA cytological slides. The proposed automatic deep learning system is shown to be capable of detecting metastases of enlarged mediastinal lymph nodes and immediately notifying clinical doctors as a guide for subsequent adequate treatment. Figure 1 presents the workflow of the system and information on the dataset in detail. In a quantitative evaluation on the segmentation performance, the proposed method is compared with three state-of-the-art deep learning models, including U-Net [31], SegNet [32], and FCN [33], since this is the first research on automatic segmentation of lymph nodes metastasis in EBUS-TBNA cytological slides.

2. Methods

In this study, we developed a fast, efficient, and fully automatic deep-convolutional-neural-network-based AI system to segment metastatic lesions in EBUS-TBNA cytological WSIs. Figure 2 presents the overview of the deep-convolutional-neural-network-based AI system. Each WSI was converted into a patch-based hierarchical structure and processed using the fast background filtering model and the proposed deep convolutional neural network for the segmentation of a metastatic lesion from EBUS-TBNA WSIs. Figure 4 illustrates the detailed architecture of the proposed deep learning model.

2.1. Whole-Slide Image Processing

Figure 2 shows the framework for the segmentation of metastatic lesion from EBUS-TBNA WSIs. For the efficiency of the data assessment, WSIs tend to be encoded and stored in a pyramid structure, that contains several layers of images at different magnification rate, by various microscopic scanners such as Leica, Hamamatsu, and Philips. In order to deal with gigapixel WSIs effectively, each WSI I ( i , j ) was converted into a patch-based structure H = h α , β v ( x , y ) v = 1 N I ( i , j ) , where α , β , x, y, and v represent the patch column index, patch row index, patch horizontal subindex, patch vertical subindex, and the image level, respectively. When v=N, α , β , x, and y were calculated as follows:
α = i / w , β = j / ψ , x = i α × w , y = j β × ψ
where w and ψ denote the patch width and patch height, respectively. The values α , β , x , and y are in the range { 0 , , γ 1 } , { 0 , , ζ 1 } , { 0 , , w 1 } , and { 0 , , ψ 1 } , respectively; ( w , ψ ) = (512,512) in this study. The operation was devised for rapid processing of WSI by discarding the border part of a slide, which tends to have a low probability of containing regions of interest.
Firstly, a WSI was processed with fast background filtering by Otsu’s method applied onto h α , β z ( x , y ) at the level closest to and greater than or equal to the size of a unit tile and then mapping the filtered image back to the highest level to efficiently discard all the background patches (patches that contains less than 70% tissue samples), significantly decreasing the computational cost per slide. The value of z is formulated in Equation (2). As the dimensionality of WSIs was tremendous, fast background filtering was applied in order to efficiently discard the background areas which contained no cells and rapidly decreased the dimension. For further analysis, previous studies [34,35] also showed that Otsu’s method performed well to remove the background regions of a WSI on the slide’s thumbnail image to efficiently discard all background tiles, thus drastically reducing the amount of computation.
z = a r g m i n v ( γ × ζ 1 c a r d ( h v ) w × ψ )
The proposed deep convolutional neural network L was applied for a fast WSI analysis. The proposed deep convolutional neural network is described in detail in the next section. Each tile of h α , β N ( x , y ) was analyzed by the proposed convolution model L to generate the probabilities of malignant cells as shown in Equation (3).
p α , β N ( x , y ) c = L ( h α , β N ( x , y ) )
where c = 0 , , C represents the number of types of tissue to be identified, and 0, 1 and 2 represent the background, the nontarget, and the target cell type, respectively.
A two-dimensional pixel-based class map was produced as the index of the cell type that had the maximum probability of the pixel, using Equation (4).
m α , β N ( x , y ) = a r g m a x c ( ( p α , β N ( x , y ) c ) )
Then, the pixel-based segmentation results of tumor cells R = { r α , β N ( x , y ) } were produced based on class map m α , β N ( x , y ) using Equation (5). Equation (5) suppressed the nontumor information, producing the tumor information as the segmentation results.
r α , β N ( x , y ) = I α , β ( x , y ) , m α , β N ( x , y ) > 1 ϕ , o t h e r w i s e
where ϕ represents a null set.

2.2. Proposed Convolution Network Architecture

A fully convolutional network (FCN) was introduced by Shelhamer et al. [33] and it was demonstrated to be successful in the tumor segmentation of breast cancer metastases [36], thyroid cancer [37], cervical cancer [38], and ovarian cancer [39]. Inspired by the fully convolutional network (FCN) framework of Shelhamer et al. [33], the proposed deep learning network architecture has two improvements, as shown in Figure 3. To begin, we employed a five-layer shallower architecture instead of the original seven-layer FCN architecture to deal with the issue of limited GPU memory in training. Secondly, we used a single-stream 32 s upsampling path instead of three upsampling paths to prevent overly fragmented segmentation results and to save computing time. The proposed deep learning network architecture consisted of six convolutional layers, five max-pooling layers, two dropout layers, one deconvolutional layer, and a cropping layer. Firstly, there were five convolutional layers, with the first two layers consisting of two convolution sequences (kernel size of 3 × 3 and stride size of 1, respectively) and the last three layers consisting of three convolution sequences (kernel size of 3 × 3 and stride size of 1, respectively), and a ReLU after each convolution layer. To downsample the feature maps, a max pooling layer (kernel size of 2 × 2 and stride size of 2) was added at the end of each convolution layer. After the five convolution layers and five max-pooling layers, there were two drop-out layers (dropout ratio of 0.5) with two convolutions (kernel size of 7 × 7 and stride size of 1 for the first convolution, and a kernel size of 1 × 1 and stride size of 1 for the second convolution, respectively), and a ReLU after each convolution. Following the two dropout layers was a convolution layer with a kernel size of 1 × 1 and a stride size of 1. Following the convolution layer, a deconvolution layer with kernel size of 64 × 64 and stride size of 32 was used to upsample the feature maps. Following the deconvolution layer, cropping was performed in order to match the input size. After cropping, softmax was employed to calculate the probability of each class. Finally, an argmax function was utilized to generate a two-dimensional pixel-based class map as the pixel’s index of the tissue type with the highest likelihood using Equation (4). The detailed architecture of the proposed deep learning network is shown in Figure 4 and Table 1.

3. Data and Results

3.1. Data Preparation

EBUS-TBNA cytology samples of patients were collected from the Department of Pathology at the Tri-Service General Hospital. This research was approved by the Institutional Review Board of the Tri-Service General Hospital (TSGH) (TSGHIRB No.1-107-05-171 and No.B202005070), and informed consent was formally waived by the approving committee. The data were deidentifed and used for a retrospective study without impacting patient care. All the methods were performed in accordance with the relevant guidelines and regulations. For this retrospective study, slides of nodal aspirates from consecutive EBUS-TBNA procedures performed at TSGH between January 2018 and December 2020 were obtained. We refer to two slides on average from a single lymph node as a case because each nodal sampling could involve more than one slide. All cases had an associated cell block that contributed to the final diagnosis. All the lung cancer patients had thoracic CT before EBUS-TBNA examination, which showed enlarged hilar or mediastinal lymph nodes. Each patient was aware of the indications, technique, and possible complications of EBUS-TBNA. EBUS-TBNA was performed under general endotracheal anesthesia and the procedure was performed only under moderate sedation. All the EBUS-TBNA procedures were performed by pulmonologists. On the advice of pulmonologists and thoracic surgeons, EBUS-TBNA was used to take selective lymph node samples.
There is still some limitation of EBUS-TBNA in mediastinal lymph nodes sampling. The size of the lymph node, the SUVmax of the lymph node, and the characteristics of the node in color Doppler image may provide information when the physician is performing the EBUS-TBNA procedure [40,41,42]. However, the yield rate of this procedure depends on the experience of the operator. The new diagnostic tool of elastography was applied to predict the lymph node status [43]. In this study, we analyzed the clinical data including the size of the lymph node, SUVmax of the tumor, and blue color proportion of the node in elastography (35%) to select the possible metastatic lymph nodes for EBUS-TBNA cytological examination.
Initially, aspirated material was pushed out by the internal stylet and smeared onto glass slides for immediate ROSE. Then, using the following staining procedures, a quick cytological Liu’s staining was carried out. The Liu A solution was added to alcohol-fixed endobronchial-ultrasound-guided transbronchial needle aspiration smears and allowed to stand for 30 s. The Liu B solution (about double the dose of A) was then added to the smears without discarding the Liu A solution. To stain for 2 min, the two solutions were completely mixed. Before microscopic analysis, the smears were rinsed with tap water and air dried or dried with filter paper. The EBUS-TBNA specimen’s cytological imaging data were evaluated. For histological assessment, the residual aspirate and additional needle routes were preserved in formalin solution. Staff pathologists examined cytology and tissue slides.
Based on the diagnosis by two cytopathologists, we classified the EBUS-TBNA samples into 47 positive slides from 24 patient cases and 75 negative slides from 38 patients. The positive group was defined as the cases of enlarged mediastinal lymph nodes accompanied by tumor cells metastasis, and the follow-up diagnosis of the positive cases was determined as adenocarcinoma (n = 20), squamous-cell carcinoma (n = 3), and small-cell carcinoma (n = 1). On the other hand, the negative group contained the cases diagnosed as having no evidence of tumor cells metastasis, and the associated follow-up diagnosis was determined as reactive lymph nodes (n = 36) and benign granulomatous lymphadenopathy (n = 2). Images were produced by a digital slide scanner (Leica AT Turbo (Leica, Germany) with a 20× objective lens and stored as a pyramid data structure with several downsampled variants of the base image in the same file encoded in the svs file format. This study employed a total of 122 slides, including 47 malignant slides and 75 benign slides. The average slide dimensions were 97,608 × 45,309 pixels with a physical size of 49.13 × 22.80 mm 2 . All the slides were anonymized during the data collection. Pixel-level annotations were generated as reference standards by two expert pathologists. Two experiments were performed in evaluation and the data division process in the first and second experiments was performed by random sampling of the slides. For the first experiment, the AI models were trained using a total of 46 WSIs, which accounted for 37.7% of the whole dataset and included 36 malignant slides and 10 benign slides, where around 0.02% benign cells and 0.03% malignant cells were sampled, and the remaining 76 WSIs, which accounted for 62.3% of the whole dataset and included 11 malignant and 65 benign slides, were employed as a separate testing set. Figure 1b shows the data distribution in further detail. For the second experiment, a three-fold cross-validation was performed. During the evaluation, each model was independently assessed on the testing set.

3.2. Experimental Settings

The proposed deep learning framework was implemented using the Caffe library on a workstation equipped with four NVIDIA GeForce GTX 1080 Ti GPU cards, Intel Xeon Gold 6134 CPU, and 128 GB memory, and a workstation equipped with one NVIDIA GeForce GTX 1080 Ti GPU card, Intel Xeon CPU E5-2650 v2, and 32 GB memory. For training, the proposed method was trained with the settings as follows: learning rate, 1 × 10 10 ; dropout ratio, 0.5; weight decay, 0.0005; and batch size, 1. A stochastic gradient descent (SGD) optimizer was utilized for the optimization, and the categorical cross-entropy function was employed as a loss function. In addition, the baseline models (U-Net, SegNet, and FCN) were implemented utilizing the keras implementation of the image segmentation models by Gupta et al. [44] on the workstation equipped with one NVIDIA GeForce GTX 1080 Ti GPU card, Intel Xeon CPU E5-2650 v2, and 32 GB memory. For training, the baseline models were trained using the default settings based on the codes provided in the literature. Furthermore, for WSI processing, the proposed deep learning model and the baseline models utilized the same framework that is described in detail in Section 2.1.

3.3. Evaluation Metrics

The quantitative evaluation of the segmentation performance was produced using four measurements, i.e., precision, sensitivity, Dice similarity coefficient (DSC), and Intersection over Union (IoU). The evaluation metrics were computed as follows:
P r e c i s i o n = T P T P + F P
S e n s i t i v i t y = T P T P + F N
D S C = 2 T P 2 T P + F P + F N
I o U = T P T P + F N + F P
where TP denotes the true positives, TN represents the true negatives, FP denotes false positives, and FN is the false negatives.

3.4. Quantitative Evaluation with Statistical Analysis

For the first experiment, we evaluated the performance of the proposed method with the three state-of-the-art deep learning models, including U-Net, SegNet, and FCN, as shown in Table 2(a). The experimental results demonstrated that the proposed method achieved a precision of 93.4%, sensitivity of 89.8%, DSC of 82.2%, and IoU of 83.2% while the U-Net model obtained a precision of 67.0%, sensitivity of 54.1%, DSC of 55.7%, and IoU of 47.3%, the SegNet model obtained a precision of 55.8%, sensitivity of 43.8%, DSC of 42.2%, and IoU of 33.6%, and the FCN model obtained a precision of 59.0%, sensitivity of 63.0%, DSC of 52.5%, and IoU of 42.7%, respectively. To further validate the robustness and effectiveness of the proposed method, we assessed the quantitative scores with Fisher’s least significant difference (LSD), using SPSS software [45] (see Table 2(b)). The LSD test results demonstrated that the proposed method significantly outperformed the baseline methods (U-Net, SegNet, and FCN) based on precision, sensitivity, DSC, and IoU (p< 0.001 ). Figure 5 presents the box plots of the quantitative evaluation results of the proposed method and three baseline methods (U-Net, SegNet, and FCN) for metastatic lesion segmentation in EBUS-TBNA, showing that the proposed method outperforms the baseline methods in terms of precision, sensitivity, DSC, and IoU (p< 0.001 ).
For the second experiment using a three-fold cross validation (see Table 3), the proposed method demonstrated a promising performance, higher than the two best performing baseline methods, including U-Net and FCN, with a precision of 91.8 ± 1.2, sensitivity of 96.3 ± 0.8, DSC of 94.0 ± 1.0, and IoU of 88.7 ± 1.8 while the U-Net model obtained a precision of 68.6 ± 17.0, sensitivity of 52.3 ± 9.2, DSC of 57.6 ± 2.8, and IOU of 40.5 ± 2.8 and the FCN model obtained a precision of 72.3 ± 19.0, sensitivity of 59.0 ± 15.4, DSC of 61.6 ± 0.5, and IOU of 44.5 ± 0.6, respectively. The experimental results of the two experiments demonstrated that the proposed method could be used for the automatic detection and segmentation of lymph node metastasis in EBUS-TBNA cytological slides robustly. To further investigate if there was any overfitting or underfitting, Figure 6 shows an epochwise train–test curve that presents the training loss with the training and the testing DSC scores through iterations/epochs, indicating that the proposed model is least likely to be underfitting or overfitting. Figure 6 was generated based on the result of the second fold in the cross-validation.
For the qualitative evaluation, Figure 7 compares the segmentation outputs by the proposed method and the three baseline methods (U-Net, SegNet, and FCN) for metastatic lesion segmentation. The results show that the proposed method generated annotations consistent with the reference standard produced by an expert pathologist while the three baseline methods performed poorly for metastatic lesions segmentation. Figure 8 presents a pathologist’s assessment of the segmentation results produced by the proposed approach. It can be seen that the automatic predictive results generated by the proposed method demonstrate the typical lung adenocarcinoma features, including syncytial tumor group, hyperchromatic nuclei, high N/C ratio, prominent nucleoli, irregular nuclear shape, pleomorphic nuclei, and intracytoplasmic vacuolation, that are required by pathologists to make a diagnostic evaluation.

3.5. Computational Time Comparison

The computational time is critical for methods to be implemented in practical clinical use. To demonstrate the computational efficiency of the proposed method, we analyzed the AI computational time utilizing different hardware specifications (see Section 3.2). We evaluated the computational efficiency of the proposed method and the three baseline methods by calculating the AI inference time on a 77,688 × 41,014 sized WSI. To process a 77,688 × 41,014 sized WSI, the proposed method took less than a minute utilizing four NVIDIA Geforce GTX 1080 Ti GPU cards and 3.6 min using a single NVIDIA Geforce GTX 1080 Ti GPU card while the U-Net model required 9.1 min, the SegNet model required 8.4 min, and the FCN model required 12.4 min, as shown in Table 4. The computational evaluation experiments demonstrated that the proposed technique could process WSI 2.5 times faster than U-Net, 2.3 times faster than SegNet, and 3.4 times quicker than FCN, even with a less expensive GPU, demonstrating that the proposed method was an effective, suitable, and cost-effective solution for processing multigigapixel histopathological images in clinical practice, where rapid diagnosis analyses are required.

4. Discussion and Conclusions

The diagnosis of mediastinal metastatic lesions is required to determine the best treatment plan. It is critical for lung cancer staging to correctly classify lymph nodes as benign or malignant. The detection of mediastinal lymph node metastases can save unnecessary procedures. Although TBNA, transesophageal-ultrasound-guided needle aspiration, computed-tomography-guided transbronchial aspiration, and mediastinoscopy can all be used to sample lymph nodes, the majority of these techniques have drawbacks, including the need for general anesthesia, a low yield, and poor accessibility. The EBUS-TBNA has emerged as a revolutionary method to estimate both benign or metastatic lymph nodes with a high diagnostic yield. It has several advantages including being a minimally invasive approach, safe, cost effective, and offering real-time image guidance [11]. In 19% of patients, EBUS-TBNA may give diagnostic information as well as staging information, reducing the need for additional invasive procedures [8]. The EBUS-TBNA may be utilized for on-site assessment because of immediate smears from needle aspiration and brushing specimens [11,12,13,14,15], and the diagnostic rate at various sites ranges from 50% to 90% [46,47,48,49,50,51,52,53,54,55]. The large variation in the diagnosis rate of different centers is due to the manual diagnosis. In this work, we presented a deep learning method that produced consistent and accurate diagnosis of mediastinal metastasis lesions for lung cancer staging. Apart from lung cancer staging, EBUS-TBNA has been used in various medical applications. The accuracy of EBUS-TBNA for detecting mediastinal metastases from extrathoracic malignancy and lymphoma was 85–95 percent [56,57] and 91–97 percent [58], respectively, while the accuracy of EBUS-TBNA for diagnosing sarcoidosis was 79 percent [59]. Our future plans include applying the proposed deep-learning-based framework to other applications such as detecting mediastinal metastasis from extrathoracic malignancy and lymphoma, diagnosing sarcoidosis, detecting endoscopic-ultrasonography-steered fine-needle aspiration for the diagnosis of pancreaticobiliary lesions, and analyzing pleural effusion or ascites. In addition, we are interested in exploring further model improvement, such as incorporating upsampling or the transpose details used for dilation in deconvolution.
Glass slides may now be converted into WSIs, allowing pathological images to be examined using computer-based approaches [22]. WSIs are obtained at very high resolution (order of 10 gigapixels) and are difficult to visually scrutinize thoroughly. Over the last decade, there has been a surge of interest in creating computational tools to assist medical practitioners with improving the efficiency of medical image analysis. There are compelling grounds to assume that digital pathology in conjunction with AI is a viable solution to this problem since it assists the establishment of more exact diagnoses, decreases examination time, and reduces the labor of pathologists as well as the examination cost. The computational cost is a substantial barrier to using computational approaches to diagnose gigapixel WSIs, and as a result, many existing algorithms are not suitable for real-world deployment. A complete and comprehensive automated inspection of WSIs with high accuracy may require additional time and computer resources. We presented in this study an efficient and effective approach for automatic detection and segmentation of lymph nodes metastasis in EBUS-TBNA cytological slides in seconds. Although it is challenging for physicians to detect metastatic lesions using PET/CT imaging, AI utilizing deep learning has shown acceptable and fast diagnostic abilities for distinguishing malignant from benign mediastinal lymph nodes. The effectiveness and robustness of the proposed method was evaluated using two experiments. The experimental results demonstrated that the proposed method achieved promising performance in the segmentation of enlarged mediastinal lymph nodes metastasis in EBUS-TBNA WSIs, significantly outperforming the state-of-the-art baseline approaches (p< 0.001 ). The high precision and recall in the experimental results demonstrated that the proposed method could aid the pathologists with aspects of their evaluation applicable to an automatic analysis.

Author Contributions

C.-W.W. and T.-K.C. conceived the idea of this work. C.-W.W. designed the methodology and the software tool of this work. Y.-C.L., Y.-J.L. and M.-A.K. carried out the validation of the methodology of this work. M.-A.K. performed the formal analysis of this work. Y.-C.L. performed the investigation. T.-W.H. and Y.-J.L. participated in the annotation or curation of the dataset. C.-W.W., M.-A.K. and T.-K.C. prepared and wrote the manuscript. C.-W.W. reviewed and revised the manuscript. M.-A.K. prepared the visualization of the manuscript. C.-W.W. supervised this work. C.-W.W. and T.-K.C. administered this work. C.-W.W. and T.-K.C. acquired funding for this work. All authors reviewed the manuscript before submission. All authors have read and agreed to the published version of the manuscript.

Funding

This study is supported by the Ministry of Science and Technology of Taiwan, under a grant (MOST109-2221-E-011-018-MY3, MOST111-2320-B-016-009), Tri-Service General Hospital, Taipei, Taiwan (TSGH-C-108086, TSGH-D-109094, TSGH-D-110036, and TSGH-A-111010), and National Taiwan University of Science and Technology-Tri-Service General Hospital, Taipei, Taiwan (NTUST-TSGH-111-05).

Institutional Review Board Statement

This research was approved by the Institutional Review Board of the Tri-Service General Hospital (TSGH) (TSGHIRB No.1-107-05-171) on 16th of December 2021 and (TSGHIRB No.B202005070) on 30th of June 2021, and informed consents were formally waived by the approving committee. The data were deidentifed and used for a retrospective study without impacting patient care.

Informed Consent Statement

Patient consent was formally waived by the approving review board, and the data were deidentified and used for a retrospective study without impacting patient care.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare that they have no known competing financial interest or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Society, A.C. Early Detection, Diagnosis, and Staging. Available online: https://www.cancer.org/cancer/lung-cancer/detection-diagnosis-staging/detection.html (accessed on 22 May 2020).
  2. Jemal, A.; Bray, F.; Center, M.M.; Ferlay, J.; Ward, E.; Forman, D. Global cancer statistics. CA Cancer J. Clin. 2011, 61, 69–90. [Google Scholar] [CrossRef] [PubMed]
  3. Hasteh, F.; Lin, G.Y.; Weidner, N.; Michael, C.W. The use of immunohistochemistry to distinguish reactive mesothelial cells from malignant mesothelioma in cytologic effusions. Cancer Cytopathol. J. Am. Cancer Soc. 2010, 118, 90–96. [Google Scholar] [CrossRef] [PubMed]
  4. de Sousa, V.M.L.; Carvalho, L. Heterogeneity in lung cancer. Pathobiology 2018, 85, 96–107. [Google Scholar] [CrossRef] [PubMed]
  5. Arias, S.; Liu, Q.; Frimpong, B.; Lee, H.; Feller-Kopman, D.; Yarmus, L.; Wang, K. Role of the endobronchial landmarks guiding TBNA and EBUS-TBNA in lung cancer staging. Can. Respir. J. 2016, 2016, 1652178. [Google Scholar] [CrossRef]
  6. Righi, L.; Franzi, F.; Montarolo, F.; Gatti, G.; Bongiovanni, M.; Sessa, F.; La Rosa, S. Endobronchial ultrasound-guided transbronchial needle aspiration (EBUS-TBNA)—From morphology to molecular testing. J. Thorac. Dis. 2017, 9, S395. [Google Scholar] [CrossRef]
  7. Vaidya, P.J.; Kate, A.H.; Chhajed, P.N. Endobronchial ultrasound-guided transbronchial needle aspiration: The standard of care for evaluation of mediastinal and hilar lymphadenopathy. J. Cancer Res. Ther. 2013, 9, 549. [Google Scholar]
  8. Czarnecka-Kujawa, K.; Yasufuku, K. The role of endobronchial ultrasound versus mediastinoscopy for non-small cell lung cancer. J. Thorac. Dis. 2017, 9, S83. [Google Scholar] [CrossRef]
  9. Suo, J.; Zhang, Q.; Chang, W.; Shi, J.; Yan, Z.; Chen, M. Evaluation of axillary lymph node metastasis by using radiomics of dual-modal ultrasound composed of elastography and b-mode. Zhongguo Liao Xie Zhi Chin. J. Med. Instrum. 2017, 41, 313–316. [Google Scholar]
  10. Fielding, D.I.; Kurimoto, N. EBUS-TBNA/staging of lung cancer. Clin. Chest Med. 2013, 34, 385–394. [Google Scholar] [CrossRef]
  11. Herth, F.; Becker, H.D.; Ernst, A. Conventional vs. endobronchial ultrasound-guided transbronchial needle aspiration: A randomized trial. Chest 2004, 125, 322–325. [Google Scholar] [CrossRef]
  12. Skov, B.G.; Baandrup, U.; Jakobsen, G.K.; Kiss, K.; Krasnik, M.; Rossen, K.; Vilmann, P. Cytopathologic diagnoses of fine-needle aspirations from endoscopic ultrasound of the mediastinum: Reproducibility of the diagnoses and representativeness of aspirates from lymph nodes. Cancer Cytopathol. Interdiscip. Int. J. Am. Cancer Soc. 2007, 111, 234–241. [Google Scholar] [CrossRef] [PubMed]
  13. Sakairi, Y.; Nakajima, T.; Iizasa, T.; Itami, M.; Yoshino, I.; Kimura, H. Diagnosis of lymph node metastasis by endobronchial ultrasound-guided transbronchial needle aspiration more than 1 year after lung cancer resection: Report of a case. Surg. Today 2011, 41, 983–985. [Google Scholar] [CrossRef] [PubMed]
  14. Kuo, C.H.; Chen, H.C.; Chung, F.T.; Lo, Y.L.; Lee, K.Y.; Wang, C.W.; Kuo, W.H.; Yen, T.C.; Kuo, H.P. Diagnostic value of EBUS-TBNA for lung cancer with non-enlarged lymph nodes: A study in a tuberculosis-endemic country. PLoS ONE 2011, 6, e16877. [Google Scholar] [CrossRef]
  15. Groth, S.S.; Whitson, B.A.; D’Cunha, J.; Maddaus, M.A.; Alsharif, M.; Andrade, R.S. Endobronchial ultrasound-guided fine-needle aspiration of mediastinal lymph nodes: A single institution’s early learning curve. Ann. Thorac. Surg. 2008, 86, 1104–1110. [Google Scholar] [CrossRef]
  16. Canneto, B.; Ferraroli, G.; Falezza, G.; Infante, M.V. Ideal conditions to perform EBUS-TBNA. J. Thorac. Dis. 2017, 9, S414. [Google Scholar] [CrossRef]
  17. Yasufuku, K.; Nakajima, T.; Motoori, K.; Sekine, Y.; Shibuya, K.; Hiroshima, K.; Fujisawa, T. Comparison of endobronchial ultrasound, positron emission tomography, and CT for lymph node staging of lung cancer. Chest 2006, 130, 710–718. [Google Scholar] [CrossRef]
  18. Yu, K.H.; Beam, A.L.; Kohane, I.S. Artificial intelligence in healthcare. Nat. Biomed. Eng. 2018, 2, 719–731. [Google Scholar] [CrossRef] [PubMed]
  19. Fu, Y.; Jung, A.W.; Torne, R.V.; Gonzalez, S.; Vöhringer, H.; Shmatko, A.; Yates, L.R.; Jimenez-Linan, M.; Moore, L.; Gerstung, M. Pan-cancer computational histopathology reveals mutations, tumor composition and prognosis. Nat. Cancer 2020, 1, 800–810. [Google Scholar] [CrossRef]
  20. Cardoso, A.; Neves, I.; Magalhães, A.; Sucena, M.; Barroca, H.; Fernandes, G. The value of rapid on-site evaluation during EBUS-TBNA. Rev. Port. Pneumol. 2015, 21, 253–258. [Google Scholar] [CrossRef]
  21. Jain, D.; Allen, T.C.; Aisner, D.L.; Beasley, M.B.; Cagle, P.T.; Capelozzi, V.L.; Hariri, L.P.; Lantuejoul, S.; Miller, R.; Mino-Kenudson, M.; et al. Rapid on-site evaluation of endobronchial ultrasound–guided transbronchial needle aspirations for the diagnosis of lung cancer: A perspective from members of the pulmonary pathology society. Arch. Pathol. Lab. Med. 2018, 142, 253–262. [Google Scholar] [CrossRef]
  22. Mukhopadhyay, S.; Feldman, M.D.; Abels, E.; Ashfaq, R.; Beltaifa, S.; Cacciabeve, N.G.; Cathro, H.P.; Cheng, L.; Cooper, K.; Dickey, G.E.; et al. Whole Slide Imaging Versus Microscopy for Primary Diagnosis in Surgical Pathology: A Multicenter Blinded Randomized Noninferiority Study of 1992 Cases (Pivotal Study). Am. J. Surg. Pathol. 2018, 42, 39–52. [Google Scholar] [CrossRef] [PubMed]
  23. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Wei, J.W.; Tafe, L.J.; Linnik, Y.A.; Vaickus, L.J.; Tomita, N.; Hassanpour, S. Pathologist-level classification of histologic patterns on resected lung adenocarcinoma slides with deep neural networks. Sci. Rep. 2019, 9, 3358. [Google Scholar] [CrossRef]
  25. Huang, P.; Lin, C.T.; Li, Y.; Tammemagi, M.C.; Brock, M.V.; Atkar-Khattra, S.; Xu, Y.; Hu, P.; Mayo, J.R.; Schmidt, H.; et al. Prediction of lung cancer risk at follow-up screening with low-dose CT: A training and validation study of a deep learning method. Lancet Digit. Health 2019, 1, e353–e362. [Google Scholar] [CrossRef]
  26. van Rijthoven, M.; Balkenhol, M.; Siliņa, K.; van der Laak, J.; Ciompi, F. HookNet: Multi-resolution convolutional neural networks for semantic segmentation in histopathology whole-slide images. Med. Image Anal. 2021, 68, 101890. [Google Scholar] [CrossRef]
  27. Wang, C.W.; Huang, S.C.; Lee, Y.C.; Shen, Y.J.; Meng, S.I.; Gaol, J.L. Deep Learning for Bone Marrow Cell Detection and Classification on Whole-Slide Images. Med. Image Anal. 2021, 75, 102270. [Google Scholar] [CrossRef] [PubMed]
  28. Dov, D.; Kovalsky, S.Z.; Assaad, S.; Cohen, J.; Range, D.E.; Pendse, A.A.; Henao, R.; Carin, L. Weakly supervised instance learning for thyroid malignancy prediction from whole slide cytopathology images. Med. Image Anal. 2021, 67, 101814. [Google Scholar] [CrossRef] [PubMed]
  29. Tang, C.; Xinwang, L.; Zheng, X.; Li, W.; Xiong, J.; Wang, L.; Zomaya, A.; Longo, A. DeFusionNET: Defocus blur detection via recurrently fusing and refining discriminative multi-scale deep features. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 955–968. [Google Scholar] [CrossRef]
  30. Masud, M.; Hossain, M.S.; Alhumyani, H.; Alshamrani, S.S.; Cheikhrouhou, O.; Ibrahim, S.; Muhammad, G.; Rashed, A.E.E.; Gupta, B. Pre-trained convolutional neural networks for breast cancer detection using ultrasound images. ACM Trans. Internet Technol. 2021, 21, 1–17. [Google Scholar] [CrossRef]
  31. Falk, T.; Mai, D.; Bensch, R.; Çiçek, Ö.; Abdulkadir, A.; Marrakchi, Y.; Böhm, A.; Deubner, J.; Jäckel, Z.; Seiwald, K.; et al. U-Net: Deep learning for cell counting, detection, and morphometry. Nat. Methods 2019, 16, 67–70. [Google Scholar] [CrossRef]
  32. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  33. Shelhamer, E.; Long, J.; Darrell, T. Fully convolutional networks for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef] [PubMed]
  34. Lin, H.; Chen, H.; Graham, S.; Dou, Q.; Rajpoot, N.; Heng, P.A. Fast scannet: Fast and dense analysis of multi-gigapixel whole-slide images for cancer metastasis detection. IEEE Trans. Med. Imaging 2019, 38, 1948–1958. [Google Scholar] [CrossRef] [PubMed]
  35. Campanella, G.; Hanna, M.G.; Geneslaw, L.; Miraflor, A.; Werneck Krauss Silva, V.; Busam, K.J.; Brogi, E.; Reuter, V.E.; Klimstra, D.S.; Fuchs, T.J. Clinical-grade computational pathology using weakly supervised deep learning on whole slide images. Nat. Med. 2019, 25, 1301–1309. [Google Scholar] [CrossRef]
  36. Khalil, M.A.; Lee, Y.C.; Lien, H.C.; Jeng, Y.M.; Wang, C.W. Fast Segmentation of Metastatic Foci in H&E Whole-Slide Images for Breast Cancer Diagnosis. Diagnostics 2022, 12, 990. [Google Scholar]
  37. Lin, Y.J.; Chao, T.K.; Khalil, M.A.; Lee, Y.C.; Hong, D.Z.; Wu, J.J.; Wang, C.W. Deep Learning Fast Screening Approach on Cytological Whole Slides for Thyroid Cancer Diagnosis. Cancers 2021, 13, 3891. [Google Scholar] [CrossRef]
  38. Wang, C.W.; Liou, Y.A.; Lin, Y.J.; Chang, C.C.; Chu, P.H.; Lee, Y.C.; Wang, C.H.; Chao, T.K. Artificial intelligence-assisted fast screening cervical high grade squamous intraepithelial lesion and squamous cell carcinoma diagnosis and treatment planning. Sci. Rep. 2021, 11, 16244. [Google Scholar] [CrossRef]
  39. Wang, C.W.; Lee, Y.C.; Chang, C.C.; Lin, Y.J.; Liou, Y.A.; Hsu, P.C.; Chang, C.C.; Sai, A.K.O.; Wang, C.H.; Chao, T.K. A Weakly Supervised Deep Learning Method for Guiding Ovarian Cancer Treatment and Identifying an Effective Biomarker. Cancers 2022, 14, 1651. [Google Scholar] [CrossRef]
  40. Ayub, I.I.; Mohan, A.; Madan, K.; Hadda, V.; Jain, D.; Khilnani, G.; Guleria, R. Identification of specific EBUS sonographic characteristics for predicting benign mediastinal lymph nodes. Clin. Respir. J. 2018, 12, 681–690. [Google Scholar] [CrossRef]
  41. Kandemir, Z.; Şentürk, A.; Özdemir, E.; Yildirim, N.; Hasanoğlu, H.C.; Keskin, M.; Türkölmez, Ş. The evaluation of hypermetabolic mediastinal-hilar lymph nodes determined byPET/CT in pulmonary and extrapulmonary malignancies: Correlation with EBUS-TBNA. Turk. J. Med. Sci. 2015, 45, 1234–1242. [Google Scholar] [CrossRef]
  42. Zivković, N.P.; Ostović, K.T.; Bosnjak, D.; Tudorić, N. Initial Results of Diagnostic Application of Endobronchial Ultrasound in Clinical Hospital Dubrava. Lijec. Vjesn. 2016, 138, 34–38. [Google Scholar] [PubMed]
  43. Trosini-Désert, V.; Jeny, F.; Maksud, P.; Giron, A.; Degos, V.; Similowski, T. Contribution of endobronchial ultrasound elastography to the characterization of mediastinal lymphadenopathy: A single-center, prospective, observational study. Respir. Med. Res. 2019, 76, 28–33. [Google Scholar] [CrossRef]
  44. Gupta, D.; Jhunjhunu, W.; Wala, R.; Juston, M.; Jaled, M.C. Image Segmentation Keras: Implementation of Segnet, FCN, UNet, PSPNet and Other Models in Keras. Available online: https://github.com/divamgupta/image-segmentation-keras (accessed on 15 October 2020).
  45. IBM Corp. In Ibm Spss Statistics for Windows; IBM Corp: Armonk, NY, USA, 2013.
  46. Gu, P.; Zhao, Y.Z.; Jiang, L.Y.; Zhang, W.; Xin, Y.; Han, B.H. Endobronchial ultrasound-guided transbronchial needle aspiration for staging of lung cancer: A systematic review and meta-analysis. Eur. J. Cancer 2009, 45, 1389–1396. [Google Scholar] [CrossRef]
  47. Dhooria, S.; Agarwal, R.; Aggarwal, A.N.; Bal, A.; Gupta, N.; Gupta, D. Differentiating tuberculosis from sarcoidosis by sonographic characteristics of lymph nodes on endobronchial ultrasonography: A study of 165 patients. J. Thorac. Cardiovasc. Surg. 2014, 148, 662–667. [Google Scholar] [CrossRef] [PubMed]
  48. Dhooria, S.; Aggarwal, A.N.; Singh, N.; Gupta, D.; Behera, D.; Gupta, N.; Agarwal, R. Endoscopic ultrasound-guided fine-needle aspiration with an echobronchoscope in undiagnosed mediastinal lymphadenopathy: First experience from India. Lung India Off. Organ Indian Chest Soc. 2015, 32, 6. [Google Scholar]
  49. Dhooria, S.; Sehgal, I.S.; Nalini Gupta, B.R.; Aggarwal, A.N.; Behera, D.; Agarwal, R. Yield of new versus reused endobronchial ultrasound-guided transbronchial needle aspiration needles: A retrospective analysis of 500 patients. Lung India Off. Organ Indian Chest Soc. 2016, 33, 367. [Google Scholar]
  50. Dhooria, S.; Sehgal, I.S.; Gupta, N.; Aggarwal, A.N.; Behera, D.; Agarwal, R. Diagnostic yield and complications of EBUS-TBNA performed under bronchoscopist-directed conscious sedation. J. Bronchol. Interv. Pulmonol. 2017, 24, 7–14. [Google Scholar] [CrossRef]
  51. Dhooria, S.; Sehgal, I.S.; Gupta, N.; Bal, A.; Prasad, K.T.; Aggarwal, A.N.; Ram, B.; Agarwal, R. A randomized trial evaluating the effect of 10 versus 20 revolutions inside the lymph node on the diagnostic yield of EBUS-TBNA in subjects with sarcoidosis. Respiration 2018, 96, 464–471. [Google Scholar] [CrossRef]
  52. Ost, D.E.; Ernst, A.; Lei, X.; Feller-Kopman, D.; Eapen, G.A.; Kovitz, K.L.; Herth, F.J.; Simoff, M. Diagnostic yield of endobronchial ultrasound-guided transbronchial needle aspiration: Results of the AQuIRE Bronchoscopy Registry. Chest 2011, 140, 1557–1566. [Google Scholar] [CrossRef]
  53. Madan, K.; Mohan, A.; Ayub, I.I.; Jain, D.; Hadda, V.; Khilnani, G.C.; Guleria, R. Initial experience with endobronchial ultrasound-guided transbronchial needle aspiration (EBUS-TBNA) from a tuberculosis endemic population. J. Bronchol. Interv. Pulmonol. 2014, 21, 208–214. [Google Scholar] [CrossRef]
  54. Madan, K.; Dhungana, A.; Mohan, A.; Hadda, V.; Jain, D.; Arava, S.; Pandey, R.M.; Khilnani, G.C.; Guleria, R. Conventional Transbronchial Needle Aspiration Versus Endobronchial Ultrasound–guided Transbronchial Needle Aspiration, With or Without Rapid On-Site Evaluation, for the Diagnosis of Sarcoidosis: A Randomized Controlled Trial. J. Bronchol. Interv. Pulmonol. 2017, 24, 48–58. [Google Scholar] [CrossRef] [PubMed]
  55. Madan, N.K.; Madan, K.; Jain, D.; Walia, R.; Mohan, A.; Hadda, V.; Mathur, S.; Iyer, V.K.; Khilnani, G.C.; Guleria, R. Utility of conventional transbronchial needle aspiration with rapid on-site evaluation (c-TBNA-ROSE) at a tertiary care center with endobronchial ultrasound (EBUS) facility. J. Cytol./Indian Acad. Cytol. 2016, 33, 22. [Google Scholar] [CrossRef] [PubMed]
  56. Sanz-Santos, J.; Cirauqui, B.; Sanchez, E.; Andreo, F.; Serra, P.; Monso, E.; Castella, E.; Llatjós, M.; Mesa, M.; Ruiz-Manzano, J.; et al. Endobronchial ultrasound-guided transbronchial needle aspiration in the diagnosis of intrathoracic lymph node metastases from extrathoracic malignancies. Clin. Exp. Metastasis 2013, 30, 521–528. [Google Scholar] [CrossRef] [PubMed]
  57. Özgül, M.A.; Çetinkaya, E.; Tutar, N.; Özgül, G.; Onaran, H.; Bilaçeroglu, S. Endobronchial ultrasound-guided transbronchial needle aspiration for the diagnosis of intrathoracic lymphadenopathy in patients with extrathoracic malignancy: A study in a tuberculosis-endemic country. J. Cancer Res. Ther. 2013, 9, 416. [Google Scholar]
  58. Erer, O.F.; Erol, S.; Anar, C.; Aydoğdu, Z.; Özkan, S.A. Diagnostic yield of EBUS-TBNA for lymphoma and review of the literature. Endosc. Ultrasound 2017, 6, 317. [Google Scholar]
  59. Agarwal, R.; Srinivasan, A.; Aggarwal, A.N.; Gupta, D. Efficacy and safety of convex probe EBUS-TBNA in sarcoidosis: A systematic review and meta-analysis. Respir. Med. 2012, 106, 883–892. [Google Scholar] [CrossRef] [Green Version]
Figure 1. System workflow and dataset information. (a) System workflow; (a, i) collection of patients’ lung smear samples through EBUS-TBNA; (a, ii) preparation and staining of EBUS smears slides using Liu’s staining; (a, iii) digitalization of Liu stained microscopic slides at the microscopic resolution (20× magnification); (a, iv) random distribution of digitized whole-slide gigapixel images into a separate training (37.7%) set and a testing (62.3%) set for the first experiment and a three-fold cross validation for the second experiment; (a, v) processing of WSIs with fast background filtering; (a, vi) rapid detection of enlarged mediastinal lymph nodes metastasis in EBUS-TBNA cytological slides using the proposed deep learning model in seconds. (b) Distribution of dataset for each class, as well as for training and testing sets for the first experiment and a three-fold cross validation for the second experiment. (c) Distribution of the number of tiles per WSI. (d) Size distribution of the WSIs, with width and height represented in blue and orange, respectively.
Figure 1. System workflow and dataset information. (a) System workflow; (a, i) collection of patients’ lung smear samples through EBUS-TBNA; (a, ii) preparation and staining of EBUS smears slides using Liu’s staining; (a, iii) digitalization of Liu stained microscopic slides at the microscopic resolution (20× magnification); (a, iv) random distribution of digitized whole-slide gigapixel images into a separate training (37.7%) set and a testing (62.3%) set for the first experiment and a three-fold cross validation for the second experiment; (a, v) processing of WSIs with fast background filtering; (a, vi) rapid detection of enlarged mediastinal lymph nodes metastasis in EBUS-TBNA cytological slides using the proposed deep learning model in seconds. (b) Distribution of dataset for each class, as well as for training and testing sets for the first experiment and a three-fold cross validation for the second experiment. (c) Distribution of the number of tiles per WSI. (d) Size distribution of the WSIs, with width and height represented in blue and orange, respectively.
Diagnostics 12 02234 g001
Figure 2. The overview of the proposed deep learning analysis system. (a) Each WSI is converted into a patch-based hierarchical structure (each tile is represented by a black rectangular box). (b) Each WSI is processed using the fast background filtering model to efficiently discard all the background. (c) Each tile is examined by the proposed deep convolutional neural network for segmentation of metastatic lesion from EBUS-TBNA WSIs. (d) The tiles are then stitched together to obtain an output image.
Figure 2. The overview of the proposed deep learning analysis system. (a) Each WSI is converted into a patch-based hierarchical structure (each tile is represented by a black rectangular box). (b) Each WSI is processed using the fast background filtering model to efficiently discard all the background. (c) Each tile is examined by the proposed deep convolutional neural network for segmentation of metastatic lesion from EBUS-TBNA WSIs. (d) The tiles are then stitched together to obtain an output image.
Diagnostics 12 02234 g002
Figure 3. The comparison of the original FCN and the proposed modified FCN. (a) The architecture of original FCN. (b) The architecture of proposed modified FCN.
Figure 3. The comparison of the original FCN and the proposed modified FCN. (a) The architecture of original FCN. (b) The architecture of proposed modified FCN.
Diagnostics 12 02234 g003
Figure 4. The detailed architecture of the proposed deep learning network.
Figure 4. The detailed architecture of the proposed deep learning network.
Diagnostics 12 02234 g004
Figure 5. The box plot of quantitative evaluation results of metastatic lesion segmentation where the outliers >3 × interquartile range are marked with an asterisk. The LSD test results demonstrates that the proposed method performs significantly better than the baseline methods (p< 0.001 ).
Figure 5. The box plot of quantitative evaluation results of metastatic lesion segmentation where the outliers >3 × interquartile range are marked with an asterisk. The LSD test results demonstrates that the proposed method performs significantly better than the baseline methods (p< 0.001 ).
Diagnostics 12 02234 g005
Figure 6. The results of the training loss, the training DSC scores, and the testing DSC scores through iterations/epochs represented in blue and orange, respectively.
Figure 6. The results of the training loss, the training DSC scores, and the testing DSC scores through iterations/epochs represented in blue and orange, respectively.
Diagnostics 12 02234 g006
Figure 7. Visual comparison of the typical segmentation results generated by the proposed method and the three baseline methods (U-Net, SegNet, and FCN) for metastatic lesion segmentation in EBUS-TBNA WSIs.
Figure 7. Visual comparison of the typical segmentation results generated by the proposed method and the three baseline methods (U-Net, SegNet, and FCN) for metastatic lesion segmentation in EBUS-TBNA WSIs.
Diagnostics 12 02234 g007
Figure 8. Pathologist’s assessment of automatic predictive results generated by the proposed method with typical lung adenocarcinoma features required to generate a diagnostic evaluation, including syncytial tumor group, hyperchromatic nuclei, high N/C ratio, prominent nucleoli, irregular nuclear shape, pleomorphic nuclei, and intracytoplasmic vacuolation.
Figure 8. Pathologist’s assessment of automatic predictive results generated by the proposed method with typical lung adenocarcinoma features required to generate a diagnostic evaluation, including syncytial tumor group, hyperchromatic nuclei, high N/C ratio, prominent nucleoli, irregular nuclear shape, pleomorphic nuclei, and intracytoplasmic vacuolation.
Diagnostics 12 02234 g008
Table 1. The architecture of the proposed deep learning network.
Table 1. The architecture of the proposed deep learning network.
LayerFeatures (Train)Features (Inference)Kernel SizeStridePadding
Input512 × 512 × 3512 × 512 × 3---
Conv1_1 + relu1_1710 × 710 × 64710 × 710 × 643 × 31same
Conv1_2 + relu1_2710 × 710 × 64710 × 710 × 643 × 31same
Pool1355 × 355 × 64355 × 355 × 642 × 22valid
Conv2_1 + relu2_1355 × 355 × 128355 × 355 × 1283 × 31same
Conv2_2 + relu2_2355 × 355 × 128355 × 355 × 1283 × 31same
Pool2178 × 178 × 128178 × 178 × 1282 × 22valid
Conv3_1 + relu3_1178 × 178 × 256178 × 178 × 2563 × 31same
Conv3_2 + relu3_2178 × 178 × 256178 × 178 × 2563 × 31same
Conv3_3 + relu3_3178 × 178 × 256178 × 178 × 2563 × 31same
Pool389 × 89 × 25689 × 89 × 2562 × 22valid
Conv4_1 + relu4_189 × 89 × 51289 × 89 × 5123 × 31same
Conv4_2 + relu4_289 × 89 × 51289 × 89 × 5123 × 31same
Conv4_3 + relu4_389 × 89 × 51289 × 89 × 5123 × 31same
Pool445 × 45 × 51245 × 45 × 5122 × 22valid
Conv5_1 + relu5_145 × 45 × 51245 × 45 × 5123 × 31same
Conv5_2 + relu5_245 × 45 × 51245 × 45 × 5123 × 31same
Conv5_3 + relu5_345 × 45 × 51245 × 45 × 5123 × 31same
Pool523 × 23 × 51223 × 23 × 5122 × 22valid
Conv6 + relu6 + Drop617 × 17 × 409617 × 17 × 40967 × 71same
Conv7 + relu7 + Drop717 × 17 × 409617 × 17 × 40961 × 11same
Conv817 × 17 × 317 × 17 × 31 × 11same
Deconv9576 × 576 × 3576 × 576 × 364 × 6432same
Cropping512 × 512 × 3512 × 512 × 3---
Softmax512 × 512 × 3512 × 512 × 3---
Output Class Map512 × 512 × 1512 × 512 × 1---
Table 2. The first experiment: quantitative evaluation and statistical analysis of the proposed method and the baseline methods in metastatic lesions segmentation on EBUS-TBNA WSIs.
Table 2. The first experiment: quantitative evaluation and statistical analysis of the proposed method and the baseline methods in metastatic lesions segmentation on EBUS-TBNA WSIs.
(a): Quantitative Segmentation Results
Method Score95% C.I. for Mean
MeanStd. DeviationStd. ErrorLower BoundUpper Bound
PrecisionProposed Method 0.9340.0940.0280.8710.997
U-Net [31] 0.6700.1230.0360.5920.748
SegNet [32] 0.5580.1580.0460.6440.656
FCN [33] 0.5900.1180.0340.5150.664
SensitivityProposed Method 0.8980.0930.0280.8360.960
U-Net [31] 0.5410.1140.0330.4690.613
SegNet [32] 0.4380.0630.0180.3970.478
FCN [33] 0.6300.2020.0580.5010.758
DSCProposed Method 0.8220.2830.0850.6321.012
U-Net [31] 0.5560.0960.2770.4950.617
SegNet [32] 0.4210.1020.2960.3560.486
FCN [33] 0.5250.1600.0460.4220.627
IoUProposed Method 0.8320.1420.0420.7360.927
U-Net [31] 0.4730.0900.0260.4160.530
SegNet [32] 0.3350.1060.0300.2680.403
FCN [33] 0.4270.1550.0440.3280.526
(b): LSD Multiple Comparisons
Dependent Variable(I) Method(J) MethodMean Difference (I−J)Std. ErrorSig.95% C.I.
Lower BoundUpper Bound
PrecisionProposed MethodU-Net [31]0.263 *0.536<0.0010.1540.511
SegNet [32]0.377 *0.536<0.0010.2680.483
FCN [33]0.343 *0.525<0.0010.2370.449
SensitivityProposed MethodU-Net [31]0.357 *0.038<0.0010.2780.278
SegNet [32]0.460 *0.386<0.0010.3820.332
FCN [33]0.268 *0.541<0.0010.1580.377
DSCProposed MethodU-Net [31]0.265 *0.074<0.0010.1130.416
SegNet [32]0.400 *0.074<0.0010.2480.551
FCN [33]0.296 *0.072<0.0010.1500.443
IoUProposed MethodU-Net [31]0.358 *0.047<0.0010.2610.455
SegNet [32]0.496 *0.047<0.0010.3990.593
FCN [33]0.404 *0.0525<0.0010.2980.510
* The proposed method is significantly better than the baseline methods (U-Net, SegNet, and FCN) using LSD test (p< 0.001 ).
Table 3. The second experiment: quantitative evaluation of the proposed method and the two best-performing baseline models using three-fold cross-validation.
Table 3. The second experiment: quantitative evaluation of the proposed method and the two best-performing baseline models using three-fold cross-validation.
ModelFoldPrecisionSensitivityDSCIoU
Fold10.9090.9540.9310.871
Proposed MethodFold20.9350.9740.9540.913
Fold30.9090.960.9340.877
Mean ± std0.918 ± 0.0120.963 ± 0.0080.940 ± 0.0100.887 ± 0.018
Fold10.5600.5500.5600.385
U-Net [31]Fold20.8800.4200.5600.392
Fold30.6200.6000.6100.438
Mean ± std0.686 ± 0.1700.523 ± 0.0920.576 ± 0.0280.405 ± 0.028
Fold10.7300.5500.6200.449
FCN [33]Fold20.9100.4600.6100.438
Fold30.5300.7600.6200.449
Mean ± std0.723 ± 0.1900.590 ± 0.1540.616 ± 0.0050.445 ± 0.006
Table 4. Computational time comparison of the proposed method and the baseline methods.
Table 4. Computational time comparison of the proposed method and the baseline methods.
MethodWSI Size (pixels)AI Inference Time (min)
Proposed Method77,688 × 41,0140.9
Proposed Method77,688 × 41,0143.6
U-Net [31]77,688 × 41,0149.1
SegNet [32]77,688 × 41,0148.4
FCN [33]77,688 × 41,01412.4
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, C.-W.; Khalil, M.-A.; Lin, Y.-J.; Lee, Y.-C.; Huang, T.-W.; Chao, T.-K. Deep Learning Using Endobronchial-Ultrasound-Guided Transbronchial Needle Aspiration Image to Improve the Overall Diagnostic Yield of Sampling Mediastinal Lymphadenopathy. Diagnostics 2022, 12, 2234. https://doi.org/10.3390/diagnostics12092234

AMA Style

Wang C-W, Khalil M-A, Lin Y-J, Lee Y-C, Huang T-W, Chao T-K. Deep Learning Using Endobronchial-Ultrasound-Guided Transbronchial Needle Aspiration Image to Improve the Overall Diagnostic Yield of Sampling Mediastinal Lymphadenopathy. Diagnostics. 2022; 12(9):2234. https://doi.org/10.3390/diagnostics12092234

Chicago/Turabian Style

Wang, Ching-Wei, Muhammad-Adil Khalil, Yi-Jia Lin, Yu-Ching Lee, Tsai-Wang Huang, and Tai-Kuang Chao. 2022. "Deep Learning Using Endobronchial-Ultrasound-Guided Transbronchial Needle Aspiration Image to Improve the Overall Diagnostic Yield of Sampling Mediastinal Lymphadenopathy" Diagnostics 12, no. 9: 2234. https://doi.org/10.3390/diagnostics12092234

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop