Next Article in Journal
Integrated Approaches to Identify miRNA Biomarkers Associated with Cognitive Dysfunction in Multiple Sclerosis Using Text Mining, Gene Expression, Pathways, and GWAS
Next Article in Special Issue
Differential Diagnosis of Alzheimer Disease vs. Mild Cognitive Impairment Based on Left Temporal Lateral Lobe Hypomethabolism on 18F-FDG PET/CT and Automated Classifiers
Previous Article in Journal
A Mini-Review Regarding the Clinical Outcomes of In Vitro Fertilization (IVF) Following Pre-Implantation Genetic Testing (PGT)-Next Generation Sequencing (NGS) Approach
Previous Article in Special Issue
A Comprehensive Performance Analysis of Transfer Learning Optimization in Visual Field Defect Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detecting Endotracheal Tube and Carina on Portable Supine Chest Radiographs Using One-Stage Detector with a Coarse-to-Fine Attention

1
Department of Computer Science and Information Engineering, National Cheng Kung University, Tainan 701401, Taiwan
2
Department of Surgery, National Cheng Kung University Hospital, College of Medicine, National Cheng Kung University, Tainan 704302, Taiwan
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(8), 1913; https://doi.org/10.3390/diagnostics12081913
Submission received: 5 July 2022 / Revised: 4 August 2022 / Accepted: 5 August 2022 / Published: 7 August 2022
(This article belongs to the Special Issue Artificial Neural Networks in Medical Diagnosis)

Abstract

:
In intensive care units (ICUs), after endotracheal intubation, the position of the endotracheal tube (ETT) should be checked to avoid complications. The malposition can be detected by the distance between the ETT tip and the Carina (ETT–Carina distance). However, it struggles with a limited performance for two major problems, i.e., occlusion by external machine, and the posture and machine of taking chest radiographs. While previous studies addressed these problems, they always suffered from the requirements of manual intervention. Therefore, the purpose of this paper is to locate the ETT tip and the Carina more accurately for detecting the malposition without manual intervention. The proposed architecture is composed of FCOS: Fully Convolutional One-Stage Object Detection, an attention mechanism named Coarse-to-Fine Attention (CTFA), and a segmentation branch. Moreover, a post-process algorithm is adopted to select the final location of the ETT tip and the Carina. Three metrics were used to evaluate the performance of the proposed method. With the dataset provided by National Cheng Kung University Hospital, the accuracy of the malposition detected by the proposed method achieves 88.82 % and the ETT–Carina distance errors are less than 5.333 ± 6.240 mm.

1. Introduction

In intensive care units (ICUs), endotracheal intubation is a common medical procedure when a patient cannot spontaneously breathe. When intubating, the position of an endotracheal tube (ETT) needs to be taken into account because some complications can be caused by its malposition. A deep position can cause tachycardia, hypertension, and/or bronchospasm. On the other hand, a shallow position can cause inadvertent extubation or larynx damage [1]. Extensive research has shown that the appropriate depth of an ETT can be determined by the distance between the ETT and Carina [2]. Therefore, to reduce the likelihood of serious complications, it is essential to develop a method to accurately detect the distance between ETT and Carina.
Recent evidence [3] suggests that chest radiography should be performed after endotracheal intubation. However, in ICUs, chest radiographs (CXRs) are usually obtained in supine anteroposterior (AP) view by a portable X-ray machine. In this situation, several challenges limit the detection accuracy; for example, external monitoring devices, tubes, or catheters can cause ambiguity in the position of an ETT and Carina. Additionally, the image quality of CXR obtained in the supine AP view is lower than the standard posteroanterior (PA) view. Therefore, over the past decade, researchers pay increased attention to computer-aided detection (CAD) methods in an attempt to improve the detection of unsuitable endotracheal intubation and reduce the burden on doctors.
Previous automatic CAD methods for ETTs detection [4,5,6] have a similar procedure, which can be simplified into four steps, i.e., preprocessing, finding neck, seed generation, and region growing. These methods use feature extraction to detect the ETT and have achieved acceptable results. However, such approaches have failed to address the problem that the manual templates and hyperparameters should be determined based on experience or morphology in the procedure. These shortages may limit the ability of generalization and increase the responsibility of researchers.
During the same period, the success of deep convolutional neural networks (CNN) in image classification catch researchers’ attention. Therefore, some studies apply CNN to detect the malposition, trying to simplify the feature extraction process. Lakhani [7] used a CNN-based classifier to identify the presence or absence and the low or normal positioning of the ETT. Frid-Adar et al. [8] used the physical properties of ETTs and a public dataset of chest radiographs to synthesize an ETT over real X-ray images. After that, they used a U-Net shaped CNN to detect and segment ETTs. Lakhani et al. [9] used Inception V3 [10] to classify CXRs into 12 categories with the distance between ETT tip and Carina, including bronchial insertion, the distance range from 0 cm to 10 cm (0.0–0.9 cm, 1.0–1.9 cm, etc.) and over 10 cm. Chen et al. [11] used Mask R-CNN [12] to predict an ETT and a tracheal bifurcation for each image at first, and then found the ETT tip and Carina by a feature extraction post-process. Their method achieves a state-of-the-art performance. Although these CNN-based methods eased the feature extraction process, the context and semantic feature representation ability of these methods are still not enough. Therefore, to further lessen the burden on researchers and doctors, we need to improve these exciting methods.
Considering the efficiency and the flexibility of a model, in this paper, an end-to-end architecture was proposed, which was composed of FCOS: Fully Convolutional One-Stage Object Detection [13] and an attention mechanism called Coarse-to-Fine Attention (CTFA). FCOS was a one-stage anchor-free object detection method based on four distance values between the predicted center point and four sides of a bounding box. Furthermore, CTFA was applied to the deepest layer of backbone in FCOS, which consisted of global-modelling attention and scale attention. CTFA captured the long-range relationships between features by global-modelling attention based on [14] and rescaled the feature values with local relationships by scale attention based on [15]. Moreover, a mask prediction branch was attached during training, keeping more useful information. After finding the candidates of ETT tips and Carinas, a post-process algorithm was utilized to decide the best feature point of the ETT tip and Carina. The experiments on the chest radiographs datasets provided by ICU in National Cheng Kung University Hospital and Tainan Hospital showed that our method outperformed the existing state-of-the-art models. In summary, the main contributions of this paper are summarized as follows:
  • A new spatial attention (SA) module was proposed, which noticed the feature distribution in spatial and channel domains simultaneously. Different from previous studies, the SA employed the weight of channel attention to refine the weight of spatial attention, which also solved the problem from [15] where it overlooked the synergies between the channel and spatial attention. Moreover, the performance of SA was better than other attention modules.
  • A new merge method of attention modules was proposed, which fused global modelling attention and scale attention. Instead of stacking the same type of attention together, this method enlarged the ability of feature extraction in a coarse-to-fine way and received better results.
  • The architecture proposed by this paper improved the outcomes of object detection tasks by adding a segmentation branch. With the segmentation branch, the architecture kept more useful information during the feature extraction.
  • The experimental result demonstrated that the proposed architecture in this paper achieves outstanding performance on the chest radiograph datasets provided by the ICU in National Cheng Kung University Hospital and Tainan Hospital.
The rest of this paper is organized as follows. Section 2 introduces previous research about object detection and attention module. Section 3 explains the proposed method for detecting endotracheal intubation malposition. Section 4 shows and discusses the experimental results. Section 5 summarizes this paper.

2. Related Works

2.1. Object Detection

Object detection is a subfield of computer vision. The purpose of object detection is to localize and recognize objects with bounding boxes; this is different from classification, which is aimed at classifying a view into a class. Generally, it can be separated into one-stage detectors and two-stage detectors, based on whether the region proposal is adopted or not. Moreover, object detection can also be separated into anchor-based detectors and anchor-free detectors based on whether predefined anchors are employed or not. This paper used a one-stage anchor-free architecture named FCOS [13] as the base.

2.1.1. Anchor-Based Approach

Anchor-based detectors set anchor boxes in each position on the feature map and predict the probability of having objects in each anchor box. This approach can be separated into two classes by whether it uses region proposals or not, i.e., two-stage detector and one-stage detector. Two-stage detectors generate region proposals and then classify and refine them. For instance, Ren et al. [16] improved Fast R-CNN [17] by proposing a region proposal network (RPN), which generated high-quality region proposals with more efficiency than the Selective Search applied in Fast R-CNN. After finding the proposals, they used the region of interest (ROI) pooling to produce feature maps with the same resolution for further refinement. Cai et al. [18] noticed the impact of intersection over union (IoU) threshold when training object detector. Therefore, they proposed Cascade R-CNN, which consists of a sequence of detectors trained with an increasing IoU threshold. He et al. proposed Mask R-CNN [12] which is based on Faster R-CNN [16] and Feature Pyramid Network (FPN) [19], trying to perform object detection and segmentation at the same time.
One-stage detectors classify and regress the location directly. These detectors are more efficient and lightweight but have lower performance compared to two-stage detectors. For instance, SSD [20] detects objects of different scales by different layers in VGG-16 [21], i.e., small objects are detected on high-resolution feature maps, and large objects are detected on low-resolution feature maps. Liu et al. [22] proposed RFBNet to strengthen the receptive fields (RFs) based on SSD and the structure of the RFs in the human visual system. The RF Block (RFB) in RFBNet is similar to a mix of the inception block [10,23] and the atrous spatial pyramid pooling (ASPP) [24]. Lin et al. proposed Focal Loss to solve the extreme imbalance between positive and negative samples during training [25]. The Focal Loss which is similar to weighted cross entropy (CE) loss provided larger weights for positive samples and lower weights for negative samples. Based on this loss function, Lin et al. also proposed a one-stage detector named RetinaNet which used ResNet [26] and FPN [19] as a backbone to demonstrate the capability of Focal Loss.

2.1.2. Anchor-Free Approach

Anchor-free detectors do not need to design anchor boxes, avoiding the complicated computations related to anchor boxes. FCOS [13,27] transformed object detection tasks into per-pixel prediction tasks and used multi-level prediction to improve the recall. Based on their approach, they regressed the distance between the location and the four sides of the bounding box. Furthermore, they also proposed the center-ness branch to suppress the low-quality detected bounding boxes. FoveaBox [28] is similar to FCOS. FoveaBox also applied FPN to solve the intractable ambiguity which is caused by the overlaps in ground truth boxes. Additionally, they also regressed the offset between the location and the four sides of the bounding box. However, the definition of the positive samples in FoveaBox is different from FCOS. They used a formula with a shrunk factor to form shrunk ground truth boxes. If a sample falls into the shrunk ground truth boxes, the sample is positive, otherwise, it is negative. Moreover, the formula for offset computation is also different from FCOS.
CornerNet [29] detected an object as a pair of the top-left corner and bottom-right corner of a bounding box with heatmaps and then grouped the pairs of corners belonging to the same instance with embedding vectors. Furthermore, to solve the problem that the corner outside the object cannot be localized based on local evidence, they proposed corner pooling which took in two feature maps to encode the locations of the corner. During training, they used unnormalized 2D Gaussian to ease the difficulty of detecting the exact position of a corner. Duan et al. noticed that the CornerNet had a high false-positive rate. They intuited that the higher IoU of bounding boxes, the higher probability of the center key points in the central region will be predicted as the same class [30]. Therefore, they improved CornerNet with two strategies named center pooling and cascade corner pooling and represented each object using a center point and a pair of corners. They defined a scale-aware central region for each bounding box, and the bounding box will be preserved if the center key point had the same class as the bounding box that was detected in its central region.

2.2. Attention Mechanism

Attention mechanisms that come from Natural Language Processing (NLP) make a huge impact on vision tasks. Generally, attention mechanisms are adopted to determine where to focus. They take feature maps as input and produce weights for each value on the feature maps. The weights could be formed by the relationships between two pixels, the relationships between channels, spatial relationships, or unmentioned relationships. Attention mechanisms can be separated into dot-product attention, channel attention, spatial attention, level attention, and some variants by the relationships they adopt. Moreover, attention mechanisms can also be separated into global modelling attention and scale attention by their purpose [31,32]. Therefore, this paper classifies the attention mechanisms from this point of view.

2.2.1. Global Modelling Attention

Global modelling attention is effective for modelling the long-range global contextual information. Wang et al. proposed a non-local module to capture long-range dependencies with deep neural networks in computer vision [33]. Fu et al. proposed a Dual Attention Network (DANet) to capture feature dependencies in the spatial and channel dimensions with two parallel attention modules [34]. One is a position attention module and the other is a channel attention module. Both of these modules are similar to the non-local module. Miao et al. proposed CCAGNet [14] that considered multi-region context information simultaneously by combining Cross-context Attention Mechanism (CCAM), Receptive Field Attention Mechanism (RFAM), and Semantic Fusion Attention Mechanism (SFAM). The CCAM modeled the long-range and adjacent relationships among the feature values by two parallel non-linear functions. Moreover, the RFAM was proposed to improve the RFB in RFBNet [22] by inserting channel attention and spatial attention implemented with CCAM. Finally, SFAM was proposed to improve traditional up-sampling by Sub-Pixel convolution up-sampling [35] followed by CCAM.

2.2.2. Scale Attention

Scale attention is adopted to determine where to focus and suppress uninformative features. This attention mechanism is widely used in computer vision. Hu et al. proposed Squeeze-and-Excitation Networks (SENet) [36] which consist of a squeezing process and an excitation process. The squeezing process used global average pooling to generate the relationships between channels. Then, in the excitation process, the relationships mentioned above pass through a bottleneck with two fully connected layers to fully capture channel-wise dependencies. Woo et al. proposed CBAM to capture channel-wise relationships and spatial-wise relationships in order [15]. Their channel attention module included two parallel SE blocks with max pooling and average pooling. Their spatial attention module also used max pooling and average pooling to grab the relationships in spatial domains. Gu et al. aggregated spatial attention, channel attention, and level attention together and proposed Comprehensive Attention Neural Networks (CA-Net) [37]. Their spatial attention included a non-local block and the two-pathway attention blocks which refined the deep feature maps with the shallow ones. Moreover, their level attention consisted of channel attention and refinement process, which integrated each scale output in the decoder.

3. Materials and Methods

3.1. Overview

The proposed overall architecture is shown in Figure 1. In general, the proposed method embedded CTFA and a mask branch into FCOS [13] and employed the ResNet50 [26] as a backbone. This architecture aimed to take chest radiographs as input and located the ETT tip and the Carina. After finding the candidates of ETT tips and Carinas, a post-process algorithm was applied to refine the best ETT tip and Carina. ResNet50 [26] without the last down-sampling operations was applied to extract semantic features and the feature maps from high resolution to low resolution are denoted as C 2 5 , respectively. Followed by C 5 , CTFA was applied to refine the feature representation and then passed into the neck. The neck was implemented by FPN [19] which up-sampled the low-resolution feature maps and merged them with the high-resolution feature maps by element-wise summation. The feature maps produced by FPN are defined as P 2 5 from high resolution to low resolution. Additionally, the P 5 would pass through a series of convolutions with strides of 2 and kernel sizes of 3 × 3 to form P 6 and P 7 . Afterwards, P 3 7 were fed into the FCOS detection head, and P 3 5 were also fed into a segmentation head with P 2 . After detecting where the ETT tip and Carina might be, the post-process algorithm would choose the best ETT tip and Carina feature point through Gaussian masks generated from the center of the predicted bounding boxes of ETT and tracheal bifurcation.

3.2. Coarse-to-Fine Attention (CTFA)

CTFA fused global-modelling attention (GA) and scale attention (SA) with a point-wise convolution as shown in Figure 2. Concretely, the feature map produced by GA would pass through a point-wise convolution to smooth the features. Then, these smoothed feature maps were fed into SA to rescale features. We found that the Cross-context Attention Mechanism (CCAM) in [14] can grab long-range relationships and its performance was good as shown in the Section 4.5.3 of the ablation study. Therefore, we adopted CCAM as our GA in this paper. Moreover, we noticed that the CBAM can grab local relationships. However, some shortages limit the performance of CBAM. Therefore, we addressed these shortages and propose a new SA based on the CBAM.

3.2.1. Global-Modelling Attention (GA)

The CCAM in [14] encoded the cross-region relationship and adjacent-region relationships by two branch. Then, it fused the relationships by concatenation and adopted a series of convolutions followed by a softmax activation function to grab the attention weight. However, we noticed that the performance of CCAM without employing a softmax activation function was better than the performance of CCAM with a softmax activation function in this paper as shown in the Section 4.5.1 of the ablation study. Therefore, GA did not adopt the softmax activation function and the following steps in the CCAM. The architecture of GA is shown in Figure 3 and it can be summarized as:
F c r = C o n v 1 × 1 ( D i C o n v 3 × 3 , 3 ( F i n ) ) , F a r = D i C o n v 3 × 3 , 3 ( C o n v 1 × 1 ( F i n ) ) ,
F G A = D i C o n v 3 × 3 , 2 ( C o n v 1 × 1 ( F c r F a r ) ) ,
where the R e L u and batch normalization are neglected, F i n R C × H × W denotes the input feature map, D i C o n v denotes the dilated convolution and the parameters are the kernel size and the dilation rate, respectively, C o n v denotes the traditional convolution and the parameter is the kernel size, F c r and F a r denote the feature maps generated from the cross-region branch and the adjacent-region branch, ‖ denotes the concatenate operation, F G A represents the output of the GA module.
Specifically, GA generated long-range relationships through two branches. The upper branch F c r captured long-range context information by a dilated convolution whose kernel size was 3 × 3 and the dilation rate was 3, and smooths the features by a point-wise convolution (Equation (1)). The lower branch F a r grabbed local context information by a point-wise convolution and expands the influence of the features by a dilated convolution whose kernel size was 3 × 3 and the dilation rate was 3 (Equation (1)). These two feature maps concatenated to form the feature maps F g a = F c r F a r R 2 C × H × W . Then, F g a passed through a point-wise convolution and a dilated convolution whose kernel size was 3 × 3 and the dilation rate was 2 to produce the final output of GA, F G A (Equation (2)). The point-wise convolution here was used to decrease the channel numbers and the dilated convolution was used to expand the influence of features again.

3.2.2. Scale Attention (SA)

There were certain problems with the use of CBAM [15]. One of these was that CBAM ignored the interaction between channel attention and spatial attention. Another was that the channel-wise pooling operations overlooked the feature distribution of the feature maps. Based on the above analysis, the SA was proposed to smooth out these defects at the same time. The construction of SA is shown in Figure 4 and its overall process can be summarized as:
F s = M c a v g ( F x ) M c m a x ( F x ) ,
F S E 1 = f c e x ( f c s q ( C o n v 3 × 3 ( M a a v g ( F s ) ) ) ) , F S E 2 = f c e x ( f c s q ( C o n v 3 × 3 ( M a m a x ( F s ) ) ) ) ,
F s a = C o n v 1 × 1 ( F s ( F S E 1 F S E 2 ) ) ,
F S A = F x ( σ ( F s a ) ) ,
where the R e L u and batch normalization are neglected, F x denotes the input feature map, M c a v g and M c m a x denote channel-wise average pooling and channel-wise max pooling, ‖ denotes the concatenate operation, F s denotes the feature map after channel-wise pooling, M a a v g and M a m a x denote adaptive average pooling and adaptive max pooling, C o n v denotes the convolutional layer and the subscript represents the kernel size, f c s q and f c e x are two fully connected layers, F S E 1 and F S E 2 denote the feature maps in two branches after two fully connected layers, ⊕ denotes the element-wise summation, F s a represents the spatial weight, σ represents the softmax function, F S A represents the output of the SA module.
Specifically, SA can be separated into four steps. First, the input feature map F x R C × H × W was fed into two channel-wise poolings and generated two feature maps F c a v g = M c a v g ( F x ) R 8 × H × W and F c m a x = M c m a x ( F x ) R 8 × H × W , respectively. Afterwards, F c a v g was concatenated with F c m a x to generate F s R 16 × H × W as the Equation (3). By preserving more pooled channels, SA kept more of the channel information of the feature maps. Second, F s passed into adaptive max pooling and adaptive average pooling simultaneously to create F a m a x = M a m a x ( F s ) R 16 × 3 × 3 and F a a v g = M a a v g ( F s ) R 16 × 3 × 3 , respectively. Then, F a m a x and F a a v g would pass through one shared-weight convolution layer and two shared-weight fully connected layers to generate F S E 1 and F S E 2 as the Equation (4). By retaining a higher resolution, SA saved more spatial information than the traditional pooling method. Third, SA directly fused F S E 1 and F S E 2 with a element-wise summation. Thereafter, the mixed feature was element-wise multiplied with F s followed by a point-wise convolution layer as Equation (5). This step integrated channel attention weight into the spatial attention process and reduced the channel of the merged features with a point-wise convolution layer. Finally, the spatial attention feature, F s a R 1 × H × W , passed through softmax to grab the spatial attention weight. Then, the output of SA, F S A R C × H × W , was generated from the element-wise multiplication with F x and the spatial attention weight as the Equation (6).

3.3. Mask Branch

We noticed that the Mask R-CNN with a special feature extraction method proposed by Chen et al. [11] refined the feature points of the ETT tip and the Carina with the mask of ETT and tracheal bifurcation and achieved state-of-the-art performance. Inspired by their feature extraction algorithm and considering the characteristics of FCOS [13], this paper adopted a mask branch into the neck, trying to refine the feature point of the ETT tip and Carina in the same way. The architecture of the mask branch is shown in Figure 1 and its operation can be summarized as:
F s e g = P 2 U p s a m e p l e 2 ( P 3 ) U p s a m e p l e 4 ( P 4 ) U p s a m e p l e 8 ( P 5 ) ,
F S E G = C o n v 3 × 3 ( U p s a m p l e 4 ( C o n v 3 × 3 4 ( F s e g ) ) ) ,
where the R e L u and batch normalization are neglected, U p s a m p l e i denotes enlarging the resolution i times with nearest interpolation, ⊕ denotes the element-wise summation, F s e g denotes the feature map after fusion, C o n v j i denotes stacking convolution with j kernel size i times, F S E G denotes the final output of the segmentation head.
Considering that the features of an object might be separated into the feature pyramid, and a large magnification may cause some problems, this paper only fused P 3 5 together (Equation (7)). Explicitly, P 3 5 would be enlarged to the same resolution with P 2 and then summarized together. Afterwards, the fused feature maps would feed into the segmentation head which includes a series of convolution and up-sampling to generate the segmentation mask (Equation (8)).

3.4. Post-Process Algorithm

The process of the post-process algorithm is shown in Figure 5. In general, the algorithm has three steps. First, it found the highest confidence bounding box (bbox) of the ETT/tracheal bifurcation. Second, it applied a Gaussian mask based on the center of the ETT/tracheal bifurcation bbox or the center of the image to determine the best ETT tip/Carina. Finally, if the confidence of the best ETT tip/Carina was less than a threshold, then the best ETT tip/Carina would be generated after comparing with the confidence of the ETT/tracheal bifurcation bbox.
Expressly, in the first step, the algorithm kept the first bbox of the ETT/tracheal bifurcation as the best one and enumerated all bboxes which were classified as ETT/tracheal bifurcation. If the confidence of a bbox was larger than the best one, then the best bbox would be replaced by the bbox. As for the second step, if the bbox of ETT/tracheal bifurcation existed, the algorithm would produce a Gaussian mask based on the center of the bbox. Otherwise, the Gaussian mask would be generated based on the center of an input image. Afterwards, the confidence of the bboxes classified as ETT tip/Carina would multiply the value which was grabbed from the same position as the center point of the bbox on the Gaussian mask. Then, the algorithm chose the best ETT tip bbox/Carina bbox as the first step based on this multiplied value. After checking whether the ETT tip bbox/Carina bbox existed or not, the algorithm adopted a threshold to filter out the ETT tip bboxes/Carina bboxes which did not need to be purified. Finally, the remaining bbox of ETT tip/Carina was compared with the corresponding bbox of ETT/tracheal bifurcation. If the confidence of the ETT tip bbox was lower than the ETT bbox, the ETT tip would be replaced by the center point of the lowest two points of the ETT bbox. Otherwise, the ETT tip was the center point of the ETT tip bbox. As for Carina, if the confidence of the Carina bbox was lower than the tracheal bifurcation bbox, the Carina would be replaced by the center point of the tracheal bifurcation bbox. Otherwise, the Carina was the center point of the Carina bbox.

3.5. Loss Function

The loss function was based on FCOS [13] appended with a segmentation loss. The formula of the loss function can be summarized as:
L t o t a l = L c l s + L r e g + L c e n t e r n e s s + L s e g ,
where the L c l s is the classification loss implemented with focal loss [25], L r e g is the regression loss implemented with GIoU loss [38], L c e n t e r n e s s is the center-ness loss implemented with binary cross entropy loss [13], L s e g is the segmentation loss implemented with cross entropy loss plus dice loss as in [39].
L c l s = F L ( p t ) = a t ( 1 p t ) γ l o g ( p t ) , p t = p , i f y = 1 1 p , o t h e r w i s e , a t = a , i f y = 1 1 a , o t h e r w i s e ,
The formula of focal loss is shown in Equation (10), where y { ± 1 } denotes the ground-truth class, p [ 0 , 1 ] denotes the predicted probability for the class with label y = 1 , a [ 0 , 1 ] denotes a weighting factor for class with label y = 1 . Additionally, the γ is set to 2 as [13] did. This loss function can mitigate the class imbalance problem by down-weight easy samples.
L r e g = L G I o U = 1 G I o U G I o U = I o U C   \   ( A B ) C I o U = A B A B
The formula of GIoU loss is shown in Equation (11), where A , B denotes two arbitrary shapes, C denotes the smallest convex shapes enclosing both A and B. GIoU solved the problem that if A B = 0 , IoU cannot reflect the relationship between A and B.
L c e n t e r n e s s = L B C E ( c e n t e r n e s s , c e n t e r n e s s * ) c e n t e r n e s s * = m i n ( l * , r * ) m a x ( l * , r * ) × m i n ( t * , b * ) m a x ( t * , b * )
The L B C E is the binary cross-entropy loss, the c e n t e r n s s * [ 0 , 1 ] denotes the ground-truth value of centerness, l * , r * , b * , t * denotes the distance from a feature point to the left, right, top, and bottom of the relative ground-truth bbox. FCOS down-weighted the confidence of low-quality predicted bboxes by adding this loss function. Concretely, during the inference phase, the classification score will multiply with the corresponding predicted centerness score. Thus, the low-quality bboxes could be filtered out by non-maximum suppression (NMS).
L s e g = L D i c e + λ L C E L D i c e = 1 D i c e D i c e = 2 A B A + B L C E ( p t ) = l o g ( p t ) p t = p , i f y = 1 1 p , o t h e r w i s e ,
The λ is a trade-off between Dice loss and cross-entropy loss, and is set to 1 in all our experiments, the L D i c e is the dice loss, A , B denotes two arbitrary shapes as in Equation (11), L C E denotes the cross-entropy loss, and p [ 0 , 1 ] denotes the predicted probability for the class with label y = 1 as the Equation (10).

4. Results and Discussion

4.1. Dataset and Evaluation Metrics

This paper was approved by the institutional review board (IRB) of the National Cheng Kung University (NCKU) Hospital (IRB number: A-ER-108-305). The chest radiograph dataset provided by NCKU Hospital includes 1,870 portable chest radiographs of intubated ICU patients in DICOM format and the ground truth (GT) annotations were labeled by two board-certified intensivists. The GT ETT was labelled by four points ( P 1 4 ), and the GT tracheal bifurcation was labelled by nine points ( P 5 13 ), as shown in Figure 6a. The purpose of this paper was to detect the malposition by locating the ETT tip and the Carina. Therefore, this paper adopted two boxes with a size of 300 × 300 to label the feature point of the ETT tip which was the middle point of P 2 and P 3 , and the feature point of the Carina which was P 9 (the feature point of ETT tip and Carina are at the center of the boxes). Furthermore, the 13 points were corrected to be sequential for generating the GT mask. In summarize, the GT became Figure 6b. In Figure 6b, the green nodes denote the original points labeled by the intensivists, the blue nodes denote the ETT tip and Carina, the green boxes denote the GT bboxes of ETT and tracheal bifurcation, the blue boxes denote the GT bboxes of ETT tip and Carina, and the red polygons denote the GT mask of ETT and tracheal bifurcation. Finally, this paper used extra 150 chest radiographs to validate the proposed approach.
This paper applied object error, distance error, recall, precision, and accuracy to evaluate the performance. Object error and distance error were implemented with the Euclidean Distance. The object error of the ETT tip was defined as the distance between the center of the predicted ETT tip bbox and the center of the GT ETT tip bbox. The object error of the Carina was defined as the distance between the center of the predicted Carina bbox and the center of the GT Carina bbox. Moreover, the distance between the ETT tip and the Carina was named the ETT–Carina distance. The distance error was defined as the absolute difference between the GT ETT–Carina distance and the predicted ETT–Carina distance. As for recall and precision, this paper defined successful detection (true positive) as when the object error was no more than 10 mm [40], a false positive indicated that a region was a specific object but the region did not include the object, and a false negative was that the model did not indicate the GT object. Therefore, the recall and precision are defined as:
Recall = T P ( T P + F N ) ,
Precision = T P ( T P + F P ) ,
Accuracy = ( T P + T N ) ( T P + T N + F P + F N )
Moreover, by the Goodman’s criteria [41,42], the ideal distance of the ETT–Carina was in the range of [30, 70] mm. However, another paper [43] recommended that the average correct tube insertion depth was 21 cm in women and 23 cm in men. In this situation, the distance of the ETT–Carina was less than 25 mm. Taking this into consideration, this paper defined that the endotracheal intubation was suitable if the distance of the ETT–Carina is in the range of [20, 70] mm, as in [9].

4.2. Implementation Details

The architecture was implemented based on Pytorch. During training, Adam [44] with the initial learning rate of 10 4 was adopted. The batch size was set to 2 and the total epochs were set to 120. Moreover, we employed a warm-up policy for the first five epochs and utilized the cosine learning rate decay policy for the rest of the epochs. Additionally, the input images were resized to have their shorter side be 800 and their longer side less or equal to 1333. In addition, random color jitter, random rotation angle within the range of [ 10 , 10 ] degrees, and random cropping were used during training. Furthermore, the dataset provided by NCKU Hospital was equally divided into five folders to execute 5-fold cross-validation. Finally, all of the experiments were executed on an Nvidia GeForce RTX 2080 Ti GPU.

4.3. Results

As shown in Table 1, by the definition of “suitable position”, this paper achieved 88.82% accuracy on the NCKU dataset and 90.67% on the external validation based on the annotations from board-certified intensivists. Moreover, the mean of the ETT–Carina distance error was 5.333 mm and the standard deviation of the ETT–Carina distance error was 6.240 mm. In [9], the mean of the ETT–Carina distance error was 6.9 mm, and the standard deviation of the ETT–Carina distance error was 7.0 mm. Table 2 shows that the distance error of 85.83% images were less than 10 mm on the NCKU dataset and 84.00% images were less than 10 mm on the external validation. Table 3 and Table 4 show the confusion matrix of diagnosis in NCKU dataset and external validation.
For the ETT tip and the Carina, Table 5 shows that the recall and precision of the ETT tip were 90.96% and 91.60%, separately. The recall and precision of the Carina were 93.90% and 94.10%, separately. Table 6 shows the mean and standard deviation of object error on the ETT tip were 4.304 mm and 5.526 mm, respectively. The mean and standard deviation of object error on the Carina were 4.118 mm and 3.655 mm, respectively. Table 7 shows that in the object error of the ETT tip, there were 90.96% images not longer than 10 mm. Table 8 shows that in the object error of the Carina, the differences of the 93.90% images were not longer than 10 mm. Although most of the performances on the external validation were dropped, the performance on ETT–Carina still showed the robustness of our architecture in detecting the malposition of the ETT.

4.4. Compare with the SOTA

This section compares the proposed method with the state-of-the-art (SOTA) method proposed by Chen et al. [11]. Both of these methods executed the experiment on the same datasets. Table 9 presents the accuracy of detecting malposition and the distance error of the ETT–Carina, and Table 10 presents the error distribution of the ETT-Carina distance. Furthermore, the percentage in the brackets denotes the improvement or degeneration rate of the performances. Based on the tables, we observed that most of the performances of the proposed method were better than SOTA. It is worth mentioning that the proposed method had higher accuracy on both data sets, which indicated that the proposed method could be more accurate in detecting the malposition of the ETT.
Table 11 and Table 12 demonstrate the performance of the object error and the error distribution on the ETT tip. Table 13 and Table 14 demonstrate the performance of the object error and the error distribution on the Carina. In general, our method was bad at detecting the ETT tip but good at detecting the Carina compared with SOTA. However, the improvement on the Carina could cover the degeneration of the ETT tip. Therefore, our method could detect the malposition more effectively.

4.5. Ablation Study

The ablation studies were performed on the NCKU dataset with folder 5 as the testing data to verify the effectiveness of our method. This section only focuses on the accuracy, the mean distance error, and the standard deviation.

4.5.1. Structure of GA

Table 15 demonstrates the performance of whether the GA adopted a softmax function. This results presented that the GA without a softmax activation function achieved a better outcome. Therefore, CTFA employed GA without a softmax function.

4.5.2. Structure of SA

In Table 16, the c and k in the brackets behind the SA denote the pooled channel number and the followed kernel size. The SE in the brackets behind the SA shows whether the SA adopts a SE block as the channel attention. From the six rows above, we observed that when the output channel number of the channel-wise max pooling and average pooling was 8, and the followed kernel size was 1, the performance was the best. Therefore, CTFA employed SA with 8 pooled channel numbers followed by a point-wise convolution. Furthermore, comparing the fourth row with the last row, we noticed that employing the SE block in SA would increase the performance.

4.5.3. Compare with Attention Modules

Table 17 and Table 18 display the results of adopting existing attention modules in FCOS [13]. This paper calculated the number of parameters and giga floating point operations (GFLOPs) for each attention module with an image size of 224 × 224 . Furthermore, the percentage in the brackets at the “Parameter (M)” and “GFLOPs” column is the degree of increment compared with FCOS. By Table 17, we discovered that the SE block [36], Nonlocal block [33], Nonlocal with CSP block [45] and the CBAM block [15] brought no benefit to the accuracy of ETT-Carina, but the CCAM block [14] and the SA block proposed by this paper could improve the accuracy. Moreover, the SA block was more lightweight than the CCAM block and achieved higher accuracy, as shown in Table 18. These results illustrated the efficiency and effectiveness of the SA block.

4.5.4. Fusion Method

This part explores the fusion method. The second row in Table 19 denotes that the C 5 with reduced channel directly passes through SA, followed by GA. The third row denotes that the C 5 with reduced channel goes through SA and GA in a parallel fashion, and then concatenates the output feature maps together followed by a point-wise convolution, batch normalization, and relu to integrate the feature map. The last row denotes that the C 5 with reduced channel directly passes through GA, followed by SA. As shown in Table 19, the last row demonstrated that the long-range relationship rescaled by the local relationship could improve the performance. Moreover, this was the best connection method in our ablation studies; thus, the CTFA adopted this method to grab better features in this paper.

4.5.5. Fusing Global Modelling Attention and Scale Attention

Table 20 presents the result of fusing a global modelling attention and a scale attention. Here, * 2 denotes stacking two of the same attention module together. We observed that when directly stacking two of the same attention module together, the accuracy would drop compared with the performance in Table 17. Specifically, when stacking two CSPnonlocal blocks together, the accuracy would degrade from 86.10 % to 85.56 % , and when stacking two CCAM blocks together, the accuracy would degrade from 86.90 % to 86.10 % . Furthermore, stacking two SA blocks also grabbed a worse result. However, if we stacked global modelling attention with scale attention, the performance would increase, as shown in the last two rows. This result showed that the global modelling attention could grab long-range relationships and the scale attention could refine the attention map effectively. Therefore, stacking these two kinds of attention modules together could achieve higher performance.

4.5.6. Mask Branch

In Table 21, the second row denotes that CTFA is adopted in FCOS, and the “Seg” denotes that the CTFA and the mask branch are adopted in FCOS at the same time. Moreover, the text in the brackets behind “Seg” in the “Method” column denotes which mask annotation our approach employed at the mask branch, and the “Fusion” in the “Method” column denotes that we directly fuse the feature maps from P 2 to P 5 . Comparing the results of the second row and third row, we noticed that the mask branch did not bring advantages to the detection model. However, if we only used the annotation of ETT and tracheal bifurcation, the performance would be improved. This phenomenon might be caused by the occlusion of the annotations. Concretely, the annotations of the ETT and the tracheal bifurcation were always occluded by the annotation of the ETT tip and the Carina. Therefore, the mask branch might focus on the ETT tip and the Carina which was similar to the original detection model (FCOS + CTFA). However, if we only adopted the annotation of the ETT and the tracheal bifurcation, the detection model might pay more attention to the ETT and the tracheal bifurcation. Thus, the detection model might grab some useful context information for detecting the ETT tip and the Carina. Comparing the last two rows, we found that the fusion policy further improves the performance. Considering the accuracy and the object error, the method of the last row achieved the best result on the internal dataset with folder 5 as the testing data. Moreover, it was the final architecture employed by this paper for detecting the malposition of the ETT.

4.5.7. The Post-Process Algorithm

Figure 7 and Figure 8 demonstrate the effect of post-processing. The red bboxes and points in these figures are the GT ETT/bifurcation bboxes and the position of GT ETT tip/Carina, respectively. The green polygon is the GT mask of the ETT and the bifurcation. The blue bbox and point are the predicted ETT bbox and ETT tip, respectively. The yellow bbox and point are the predicted bifurcation and Carina, respectively. Specifically, without the post-process, the model might leave more than one predicted ETT tip/Carina, such as where the red arrow points in Figure 7a. However, with the post-process, the extra points would be removed as shown in Figure 7b. Besides, with the refinement process in the post-process, the feature point of ETT tip/Carina could be further refined as shown in Figure 8. Concretely, the object error of Carina was corrected from 8.469 mm to 1.319 mm.
We observed that the segmentation results of the ETT tip were always near the GT ETT tip. However, although the mask branch could improve the detection result, the segmentation result was not good enough to replace the detection result. As shown in Table 22, if we corrected the detection result of the ETT tip by the segmentation result of the ETT tip when the distance between them was longer than 100 pixels, the performance of ETT would decrease and thus the performance of the ETT-Carina would also decline. Therefore, in the post-process, we only used the bbox result of the ETT and the tracheal bifurcation to refine the position of the ETT tip and the Carina when the ETT tip and the Carina were not good enough.

4.5.8. Visualization

In this part, the red bboxes and points denote the GT ETT/bifurcation bboxes and the position of GT ETT tip/Carina, respectively. The blue bbox and point are the predicted ETT bbox and ETT tip, respectively. The green bbox and point are the predicted bifurcation and Carina, respectively. The light blue point is the position of the ETT tip produced by the mask branch. In the Table 23, the first row demonstrates the good results, the second row shows the medium results, and the third row presents the bad results. We noticed that if an image had a clear location of the ETT tip and the Carina, the performance would be better. Moreover, the life-supporting device might blur the location of the ETT tip and the shadow of the heart might occlude the position of the Carina as shown in the medium case. Apart from the problems mentioned above, the angle of the CXR might also degrade the performance of the proposed method, as shown in the worse results when it is applied.

5. Conclusions

This paper proposed an end-to-end architecture to improve the FCOS [13] for detecting the position of the ETT tip and the Carina. First, a Coarse-to-Fine (CTFA) attention module was designed to capture long-range relationships by global-modelling attention (GA) and rescale the feature value with local relationships grabbed by scale attention (SA). Then, a mask branch was adopted to enhance the feature representation of the “backbone” and “neck”. After that, a post-process algorithm was employed to ensure that only one detection result for each class would be left in an image and further refine the feature point of ETT tip and Carina. Experiments on the chest radiograph datasets provided by National Cheng Kung University Hospital demonstrated that the proposed architecture provides an effective solution to improve the performance of the FCOS. The mean malposition accuracy achieved 88.82% and the ETT-Carina distance error was less than 5.333 ± 6.240 mm on the internal dataset. Furthermore, our method indicates the position of the ETT, ETT tip, tracheal bifurcation, and the Carina on an image. Therefore, it is more reasonable for detecting the malposition of the ETT compared with classification-based methods. This end-to-end deep learning-based method could help the intensivists, helping them to keep an eye on the position of the ETT in ICU patients.
Although our proposed methods have exciting results, two shortages still exist that might need to be overcome. First, the annotation of the ETT tip is labelled on the slash at the bottom of the ETT. However, the feature point on the edge is not a good feature point compared with the corner point. Therefore, an accurate ETT tip annotation might further improve the performance. Second, the fusion method adopted for GA and SA is simple, as is the additional mask branch, so further investigations of the interaction between them might improve the performance. In the future, we will focus on these shortages, trying to provide a more reliable model for intensivists.

Author Contributions

Conceptualization, C.-Y.C. and L.-K.M.; methodology, C.-Y.C. and L.-K.M.; software, L.-K.M.; validation, C.-Y.C. and L.-K.M.; formal analysis, L.-K.M.; investigation, C.-Y.C., M.-H.H., C.-H.L. and L.-K.M.; resources, C.-Y.C., M.-H.H., C.-H.L. and Y.-N.S.; data curation, C.-Y.C., M.-H.H., C.-H.L. and L.-K.M.; writing—original draft preparation, L.-K.M.; writing—review and editing, C.-Y.C., M.-H.H., C.-H.L., Y.-N.S. and L.-K.M.; visualization, L.-K.M.; supervision, C.-Y.C., Y.-N.S. All authors have read and agreed to the published version of the manuscript.

Funding

Chen received support for article research from the Higher Education Sprout Project, Ministry of Education to the Headquarters of University Advancement at National Cheng Kung University (NCKU) and the Ministry of Science and Technology, Executive Yuan, Taiwan (MOST 111-2221-E-006-125-MY2). Lai received support for article research from the Ministry of Science and Technology, Executive Yuan, Taiwan (MOST 109-2634-F-006-023) and from National Cheng Kung University Hospital, Tainan, Taiwan (NCKUH-10901003).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board of National Cheng Kung University Hospital (A-ER-108-305; 26 December 2019).

Informed Consent Statement

Consent from each patient was waived as this clinical study is a retrospective review of medical records produced during the patient’s treatment process.

Data Availability Statement

The data that support the findings of this study are not publicly available due to the information that could compromise the privacy of research participants.

Acknowledgments

The research was supported by the Higher Education Sprout Project, Ministry of Education to the Headquarters of University Advancement at National Cheng Kung University (NCKU) and also by the Ministry of Science and Technology, Executive Yuan, Taiwan (MOST 111-2221-E-006-125-MY2 and MOST 109-2634-F-006-023), by National Cheng Kung University Hospital, Tainan, Taiwan (NCKUH-10901003).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Hunter, T.B.; Taljanovic, M.S.; Tsau, P.H.; Berger, W.G.; Standen, J.R. Medical devices of the chest. Radiographics 2004, 24, 1725–1746. [Google Scholar] [CrossRef] [PubMed]
  2. Varshney, M.; Sharma, K.; Kumar, R.; Varshney, P.G. Appropriate depth of placement of oral endotracheal tube and its possible determinants in Indian adult patients. Indian J. Anaesth. 2011, 55, 488. [Google Scholar] [CrossRef] [PubMed]
  3. Bentz, M.R.; Primack, S.L. Intensive care unit imaging. Clin. Chest Med. 2015, 36, 219–234. [Google Scholar] [CrossRef] [PubMed]
  4. Chen, S.; Zhang, M.; Yao, L.; Xu, W. Endotracheal tubes positioning detection in adult portable chest radiography for intensive care unit. Int. J. Comput. Assist. Radiol. Surg. 2016, 11, 2049–2057. [Google Scholar] [CrossRef]
  5. Kao, E.F.; Jaw, T.S.; Li, C.W.; Chou, M.C.; Liu, G.C. Automated detection of endotracheal tubes in paediatric chest radiographs. Comput. Methods Programs Biomed. 2015, 118, 1–10. [Google Scholar] [CrossRef]
  6. Ramakrishna, B.; Brown, M.; Goldin, J.; Cagnon, C.; Enzmann, D. An improved automatic computer aided tube detection and labeling system on chest radiographs. In Proceedings of the Medical Imaging 2012: Computer-Aided Diagnosis, San Diego, CA, USA, 4–9 February 2012; Volume 8315, p. 83150R. [Google Scholar]
  7. Lakhani, P. Deep convolutional neural networks for endotracheal tube position and X-ray image classification: Challenges and opportunities. J. Digit. Imaging 2017, 30, 460–468. [Google Scholar] [CrossRef]
  8. Frid-Adar, M.; Amer, R.; Greenspan, H. Endotracheal tube detection and segmentation in chest radiographs using synthetic data. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 784–792. [Google Scholar]
  9. Lakhani, P.; Flanders, A.; Gorniak, R. Endotracheal tube position assessment on chest radiographs using deep learning. Radiol. Artif. Intell. 2020, 3, e200026. [Google Scholar] [CrossRef]
  10. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  11. Chen, C.Y.; Huang, M.H.; Sun, Y.N.; Lai, C.H. Development of Automatic Endotracheal Tube and Carina Detection on Portable Supine Chest Radiographs using Artificial Intelligence. arXiv 2022, arXiv:2206.03017. [Google Scholar] [CrossRef]
  12. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  13. Tian, Z.; Shen, C.; Chen, H.; He, T. Fcos: Fully convolutional one-stage object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 9627–9636. [Google Scholar]
  14. Miao, S.; Du, S.; Feng, R.; Zhang, Y.; Li, H.; Liu, T.; Zheng, L.; Fan, W. Balanced single-shot object detection using cross-context attention-guided network. Pattern Recognit. 2022, 122, 108258. [Google Scholar] [CrossRef]
  15. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  16. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef]
  17. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  18. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  19. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  20. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  21. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  22. Liu, S.; Huang, D.; Wang, Y. Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 385–400. [Google Scholar]
  23. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI’17, San Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar] [CrossRef]
  24. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  25. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  27. Tian, Z.; Shen, C.; Chen, H.; He, T. FCOS: A Simple and Strong Anchor-Free Object Detector. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 1922–1933. [Google Scholar] [CrossRef] [PubMed]
  28. Kong, T.; Sun, F.; Liu, H.; Jiang, Y.; Li, L.; Shi, J. Foveabox: Beyound anchor-based object detection. IEEE Trans. Image Process. 2020, 29, 7389–7398. [Google Scholar] [CrossRef]
  29. Law, H.; Deng, J. Cornernet: Detecting objects as paired keypoints. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 734–750. [Google Scholar]
  30. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  31. Li, R.; Zheng, S.; Zhang, C.; Duan, C.; Wang, L.; Atkinson, P.M. ABCNet: Attentive bilateral contextual network for efficient semantic segmentation of Fine-Resolution remotely sensed imagery. ISPRS J. Photogramm. Remote Sens. 2021, 181, 84–98. [Google Scholar] [CrossRef]
  32. Shen, Z.; Zhang, M.; Zhao, H.; Yi, S.; Li, H. Efficient attention: Attention with linear complexities. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 3531–3539. [Google Scholar]
  33. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  34. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  35. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  36. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  37. Gu, R.; Wang, G.; Song, T.; Huang, R.; Aertsen, M.; Deprest, J.; Ourselin, S.; Vercauteren, T.; Zhang, S. CA-Net: Comprehensive attention convolutional neural networks for explainable medical image segmentation. IEEE Trans. Med. Imaging 2020, 40, 699–711. [Google Scholar] [CrossRef]
  38. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  39. Feng, S.; Zhao, H.; Shi, F.; Cheng, X.; Wang, M.; Ma, Y.; Xiang, D.; Zhu, W.; Chen, X. CPFNet: Context pyramid fusion network for medical image segmentation. IEEE Trans. Med Imaging 2020, 39, 3008–3018. [Google Scholar] [CrossRef]
  40. Huo, Z.; Mao, H.; Zhang, J.; Sykes, A.M.; Munn, S.; Wandtke, J. Computer-aided detection of malpositioned endotracheal tubes in portable chest radiographs. In Proceedings of the Medical Imaging 2014: Computer-Aided Diagnosis, San Diego, CA, USA, 15–20 February 2014; Volume 9035, pp. 151–156. [Google Scholar]
  41. Ong, K.C.; A’Court, G.D.; Eng, P.; Ong, Y.Y. Ideal endotracheal tube placement by referencing measurements on the tube. Ann. Acad. Med. Singap. 1996, 25, 550–552. [Google Scholar]
  42. Goodman, L.R.; Conrardy, P.A.; Laing, F.; Singer, M.M. Radiographic evaluation of endotracheal tube position. Am. J. Roentgenol. 1976, 127, 433–434. [Google Scholar] [CrossRef]
  43. Sitzwohl, C.; Langheinrich, A.; Schober, A.; Krafft, P.; Sessler, D.I.; Herkner, H.; Gonano, C.; Weinstabl, C.; Kettner, S.C. Endobronchial intubation detected by insertion depth of endotracheal tube, bilateral auscultation, or observation of chest movements: Randomised trial. BMJ 2010, 341, c5943. [Google Scholar] [CrossRef]
  44. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  45. Wang, C.Y.; Liao, H.Y.M.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A new backbone that can enhance learning capability of CNN. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 390–391. [Google Scholar]
Figure 1. An overview of the proposed architecture for detecting the malposition. It consists of a ResNet-based backbone, Coarse-to-Fine Attention (CTFA), FPN-based neck, FCOS-based detection head, and segmentation head. The legends below demonstrate the operations above.
Figure 1. An overview of the proposed architecture for detecting the malposition. It consists of a ResNet-based backbone, Coarse-to-Fine Attention (CTFA), FPN-based neck, FCOS-based detection head, and segmentation head. The legends below demonstrate the operations above.
Diagnostics 12 01913 g001
Figure 2. An illustration of coarse-to-fine attention (CTFA). CTFA consisted of a global-modelling attention (GA) and a scale attention (SA). GA was aimed at capturing long-range relationships and SA was aimed at reweighting with local relationships.
Figure 2. An illustration of coarse-to-fine attention (CTFA). CTFA consisted of a global-modelling attention (GA) and a scale attention (SA). GA was aimed at capturing long-range relationships and SA was aimed at reweighting with local relationships.
Diagnostics 12 01913 g002
Figure 3. An illustration of global-modelling attention (GA). GA generated long-range relationships through two branches. The upper branch was aimed at capturing long-range context information and the lower branch was aimed at grabbing local context information. Then, this information is integrated by a series of operations.
Figure 3. An illustration of global-modelling attention (GA). GA generated long-range relationships through two branches. The upper branch was aimed at capturing long-range context information and the lower branch was aimed at grabbing local context information. Then, this information is integrated by a series of operations.
Diagnostics 12 01913 g003
Figure 4. An illustration of scale attention (SA). SA addressed the defects of convolutional block attention module (CBAM) by adaptive channel pooling and squeeze-and-excitation (SE) block.
Figure 4. An illustration of scale attention (SA). SA addressed the defects of convolutional block attention module (CBAM) by adaptive channel pooling and squeeze-and-excitation (SE) block.
Diagnostics 12 01913 g004
Figure 5. An illustration of post-process.
Figure 5. An illustration of post-process.
Diagnostics 12 01913 g005
Figure 6. Ground Truth. (a) Original ground truth. (b) Pre-processed ground truth.
Figure 6. Ground Truth. (a) Original ground truth. (b) Pre-processed ground truth.
Diagnostics 12 01913 g006
Figure 7. Ensuring at most one ETT tip/Carina left. (a) Without post-process. (b) With post-process.
Figure 7. Ensuring at most one ETT tip/Carina left. (a) Without post-process. (b) With post-process.
Diagnostics 12 01913 g007
Figure 8. Refining the feature point of ETT tip/Cairna by the bbox of ETT/Bifurcation. (a) Without post-process. (b) With post-process.
Figure 8. Refining the feature point of ETT tip/Cairna by the bbox of ETT/Bifurcation. (a) Without post-process. (b) With post-process.
Diagnostics 12 01913 g008
Table 1. The performance in ETT–Carina distance error.
Table 1. The performance in ETT–Carina distance error.
Test FolderAcc. (%)Mean (mm)Std. (mm)
Folder 190.375.1305.609
Folder 287.705.9698.325
Folder 388.245.2565.491
Folder 486.635.4376.663
Folder 591.184.8745.111
Average88.825.3336.240
External val.90.675.0155.147
Table 2. The distance error distribution in ETT–Carina.
Table 2. The distance error distribution in ETT–Carina.
Test Folder≤5 mm (%)≤10 mm (%)≤15 mm (%)≤20 mm
Folder 162.5785.2993.5896.79
Folder 263.1084.2292.2595.72
Folder 363.9083.9692.2595.45
Folder 463.9087.1792.7897.06
Folder 564.4488.5093.8597.06
Average63.5885.8392.9496.42
External val.66.0084.0092.6797.33
Table 3. The confusion matrix of diagnosis.
Table 3. The confusion matrix of diagnosis.
GTSuitableUnsuitable
Predict
Suitable1350126
Unsuitable66311
Undetection125
Table 4. The confusion matrix of diagnosis (external val.).
Table 4. The confusion matrix of diagnosis (external val.).
GTSuitableUnsuitable
Predict
Suitable1108
Unsuitable526
Undetection10
Table 5. The performance in recall and precision.
Table 5. The performance in recall and precision.
Recall and PrecisionETT TipCarina
Test FolderRecall (%)Precision (%)Recall (%)Precision (%)
Folder 190.6491.3794.6594.91
Folder 289.3089.5493.5893.58
Folder 390.9192.1492.2592.49
Folder 491.1891.4294.9295.17
Folder 592.7893.5394.1294.37
Average90.9691.6093.9094.10
External val.92.6793.2988.0088.59
Table 6. The performance in object error.
Table 6. The performance in object error.
Object ErrorETT TipCarina
Test FolderMean (mm)Std. (mm)Mean (mm)Std. (mm)
Folder 14.4155.2813.9523.345
Folder 24.8587.8694.2363.663
Folder 33.9744.4054.3223.947
Folder 44.5846.2733.8953.527
Folder 53.6903.8004.1853.793
Average4.3045.5264.1183.655
External val.3.7334.6134.6884.043
Table 7. The object error distribution in ETT tip.
Table 7. The object error distribution in ETT tip.
Test Folder≤5 mm (%)≤10 mm (%)≤15 mm (%)≤20 mm (%)
Folder 175.9490.6494.3997.06
Folder 275.4089.3094.6596.79
Folder 378.6190.9195.1997.06
Folder 473.2691.1894.9297.86
Folder 581.0292.7897.0697.59
Average76.8590.9695.2497.27
External val.83.3392.6794.6796.67
Table 8. The object error distribution in Carina.
Table 8. The object error distribution in Carina.
Test Folder≤5 mm (%)≤10 mm (%)≤15 mm (%)≤20 mm (%)
Folder 174.6094.6598.1399.20
Folder 274.0693.5897.5999.20
Folder 373.5392.2596.5298.40
Folder 478.3494.9297.8698.93
Folder 574.0694.1298.1398.40
Average74.9293.9097.6598.83
External val.68.6788.0096.6798.00
Table 9. The comparison results of accuracy and ETT–Carina distance error.
Table 9. The comparison results of accuracy and ETT–Carina distance error.
MethodMalposition Accuracy (%)ETT-Carina Distance Error
Mean (mm)Std. (mm)
SOTA average [11]88.115.5436.310
Ours average88.82 (+0.81%)5.333 (−3.79%)6.240 (−1.11%)
SOTA external val. [11]87.335.6686.651
Ours external val.90.67 (+3.82%)5.015 (−11.52%)5.147 (−22.61%)
Table 10. The comparison results of error distribution on the ETT–Carina distance.
Table 10. The comparison results of error distribution on the ETT–Carina distance.
MethodETT-Carina Distance Error Distribution
≤5 mm (%)≤10 mm (%)≤15 mm (%)≤20 mm (%)
SOTA average [11]60.3784.2092.7895.39
Ours average63.58 (+5.32%)85.83 (+1.94%)92.94 (+0.17%)96.42 (+1.08%)
SOTA external val. [11]64.0082.0090.6794.67
Ours external val.66.00 (+3.13%)84.00 (+2.44%)92.67 (+2.21%)97.33 (+2.81%)
Table 11. The comparison results of recall, precision, and object error on the ETT tip.
Table 11. The comparison results of recall, precision, and object error on the ETT tip.
MethodETT Tip
Recall (%)Precision (%)Mean (mm)Std. (mm)
SOTA average [11]93.3193.494.1224.402
Ours average90.96 (−2.52%)91.60 (−2.02%)4.304 (+4.42%)5.526 (+25.53%)
SOTA external val. [11]90.2790.274.2865.943
Ours external val.92.67 (+2.66%)93.29 (+3.35%)3.733 (−12.90%)4.613 (−22.38%)
Table 12. The comparison results of error distribution on the ETT tip.
Table 12. The comparison results of error distribution on the ETT tip.
MethodETT Tip Object Error Distribution
≤5 mm (%)≤10 mm (%)≤15 mm (%)≤20 mm (%)
SOTA average [11]75.0893.3196.3698.21
Ours average76.85 (+2.36%)90.96 (−2.52%)95.24 (−1.16%)97.29 (−0.94%)
SOTA external val. [11]79.3390.2795.3396.97
Ours external val.83.33 (+5.04%)92.67 (+2.66%)94.67 (−0.69%)96.67 (−0.31%)
Table 13. The comparison results of recall, precision, and object error on the Carina.
Table 13. The comparison results of recall, precision, and object error on the Carina.
MethodCarina
Recall (%)Precision (%)Mean (mm)Std. (mm)
SOTA average [11]94.7095.234.7755.342
Ours average93.90 (−0.84%)94.10 (−1.19%)4.118 (−13.76%)3.655 (−31.58%)
SOTA external val. [11]91.6491.964.5674.513
Ours external val.88.00 (−3.97%)88.59 (−3.66%)4.688 (+2.65%)4.043 (−10.41%)
Table 14. The comparison results of error distribution on the Carina.
Table 14. The comparison results of error distribution on the Carina.
MethodCarina Object Error Distribution
≤5 mm (%)≤10 mm (%)≤15 mm (%)≤20 mm (%)
SOTA average [11]68.8494.7095.5597.12
Ours average74.92 (+8.83%)93.90 (−0.84%)97.65 (+2.20%)98.83 (+1.76%)
SOTA external val. [11]73.3391.6495.3396.54
Ours external val.68.67 (−6.35%)88.00 (−3.97%)96.67 (+1.41%)98.00 (+1.51%)
Table 15. The effect of softmax in GA.
Table 15. The effect of softmax in GA.
MethodMalposition Accuracy (%ETT-CarinaETT TipCarina
Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)
w/softmax90.115.2096.6283.9685.8004.2034.097
w/o softmax91.184.9115.1143.6893.8024.2383.862
Table 16. The effect of channel, kernel and SE block of SA.
Table 16. The effect of channel, kernel and SE block of SA.
MethodMalposition Accuracy (%)ETT-CarinaETT TipCarina
Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)
SA (c1 + k7)83.694.9044.8133.9983.6254.3863.750
SA (c1 + k1)85.835.6487.6284.9118.6054.1853.674
SA (c4 + k1)85.835.1826.2454.1884.0674.6115.759
SA (c8 + k1)87.705.0675.2484.2734.4184.3054.016
SA (c8 + k7)83.694.6444.4014.0073.6154.0283.372
SA (c16 + k1)85.564.8834.7783.9853.6964.3513.969
SA (w/o SE)86.365.4919.6974.61911.9974.3913.956
Table 17. The comparison results of attention modules.
Table 17. The comparison results of attention modules.
MethodMalposition Accuracy (%)ETT-CarinaETT TipCarina
Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)
FCOS [13]86.105.3357.8314.2545.4274.6597.497
FCOS + SE [36]85.035.4245.8544.2844.1564.5434.943
FCOS + CSPnonlocal [45]86.105.4045.8173.9803.7084.3324.416
FCOS + nonlocal [33]86.105.4226.1394.52110.0594.4114.423
FCOS + CBAM [15]86.105.3035.6544.3814.8704.3804.260
FCOS + CCAM [14]86.904.6324.4914.0253.6414.0353.517
FCOS + SA87.705.0675.2484.2734.4184.3054.016
Table 18. The comparison results of attention modules in parameters and GFLOPs.
Table 18. The comparison results of attention modules in parameters and GFLOPs.
MethodParameters (M)GFLOPs
FCOS [13]32.11819.764
FCOS + SE [36]32.126 (+0.02%)19.764 (+0%)
FCOS + CSPnonlocal [45]32.284 (+0.52%)19.782 (+0.09%)
FCOS + nonlocal [33]33.302 (+3.69%)19.882 (+0.60%)
FCOS + CBAM [15]32.127 (+0.03%)19.764 (+0%)
FCOS + CCAM [14]34.154 (+6.34%)19.964 (+1.01%)
FCOS + SA32.253 (+0.42%)19.778 (+0.07%)
Table 19. The results of GA and SA fusion method.
Table 19. The results of GA and SA fusion method.
MethodMalposition Accuracy (%)ETT-CarinaETT TipCarina
Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)
FCOS86.105.3357.8314.2545.4274.6597.497
FCOS + SA + GA83.965.2255.3064.3044.3764.4254.032
FCOS + GA || SA87.175.4926.5834.9569.5494.1644.158
FCOS + GA + SA87.974.8684.9534.1434.1574.0163.350
Table 20. The effect of fusing global modelling attention and scale attention.
Table 20. The effect of fusing global modelling attention and scale attention.
MethodMalposition Accuracy (%)ETT-CarinaETT TipCarina
Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)
FCOS86.105.3357.8314.2545.4274.6597.497
FCOS + nonlocal*2OOMOOMOOMOOMOOMOOMOOM
FCOS + CSPnonlocal*285.565.8008.9914.3915.5434.7037.512
FCOS + CCAM*286.104.8554.9884.0204.0374.3013.835
FCOS + SA*287.175.6436.8204.4325.0214.4225.387
FCOS + CSPnonlocal + SA86.905.7276.5184.6465.2674.6854.732
FCOS + CCAM + SA87.974.8684.9534.1434.1574.0163.350
Table 21. The results of employing mask branch into FCOS.
Table 21. The results of employing mask branch into FCOS.
MethodMalposition Accuracy (%)ETT-CarinaETT TipCarina
Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)
FCOS [13]86.105.3357.8314.2545.4274.6597.497
CTFA87.974.8684.9534.1434.1574.0163.350
Seg (All)87.974.9095.1793.9394.4684.0433.121
Seg (ETT)89.045.4867.6824.3986.7544.5214.244
Seg (ETT + Carina)90.115.3346.7524.0885.9894.3294.253
Seg (ETT + Carina) + Fusion91.184.9115.1143.6893.8024.2383.862
Table 22. The results of adopting mask prediction or not in the post-process algorithm.
Table 22. The results of adopting mask prediction or not in the post-process algorithm.
MethodMalposition Accuracy (%)ETT-CarinaETT TipCarina
Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)Mean Err. (mm)Err Std. (mm)
w/mask85.567.43810.5526.38910.6724.3294.253
w/o mask90.115.3346.7524.0885.9894.3294.253
Table 23. The visualization results.
Table 23. The visualization results.
Good Diagnostics 12 01913 i001 Diagnostics 12 01913 i002
Medium Diagnostics 12 01913 i003 Diagnostics 12 01913 i004
Worse Diagnostics 12 01913 i005 Diagnostics 12 01913 i006
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mao, L.-K.; Huang, M.-H.; Lai, C.-H.; Sun, Y.-N.; Chen, C.-Y. Detecting Endotracheal Tube and Carina on Portable Supine Chest Radiographs Using One-Stage Detector with a Coarse-to-Fine Attention. Diagnostics 2022, 12, 1913. https://doi.org/10.3390/diagnostics12081913

AMA Style

Mao L-K, Huang M-H, Lai C-H, Sun Y-N, Chen C-Y. Detecting Endotracheal Tube and Carina on Portable Supine Chest Radiographs Using One-Stage Detector with a Coarse-to-Fine Attention. Diagnostics. 2022; 12(8):1913. https://doi.org/10.3390/diagnostics12081913

Chicago/Turabian Style

Mao, Liang-Kai, Min-Hsin Huang, Chao-Han Lai, Yung-Nien Sun, and Chi-Yeh Chen. 2022. "Detecting Endotracheal Tube and Carina on Portable Supine Chest Radiographs Using One-Stage Detector with a Coarse-to-Fine Attention" Diagnostics 12, no. 8: 1913. https://doi.org/10.3390/diagnostics12081913

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop