Next Article in Journal
Harmful Microalgae Detection: Biosensors versus Some Conventional Methods
Previous Article in Journal
Adversarial Resolution Enhancement for Electrical Capacitance Tomography Image Reconstruction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Segmentation Performance Comparison Considering Regional Characteristics in Chest X-ray Using Deep Learning

1
Department of Biomedical Engineering, College of Health Science, Gachon University, Incheon 21936, Korea
2
Department of Health Sciences and Technology, Gachon Advanced Institute for Health Sciences and Technology (GAIHST), Gil Medical Center, Gachon University, Incheon 21936, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(9), 3143; https://doi.org/10.3390/s22093143
Submission received: 9 March 2022 / Revised: 6 April 2022 / Accepted: 14 April 2022 / Published: 20 April 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Chest radiography is one of the most widely used diagnostic methods in hospitals, but it is difficult to read clearly because several human organ tissues and bones overlap. Therefore, various image processing and rib segmentation methods have been proposed to focus on the desired target. However, it is challenging to segment ribs elaborately using deep learning because they cannot reflect the characteristics of each region. Identifying which region has specific characteristics vulnerable to deep learning is an essential indicator of developing segmentation methods in medical imaging. Therefore, it is necessary to compare the deep learning performance differences based on regional characteristics. This study compares the differences in deep learning performance based on the rib region to verify whether deep learning reflects the characteristics of each part and to demonstrate why this regional performance difference has occurred. We utilized 195 normal chest X-ray datasets with data augmentation for learning and 5-fold cross-validation. To compare segmentation performance, the rib image was divided vertically and horizontally based on the spine, clavicle, heart, and lower organs, which are characteristic indicators of the baseline chest X-ray. Resultingly, we found that the deep learning model showed a 6–7% difference in the segmentation performance depending on the regional characteristics of the rib. We verified that the performance differences in each region cannot be ignored. This study will enable a more precise segmentation of the ribs and the development of practical deep learning algorithms.

1. Introduction

A chest X-ray (CXRs) is essential for diagnosing rib fracture, heart shape, blood vessel expansion, pulmonary vessel size, and pulmonary edema [1]. However, X-ray images are difficult to read because several anatomic structures cross and overlap in the chest [2], making it is difficult for radiologists to identify diseases hidden in a chest background where organs overlap [3,4]. Foreground bones such as ribs and clavicles have been reported to cover 82%–95% of undiscovered lung cancers [5].
Thus, chest background structure removal or suppression algorithm is used to observe the chest area more clearly to detect lung nodules and cancer tissues. The difference-image processing proposed by Xu, which corresponds to the subtraction of a signal-enhanced image and a signal-suppressed image [6], is one method for suppressing the background structures. Shiraishi et al. proposed a removal segmentation method that first segmented the lung field and then filtered the background using the average radial-gradient technique [7]. The filter variable adaptively depends on the location of the region of interest (ROI) and the anatomical classification.
With the advancement of image processing studies, computer-aided diagnosis (CAD) studies and segmentation/classification of the medical image have been actively conducted to eliminate obstacles that interfere with diagnosis [8].
In biology fields, cell segmentation, and classification, the task of uniquely identifying each tumor cell in an image has shown significant progress [9]. Greenwald created the “Mesmer” method, a deep learning-enabled segmentation algorithm that performs nuclear and whole-cell segmentation in tissue imaging data. Mesmer, which achieves human-level performance for whole-cell segmentation, has enabled the automated extraction of key cellular features such as the subcellular localization of protein signals [10]
Staal proposed a 3D image segmentation framework that enables automatic segmentation and labeling of the entire rib on chest computed tomography (CT) [11]. For segmentation and labeling of complete ribs, the center of gravity of all ribs is calculated to decide centerlines. The ribs are labeled according to the calculated centerlines. This method was validated with 96.8% sensitivity, 97.8% specificity and 80% accuracy.
Similar to previous studies [6,7,8,9,10,11], rib segmentation can facilitate the accurate diagnosis of lesions. However, separating ribs on chest X-rays requires professional knowledge in medical anatomy. Additionally, it is difficult for experts to find all 12 pairs of ribs individually, and manually labeling them is laborious and time-consuming [12].
As a solution, rib segmentation using deep learning has been proposed. Segmentation using deep learning aids in obtaining useful information by performing rib suppression in CXRs, resulting in a more accurate diagnosis [13]. Due to the growing demand for CAD, deep learning has shown significant progress in medical imaging [14,15].
Arif et al. performed automatic segmentation of cervical spine X-ray images using a novel shape-aware deep segmentation network (Unet-S) and achieved a dice similarity coefficient (DSC) of 0.84 [16]. In 2020, Oliveira proposed a segmentation pipeline using a 3D model to aid in the rib segmentation of 2D X-ray images. It consisted of the MIP procedure for acquiring bone labels from CT scans, lung segmentation from CXR labels, and CoDAGAN for segmenting CXR ribs, achieving a maximum of 0.934 AUC [17].
Therefore, segmentation and image processing studies for rib diagnosis are well established. However, ribs overlap with various internal organs such as the lungs, heart, clavicle, and lower organs, making different characteristics for each rib region [18]. Using deep learning, these regional characteristics can be automatically extracted from the medical image [19], and we termed the image as having “regional characteristics”, which affect the deep learning process. Current studies have limitations in that deep learning has been conducted without considering the regional characteristics of ribs.
Hence, to minimize the difference in characteristics by rib regions, Wang proposed a multitask dense connection U-net (MDU-Net) that enabled multiple dense connections by combining Dense-Net with the U-Net model. The method separately divides the clavicle, anterior ribs, and posterior ribs, and each part passes multiple dense layers and then combines the segmentation results into one. This method succeeded in segmenting the ribs with 88% DSC [20].
Although rib images can be more clearly expressed by using image processing, filters, and 3D frames, it is necessary to verify whether a deep learning model delicately reflects the rib’s various features. However, there have been no reports verifying whether deep learning has completely reflected each region’s characteristics.
For complete segmentation of the entire rib, it is essential to consider the characteristics of each rib region. We hypothesized that deep learning, which did not reflect regional characteristics, causes regional performance differences. Therefore, this study quantitatively analyzed deep learning performance differences based on rib region to verify whether deep learning reflects the characteristics of each part. To compare segmentation performance, the rib image was divided based on the spine, clavicle, heart, and lower organs, which are characteristic indicators of the baseline chest X-ray.
As a result of deep learning, the predicted ribs masks were divided horizontally into superior, middle, and inferior parts based on the clavicle and heart, and vertically into medial and lateral parts around the spine to determine which region specifically showed high performance (Figure 1).
Previous studies so far have focused on preprocessing rib images to provide reliable segmented images, and it is now necessary to verify how exactly deep learning reflects the characteristics of each region-specific feature.
Our study should provide essential pipelines for producing more strong rib segmentation algorithms in the future, and help experts reflect regional characteristics more accurately when applying deep learning to other medical images.

2. Materials and Methods

2.1. Data Collection and Annotation

A total of 195 chest X-ray images formatted with digital imaging and communications in medicine (DICOM) were collected from Gil Hospital. The institutional review board of GUGMC approved this retrospective study and waived the requirement for informed patient consent (approval number: GBIRB-2019-337). Five experts who were trained to identify ribs for specialists demarcated the ROI, 12 pairs of ribs, on the images using ImageJ (NIH, Bethesda, MD, USA), and a specialist determined the final ROI after revision and modification.

2.2. Data Pre-Processing

For deep learning, resizing and zero-padding were applied to all the input images. The size was set to 1024 × 1024 pixels, with a value of 8 bits per pixel. Most deep learning methods show high performance when the number of training datasets is large. Therefore, to prevent overfitting during training and improve model performance, data augmentation was used for training data [21].
Three transformed images were produced from one training image by applying a rotation of approximately 10° and space augmentation or shifting translation of approximately 5%. Thus, 779 training images were generated with data augmentation, and the remaining 39 test images were used for verification. We used a five-fold cross-validation method to present the model’s performance quantitatively [22].

2.3. U-Net Architecture

A U-Net architecture, a generic deep learning solution for biomedical image segmentation based on an encoder–decoder structure, was used to extract the ribs’ image features [23,24,25,26]. As shown in Figure 2, the U-net has a U-shaped symmetrical form that includes a contracting path called an encoder to extract image features and an expansion path called a decoder to expand the feature map [27]. In particular, it is characterized by a U-net architecture with skip connections at the same depths to skip features from the contracting path to the expanding path to minimize spatial information lost during down-sampling [28].
The rectified linear unit (ReLu) function was used as the activation function, whereas the sigmoid function was used in the last convolutional layer. The hyperparameter was set to four batch sizes, 100 epochs, and a 0.001 learning rate. Additional information of the U-Net algorithm is presented in Table 1.

2.4. Criteria of Dividing Ribs

Original rib images (Figure 3b) were divided based on two criteria to compare the deep learning performance by region. The first criterion was to divide vertically into the medial and lateral sides along the lines of the spine (Figure 3a). The second criterion was to divide horizontally into the superior, middle, and inferior sides based on the clavicle, heart, and lower organs that consist of soft tissue (Figure 3b).

3. Results

To verify the performance of the rib segmentation model, 39 test images that were not used in training were used. The segmentation model performance was assessed by comparing the segmented images with the ground truth mask pixel-by-pixel. Three performance indicators were used: precision, recall, sensitivity, and DSC. True positive (TP) cases were those in which the rib was recognized as a rib, whereas false positive (FP) cases were those in which non-rib regions were recognized as the rib. Furthermore, true negatives (TNs) implied that non-rib regions were recognized as non-ribs, whereas false negatives (FNs) implied that ribs were incorrectly recognized as non-ribs. The precision, recall, sensitivity, and DSC were determined using Equations (1)–(4), respectively.
Precision = T P T P + F P
Recall = T P T P + F N
Sensitivity = T P T P + F N
DSC = T P T P + T N + F P + F N × 2

3.1. Whole Ribs Segmentation Model Performance

Figure 4 depicts the segmented result images.
The segmentation result images were compared with the ground-truth images in the test dataset (39). On average, we observed a precision of 93.51 ± 0.04%, recall of 88.56 ± 0.86%, sensitivity of 88.14 ± 0.05%, and DSC of 89.65 ± 0.04 (Table 2).

3.2. Comparison Performance Difference between Medial and Lateral Ribs

Thirty-nine test sets were created, with entire rib images divided into medial and lateral regions based on the spine. The ground truth and predicted images were compared to show the differences between the two regions. This process is shown in Figure 5.
On average, medial ribs had a precision of 89.756 ± 0.07%, a recall of 82.66 ± 0.82%, a sensitivity of 82.526 ± 0.08%, and a DSC of 84.756 ± 0.04%, while lateral ribs had a precision of 95.22 ± 0.03%, a recall of 90.89 ± 0.9%, a sensitivity of 91.03 ± 0.04%, and a DSC of 92.03 ± 0.03% (Table 3).

3.3. Comparison Performance Difference between Superior, Middle, and Inferior Ribs

Thirty-nine test sets were created, with all rib images divided into superior, middle, and inferior regions based on the clavicle, heart, and lower organs. The ground truth and predicted images were compared to show the differences between the three regions. This process is shown in Figure 6.
On average, precision, recall, sensitivity, and dice similarity coefficient values in the superior ribs were 92.69 ± 0.04%, 88.92 ± 0.88%, 88.92 ± 0.05%, 89.94 ± 0.04%; 94.11 ± 0.04%, 90.31 ± 0.90%, 89.88 ± 0.05%, 91.01 ± 0.04% in the middle ribs; and 93.45 ± 0.04%, 83.53 ± 0.83%, 83.294 ± 0.1%, 85.8 ± 0.2% in the inferior ribs, respectively (Table 4).

4. Discussion and Conclusions

In this work, we postulated that deep learning training without considering the regional characteristics of ribs would cause differences in performance for each region. To verify this, we investigated whether significant differences occurred by comparing the deep learning performance of each region by dividing the ribs vertically and horizontally based on specific structures.
In the rib image, which was divided into medial and lateral ribs based on the spine, the performance was 7% higher in the lateral rib than in the medial rib. Additionally, rib images were divided into superior, middle, and inferior regions based on the clavicle, heart, and lower organs. The middle ribs showed the highest performance among them and presented an average of 6% higher performance compared with the inferior region, which showed the lowest performance.
Therefore, it was confirmed that even if the same deep learning model was used, a performance difference of at least 6% was observed for each region. To comprehend these results, the characteristics of X-rays need to be considered.
The X-ray equipment commonly used in hospitals works on a principle in which particles of X-rays are irradiated to the body and transmitted, and the remaining X-rays are detected in the film [29]. Because the radiation density in our body is higher in the order of air, fat, organs, and bones, the lung area containing the most air appears dark, and the bone with the highest density appears bright in the X-ray image [30].
Therefore, in the image divided into medial and lateral based on the spine, it is difficult to distinguish between the spine and ribs because the density of the spine and ribs is similar. Moreover, because the boundary of the costal cartilage, which is set between the spine and ribs, is not clear to distinguish, it is considered that the medial region including the spine performed worse than the lateral region.
In rib segmentation, which is divided into superior, middle, and inferior, the inferior rib of the lower organ region, which contains more soft tissue, appears brighter, similar to other bones in the image. This makes the inferior region, which includes various organs, perform worse than the others.
Here, we used a normal chest X-ray image dataset at Gil Hospital. In the future, reinforcement learning with additional datasets of rib fractures and diseases such as complications of external fixation/traction or fracture union [31] can improve the developed model.
We divided the entire ribs into two criteria based on the specific structures of the spine, clavicle, heart, and lower organs. In the future, comparing the performance by dividing the region of every 12 pairs of ribs will enable the discovery of new rib region segmentation criteria and serve as the foundation for studying rib fracture detection.
Figure 7 shows the error pixel, which makes false positives and false negatives. Therefore, to minimize these error pixels, we are currently further studying algorithms for rib labeling for dividing the region of every 12 pairs. It is expected that the revised algorithm will minimize FP and FN ratio by recognizing pixels more delicately in the future. We expect that our rib comparison study will improve the reliability of the segmentation method in digital radiography.
In this study, only U-Net was used among the CNN architectures, but regional performance comparisons with other CNN models are essential for the normalization regional performance differences [32]. We will conduct investigations to quantify and examine regional performance differences according to the architecture by using more advanced architectures such as the ResNet [33], GoogLeNet [34], and AlexNet [35] series in the future.
In conclusion, the chest X-ray deep learning model showed a 6%–7% difference in segmentation performance depending on the regional characteristics of the rib. This result shows that feature extraction technology that considers region specific characteristics is required in the deep learning process.
Therefore, this study is meaningful in that it provides the importance of reflecting the characteristics of each region as well as presents the direction of erstwhile rib research. If feature extraction technology is developed through further research in the future, it is expected that the characteristics of each region can be expressed more precisely in medical images.

Author Contributions

Drafted the manuscript, conducted the deep learning analysis, and performed statistical analyses, H.M.L.; designed the study, supervised the deep learning analysis, and revised the manuscript, Y.J.K. and K.G.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the GRRC program of Gyeonggi Province (GRRC-Gachon2020(B01)), AI-based Medical Image Analysis, and the Ministry of Science and ICT (MSIT), Korea, under the Information Technology Research Center (ITRC) support program (IITP-2022-2017-0-01630) supervised by the Institute for Information & Communications Technology Promotion (IITP), which is supported by the Gachon University Gil Medical Center (FRD2019-11-02(3)).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Institutional Review Board of Gil Medical Center (GBIRB-2019-337).

Informed Consent Statement

All the records from patients were retrospectively reviewed.

Data Availability Statement

The datasets generated and/or analyzed during the current study are not publicly available because the institutional review board did not grant permission to share patient data, but they are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pelicano, A.C.; Gonçalves, M.C.T.; Godinho, D.M.; Castela, T.; Orvalho, M.L.; Ara, N.A.M.; Porter, E.; Conceiç, R.C. Development of 3D MRI-Based Anatomically Realistic Models of Breast Tissues and Tumors for Microwave Imaging Diagnosis. Sensors 2021, 21, 8265. [Google Scholar] [CrossRef]
  2. Tsoumakidou, M.; Chrysofakis, G.; Tsiligianni, I.; Maltezakis, G.; Siafakas, N.M.; Tzanakis, N. A Prospective Analysis of 184 Hemoptysis Cases—Diagnostic Impact of Chest X-ray, Computed Tomography, Bronchoscopy. Respiration 2006, 73, 808–814. [Google Scholar] [CrossRef] [PubMed]
  3. Carmody, D.P.; Nodine, C.F.; Kundel, H.L. An Analysis of Perceptual and Cognitive Factors in Radiographic Interpretation. Perception 1980, 9, 339–344. [Google Scholar] [CrossRef]
  4. O’Connor, S.D.; Yao, J.; Summers, S.R. Lytic Metastases in Thoracolumbar Spine: Computer-Aided Detection at CT—Preliminary Study. Radiology 2007, 242, 811–816. [Google Scholar] [CrossRef] [PubMed]
  5. Austin, J.H.; Romney, B.M.; Missed, L.S. Bronchogenic Carcinoma: Radiographic Findings in 27 Patients with a Potentially Resectable Lesion Evident in Retrospect. Radiology 1992, 182, 115–122. [Google Scholar] [CrossRef] [PubMed]
  6. Xu, X.; Doi, K.; Kobayashi, T.; MacMahon, H.; Giger, M.L. Development of an Improved CAD Scheme for Automated Detection of Lung Nodules in Digital Chest Images. Med. Phys. 1997, 24, 1395–1403. [Google Scholar] [CrossRef]
  7. Shiraishi, J.; Li, Q.; Suzuki, K.; Engelmann, R.; Doi, K. Computer-Aided Diagnostic Scheme for the Detection of Lung Nodules on Chest Radiographs: Localized Search Method Based on Anatomical Classification. Med. Phys. 2006, 33, 2642–2653. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Doi, K. Computer-Aided Diagnosis in Medical Imaging: Historical Review, Current Status and Future Potential. Comput. Med. Imaging Graph. 2007, 31, 198–211. [Google Scholar] [CrossRef] [Green Version]
  9. Verma, R.; Kumar, N.; Patil, A. MoNuSAC2020: A Multi-Organ Nuclei Segmentation and Classification Challenge. IEEE Trans. Med. Imaging 2021, 40, 3413–3423. [Google Scholar] [CrossRef]
  10. Greenwald, N.F.; Miller, G.; Moen, E.; Kong, A.; Kagel, A.; Dougherty, T.; Fullaway, C.S.; McIntosh, B.J.; Leow, K.X.; Schwartz, M.S. Whole-Cell Segmentation of Tissue Images with Human-Level Performance Using Large-Scale Data Annotation and Deep Learning. Nat. Biotechnol. 2021, 40, 1–29. [Google Scholar] [CrossRef]
  11. Staal, J.; van Ginneken, B.; Viergever, M.A. Automatic Rib Segmentation and Labeling in Computed Tomography Scans Using a General Framework for Detection, Recognition and Segmentation of Objects in Volumetric Data. Med. Image Anal. 2007, 11, 35–46. [Google Scholar] [CrossRef]
  12. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A Survey on Deep Learning in Medical Image Analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  13. Song, H.-J.; Lee, E.-B.; Jo, H.-J.; Park, S.-Y.; Kim, S.-Y.; Kim, H.-J.; Hong, J.-W. Evaluation of Classification and Accuracy in Chest X-ray Images Using Deep Learning with Convolution Neural Network. J. Korean Soc. Radiol. 2020, 14, 39–44. [Google Scholar]
  14. Zhang, Y.; Gorriz, J.M.; Dong, Z. Deep Learning in Medical Image Analysis. J. Imaging 2021, 7, 74. [Google Scholar] [CrossRef]
  15. Sharma, H.B.; Panigrahi, S.; Sarmah, A.K.; Dubey, B.K. ULNet for the Detection of Coronavirus (COVID-19) from Chest X-ray Images. Sci. Total Environ. 2021, 137, 104834. [Google Scholar]
  16. Al Arif, S.M.M.R.; Knapp, K.; Slabaugh, G. Fully Automatic Cervical Vertebrae Segmentation Framework for X-ray Images. Comput. Methods Programs Biomed. 2018, 157, 95–111. [Google Scholar] [CrossRef] [Green Version]
  17. Oliveira, H.; Mota, V.; Machado, A.M.; dos Santos, J.A. From 3D to 2D: Transferring Knowledge for Rib Segmentation in Chest X-rays. Pattern Recognit. Lett. 2020, 140, 10–17. [Google Scholar] [CrossRef]
  18. Kundel, H.L.; Nodine, N.C. Interpreting Chest Radiographs without Visual Search. Radiology 1975, 116, 527–532. [Google Scholar] [CrossRef]
  19. Liang, C.; Xin, S. Research Status and Prospects of Deep Learning in Medical Images. In Proceedings of the 2020 International Conference on Communications, Information System and Computer Engineering (CISCE), Kuala Lumpur, Malaysia, 3–5 July 2020. [Google Scholar]
  20. Wang, W.; Feng, H.; Bu, Q.; Cui, L.; Xie, Y.; Zhang, A.; Feng, J.; Zhu, Z.; Chen, Z. MDU-Net: A Convolutional Network for Clavicle and Rib Segmentation from a Chest Radiograph. J. Healthc. Eng. 2020, 2020, 9. [Google Scholar] [CrossRef]
  21. Shorten, C.; Khoshgoftaar, T.M. A Survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 1–48. [Google Scholar] [CrossRef]
  22. Yadav, S.; Shukla, S. Analysis of K-Fold Cross-Validation over Hold-Out Validation on Colossal Datasets for Quality Classification. In Proceedings of the 6th International Conference on Advanced Computing (IACC), Bhimavaram, India, 27–28 February 2016. [Google Scholar]
  23. Kim, Y.J.; Ganbold, B.; Kim, K.G. Web-Based Spine Segmentation Using Deep Learning in Computed Tomography Images. Healthc. Inform. Res. 2020, 26, 61–67. [Google Scholar] [CrossRef]
  24. Horng, M.H.; Kuok, C.P.; Fu, M.J.; Lin, C.J.; Sun, Y.N. Cobb Angle Measurement of Spine from X-ray Images Using Convolutional Neural Network. Comput. Math. Methods Med. 2019, 2019, 18. [Google Scholar] [CrossRef] [Green Version]
  25. Li, X.; Chen, H.; Qi, X.; Dou, Q.; Fu, C.W.; Heng, P.A. H-DenseUNet: Hybrid Densely Connected UNet for Liver and Tumor Segmentation from CT Volumes. IEEE Trans. Med. Imaging 2018, 37, 2663–2674. [Google Scholar] [CrossRef] [Green Version]
  26. Albishri, A.A.; Shah, S.J.H.; Lee, Y. CU-Net: Cascaded U-Net Model for Automated Liver and Lesion Segmentation and Summarization. In Proceedings of the 2019 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), San Diego, CA, USA, 18–21 November 2019. [Google Scholar]
  27. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar] [CrossRef]
  28. Drozdzal, M.; Vorontsov, E.; Chartrand, G.; Kadoury, S.; Pal, C. Importance of skip connections in biomedical image segmentation. arXiv 2016, arXiv:1608.04117. [Google Scholar] [CrossRef] [Green Version]
  29. Seibert, J.A. X-ray Imaging Physics for Nuclear Medicine Technologists. J. Nucl. Med. Technol. 2004, 32, 139–147. [Google Scholar]
  30. Wolde Hawariat, B.Y.; Fenta, B.D.; Gebreselassie, H.A. Subcutaneous Chest Wall Hamartoma: Case Report. J. Pediatr. Surg. Case Rep. 2021, 73, 102009. [Google Scholar] [CrossRef]
  31. Nirula, R.; Diaz, J.J.; Trunkey, D.D.; Mayberry, J.C. Rib Fracture Repair: Indications, Technical Issues, and Future Directions. World J. Surg. 2009, 33, 14–22. [Google Scholar] [CrossRef]
  32. Zhang, R.; Du, L.; Xiao, Q.; Liu, J. Comparison of Backbones for Semantic Segmentation Network. J. Phys. Conf. Ser. 2020, 1544, 012196. [Google Scholar] [CrossRef]
  33. Cai, W.; Wei, Z.; Song, Y.; Li, M.; Yang, X. Residual-Capsule Networks with Threshold Convolution for Segmentation of Wheat Plantation Rows in UAV Images. Multimed. Tools Appl. 2021, 80, 32131–32147. [Google Scholar] [CrossRef]
  34. Alqudah, A.M.; Alquraan, H.; Qasmieh, I.A. Segmented and Non-Segmented Skin Lesions Classification Using Transfer Learning and Adaptive Moment Learning Rate Technique Using Pretrained Convolutional Neural Network. J. Biomim. Biomater. Biomed. Eng. 2019, 42, 67–78. [Google Scholar]
  35. Chen, J.; Wan, Z.; Zhang, J.; Li, W.; Chen, Y.; Li, Y.; Duan, Y. Medical Image Segmentation and Reconstruction of Prostate Tumor Based on 3D AlexNet. Comput. Methods Programs Biomed. 2021, 200, 105878. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Flowchart that presents the total process of comparing region performance by dividing the predicted ribs’ mask.
Figure 1. Flowchart that presents the total process of comparing region performance by dividing the predicted ribs’ mask.
Sensors 22 03143 g001
Figure 2. U-Net architecture for segmentation of ribs.
Figure 2. U-Net architecture for segmentation of ribs.
Sensors 22 03143 g002
Figure 3. Criteria of dividing ribs: (a) Ribs vertically divide into medial and lateral of spine; (b) Original chest X-ray image; (c) Ribs horizontally divided into superior, middle, and inferior based on the clavicle, heart, and lower organs.
Figure 3. Criteria of dividing ribs: (a) Ribs vertically divide into medial and lateral of spine; (b) Original chest X-ray image; (c) Ribs horizontally divided into superior, middle, and inferior based on the clavicle, heart, and lower organs.
Sensors 22 03143 g003
Figure 4. Segmentation model results. (a) Original images, (b) Ground truth images, (c) U-Net predicted images, and (d) Overlay images with (a,c).
Figure 4. Segmentation model results. (a) Original images, (b) Ground truth images, (c) U-Net predicted images, and (d) Overlay images with (a,c).
Sensors 22 03143 g004
Figure 5. Vertical crop image divided into medial and lateral based on spine: (a) Entire rib image before cropping; (b) Ground truth images of medial ribs; (c) Ground truth images of lateral ribs; (d) Predicted images of medial ribs; and (e) Predicted images of lateral ribs.
Figure 5. Vertical crop image divided into medial and lateral based on spine: (a) Entire rib image before cropping; (b) Ground truth images of medial ribs; (c) Ground truth images of lateral ribs; (d) Predicted images of medial ribs; and (e) Predicted images of lateral ribs.
Sensors 22 03143 g005
Figure 6. Horizontal crop image divided into superior, middle, and below based on the clavicle, heart, and lower organs: (a) Entire rib image before cropping; (b) Ground truth images of superior ribs; (c) Ground truth images of middle ribs; (d) Ground truth images of inferior ribs; (e) Predicted images of superior ribs; (f) Predicted images of middle ribs; and (g) Predicted images of inferior ribs.
Figure 6. Horizontal crop image divided into superior, middle, and below based on the clavicle, heart, and lower organs: (a) Entire rib image before cropping; (b) Ground truth images of superior ribs; (c) Ground truth images of middle ribs; (d) Ground truth images of inferior ribs; (e) Predicted images of superior ribs; (f) Predicted images of middle ribs; and (g) Predicted images of inferior ribs.
Sensors 22 03143 g006
Figure 7. Segmentation predicted result images along with the false positives (FP: red color) and false negatives (FN: blue color).
Figure 7. Segmentation predicted result images along with the false positives (FP: red color) and false negatives (FN: blue color).
Sensors 22 03143 g007
Table 1. U-Net architecture and hyperparameters.
Table 1. U-Net architecture and hyperparameters.
Unit LevelConv LayerFilterActivation FunctionOutput Size
Input 1024 × 1024 × 1
EncodingLevel 1Conv 13 × 3/32ReLu1024 × 1024 ×32
Conv 23 × 3/321024 × 1024 × 32
Level 2Conv 33 × 3/64ReLu512 × 512 × 64
Conv 43 × 3/64512 × 512 × 64
Level 3Conv 53 × 3/128ReLu256 × 256 × 128
Conv 63 × 3/128256 × 256 × 128
Level 4Conv 73 × 3/256ReLu128 × 128 × 256
Conv 83 × 3/256128 × 128 × 256
BridgeLevel 5Conv 93 × 3/512ReLu64 × 64 × 512
Conv 103 × 3/51264 × 64 × 512
DecodingLevel 6Conv 113 × 3/256ReLu128 × 128 × 256
Conv 123 × 3/256128 × 128 × 256
Level 7Conv 133 × 3/128ReLu256 × 256 × 128
Conv 143 × 3/128256 × 256 × 128
Level 8Conv 153 × 3/64ReLu512 × 512 × 64
Conv 163 × 3/64512 × 512 × 64
Level 9Conv 173 × 3/32ReLu1024 × 1024 × 32
Conv 183 × 3/321024 × 1024 × 32
Output Conv 191 × 1Sigmoid1024 × 1024 × 1
Table 2. Five-fold cross validation results of the trained segmentation model.
Table 2. Five-fold cross validation results of the trained segmentation model.
FoldsPrecision (%)Recall (%)Sensitivity (%)DSC (%)
Fold-192.0687.5487.1788.95
Fold-292.7287.6187.6289.1
Fold-397.7488.4487.4189.74
Fold-497.7690.1489.1391.08
Fold-587.2589.0489.3989.39
Average93.51 ± 0.0488.56 ± 0.8688.14 ± 0.0589.65 ± 0.04
Table 3. Five-fold cross validation results of the medial and lateral ribs.
Table 3. Five-fold cross validation results of the medial and lateral ribs.
Precision (%)Recall (%)Sensitivity (%)DSC (%)
Medial Ribs89.76 ± 0.0782.66 ± 0.8282.53 ± 0.0884.76 ± 0.04
Lateral Ribs95.22 ± 0.0390.89 ± 0.991.03 ± 0.0492.03 ± 0.03
Table 4. Five-fold cross validation results of the superior, middle, and inferior ribs.
Table 4. Five-fold cross validation results of the superior, middle, and inferior ribs.
Precision (%)Recall (%)Sensitivity (%)DSC (%)
Superior Ribs92.69 ± 0.0488.92 ± 0.8888.92 ± 0.0589.94 ± 0.04
Middle Ribs94.11 ± 0.0490.31 ± 0.9089.88 ± 0.0591.01 ± 0.04
Inferior Ribs93.45 ± 0.0483.53 ± 0.8383.294 ± 0.185.8 ± 0.2
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, H.M.; Kim, Y.J.; Kim, K.G. Segmentation Performance Comparison Considering Regional Characteristics in Chest X-ray Using Deep Learning. Sensors 2022, 22, 3143. https://doi.org/10.3390/s22093143

AMA Style

Lee HM, Kim YJ, Kim KG. Segmentation Performance Comparison Considering Regional Characteristics in Chest X-ray Using Deep Learning. Sensors. 2022; 22(9):3143. https://doi.org/10.3390/s22093143

Chicago/Turabian Style

Lee, Hyo Min, Young Jae Kim, and Kwang Gi Kim. 2022. "Segmentation Performance Comparison Considering Regional Characteristics in Chest X-ray Using Deep Learning" Sensors 22, no. 9: 3143. https://doi.org/10.3390/s22093143

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop