Next Article in Journal
Host Defence Peptides in Diabetes Mellitus Type 2 Patients with Periodontal Disease. A Systematic Review
Next Article in Special Issue
Shear Wave Elastography and Thyroid Imaging Reporting and Data System (TIRADS) for the Risk Stratification of Thyroid Nodules—Results of a Prospective Study
Previous Article in Journal
VGG19 Network Assisted Joint Segmentation and Classification of Lung Nodules in CT Images
Previous Article in Special Issue
Whole-Body Radioiodine Effective Half-Life in Patients with Differentiated Thyroid Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Based Fast Screening Approach on Ultrasound Images for Thyroid Nodules Diagnosis

1
Department of Mechanical Engineering, National Taiwan University of Science and Technology, Taipei 106, Taiwan
2
Department of Electrical Engineering, National Taiwan University of Science and Technology, Taipei 106, Taiwan
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(12), 2209; https://doi.org/10.3390/diagnostics11122209
Submission received: 26 October 2021 / Revised: 21 November 2021 / Accepted: 25 November 2021 / Published: 26 November 2021

Abstract

:
Thyroid nodules are widespread in the United States and the rest of the world, with a prevalence ranging from 19 to 68%. The problem with nodules is whether they are malignant or benign. Ultrasonography is currently recommended as the initial modality for evaluating thyroid nodules. However, obtaining a good diagnosis from ultrasound imaging depends entirely on the radiologists levels of experience and other circumstances. There is a tremendous demand for automated and more reliable methods to screen ultrasound images more efficiently. This research proposes an efficient and quick detection deep learning approach for thyroid nodules. An open and publicly available dataset, Thyroid Digital Image Database (TDID), is used to determine the robustness of the suggested method. Each image is formatted into a pyramid tile-based data structure, which the proposed VGG-16 model evaluates to provide segmentation results for nodular detection. The proposed method adopts a top-down approach to hierarchically integrate high- and low-level features to distinguish nodules of varied sizes by employing fuse features effectively. The results demonstrated that the proposed method outperformed the U-Net model, achieving an accuracy of 99%, and was two times faster than the competitive model.

1. Introduction

The thyroid gland is a butterfly-shaped endocrine gland in the lower front of the neck [1]. This endocrine gland produces thyroid hormones, which are then released into the bloodstream and help maintain the human body’s metabolism [2]. Thyroid cancer is increasing worldwide, while the fatality rate remains steady [3]. Thyroid nodules are relatively common with a prevalence of 19–68% in the general population and are usually discovered incidentally in the first routine neck imaging scan [3,4]. In 2019, around 52,070 people in the United States were diagnosed with thyroid cancer; among them, females are more prevalent than males [5].
There have been numerous initiatives in the last ten years to reduce the prevalence of thyroid cancer [6,7]. Ultrasound imaging is the most commonly used technique in thyroid radiology for thyroid nodule assessment due to its non-invasive nature and cost-effectiveness [8]. The most significant factor in a successful surgery outcome is accurate thyroid nodule diagnosis. Besides conventional methods in clinical diagnostics, computer-aided diagnosis (CAD) systems have become more common [9]. Digitalizing disease detection aims to achieve high accuracy for the diagnosis process and reduce patient cost and time. Various Machine Learning (ML) methods have been utilized for the advancement of the thyroid ultrasound CAD system. Deep Learning (DL), a subdomain of ML, has rapidly grown in medical imaging analysis and computer vision and is frequently seen as a viable alternative for evaluating ultrasound images [10].
Previous researchers have presented various approaches to detect nodules in ultrasound images. The authors of [11] presented a CAD system that applied a histogram analysis and segmentation-based fractal texture analysis algorithm for measuring the direction-independent features in ultrasound images to detect thyroid nodules. To distinguish between malignant and benign thyroid nodules, a support vector machine (SVM) and a random forest classifier were employed to extract characteristics. The authors also used segmentation techniques to improve nodules’ classification for more accurate diagnoses. Researchers Nougroho et al. [12] developed a CAD system to diagnose thyroid cancer. The primary purpose was to facilitate the radiologists in analyzing important characteristics of ultrasound images using a digital image processing method. Their proposed method had four stages: image enhancement, segmentation, feature extraction, and classifying each characteristic using multilayer perceptron (MLP) and SVM, and determining whether the tumor is benign or malignant. Song et al., in [13], introduced the InceptionV3-based approach for detecting thyroid nodules. The primary purpose of their research was to assist medical experts in identifying benign nodules and avoiding unnecessary Fine Needle Aspiration (FNA). They trained their algorithm on a shallow cropped nodule dataset created with the help of a physician. The performance of their experiment showed that their model might assist radiologists in recognizing malignant nodules with promising results. Authors Ko, S.Y. et al. [14] presented the convolutional neural network (CNN) model for thyroid cancer malignancy detection and compared the model output with the radiologists’ diagnostic performance. Two pre-trained models, i.e., “imagenet-vgg-verydeep16” and “imagenet-vgg-f” were used. A radiologist extracted the region of interest (ROI) from each ultrasound image to train the CNN with their local data. Results showed that both CNN performed similarly to expert radiologists’ images in differentiating thyroid cancer. The authors of [15] present a novel CAD system for categorizing and detecting thyroid ultrasound images driven by task-specific knowledge. The approach they proposed is divided into two parts. First, a multi-scale region-based detection network was built to learn pyramidal features for recognizing nodules at different scales. The following step was to create a multi-branch classification network with multi-view diagnosis-oriented features. Each network branch improved on a specific set of features that radiologists commonly employ. The authors claimed that their proposed CAD system outperformed 8% of expert radiologists’ findings. Vasile, M.C. et al. [16] published another study on diagnosing and classifying four different forms of thyroid nodules. They used an ensemble approach that combined two deep learning models for this purpose. Results showed that the proposed ensemble CNN–VGG technique outperformed the 5-CNN and VGG-19 models, achieving an overall accuracy of 97.35%. For an automatic diagnosis of thyroid nodules, a multitask cascade deep learning model (MCDLM) was presented by Yang, W. et al. [17], which integrated radiologists’ diverse domain knowledge (DK) and leveraged multimodal ultrasound images. The authors used the U-Net model and the dual-path semi-supervised conditional generative adversarial network (DScGAN) model for the precise segmentation results to generate high-quality images for discriminative purposes. After that, DScGAN generated images trained for a supervised support vector machine (S3VM) for thyroid nodule classification. Results showed that MCDLM achieved 90.01% classification accuracy. Another author, Abdolali, F. et al. [18], proposed an approach capable of detecting a variety of thyroid nodules. The proposed multitask model, Mask R-CNN, used regularization with a loss function and prioritized detection over-segmentation. Their suggested model outperformed Faster R-CNN’s and the traditional Mask R-CNN’s results.
Automatic precise detection of thyroid nodules is a crucial but challenging step for several reasons, i.e., blurry appearance, vague margin, irregular shape, and difficulty distinguishing between healthy tissues and nodule region. This research presented an automated technique for detecting and segmenting thyroid nodules using ultrasound images. The suggested approach uses a DL model with a fully convolutional neural network and a VGG-16 backbone to improve detection accuracy, and utilizing customized VGG-16 model results in achieving higher accuracy with a comparably simple model. We tested our model using several quantitative measures using a free and publicly available thyroid nodule dataset. Ground truth analysis was used to validate the thyroid nodule segmentation results.
The main contributions of our research consist of:
  • The proposed approach can precisely segment the thyroid nodule from ultrasound images despite blurring and noise effect fluctuations in input images.
  • The dataset employed for this study has diverse characteristics, consisting of 400 thyroid ultrasound images from five separate diagnosis stages, which are indicated by Thyroid Imaging Reporting and Data System (TIRADS)-1 to TIRADS-5.
  • The annotation is created for the ground truth masks because the current dataset partially lacks labeling work.
  • To validate the effectiveness of the proposed method, a statistical analysis of the proposed model in comparison with the U-Net model was provided.
The remaining sections are described as follows. Section 2 explains the methodology of the proposed convolutional neural network and its architecture. Subsequently, in Section 3, results with quantitative analysis of the proposed method and its comparison with other studies are presented. Finally, Section 4 concludes the paper with a future perspective.

2. Methodology

This Section presents the proposed method employed for thyroid nodule detection. The proposed convolutional neural network (CNN) used VGG-16 architecture as the backbone to process the ultrasound images [19]. Figure 1 depicts the methodology workflow. Each nodule image is first converted into a hierarchical tile-based data structure and processed to access the results of the nodule segmentation via the proposed CNN. Figure 2 shows the detailed architecture of the applied VGG-16 model. Furthermore, the effectiveness of the proposed approach was compared to that of another popular modern U-Net model [20].

2.1. Dataset

Collecting a significant amount of thyroid nodules-based ultrasound images is challenging due to time constraints and patient cooperation. Therefore, we chose to use a publicly available thyroid nodules images dataset. The Thyroid Digital Image Database (TDID) is an open and public dataset of Universidad Nacional de Colombia [21]. The TDID dataset, consisting of 400 ultrasonography thyroid images from 298 patients, was published in 2015. For each patient, one or more ultrasound images of the thyroid were obtained. The image size is 560 × 360 pixels, and it includes a detailed explanation and diagnostic description of the suspected thyroid lesions written by radiologists. The Thyroid Imaging Reporting and Data System (TI-RADS) of the American College of Radiology [22] scores were awarded to each image to predict the risk of thyroid nodule malignancy based on ultrasound parameters. TI-RADS level is a benchmark for evaluating the stage of the thyroid nodules, which aids in placing them in one of the five different stages. TIRADS-1 indicates the benign class, whereas TIRADS-5 indicates a significant risk of thyroid cancer.

2.2. Annotation

The ground truth (GT) mask associated with each thyroid ultrasound picture is required to differentiate the nodule part for the training procedure. The labelme [23] annotates the thyroid images before creating a polygon mask for each image. Figure 3 shows an example of the original image and its GT counterpart. The interpretations of ultrasound images by labelme software are saved as JSON files, comprised of polygon points for the nodule region attributed to 0 or 1. The pixels inside the enclosing polygon associated with the nodule region have a value of 1, while the remainder are considered background having a value of 0.

2.3. Training Methodology

Table 1 demonstrates the distribution of the dataset for training and testing purposes in the proposed method. The proposed VGG-16 model employs a stochastic gradient descent (SGD) optimizer and a cross-entropy loss function for training. In contrast, the benchmark U-Net model employed an Adadelta optimizer and a cross-entropy loss function and used the Keras [24] framework for implementation. The learning rate adjustment throughout the training phase ensures maximum training accuracy and less training loss. The ideal learning rate would result in a rapid decrease in training loss until it reaches the minimum level. Table 2 lists the details of the proposed and benchmark models’ training parameters, such as learning rate, drop rate, weight decay, and optimizer.

2.4. Proposed Convolutional Neural Network (CNN) Architecture

The extraction of the core features can be carried out automatically using the convolutional neural network filters (CNN). For better training results, the implementation and adjustment of weights are essential. The proposed model architecture and its detailed configuration about filter size, padding, stride, and pooling are explained in Table 3. The input image given to CNN has 712 × 712 × 1 size after applying the padding operation. Five convolutional layers process the input images with Relu functions and pooling layers. The first two layers of CNN contain a sequence of two convolutional layers with 64 and 128 filters, respectively, with 3 × 3 kernel size and 1 × 1 stride size. However, the last three convolutional layers for three convolution sequences contain 3 × 3 kernel size and 1 × 1 stride size and use 256 filters for the third layer and 512 filters for the fourth and fifth layers. Before passing through the sequence of two drop-out layers, a feature map of 23 × 23 × 512 is obtained. The size of the output feature map can be calculated by using the following formula
q h , q w , q r =   n h + 2 e k   s + 1 , n w + 2 e k   s + 1 ,   n k
n h and n w represent the height and width of the input image size, respectively, whereas k denotes the kernel size, while e , s ,   and   n k indicate the padding size, stride size, and the number of filters used, respectively. q h , q w , and   q r represent the output height, width, and channel number after each convolutional layer, respectively. Pool-1 to Pool-5 layers performed the max-pooling function (kernel size 2 × 2, and stride 2 × 2) to reduce the feature map size. The output size of the image after each pooling layer can be formulated as follows
d h , d w , d r =   n h + 2 e k   s + 1 , n w + 2 e k   s + 1 ,   n c
where n c is the number of channels of the input. d h ,   d w ,   and   d r are the output height, width, and number of channel after pooling layers, respectively.
The two drop-out layers consisted of 4096 filters, kernel size of 7 × 7 and stride size 1 × 1. Following the drop-out layers, a convolution layer was used to decrease the number of output channels with 1 × 1 kernel size and 1 × 1 stride size. To resize the feature maps into the same padding images and predict each pixel while keeping the spatial information intact in both the original images and the upsampled feature maps, a deconvolutional layer with the configuration of 64 × 64 kernel size and stride size 1 × 1 is used. A cropping operation was performed after the deconvolution layer to fit the input size.

3. Results

This Section describes the testing and validation performance of the proposed CNN model based on accuracy, precision, and other standard classification metrics.

3.1. Evaluation Metrics

The segmentation results are quantitatively evaluated using parameters such as accuracy (Acc), intersection-over-union (IoU), precision, recall, and dice score (DSC). The explanation of these parameters is as follows
A c c u r a c y   A c c = T P + T N T P + T N + F P + F N  
S e n s i t i v i t y   o r   R e c a l l   T P R = T P T P + F N  
P r e c i s i o n = T P T P + F P  
D i c e   S c o r e   D S C = 2 × T P 2 × T P + F N + F P  
I n t e r s e c t i o n - o v e r - U n i o n   I o U = A r e a   o f   O v e r l a p A r e a   o f   U n i o n = T P T P + F P + F N  
T P represents the true positive, T N is true negative, whereas F P   a n d   F N are the false positive and false negative, respectively.

3.2. Performance Evaluation Analysis

This research aims to create a deep learning framework that can recognize thyroid nodules in ultrasound images effectively. Despite the considerable variances in the ultrasound data, the proposed model outperformed U-Net quantitatively and produced highly accurate detection results. The testing results revealed that the proposed model achieved an overall accuracy of 99%, dice score of 97.5%, sensitivity of 98%, IoU of 97.1%, and precision of 97%. In comparison, the benchmark U-Net approach obtained an accuracy of 96%, a precision of 96%, sensitivity of 95.2%, dice score of 95.4%, and IoU of 95.3%. Table 4 shows the comparison of both methods and their respective parameters. The experimental findings show that the proposed method is very accurate, efficient, and reliable. Figure 4 shows the qualitative segmentation results of the proposed method and the benchmark method (U-Net), demonstrating that the proposed method can segment thyroid nodules following the reference standard. In contrast, the state-of-the-art benchmark method (U-Net) cannot detect in some cases.

3.3. System Description and Time Analysis

The training phase was completed on the hardware unit with a Core [email protected] GHz processor with 16 GB DDR4 RAM. The graphics card used was NVIDIA RTX 2070. Furthermore, our suggested method exceeds the benchmark method in terms of computing efficiency. According to the processing time, the suggested VGG-16 and benchmark algorithms consume 0.2 and 0.47 s for inference time per test image, respectively. Furthermore, existing research shows that the processing time of 47 s for the test image is insufficient for real-time findings.

3.4. Performance Comparison with Other State-of-the-Art Methods

Several researchers have addressed the thyroid nodule problem in their studies and proposed the best possible solution. Table 5 summarizes the performance of previous researches in comparison to our proposed technique. The results in this table show that our proposed study outperforms earlier studies in detecting thyroid nodules using ultrasound imaging.

4. Conclusions

Ultrasonic accurate segmentation of the thyroid nodule area is an indispensable prerequisite for the diagnosis of thyroid cancer. For this purpose, we developed a deep learning model that uses the VGG-16 framework as the backbone, which is extensively used in medicine for the automatic detection and segmentation of thyroid nodule images. We evaluated our method on a TDID challenging thyroid dataset having high noise, blurry boundaries, and no calipers. The experimental results showed that the proposed method outperformed the state-of-the-art U-Net model. The proposed segmentation network segmented the thyroid nodule accurately with an accuracy of 99% and provided more precise predictions. Although artificial intelligence will not replace physicians in the near years, clinical specialists can study the principles of AI innovation and how AI-based structures can assist them in giving more benefits to their patients at work. In general practice, our deep learning model could help endocrinologists by providing a second opinion throughout the diagnosing process.

Author Contributions

Conceptualization, H.A.U.R. and S.-F.S.; methodology, H.A.U.R.; validation, H.A.U.R.; formal analysis, H.A.U.R.; Project administration, S.-F.S.; Resources, S.-F.S.; Software, H.A.U.R.; Validation, H.A.U.R.; Visualization, H.A.U.R.; review and editing, S.-F.S.; supervision, C.-Y.L. and S.-F.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Abbad Ur Rehman, H.; Lin, C.Y.; Mushtaq, Z.; Su, S.F. Performance Analysis of Machine Learning Algorithms for Thyroid Disease. Arab. J. Sci. Eng. 2021. [Google Scholar] [CrossRef]
  2. Abbad Ur Rehman, H.; Lin, C.Y.; Mushtaq, Z. Effective K-Nearest Neighbor Algorithms Performance Analysis of Thyroid Disease. J. Chin. Inst. Eng. Trans. Chin. Inst. Eng. A 2021, 44, 77–87. [Google Scholar] [CrossRef]
  3. Fresilli, D.; David, E.; Pacini, P.; Del Gaudio, G.; Dolcetti, V.; Lucarelli, G.; Di Leo, N.; Bellini, M.; D’Andrea, V.; Sorrenti, S.; et al. Thyroid Nodule Characterization: How to Assess the Malignancy Risk. Update of the Literature. Diagnostics 2021, 11, 1374. [Google Scholar] [CrossRef] [PubMed]
  4. Kumar, V.; Webb, J.; Gregory, A.; Meixner, D.D.; Knudsen, J.M.; Callstrom, M.; Fatemi, M.; Alizad, A. Automated Segmentation of Thyroid Nodule, Gland, and Cystic Components from Ultrasound Images Using Deep Learning. IEEE Access 2020, 8, 63482–63496. [Google Scholar] [CrossRef]
  5. Siegel, R.L.; Mph, K.D.M.; Jemal, A. Cancer statistics, 2019. CA Cancer J. Clin. 2018, 69, 7–34. [Google Scholar] [CrossRef] [Green Version]
  6. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. Lect. Notes Comput. Sci. 2015, 9351, 234–241. [Google Scholar] [CrossRef] [Green Version]
  7. Ma, L.; Ma, C.; Liu, Y.; Wang, X. Thyroid Diagnosis from SPECT Images Using Convolutional Neural Network with Optimization. Comput. Intell. Neurosci. 2019, 2019, 6212759. [Google Scholar] [CrossRef] [Green Version]
  8. Nguyen, D.T.; Kang, J.K.; Pham, T.D.; Batchuluun, G.; Park, K.R. Ultrasound Image-Based Diagnosis of Malignant Thyroid Nodule Using Artificial Intelligence. Sensors 2020, 20, 1822. [Google Scholar] [CrossRef] [Green Version]
  9. Li, X.; Zhang, S.; Zhang, Q.; Wei, X.; Pan, Y.; Zhao, J.; Xin, X.; Qin, C.; Wang, X.; Li, J.; et al. Diagnosis of thyroid cancer using deep convolutional neural network models applied to sonographic images: A retrospective, multicohort, diagnostic study. Lancet Oncol. 2019, 20, 193–201. [Google Scholar] [CrossRef]
  10. Sharifi, Y.; Bakhshali, M.A.; Dehghani, T.; DanaiAshgzari, M.; Sargolzaei, M.; Eslami, S. Deep learning on ultrasound images of thyroid nodules. Biocybern. Biomed. Eng. 2021, 41, 636–655. [Google Scholar] [CrossRef]
  11. Prochazka, A.; Gulati, S.; Holinka, S.; Smutek, D. Classification of Thyroid Nodules in Ultrasound Images Using Direction-Independent Features Extracted by Two-Threshold Binary Decomposition. Technol. Cancer Res. Treat. 2019, 18, 1533033819830748. [Google Scholar] [CrossRef] [PubMed]
  12. Nugroho, H.A.; Frannita, E.L.; Ardiyanto, I.; Choridah, L. Computer aided diagnosis for thyroid cancer system based on internal and external characteristics. J. King Saud Univ. Comput. Inf. Sci. 2021, 33, 329–339. [Google Scholar] [CrossRef]
  13. Song, J.; Chai, Y.J.; Masuoka, H.; Park, S.-W.; Kim, S.-J.; Choi, J.Y.; Kong, H.-J.; Lee, K.E.; Lee, J.; Kwak, N.; et al. Ultrasound image analysis using deep learning algorithm for the diagnosis of thyroid nodules. Medicine 2019, 98, e15133. [Google Scholar] [CrossRef] [PubMed]
  14. Ko, S.Y.; Lee, J.H.; Yoon, J.H.; Na, H.; Hong, E.; Han, K.; Jung, I.; Kim, E.; Moon, H.J.; Park, V.Y.; et al. Deep convolutional neural network for the diagnosis of thyroid nodules on ultrasound. Head Neck 2019, 41, 885–891. [Google Scholar] [CrossRef] [PubMed]
  15. Liu, T.; Guo, Q.; Lian, C.; Ren, X.; Liang, S.; Yu, J.; Niu, L.; Sun, W.; Shen, D. Automated detection and classification of thyroid nodules in ultrasound images using clinical-knowledge-guided convolutional neural networks. Med. Image Anal. 2019, 58, 101555. [Google Scholar] [CrossRef] [PubMed]
  16. Vasile, C.; Udriștoiu, A.; Ghenea, A.; Popescu, M.; Gheonea, C.; Niculescu, C.; Ungureanu, A.; Udriștoiu, Ș.; Drocaş, A.; Gruionu, L.; et al. Intelligent Diagnosis of Thyroid Ultrasound Imaging Using an Ensemble of Deep Learning Methods. Medicine 2021, 57, 395. [Google Scholar] [CrossRef]
  17. Yang, W.; Dong, Y.; Du, Q.; Qiang, Y.; Wu, K.; Zhao, J.; Yang, X.; Zia, M.B. Integrate domain knowledge in training multi-task cascade deep learning model for benign–malignant thyroid nodule classification on ultrasound images. Eng. Appl. Artif. Intell. 2021, 98, 104064. [Google Scholar] [CrossRef]
  18. Abdolali, F.; Kapur, J.; Jaremko, J.L.; Noga, M.; Hareendranathan, A.R.; Punithakumar, K. Automated thyroid nodule detection from ultrasound imaging using deep convolutional neural networks. Comput. Biol. Med. 2020, 122, 103871. [Google Scholar] [CrossRef]
  19. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 640–651. [Google Scholar] [CrossRef]
  20. Falk, T.; Mai, D.; Bensch, R.; Çiçek, Ö.; Abdulkadir, A.; Marrakchi, Y.; Böhm, A.; Deubner, J.; Jäckel, Z.; Seiwald, K.; et al. U-Net: Deep learning for cell counting, detection, and morphometry. Nat. Methods 2019, 16, 67–70. [Google Scholar] [CrossRef]
  21. Pedraza, L.; Vargas, C.; Narváez, F.; Durán, O.; Muñoz, E.; Romero, E. An open access thyroid ultrasound image database. In Proceedings of the 10th International Symposium on Medical Information Processing and Analysis, Cartagena de Indias, Colombia, 28 January 2015; Volume 9287, pp. 1–6. [Google Scholar]
  22. Tessler, F.N.; Middleton, W.D.; Grant, E.G. Thyroid Imaging Reporting and Data System (TI-RADS): A User’s Guide. Radiology 2018, 287, 29–36. [Google Scholar] [CrossRef] [PubMed]
  23. Russell, B.C.; Torralba, A.; Murphy, K.P.; Freeman, W.T. LabelMe: A Database and Web-Based Tool for Image Annotation. Int. J. Comput. Vis. 2008, 77, 157–173. [Google Scholar] [CrossRef]
  24. Gupta, D.; Jhunjhunu, W.; Wala, R.; Juston, M.; Jaled, M.C. Image Segmentation Keras: Implementation of Segnet, FCN, UNet, PSPNet and Other Models in Keras. Available online: https://github.com/divamgupta/image-segmentation-keras (accessed on 20 October 2021).
  25. Pan, H.; Zhou, Q.; Latecki, L.J. SGUNET: Semantic Guided UNET For Thyroid Nodule Segmentation. In Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), Nice, France, 13–16 April 2021; pp. 630–634. [Google Scholar] [CrossRef]
  26. Wu, J.; Zhang, Z.; Zhao, J.; Qiang, Y. Ultrasound Image Segmentation of Thyroid Nodules Based on Joint Up-sampling. J. Phys. Conf. Ser. 2020, 1651, 012157. [Google Scholar] [CrossRef]
  27. Haji, S.O.; Yousif, R.Z. A Novel Neutrosophic Method for Automatic Seed Point Selection in Thyroid Nodule Images. BioMed Res. Int. 2019, 2019, 7632308. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Schematic block diagram of the proposed methodology.
Figure 1. Schematic block diagram of the proposed methodology.
Diagnostics 11 02209 g001
Figure 2. Proposed backbone VGG-16 architecture.
Figure 2. Proposed backbone VGG-16 architecture.
Diagnostics 11 02209 g002
Figure 3. Original image and its corresponding GT masks.
Figure 3. Original image and its corresponding GT masks.
Diagnostics 11 02209 g003
Figure 4. Segmentation output results of the proposed and benchmark methods.
Figure 4. Segmentation output results of the proposed and benchmark methods.
Diagnostics 11 02209 g004
Table 1. Distribution of dataset for Testing and Training.
Table 1. Distribution of dataset for Testing and Training.
Diagnosis ClassTrainingTestingTotal
TIRADS-18444128
TIRADS-2613293
TIRADS-3471966
TIRADS-4431154
TIRADS-5451459
Total280120400
Table 2. Network Parameters Detail.
Table 2. Network Parameters Detail.
ModelLearning RateDrop RateWeight Decay
Proposed Method1 × 10−90.50.0005
U-Net0.00010.20.0002
Table 3. Architecture and configuration of the proposed network.
Table 3. Architecture and configuration of the proposed network.
LayerFeature (Train)Feature (Inference)Kernel SizeStride
Input Image512 × 512 × 1512 × 512 × 1--
Padding712 × 712 × 1712 × 712 × 1--
Convolutional 1
Conv-1 + Relu-1710 × 710 × 64710 × 710 × 643 × 31
Conv-2 + Relu-2710 × 710 × 64710 × 710 × 643 × 31
Pool-1355 × 355 × 64355 × 355 × 642 × 22
Convolutional 2
Conv-1 + Relu-1355 × 355 × 128355 × 355 × 1283 × 31
Conv-2 + Relu-2355 × 355 × 128355 × 355 × 1283 × 31
Pool-2178 × 178 × 128178 × 178 × 1282 × 22
Convolutional 3
Conv-1 + Relu-1178 × 178 × 256178 × 178 × 2563 × 31
Conv-2 + Relu-2178 × 178 × 256178 × 178 × 2563 × 31
Conv-3 + Relu-3178 × 178 × 256178 × 178 × 2563 × 31
Pool-389 × 89 × 25689 × 89 × 2562 × 22
Convolutional 4
Conv-1 + Relu-189 × 89 × 51289 × 89 × 5123 × 31
Conv-2 + Relu-289 × 89 × 51289 × 89 × 5123 × 31
Conv-3 + Relu-389 × 89 × 51289 × 89 × 5123 × 31
Pool-445 × 45 × 51245 × 45 × 5122 × 22
Convolutional 5
Conv-1 + Relu-145 × 45 × 51245 × 45 × 5123 × 31
Conv-2 + Relu-245 × 45 × 51245 × 45 × 5123 × 31
Conv-3 + Relu-345 × 45 × 51245 × 45 × 5123 × 31
Pool-523 × 23 × 51223 × 23 × 5122 × 22
Conv-6 + Relu-6 + Drop617 × 17 × 409617 × 17 × 40967 × 71
Conv-7 + Relu-7 + Drop717 × 17 × 409617 × 17 × 40961 × 11
Conv-817 × 17 × 117 × 17 × 11 × 11
Deconv-9576 × 576 × 1576 × 576 × 164 × 6432
Cropping512 × 512 × 1512 × 512 × 1--
Output512 × 512 × 1512 × 512 × 1--
Table 4. Evaluation Analysis of the Proposed and benchmark method.
Table 4. Evaluation Analysis of the Proposed and benchmark method.
ParametersProposed MethodU-Net
Accuracy9996
Precision9796
Sensitivity9895.2
DSC97.595.4
IoU97.195.3
Table 5. Comparison of the proposed method with other existing studies.
Table 5. Comparison of the proposed method with other existing studies.
AuthorsMethodDatasetAccuracy (%)
Huitong et al. [25]SGUNETTDID93.6
Wu et al. [26]U-Net (backbone)Private Dataset93.19
Haji et al. [27]SSHOSTDID96
Abdolali et al. [18]Mask R-CNNPrivate Dataset84
Liu et al. [15]ResNet-50 (backbone)Private Dataset97.1
Nguyen et al. [8]ResNet + InceptionNetTDID92.05
Proposed MethodVGG-16 (backbone)TDID99
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rehman, H.A.U.; Lin, C.-Y.; Su, S.-F. Deep Learning Based Fast Screening Approach on Ultrasound Images for Thyroid Nodules Diagnosis. Diagnostics 2021, 11, 2209. https://doi.org/10.3390/diagnostics11122209

AMA Style

Rehman HAU, Lin C-Y, Su S-F. Deep Learning Based Fast Screening Approach on Ultrasound Images for Thyroid Nodules Diagnosis. Diagnostics. 2021; 11(12):2209. https://doi.org/10.3390/diagnostics11122209

Chicago/Turabian Style

Rehman, Hafiz Abbad Ur, Chyi-Yeu Lin, and Shun-Feng Su. 2021. "Deep Learning Based Fast Screening Approach on Ultrasound Images for Thyroid Nodules Diagnosis" Diagnostics 11, no. 12: 2209. https://doi.org/10.3390/diagnostics11122209

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop