Next Article in Journal
Up-to-Date Imaging and Diagnostic Techniques for Prostate Cancer: A Literature Review
Next Article in Special Issue
The Use of Artificial Intelligence in the Liver Histopathology Field: A Systematic Review
Previous Article in Journal
Bed Bugs (Hemiptera, Cimicidae): A Global Challenge for Public Health and Control Management
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance Analysis of Segmentation and Classification of CT-Scanned Ovarian Tumours Using U-Net and Deep Convolutional Neural Networks

by
Ashwini Kodipalli
1,*,
Steven L. Fernandes
2,*,
Vaishnavi Gururaj
3,
Shriya Varada Rameshbabu
4 and
Santosh Dasar
5
1
Department of Artificial Intelligence & Data Science, Global Academy of Technology, Bangalore 560098, India
2
Department of Computer Science, Design, Journalism, Creighton University, Omaha, NE 68178, USA
3
Department of Computer Science, George Mason University, Fairfax, VA 22030, USA
4
Department of Computer Science & Engineering, Global Academy of Technology, Bangalore 560098, India
5
Department of Radiologist, SDM College of Medical Sciences and Hospital, Dharwad 580009, India
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(13), 2282; https://doi.org/10.3390/diagnostics13132282
Submission received: 8 June 2023 / Revised: 29 June 2023 / Accepted: 3 July 2023 / Published: 5 July 2023

Abstract

:
Difficulty in detecting tumours in early stages is the major cause of mortalities in patients, despite the advancements in treatment and research regarding ovarian cancer. Deep learning algorithms were applied to serve the purpose as a diagnostic tool and applied to CT scan images of the ovarian region. The images went through a series of pre-processing techniques and, further, the tumour was segmented using the UNet model. The instances were then classified into two categories—benign and malignant tumours. Classification was performed using deep learning models like CNN, ResNet, DenseNet, Inception-ResNet, VGG16 and Xception, along with machine learning models such as Random Forest, Gradient Boosting, AdaBoosting and XGBoosting. DenseNet 121 emerges as the best model on this dataset after applying optimization on the machine learning models by obtaining an accuracy of 95.7%. The current work demonstrates the comparison of multiple CNN architectures with common machine learning algorithms, with and without optimization techniques applied.

1. Introduction

Ovarian cancer stands out as a commonly diagnosed type of cancer worldwide. Considering the fact that it usually goes unrecognised until it reaches terminal stages, ovarian cancer is a leading reason for high mortality rates among women as a gynaecological illness. Ranking fifth in deaths due to cancer among women, the risk of being diagnosed with ovarian cancer peaks between the ages of 55 and 64, on average [1]. Silent symptoms and undetermined causes act as major factors for late diagnosis and ineffective screening methods.
The American Cancer Society claims that around 19,710 women will be diagnosed with ovarian cancer and that around 13,270 deaths will occur from ovarian cancer in 2023 in the United States [2]. In the past few years, significant developments in the field of biomedical imaging have contributed to the domain of cancer detection. With interdisciplinary approaches being popularized to solve objectives, Medical Imaging can be combined with Machine Learning and Deep Learning disciplines to effectively detect and categorize tumours. Ultrasound and CT scan images contain large amounts of information, making them ideal to use in the case of the implementation of Deep Learning algorithms.
The symptoms are not only vague in nature, such as bloating, abdominal pain, fatigue, etc., but they are also noticeable only in the later stages. The lack of symptoms showing up early often leads to delayed medical examination of the subject and therefore late detection of the tumour in general. Also, unlike in a few other cancers, there is no ascertained screening process for ovarian cancer. The most common approach is to have the subject undergo a scanning process, usually a CT scan, where the radiologist would go through hundreds of images and determine if there is a tumour present or not. Not only is this time-consuming, but also prone to error due to possible false positives. All of the above reasons can be considered as the main challenges to detect ovarian tumours. The advent of technology like Deep Learning mechanisms can aid in the detection of a tumour by increasing its efficiency in terms of time and accuracy.
This paper aims to provide a comparative study of the detection and classification of ovarian tumours using Machine Learning and Deep Learning algorithms, using CT images of ovaries. Multiple ML models and CNN variants were used for this purpose and the comparison was carried out inter-categorically as well as intra-categorically. The Literature Survey section covers the latest developments and ongoing research, not only in the detection of ovarian cancer but also in how state-of-the-art Deep Learning algorithms are used in other medical scenarios.
The organization of the paper is as follows: Section 2 reports the literature survey of the latest research performed in the area of biomedical imaging and of several learning algorithms used. Section 3 demonstrates the methodology, along with the steps and the models used in the current work. Section 4 presents the experimental results and the discussion. Section 5 provides the concluding thoughts.

2. Literature Review

2.1. Medical Imaging Classification Using CNN

Jung et al. [3] used ultrasound images of the lower body region of females to remove unwanted information in the frame and classify the ovaries into five classes—normal, cystadenoma, mature cystic teratoma, endometrioma, and malignant tumour. They used a texture-based analysis for tumour detection and trained a convolutional autoencoder or CNN-CAE. The images before and after the autoencoder are both fed into CNNs such as Inception, ResNet and different variants of DenseNets. Weighted class activation mapping or Grad-CAM was used to visualise the result. It can be noted that the model classified better when unnecessary data were removed using CNN-CAE. DenseNet121 and DenseNet161 were the better performers amongst all the algorithms used when parameters like accuracy, sensitivity, specificity, and the area under the curve (AUC) were considered to be metrics of performance.
Wang et al. [4] used pelvic CT scan images to detect and segment out ovarian cancer tumours simultaneously, i.e., creating a multi-task deep learning model. They proposed a model called YOLO-OCv2, which was an enhancement of their previously proposed algorithm. Mosaic enhancement was also used here, in order to improve the background information of the object. However, the multitask model YOLO-OCv2 outperformed other algorithms like Faster-RCNN, SSD and RetinaNet, which were trained on the COCO dataset. In this work, Mahmood et al. [5] created a Nuclei segmentation model that could be used to segment out the nuclei in multiple locations of the body. The authors used Conditional Generative Adversarial Networks, or cGAN, as they can control the GAN training output depending on a class. The model was trained on synthetically generated data along with real data in order to make sure that sufficient input was present. The model was trained with data from nine organs and was tested on four organs, where it outperformed its peers, such as FCN, U-Net and Mask R-CNN. Guan et al. [6] used mammographic images to detect breast cancer using CNN models. The authors focused on Affine transformations and synthetic data generation using GANs.

2.2. Medical Imaging Classification Using Ensemble Deep Learning

According to Karimi et al. [7], a Vision Transformers (ViT) algorithm was proposed which divided images into Image Patches. The proposed algorithm using transformers did not use any convolution operations to segment the brain cortical plate and the hippocampus in MRI images of the brain. The results were compared with FCN architectures like 3D UNet++, Attention UNet, SE-FCN and DSRNet. The proposed network performed segmentation accurately when compared to the other, and with a significantly smaller number of labelled training images. Xu et al. [8] worked on histopathological whole-slide images (WSIs) to detect ovarian cancer using CNNs trained on images of multiple resolutions. The authors proposed a new modified version of ResNet50 called the Heatmap ResNet50 algorithm for CNN-based patch selection, and ResNet18 along with MR-ViT was used for ViT-based slide classification. Li et al. [9] introduced a variation of UNet known as CR-UNet to simultaneously segment out ovaries and follicles from transvaginal ultrasound (TVUS) images. The proposed model was then compared with models like DeepLabV3+, PSPNet-1, PSPNet-2 and U-Net to find that the proposed model outperformed them all. In the proposed work by Goodfellow et al. [10], an adversarial net framework was suggested that loosely resembles a minimax two-player game. Nagarajan et al. [11] and Zhao et al. [12], in their research work, provided three approaches that were used to classify ovarian cancer types using CT images. The first approach used a deep convolutional neural network (DCNN) based on AlexNet, which did not provide satisfactory results. The second approach had an overfitting problem. To overcome this, GAN was used in the third approach to augment the image samples along with the DCNN, which provided the best results out of the three approaches in metrics such as precision, recall, f-measure and accuracy. The research work of Saha et al. [13] included a novel 2D segmentation network called MU-net, which was a combination of MobileNetV2 and U-Net used to segment out follicles in ovarian ultrasound images. An USOVA3D Training Set 1 dataset was used. The proposed model was evaluated against several other models from previous works in the literature, and was shown to be more accurate, with an accuracy of 98.4%. Jin, J et al. [14], in their work, used four UNet models: U-net, U-net++, U-net with Resnet and CE-Net to perform automatic segmentation. In Thangamma et al. [15], the k-means algorithm and fuzzy c-means algorithm were used on ultrasound images of ovaries. It was concluded that the fuzzy c-means algorithm provided a better result than the k-means algorithm The work by Hema et al. [16] involved FaRe-ConvNN, which applied annotations on the image dataset, where the images had three categories: epithelial, germ and stroma cells. In order to avoid overfitting and other issues due to the small dataset size, image augmentation using image enhancement and transformation techniques like resizing, masking, segmentation, normalization, vertical or horizontal flips and rotation was undertaken. FaRe-ConvNN was used to compensate for manual annotation. After the region-based training in FaRe-ConvNN, a combination of SVC and Gaussian NB classifiers was used to classify the images, which resulted in impressive precision and recall values [17]. In the works carried out by Ashwini et al. [18,19,20], various Deep Learning models were used to segment the CT scanned images and classify them using variants of CNN. In the work [18,19], Otsu’s method was used to segment the tumour and a dice score of 0.82 and Jaccard score of 0.8356 were obtained. Further, to perform segmentation, cGAN was used [20] and, in this study, the segmentation and classification of tumours were carried out in a single pipeline, which obtained the dice score of 0.91 and the Jaccard score of 0.89. Similarly, in the works carried out by Fernandes et al. [21,22], according to the work [21], the authors proposed the segmentation of brain MRI images using entropy-based techniques. As per [22], the detection and classification of brain tumours by parallel processing was carried out using big data tools such as Kafka and PySpark.

2.3. Deep Learning in Medical Imaging Segmentation

Koonce et al. [23] shed light on EfficientNet, which comprised the inverted residual blocks of MobileNet v2 combined with the MnasNet architecture to form a robust model for performing Image Recognition. Rehman et al. [24] at BU-Net used a Residual Extended Skip (RES) block and a Wide Context (WC) block in a U-Net architecture to implement the proposed model, BU-Net, to segment Brain tumour cells in MRI scanned images. In the current work by Rehman et al. [25], the authors proposed a model named BrainSeg-Net to achieve the segmentation of tumour. The proposed model included a Feature Enhancer (FE) block at every encoder stage to protect critical information that could be tampered with during the convolution and transformation processes. Jalali et al. [26] proposed ResBCDU-Net for lung segmentation in CT images, which was used in applications such as in detecting lung cancer. To form the ResBCDU-Net, a pre-trained ResNet-34 network was used in place of an encoder in a typical U-Net model. The proposed method performed better than models like U-Net, RU-Net, ResNet34-UNet and BCDU-Net when measured using several evaluation metrics. Maureen et al. [27] and Neelima et al. [28] carried out an extensive review of bone image segmentation by considering the methods used in medical additive manufacturing. According to this review, global thresholding is the most commonly used method for segmentation and has obtained an accuracy of under 0.6 mm. Further, the authors have proposed using other advanced thresholding methods that may improve the accuracy to 0.38 mm. In the work carried out by Minnema et al. [29], the CNN-based STL method was applied for bone segmentation in CT scan images, which was able to accurately segment the skull and obtain a mean dice value of 0.92 ± 0.4. As per [30], a residual spatial pyramid pooling (RASPP) module was proposed to minimize the loss of location information in different modules. On similar lines, the work proposed by [31] optimized the CNN UNet model by applying it on a CT dataset generated from the MRI images. The results showed that the model performed well on the CT images when compared with the MRI images.

3. Methodology

3.1. Dataset Description

The current research work was carried out in collaboration with SDM Dharwad Hospital, Dharwad, Karnataka, India. The dataset used in the research work was obtained from SDM Dharwad College and Hospital. The entire work has been approved by the ethical approval committee of the hospital. The radiologist annotated the lesions, which acted as the ground truth for the study. The size of the dataset is as follows: there were 2560 benign and 2370 malignant images. For the complete study of the tumours, all three orientations of the images were considered—axial, coronal and sagittal views. The images were 2D.

3.2. Segmentation Using U-Net Model

U-Net is one among the oldest image segmentation models, first introduced in the paper U-Net: Convolutional Networks for Biomedical Image Segmentation. It comprises an encoder–decoder architecture for down-sampling and up-sampling, respectively. The U-Net [32] architecture is displayed in Figure 1.
The connections between encoder and decoder were the skip connection, which concatenated the encoder feature map with the decoder. This helped in the training process due to the backward flow of gradients. Image segmentation could be assumed as a combination of classification and localization tasks. The skip connections and decoder networks constituted the important aspects of the U-Net. The encoder network, also known as the contracting network, learned a feature map of the input layer. The task of the encoder was very similar to a classification task, in identifying the objects present in the image. In between the encoder and decoder networks lies the bottleneck layer, which comprised two convolutional layers followed by the ReLU activation layer. This layer produced the final feature map representation. The decoder network, also called the expansive network, took the feature map, as outputted by the bottleneck layer, as an input and outputted a segmentation mask. This was accomplished with the aid of skip connections. The task of the decoder is mainly to localize the object in the image. Skip connections, indicated with the grey arrows in the architecture, used contextual features learned from the encoder block and generated the segmentation map.

3.3. Detailed Methodology

Figure 2 describes the implementation flowchart. The CT scan images of the ovarian region collected were used as input images for the algorithms. These images underwent pre-processing techniques before they were fed into these models as input. Pre-processing includes a series of digital image processing techniques to support the segmentation of the input image. Intensity transformation was performed to obtain higher-quality images, by enhancing the pixel intensity of the image. Median filter was applied to reduce speckle-like patterns formed by noise in the CT scanned image and for better edge detection. Histogram equalization helped in highlighting fine details in the image to segment out the Region of Interest (ROI), which in our case was the ovarian tumour, with ease.
After undergoing the pre-processing steps, the images were fed as input to a UNet model for segmenting out the tumour in the CT scanned image. UNet [32] is a Convolutional Neural Network (CNN) architecture applied to medical images to perform segmentation tasks. It is mainly used to segment images with complex shapes and sizes, such as tumours. With its increasing popularity in biomedical imaging, it is also being used in other fields such as satellite imaging, etc. The segmented images were then used as input in several deep learning CNN variants and machine learning models for classification. The input instances were classified into two categories, namely, benign and malignant tumours, to detect whether the subject is at risk of Ovarian cancer or not. The CNN models implemented included the simple CNN model, ResNet 152, DenseNet 121 and Inception—ResNet v4. VGG16 and Xception. These are the state-of-the-art models. The reason for choosing smaller layers of these models was due to small dataset available for this research work. Machine learning models like Random Forest, Gradient Boost algorithm, AdaBoost and XGBoost were also used. The outcomes of all these algorithms were then compared.

4. Results and Discussions

The results section is divided into three parts. The first section describes the segmentation results using UNet, the second section describes the classification results using variants of CNN and the third section shows the comparison of the deep learning results with machine learning results.

4.1. Experimental Settings

Using bicubic interpolation, all the images were resized to 512 × 512 pixels. The batch size was set to 64 and every model was trained up to 1000 epochs. A final fully connected layer with ReLU activation function had 256 hidden neurons, followed by a dropout layer with a probability of 0.5 to prevent overfitting. Adam optimizers were used with the parameter values (beta 1 and beta 2) set to 0.6 and 0.8 and the learning rate as 0.0001. The last dense layers of all architecture were modified to output two classes corresponding to benign and malignant. All pre-trained CNN models were fine-tuned separately. The training and testing of the proposed architecture was implemented using Python using the keras package and run on Nvidia RTX 3060 GPU with 32 GB RAM.

4.2. Segmentation Results Using the UNet Model

The performance metrics used to evaluate the segmentation process were Dice and Jaccard scores. The loss function used in the UNet was Softmax.
Dice score: This metric was used to determine the similarity between two images. Here, two images refer to the ground truth and the segmented image. The equation for the Dice score is given below:
D i c e = 2 | S   G | | S | + | G |
where S indicates the segmented region that needs to be evaluated and G indicates the ground truth of the image. | | indicates the cardinality of the set. The Dice score always has a value between 0 and 1. The greater the value, the better the segmentation.
Jaccard score: This metric is used to calculate the overlap area between the segmented and the ground truth. The equation for the Jaccard score is given below:
J a c c a r d = d i c e 2 d i c e
The evaluated values of the Jaccard score range lie between 0 and 1; the greater the value, the better the segmentation results obtained.
Figure 3 presents the images of the input, ground truth (label) and the segmented images for benign images. It is observed from Table 1 that UNet performed well for the benign images, as the shape of the tumour was well-grown and the borders were clear. Table 1 has the 20 sample results from the testing set.
From Figure 3, it can be observed that the ground truth and the segmented images were very close and the same is reflected in the Dice and the Jaccard score tabulated in Table 1. The column “class0” indicated the background and the column “class1” indicated the tumour. The average Dice score was 0.998 ± 0.12 for class0 and 0.981 ± 0.19 for class1. The average Jaccard score was 0.995 ± 0.22 for class0 and 0.964 ± 0.20 for class1.
Figure 4 presents the images of the input, ground truth (label) and the segmented version of malignant images. It can be observed from Table 2 that UNet did not perform well in detecting malignant instances, as the shape of the tumour was very uneven and small and the borders were very uncertain due to the characteristics of the malignant tumours. Table 2 has 20 sample results from the testing set.
From Figure 4, it can be observed, from the malignant image in the first row, that the tumour size was very small and that the pixel intensity is much less. Due to size and intensity, the UNet model performance was low, with the dice and Jaccard score as follows: class0_dice was 0.98 and class1_dice was 0.87, and class0_jaccard was 0.99 and class1_jaccard was 0.84.
On similar lines, in the malignant image present in the middle row, we observed that the tumour had grown fully and the intensity of the pixel was high. Due to this, UNet performed well, with the Dice and Jaccard scores as follows: class0_dice was 0.996 and class1_dice was 0.950, and class0_jaccard was 0.993 and class1_jaccard was 0.926. The malignant image in the last row shared a similar characteristic to the first row image. The tumour was very uncertain with respect to size, shape and border. Hence, lower values were obtained in the segmentation results.
Table 2 shows 20 sample image performances from the test dataset. Due to the nature of the malignant tumours’ size, shape, border and intensity of the pixels, the performance of UNet on malignant tumours was overall less when compared to the performance on benign tumours.
The mean of Table 1 and Table 2 is shown in Figure 5. From the graph below, depicted in Figure 6, we can see that UNet performed well on benign images when compared to malignant images. The dice score range for benign images was between 0.992 and 0.998, whereas the dice score for the malignant category ranged between 0.70 and 0.91. Figure 7 shows the Jaccard score comparison for benign and malignant images.

4.3. Classification Results Using Variants of the CNN Model

Since the current research work is in the medical domain, it is extremely important to measure TP, TN, FP and FN and take appropriate measures for the necessary treatment. The different performance metrics used to analyse the classification results in our experiment were accuracy, precision, recall and F1 score. The metrics were expressed mathematically, as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N × 100
P r e c i s i o n = T P T P + F P × 100
R e c a l l = T P T P + F N × 100
F 1   s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
From Table 3 and Figure 8, it was observed that DenseNet121 outperformed the other CNN models used. Owing to the connectivity patterns in the DenseNet architecture, the information would not be lost or vanished by the time it reached the last layer.
The results were then compared with ensemble machine learning models using the same data, for classification. From Table 4, it can be noted that AdaBoosting outperformed the other ensemble machine learning models. From Table 5 and Figure 9, it can be observed that the performance of the learning algorithms was improved with the tuning techniques.
The different optimization techniques used for fine tuning the parameters were HyperOpt, Optuna and Multi-Fidelity Optimization. After fine tuning the parameters, the machine learning algorithms’ performances had improved.
HyperOpt works as follows: When compared with other hyperparameter techniques like randomised search and grid search, HyperOpt reduces the number of trials to find the best parameters by defining the search algorithm which selects the best input values from each new iteration. The key feature of the Optuna hyperparameter is its automated search for optimized hypermeters even in the large search space by pruning the unpromising trails for faster results. This technique also has a unique characteristic of parallel search, which makes searching faster. The multi-fidelity technique is well suited when the model is very large or when the size of the dataset is very large.
After fine tuning the parameters, the improved results using ML algorithms are as follows.

5. Conclusions and Discussion

Ovarian cancer is one of the most dangerous diseases found in women. An alarming number of deaths are caused every year due to the diagnosis of ovarian cancer only in the third and fourth stages. Advancements in the field of deep learning have helped to solve several issues in the medical field, as well. Although a completely dependable solution, we can find a significant amount of work carried out in helping doctors and radiologists to detect tumours.
The findings of this work can be generalised and extended to other clinical settings, but only with datasets having similar characteristics, for example, to detect the presence of tumours in other organs of the body, like breast mammographic, whose images detect breast cancer, or MRI or CT scans to segment follicles or parts of the brain using MRI images, nuclei cell segmentation, etc., but only after undergoing the process of fine tuning of parameters. Regarding the concept of transfer learning, deep learning models can increase the performance. By means of transfer learning, the efficiency of the models implemented in other cases can be improved, as it would leverage the knowledge gained by being trained on a larger and more diverse dataset. Using ensemble architectures could also improve the accuracy and robustness of the model. However, it is important to note that there are a few limitations to extending the work. The type of data, the objective to be achieved and the environment of the experiment could be very different from the current work. The resolution of the images, noise levels and parameters considered in consolidating the dataset also affects the performance. The type of model used in the experiment would also be one of the factors.
These computer screening methods take CT scan images of the ovarian region as input and perform pre-processing, segmentation and classification on them, using various algorithms. In the current work, multiple algorithms in machine learning, as well as in deep learning fields, have been implemented and compared. Segmentation was carried out using the UNet model. A total of 20 images were sampled from both the categories—benign and malignant—on which the UNet algorithm was applied. The performance was measured using the Jaccard score and Dice score. It was found that the model performed better when segmenting benign tumours. The UNet model did not perform well for the extremely small tumours, as seen in the malignant results. Therefore, more complexed model such as Transformers can be tried for the segmentation of smaller and uncertain tumours.
Deep learning models include several variants of CNNs such as CNN, ResNet 152, DenseNet 121, Inception-ResNet V4, VGG16 and Xception. The evaluation metrics used are accuracy, precision, recall and F1-score. It was observed that DenseNet 121 surpasses the other CNNs in all the metrics, while scoring an accuracy of 95.70%. The machine learning algorithms involved were Random Forest, Gradient Boosting, AdaBoosting and XGBoosting. The same metrics used above were used to evaluate these models, as well. Values obtained clearly showed that AdaBoosting outperformed the other Machine Learning algorithms considered, scoring an accuracy of 89.40% and leading the other metrics as well.
To boost the performance of the machine learning models, hyperparameter tuning was performed using HyperOpt, Optuna and Multi-Fidelity Optimization techniques. The resulting values were significantly better than those sans optimization. However, AdaBoosting remained the best performing technique among the ML models considered. We noted that deep learning architectures performed more efficiently than machine learning models with optimization techniques when classifying benign and malignant tumours in the CT scanned images, which proves them to be an aiding tool.
This experiment can be used as a Computer Aided Detection (CAD) system which helps doctors and radiologists by narrowing the region of the tumour present, segmenting it precisely and classifying the tumour into benign and malignant categories. The person in charge could use this as an assisting tool, as it would reduce effort and time while increasing accuracy.

Author Contributions

Methodology, A.K. and V.G.; Validation, S.L.F. and S.D.; Formal analysis, A.K.; Investigation, A.K.; Writing—original draft, A.K.; Writing—review & editing, S.L.F., V.G. and S.V.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in collaboration with SDM College of Medical Sciences and Hospitals, and approved by the Hospital Ethical Approval Committee on 18 March 2021, approval code: Ref: SDMCMS&H/Principal Office/061/2021.

Informed Consent Statement

The work is carried out in collaboration with SDM College of Medical Sciences and Hospital. The dataset is obtained from the hospital with the proper ethical approvals from the hospital authority.

Data Availability Statement

Dataset is available on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Labidi-Galy, S.I.; Treilleux, I.; Goddard-Leon, S.; Combes, J.-D.; Blay, J.-Y.; Ray-Coquard, I.; Caux, C.; Bendriss-Vermare, N. Plasmacytoid dendritic cells infiltrating ovarian cancer are associated with poor prognosis. Oncoimmunology 2012, 1, 380–382. [Google Scholar] [CrossRef] [Green Version]
  2. Siegel, R.L.; Miller, K.D.; Wagle, N.S.; Jemal, A. Cancer statistics. CA Cancer J. Clin. 2023, 73, 17–48. [Google Scholar] [CrossRef]
  3. Jung, Y.; Kim, T.; Han, M.R.; Kim, S.; Kim, G.; Lee, S.; Choi, Y.J. Ovarian tumour diagnosis using deep convolutional neural networks and a denoising convolutional autoencoder. Sci. Rep. 2022, 12, 17024. [Google Scholar] [CrossRef]
  4. Wang, X.; Li, H.; Zheng, P. Automatic Detection and Segmentation of Ovarian Cancer Using a Multitask Model in Pelvic CT Images. Oxidative Med. Cell. Longev. 2022, 2022, 6009107. [Google Scholar] [CrossRef]
  5. Mahmood, F.; Borders, D.; Chen, R.J.; Mckay, G.N.; Salimian, K.J.; Baras, A.; Durr, N.J. Deep Adversarial Training for Multi-Organ Nuclei Segmentation in Histopathology Images. IEEE Trans. Med. Imaging 2019, 39, 3257–3267. [Google Scholar] [CrossRef]
  6. Guan, S.; Loew, M. Breast cancer detection using synthetic mammograms from generative adversarial networks in convolutional neural networks. J. Med. Imaging 2019, 6, 031411. [Google Scholar] [CrossRef] [PubMed]
  7. Karimi, D.; Dou, H.; Gholipour, A. Medical Image Segmentation Using Transformer Networks. IEEE Access 2022, 10, 29322–29332. [Google Scholar] [CrossRef] [PubMed]
  8. Xu, T.; Farahani, H.; Bashashati, A. Multi-Resolution Vision Transformer for Subtype Classification in Ovarian Cancer Whole-Slide Histopathology Images. 2022. Available online: http://hdl.handle.net/2429/81390 (accessed on 2 February 2023).
  9. Li, H.; Fang, J.; Liu, S.; Liang, X.; Yang, X.; Mai, Z.; Van, M.T.; Wang, T.; Chen, Z.; Ni, D. CR-Unet: A Composite Network for Ovary and Follicle Segmentation in Ultrasound Images. IEEE J. Biomed. Health Inform. 2019, 24, 974–983. [Google Scholar] [CrossRef] [PubMed]
  10. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014, arXiv:1406.2661v1. [Google Scholar] [CrossRef]
  11. Nagarajan, P.H.; Tajunisha, N. Automatic Classification of Ovarian Cancer Types from CT Images Using Deep Semi-Supervised Generative Learning and Convolutional Neural Network. Rev. D’intelligence Artif. 2021, 35, 273–280. [Google Scholar] [CrossRef]
  12. Zhao, Q.; Lyu, S.; Bai, W.; Cai, L.; Liu, B.; Wu, M.; Sang, X.; Yang, M.; Chen, L. A Multi-Modality Ovarian Tumour Ultrasound Image Dataset for Unsupervised Cross-Domain Semantic Segmentation. arXiv 2022, arXiv:2207.06799. [Google Scholar]
  13. Saha, D.; Mandal, A.; Ghosh, R. MU Net: Ovarian Follicle Segmentation Using Modified U-Net Architecture. Int. J. Eng. Adv. Technol. 2022, 11, 30–35. [Google Scholar] [CrossRef]
  14. Jin, J.; Zhu, H.; Zhang, J.; Ai, Y.; Zhang, J.; Teng, Y.; Xie, C.; Jin, X. Multiple U-Net-Based Automatic Segmentations and Radiomics Feature Stability on Ultrasound Images for Patients With Ovarian Cancer. Front. Oncol. 2021, 10, 614201. [Google Scholar] [CrossRef] [PubMed]
  15. Thangamma, N.G.; Prasanna, D.S. Analyzing ovarian tumour and cancer cells using image processing algorithms K means & fuzzy C-means. Int. J. Eng. Technol. 2018, 7, 510–512. [Google Scholar]
  16. Hema, L.K.; Manikandan, R.; Alhomrani, M.; Pradeep, N.; Alamri, A.S.; Sharma, S.; Alhassan, M. Region-Based Segmentation and Classification for Ovarian Cancer Detection Using Convolution Neural Network. Contrast Media Mol. Imaging 2022, 2022, 5968939. [Google Scholar] [CrossRef]
  17. Ahamad, M.; Aktar, S.; Uddin, J.; Rahman, T.; Alyami, S.A.; Al-Ashhab, S.; Akhdar, H.F.; Azad, A.; Moni, M.A. Early-Stage Detection of Ovarian Cancer Based on Clinical Data Using Machine Learning Approaches. J. Pers. Med. 2022, 12, 1211. [Google Scholar] [CrossRef]
  18. Kodipalli, A.; Guha, S.; Dasar, S.; Ismail, T. An inception-ResNet deep learning approach to classify tumours in the ovary as benign and malignant. Expert Syst. 2022, e13215. [Google Scholar] [CrossRef]
  19. Kodipalli, A.; Devi, S.; Dasar, S.; Ismail, T. Segmentation and classification of ovarian cancer based on conditional adversarial image to image translation approach. Expert Syst. 2022, e13193. [Google Scholar] [CrossRef]
  20. Ruchitha, P.J.; Sai, R.Y.; Kodipalli, A.; Martis, R.J.; Dasar, S.; Ismail, T. Comparative analysis of active contour random walker and watershed algorithms in segmentation of ovarian cancer. In Proceedings of the 2022 International Conference on Distributed Computing, VLSI, Electrical Circuits and Robotics (DISCOVER), Shivamogga, India, 14–15 October 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 234–238. [Google Scholar]
  21. Liaqat, A.; Khan, M.A.; Shah, J.H.; Sharif, M.; Yasmin, M.; Fernandes, S.L. Automated ulcer and bleeding classification from WCE images using multiple features fusion and selection. J. Mech. Med. Biol. 2018, 18, 1850038. [Google Scholar] [CrossRef]
  22. Fernandes, S.L.; Tanik, U.J.; Rajinikanth, V.; Karthik, K.A. A reliable framework for accurate brain image examination and treatment planning based on early diagnosis support for clinicians. Neural Comput. Appl. 2020, 32, 15897–15908. [Google Scholar] [CrossRef]
  23. Koonce, B. EfficientNet. In Convolutional Neural Networks with Swift for Tensorflow; Apress: Berkeley, CA, USA, 2021. [Google Scholar] [CrossRef]
  24. Rehman, M.U.; Cho, S.; Kim, J.H.; Chong, K.T. Bu-net: Brain tumour segmentation using modified u-net architecture. Electronics 2020, 9, 2203. [Google Scholar] [CrossRef]
  25. Rehman, M.U.; Cho, S.; Kim, J.; Chong, K.T. Brainseg-net: Brain tumour mr image segmentation via enhanced encoder–decoder network. Diagnostics 2021, 11, 169. [Google Scholar] [CrossRef] [PubMed]
  26. Jalali, Y.; Fateh, M.; Rezvani, M.; Abolghasemi, V.; Anisi, M.H. ResBCDU-Net: A deep learning framework for lung CT image segmentation. Sensors 2021, 21, 268. [Google Scholar] [CrossRef] [PubMed]
  27. van Eijnatten, M.; van Dijk, R.; Dobbe, J.; Streekstra, G.; Koivisto, J.; Wolff, J. CT image segmentation methods for bone used in medical additive manufacturing. Med. Eng. Phys. 2018, 51, 6–16. [Google Scholar] [CrossRef]
  28. Neelima, G.; Chigurukota, D.R.; Maram, B.; Girirajan, B. Optimal DeepMRSeg based tumour segmentation with GAN for brain tumour classification. Biomed. Signal Process. Control 2022, 74, 103537. [Google Scholar] [CrossRef]
  29. Minnema, J.; van Eijnatten, M.; Kouw, W.; Diblen, F.; Mendrik, A.; Wolff, J. CT image segmentation of bone for medical additive manufacturing using a convolutional neural network. Comput. Biol. Med. 2018, 103, 130–139. [Google Scholar] [CrossRef] [Green Version]
  30. Rehman, M.U.; Ryu, J.; Nizami, I.F.; Chong, K.T. RAAGR2-Net: A brain tumour segmentation network using parallel processing of multiple spatial frames. Comput. Biol. Med. 2023, 152, 106426. [Google Scholar] [CrossRef]
  31. Islam, K.T.; Wijewickrema, S.; O’leary, S. A deep learning framework for segmenting brain tumours using MRI and synthetically generated CT images. Sensors 2022, 22, 523. [Google Scholar] [CrossRef]
  32. Rukundo, O. Effects of Image Size on Deep Learning. Electronics 2023, 12, 985. [Google Scholar] [CrossRef]
Figure 1. U-Net model architecture for segmentation of benign and malignant tumours.
Figure 1. U-Net model architecture for segmentation of benign and malignant tumours.
Diagnostics 13 02282 g001
Figure 2. Flow of the implementation.
Figure 2. Flow of the implementation.
Diagnostics 13 02282 g002
Figure 3. Benign dataset. Towards the (left) is the input image, the (middle) is the label and the (right) is the segmented image.
Figure 3. Benign dataset. Towards the (left) is the input image, the (middle) is the label and the (right) is the segmented image.
Diagnostics 13 02282 g003aDiagnostics 13 02282 g003b
Figure 4. Malignant dataset. Towards the (left) is the input image, the (middle) is the label and the (right) is the segmented image.
Figure 4. Malignant dataset. Towards the (left) is the input image, the (middle) is the label and the (right) is the segmented image.
Diagnostics 13 02282 g004
Figure 5. Mean plot of benign and malignant performance.
Figure 5. Mean plot of benign and malignant performance.
Diagnostics 13 02282 g005
Figure 6. Dice score comparison between benign and malignant tumours.
Figure 6. Dice score comparison between benign and malignant tumours.
Diagnostics 13 02282 g006
Figure 7. Jaccard score comparison between benign and malignant tumours.
Figure 7. Jaccard score comparison between benign and malignant tumours.
Diagnostics 13 02282 g007
Figure 8. Diagrammatic representation of the performance of the CNN models.
Figure 8. Diagrammatic representation of the performance of the CNN models.
Diagnostics 13 02282 g008
Figure 9. Diagrammatic representation of the comparison of the results with tuning techniques.
Figure 9. Diagrammatic representation of the comparison of the results with tuning techniques.
Diagnostics 13 02282 g009
Table 1. Dice score and Jaccard score for benign tumour.
Table 1. Dice score and Jaccard score for benign tumour.
ImageBenign
Class0_DiceClass1_DiceClass0_JaccardClass1_Jaccard
CT_10.997560.981930.995140.96451
CT_20.997750.989750.995520.97972
CT_30.997580.987100.995180.97454
CT_40.997380.989740.994790.97969
CT_50.997770.973340.995560.94807
CT_60.997010.983480.994040.96751
CT_70.997610.972690.995230.94684
CT_80.997290.990240.994590.98068
CT_90.998950.987400.997900.97512
CT_100.998820.928170.997650.86597
CT_110.997860.974700.995770.95066
CT_120.998140.979560.996370.95994
CT_130.998620.968640.997240.93920
CT_140.997180.957980.994390.91935
CT_150.994600.885550.989260.89461
CT_160.996960.967090.993950.93623
CT_170.996860.966780.993750.93569
CT_180.997770.969980.995530.94171
CT_190.996430.941710.992900.88984
CT_200.997240.969650.994500.94101
Table 2. Dice score and Jaccard score for malignant tumour.
Table 2. Dice score and Jaccard score for malignant tumour.
ImageMalignant
Class0_DiceClass1_DiceClass0_JaccardClass1_Jaccard
CT_10.994850.9172090.9897530.847079
CT_20.9965330.9507310.9930910.906089
CT_30.9948430.9127690.9897380.839535
CT_40.9964120.8816440.9928490.788339
CT_50.9985030.91790.997010.848258
CT_60.996610.9516910.9932420.907835
CT_70.9977970.9275980.9956030.864972
CT_80.9975020.9696510.9950170.941091
CT_90.9966530.9619370.9933280.926665
CT_100.9952130.9266770.9904710.863373
CT_110.9835320.8362380.9675980.718565
CT_120.9951180.8667930.9902840.764902
CT_130.9990530.9667290.9981070.9356
CT_140.9953170.8617020.9906770.757008
CT_150.9970930.9525540.9942030.909406
CT_160.9971170.9652290.994250.932794
CT_170.9954570.7272090.9909540.57135
CT_180.9861630.7016140.9727040.540374
CT_190.9984940.9145190.9969920.842501
CT_200.9987750.9182280.9975520.848818
Table 3. Performance of various CNN in classifying ovarian tumours.
Table 3. Performance of various CNN in classifying ovarian tumours.
Sl. NoCNN ArchitecturesAccuracyPrecisionRecallF1 Score
1.CNN89.7%89.1%88.0%89.4%
2.ResNet 15292.7%92.4%91.5%92.7%
3.DenseNet12195.7%95.2%94.3%95.6%
4.Inception-ResNet V494.3%94.1%93.2%94.2%
5.VGG 1691.5%91.6%90.5%91.4%
6.Xception87.2%87.3%86.7%86.3%
Table 4. Performance of the ensemble learning model for classification.
Table 4. Performance of the ensemble learning model for classification.
Sl. NoEnsemble ML ModelsAccuracyPrecisionRecallF1 Score
1.Random Forest82.4%82.7%83.6%81.5%
2.Gradient boosting80.9%80.3%80.88%79.5%
3.AdaBoosting89.4%88.5%89.1%87.6%
4.XGBoosting79.5%80.2%79.8%80.3%
Table 5. Improved performance with hyperparameter tuning techniques.
Table 5. Improved performance with hyperparameter tuning techniques.
Sl. NoEnsemble ML ModelsAccuracyPrecisionRecallF1 Score
1.Random Forest86.3%86.3%87.9%86.2%
2.Gradient boosting82.7%82.7%83.7%81.3%
3.AdaBoosting91.37%91.87%90.3%91.78%
4.XGBoosting80.34%81.67%80.48%81.78%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kodipalli, A.; Fernandes, S.L.; Gururaj, V.; Varada Rameshbabu, S.; Dasar, S. Performance Analysis of Segmentation and Classification of CT-Scanned Ovarian Tumours Using U-Net and Deep Convolutional Neural Networks. Diagnostics 2023, 13, 2282. https://doi.org/10.3390/diagnostics13132282

AMA Style

Kodipalli A, Fernandes SL, Gururaj V, Varada Rameshbabu S, Dasar S. Performance Analysis of Segmentation and Classification of CT-Scanned Ovarian Tumours Using U-Net and Deep Convolutional Neural Networks. Diagnostics. 2023; 13(13):2282. https://doi.org/10.3390/diagnostics13132282

Chicago/Turabian Style

Kodipalli, Ashwini, Steven L. Fernandes, Vaishnavi Gururaj, Shriya Varada Rameshbabu, and Santosh Dasar. 2023. "Performance Analysis of Segmentation and Classification of CT-Scanned Ovarian Tumours Using U-Net and Deep Convolutional Neural Networks" Diagnostics 13, no. 13: 2282. https://doi.org/10.3390/diagnostics13132282

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop