Next Article in Journal
Contemporary Role of Positron Emission Tomography (PET) in Endocarditis: A Narrative Review
Previous Article in Journal
Fully Automated Measurement of Cobb Angles in Coronal Plane Spine Radiographs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Ovarian Tumor Diagnosis: Performance of Convolutional Neural Networks in Classifying Ovarian Masses Using Ultrasound Images

by
Maria Giourga
1,*,
Ioannis Petropoulos
2,
Sofoklis Stavros
3,
Anastasios Potiris
3,
Angeliki Gerede
4,
Ioakeim Sapantzoglou
1,
Maria Fanaki
1,
Eleni Papamattheou
1,
Christina Karasmani
1,
Theodoros Karampitsakos
3,
Spyridon Topis
3,
Athanasios Zikopoulos
3,
Georgios Daskalakis
1 and
Ekaterini Domali
1
1
1st Department of Obstetrics and Gynecology, National and Kapodistrian University of Athens, 11528 Athens, Greece
2
School of Electrical & Computer Engineering, National Technical University of Athens, 15772 Athens, Greece
3
Third Department of Obstetrics and Gynecology, University Hospital “ATTIKON”, Medical School of the National and Kapodistrian University of Athens, 12462 Athens, Greece
4
Department of Obstetrics and Gynecology, University of Thrace, 68100 Alexandroupolis, Greece
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2024, 13(14), 4123; https://doi.org/10.3390/jcm13144123
Submission received: 29 May 2024 / Revised: 29 June 2024 / Accepted: 5 July 2024 / Published: 15 July 2024
(This article belongs to the Section Obstetrics & Gynecology)

Abstract

:
Background/Objectives: This study aims to create a strong binary classifier and evaluate the performance of pre-trained convolutional neural networks (CNNs) to effectively distinguish between benign and malignant ovarian tumors from still ultrasound images. Methods: The dataset consisted of 3510 ultrasound images from 585 women with ovarian tumors, 390 benign and 195 malignant, that were classified by experts and verified by histopathology. A 20% to80% split for training and validation was applied within a k-fold cross-validation framework, ensuring comprehensive utilization of the dataset. The final classifier was an aggregate of three pre-trained CNNs (VGG16, ResNet50, and InceptionNet), with experimentation focusing on the aggregation weights and decision threshold probability for the classification of each mass. Results: The aggregate model outperformed all individual models, achieving an average sensitivity of 96.5% and specificity of 88.1% compared to the subjective assessment’s (SA) 95.9% sensitivity and 93.9% specificity. All the above results were calculated at a decision threshold probability of 0.2. Notably, misclassifications made by the model were similar to those made by SA. Conclusions: CNNs and AI-assisted image analysis can enhance the diagnosis and aid ultrasonographers with less experience by minimizing errors. Further research is needed to fine-tune CNNs and validate their performance in diverse clinical settings, potentially leading to even higher sensitivity and overall accuracy.

1. Introduction

Ovarian cancer is one of the leading causes of cancer-related deaths among women worldwide. It is calculated as the deadliest gynecological cancer with the highest morbidity and mortality, accounting for more than 12,700 deaths each year [1]. It is often called a silent killer due to the lack of symptoms, especially in early stages, leading to late diagnosis, thus limiting treatment options [2]. This highlights the importance of early detection and immediate referral of such patients in tertiary oncology centers [3]. The most effective and easily accessible tool for the categorization of ovarian masses is ultrasonography [4]. Subjective assessment (SA) by expert ultrasonographers achieves the highest sensitivity and specificity rates, although there is a shortage of experts in this field [5]. Proposed models, such as Data System Ultrasound (ORADS-US), and models of the International Ovarian Tumor Analysis (IOTA), such as Simple Rules Risk (SRR), Risk of Malignancy Algorithm (RMI), logistic regression models LR1 and LR2, and Assessment of Different NEoplasias in the adneXa (ADNEX model), are widely used internationally and can achieve high rates of overall accuracy [6,7,8,9,10,11]. Although these tools have been validated externally and can achieve comparable results with subjective assessment, their performance depends on the examiner [4,12]. This fact underscores the critical role of human expertise in achieving the optimal diagnostic accuracy, though the risk of human error remains. According to published and ongoing research, an addition to the medical field is the use of artificial intelligence and deep learning, which has the potential to enhance diagnostic precision, decision-making, and limit human error [13,14,15,16]. Convolutional neural networks (CNNs), a subset of deep learning algorithms, offer automated image analysis that can extract complex features and patterns of medical images, such as X-rays, MRI, CT, and ultrasound images [17,18,19].
Researchers have already utilized pre-trained CNNs and reported promising results equivalent to experienced sonographers’ subjective assessment [14,17,20,21,22,23,24]. However, using neural networks in clinical practice necessitates substantial validation and the creation of large datasets of annotated ovarian images that can be utilized to train the models appropriately. The fusion of CNNs and medical imaging has the potential to transform everyday clinical practice and ovarian tumor diagnosis [13].

2. Materials and Methods

2.1. Dataset

We utilized a comprehensive dataset comprising still ultrasound images from patients with ovarian tumors treated in our tertiary gynecological oncology center from 2011 to 2023. The dataset consisted of 8324 collected ultrasound images from 585 women with ovarian tumors. Out of these masses, 390 were benign, and 195 were malignant ovarian tumors. For training and testing the model, 3510 images were used in total by choosing 6 representative still ultrasound images from each patient. Out of these images, 539 had calipers. The cases were selected from our center archives to represent different histopathology types. There was a slight imbalance in our dataset since 66.7% of the tumors were benign while the remaining 33.3% were malignant. This difference is reasonable, as it depicts the actual probability of malignancy in patients presenting with ovarian cysts [25]. The explant pathology report was used as a reference diagnosis and the masses were classified according to the International Federation of Gynecology and Obstetrics (FIGO) classifications for ovarian cancer staging [26]. The histopathology of the ovarian masses used can be found in Table 1.

2.2. Ultrasound Assessment

Based on our center protocol, women with ovarian lesions undergo an ultrasound examination preoperatively. Patients presenting with ovarian tumors that underwent ultrasound assessment up to 120 days before surgery were eligible for inclusion in the study. All ultrasound examinations were performed according to IOTA rules and definitions by 5 individual gynecologists with 6 to 21 years of experience. All ultrasound examiners are IOTA-certified and they perform more than 600 ultrasound examinations per year. The images used in our study were collected by 6 different ultrasound machines (General Electric HealthCare LOGIQ P6, Voluson S6, S8, E6, E10, and Samsung HS40, Athens, Greece). All cases were prospectively classified by our experts, while being blind to histopathology reports. They categorized the masses as benign, malignant, and inconclusive.

2.3. Data Processing

After the extraction from the ultrasound system, the images were cropped to the region of interest (ROI). A combinational approach was employed for this task, which utilized both manual and auto-cropping. The simple algorithm that was developed, after deidentifying the images by removing edge areas, isolated the ovaries and excluded unwanted structures (uterus, cervix, bladder, etc.). The key idea was to use pixel value comparisons and assume most of the ultrasound images had similar structures, meaning that deviations from the expected structure would translate into unwanted regions that were removed. Although the results had to be manually reviewed afterwards, the margin of error was low enough to deem the algorithm effective. The input layer of each CNN model has a shape of (150, 150) and ReLU activation. The images were thus resized accordingly, into 150-by-150 pixel-sized square images. Since the original images were in RGB format (three color channels, size of x, y, 3), a flattened layer was also added before passing the images as input to the model. The pixel values were also normalized over the entire dataset. Data augmentation was employed to artificially increase the dataset size and prevent model overfitting. Several geometric transformations were applied, including image flipping, rotation, shearing, and zoom. The color properties of the images were also modified by using brightness and contrast adjustments. After experimentation, noise injection was not incorporated, as the results showed it hindered the overall system performance.

2.4. Model Building

The core principle we adhered to for building the CNN models was transfer learning. Three different pre-trained models from ImageNet were used as building blocks for our final classifier system, VGG16, Resnet50, and InceptionNet3. Each of the three models consisted of the pre-trained base, with the modified input layer mentioned above, followed by a flattened layer and two dense layers. The output of each model was a single neuron, representing the probability of malignancy, for each image passed as input. The final classifier for each image was an aggregate of the three different networks, essentially a weighted average of the three output probabilities. The exact value of the weights was one of the primary parameters of experimentation in our study. The criterion by which the final weights were chosen was maximizing the model’s accuracy. The values used in our study were 0.5, 0.25, and 0.25, for VGG15, Resnet50, and InceptionNet3, respectively. However, the weights were purposedly left as a tunable hyperparameter during the final system design. After every image was passed through the full model, the probabilities for the representative images of each patient were averaged out, and the final probability of malignancy was calculated. This value was then compared with the decision threshold and the case was finally labeled as benign or malignant. The exact value of the decision threshold was the second parameter greatly examined in this study. Similar to the weights, different decision thresholds can affect which type of cases were successfully labeled and which were not. The value of the decision threshold was 0.2. As with the case of the pre-trained model weights, the decision threshold is modifiable, according to the requirements and needs of different scenarios. A cross-validation approach based on the k-fold technique was employed to evaluate the performance of the CNN models. A 20% to 80% split was utilized for training and validation, ensuring comprehensive data utilization across multiple folds. The final classifier was constructed through a weighted aggregation of three pre-trained CNN architectures: VGG16, ResNet50, and InceptionNet. Parameters such as aggregation weights and decision threshold probability were fine-tuned to optimize classification performance.

2.5. Training Process

As stated before, each model initially had its input layer modified to a shape of 150-by-150 neurons with ReLU activation. After the convolutional base, a flattened layer consisting of 8192 neurons was added. Two dense layers followed, one with 256 neurons and ReLU activation, and the final layer of the model with a single neuron (representing the binary classifier decision) and sigmoid activation function. The convolutional base was trainable for all the training epochs and, therefore, the pre-trained weights served as a solid starting point. A batch size of 20 images was used, in conjunction with batch normalization. Binary cross-entropy was employed as the loss function while RMSprop (root mean squared propagation) was used for optimization, with an adaptive learning rate of 2x10-5. The training went on for 50 epochs each time, using callbacks and early stopping. The main python package used was Keras from TensorFlow and the entirety of the training process ran on an NVIDIA Quadro M2000 graphics card.

3. Results

The performances of the three different CNN models (VGG16, ResNet50, and InceptionNet) and the aggregate model were evaluated and based on overall accuracy, sensitivity, specificity, and AUC, with their corresponding 95% confidence interval (CI). Bootstrapping was used to calculate the CI. The results can be found in Table 2 and Figure 1. The corresponding ROC curves appear in Figure 2.
VGG16 scored an overall accuracy of 87.50% averaged out over all the k-folds. It achieved an average sensitivity of 95.5% at a specificity of 83.60%. The best-performing fold had a sensitivity of 100%, while the worst-performing fold had a sensitivity of 90.5% with a specificity of 80.0%. It was the overall best-performing CNN among the three pre-trained models, while considering all the performance metrics. ResNet50 had an overall accuracy of 86.8%. The average sensitivity was 90.2% with specificity equal to 84.9%. Even though the highest-scoring fold reached an accuracy of 95.7%, ResNet50 had a bad-performing fold in terms of specificity (68.0%). It had the lowest overall performance among the pre-trained models, except for a marginally higher average specificity when compared to VGG16. InceptionNet achieved accuracy equal to 88.9%, a sensitivity of 88.7%, and a specificity of 88.9%. It is the only pre-trained model that had a higher average specificity than sensitivity, with the highest specificity among the three. It is essentially a “failsafe” in the aggregate model, increasing its ability to correctly label benign tumors and reduce false positives. The aggregate model outperformed all models, only being marginally second to InceptionNet in terms of specificity. It achieved an average sensitivity of 96.5% with a specificity of 88.1%, and the overall accuracy was 90.9%. All the above results were calculated at a decision threshold probability of 0.2.
The constraint applied to choose a value for the decision threshold was to maximize sensitivity while also maintaining high specificity. After enforcing a requirement of having a minimum sensitivity of 98.0%, the value of 0.2 was chosen as the benchmark for all our evaluation runs. At the default decision value of 0.5, the specificity of the aggregate model increased significantly at 98.8%; however, the sensitivity dropped to a below acceptable value, considering the implications in a diagnosis of malignancy. Table 3 and Figure 3 portray the values of all the performance metrics of the aggregate model (for the best-performing k-fold) for decision thresholds between 0.1 and 0.9 with a step of 0.1, with a decision threshold probability of 0.2.
The same principle was applied when selecting the weights of the three pre-trained models in the aggregate network. As shown by our results, the best-performing model in terms of sensitivity and overall performance was VGG16. Even though InceptionNet performed worse than the other two models in terms of sensitivity, its presence is required in the aggregate model to ensure a high specificity. We executed several evaluation runs of the aggregate model, while modifying the weight of the VGG16 model and evenly splitting the remaining weight between the other two models. After experimenting with various values between 0.1 and 0.8, the optimal combination of weights according to our results was 0.5, 0.25, and 0.25, for VGG16, ResNet50, and InceptionNet, respectively. Table 4 and Figure 4 depict the performance metrics with adjusted weights for each model, for an average k-fold with a test size of 117 patients and decision threshold equal to 0.2.
The ultrasonographers achieved a sensitivity of 95.9%, a specificity of 93.6%, and an overall accuracy of 94.2% while classifying the masses. Out of 585 cases, 4 cases were classified as inconclusive by subjective assessment. Out of these, three were malignant according to the histopathology report, and one was benign. Cases classified as inconclusive by the ultrasonographers were categorized as malignant for the purpose of calculating the performance metrics.
The most frequent histological type in false-positive cases by the aggregate model were cystadenomas and endometriomas. In total, 14 cystadenomas and 14 endometriomas were wrongfully classified as malignant by the model. Additionally, 10 cystadenomas and 5 endometriomas were also misclassified as malignant by subjective assessment. Out of these, eight cystadenomas and four endometriomas were misclassified by both the aggregate model and the SA. The total amount of mislabeled cases by the model was calculated as the sum of false positives and false negatives, added over all k-folds. The histopathology of misclassified cases by the aggregate model and subjective assessment can be found in Table 5. Ultrasound images of misclassified images by the CNNs can be found in Figure 5 and Figure 6.

4. Discussion

The use of AI in the evaluation of ultrasound images of ovarian masses seems to be promising. The first steps in assessing the risk of ovarian cancer using AI started as early as 1999 by various study groups [27,28,29,30]. By 2010, Lucidame et al. used HistoScanning, a technique based on the quantification of tissue disorganization in malignancies of backscattered ultrasound 3D raw volumes, achieving a sensitivity of 98% using a small dataset [31]. Multiple subsequent studies have demonstrated high levels of sensitivity specificity and overall accuracy, often surpassing the outcomes of subjective assessment [13,17,20,21,22,23,24,32,33,34].
Several recent studies in the field have also utilized radiomics combined with deep learning to develop a fused predictive model for malignancy in ovarian tumors [20,35,36,37,38]. The key procedure employed is image segmentation followed by feature extraction. The results are then propagated and used in the traditional deep learning techniques we followed in our study. Barcroft et al. used the Dice surface coefficient to measure the segmentation performance, achieving remarkable results [20]. Du et al. fed the extracted features into a logistic regression (LR) model and used a nomogram to evaluate and visualize their outcomes [36]. Despite promising results in the bibliography, we opted out of incorporating radiomics in our study for two main reasons. Firstly, it can severely increase the complexity and development time of the model. Secondly, it has been documented that combining radiomics with deep learning often leads to overfitting the model, due to specific features being included in the training [39].
Our aggregate model achieved comparable results to previously published studies, demonstrating a sensitivity of 96.5% and a specificity of 88.1%. Notably, our model’s sensitivity was higher than the SA in our dataset, which was 95.9%, though it exhibited lower specificity compared to SA (88.1% vs. 93.6%). When compared to the International Ovarian Tumor Analysis (IOTA) studies with a sensitivity of 90.4% and specificity of 92.7%, the SA indicated slightly higher performance metrics than the SA in our dataset [40].
Our findings align with those of Christiansen et al., who developed and trained deep neural networks, achieving a sensitivity of 97.1% and specificity of 89.3%, which were equivalent to subjective assessments using data from 785 ovarian masses obtained from two different centers [21]. Similarly, Hsu et al. reported a mean sensitivity and specificity of 91.37% and 92.92%, respectively, by combining different neural networks [23]. Gao et al., using a diverse dataset and two external validation sets, demonstrated that CNN models could significantly aid non-expert ultrasonographers, improving sensitivity by 12.3% (82.7% vs. 70.4%) and specificity by 8.6% (88.7% vs. 80.1%), underscoring the potential of AI-assisted ultrasonography to reduce human error and facilitate classification [22]. Although in our study the images used were obtained by ultrasonographers with substantial experience, we are confident that inexperienced examiners, who tend to perform with lower accuracy, can benefit from using our model. It can raise suspicion of malignancy, allowing high-risk cases to then be referred to specialized centers for appropriate treatment.
In the present study, we decided to exclude borderline cases, as they are rare and hard to diagnose, even by experts [41]. They often require different treatment with fertility-sparing procedures and extensive follow-up since they appear in younger women [42]. Labeling only malignant and benign cases ensured that inadequate data from borderline cases did not affect the system’s performance. The investigation of such cases was thus left outside the scope of this study and deemed as the central topic for our future research.
We observed a pattern where the most frequent benign histopathological types, particularly cystadenomas and endometriomas, were misclassified as malignant by our model. Specifically, 14 cystadenomas and 14 endometriomas were incorrectly classified as malignant. Cystadenomas appear as large and multilocular cysts with mixed echogenicity, which can lead even experts to false-positive results [43]. We noticed that these cystadenomas indeed presented atypical features, leading both the examiners and the CNNs to a false diagnosis. Endometriomas can appear with low-level echogenicity and atypical features, such as papillary projections, also posing challenges for accurate classification [43,44]. The inclusion of women who specifically underwent surgical treatment and the fact that we are an oncology center with many referrals can lead to the assumption that the cases treated are more likely to be complex. Further research is needed to refine these results.
A major consideration during this study was deciding on the validation and testing techniques that would be applied. An issue somewhat overlooked in similar studies is the independence of the test set and the performance of the system on datasets obtained externally [21,24]. Most of the traditional validation methods used while training CNN models suffer from bias, resulting in skewed results in terms of accuracy. Due to the objective difficulties in obtaining data from different examination centers, the data used to train and test the system usually originate from the same facility. To enhance the diversity of our dataset, we included images obtained over a 12-year period by different examiners and ultrasound machines, resulting in a more heterogeneous dataset.
Despite only having data from a single examination center, we also developed a tailored validation technique by modifying the traditional k-fold cross-validation method. The basis behind our approach was ensuring that the training sets for all the folds had high enough variance to train the models on all possible features that can be found in ovarian ultrasound images. To accomplish this, instead of the usual random indexing of the dataset before the training and validation split for each fold, we further divided the dataset into multiple subcategories, labeled with various indicative features. Those included histological type, existence of calipers or not, high or low resolution, etc. We then proceeded to split the dataset into k-folds, ensuring that not only every image appeared at least once in a validation set, but also that each subcategory was adequately represented in each training set.
The second noteworthy differentiation was modifying the validation set for each fold to emulate testing the model with an external dataset. Several modifications were made to the images, mainly consisting of resizing, color corrections, saturation adjustments, and feature repositioning. The goal was to alter the validation dataset as much as possible while maintaining the key features related to the diagnosis of malignancies.
Our study has certain limitations. It was conducted at a single center, which may limit the generalizability of our results to other settings or populations, and it lacked external validation. Moreover, the performance of the model relies on the quality of the ultrasound images, which were obtained by experienced ultrasonographers, potentially limiting its effectiveness with images from less experienced practitioners but also ensuring high-quality and reliable data. To further address these issues, we obtained a heterogenous dataset and employed advanced data augmentation techniques to diversify our data. The use of a modified k-fold cross-validation also enhanced the robustness and reliability of the aggregate model. Lastly, the inclusion of cases with histology reports serves as a gold standard for validation, certifies the accuracy of our ground truth, and strengthens our study’s reliability.
The scheme we designed and developed is meant to be used as a tool in the field of ovarian tumor classification that can be easily used to enhance correct diagnosis and assist non-expert sonographers. Therefore, it can be beneficial for end-users to be able to differentiate and specialize it in detecting or ruling out specific type of tumors, depending on what other systems, tools, and diagnosis techniques it is being paired with. However, future research and validation is needed to generalize the use of the model in different centers since we could not provide external validation to test the model’s performance.

5. Conclusions

The results demonstrated the potential of deep learning and AI-assisted image analysis in enhancing the diagnostic process for ovarian tumors. By leveraging pre-trained CNN models, the study achieved high levels of sensitivity and specificity, even surpassing subjective assessment’s sensitivity. The robust performance across diverse k-folds underscores the reliability of the developed classifier. Importantly, the study highlighted the role of automated image analysis in minimizing errors associated with subjective interpretation, thereby improving the diagnostic accuracy. Room for further optimization is there should reducing the time and resources of collecting data become a pressing issue.

Author Contributions

M.G. was responsible for the conceptualization, methodology, validation, investigation, writing—review and editing, and visualization of this article. I.P. was responsible for data curation, investigation, methodology, and writing review and editing. S.S. was responsible for Data Curation. A.P., A.G., C.K., T.K., S.T. and A.Z. were responsible for data curation. I.S. was responsible for the development of software, writing, reviewing and the investigation. M.F. was responsible for data curation, editing, reviewing, and investigating. E.P. was responsible for writing, editing, and investigating. G.D. was responsible for the project administrations and supervision. E.D. was responsible for the supervision and project administration. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of Alexandra General Hospital, National and Kapodistrian University of Athens (336/15-06-2022).

Informed Consent Statement

Patient consent was waived since anonymous, prestored ultrasound images were used for training the models.

Data Availability Statement

The dataset used in our study is not publicly available due to patients’ privacy and research permissions, but data and results are available upon reasonable request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Siegel, R.L.; Giaquinto, A.N.; Jemal, A. Cancer statistics, 2024. CA Cancer J. Clin. 2024, 74, 12–49. [Google Scholar] [CrossRef] [PubMed]
  2. McCorkle, R.; Pasacreta, J.; Tang, S.T. The silent killer: Psychological issues in ovarian cancer. Holist. Nurs. Pract. 2003, 17, 300–308. [Google Scholar] [CrossRef] [PubMed]
  3. Forstner, R. Early detection of ovarian cancer. Eur. Radiol. 2020, 30, 5370–5373. [Google Scholar] [CrossRef] [PubMed]
  4. Froyman, W.; Timmerman, D. Methods of Assessing Ovarian Masses: International Ovarian Tumor Analysis Approach. Obstet. Gynecol. Clin. N. Am. 2019, 46, 625–641. [Google Scholar] [CrossRef] [PubMed]
  5. Timmerman, D.; Schwärzler, P.; Collins, W.P.; Claerhout, F.; Coenen, M.; Amant, F.; Vergote, I.; Bourne, T.H. Subjective assessment of adnexal masses with the use of ultrasonography: An analysis of interobserver variability and experience. Ultrasound Obstet. Gynecol. 1999, 13, 11–16. [Google Scholar] [CrossRef] [PubMed]
  6. Timmerman, D. The use of mathematical models to evaluate pelvic masses; can they beat an expert operator? Best Pract. Res. Clin. Obstet. Gynaecol. 2004, 18, 91–104. [Google Scholar] [CrossRef] [PubMed]
  7. Timmerman, D.; Testa, A.C.; Bourne, T.; Ameye, L.; Jurkovic, D.; Van Holsbeke, C.; Paladini, D.; Van Calster, B.; Vergote, I.; Van Huffel, S.; et al. Simple ultrasound-based rules for the diagnosis of ovarian cancer. Ultrasound Obstet. Gynecol. 2008, 31, 681–690. [Google Scholar] [CrossRef]
  8. Valentin, L.; Hagen, B.; Tingulstad, S.; Eik-Nes, S. Comparison of ‘pattern recognition’ and logistic regression models for discrimination between benign and malignant pelvic masses: A prospective cross validation. Ultrasound Obstet. Gynecol. 2001, 18, 357–365. [Google Scholar] [CrossRef] [PubMed]
  9. Van Calster, B.; Van Hoorde, K.; Valentin, L.; Testa, A.C.; Fischerova, D.; Van Holsbeke, C.; Savelli, L.; Franchi, D.; Epstein, E.; Kaijser, J.; et al. Evaluating the risk of ovarian cancer before surgery using the ADNEX model to differentiate between benign, borderline, early and advanced stage invasive, and secondary metastatic tumours: Prospective multicentre diagnostic study. BMJ 2014, 349, g5920. [Google Scholar] [CrossRef]
  10. Zhang, Q.; Dai, X.; Li, W. Systematic Review and Meta-Analysis of O-RADS Ultrasound and O-RADS MRI for Risk Assessment of Ovarian and Adnexal Lesions. AJR Am. J. Roentgenol. 2023, 221, 21–33. [Google Scholar] [CrossRef]
  11. Kaijser, J.; Bourne, T.; Valentin, L.; Sayasneh, A.; Van Holsbeke, C.; Vergote, I.; Testa, A.C.; Franchi, D.; Van Calster, B.; Timmerman, D. Improving strategies for diagnosing ovarian cancer: A summary of the International Ovarian Tumor Analysis (IOTA) studies. Ultrasound Obstet. Gynecol. 2013, 41, 9–20. [Google Scholar] [CrossRef]
  12. Giourga, M.; Pouliakis, A.; Vlastarakos, P.; Stavrou, S.; Tsiriva, M.; Gerede, A.; Daskalakis, G.; Voros, C.; Drakakis, P.; Domali, E. Evaluation of IOTA-ADNEX Model and Simple Rules for Identifying Adnexal Masses by Operators with Varying Levels of Expertise: A Single-Center Diagnostic Accuracy Study. Ultrasound Int. Open 2023, 9, E11–E17. [Google Scholar] [CrossRef] [PubMed]
  13. Akazawa, M.; Hashimoto, K. Artificial intelligence in gynecologic cancers: Current status and future challenges—A systematic review. Artif. Intell. Med. 2021, 120, 102164. [Google Scholar] [CrossRef] [PubMed]
  14. Bini, S.A. Artificial Intelligence, Machine Learning, Deep Learning, and Cognitive Computing: What Do These Terms Mean and How Will They Impact Health Care? J. Arthroplast. 2018, 33, 2358–2361. [Google Scholar] [CrossRef]
  15. Chen, X.; Huo, X.F.; Wu, Z.; Lu, J.J. Advances of Artificial Intelligence Application in Medical Imaging of Ovarian Cancers. Chin. Med. Sci. J. 2021, 36, 196–203. [Google Scholar] [CrossRef]
  16. Malani, S.N.; Shrivastava, D.; Raka, M.S. A Comprehensive Review of the Role of Artificial Intelligence in Obstetrics and Gynecology. Cureus 2023, 15, e34891. [Google Scholar] [CrossRef] [PubMed]
  17. Acharya, U.R.; Mookiah, M.R.; Vinitha Sree, S.; Yanti, R.; Martis, R.J.; Saba, L.; Molinari, F.; Guerriero, S.; Suri, J.S. Evolutionary algorithm-based classifier parameter tuning for automatic ovarian cancer tissue characterization and classification. Ultraschall Med. 2014, 35, 237–245. [Google Scholar] [CrossRef]
  18. Akazawa, M.; Hashimoto, K. Artificial Intelligence in Ovarian Cancer Diagnosis. Anticancer Res. 2020, 40, 4795–4800. [Google Scholar] [CrossRef]
  19. Sone, K.; Toyohara, Y.; Taguchi, A.; Miyamoto, Y.; Tanikawa, M.; Uchino-Mori, M.; Iriyama, T.; Tsuruga, T.; Osuga, Y. Application of artificial intelligence in gynecologic malignancies: A review. J. Obstet. Gynaecol. Res. 2021, 47, 2577–2585. [Google Scholar] [CrossRef]
  20. Barcroft, J.F.; Linton-Reid, K.; Landolfo, C.; Al-Memar, M.; Parker, N.; Kyriacou, C.; Munaretto, M.; Fantauzzi, M.; Cooper, N.; Yazbek, J.; et al. Machine learning and radiomics for segmentation and classification of adnexal masses on ultrasound. NPJ Precis Oncol. 2024, 8, 41. [Google Scholar] [CrossRef]
  21. Christiansen, F.; Epstein, E.L.; Smedberg, E.; Åkerlund, M.; Smith, K.; Epstein, E. Ultrasound image analysis using deep neural networks for discriminating between benign and malignant ovarian tumors: Comparison with expert subjective assessment. Ultrasound Obstet. Gynecol. 2021, 57, 155–163. [Google Scholar] [CrossRef] [PubMed]
  22. Gao, Y.; Zeng, S.; Xu, X.; Li, H.; Yao, S.; Song, K.; Li, X.; Chen, L.; Tang, J.; Xing, H.; et al. Deep learning-enabled pelvic ultrasound images for accurate diagnosis of ovarian cancer in China: A retrospective, multicentre, diagnostic study. Lancet Digit. Health 2022, 4, e179–e187. [Google Scholar] [CrossRef] [PubMed]
  23. Hsu, S.T.; Su, Y.J.; Hung, C.H.; Chen, M.J.; Lu, C.H.; Kuo, C.E. Automatic ovarian tumors recognition system based on ensemble convolutional neural network with ultrasound imaging. BMC Med. Inform. Decis. Mak. 2022, 22, 298. [Google Scholar] [CrossRef] [PubMed]
  24. Jung, Y.; Kim, T.; Han, M.R.; Kim, S.; Kim, G.; Lee, S.; Choi, Y.J. Ovarian tumor diagnosis using deep convolutional neural networks and a denoising convolutional autoencoder. Sci. Rep. 2022, 12, 17024. [Google Scholar] [CrossRef]
  25. Farag, N.H.; Alsaggaf, Z.H.; Bamardouf, N.O.; Khesfaty, D.M.; Fatani, M.M.; Alghamdi, M.K.; Saharti, S.N. The Histopathological Patterns of Ovarian Neoplasms in Different Age Groups: A Retrospective Study in a Tertiary Care Center. Cureus 2022, 14, e33092. [Google Scholar] [CrossRef]
  26. Prat, J. FIGO’s staging classification for cancer of the ovary, fallopian tube, and peritoneum: Abridged republication. J. Gynecol. Oncol. 2015, 26, 87–89. [Google Scholar] [CrossRef]
  27. Timmerman, D.; Verrelst, H.; Bourne, T.H.; De Moor, B.; Collins, W.P.; Vergote, I.; Vandewalle, J. Artificial neural network models for the preoperative discrimination between malignant and benign adnexal masses. Ultrasound Obstet. Gynecol. 1999, 13, 17–25. [Google Scholar] [CrossRef]
  28. Clayton, R.D.; Snowden, S.; Weston, M.J.; Mogensen, O.; Eastaugh, J.; Lane, G. Neural networks in the diagnosis of malignant ovarian tumours. Br. J. Obstet. Gynaecol. 1999, 106, 1078–1082. [Google Scholar] [CrossRef] [PubMed]
  29. Tailor, A.; Jurkovic, D.; Bourne, T.H.; Collins, W.P.; Campbell, S. Sonographic prediction of malignancy in adnexal masses using an artificial neural network. Br. J. Obstet. Gynaecol. 1999, 106, 21–30. [Google Scholar] [CrossRef]
  30. Szpurek, D.; Moszynski, R.; Smolen, A.; Sajdak, S. Artificial neural network computer prediction of ovarian malignancy in women with adnexal masses. Int. J. Gynaecol. Obstet. 2005, 89, 108–113. [Google Scholar] [CrossRef]
  31. Lucidarme, O.; Akakpo, J.P.; Granberg, S.; Sideri, M.; Levavi, H.; Schneider, A.; Autier, P.; Nir, D.; Bleiberg, H.; Ovarian HistoScanning Clinical Study Group. A new computer-aided diagnostic tool for non-invasive characterisation of malignant ovarian masses: Results of a multicentre validation study. Eur. Radiol. 2010, 20, 1822–1830. [Google Scholar] [CrossRef] [PubMed]
  32. Li, J.; Chen, Y.; Zhang, M.; Zhang, P.; He, K.; Yan, F.; Xu, H.; Burkhoff, D.; Luo, Y.; Wang, L.; et al. A Deep Learning Model System for Diagnosis and Management of Adnexal Masses. Cancers 2022, 14, 5291. [Google Scholar] [CrossRef] [PubMed]
  33. Aramendía-Vidaurreta, V.; Cabeza, R.; Villanueva, A.; Navallas, J.; Alcázar, J.L. Ultrasound Image Discrimination between Benign and Malignant Adnexal Masses Based on a Neural Network Approach. Ultrasound Med. Biol. 2016, 42, 742–752. [Google Scholar] [CrossRef] [PubMed]
  34. Martínez-Más, J.; Bueno-Crespo, A.; Khazendar, S.; Remezal-Solano, M.; Martínez-Cendán, J.P.; Jassim, S.; Du, H.; Al Assam, H.; Bourne, T.; Timmerman, D. Evaluation of machine learning methods with Fourier Transform features for classifying ovarian tumors based on ultrasound images. PLoS ONE 2019, 14, e0219388. [Google Scholar] [CrossRef] [PubMed]
  35. Adusumilli, P.; Ravikumar, N.; Hall, G.; Swift, S.; Orsi, N.; Scarsbrook, A. Radiomics in the evaluation of ovarian masses—A systematic review. Insights Imaging 2023, 14, 165. [Google Scholar] [CrossRef] [PubMed]
  36. Du, Y.; Xiao, Y.; Guo, W.; Yao, J.; Lan, T.; Li, S.; Wen, H.; Zhu, W.; He, G.; Zheng, H.; et al. Development and validation of an ultrasound-based deep learning radiomics nomogram for predicting the malignant risk of ovarian tumours. Biomed. Eng. Online 2024, 23, 41. [Google Scholar] [CrossRef] [PubMed]
  37. Tang, Z.P.; Ma, Z.; He, Y.; Liu, R.C.; Jin, B.B.; Wen, D.Y.; Wen, R.; Yin, H.H.; Qiu, C.C.; Gao, R.Z.; et al. Ultrasound-based radiomics for predicting different pathological subtypes of epithelial ovarian cancer before surgery. BMC Med. Imaging 2022, 22, 147. [Google Scholar] [CrossRef]
  38. Liu, L.; Cai, W.; Zhou, C.; Tian, H.; Wu, B.; Zhang, J.; Yue, G.; Hao, Y. Ultrasound radiomics-based artificial intelligence model to assist in the differential diagnosis of ovarian endometrioma and ovarian dermoid cyst. Front. Med. 2024, 11, 1362588. [Google Scholar] [CrossRef] [PubMed]
  39. Afshar, P.; Mohammadi, A.; Plataniotis, K.N.; Oikonomou, A.; Benali, H. From handcrafted to deep-learning-based cancer radiomics: Challenges and opportunities. IEEE Signal Process. Mag. 2019, 36, 132–160. [Google Scholar] [CrossRef]
  40. Timmerman, D.; Valentin, L.; Bourne, T.H.; Collins, W.P.; Verrelst, H.; Vergote, I.; Group, I.O.T.A.I. Terms, definitions and measurements to describe the sonographic features of adnexal tumors: A consensus opinion from the International Ovarian Tumor Analysis (IOTA) Group. Ultrasound Obstet. Gynecol. 2000, 16, 500–505. [Google Scholar] [CrossRef]
  41. Exacoustos, C.; Romanini, M.E.; Rinaldo, D.; Amoroso, C.; Szabolcs, B.; Zupi, E.; Arduini, D. Preoperative sonographic features of borderline ovarian tumors. Ultrasound Obstet. Gynecol. 2005, 25, 50–59. [Google Scholar] [CrossRef] [PubMed]
  42. Daraï, E.; Fauvet, R.; Uzan, C.; Gouy, S.; Duvillard, P.; Morice, P. Fertility and borderline ovarian tumor: A systematic review of conservative management, risk of recurrence and alternative options. Hum. Reprod. Update 2013, 19, 151–166. [Google Scholar] [CrossRef] [PubMed]
  43. Sayasneh, A.; Ekechi, C.; Ferrara, L.; Kaijser, J.; Stalder, C.; Sur, S.; Timmerman, D.; Bourne, T. The characteristic ultrasound features of specific types of ovarian pathology (review). Int. J. Oncol. 2015, 46, 445–458. [Google Scholar] [CrossRef] [PubMed]
  44. Tanase, Y.; Kawaguchi, R.; Takahama, J.; Kobayashi, H. Factors that Differentiate between Endometriosis-associated Ovarian Cancer and Benign Ovarian Endometriosis with Mural Nodules. Magn. Reson. Med. Sci. 2018, 17, 231–237. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Performance metrics for diagnosing ovarian tumors as benign or malignant for the CNN models, VGG16 (-), ResNet50 (-), and InceptionNet (-), and the aggregate model (-).
Figure 1. Performance metrics for diagnosing ovarian tumors as benign or malignant for the CNN models, VGG16 (-), ResNet50 (-), and InceptionNet (-), and the aggregate model (-).
Jcm 13 04123 g001
Figure 2. ROC curves for diagnosing ovarian tumors as benign or malignant for the CNN models, VGG16 (-), ResNet50 (-), and InceptionNet (-), and the aggregate model (-).
Figure 2. ROC curves for diagnosing ovarian tumors as benign or malignant for the CNN models, VGG16 (-), ResNet50 (-), and InceptionNet (-), and the aggregate model (-).
Jcm 13 04123 g002
Figure 3. Performance metrics as a function of the decision threshold probability for the aggregate CNN model, the best-performing k-fold.
Figure 3. Performance metrics as a function of the decision threshold probability for the aggregate CNN model, the best-performing k-fold.
Jcm 13 04123 g003
Figure 4. Graph of the performance metrics for the aggregate model in an average k-fold with a test-set size of 117 cases, for varying weights of each pre-trained model. Decision threshold of 0.2.
Figure 4. Graph of the performance metrics for the aggregate model in an average k-fold with a test-set size of 117 cases, for varying weights of each pre-trained model. Decision threshold of 0.2.
Jcm 13 04123 g004
Figure 5. Still ultrasound images of malignant ovarian masses misclassified as benign by the aggregate model. (a) Endometrioid ovarian carcinoma, (b) high-grade serous ovarian carcinoma, and (c) high-grade serous ovarian carcinoma.
Figure 5. Still ultrasound images of malignant ovarian masses misclassified as benign by the aggregate model. (a) Endometrioid ovarian carcinoma, (b) high-grade serous ovarian carcinoma, and (c) high-grade serous ovarian carcinoma.
Jcm 13 04123 g005
Figure 6. Still ultrasound images of benign ovarian masses misclassified as malignant by the aggregate model. (a) Serous cystadenoma, (b) endometrioma, and (c) mature cystic teratoma.
Figure 6. Still ultrasound images of benign ovarian masses misclassified as malignant by the aggregate model. (a) Serous cystadenoma, (b) endometrioma, and (c) mature cystic teratoma.
Jcm 13 04123 g006
Table 1. Histopathology of the ovarian masses included.
Table 1. Histopathology of the ovarian masses included.
HistopathologyN (%)
BenignCystadenoma (serous, mucinous, and sero-mucinous)153 (26.2)
Endometrioma72 (12.3)
Mature teratoma60 (10.3)
Benign tubal45 (7.7)
Corpus luteum25 (4.2)
Cystadenofibroma (serous and mucinous)18 (3.1)
Thecoma12 (2.1)
Fibroma2 (0.3)
Other benign3 (0.5)
Total benign390 (66.7)
MalignantSerous carcinoma101 (17.3)
Metastatic26 (4.4)
Germ cell malignant tumor15 (2.6)
Mucinous carcinoma15 (2.6)
Endometrioid carcinoma12 (2.0)
Clear cell carcinoma9 (1.5)
Sex cord malignant tumor6 (1)
Carcinosarcoma4 (0.7)
Other malignant tumors7 (1.2)
Total malignant195 (33.3)
Table 2. Performance metrics for all the CNNs individually, the aggregate model, and subjective assessment. SA: subjective assessment. AUC: area under the curve. CI: confidence interval.
Table 2. Performance metrics for all the CNNs individually, the aggregate model, and subjective assessment. SA: subjective assessment. AUC: area under the curve. CI: confidence interval.
MetricsVGG16% (95% CI)ResNet50%
(95% CI)
InceptionNet% (95% CI)Aggregate% (95% CI)SA%
(95% CI)
Accuracy87.50
(82.3–91.9)
86.80
(82.6–89.6)
88.90
(83.7–93.5)
90.90
(85.6–93.1)
94.2
(92.3–98.3)
Sensitivity95.50
(91.1–97.3)
90.20
(86.2–93.2)
88.70
(83.9–91.9)
96.50
(91.2–98.5)
95.90
(93.9–99.1)
Specificity83.60
(78.7–85.6)
84.90
(79.7–88.1)
88.90
(83.7–93.5)
88.10
(85.1–90.2)
93.60
(88.6–95.9)
AUC89.50
(84.8–91.9)
87.50
(80.1–90.3)
88.70
(83.9–91.9)
92.20
(90.8–97.1)
-
Table 3. Performance metrics for all the CNNs individually, the aggregate model, and subjective assessment for different decision thresholds for the best-performing k-fold.
Table 3. Performance metrics for all the CNNs individually, the aggregate model, and subjective assessment for different decision thresholds for the best-performing k-fold.
Threshold0.10.20.30.40.50.60.70.80.9
Accuracy0.9320.9740.9740.9490.9060.8800.8550.8460.803
Specificity0.9000.9630.9750.9750.9880.9881.0001.0001.000
Sensitivity1.0001.0000.9730.8920.7300.6490.5410.5140.379
AUC0.9500.9810.9740.9330.8590.8180.7700.7570.689
Table 4. Performance metrics for the aggregate model in an average k-fold with a test-set size of 117 cases, for varying weights of pre-trained models. The decision threshold used was 0.2. AUC: area under the curve. FP: false positives. FN: false negatives.
Table 4. Performance metrics for the aggregate model in an average k-fold with a test-set size of 117 cases, for varying weights of pre-trained models. The decision threshold used was 0.2. AUC: area under the curve. FP: false positives. FN: false negatives.
VGG16ResNetInceptionNetVGG16ResNetInceptionNetVGG16ResNetInceptionNetVGG16ResNetInceptionNet
Weights0.10.450.450.20.40.40.30.350.350.40.30.3
Accuracy0.9150.9150.9150.915
Specificity0.9020.890.890.89
Sensitivity0.9430.9430.9430.943
AUC0.9230.9310.9310.931
FP8999
FN2222
VGG16ResNetInceptionNetVGG16ResNetInceptionNetVGG16ResNetInceptionNetVGG16ResNetInceptionNet
Weights0.50.250.250.60.20.20.70.150.150.80.10.1
Accuracy0.9060.8970.880.863
Specificity0.8780.8660.8410.829
Sensitivity0.9710.9710.9710.943
AUC0.9250.9190.9060.886
FP10111314
FN1112
Table 5. Misclassified cases by the aggregate model and subjective assessment.
Table 5. Misclassified cases by the aggregate model and subjective assessment.
HistopathologyAggregate ModelSA
Benign (Total)4626
Cystadenoma1410
Endometrioma145
Mature teratoma6-
Abscess31
Corpus Luteum31
Hydrosalpinx2-
Cystadenofibroma25
Serous Cyst11
Rete ovarii1-
Brenner tumor-2
Thecoma-2
Malignant (Total)78
Serous carcinoma41
Endometrioid carcinoma22
Metastatic12
Ovarian Schwannoma-1
Immature teratoma-2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Giourga, M.; Petropoulos, I.; Stavros, S.; Potiris, A.; Gerede, A.; Sapantzoglou, I.; Fanaki, M.; Papamattheou, E.; Karasmani, C.; Karampitsakos, T.; et al. Enhancing Ovarian Tumor Diagnosis: Performance of Convolutional Neural Networks in Classifying Ovarian Masses Using Ultrasound Images. J. Clin. Med. 2024, 13, 4123. https://doi.org/10.3390/jcm13144123

AMA Style

Giourga M, Petropoulos I, Stavros S, Potiris A, Gerede A, Sapantzoglou I, Fanaki M, Papamattheou E, Karasmani C, Karampitsakos T, et al. Enhancing Ovarian Tumor Diagnosis: Performance of Convolutional Neural Networks in Classifying Ovarian Masses Using Ultrasound Images. Journal of Clinical Medicine. 2024; 13(14):4123. https://doi.org/10.3390/jcm13144123

Chicago/Turabian Style

Giourga, Maria, Ioannis Petropoulos, Sofoklis Stavros, Anastasios Potiris, Angeliki Gerede, Ioakeim Sapantzoglou, Maria Fanaki, Eleni Papamattheou, Christina Karasmani, Theodoros Karampitsakos, and et al. 2024. "Enhancing Ovarian Tumor Diagnosis: Performance of Convolutional Neural Networks in Classifying Ovarian Masses Using Ultrasound Images" Journal of Clinical Medicine 13, no. 14: 4123. https://doi.org/10.3390/jcm13144123

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop