Next Article in Journal
Consensus-Based Information Filtering in Distributed LiDAR Sensor Network for Tracking Mobile Robots
Previous Article in Journal
Rehabilitation Assessment System for Stroke Patients Based on Fusion-Type Optoelectronic Plethysmography Device and Multi-Modality Fusion Model: Design and Validation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Distinguishing the Uterine Artery, the Ureter, and Nerves in Laparoscopic Surgical Images Using Ensembles of Binary Semantic Segmentation Networks

1
Faculty of Informatics, University of Debrecen, 4028 Debrecen, Hungary
2
Department of Obstetrics and Gynaecology, Faculty of Medicine, University of Debrecen, 4032 Debrecen, Hungary
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(9), 2926; https://doi.org/10.3390/s24092926
Submission received: 5 March 2024 / Revised: 26 April 2024 / Accepted: 2 May 2024 / Published: 4 May 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Performing a minimally invasive surgery comes with a significant advantage regarding rehabilitating the patient after the operation. But it also causes difficulties, mainly for the surgeon or expert who performs the surgical intervention, since only visual information is available and they cannot use their tactile senses during keyhole surgeries. This is the case with laparoscopic hysterectomy since some organs are also difficult to distinguish based on visual information, making laparoscope-based hysterectomy challenging. In this paper, we propose a solution based on semantic segmentation, which can create pixel-accurate predictions of surgical images and differentiate the uterine arteries, ureters, and nerves. We trained three binary semantic segmentation models based on the U-Net architecture with the EfficientNet-b3 encoder; then, we developed two ensemble techniques that enhanced the segmentation performance. Our pixel-wise ensemble examines the segmentation map of the binary networks on the lowest level of pixels. The other algorithm developed is a region-based ensemble technique that takes this examination to a higher level and makes the ensemble based on every connected component detected by the binary segmentation networks. We also introduced and trained a classic multi-class semantic segmentation model as a reference and compared it to the ensemble-based approaches. We used 586 manually annotated images from 38 surgical videos for this research and published this dataset.

1. Introduction

The hysterectomy is one of the most frequently performed gynecological surgeries for women [1]. This surgery can be completed in different ways. The traditional techniques comprise vaginal and abdominal hysterectomies [2], but with the improvement of technology over time, the laparoscopic hysterectomy technique was introduced in 1988 [3] to replace the abdominal technique. Although the laparoscopic procedure proved to be superior in terms of postoperative recovery and operative blood loss [4], ureteral injury became a frequent complication of laparoscopic hysterectomies [5,6,7,8].
One of the main reasons for the higher number of ureter injuries is the lack of clear guidelines during a laparoscopic hysterectomy. The risk of injury highly depends on the surgeon’s experience [9]. The technique commonly used to prevent damage to the ureter is etching to the posterior broad ligament and then identifying and moving the ureter laterally [10]. However, this technique can be performed only after the clear identification of the ureter, which is demanding when the surgeon only has visual information available. Apart from this visual information, other regularly used techniques to find the ureter include the placement of a ureteral stent or using x-ray fluoroscopy [11,12]. Nonetheless, these foreign materials lead to additional risks for the patient during the surgery.
In recent years, biomedical image processing has developed rapidly, allowing the application of imaging science to multi-modal biomedical images and providing useful support for examining and diagnosing human patients [13]. Machine learning and artificial intelligence has also boosted the evolution of medical imaging services such as computer-aided diagnosis, image fusion, or image segmentation [14]. Nowadays, these techniques are used in many areas, including in the analysis of CT or MRI images [13], the detection of tumors or cardiac diseases [14], and the registration of multi-modal medical images [15].

1.1. Related Works

1.1.1. Ureter and Uterine Artery Detection

As mentioned in the introduction, detecting the ureter and uterine arteries and their differentiation during laparoscopic surgeries is challenging. Still, a few research efforts have been made to address this issue. Accordingly, in our former work [16], we presented a semi-automatic solution to differentiate the ureter from the uterine artery from images captured during laparoscopic surgeries. In this system, a medical expert or assistant manually selected the region of interest (ROI) from each examined image. Then, 224 × 224 sub-images were cut off from the highlighted ROI and classified using GoogLeNet [17]. The final classification of the ROI was decided through majority voting.
The authors of [18] developed and built an image-guided endoscope system using infrared light to detect the ureter during surgery. They proposed a self-adaptive threshold-based segmentation algorithm to enhance the accuracy of ureter detection. Finally, they fused the detection results with the RGB images captured with the endoscopic camera. In [19], a convolutional neural network (CNN) called Kid-Net was proposed to semantically segment arteries, veins, and the ureter. However, this network was trained to segment these structures from volumetric CT images, not the RGB images used in laparoscopic surgeries.
Multi-class semantic segmentation is a reasonable solution for the task of processing images from laparoscopic surgeries. Ref. [20] used the Mask R-CNN [21] to segment multiple instruments from a video of a laparoscopic surgery. Their proposed network uses a region-based CNN to segment the labeled instruments from the images. The work in [22] presented a modified U-Net architecture named U-NetPlus, with the same aim of segmenting the tools used during laparoscopic surgeries. They initiated pre-trained VGG-11 and VGG-16 architectures with batch normalization for the encoder part. In [23], a deep CNN and a dataset named m2caiSeg were provided. This annotated dataset contains labeled organs, instruments, fluids, and arteries. The presented network segments images from laparoscopic surgeries into 19 categories, pixel-wise.

1.1.2. Ensemble Models

Ensemble learning aims to solve a particular machine learning task by combining the knowledge of multiple members. The idea behind this learning method is partly the simulation of human nature, gathering meaningful, diverse knowledge to achieve an enhanced solution to a complex problem [24]. Usually, these members are trained on the same dataset for the same aim but using different approaches. Furthermore, when dealing with a multi-class segmentation problem, binarization may help to solve it. The most common techniques for decomposition are the “one-vs-one” (OVO) [25] and “one-vs-all” (OVA) [26] techniques. The OVO scheme divides the classification problem of m classes into m ( m 1 ) / 2 binary classifications. During the training of the binary models, a subset of the original dataset is used, containing only the corresponding classes. Each of these binary models creates a prediction matrix P ( i , j )   ( x , y ) , where an entry of the matrix indicates the confidence of the classifier for the i -th class over the j -th one for a given image pixel ( x , y ) with x = 1 , , W and y = 1 , , H , where the input image has the width W and height H . To aggregate the P ( i , j ) matrices for a final prediction, a voting strategy is a simple possibility. Namely, each classifier votes for its predicted class, and the class with the most votes will be suggested finally [27].
On the other hand, the OVA decomposition technique breaks down the m -class classification problem into m binary ones. Usually, the training phase uses the whole dataset, labeling one class as positive and all other classes as negative. Each of these m binary classifiers creates a matrix P i   ( x , y ) describing the confidence level of class i for each image pixel ( x , y ) . A standard aggregation scheme for the OVA approach is the Maximum Confidence Strategy, where the class with the highest confidence level will be chosen as the final prediction [27].
The output of an ensemble system can be calculated in many ways. The first and most straightforward way is to aggregate the member outputs by calculating their average, weighted average [28], or arithmetic mean [29]. Another method used to fuse the ensemble members’ outputs is majority voting [30]. This decision model can be boosted by assigning voter confidence levels [31] or weights [32]. Typically, weighted voting ensembles perform better than non-weighted ones since the weights or any other additional parameter let the ensemble be tuned further, which generally leads to a better prediction.
Our work uses the OVA technique to decompose the traditional multi-class U-Net semantic segmentation model into three binary semantic segmentation models. Then, we merge these models to compose both a pixel-wise and a region-based ensemble. We also present a weighted version of these ensembles, where the binary members’ accuracies or confidence levels are considered as weights.

2. Materials and Methods

2.1. Dataset

The dataset used for this research contains 586 images from 38 surgical videos captured during actual laparoscopic surgeries with a resolution of 1340 × 648 pixels. Each image has a corresponding manually annotated mask, which indicates the pixel-level presence of the ureter, the uterine artery, or nerves on the selected frame. In collaboration with the University of Debrecen, gynecological experts carried out these annotations. For this task, an annotation software was developed, where the gynecologists could draw the boundaries of the occurring ureter, uterine artery, or nerve. Later, these annotated boundaries were filled to create a pixel-level mask of the annotated organs and merged into a single mask to make the final ground-truth mask for the input image. Figure 1 demonstrates the annotation process with its main steps.
From the total of 586 images, we obtained 470 containing uterine arteries, 254 containing the ureter, and 145 containing the nerves. It is also important to note that from these images, 285 frames included both the uterine artery and the ureter. During our research, the dataset was separated into two parts, where 530 images were used to train the semantic segmentation models, and 56 images were separated for testing purposes. During the training of the multi-class and binary networks, 5-fold cross-validation was used, where each split was based on the videos. Therefore, images from the same videos were always kept in the same split. The frames of the test dataset were collected from four surgical videos, which were excluded from the training videos to make our final evaluation reliable. The dataset containing the images from the surgical videos with the corresponding masks (both with separate mask images for every present class and with multi-label mask files well) has been published and is available for further research [33].

2.2. U-Net Deep Convolutional Network

The reference model for this task was a multi-class U-Net semantic segmentation model [34]. This model takes the whole dataset with multi-class labeling (0—background; 1—uterine artery; 2—ureter; 3—nerve) and returns a pixel-level prediction map for each image. The U-Net architecture can take a pre-trained encoder and use it during training. U-Net architecture is widely used for semantic segmentation tasks, since the architecture can be modified and improved through different operations quite easily, which can lead to new network architectures based on U-Net such as those seen in [35,36]. For our research, the EfficientNet-b3 [37] encoder was selected and used for the multi-class and binary segmentation networks in this case. Figure 2 illustrates the architecture of the used U-Net semantic segmentation network with the EfficientNet backbone. As this figure demonstrates, the downsampling part of the network represents an EfficientNet-b3 encoder, which uses a 5-staged downsampling method plus an additional downsampling layer containing a Leaky ReLU, a 2 × 2 Max-Pooling layer, and a Dropout layer as well. The upsampling is performed via the default U-Net operations using upsampling layers with transposed 2D convolutional operations. Each final upsampling layer is a 1 × 1 convolution layer attached with a Sigmoid or SoftMax layer, depending on the number of output classes.
As shown in the discussion of the performance of this network in the Results section, this multi-class network could not efficiently find the differences between the ureter, the uterine artery, and nerves. This poor performance was achieved because of the nature of the dataset and the visually similar classes. The distribution of classes in the dataset is highly imbalanced; images with the ureter or uterine arteries are over-represented compared to those with nerves. The problem with this imbalanced distribution is that even with a higher batch size, it is possible that not all the classes were represented in the batch. This resulted in confusion about the loss function and the correct convergence to its minimum.

2.3. Binary Semantic Segmentation Models and Their Ensembles

Due to the abovementioned problems, the multi-class segmentation model had to be decomposed into three separate binary semantic segmentation models. So, instead of having one model which has the goal of learning to segment the ureter, uterine artery, and nerves at the same time, the idea was to create separate models and train each of them to only have the capability to segment one of the three organs. Afterward, these binary segmentation models were merged to predict the existence of the classes on images that could contain any of these classes. All of the binary models, like the original multi-class model, used the U-Net architecture with the EfficientNet encoder. This ensemble of binary models was required to build the ability to predict multi-class images accurately. This section will discuss four possible techniques to build ensemble models that can be applied to our task.

2.3.1. Pixel-Wise Ensemble

The first approach uses a pixel-wise ensemble of the models. In this ensemble method, each binary model creates a prediction for the input image of size W × H , resulting in three distinct prediction maps P i x , y : x = 1 , ,   W ; y = 1 , ,   H ; and i = 1 , ,   3 , representing the artery, ureter, and nerve segmentations, respectively. Then, the merging algorithm checks every pixel separately in each of these maps. Here, the algorithm has three choices. Firstly, when each binary model predicts the pixel as the background, the final prediction will also be the background class. Secondly, when one of the models predicts only the pixel as positive, it is classified into the corresponding organ’s class. The third scenario is when at least two models predict the same pixel as positive. In this case, the class prediction with the highest confidence level is accepted as the final class for that pixel. That is, this ensemble algorithm provides a final labeled output image P x , y 0,1 , 2,3 , x = 1 , ,   W , y = 1 , ,   H , from the binary segmentation members’ prediction matrices P i x , y ,   i = 1 , , 3 , as follows:
P x , y = 0 ,    i f    i P i x , y < 0.5 ,   argmax i { 1,2 , 3 } P i x , y ,    o t h e r w i s e .
Notice that the label 0 in P x , y corresponds to the background class, and the further limitation in (1) that only pixels above the confidence level 0.5 are considered as positive, that is, accepted to correspond to one of the investigated organs. Figure 3 illustrates the results of the pixel-wise ensemble technique. As it can be seen in this figure, the input image flows through the binary segmentation networks that were trained for separate organs, providing three segmentation maps. Afterwards, these maps are fed into the above-mentioned algorithm and equation, which results in the final segmentation map as an output image.

2.3.2. Weighted Pixel-Wise Ensemble

The ensemble technique mentioned above was improved by weighing the member binary models with their respective accuracies w 1 , w 2 , and w 3 . These accuracy values indicated how well each binary member of the ensemble could predict its appropriate class. Using the former notations, the decision model given in (1) can be simply updated to incorporate this weighting approach as
w P x , y = 0 ,    i f    i w i P i x , y < 0.5 ,   argmax i { 1,2 , 3 } w i P i x , y ,    o t h e r w i s e .

2.3.3. Region-Based Ensemble

The region-based ensemble method was added as an extra step to the previously discussed pixel-wise ensemble. The first steps of the algorithm remained the same by selecting the final prediction for each pixel. To ensure that the final prediction map does not contain inconsistent regions within the classified organs, this ensemble algorithm examines every connected component in it.
Namely, we took all of the connected components separately and determined their final labels. We let R = [ r k ] denote the list of all r k connected components of the prediction matrix P x , y (or W P x , y ), noting that the cardinality of R (denoted by | R | ) may vary image by image since the investigated organs can appear as multiple connected components. Then, the algorithm iterated through all of these r k , k = 1 , , | R | regions to re-label all of their pixels ( x , y ) r k with the most dominant labels they enclosed, as follows:
R P ( x , y ) = argmax i { 1,2 , 3 } | { ( x , y ) r k : P x , y = i } | .
That is, each r k was re-labeled to the most frequent organ label it contained. Figure 4 depicts a flowchart of this region-based ensemble technique for an example input image. From the results of the pixel-wise ensemble process, it created a binary map of connected components; then, the equation above was used to examine the majority class of each region, which provided the final classes for all separable regions.

2.3.4. Weighted Region-Based Ensemble

Furthermore, similarly to the pixel-wise ensemble method, a weighted version can also be introduced for the region-wise one. For this purpose, using the former notations, the prediction matrix of the weighted pixel-base ensemble w P x , y can be re-labeled to that of the weighted region-based ensemble, as follows:
w R P ( x , y ) = argmax i { 1,2 , 3 } | { ( x , y ) r k : w P x , y = i } | .

2.4. Model Evaluation

We used two metrics to measure each model’s segmentation accuracy in our application, both globally at the image level and at the class level. This overall evaluation included the multi-class segmentation model (see Section 2.2), the binary semantic segmentation ones, and the pixel-wise/region-wise, non-weighted/weighted ensemble variants. The first applied metric was the Jaccard score [38], also known as the Intersection over Union (IoU), and this was calculated for each segmentation technique at the image level as
J a c a r d g t , P = | g t P | | g t P | ,
where g t stands for the manually drawn ground-truth mask of the input image and P denotes the prediction mask of the given segmentation model. Since the image backgrounds were quite large, they could have distorted the segmentation score. Thus, we also calculated the score at the class level, restricting to the corresponding organ only via
J a c c a r d c g t c , P c = | g t c P c | | g t c P c | ,
where g t c and P c are the masks of the ground truth and the segmentation algorithm restricted to the corresponding organ. By following our previous labeling, we also wrote J a c c a r d 1 , J a c c a r d 2 , and J a c c a r d 3 for the ureter artery, ureter, and nerve classes, respectively. Similar to the Jaccard score, we also calculated the Dice coefficient for all of the segmentation methods for both the image and class levels, as follows:
D i c e g t , P = | g t P | | g t | + | P |
D i c e c g t c , P c = | g t c P c | g t c + | P c | .
Moreover, we used the D i c e 1 , D i c e 2 , and D i c e 3 notations when we directly wanted to refer to the ureter artery, ureter, and nerves. We also noted that the number of images in the dataset was insufficient to train the network with multi-class segmentation properly. To overcome this problem, the images were augmented by cropping a total of 24 sub-images with a size of 512 × 512 pixels from each. This slicing algorithm randomly selected sub-images from the original one with a special condition. Namely, the selected sub-images needed to contain at least 1 / 3 of the labeled region. After performing the slicing algorithm on every image, the dataset expanded from its original size of 586 images to 14,100 images. The test dataset was kept in its original form, and all models were evaluated using the original test dataset of 56 images.

3. Results

This section summarizes and compares the performance of the proposed segmentation models. For the implementation of this research, we employed Python 3.9 using Tensorflow 2.12.0 and Keras 2.12.0 machine learning frameworks. The training and evaluation of the semantic segmentation models and its ensembles were performed using a NVIDIA Tesla K80 graphics card with 12 GB of memory. During our research, we tested several hyperparameters for training with the learning rate varying from 10 3 to 10 4 with five steps. At the end, the best results were achieved using a learning rate of 5 × 10 4 with a batch size of 16. Due to the lack of computation capacity, this was the maximum batch size we could test for that size of images. For each case, we used the Adam optimizer and trained the network until the validation loss converged to its minimum.

3.1. Multi-Class Semantic Segmentation Model

According to the evaluation of the test dataset, the original multi-class model does not show the capability to distinguish the essential regions. The scores corresponding to the class-level performance demonstrate well the inability of the model to differentiate the investigated structures. The only class where this model showed signs of having a predicting capability was the ureter (class 2), where it achieved much higher scores than those it achieved for the other two structures. Still, we can use these multi-class U-Net semantic segmentation results as a benchmark to conduct comparisons with the performances of the ensemble models later.

3.2. Decomposed Binary Models

The binary semantic segmentation models were trained five times each by splitting the dataset into five portions. The annotations for the binary segmentation networks were binarized regarding the class that the binary networks were trained for. In every training cycle, different parts of the dataset were used as validation datasets to fine-tune the model. The final evaluation of the binary semantic segmentation models was conducted on the test dataset. At both the image and class levels, we observed certain improvements regarding the multi-class semantic segmentation model.

3.3. Ensemble Models

The binary segmentation models were merged into a multi-class semantic segmentation model using the ensemble creation techniques discussed in Section 2.3. If we compare the ensemble techniques with each other, the pixel-wise technique has the lowest accuracy since it creates its predictions with the highest noise over the images. By applying weights, this noise reduces, which leads to a higher segmentation accuracy for the separate classes. By using the region-based ensemble technique, each connected region is assigned with a single label. This removes the inconsistent predictions within the regions, which enhances the performance by a significant margin. The summarized results for all tested models and ensembles are demonstrated in Table 1. If we take a closer look at the performance metrics, the ensemble approach outperforms the classic multi-class segmentation network with a large margin. For the task of uterine artery segmentation, our ensemble techniques scored 0.2260 more for the Jaccard scores and 0.2632 more for Dice on average. The same significant improvement can be seen for the nerve segmentation, with higher average Jaccard and Dice scores by 0.4834 and 0.5, respectively. The multi-class segmentation network achieved the best performance in ureter segmentation; however, the ensemble segmentation networks still made an improvement of 0.0221 for the Jaccard score and performed slightly worse for the Dice score, with a 0.0207 lower average score. These figures indicate that the overall segmentation performance improved a lot, with 0.1614 for the Jaccard score and 0.1051 for the Dice score demonstrating the improvement made by the ensemble segmentation techniques. Figure 5 visualizes the performance improvements achieved for the different organ detection tasks.
As part of our evaluation, we conducted a visual inspection to compare the ensemble models’ segmentation results with the test images’ ground-truth masks. Figure 6 illustrates how well the outputs of the ensembles fit after superimposing them onto the original RGB images. In these examples, we used the following color mapping: uterine artery—red; ureter—green; nerves—blue.

4. Discussion

In the previous sections, we presented an evaluation of our proposed ensemble techniques. Our work focused on capturing the differences between three different organs. Using decomposed segmentation networks was an efficient way to carry out this work since we could train each network to learn the features of the separate organs and try to differentiate them from all of the other organs as well. However, this process is harsh on time and computational complexity, which can be considered a weakness of the proposed solution since the evaluation contains four separate steps: the prediction of the three binary segmentation network and the ensemble method. As the ensemble of the multiple segmentation networks is only possible after the prediction of all separate networks, the time consumption is equivalent to that of three basic U-Net segmentation networks, which can be major setback for real-time segmentation. In our future work, we need to optimize this process to make the segmentation more efficient. However, the ensemble algorithms do not contain complex computational operations, so if the predictions of the binary networks are optimized or parallelized, the time consumption of the whole segmentation package can be reduced significantly.
The dataset we used for this work is unique for semantic segmentation tasks based on gynecological surgeries. This is one reason why our proposed solution was evaluated on this dataset alone. This may result in a bias in the trained segmentation models as they were only trained on this single dataset. However, the dataset used contains 38 different examinations, which can help the generalization of the models as they are able to see the organs with enough deviation in terms of shape, texture, or color.

5. Conclusions

In this work, we propose a method for semantically segmenting human organs during laparoscopic surgeries. In such datasets as the presented one, the original type of multi-class segmentation model fails to predict each represented structure accurately. To overcome this problem, binary semantic segmentation models were trained to distinguish one dedicated class from all others. Finally, these models were ensembled using two techniques that outperformed the original multi-class network by a large margin. Expanding the dataset could also make the multi-class segmentation model work. Still, since the annotation of surgery images is time-consuming and can only be performed by experts, other methods must be considered, like the presented one. Furthermore, more surgical images would enhance the performance of these binary segmentation models as well. The other main area for improvement in this current work involves lowering the time consumption of the segmentation models. The current research focused on image data; however, optimizing our method to be able to handle image sequences or videos of surgical experiments would make our solution a great benchmark for real-time decision support systems for hysterectomies. However, creating a real-time analysis system takes a lot of effort in terms of optimization logic, and implementation as well. One of the ways to achieve this could be through the fusion of binary models, so that the ensemble is effectively completed within a single segmentation model, which should reduce the computation time by a significant amount.

Author Contributions

Conceptualization, N.S. and B.H.; Methodology, N.S., D.K., A.H. and B.H.; Software, N.S. and D.K.; Validation, A.H. and B.H.; Resources, P.T.; Data curation, P.T.; Writing—original draft, N.S.; Writing—review & editing, A.H. and B.H.; Visualization, N.S.; Supervision, A.H. and B.H.; Project administration, B.H.; Funding Acquisition, B.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the project TKP2021-NKTA-34, implemented with the support provided by the National Research, Development, and Innovation Fund of Hungary under the TKP2021-NKTA funding scheme.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Ethics Committee of VKSZ 14-1-2015-0072, SCOPIA: Development of diagnostic tools based on endoscope technology. The data used for the samples are recorded anonymously and cannot be used to establish the identity of the patients. In this way, the participants were not required to provide consent for data sharing.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Aarts, J.W.; Nieboer, T.E.; Johnson, N.; Tavender, E.; Garry, R.; Mol, B.W.J.; Kluivers, K.B. Surgical approach to hysterectomy for benign gynaecological disease. Cochrane Database Syst. Rev. 2015, 8. [Google Scholar] [CrossRef] [PubMed]
  2. Clayton, R.D. Hysterectomy. Best Pract. Res. Clin. Obstet. Gynaecol. 2006, 20, 73–87. [Google Scholar] [CrossRef] [PubMed]
  3. Reich, H.; Decaprio, J.O.H.N.; Mcglynn, F.R.A.N. Laparoscopic hysterectomy. J. Gynecol. Surg. 1989, 5, 213–216. [Google Scholar] [CrossRef]
  4. Kluivers, K.B.; Johnson, N.P.; Chien, P.; Vierhout, M.E.; Bongers, M.; Mol, B.W. Comparison of laparoscopic and abdominal hysterectomy in terms of quality of life: A systematic review. Eur. J. Obstet. Gynecol. Reprod. Biol. 2008, 136, 3–8. [Google Scholar] [CrossRef] [PubMed]
  5. Councell, R.B.; Thorp, J.M., Jr.; Sandridge, D.A.; Hill, S.T. Assessments of laparoscopic-assisted vaginal hysterectomy. J. Am. Assoc. Gynecol. Laparosc. 1994, 2, 49–56. [Google Scholar] [CrossRef] [PubMed]
  6. Jones, R.A. Laparoscopic hysterectomy: A series of 100 cases. Med. J. Aust. 1993, 159, 447–449. [Google Scholar] [CrossRef] [PubMed]
  7. Woodland, M.B. Ureter injury during laparoscopy-assisted vaginal hysterectomy with the endoscopic linear stapler. Am. J. Obstet. Gynecol. 1992, 167, 756–757. [Google Scholar] [CrossRef] [PubMed]
  8. Manoucheri, E.; Cohen, S.L.; Sandberg, E.M.; Kibel, A.S.; Einarsson, J. Ureteral injury in laparoscopic gynecologic surgery. Rev. Obstet. Gynecol. 2012, 5, 106. [Google Scholar] [PubMed]
  9. Janssen, P.F.; Brölmann, H.A.M.; Huirne, J.A.F. Causes and prevention of laparoscopic ureter injuries: An analysis of 31 cases during laparoscopic hysterectomy in the Netherlands. Surg. Endosc. 2013, 27, 946–956. [Google Scholar] [CrossRef]
  10. Harris, W.J.; Daniell, J.F. Early complications of laparoscopic hysterectomy. Obstet. Gynecol. Surv. 1996, 51, 559–567. [Google Scholar] [CrossRef]
  11. Wood, E.C.; Maher, P.; Pelosi, M.A. Routine use of ureteric catheters at laparoscopic hysterectomy may cause unnecessary complications. J. Am. Assoc. Gynecol. Laparosc. 1996, 3, 393–397. [Google Scholar] [CrossRef] [PubMed]
  12. Tanaka, E.; Ohnishi, S.; Laurence, R.G.; Choi, H.S.; Humblet, V.; Frangioni, J.V. Real-time intraoperative ureteral guidance using invisible near-infrared fluorescence. J. Urol. 2007, 178, 2197–2202. [Google Scholar] [CrossRef] [PubMed]
  13. Dougherty, G. Digital Image Processing for Medical Applications; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  14. Bankman, I. (Ed.) Handbook of Medical Image Processing and Analysis; Elsevier: Amsterdam, The Netherlands, 2008. [Google Scholar]
  15. Miao, S.; Wang, Z.J.; Liao, R. A CNN regression approach for real-time 2D/3D registration. IEEE Trans. Med. Imaging 2016, 35, 1352–1363. [Google Scholar] [CrossRef] [PubMed]
  16. Harangi, B.; Hajdu, A.; Lampe, R.; Torok, P. Differentiating ureter and arteries in the pelvic via endoscope using deep neural network. In Proceedings of the 10th International Symposium on Image and Signal Processing and Analysis, Ljubljana, Slovenia, 18–20 September 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 86–89. [Google Scholar]
  17. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  18. Yu, F.; Song, E.; Liu, H.; Zhu, J.; Hung, C.-C. Laparoscopic image-guided system based on multispectral imaging for the ureter detection. IEEE Access 2018, 7, 3800–3809. [Google Scholar] [CrossRef]
  19. Taha, A.; Lo, P.; Li, J.; Zhao, T. Kid-net: Convolution networks for kidney vessels segmentation from ct-volumes. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Cham, Switzerland, 2018; pp. 463–471. [Google Scholar]
  20. Kletz, S.; Schoeffmann, K.; Benois-Pineau, J.; Husslein, H. Identifying surgical instruments in laparoscopy using deep learning instance segmentation. In Proceedings of the 2019 International Conference on Content-Based Multimedia Indexing (CBMI), Dublin, Ireland, 4–6 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–6. [Google Scholar]
  21. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 September 2017; pp. 2961–2969. [Google Scholar]
  22. Hasan, S.K.; Linte, C.A. U-NetPlus: A modified encoder-decoder U-Net architecture for semantic and instance segmentation of surgical instruments from laparoscopic images. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 7205–7211. [Google Scholar]
  23. Maqbool, S.; Riaz, A.; Sajid, H.; Hasan, O. m2caiseg: Semantic segmentation of laparoscopic images using convolutional neural networks. arXiv 2020, arXiv:2008.10134. [Google Scholar]
  24. Sagi, O.; Rokach, L. Ensemble learning: A survey. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2018, 8, e1249. [Google Scholar] [CrossRef]
  25. Knerr, S.; Personnaz, L.; Dreyfus, G. Single-layer learning revisited: A stepwise procedure for building and training a neural network. In Neurocomputing; Springer: Berlin, Heidelberg, 1990; pp. 41–50. [Google Scholar]
  26. Anand, R.; Mehrotra, K.; Mohan, C.; Ranka, S. Efficient classification for multiclass problems using modular neural networks. IEEE Trans. Neural Netw. 1995, 6, 117–124. [Google Scholar] [CrossRef]
  27. Galar, M.; Fernández, A.; Barrenechea, E.; Bustince, H.; Herrera, F. An overview of ensemble methods for binary classifiers in multi-class problems: Experimental study on one-vs-one and one-vs-all schemes. Pattern Recognit. 2011, 44, 1761–1776. [Google Scholar] [CrossRef]
  28. Lindsey, T.; Lee, J.-J. Automated cardiovascular pathology assessment using semantic segmentation and ensemble learning. J. Digit. Imaging 2020, 33, 607–612. [Google Scholar] [CrossRef]
  29. Khan, I.A.; Asaduzzaman, F.; Fattah, S.A. An Automatic Ocular Disease Detection Scheme from Enhanced Fundus Images Based on Ensembling Deep CNN Networks. In Proceedings of the 2020 11th International Conference on Electrical and Computer Engineering (ICECE), Dhaka, Bangladesh, 17–19 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 491–494. [Google Scholar]
  30. Chen, Y.; Wang, Y.; Gu, Y.; He, X.; Ghamisi, P.; Jia, X. Deep learning ensemble for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1882–1897. [Google Scholar] [CrossRef]
  31. Alosaimi, N.; Alhichri, H. Fusion of CNN ensemble for Remote Sensing Scene Classification. In Proceedings of the 2020 3rd International Conference on Computer Applications & Information Security (ICCAIS), Riyadh, Saudi Arabia, 19–21 March 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1–6. [Google Scholar]
  32. Yazdizadeh, A.; Patterson, Z.; Farooq, B. Ensemble convolutional neural networks for mode inference in smartphone travel survey. IEEE Trans. Intell. Transp. Syst. 2019, 21, 2232–2239. [Google Scholar] [CrossRef]
  33. UD Ureter-Artery-Nerve Dataset. Norbert Serban. UD Ureter-Uterine Artery-Nerve Dataset. IEEE Dataport. 2023. Available online: https://dx.doi.org/10.21227/q2dd-yt09 (accessed on 4 March 2024).
  34. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  35. Li, W.; Lambert-Garcia, R.; Getley, A.C.; Kim, K.; Bhagavath, S.; Majkut, M.; Rack, A.; Lee, P.D.; Leung, C.L.A. AM-SegNet for additive manufacturing in situ X-ray image segmentation and feature quantification. Virtual Phys. Prototyp. 2024, 19, e2325572. [Google Scholar] [CrossRef]
  36. Ouyang, S.; He, B.; Luo, H.; Jia, F. SwinD-Net: A lightweight segmentation network for laparoscopic liver segmentation. Comput. Assist. Surg. 2024, 29, 2329675. [Google Scholar] [CrossRef] [PubMed]
  37. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; PMLR: Cambridge, MA, USA, 2019; pp. 6105–6114. [Google Scholar]
  38. Jaccard, P. Distribution de la flore alpine dans le bassin des Dranses et dans quelques régions voisines. Bull Soc. Vaudoise Sci. Nat. 1901, 37, 241–272. [Google Scholar]
Figure 1. Annotation process for the frames extracted from surgical videos.
Figure 1. Annotation process for the frames extracted from surgical videos.
Sensors 24 02926 g001
Figure 2. The architecture of the U-Net network with the EfficientNet-b3 encoder.
Figure 2. The architecture of the U-Net network with the EfficientNet-b3 encoder.
Sensors 24 02926 g002
Figure 3. Flowchart of the pixel-wise ensemble technique.
Figure 3. Flowchart of the pixel-wise ensemble technique.
Sensors 24 02926 g003
Figure 4. Flowchart of the region-based ensemble technique.
Figure 4. Flowchart of the region-based ensemble technique.
Sensors 24 02926 g004
Figure 5. Comparison between the reference multi-class semantic segmentation method and the proposed ensemble methods according to the Jaccard (a) and Dice (b) scores.
Figure 5. Comparison between the reference multi-class semantic segmentation method and the proposed ensemble methods according to the Jaccard (a) and Dice (b) scores.
Sensors 24 02926 g005
Figure 6. Visual inspection of the segmentation results: input image (a); ground truth (b); pixel-wise ensemble (c); weighted pixel-wise ensemble (d); region-based ensemble (e); weighted region-based ensemble (f); uterine artery—red; ureter—green; nerve—blue.
Figure 6. Visual inspection of the segmentation results: input image (a); ground truth (b); pixel-wise ensemble (c); weighted pixel-wise ensemble (d); region-based ensemble (e); weighted region-based ensemble (f); uterine artery—red; ureter—green; nerve—blue.
Sensors 24 02926 g006
Table 1. Performance of the different models and ensembles on the test dataset.
Table 1. Performance of the different models and ensembles on the test dataset.
Jaccard1Jaccard2Jaccard3Dice1Dice2Dice3JaccardDice
Multi-class U-Net0.02060.31110.06530.05140.38860.09970.70960.8252
Binary U-Net model for the ureter class0.2272 ±
0.0173
--0.3028 ±
0.0108
--0.8192 ±
0.0071
0.8928 ±
0.0041
Binary U-Net model for the uterine artery class-0.3190 ±
0.0849
--0.3880 ±
0.0822
-0.8565 ± 0.02610.9189 ±
0.0158
Binary U-Net model for the nerve class--0.1262 ±
0.0955
--0.1702 ±
0.1023
0.8173 ±
0.0690
0.8972 ±
0.0438
Pixel-wise ensemble0.19220.25030.53320.25830.30120.59360.87310.9272
Weighted pixel-wise ensemble0.25620.42010.55330.33100.49990.60100.88830.9358
Region-based ensemble0.26630.33000.55360.33160.36450.60190.84610.9290
Weighted region-based ensemble0.27180.33250.55480.33720.36810.60250.87640.9292
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Serban, N.; Kupas, D.; Hajdu, A.; Török, P.; Harangi, B. Distinguishing the Uterine Artery, the Ureter, and Nerves in Laparoscopic Surgical Images Using Ensembles of Binary Semantic Segmentation Networks. Sensors 2024, 24, 2926. https://doi.org/10.3390/s24092926

AMA Style

Serban N, Kupas D, Hajdu A, Török P, Harangi B. Distinguishing the Uterine Artery, the Ureter, and Nerves in Laparoscopic Surgical Images Using Ensembles of Binary Semantic Segmentation Networks. Sensors. 2024; 24(9):2926. https://doi.org/10.3390/s24092926

Chicago/Turabian Style

Serban, Norbert, David Kupas, Andras Hajdu, Peter Török, and Balazs Harangi. 2024. "Distinguishing the Uterine Artery, the Ureter, and Nerves in Laparoscopic Surgical Images Using Ensembles of Binary Semantic Segmentation Networks" Sensors 24, no. 9: 2926. https://doi.org/10.3390/s24092926

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop