Next Article in Journal
A High-Sensitivity Fiber Optic Soil Moisture Sensor Based on D-Shaped Fiber and Tin Oxide Thin Film Coatings
Previous Article in Journal
ARSOD-YOLO: Enhancing Small Target Detection for Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MugenNet: A Novel Combined Convolution Neural Network and Transformer Network with Application in Colonic Polyp Image Segmentation

1
School of Mechanical and Power Engineering, East China University of Science and Technology, Shanghai 200237, China
2
Department of Engineering, Purdue University, West Lafayette, IN 47907, USA
3
Department of Mechanical Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(23), 7473; https://doi.org/10.3390/s24237473
Submission received: 30 September 2024 / Revised: 18 November 2024 / Accepted: 21 November 2024 / Published: 23 November 2024
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Accurate polyp image segmentation is of great significance, because it can help in the detection of polyps. Convolutional neural network (CNN) is a common automatic segmentation method, but its main disadvantage is the long training time. Transformer is another method that can be adapted to the automatic segmentation method by employing a self-attention mechanism, which essentially assigns different importance weights to each piece of information, thus achieving high computational efficiency during segmentation. However, a potential drawback with Transformer is the risk of information loss. The study reported in this paper employed the well-known hybridization principle to propose a method to combine CNN and Transformer to retain the strengths of both. Specifically, this study applied this method to the early detection of colonic polyps and to implement a model called MugenNet for colonic polyp image segmentation. We conducted a comprehensive experiment to compare MugenNet with other CNN models on five publicly available datasets. An ablation experiment on MugenNet was conducted as well. The experimental results showed that MugenNet can achieve a mean Dice of 0.714 on the ETIS dataset, which is the optimal performance on this dataset compared to other models, with an inference speed of 56 FPS. The overall outcome of this study is a method to optimally combine two methods of machine learning which are complementary to each other.

1. Introduction

Colonoscopy is an effective technique for detecting colorectal polyps, which are polypoid lesions on the surface of the colonic mucosa within the intestinal lumen. Colorectal polyps commonly lead to intestinal bleeding and inflammatory responses, and in severe cases, they may progress to colon cancers. Colonoscopy is the most effective way to prevent colon cancers, but the procedure may increase the risk of perforation and cause severe discomfort to the patient [1,2,3]. The detection of abnormal polyps is carried out by means of image processing in colonoscopy.
Previous research indicated that the clinical miss rate for colonic polyps can be as high as 25% [4]. The segmentation of polyps from colonoscopy images is the first step in the early detection of colon cancers. Polyp segmentation can be accomplished using graph-based methods (e.g., Grabcut [5], Graphcut [6]), which model an image as a graph structure and employ graph cut operations to separate foreground and background regions. These methods rely on pixel similarity, calculating connection weights between pixels based on low-level features such as color and brightness. Typically, these methods require user intervention, (such as selecting a region or initial labeling) to help establish foreground and background regions for segmentation. While these methods perform well on small images, their computational efficiency decreases as the image resolution increases, limiting their suitability for real-time processing. In addition, due to the variation in size, color, and texture among polyps of similar types, and the unclear boundaries between polyps and the surrounding mucosa, polyp segmentation remains a challenging task. The accuracy and efficiency of polyp segmentation can be improved by deep learning methods [7].
The basic idea behind the study reported in this paper is to combine or fuse Transformer and CNN due to their nearly complementary ability, according to the hybridization principle [8]. The study reported in this paper proposed a new neural network coined MugenNet. Comprehensive experiments were conducted with the MugenNet on five public colon polyp datasets using six performance indicators or indexes. The study also compared MugenNet with other ten neural networks. These models are listed in Section 3. The main contributions of our work are summarized as follows. (1) A novel deep network is proposed for real-time, accurate polyp segmentation. By combining the CNN and Transformer architectures, the network leverages the CNN’s capacity for capturing local information alongside the Transformer’s strength in processing global context, significantly improving segmentation accuracy and speed. (2) We designed a fusion module called the Mugen module, which can dynamically adjust importance weights for each channel. Experiments demonstrate that this module effectively integrates global features while focusing on pixel-level edge features, which is particularly crucial for accurate polyp segmentation. (3) On five public colon polyp datasets, MugenNet achieves an outstanding performance in comparison with the traditional polyp segmentation methods in terms of training and processing speed, making MugenNet suitable for real-time clinical video data streaming applications.
The remaining part of the paper is organized as follows. Section 2 discusses the idea, method, and their implementation into the model MugenNet. Section 3 presents the experiment to validate MugenNet. Section 4 details the conclusion, along with the discussion of some future works.

2. Related Work

2.1. CNNs for Image Segmentation

Convolutional neural network (CNN) is a well-known image segmentation method that extracts features from pixels or raw information blocks using convolutional operators [9]. U-Net is a classic model for medical image segmentation, with its encoder–decoder structure effectively capturing multi-scale features, making it suitable for a wide range of medical applications [10]. Numerous researchers have proposed improvements upon the U-Net architecture, leading to variants such as ResU-Net [11], EGE-UNet [12], and U-Net v2 [13], which offer enhanced feature extraction capabilities and optimized information transmission. However, the shortcomings of CNN include (1) computational overhead in training of CNN, (2) not effectiveness in dealing with the biomedical images with low resolution, and (3) potential to be over-fitting [14].

2.2. Self-Attention Mechanism/Transformers

Transformer is a machine learning process, which takes a series of symbols as the input, and produces the semantics from this series as the output [15]. Transformer is based on the self-attention mechanism. The self-attention mechanism refers to placing more attention on the relevance of a chunk of information. Suppose that there are three pieces of information, A, B, C. If A is closer to B than to C, a higher attention would be given to B than to C (with respect to A). The Vision Transformer (ViT) represents a significant advancement in the application of Transformer architectures to computer vision tasks [16]. In segmentation tasks, ViT can be combined with upsampling or decoding modules to model dependencies between regions, producing more refined segmentation results.

2.3. Transformers to Complement CNNs

To a 2D image, CNN does not have different attention values in the scanning process of the chunks of information [17]. This feature of CNN has two disadvantages compared to Transformer: (1) the computational cost with Transformer will be much lower than that with CNN [18], because CNN must have a considerable amount of redundant computations; (2) Transformer is free of the overfitting problem, suggesting that it be more accurate than CNN. However, one shortcoming of the Transformer/self-attention mechanism is that the approach is highly dependent on the assessment of the relevance, which is further specific to different applications. For instance, in the database called ImageNet [19], such an assessment must rely on external datasets, which is in essence a kind of pre-training process. This shortcoming of Transformer contrasts with the merit of CNN (i.e., the equity among all chunks of information). Overall, CNN-based models [20,21] are unable to capture the global contextual information but are able to capture local information well. However, Transformer-based models [22,23] behave vice versa. Therefore, combining CNN and Transformer is a promising idea.
Currently, the research on effectively combining CNN and Transformer models is still limited. For instance, TransUNet [24] accomplished the combination of CNN and Transformer by employing a chain-like architecture. The method is well-suited for organ recognition tasks because organ boundaries are relatively clear, and uncertainty levels are relatively low. For the definition of uncertainty, we refer to [25]. The system also cannot perform parallel processing, which is one of the important merits of the Transformer. Swin Transformer introduced a multi-scale feature extraction mechanism with hierarchical windows, establishing local attention within each window and using cross-window connections [26]. This approach addressed the high computational cost issue found in ViT. Although the model facilitates information propagation through cross-window connections, the capture of long-range dependencies remains limited.

3. The Proposed MugenNet Model

MugenNet is mainly composed of a CNN decoder based on Resnet-34 and a Transformer decoder based on ViT, two of which are run in parallel, as shown in Figure 1. The following sections will discuss MugenNet in detail.

3.1. Design of the CNN Branch

ResNet-34 [27] was chosen as the CNN branch, as a trade-off between retaining and losing more local information during training. The proposed model used two up-sampling operations to generate three matrices, which stored the feature maps of different scales; as depicted in the inset in the upper right part of Figure 1. These feature maps can be utilized to progressively restore the spatial resolution in subsequent processes. The matrix obtained through the convolution operation in the CNN branch (the right side of Figure 1 served as the input to the Mugen module and fused with the result obtained from the Transformer branch, which will be discussed in the next section.

3.2. Design of the Transformer Branch

A pre-trained Vision Transformer (Vit-B/16) was used as the Transformer branch in our model. It is a deep learning model that segments images by dividing them into patches and using self-attention mechanisms to capture global relationships [16]. The input image was divided into 16 blocks, and a self-attention operation was performed on each block. We used two up-sampling operations to generate feature maps, matching their dimensions with the results obtained from the CNN branch. Assuming the input image is x R H × W × C , where H and W represent the length and the width of the original image, respectively, and C is the channel of the image (if it is a color image, C = 3 , otherwise C = 1 ). First, divide x into 4 × 4 patches, and every patch is x p R N × ( P 2 × C ) , where P is the side length of x p , ( P , P ) is the resolution of each patch, and N = H W / P 2 , indicating the number of patches. Then, add a location code to every patch and put each patch into a sequence as the Transformer modules (see the left part of Figure 1). Therefore, each Transformer module consisted of layer normalization (LN), multi-head self-attention (MSA) and multi-layer perceptron (MLP), as depicted in the left part of Figure 1. The use of LN instead of batch normalization was because LN helped us to increase the stability during training with a high learning rate [18,28]. MSA was used to calculate the relevant information learned in different subspaces. MLP was a stack of activation functions and served as the output layer of CNN.
The activation function used is the so-called softmax, which was used to compute the three feature maps (Q, K and V), i.e.,
N o u t = s o f t m a x ( q i k T D h ) v
In Equation (1), N o u t represents the output of each attention module, which was then input into the Mugen module. For the post-processing process, we refer to [29]. Specifically, two consecutive up-sampling convolution layers were connected to the output layer of the Transformer branch to recover the spatial resolution. The feature maps of different dimensions were preserved and fused with the output of the CNN branch.

3.3. Mugen Module

The Mugen module was used to combine the global chunks of information from Transformer and CNN, respectively. The process of the combination was composed of two parts. The first part was the Squeeze and Excitation (SE) block [29], and the second part was the module that was built based on the channel attention. The SE block can recalibrate the corresponding characteristics of the channels by explicitly modeling the inter-dependence between the channels and can improve the performance of the neural network without an increase in the computational overhead. The channel attention mechanism learns the weights of importance for each channel, allowing it to dynamically adjust the feature responses of each channel. This enables the network to effectively capture and utilize information from different channels. Such a mechanism aids in improving the performance of the network for specific tasks, and has shown significant improvements in various visual tasks, including image classification, object detection, and semantic segmentation.
In the CNN branch, the channel attention module was adopted [30]. After global max-pooling, the feature map was input into the neural network consisting of two hidden layers. The number of neurons in the first layer was C / r and the number of neurons in the second layer was C, where C was the number of channels and r was the reduction rate. The proposed model removed the global average-pooling part from the Convolutional Block Attention Module (CBAM) to focus more on the pixel-level edge information in the image segmentation task. Finally, the sigmoid function was used to generate the channel attention features.
After the above treatment, two feature maps were obtained: one from the Transformer branch and the other from the CNN branch, respectively. Inspired by ResNet, we added a residual connection architecture (Figure 1) before the final output from the Mugen module. ResNet helps to alleviate the vanishing or exploding gradient problem during the training process.
The steps in the Mugen module include the following computations:
t ^ i = S e A t t e n ( t i ) r ^ i = C h a n n e l A t t e n ( r i ) z i o u t = R e s i d u a l ( t ^ i , r ^ i , t i , r i )
After the Mugen module, a stack of hidden layers with convolution and batch normalization was used to reduce the output dimension to one. The characteristic graph was used as the intermediate output z i o u t . We used the progressive up-sampling method to process the characteristic graph obtained from the Transformer and CNN branches.
As shown in Figure 1, initially there were t 0 , r 0 R H 16 W 16 D 0 . After two consecutive up-sampling operations, we obtained t 1 , r 1 R H 8 W 8 D 1 and t 2 , r 2 R H 4 W 4 D 2 , which restored the spatial resolution to the same dimension as the input image. Finally, we input t i , r i into the three Mugen modules, respectively, perform feature fusion, and obtain three feature maps, forming the semantic segmentation map.

3.4. Details of the Transfer and Loss Function

In the Mugen module, we adopted an architecture to combine Transformer and CNN, called Feature Pyramid (FP), along with the residual, by refining the one proposed by [31]. After the top-down aggregation, the residual architecture was used to gather strong features, which was expected to improve the accuracy of the model. It is noted that compared with other bi-directional methods, the residual FP can more effectively achieve high-precision detection of targets.
From the Mugen module, three feature maps of different dimensions can be obtained. Then, attention gate (AG) was used to restore the resolution of attention. It is worth noting that, according to [32], AG can focus on target architectures of different shapes and sizes, suppress irrelevant regions in the input image, and highlight the target features of specific tasks [33]. The architecture of AG in the Mugen is shown in Figure 1.
From Figure 2, it can be seen that we used two up-sampling operations to gradually restore the original resolution of the feature map. First, we set the output of the first Mugen module ( y 1 o u t ) equal to z 1 o u t R H 8 W 8 D 1 . Then, z 1 o u t and the output of the second Mugen module ( y 2 o u t ) were put into the first attention gate to obtain z 2 o u t R H 4 W 4 D 2 . After that, we put z 2 o u t and the output of the third Mugen module ( y 3 o u t ) into the second attention gate to obtain z 3 o u t R H 2 W 2 D 3 . Finally, a separate up-sampling of z 3 o u t was conducted to obtain the final segmentation graph z o u t R H W D .
The loss function was defined as L = L I o U ω + n L B C E ω during the training process, according to [32,34], where L I o U ω was the global weighted average I o U loss (added to the global constraint) and L B C E ω was the weighted binary cross entropy loss (added to the pixel level constraint). Further, n was a hyperparameter used to allocate the proportion of the I o U loss and the B C E loss in the total loss L. Compared with the standard I o U loss, which is commonly used in the segmentation tasks, the global average I o U loss highlights its the importance of complex sample pixels by increasing the their weight of the complex sample pixels. At the same time, the weighted binary cross entropy loss pays more attention to the complex sample pixels, which can effectively extract the pixel-level edge features at the pixel level. Considering that the goal of the model was of the colonic polyp segmentation and the edge features were not obvious, we took set n to be 6 / 5 to increase the proportion of the pixel-level constraints in the overall total loss function. S x u p was up-sampled to the same size as the ground-truth graph image G. The proposed overall-loss function of MugenNet for colonic polyp detection can then be expressed by:
L t o t a l = α L G , S t u p + β L G , S r u p + γ L G , S z u p
where S t u p and S r u p represent the characteristic graph obtained from the Transformer branch and CNN branch, respectively. S z u p represents the output of the semantic segmentation graph after up-sampling. The parameters α , β , γ are adjustable hyperparameters.

3.5. Implementation

We implemented the MugenNet module (referred to as the model hereafter) in the PyTorch framework on an NVIDIA RTX 3080 GPU. The resolution of input images was shaped to 256 × 192 . The dataset was divided into training, validation, and test sets in the ratio of 7:2:1. The Adam optimizer was used to adjust the overall parameters, with a learning rate of 1 × 10 4 . The batch size was set to 16. The final prediction result was generated after the sigmoid operation.

4. Experiment

We compared the MugenNet model with some existing CNN models, which belong to the category of CNN models in terms of abilities such as learning, generalization, and qualitative segmentation.

4.1. The Dataset and Model

The experiment was conducted on five publicly available polyp segmentation datasets, namely CVC-300 [35], CVC-ClinicDB [36], CVC-ColonDB [34], ETIS [37], and Kvasir [38]. These datasets contain precisely annotated endoscopic images, covering various resolutions, polyp sizes, and complexities, making them well-suited for polyp detection, segmentation, and model performance evaluation in colorectal imaging. The CVC-300, CVC-ClinicDB, CVC-ColonDB, ETIS, and Kvasir datasets contain 300, 612, 380, 196, and 1000 images, respectively. The first four datasets consist entirely of polyp images, while the Kvasir dataset includes 500 images with polyps and 500 images without polyps. We randomly selected 1450 images from Kvasir and CVC-ClinicDB to train our model, with the weights initialized based on the training weights of DeiT-small [39] and resnet-34 [27]. We then compared our model with ten other convolutional neural network models for biomedical image segmentation, namely U-Net [10], U-Net++ [40], ResUNet++ [41], SFA [42], PraNet [43], DCRNet [44], EU-Net [9], SANet [45], ACSNet [9], and HaeDNet [46].

4.2. Training Setting and Performance Indicators

To assess the generalization ability of the model, besides the two datasets used for training the model, we also tested the model on three additional datasets, namely CVC-300, CVC-ColonDB, and ETIS.

4.3. Performance Indicators

To evaluate the proposed MugenNet for the segmentation of colon polyp images, we used the performance indicators of m I o U , m D i c e , M A E loss, F β ω , S α , and E ξ . The mathematical expression for the definition of each performance indicator along with its significance in the image segmentation task is introduced below.
The full name of I o U is Intersection over Union, which represents the ratio of intersection and union between the bounding box and ground truth. The average I o U ( m I o U ) can be calculated by
m I o U = 1 n i = 0 n A i B i A i B i
It is noted that during the training process of a segmentation model, the value of the I o U loss will be used to evaluate whether the prediction result of the model is good or valid. If I o U > 0.5 , the prediction is considered valid. Additionally, we also calculated the I o U value relative to the ground truth for each training batch, and we then took the average as the m I o U loss for this specific training batch.
The Dice coefficient is a measure of similarity between two sets (for example, A and B) and is used for medical image segmentation. The formula for the mean Dice coefficient ( m D i c e ) is as follows:
m D i c e = 1 n i = 0 n 2 A i B i A i + B i
The range of the Dice coefficient is (0, 1). The closer the coefficient is to 1, the higher the similarity between sets A and B. We calculated the Dice coefficient between the model’s predicted results and the ground truth to evaluate the reliability of the model in colon polyp image segmentation.
The full name of M A E is the Mean Absolute Error, which is used to calculate the error between predicted values and the ground truth. M A E is calculated by
M A E = 1 n i = 0 n A i B i
The advantage of M A E is that it is insensitive to outliers. This study took it as one of the criteria to evaluate the reliability of prediction results.
The parameters for the precision and recall metrics include: TP (True Positive), TN (True Negative), FP (False Positive), and FN (False Negative). Each pixel was compared in the map of the prediction result with the ground truth and the prediction result was classified into (i) correct ( D i = G ( i ) ) or (ii) incorrect ( D i G ( i ) ) . The four attributes are then calculated by
T P = D G T N = ( 1 D ) ( 1 G ) F P = D ( 1 G ) F N = ( 1 D ) G
Then, the precision and recall were calculated by
P r e c i s i o n = T P T P + F P R e c a l l = T P T P + F N
We also used the weighted F-measure ( F β ω ) [47], S-measure ( S α ) [48] and E-measure ( E ξ ) [49] to evaluate the model performances, similar to [43]. These three metrics are selected to evaluate the precision and recall of the model during testing, as well as to assess the model’s performance. The formulas for these three metrics are as follows.
The weighted F-measure can be calculated by
F β ω = 1 + β 2 P r e c i s i o n ω · R e c a l l ω β 2 · P r e c i s i o n ω + R e c a l l ω
where ω represents weight, and β is the adjustable coefficient. In this study, we took ω = 1 , β = 1 / 2 . It is worth noting that the weighted F-measure is a robust metric because it incorporates both precision and recall. The S-measure is calculated by
S = α · S 0 + ( 1 α ) · S r S 0 = μ · S F G + ( 1 μ ) · S B G S F G = 2 x ¯ F G ( x ¯ F G ) 2 + 1 + 2 λ · σ x F G S B G = 2 x ¯ B G ( x ¯ B G ) 2 + 1 + 2 λ · σ x B G
where FG stands for foreground and BG stands for background. We set α = 0.5 , μ = 0.5 , λ = 1 . x F G represents the probability of the foreground region in the predicted result and the ground truth. x B G represents the probability of the background region in the predicted result and the ground truth. The S-measure will be used to assess the structural similarity of targets in the region.
The E-measure is calculated by
E ξ = 1 w · h x = 1 w y = 1 h 2 φ G T · φ F M φ G T · φ G T + φ F M · φ F M φ I = I μ I · A
where I is the input binary foreground map. μ I is the global mean of I. A is a matrix where all elements have a value of 1, and its size is to the same as that of matric I. The E-measure was used to evaluate the performance of the model at the pixel level. We took the mean value of E-measure.

4.4. Comparison of the Performance

The results of the comparison are shown in Table 1, where “Null” means that the corresponding results are not provided in the literature. The best performance in each column of the table is highlighted in bold. The data of other existing models were obtained from the corresponding literature and their implementation was carried out using their published code.
From the results on the five different datasets (Table 1, Table 2, Table 3, Table 4 and Table 5, it can be found that the proposed MugenNet shows execellent performance in the CVC-ColonDB and ETIS polyp segmentation datasets (Table 2 and Table 3), showing the best overall performance across the three metrics. It is worth mentioning that our training datasets are sourced from Kvasir and CVC-ClinicDB (Table 4 and Table 5), indicating that our model has good generalization ability and is suitable for the predicting segmentation results on unknown datasets. Although the results on the other three datasets indicate that our model has not achieved the best performance, it is not far from PraNet (only short by 1.6%), which is the current SOTA (the state of the art) performance. We compared the inference time of the different models in Table 6. The results show that with the same training dataset, our model has reduced the inference time by 12% compared with PraNet. At the same time, our model significantly outperforms traditional biomedical image semantic segmentation model (such as U-Net); especially on the ETIS dataset, our model’s performance improvement reaches as high as 13.7% compared with the PraNet model.
The poor performance of the models such as SFA and U-Net on the ETIS dataset suggests that they have a weak generalization ability on unknown datasets. In contrast, our model outperforms the traditional biomedical image segmentation models.
Figure 3 displays the polyp segmentation results of MugenNet on the Kvasir dataset, comparing our model with U-Net, U-Net++, SFA and PraNet, where GT represents the ground truth. Our model can accurately locate polyps of different sizes and textures, then generate semantic segmentation maps. It is worth noting that the results of other models are compared with the results from [43].
In addition to the comparison with the four existing models—U-Net, U-Net++, SFA, PraNet in Figure 3—on the Kvasir dataset, we also tested our model on the other four datasets (CVC-300, CVC-ClinicDB, CVC-ColonDB, ETIS-LaribPolypDB,). The results are shown in Figure 4. From Figure 4, our model (MugenNet) can accurately locate the position and size of polyps on the five different datasets. The results show that our model possesses stable and good generalization ability.
We also trained our model (MugenNet) on the datasets of CVC-300, CVC-ClinicDB, CVC-ColonDB, ETIS-LaribPolypDB, with the results shown in Figure 5. The examples of the datasets can be found in Figure 5 under original image and ground truth. It can be seen from Figure 5 that our model performs well when trained on the other four polyp datasets. This shows that our model has good robustness (because its performance is insensitive to the training datasets) and can quickly adapt to multiple different datasets.
From the pre-training weight of the DeiT distillation neural network, as shown in Table 6, our model can converge at less than 30 epochs, and the training time is only about 15 min when the batch size is set to 16. L R represents Learning Rate, F P S represents Frames Per Second. Our model’s real-time running speed is approximately 56 frames per second (FPS), which means that our model can be used to process video streaming data during colonoscopy examinations to perform real-time polyp image segmentation tasks.
We tested the Frames Per Second ( F P S ) of ten models on the training dataset to evaluate the inference speed of the models, as shown in Figure 6. Since the FPS of U-Net and U-Net++ are both below 10 (as shown in Table 6), their inference speed is too slow; hence, they are not shown in Figure 6. It can be observed from Figure 6 that the average F P S of our model reaches 56. It is noted that when F P S is greater than 30, the model was considered reasonable for video streaming data [43]. It can be seen from Figure 6 and Table 6 that our model (MugenNet) has advantages in processing video stream data for the colonic polyp image segmentation.
The comparative results indicate that our model performs well on three unseen datasets, far surpassing traditional CNN-based biomedical image segmentation models such as U-Net, U-Net++, SFA and ResUNet++, and even slightly outperforming PraNet on certain datasets.
It is worth mentioning that the models we compared are all based on convolutional neural networks, while our model (MugenNet) combines Transformer and CNN for colon polyp image segmentation. The experiment results show that the incorporation of Transformer can effectively improve the performance of neural networks, achieving a faster processing speed. In essence, our model (MugenNet) possesses both the global learning capability of Transformer for capturing overall information and the local learning capability of CNN for focusing on specific regions. Therefore, our model achieves an excellent performance in the colon polyp image segmentation. The statistical test on eight models are shown in Figure 7.

4.5. Ablation Studies

In this section, five sets of ablation experiments were conducted to evaluate the effectiveness of each component of MugenNet on two different datasets. The results are shown in Table 7, where TB represents the Transformer branch, CB represents the CNN branch, and MM represents the Mugen module.
We used three indicators to evaluate the models, i.e., m D i c e , m I o U and F β ω . Among them, m D i c e and m I o U were used to evaluate the accuracy of the models, while F β ω was used to evaluate the precision and tecall of the models.
We removed the Transformer branch, the CNN branch, and the Mugen module from MugenNet, respectively, to examine their performances. The result shows that our model (MugenNet) significantly improves the accuracy of semantic segmentation. Compared with the model without the Transformer branch, our model performs better on the ColonDB dataset than the other two cases. The result shows that m I o U increased from 0.591 to 0.678, and m D i c e increased from 0.667 to 0.758.
The results in Table 7 show that our model can segment colonic polyp image accurately. Our model outperforms the other two models in the ablation study, where either the CNN or the Transformer was omitted. Compared with the backbone neural network, the performance of our model in m I o U improved about 43.34% on the tested dataset (CVC-ColonDB). The results prove the superiority of our model (MugenNet).

5. Conclusions

This study proposed a new neural network (MugenNet), which combines Transformer and CNN. MugenNet shows excellent results in performing colonoscopy polyp segmentation tasks. MugenNet was tested on five polyp segmentation datasets (CVC-300, CVC-ClinicDB, CVC-ColonDB, ETIS-LaribPolypDB, Kvasir). Compared with traditional CNNs, our model achieves higher accuracy and faster processing speed. The results of the experiments show that (1) our model can achieve a mean Dice of 0.714 on the most challenging ETIS dataset, specifically, a 13.7% improvement over the current state-of-the-art CNN model (i.e., PraNet), and (2) our model can achieve a better accuracy and faster image processing speed than the traditional CNN, specifically, with the speed of our model being 12% higher than that of PraNet (only taking 30 epochs within 15 min to complete the fitting of the model on the training dataset). Further, the experiments have shown that our model has a good generalization ability and has a robust performance on different datasets, compared with the other models (U-Net, U-Net++, ResUNet++, SFA, PraNet, DCRNet, EU-Net, SANet, ACSNet, HarDNet). Compared with traditional CNNs, our model (MugenNet) achieves faster training speeds. With the use of pre-trained weights from the DeiT distilled neural network, MugenNet converged in just 30 epochs. With a batch size of 16, it only takes 15 min, and the processing speed is about 56 FPS. This performance in processing speed means that our model can be applied to video streaming data processing and real-time clinical polyp segmentation, laying the foundation for the next generation colonoscopy examination techniques.
However, there is still room for improvement. MugenNet is highly dependent on training data, which may impact its performance on previously unseen, complex samples. Additionally, MugenNet’s computational demands are relatively high, which may limit its real-time performance on low-spec hardware, such as in medical endoscopes. Future work will focus on optimizing MugenNet’s architecture and parameters to reduce computational costs and further enhance its generalization capabilities. We also plan to apply MugenNet to other medical imaging segmentation tasks (such as lung infection diagnosis) to validate its applicability and robustness across various fields.

Author Contributions

Conceptualization, C.P. and W.Z.; methodology, C.P.; software, C.P.; validation, K.W., Z.Q. and W.Z.; formal analysis, K.W.; investigation, Q.L.; resources, Z.Q., Z.B. and W.Z.; writing—original draft preparation, C.P. and K.W.; writing—review and editing, W.Z.; supervision, Z.Q., L.Z. and W.Z.; project administration, Z.Q.; funding acquisition, Z.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Natural Science Foundation of Shanghai (Grant No. 23ZR1416200) to Z.Q. and NSERC (Natural Sciences and Engineering Research Council of Canada) CREATE (Collaborative Research and Training Experience) program grant (grant number: 565429-2022) to W.Z.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cheng, W.; Di, Y.; Zhang, E.; Moser, M.; Kanagaratnam, S.; Korman, L.; Sarvazyan, N.; Zhang, W. Modeling and in vitro experimental validation for kinetics of the colonoscope in colonoscopy. Ann. Biomed. Eng. 2018, 41, 1084–1093. [Google Scholar] [CrossRef] [PubMed]
  2. Qian, Z.; Lv, Y.; Lv, D.; Gu, H.; Wang, K.; Zhang, W.; Gupta, M. A new approach to polyp detection by pre-processing of images and enhanced faster R-CNN. IEEE Sens. J. 2020, 21, 11374–11381. [Google Scholar] [CrossRef]
  3. Ahamed, M.F.; Syfullah, M.K.; Sarkar, O.; Islam, M.T.; Nahiduzzaman, M.; Islam, M.R.; Khandakar, A.; Ayari, M.A.; Chowdhury, M.E. Irv2-net: A deep learning framework for enhanced polyp segmentation performance integrating inceptionresnetv2 and unet architecture with test time augmentation techniques. Sensors 2023, 23, 7724. [Google Scholar] [CrossRef] [PubMed]
  4. Gschwantler, M.; Kriwanek, S.; Langner, E.; Göritzer, B.; Schrutka-Kölbl, C.; Brownstone, E.; Feichtinger, H.; Weiss, W. High-grade dysplasia and invasive carcinoma in colorectal adenomas: A multivariate analysis of the impact of adenoma and patient characteristics. Eur. J. Gastroenterol. Hepatol. 2002, 14, 183–188. [Google Scholar] [CrossRef]
  5. Rother, C.; Kolmogorov, V.; Blake, A. “GrabCut” interactive foreground extraction using iterated graph cuts. ACM Trans. Graph. (TOG) 2004, 23, 309–314. [Google Scholar] [CrossRef]
  6. Boykov, Y.; Jolly, M.; Blake, A. Interactive graph cuts for optimal boundary & region segmentation of objects in ND images. In Proceedings of the Eighth IEEE International Conference on Computer Vision, ICCV 2001, Vancouver, BC, Canada, 7–14 July 2001; pp. 105–112. [Google Scholar]
  7. Nie, T.; Zhao, Y.; Yao, S. ELA-Net: An Efficient Lightweight Attention Network for Skin Lesion Segmentation. Sensors 2024, 24, 4302. [Google Scholar] [CrossRef]
  8. Zhang, W.; Ouyang, P.; Sun, Z. A novel hybridization design principle for intelligent mechatronics systems. In The Abstracts of the International Conference on Advanced Mechatronics: Toward Evolutionary Fusion of IT and Mechatronics: ICAM 2010.5; The Japan Society of Mechanical Engineers: Tokyo, Japan, 2010; pp. 67–74. [Google Scholar]
  9. Patel, K.; Bur, A.; Wang, G. Enhanced u-net: A feature enhancement network for polyp segmentation. In Proceedings of the 2021 18th Conference on Robots and Vision (CRV), Burnaby, BC, Canada, 26–28 May 2021; pp. 181–188. [Google Scholar]
  10. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  11. Baldeon-Calisto, M.; Lai-Yuen, S. Resu-net: Residual convolutional neural network for prostate mri segmentation. IIE Annual Conference. Proceedings; Institute of Industrial and Systems Engineers (IISE): Peachtree Corners, GA, USA, 2018; pp. 731–736. [Google Scholar]
  12. Ruan, J.; Xie, M.; Gao, J.; Liu, T.; Fu, Y. Ege-unet: An efficient group enhanced unet for skin lesion segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Vancouver, BC, Canada, 8–12 October 2023; pp. 481–490. [Google Scholar]
  13. Peng, Y.; Sonka, M.; Chen, D. U-Net v2: Rethinking the skip connections of U-Net for medical image segmentation. arXiv 2023, arXiv:2311.17791. [Google Scholar]
  14. Zhang, R.; Li, G.; Li, Z.; Cui, S.; Qian, D.; Yu, Y. Adaptive context selection for polyp segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2020: 23rd International Conference, Lima, Peru, 4–8 October 2020; Proceedings, Part VI 23; Springer: Berlin/Heidelberg, Germany, 2020; pp. 253–262. [Google Scholar]
  15. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  16. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv, 2020; arXiv:2010.11929. [Google Scholar]
  17. Zhang, W.; Yang, G.; Lin, Y.; Ji, C.; Gupta, M. On Definition of Deep Learning. In Proceedings of the 2018 World Automation Congress (WAC), Stevenson, WA, USA, 3–6 June 2018; pp. 1–5. [Google Scholar]
  18. Zhang, Y.; Liu, H.; Hu, Q. Transfuse: Fusing transformers and cnns for medical image segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Proceedings, Part I 24; Springer: Berlin/Heidelberg, Germany, 2021; pp. 14–24. [Google Scholar]
  19. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; Li, F. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  20. Huang, B.; Huang, T.; Xu, J.; Min, J.; Hu, C.; Zhang, Z. RCNU-Net: Reparameterized convolutional network with convolutional block attention module for improved polyp image segmentation. Biomed. Signal Process. Control 2023, 93, 106138. [Google Scholar] [CrossRef]
  21. Ta, N.; Chen, H.; Lyu, Y.; Wu, T. Ble-net: Boundary learning and enhancement network for polyp segmentation. Multimed. Syst. 2023, 29, 3041–3054. [Google Scholar] [CrossRef]
  22. Mandujano-Cornejo, V.; Montoya-Zegarra, J.A. Polyp2Seg: Improved polyp segmentation with vision transformer. In Proceedings of the Annual Conference on Medical Image Understanding and Analysis, Cambridge, UK, 27–29 July 2022; pp. 519–534. [Google Scholar]
  23. Qin, Y.; Xia, H.; Song, S. RT-Net: Region-Enhanced Attention Transformer Network for Polyp Segmentation. Neural Process. Lett. 2023, 55, 11975–11991. [Google Scholar] [CrossRef]
  24. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar]
  25. Cai, M.; Lin, Y.; Han, B.; Liu, C.; Zhang, W. On a simple and efficient approach to probability distribution function aggregation. IEEE Trans. Syst. Man Cybern. Syst. 2016, 47, 2444–2453. [Google Scholar] [CrossRef]
  26. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Yang, S.; Lv, S.; Zhang, W.; Cui, Y. Microfluidic point-of-care (POC) devices in early diagnosis: A review of opportunities and challenges. Sensors 2022, 22, 1620. [Google Scholar] [CrossRef] [PubMed]
  29. Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; Jagers, M. Basnet: Boundary-aware salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7479–7489. [Google Scholar]
  30. Woo, S.; Park, J.; Lee, J.; Kweon, I. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  31. Chen, P.; Chang, M.; Hsieh, J.; Chen, Y. Parallel residual bi-fusion feature pyramid network for accurate single-shot object detection. IEEE Trans. Image Process. 2021, 30, 9099–9111. [Google Scholar] [CrossRef]
  32. Schlemper, J.; Oktay, O.; Schaap, M.; Heinrich, M.; Kainz, B.; Glocker, B.; Rueckert, D. Attention gated networks: Learning to leverage salient regions in medical images. Med. Image Anal. 2019, 53, 197–207. [Google Scholar] [CrossRef]
  33. Wei, J.; Wang, S.; Huang, Q. F3Net: Fusion, feedback and focus for salient object detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12321–12328. [Google Scholar]
  34. Tajbakhsh, N.; Gurudu, S.; Liang, J. Automated polyp detection in colonoscopy videos using shape and context information. IEEE Trans. Med. Imaging 2015, 35, 630–644. [Google Scholar] [CrossRef]
  35. Vázquez, D.; Bernal, J.; Sánchez, F.; Fernández-Esparrach, G.; López, A.M.; Romero, A.; Drozdzal, M.; Courville, A. A benchmark for endoluminal scene segmentation of colonoscopy images. J. Healthc. Eng. 2017, 2017, 4037190. [Google Scholar] [CrossRef]
  36. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 2481–2495. [Google Scholar] [CrossRef]
  37. Silva, J.; Histace, A.; Romain, O.; Dray, X.; Granado, B. Toward embedded detection of polyps in wce images for early diagnosis of colorectal cancer. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 283–293. [Google Scholar] [CrossRef]
  38. Jha, D.; Smedsrud, P.; Riegler, M.; Halvorsen, P.; de Lange, T.; Johansen, D.; Johansen, H. Kvasir-seg: A segmented polyp dataset. In Proceedings of the MultiMedia Modeling: 26th International Conference, MMM 2020, Daejeon, Republic of Korea, 5–8 January 2020; Proceedings, Part II 26. Springer: Berlin/Heidelberg, Germany, 2020; pp. 451–462. [Google Scholar]
  39. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
  40. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; Proceedings 4; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  41. Jha, D.; Smedsrud, P.; Riegler, M.; Johansen, D.; De Lange, T.; Halvorsen, P.; Johansen, H. Resunet++: An advanced architecture for medical image segmentation. In Proceedings of the 2019 IEEE International Symposium on Multimedia (ISM), San Diego, CA, USA, 9–11 December 2019; pp. 225–2255. [Google Scholar]
  42. Fang, Y.; Chen, C.; Yuan, Y.; Tong, K. Selective feature aggregation network with area-boundary constraints for polyp segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2019: 22nd International Conference, Shenzhen, China, 13–17 October 2019; Proceedings, Part I 22; Springer: Berlin/Heidelberg, Germany, 2019; pp. 302–310. [Google Scholar]
  43. Fan, D.; Ji, G.; Zhou, T.; Chen, G.; Fu, H.; Shen, J.; Shao, L. Pranet: Parallel reverse attention network for polyp segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; pp. 263–273. [Google Scholar]
  44. Yin, Z.; Liang, K.; Ma, Z.; Guo, J. Duplex contextual relation network for polyp segmentation. In Proceedings of the 2022 IEEE 19th International Symposium on Biomedical Imaging (ISBI), Kolkata, India, 28–31 March 2022; pp. 1–5. [Google Scholar]
  45. Wei, J.; Hu, Y.; Zhang, R.; Li, Z.; Zhou, S.; Cui, S. Shallow attention network for polyp segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021, Proceedings, Part I 24; Springer: Berlin/Heidelberg, Germany, 2021; pp. 699–708. [Google Scholar]
  46. Huang, C.; Wu, H.; Lin, Y. Hardnet-mseg: A simple encoder-decoder polyp segmentation neural network that achieves over 0.9 mean dice and 86 fps. arXiv 2021, arXiv:2101.07172. [Google Scholar]
  47. Margolin, R.; Zelnik-Manor, L.; Tal, A. How to evaluate foreground maps? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 248–255. [Google Scholar]
  48. Cheng, M.; Fan, D.P. Structure-measure: A new way to evaluate foreground maps. Int. J. Comput. Vis. 2021, 129, 2622–2638. [Google Scholar] [CrossRef]
  49. Fan, D.; Gong, C.; Cao, Y.; Ren, B.; Cheng, M.; Borji, A. Enhanced-alignment measure for binary foreground map evaluation. arXiv 2018, arXiv:1805.10421. [Google Scholar]
Figure 1. The architecture of MugenNet: combine the Transformer branch (left) with the CNN branch (right) in the Mugen module (middle). FC: Fully Connected. AG: attention gate. SE: Squeeze and Excitation.
Figure 1. The architecture of MugenNet: combine the Transformer branch (left) with the CNN branch (right) in the Mugen module (middle). FC: Fully Connected. AG: attention gate. SE: Squeeze and Excitation.
Sensors 24 07473 g001
Figure 2. The architecture of the Vision Transformer [16].
Figure 2. The architecture of the Vision Transformer [16].
Sensors 24 07473 g002
Figure 3. The architecture of the attention gate.
Figure 3. The architecture of the attention gate.
Sensors 24 07473 g003
Figure 4. A comparison of the MugenNet with the other four nets (U-Net, U-net++, SFA, Pranet) on the Kvasir dataset.
Figure 4. A comparison of the MugenNet with the other four nets (U-Net, U-net++, SFA, Pranet) on the Kvasir dataset.
Sensors 24 07473 g004
Figure 5. A comparison of the performance of MugenNet on the five datasets (ClinicDB, ColonDB, CVC 300, ETIS, Kvasir).
Figure 5. A comparison of the performance of MugenNet on the five datasets (ClinicDB, ColonDB, CVC 300, ETIS, Kvasir).
Sensors 24 07473 g005
Figure 6. Training process on four datasets.
Figure 6. Training process on four datasets.
Sensors 24 07473 g006
Figure 7. Statistical test on eight models.
Figure 7. Statistical test on eight models.
Sensors 24 07473 g007
Table 1. Comparison results of CVC-300.
Table 1. Comparison results of CVC-300.
MethodsCVC-300 Dataset
mDicemIoUMAE F β ω S α E ξ
U-Net0.7100.6270.0220.5670.7930.826
U-Net++0.7070.6240.0180.5810.7960.831
ResUNet++0.7630.7010.0210.6130.8010.867
SFA0.4670.3290.0650.3410.6400.644
PraNet0.8710.7970.0100.8430.9250.950
DCRNet0.8560.7880.0160.8300.9210.943
EU-Net0.8370.7650.0150.8050.9040.919
SANet0.8880.8150.0180.8590.9280.962
ACSNet0.7320.6270.0160.7030.8370.871
HarDNet0.8740.8040.0190.8520.9240.948
MugenNet (Ours)0.8630.7950.0100.8640.9210.972
Table 2. Comparison Results of CVC-ColonDB.
Table 2. Comparison Results of CVC-ColonDB.
MethodsCVC-ColonDB Dataset
mDicemIoUMAE F β ω S α E ξ
U-Net0.5120.4440.0610.4980.7120.696
U-Net++0.4830.4100.0640.4670.6910.680
ResUNet++0.5720.5010.0580.5760.7060.701
SFA0.4690.3470.0960.3790.6340.675
PraNet0.7120.6400.0430.6990.8200.847
DCRNet0.7040.6310.0520.6840.8210.840
EU-Net0.7560.6810.0450.7300.8310.863
SANet0.7530.6700.0430.7260.8370.869
ACSNet0.7160.6490.0390.6970.8290.839
HarDNet0.7350.6660.0380.7240.8340.834
MugenNet (Ours)0.7580.6780.0340.8090.9180.875
Table 3. Comparison results of ETIS.
Table 3. Comparison results of ETIS.
MethodsETIS Dataset
mDicemIoUMAE F β ω S α E ξ
U-Net0.3980.3350.0360.3660.6840.643
U-Net++0.4010.3440.0350.6830.6830.629
ResUNet++0.4230.4750.0290.6280.7160.693
SFA0.2970.2170.1090.2310.5570.531
PraNet0.6280.5670.0310.6000.7940.808
DCRNet0.5560.4960.0960.5060.7360.742
EU-Net0.6870.6090.0670.6360.7930.807
SANet0.7050.6340.0250.6850.8490.881
ACSNet0.5780.5780.0590.5300.7540.737
HarDNet0.7000.6300.0250.6710.8280.854
MugenNet (Ours)0.7140.6360.0190.6060.6770.831
Table 4. Comparison results of Kvasir.
Table 4. Comparison results of Kvasir.
MethodsKvasir Dataset
mDicemIoUMAE F β ω S α E ξ
U-Net0.8180.7460.0550.7940.8580.881
U-Net++0.8210.7430.0480.8080.8620.886
ResUNet++0.8130.7930.0530.8310.8730.878
SFA0.7230.6110.0750.6700.7820.834
PraNet0.8980.8400.0300.8850.9150.944
DCRNet0.8860.8250.0350.8680.9110.933
EU-Net0.9080.8540.0380.8930.9170.951
SANet0.9040.8470.0360.8920.9150.949
ACSNet0.8980.8380.0320.8820.9200.941
HarDNet0.8970.8390.0380.8850.9120.942
MugenNet (Ours)0.8880.8280.0300.9280.9390.947
Table 5. Comparison results of CVC-ClinicDB.
Table 5. Comparison results of CVC-ClinicDB.
MethodsCVC-ClinicDB Dataset
mDicemIoUMAE F β ω S α E ξ
U-Net0.8230.7550.0190.8110.8890.913
U-Net++0.7940.7290.0220.7850.8730.891
ResUNet++0.7960.7320.0360.8030.8600.878
SFA0.7000.6070.0420.6470.7930.840
PraNet0.8990.8490.0090.8960.9360.963
DCRNet0.8960.8440.0100.8900.9330.964
EU-Net0.8920.8460.0110.8910.9360.959
SANet0.9160.8590.0120.9090.9390.971
ACSNet0.8820.8260.0110.8730.9270.947
HarDNet0.9090.8640.0180.9070.9380.961
MugenNet (Ours)0.8840.8230.0170.9470.9460.969
Table 6. Comparison of the running speed of different models.
Table 6. Comparison of the running speed of different models.
MethodsEpochLRTimeFPSmDice
U-Net30 3 × 10 4 ∼40 min∼8 FPS0.823
U-Net++30 3 × 10 4 ∼45 min∼7 FPS0.794
SFA500 1 × 10 2 >60 min∼40 FPS0.700
PraNet20 1 × 10 4 ∼30 min∼50 FPS0.899
MugenNet30 1 × 10 4 ∼15 min∼56 FPS0.875
Table 7. Ablation studies of mugennet.
Table 7. Ablation studies of mugennet.
SettingsCVC-ColonDBETIS-LaribPolypDB
mDicemIoU F β ω mDicemIoU F β ω
Backbone0.5190.4730.6370.4240.3960.381
Backbone + TB0.7260.6160.7420.6490.6050.578
Backbone + CB0.6670.5910.6800.6210.5630.497
Complete Model0.7580.6780.8090.7140.6360.606
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, C.; Qian, Z.; Wang, K.; Zhang, L.; Luo, Q.; Bi, Z.; Zhang, W. MugenNet: A Novel Combined Convolution Neural Network and Transformer Network with Application in Colonic Polyp Image Segmentation. Sensors 2024, 24, 7473. https://doi.org/10.3390/s24237473

AMA Style

Peng C, Qian Z, Wang K, Zhang L, Luo Q, Bi Z, Zhang W. MugenNet: A Novel Combined Convolution Neural Network and Transformer Network with Application in Colonic Polyp Image Segmentation. Sensors. 2024; 24(23):7473. https://doi.org/10.3390/s24237473

Chicago/Turabian Style

Peng, Chen, Zhiqin Qian, Kunyu Wang, Lanzhu Zhang, Qi Luo, Zhuming Bi, and Wenjun Zhang. 2024. "MugenNet: A Novel Combined Convolution Neural Network and Transformer Network with Application in Colonic Polyp Image Segmentation" Sensors 24, no. 23: 7473. https://doi.org/10.3390/s24237473

APA Style

Peng, C., Qian, Z., Wang, K., Zhang, L., Luo, Q., Bi, Z., & Zhang, W. (2024). MugenNet: A Novel Combined Convolution Neural Network and Transformer Network with Application in Colonic Polyp Image Segmentation. Sensors, 24(23), 7473. https://doi.org/10.3390/s24237473

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop