Next Article in Journal
Creating an AI-Enhanced Morse Code Translation System Based on Images for People with Severe Disabilities
Previous Article in Journal
RCKD: Response-Based Cross-Task Knowledge Distillation for Pathological Image Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Pixel Diffuser: Practical Interactive Medical Image Segmentation without Ground Truth

1
Major in Bio Artificial Intelligence, Department of Applied Artificial Intelligence, Hanyang University at Ansan, Ansan 15588, Republic of Korea
2
Major in Bio Artificial Intelligence, Department of Computer Science & Engineering, Hanyang University at Ansan, Ansan 15588, Republic of Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Bioengineering 2023, 10(11), 1280; https://doi.org/10.3390/bioengineering10111280
Submission received: 13 October 2023 / Revised: 30 October 2023 / Accepted: 31 October 2023 / Published: 2 November 2023

Abstract

:
Medical image segmentation is essential for doctors to diagnose diseases and manage patient status. While deep learning has demonstrated potential in addressing segmentation challenges within the medical domain, obtaining a substantial amount of data with accurate ground truth for training high-performance segmentation models is both time-consuming and demands careful attention. While interactive segmentation methods can reduce the costs of acquiring segmentation labels for training supervised models, they often still necessitate considerable amounts of ground truth data. Moreover, achieving precise segmentation during the refinement phase results in increased interactions. In this work, we propose an interactive medical segmentation method called PixelDiffuser that requires no medical segmentation ground truth data and only a few clicks to obtain high-quality segmentation using a VGG19-based autoencoder. As the name suggests, PixelDiffuser starts with a small area upon the initial click and gradually detects the target segmentation region. Specifically, we segment the image by creating a distortion in the image and repeating it during the process of encoding and decoding the image through an autoencoder. Consequently, PixelDiffuser enables the user to click a part of the organ they wish to segment, allowing the segmented region to expand to nearby areas with pixel values similar to the chosen organ. To evaluate the performance of PixelDiffuser, we employed the dice score, based on the number of clicks, to compare the ground truth image with the inferred segment. For validation of our method’s performance, we leveraged the BTCV dataset, containing CT images of various organs, and the CHAOS dataset, which encompasses both CT and MRI images of the liver, kidneys and spleen. Our proposed model is an efficient and effective tool for medical image segmentation, achieving competitive performance compared to previous work in less than five clicks and with very low memory consumption without additional training.

1. Introduction

Medical image segmentation has been a crucial task for identifying structures of specific organs. In recent years, deep learning has emerged as a promising approach for solving medical segmentation problems [1,2,3,4,5]. Despite the excellent performance of deep-learning-based models, training them requires a significant amount of data with ground truth labels, which can be time-consuming and resource-intensive.
Interactive medical segmentation can alleviate these challenges by helping us easily obtain segmentation labels. Unlike traditional image segmentation, interactive medical segmentation takes various user interactions (e.g., bounding boxes [6], scribbles [7] or clicks [8]) to compute the region of segmentation, leading to a substantial reduction in labeling cost. However, existing works [6,7,8] still require the use of segmentation labels during training.
Recently, general interactive segmentation models such as RITM [9], SimpleClick [10], and SAM [11] have shown to be effective even for medical image segmentation. These models generate and refine segmentation predictions using user clicks or points provided by the user. Despite being trained with general image segmentation data, they can also be applied to medical image segmentation without requiring additional training on medical images. However, alongside their impressive results in general and medical image segmentation, we observed that they tend to capture much wider regions than the ground truth masks, which leads to an increased number of clicks. We analyzed that this is due to the fact that they are attempting to predict the segmentation mask in a single step.
This study proposes an interactive medical segmentation model named PixelDiffuser, which consists of a light weight pre-trained autoencoder trained on general image data. Using the proposed method, a user can obtain a segmentation mask by simply clicking the location of interest. When the click is given, our model first transforms it into a binary circle image centered at the click location with a 1-pixel radius. Subsequently, the circle image is gradually transformed into the segmentation mask, as the name suggests.
Our segmentation procedure is based on our key observation, reconstruction noises, which are distortions that arise in decoded images generated through the encoding and decoding process of an autoencoder. The most important aspect is that if we mix two images in the encoding space (feature space) and decode the mixed latent, the resulting image appears as an intermediate between two source images. Motivated by this observation, our approach iteratively mixes the click image and the medical image to make the click image cover the organ of interest.
Since our method utilizes a simple autoencoder, it does not require any medical or general segmentation data, which is converse to the previous works, such as SimpleClick, SAM, and MedSAM which rely on either medical or general image segmentation data.
We employed The Beyond The Crutial Vault (BTCV) dataset [12], a multi-organ abdominal CT reference, and CHAOS dataset [13], which aims at the segmentation of liver, kidneys and spleen from CT and MRI data, to validate the model. For evaluation, we utilized the Dice score, comparing our model’s performance to other interactive segmentation approaches. It is important to note that MedSAM [14], which specializes in medical image segmentation based on the SAM method, is trained on large-scale medical image segmentation data. Therefore, it differs from our approach, which is an interactive medical segmentation model that does not require training using medical image segmentation data. For this reason, we exclude MedSAM [14] from the evaluation. The experimental results indicate that PixelDiffuser generates more accurate segmentation masks with only a few clicks. In addition, our method achieves comparable results to the SAM Huge model while being 91.4 × lighter than SAM-Huge.
Our contributions can be listed as follows:
  • In this work, we present an innovative interactive medical segmentation model called PixelDiffuser. Unlike traditional approaches, our model does not necessitate specialized training on medical segmentation datasets. Instead, we leverage a simple autoencoder trained on general image data.
  • The PixelDiffuser operates interactively by initiating the segmentation process from a single click point. This strategic approach ensures that the model gradually identifies the target mask, thereby mitigating the risk of overstepping the actual ground truth region. Consequently, the need for excessive user inputs is significantly reduced.
  • Through comprehensive experiments, we showcase the remarkable efficiency of our proposed algorithm. With only a minimal number of clicks, our approach outperforms alternative methods in achieving accurate target segmentation.

2. Related Works

2.1. Medical Image Segmentation

Medical image segmentation is an important task in assisting diagnosis and surgeries. As convolutional neural networks [15] have shown high performance, many other works utilized them to solve medical image segmentation. U-Net [5] is inspired by [15] and achieved state-of-the-art in medical image segmentation. After that, many variants of U-Net [1,2,3,4] have been developed. TransUNet [1] and UNETR [3] employed transformer [16] as a feature extractor, before predicting segmentation maps. Swin-UNet [2] and Swin-UNETR utilized Swin-Transformer architecture [17] to acquire contextual features from the medical images.
Additionally, to improve the performance of medical segmentation, ensemble-based methods [18,19,20,21] have been also invented. DivergentNets [18] proposed TriUnet which consists of three separate U-Net, and combined it with an ensemble of previous state-of-the-art medical segmentation models. This ensemble, referred to as DivergentNets, has an advantage in improving generalization in medical image segmentation. ALMN (Affinity Learning-based Multi-branch Ensemble Network) [19] employed an affinity-learning-based ensemble learning framework to achieve high performance in 3D medical segmentation. Dang, Truong et al. [20] introduced a combining strategy of several segmentation models. Georgescu, Mariana-Iuliana et al. [21] leveraged the diversity of several segmentation models to generate an efficient ensemble combining strategy. However, the methods discussed earlier require ground truth labels for training the segmentation model, which comes with the drawback of demanding substantial effort and time to collect.

2.2. Medical Interactive Image Segmentation

Traditional interactive segmentation methods can be categorized depending on the type of user interaction, as follows: contour-based methods [22], bounding-box-based methods [23], and scribble-based methods [24,25,26,27]. Recently, deep-learning-based methods have improved performance with fewer user interactions. BIFSeg [6] takes a bounding box from a user to enable zero-shot segmentation. DeepIGeoS [7] requires a scribble-type user interaction to find segmentation predictions of fetal MRI or 3D brain tumor. MIDeepSeg [8] performed 2D and 3D medical image segmentation through user-provided interior margin points. Moreover, there has been research that utilizes models trained with non-medical image data for interactive medical image segmentation. For instance, models like SimpleClick [10] and SAM [11] demonstrate the capability of segmentation even for medical images despite being trained with general image segmentation data.

3. Motivation

In contrast to the previous methods that predict a full segmentation mask in a single step, our approach predicts the final segmentation mask by iteratively expanding from the initial point. This method is inspired by our key observation named reconstruction noises, distortions in reconstructed images stemming from the bottleneck of the autoencoder architecture. In this section, we introduce the reconstruction noises and describe how we utilize them to compute the segmentation mask.

3.1. The Reconstruction Noises

An autoencoder architecture comprises two primary components: an encoder (E) and a decoder (D). For a given grayscale image a, the encoder calculates a latent feature z a = E ( a ) , and the decoder reconstructs the image as a ˜ = D ( E ( a ) ) . Due to the bottleneck in autoencoders, there is an information loss which inevitably leads to distortions in the reconstructed image.
This distortion effect is even more evident when mixing two images in the feature space. Consider another grayscale image b with its latent feature z b = E ( b ) . By averaging z a and z b , we obtain an averaged latent feature z m . When decoded, this produces a blended image m, which is a mix of a and b. However, due to the information loss stemming from a bottleneck, m does not perfectly align with a direct pixel-level mix of a and b. As illustrated in Figure 1, mixing at deeper layers introduces more distortions. We term these distortions as reconstruction noises.

3.2. From the Reconstruction Noises to Segmentation Mask

Based on our observation, we introduce our intuition on how the reconstruction noises can be utilized to generate segmentation masks. As depicted in Figure 1, it becomes evident that when the features of two images are mixed at deeper layers and then reconstructed, the resulting shape appears to be an intermediate between the two source images, a and b. This blending effect of the images is more pronounced in deeper layers. Drawing on this intuition, we posed the following question: If we are provided with a CT image and a click image (the dot image), could we potentially obtain segmentation masks through their iterative mixing? This is our motivation for the proposed method, PixelDiffuser, which generates segmentation masks by iteratively mixing a CT image with a click image (the dot image). We describe the method in more detail in the following section.

4. The PixelDiffuser

We propose the PixelDiffuser, an interactive medical segmentation method. Given a CT image I and a user’s click, our model produces an initial segmentation by the initial segmentation process. This iterative procedure transforms the user’s click into a segmentation. Within this process, the click is represented as a circle image, g 0 , centered at the click location with a 1-pixel radius. This circle image is then gradually expanded into the initial segmentation mask, g T , by iteratively blending with the CT image, I . If the initial segmentation is not satisfied, additional clicks can be provided to refine the segmentation using refinement process, which adds the additional region to g T or removes the irrelevant region from the g T . Each step of this process is described in the following subsections.
Motivated by reconstruction noises, our method computes the segmentation mask by iteratively mixing the CT image and the dot image in the feature space to distort the dot image. Our method utilizes the VGG19-based encoder E and the decoder D [28] to encode and mix the images. The encoder is the convolutional part of VGG19, while the decoder is a neural network symmetrically shaped to the encoder. The autoencoder is pre-trained using the general images (e.g., COCO dataset [29]), and we do not tune it with any medical images.

4.1. Initial Segmentation Process

Given the CT image, I , and the dot image, g 0 , created based on the user’s click, the initial segmentation process calculates the initial segmentation mask, g T . This is achieved by iteratively blending the CT image and the dot image in the feature space. Figure 2 illustrates an overview of this initial segmentation process. Since the procedure is iterative, we detail the steps involved in computing the intermediate segmentation mask, g t + 1 , given the CT image, I , and the preceding mask, g t .
With the two images, I and g t , the singular step for computing the intermediate mask, g t + 1 , consists of three sub-steps: (1) density-based intensity adjustment (Section 4.1.1), (2) feature mixing (Section 4.1.2) and (3) segmentation (Section 4.1.3). Density-based intensity adjustment step adjusts the pixel values of I to suppress pixels from irrelevant regions. As a result, reconstruction noises spread to the only target region. In the feature mixing step, both images are blended in the feature space, producing a combined image that carries reconstruction noises. In the segmentation step, the merged image is binarized using the Otsu threshold method [30], yielding the result, g t + 1 .

4.1.1. Density-Based Intensity Adjustment

Consider the appearance of a CT image. Within such an image, numerous white regions, each representing different organs, coexist alongside the target region, which represents the specific organ of interest. This implies that when distorting the dot image, it can be mixed with the other organs. To avoid this, we first need to transform the intensities of the CT image, I , to minimize the influence of pixels from irrelevant regions. The density-based intensity adjustment transforms the CT image, I , to the intensity-adjusted image, x . Let F I : ( I , g t ) x represent the density-based intensity adjustment function. This function accepts two images, I and g t , and outputs the transformed image, x . The definition of F I is as follows:
F I ( I , g t ) = kde ( crop ( I , g t ) ) ( I ) ,
where crop denotes the cropping operator extracting regions from I masked by g t as shown in Figure 3. The size of this cropped region is determined after the dilation of g t for 7 iterations using a 3 × 3 kernel, resulting in an effective padding of approximately 14 pixels on each boundary: top, bottom, left, and right. Additionally, kde ( i ) is a kernel density estimator function [31] which computes the density of pixel values of an image, i. Specifically, kde ( i ) ( · ) is a function that receives pixel values as input and returns the corresponding density value for those pixels. As illustrated in Figure 3, upon the execution of kde, all pixel intensities are substituted with normalized pixel histogram values. Subsequently, as depicted in Figure 4, given that pixel intensities in the region corresponding to the organ of interest show the highest density, we conduct a density-based intensity adjustment, retaining only those pixels similar to the organ with elevated values.

4.1.2. Feature Mixing

Given the intensity-adjusted image, x , and the intermediate mask, g t , the feature mixing step encodes them into the latent features, and subsequently, averages those features to produce the mixed features f a ( x , g t ) .
f x = E ( x ) ,
f g t = E ( g t ) ,
f a ( x , g t ) = f x + f g t 2 ,
where f x and f g t represent the latent features of x and g t , respectively. Using this mixed latent feature, we decode it to obtain the mixed and distorted image, a ( x , g t ) = D ( f a ( x , g t ) ) , as illustrated in Figure 2.

4.1.3. Segmentation

While the density-based intensity adjustment step eliminates many irrelevant regions, the mixed image, a ( x , g t ) , still contains regions outside the target organ. To remove the irrelevant regions, we compute another mixed image, a ( x , o ) , where o is a completely black image.
f x = E ( x ) ,
f o = E ( o ) ,
f a ( x , o ) = f x + f o 2 ,
a ( x , o ) = D ( f a ( x , o ) ) .
We then compute the difference between a ( x , o ) and a ( x , g t ) to eliminate irrelevant regions. Following this, the Otsu thresholding method [30] is employed to extract the binary segmentation mask g t + 1 . The Otsu method determines an optimal threshold by differentiating two mixed distributions, making it suitable for dividing light and dark regions in the image. During the initial iterations, the computed threshold might exhibit instability, leading to variations in the segmented region. This iteration continues until g t and g t + 1 are identical. Once the iterative change has stabilized or converged, the process is terminated, and the resulting segment is produced.

4.2. Refinement Process

Once the segmentation g t is obtained, the user has the option to provide additional clicks, either positive or negative, for revisions. If g t covers a smaller area than expected, users can offer positive clicks to expand the region. Conversely, if the segmentation mask contains irrelevant areas, negative clicks can be used to eliminate them. Upon receiving a positive click, the PixelDiffuser produces another segmentation mask using the same method as described in Section 4.1 and then combines it with g t .
While handling positive clicks is simple, managing negative clicks is a little bit more complex. Upon receiving a negative click, a methodology similar to the initial segmentation process (Section 4.1) is employed to identify the negative region within the positive region. As depicted in the rightmost image of Figure 5, both the positive and negative densities are computed based on their respective click points. The standard deviation of a designated region, whether positive or negative, is determined by the standard deviation of the distances between the points in that region and its corresponding click point (either positive or negative). Given these densities, any region where the negative density surpasses the positive density is removed from the segmentation mask, g t .

5. Experiments

5.1. Dataset

5.1.1. BTCV

We utilize The Beyond The Crutial Vault (BTCV) [12] dataset, which contains the CT volumes for 13 organs—(1) spleen, (2) right kidney, (3) left kidney, (4) gallbladder, (5) esophagus, (6) liver, (7) stomach, (8) aorta, (9) inferior vena cava, (10) portal vein and splenic vein, (11) pancreas, (12) right adrenal gland and (13) left adrenal gland. For each organ, there are 864, 897, 905, 359, 689, 1542, 1015, 2042, 1884, 545, 746, 319 and 369 images, respectively.
For preprocessing, all CT images are resized to 512 by 512 pixels, except in the case of the SimpleClick experiments where a size of 448 by 448 was used. Additionally, pixel values in all images were clipped to a range of [−135, 215] and normalized to [0, 1].

5.1.2. CHAOS

We also evaluate our model’s performance using CHAOS [13] dataset. This is an abdominal medical segmentation dataset, including CT and MRI images of 40 different patients. While the CT dataset only includes segmentation masks for the liver, the MRI dataset contains segmentation masks for four organs—the liver, left and right kidney, and spleen.
For preprocessing of CT images, we resize all images to 512 by 512. Additionally, we clip the pixel values to a range of [−135, 215], which is the same as BTCV. For the MRI dataset, all images are clipped to a range of [100, 1000]. After clipping, we normalized the pixel values to [0, 1].

5.2. User Interaction

To conduct a quantitative evaluation, we simulated user clicks using [9] which provides a simulation algorithm to sample both positive clicks and negative clicks. During the segmentation process, it starts with a single positive click and utilizes both negative and positive clicks for revision. It is important to note that while the sampling process requires the true segmentation masks, our proposed model does not rely on these masks during training or real-world applications.

5.3. Implementation Details

To implement our approach, we employ a VGG-based encoder E and a decoder D, where the encoder and decoder are pre-trained on the COCO dataset [29]. Instead of pre-training them from scratch, we adopt pretrained weights from Texture Reformer [32] implementation. The encoder consists of the convolutional components of VGG19 [28], but we only utilize 3 out of the 5 convolutional blocks from VGG19, where a convolutional block represents a sequence of convolutional layers followed by one max pooling. We empirically chose to use only 3 blocks for the encoder, as deeper encoders with more than 4 blocks tend to introduce more distortion. The decoder’s architecture mirrors the encoder’s but uses the nearest upsampling layer instead of max pooling. Just like the encoder, the decoder is implemented using only 3 convolutional blocks. The detailed architectures of them are indicated in Table 1.
Before feeding CT images into the segmentation process, all images are resized to a resolution of 512 × 512 . Additionally, we clip pixel values outside of the range of [ 135 , 215 ] to adjust intensities. Subsequently, we normalize the images to have pixel values in the range of [ 0 , 1 ] .
All experimental evaluations were conducted on a Linux workstation running Ubuntu 20.04, equipped with 16 GB of RAM, an Intel i9 CPU, and an NVIDIA GTX 1080 Ti GPU. The PyTorch 1.12 [33] framework was used for deep learning tasks. Since we employed pretrained weights for the encoder and decoder, we did not implement any code relevant to training, such as an optimizer or training iterations.

5.4. Description of Other Methods

We compare our approach with two existing methods, RITM [9], SimpleClick [10] and SAM [11]. Prior to proceeding to the evaluation, we briefly present the description of them. We exclude MIDeepSeg [8], DeepIGeoS [7], and MedSAM [14], which require careful clicks, such as clicks on boundary lines, or rely on medical segmentation data since we are focusing on developing interactive segmentation method when no medical segmentation data for training is available for various reasons.

5.4.1. RITM

This approach [9] is a click-based interactive segmentation model that iteratively learns from user clicks using a CNN-based backbone structure. It receives positive and negative clicks from the user, scales them to match the first layer of the model, and processes them element-wise before inputting them into the model. RITM then makes external masks, corrects the segmentation process, and repeats this procedure.

5.4.2. SimpleClick

SimpleClick [10] is a transformer-based general interactive segmentation model that takes an image and a click as an input. SimpleClick exploits the ViT as a backbone network to extract features from the concatenation of the image and the click (dot image). Utilizing the features, the simple feature pyramid [10] and subsequent segmentation head decode the feature into a segmentation mask.

5.4.3. SAM

SAM, denoted as Segment Anything Model [11], accepts various inputs such as points, boxes, and text in addition to images. Each piece of information is processed by the encoder, and the decoder subsequently produces the segmentation mask. This architecture facilitates segmentation tasks for a multitude of prompts without necessitating further training. Specifically, to achieve zero-shot performance, it undergoes training on over 1 billion mask datasets named SA-1B repeatedly.

5.4.4. FastSAM

As the name suggests, FastSAM [34] enhances the inference speed of SAM by substituting the Vision Transformer [35] with YOLOv8-seg [36] for the backbone network. This change greatly reduces the encoding time for images, resulting in a much faster inference speed compared to SAM. Similar to SAM, it also supports both unconditional and conditional segmentation, using either bounding boxes or click points.

5.5. Qualitative Evaluation

Figure 6 shows the segmentation examples produced by the PixelDiffuser for the 13 organs. In this figure, the top row represents the CT images, the second row represents the true segmentation masks (cropped), and the next 5 rows represent the segmentation (cropped) results given the number of clicks. As shown in this figure, the proposed method can discover almost the complete region of the target organ in just 3 to 5 click stages.

5.6. Quantitative Evaluation

As illustrated in Figure 7, we quantitatively evaluate the proposed model by comparing it with other existing models, SimpleClick [10], RITM [9], SAM [11], and FastSAM [34]. These models, like ours, can be trained without medical segmentation masks and leverage user clicks for interaction. The Dice score serves as our primary evaluation metric. For every n user interactions, we record the highest Dice score from the n segmentation masks.
Table 2 presents the average Dice score for each individual organ, while Figure 7 illustrates the average Dice score across all organs, correlated with the number of user clicks. These results indicate that our model outperforms the other methods when fewer clicks are employed. However, as the number of clicks increases, our model remains competitive with the other methodologies.
As demonstrated in Table 3, SAM-huge achieves competitive performance compared to our approach. However, SAM-huge involves a large number of parameters, leading to the necessity of substantial resources. As Table 3 presents, our model possesses significantly fewer parameters than SAM, making it computationally cost-effective and highly efficient.
Table 4 indicates a performance comparison of the approaches using the CHAOS CT dataset. As shown, our approach achieves superior performance to the other approaches, except for SAM-huge. Even though our model has about 90 × fewer parameters, we can find that it achieves competitive performance with SAM-huge.
For the MRI modalities, we organize the evaluation results in Table 5 and Table 6. They indicate the DICE performance using CHAOS MRI datasets. In these cases, our method outperforms most of the comparison models; however, it is not superior to SimpleClick-huge and SAM-huge. Furthermore, even when only a single click was provided, our approach is inferior to SAM-huge. We analyze that while they are pretrained with a segmentation task, our model does not rely on the segmentation task. Using information from the segmentation task, SimpleClick and SAM are likely better at capturing the structure of objects within images compared to our approach. Nevertheless, our model contains significantly fewer parameters than SAM-huge and, considering its reduced parameter count, still demonstrates commendable performance.

5.7. Limitations

We showcased that our approach achieves high performance with fewer interactions compared to the previous works. In this section, we discuss some limitations involved in our method.

5.7.1. Failure Cases Analysis

We observed that certain organs, such as the stomach, portal vein, splenic vein, and adrenal gland, exhibit low Dice scores, as depicted in the top left of Figure 8. Our analysis identified the primary causes for these failures: (1) The target organ appears small because it comprises only the head or tail portion of another organ. (2) The target organ is surrounded by another organ with a similar color. In such scenarios, we found that other methods produce more reliable results than our approach, as illustrated in Figure 8. We believe this discrepancy arises because our approach does not utilize any segmentation information during pre-training. This can make capturing the organ structures more challenging compared to other methods which undergo training with an image segmentation task.

5.7.2. Inference Time Analysis

While performance is undoubtedly crucial in interactive medical segmentation, the inference speed at which segmentation results are produced, especially after receiving user interaction, is also important. Table 7 shows the evaluation of the inference time. As it indicates, we found that our approach takes longer time compared to other methods, mainly due to our iterative process. Although our iterative approach can yield more accurate segmentation results with fewer interactions, it requires more time to compute the masks. In our future work, reducing inference time without compromising segmentation accuracy will be our primary objective.

6. Conclusions

In this research paper, we introduce PixelDiffuser, a novel interactive medical segmentation approach that leverages a pretrained autoencoder, thereby eliminating the need for segmentation data. Compared to other methods [9,10,11], our proposed model can predict highly accurate segmentation with just a few clicks, a process made simple by merely clicking on the target organ area. Additionally, since our model is composed of the VGG19 encoder and the convolutional decoder, it occupies less memory than previous works which utilize huge transformer architectures as a backbone. On the other hand, we found that our method takes a long time to calculate the segmentation due to its iterative process. To speed up the inference time and to solve the failure cases, described in Section 5.7.1, will be our future work.

Author Contributions

Conceptualization, Y.K. and M.J.; methodology, Y.K., M.J. and J.Y.; software, M.J. and J.Y.; validation, M.J., J.Y., J.L. and J.J.; formal analysis, Y.K., M.J. and J.Y.; investigation, J.L., M.L. and J.J.; data curation, M.L. and J.J.; writing—original draft preparation, M.J., J.L. and M.L.; writing—review and editing, all authors; visualization, M.J., J.Y. and J.L.; supervision, Y.K.; project administration, Y.K.; funding acquisition, Y.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grantfunded by the Korea government (MSIT) (No. RS-2022-00155885, Artificial Intelligence Convergence Innovation Human Resources Development (Hanyang University ERICA)) and the Ministry of Trade, Industry and Energy (MOTIE) and Korea Institute for Advancement of Technology (KIAT) through the International Cooperative R&D program (Project No. P0025661).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All datasets used in this research are publicly available online. The BTCV dataset can be accessed from https://www.synapse.org/#!Synapse:syn3193805/wiki/217789 (accessed on 6 January 2023), and the CHAOS dataset can be downloaded from https://chaos.grand-challenge.org/ (accessed on 9 October 2023).

Conflicts of Interest

The authors declare that they have no conflicts of interest. Additionally, the funders played no role in the study’s design, data collection, analysis, or interpretation, the writing of the manuscript, nor the decision to publish the results.

References

  1. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar]
  2. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-unet: Unet-like pure transformer for medical image segmentation. arXiv 2021, arXiv:2105.05537. [Google Scholar]
  3. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.; Roth, H.R.; Xu, D. Unetr: Transformers for 3d medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 574–584. [Google Scholar]
  4. Hatamizadeh, A.; Nath, V.; Tang, Y.; Yang, D.; Roth, H.R.; Xu, D. Swin Unetr: Swin transformers for semantic segmentation of brain tumors in MRI images. In Proceedings of the International MICCAI Brainlesion Workshop, Virtual Event, 27 September 2021; pp. 272–284. [Google Scholar]
  5. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  6. Wang, G.; Li, W.; Zuluaga, M.A.; Pratt, R.; Patel, P.A.; Aertsen, M.; Doel, T.; David, A.L.; Deprest, J.; Ourselin, S.; et al. Interactive medical image segmentation using deep learning with image-specific fine tuning. IEEE Trans. Med. Imaging 2018, 37, 1562–1573. [Google Scholar] [CrossRef] [PubMed]
  7. Wang, G.; Zuluaga, M.A.; Li, W.; Pratt, R.; Patel, P.A.; Aertsen, M.; Doel, T.; David, A.L.; Deprest, J.; Ourselin, S.; et al. DeepIGeoS: A deep interactive geodesic framework for medical image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1559–1572. [Google Scholar] [CrossRef] [PubMed]
  8. Luo, X.; Wang, G.; Song, T.; Zhang, J.; Aertsen, M.; Deprest, J.; Ourselin, S.; Vercauteren, T.; Zhang, S. MIDeepSeg: Minimally interactive segmentation of unseen objects from medical images using deep learning. Med. Image Anal. 2021, 72, 102102. [Google Scholar] [CrossRef]
  9. Sofiiuk, K.; Petrov, I.A.; Konushin, A. Reviving Iterative Training with Mask Guidance for Interactive Segmentation. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 3141–3145. [Google Scholar]
  10. Liu, Q.; Xu, Z.; Bertasius, G.; Niethammer, M. SimpleClick: Interactive Image Segmentation with Simple Vision Transformers. arXiv 2022, arXiv:2210.11006. [Google Scholar]
  11. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment anything. arXiv 2023, arXiv:2304.02643. [Google Scholar]
  12. Landman, B.; Xu, Z.; Igelsias, J.; Styner, M.; Langerak, T.; Klein, A. MICCAI multi-atlas labeling beyond the cranial vault–workshop and challenge. In Proceedings of the MICCAI Multi-Atlas Labeling beyond Cranial Vault—Workshop Challenge, Munich, Germany, 5–9 October 2015; Volume 5, p. 12. [Google Scholar]
  13. Kavur, A.E.; Gezer, N.S.; Barış, M.; Aslan, S.; Conze, P.H.; Groza, V.; Pham, D.D.; Chatterjee, S.; Ernst, P.; Özkan, S.; et al. CHAOS challenge-combined (CT-MR) healthy abdominal organ segmentation. Med. Image Anal. 2021, 69, 101950. [Google Scholar] [CrossRef] [PubMed]
  14. Ma, J.; Wang, B. Segment anything in medical images. arXiv 2023, arXiv:2304.12306. [Google Scholar]
  15. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  17. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  18. Thambawita, V.; Hicks, S.A.; Halvorsen, P.; Riegler, M.A. Divergentnets: Medical image segmentation by network ensemble. arXiv 2021, arXiv:2107.00283. [Google Scholar]
  19. Zhou, L.; Wang, S.; Sun, K.; Zhou, T.; Yan, F.; Shen, D. Three-dimensional affinity learning based multi-branch ensemble network for breast tumor segmentation in MRI. Pattern Recognit. 2022, 129, 108723. [Google Scholar] [CrossRef]
  20. Dang, T.; Luong, A.V.; Liew, A.W.C.; McCall, J.; Nguyen, T.T. Ensemble of deep learning models with surrogate-based optimization for medical image segmentation. In Proceedings of the 2022 IEEE Congress on Evolutionary Computation (CEC), Padua, Italy, 18–23 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–8. [Google Scholar]
  21. Georgescu, M.I.; Ionescu, R.T.; Miron, A.I. Diversity-Promoting Ensemble for Medical Image Segmentation. In Proceedings of the 38th ACM/SIGAPP Symposium on Applied Computing, Tallinn, Estonia, 27–31 March 2023; pp. 599–606. [Google Scholar]
  22. Xu, C.; Prince, J.L. Snakes, shapes, and gradient vector flow. IEEE Trans. Image Process. 1998, 7, 359–369. [Google Scholar] [PubMed]
  23. Rother, C.; Kolmogorov, V.; Blake, A. GrabCut: Interactive foreground extraction using iterated graph cuts. ACM Trans. Graph. (TOG) 2004, 23, 309–314. [Google Scholar] [CrossRef]
  24. Boykov, Y.Y.; Jolly, M.P. Interactive graph cuts for optimal boundary & region segmentation of objects in ND images. In Proceedings of the Eighth IEEE International Conference on Computer Vision (ICCV 2001), Vancouver, BC, Canada, 7–14 July 2001; IEEE: Piscataway, NJ, USA, 2001; Volume 1, pp. 105–112. [Google Scholar]
  25. Grady, L. Random walks for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1768–1783. [Google Scholar] [CrossRef] [PubMed]
  26. Vezhnevets, V.; Konouchine, V. GrowCut: Interactive multi-label ND image segmentation by cellular automata. In Proceedings of the Graphicon. Citeseer, Novosibirsk Akademgorodok, Russia, 20–24 June 2005; Volume 1, pp. 150–156. [Google Scholar]
  27. Criminisi, A.; Sharp, T.; Blake, A. Geos: Geodesic image segmentation. In Proceedings of the European Conference on Computer Vision, Marseille, France, 12–18 October 2008; Springer: Berlin/Heidelberg, Germany, 2008; pp. 99–112. [Google Scholar]
  28. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  29. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  30. Otsu, N. A threshold selection method from gray level histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  31. Scott, D.W. Multivariate Density Estimation: Theory, Practice, and Visualization; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  32. Wang, Z.; Zhao, L.; Chen, H.; Li, A.; Zuo, Z.; Xing, W.; Lu, D. Texture Reformer: Towards Fast and Universal Interactive Texture Transfer. arXiv 2021, arXiv:2112.02788. [Google Scholar] [CrossRef]
  33. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  34. Zhao, X.; Ding, W.; An, Y.; Du, Y.; Yu, T.; Li, M.; Tang, M.; Wang, J. Fast Segment Anything. arXiv 2023, arXiv:2306.12156. [Google Scholar]
  35. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  36. Glenn, J.; Chaurasia, A.; Qiu, J. YOLOv8-seg by Ultralytics. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 10 October 2023).
Figure 1. Illustration of reconstruction noises.
Figure 1. Illustration of reconstruction noises.
Bioengineering 10 01280 g001
Figure 2. An overall illustration of our initial segmentation process. The encoder and decoder derive from the VGG19 autoencoder architecture and are pre-trained on the COCO dataset [29].
Figure 2. An overall illustration of our initial segmentation process. The encoder and decoder derive from the VGG19 autoencoder architecture and are pre-trained on the COCO dataset [29].
Bioengineering 10 01280 g002
Figure 3. An example of kernel density estimation used to adjust the intensity values of the CT image. First, the masked region is cropped using the mask. Then, the density of pixel intensities within this cropped area is computed using the kernel density estimator. In computing the density, we exclude pixels whose value is zero.
Figure 3. An example of kernel density estimation used to adjust the intensity values of the CT image. First, the masked region is cropped using the mask. Then, the density of pixel intensities within this cropped area is computed using the kernel density estimator. In computing the density, we exclude pixels whose value is zero.
Bioengineering 10 01280 g003
Figure 4. An example of the density-based intensity adjustment process. Given the density, the pixel intensities in the CT image are transformed based on the density values corresponding to each pixel value. (a) in this figure represents the CT image before transformation, and (b) is the transformed image using the density values.
Figure 4. An example of the density-based intensity adjustment process. Given the density, the pixel intensities in the CT image are transformed based on the density values corresponding to each pixel value. (a) in this figure represents the CT image before transformation, and (b) is the transformed image using the density values.
Bioengineering 10 01280 g004
Figure 5. An illustration of our refinement process. The two leftmost images depict the ground truth segmentation and an initial segmentation prediction, respectively. With the initial segmentation, a user can provide additional clicks to refine the segmentation mask or choose to finalize it. If a positive click is given, our method appends an additional area to the initial mask. Conversely, a negative click discards parts of certain mask regions. The third image showcases the utilization of a negative click, where parts of the mask region are discarded when the negative Gaussian density surpasses the positive density, as demonstrated in the fourth and rightmost images.
Figure 5. An illustration of our refinement process. The two leftmost images depict the ground truth segmentation and an initial segmentation prediction, respectively. With the initial segmentation, a user can provide additional clicks to refine the segmentation mask or choose to finalize it. If a positive click is given, our method appends an additional area to the initial mask. Conversely, a negative click discards parts of certain mask regions. The third image showcases the utilization of a negative click, where parts of the mask region are discarded when the negative Gaussian density surpasses the positive density, as demonstrated in the fourth and rightmost images.
Bioengineering 10 01280 g005
Figure 6. Examples of interactive segmentation for the 13 organs, varying by the number of clicks.
Figure 6. Examples of interactive segmentation for the 13 organs, varying by the number of clicks.
Bioengineering 10 01280 g006
Figure 7. Quantitative evaluation. The solid lines represent the average dice score of the 13 organs, while dashed lines represent the 25 and 75 quantile lines.
Figure 7. Quantitative evaluation. The solid lines represent the average dice score of the 13 organs, while dashed lines represent the 25 and 75 quantile lines.
Bioengineering 10 01280 g007
Figure 8. An illustration of failure cases of our approaches, and results of other methods. When a target organ is too small or surrounded by another organ with a similar texture, our method tends to fail to capture a correct segmentation mask. This is a limitation of label-free texture-based segmentation methods, involving our approach.
Figure 8. An illustration of failure cases of our approaches, and results of other methods. When a target organ is too small or surrounded by another organ with a similar texture, our method tends to fail to capture a correct segmentation mask. This is a limitation of label-free texture-based segmentation methods, involving our approach.
Bioengineering 10 01280 g008
Table 1. The autoencoder layers of PixelDiffuser. For each Conv layer, parameters are listed in order: number of input channels, number of output channels, kernel size, stride, and padding, respectively. The MaxPool layer parameters, listed, respectively, are kernel size and stride. The Upsampling layer is parameterized by its scale factor. Additionally, all Conv layers involve subsequent ReLU activation functions.
Table 1. The autoencoder layers of PixelDiffuser. For each Conv layer, parameters are listed in order: number of input channels, number of output channels, kernel size, stride, and padding, respectively. The MaxPool layer parameters, listed, respectively, are kernel size and stride. The Upsampling layer is parameterized by its scale factor. Additionally, all Conv layers involve subsequent ReLU activation functions.
EncoderDecoder
Conv 0 (3, 3, 1, 1, 0)
Conv 1_1 (3, 64, 3, 1, 0)
Conv 1_2 (64, 64, 3, 1, 0)
MaxPool (2, 2)
Conv 3_1 (256, 128, 3, 1, 0)
Upsampling (2)
Conv 2_1 (64, 128, 3, 1, 0)
Conv 2_2 (128, 128, 3, 1, 0)
MaxPool (2, 2)
Conv 2_2 (128, 128, 3, 1, 0)
Conv 2_1 (128, 64, 3, 1, 0)
Upsampling (2)
Conv 3_1 (128, 256, 3, 1, 0)Conv 1_2 (64, 64, 3, 1, 0)
Table 2. Dice score evaluation for the 13 organs using BTCV dataset.
Table 2. Dice score evaluation for the 13 organs using BTCV dataset.
SpleenRight KidneyLeft Kidney
Dice Score per Click
Methods123451234512345
RITM-HRNet-180.090.570.820.900.930.090.460.760.860.900.090.680.840.890.91
RITM-HRNet-320.100.590.850.910.940.060.430.800.890.920.050.490.780.870.91
SimpleClick-Base0.080.330.580.750.840.070.410.640.770.840.060.340.620.770.83
SimpleClick-Huge0.150.600.820.890.910.540.780.860.890.910.510.800.870.890.91
SAM-base0.090.090.090.110.150.050.050.070.190.480.050.050.050.080.16
SAM-Large0.300.450.560.640.670.290.440.520.580.640.320.450.530.590.65
SAM-Huge0.880.910.910.920.920.870.880.860.840.860.870.870.860.840.86
FastSAM0.030.450.100.040.410.020.600.190.050.570.020.660.150.020.62
Ours0.810.910.930.940.940.650.810.870.940.940.680.840.890.910.92
GallbladderEsophagusLiver
Dice Score per Click
Methods123451234512345
RITM-HRNet-180.040.240.540.730.820.080.440.660.770.830.210.320.560.740.84
RITM-HRNet-320.030.430.760.840.880.030.550.740.820.860.220.420.730.850.89
SimpleClick-Base0.020.110.390.560.660.040.150.330.480.590.210.310.490.660.78
SimpleClick-Huge0.180.470.650.760.810.090.360.590.710.770.270.550.770.870.91
SAM-base0.020.020.020.030.070.020.020.050.140.250.230.230.260.280.40
SAM-Large0.120.180.230.270.310.180.200.210.210.220.420.520.580.630.66
SAM-Huge0.620.660.700.740.760.490.570.640.700.730.840.880.900.890.90
FastSAM0.010.040.050.020.050.000.010.000.000.010.120.240.170.090.27
Ours0.710.790.810.830.840.560.660.710.730.760.700.870.920.940.94
StomachAortaInferior Vena Cava
Dice Score per Click
Methods123451234512345
RITM-HRNet-180.160.580.770.850.880.060.590.780.850.890.060.300.570.740.83
RITM-HRNet-320.130.550.760.850.890.050.640.830.870.900.020.390.730.840.88
SimpleClick-Base0.150.430.630.750.820.100.260.510.680.770.020.120.330.520.65
SimpleClick-Huge0.410.740.840.890.910.230.620.790.850.880.110.460.690.800.85
SAM-base0.100.100.120.180.290.030.030.060.210.420.020.020.030.080.19
SAM-Large0.370.460.490.520.540.240.340.470.570.600.140.200.250.300.33
SAM-Huge0.670.730.800.840.860.900.900.920.920.900.680.720.750.790.81
FastSAM0.040.340.090.230.160.010.430.200.030.400.010.150.130.010.15
Ours0.460.610.680.730.760.870.910.910.920.920.640.710.750.780.81
Portal Vein and Splenic VeinPancreasRight Adrenal Gland
Dice Score per Click
Methods123451234512345
RITM-HRNet-180.040.230.460.620.710.030.270.510.650.740.000.030.210.440.58
RITM-HRNet-320.020.330.590.710.770.030.270.570.720.790.000.170.500.640.71
SimpleClick-Base0.030.120.280.410.520.040.130.300.440.550.000.020.080.160.25
SimpleClick-Huge0.110.350.530.630.700.170.480.640.730.780.010.060.220.370.49
SAM-base0.020.020.040.090.150.030.030.030.040.080.000.000.000.000.02
SAM-Large0.150.190.210.220.210.150.200.240.270.290.010.010.020.020.03
SAM-Huge0.590.620.650.690.710.530.600.650.700.720.140.220.300.370.41
FastSAM0.010.130.040.070.080.010.030.020.020.020.000.000.000.000.00
Ours0.600.700.740.760.780.560.690.750.780.800.330.400.450.490.53
Left Adrenal Gland
Dice Score per Click
Methods12345
RITM-HRNet-180.000.080.310.520.65
RITM-HRNet-320.000.170.500.640.72
SimpleClick-Base0.000.020.080.170.27
SimpleClick-Huge0.010.110.280.420.52
SAM-base0.000.000.000.010.02
SAM-Large0.020.040.050.050.04
SAM-Huge0.350.420.480.540.56
FastSAM0.000.000.000.000.00
Ours0.560.630.660.680.69
Table 3. The number of parameters of SimpleClick, SAM and PixelDifuser. SC-huge and SC-base represent SimpleClick-huge and SimpleClick-base, respectively.
Table 3. The number of parameters of SimpleClick, SAM and PixelDifuser. SC-huge and SC-base represent SimpleClick-huge and SimpleClick-base, respectively.
SC-HugeSC-BaseSAM-HugeSAM-BaseFastSAMOurs
params662,184,79398,025,513641,090,60893,735,47272,234,1497,010,959
Table 4. Dice score evaluation for the liver using CHAOS CT dataset.
Table 4. Dice score evaluation for the liver using CHAOS CT dataset.
Liver
Dice score per click
Methods12345
RITM-HRNet-180.230.350.570.760.85
RITM-HRNet-320.250.480.770.880.91
SimpleClick-Base0.240.330.510.680.79
SimpleClick-Huge0.280.560.790.890.93
SAM-base0.260.270.300.380.52
SAM-Large0.510.620.680.710.74
SAM-Huge0.800.890.910.900.91
FastSAM0.140.450.120.140.44
Ours0.730.890.930.950.96
Table 5. Dice score evaluation for the 4 organs using CHAOS T1-DUAL (MRI) dataset.
Table 5. Dice score evaluation for the 4 organs using CHAOS T1-DUAL (MRI) dataset.
LiverRight Kidney
Dice Score per Click
Methods1234512345
RITM-HRNet-180.220.450.710.830.880.050.370.630.740.81
RITM-HRNet-320.220.380.700.840.900.050.200.600.750.82
SimpleClick-Base0.230.300.500.700.810.050.280.580.740.83
SimpleClick-Huge0.310.740.870.910.930.120.660.810.870.90
SAM-base0.230.230.240.270.360.050.060.060.060.08
SAM-Large0.380.440.470.480.500.170.210.250.270.30
SAM-Huge0.740.820.860.860.880.780.820.850.850.86
FastSAM0.090.140.090.050.170.020.120.020.030.10
Ours0.420.600.710.770.810.560.690.760.790.82
Left KidneySpleen
Dice Score per Click
Methods1234512345
RITM-HRNet-180.050.240.550.700.790.070.350.660.800.86
RITM-HRNet-320.050.220.610.750.820.070.390.720.840.89
SimpleClick-Base0.060.310.670.790.850.080.270.600.750.83
SimpleClick-Huge0.170.610.810.870.900.130.560.780.860.90
SAM-base0.050.060.060.090.160.080.080.080.100.14
SAM-Large0.240.270.330.370.360.200.290.330.380.37
SAM-Huge0.760.820.840.860.860.760.780.820.840.85
FastSAM0.020.160.040.070.110.030.180.020.110.10
Ours0.560.700.750.780.810.510.690.770.810.84
Table 6. Dice score evaluation for the 4 organs using CHAOS T2-SPIR (MRI) dataset.
Table 6. Dice score evaluation for the 4 organs using CHAOS T2-SPIR (MRI) dataset.
LiverRight Kidney
Dice Score per Click
Methods1234512345
RITM-HRNet-180.230.440.690.820.880.060.570.780.850.89
RITM-HRNet-320.240.420.710.840.890.060.420.780.850.89
SimpleClick-Base0.280.380.610.760.830.210.410.640.750.83
SimpleClick-Huge0.330.610.830.900.920.630.830.890.920.93
SAM-base0.280.290.300.330.390.080.080.100.140.22
SAM-Large0.510.570.620.650.670.540.620.670.730.74
SAM-Huge0.640.760.820.850.870.800.860.860.880.89
FastSAM0.120.280.120.090.310.020.670.140.090.56
Ours0.450.630.740.790.830.790.870.900.910.91
Left KidneySpleen
Dice Score per Click
Methods1234512345
RITM-HRNet-180.050.430.720.830.880.080.540.800.870.91
RITM-HRNet-320.050.460.760.850.880.090.660.840.890.91
SimpleClick-Base0.330.560.710.800.860.390.640.790.860.90
SimpleClick-Huge0.660.830.880.900.920.640.830.880.920.93
SAM-base0.070.080.090.130.240.110.110.130.160.23
SAM-Large0.550.630.700.760.760.410.520.640.680.69
SAM-Huge0.780.820.860.870.870.860.890.900.900.92
FastSAM0.020.730.120.080.590.040.650.090.060.55
Ours0.790.880.900.910.910.700.840.880.900.91
Table 7. Inference time comparison. For each method, we measure the inference time for each organ and calculate the min, max, and mean values. In these results, we can find the difference between the min and max time of the PixelDiffuser.
Table 7. Inference time comparison. For each method, we measure the inference time for each organ and calculate the min, max, and mean values. In these results, we can find the difference between the min and max time of the PixelDiffuser.
Infer Time (s)RITMSimpleClickSAMFastSAMPixelDiffuser
Best0.11340.72581.05020.08982.0513
Worst0.12300.77081.05190.75365.6538
Mean0.11830.73831.05080.14383.2868
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ju, M.; Yang, J.; Lee, J.; Lee, M.; Ji, J.; Kim, Y. Pixel Diffuser: Practical Interactive Medical Image Segmentation without Ground Truth. Bioengineering 2023, 10, 1280. https://doi.org/10.3390/bioengineering10111280

AMA Style

Ju M, Yang J, Lee J, Lee M, Ji J, Kim Y. Pixel Diffuser: Practical Interactive Medical Image Segmentation without Ground Truth. Bioengineering. 2023; 10(11):1280. https://doi.org/10.3390/bioengineering10111280

Chicago/Turabian Style

Ju, Mingeon, Jaewoo Yang, Jaeyoung Lee, Moonhyun Lee, Junyung Ji, and Younghoon Kim. 2023. "Pixel Diffuser: Practical Interactive Medical Image Segmentation without Ground Truth" Bioengineering 10, no. 11: 1280. https://doi.org/10.3390/bioengineering10111280

APA Style

Ju, M., Yang, J., Lee, J., Lee, M., Ji, J., & Kim, Y. (2023). Pixel Diffuser: Practical Interactive Medical Image Segmentation without Ground Truth. Bioengineering, 10(11), 1280. https://doi.org/10.3390/bioengineering10111280

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop