Next Article in Journal
Integrating Wearable Sensor Signal Processing with Unsupervised Learning Methods for Tremor Classification in Parkinson’s Disease
Previous Article in Journal
An Effective Methodology for Diabetes Prediction in the Case of Class Imbalance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Uncertainty-Aware Semi-Supervised Method for Pectoral Muscle Segmentation

1
School of Computer Science and Engineering, Sun-Yat sen University, Guanghzou 510006, China
2
The Third Affiliated Hospital of Guangzhou Medical University, Guanghzou 510150, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Bioengineering 2025, 12(1), 36; https://doi.org/10.3390/bioengineering12010036
Submission received: 19 November 2024 / Revised: 19 December 2024 / Accepted: 27 December 2024 / Published: 6 January 2025
(This article belongs to the Section Biosignal Processing)

Abstract

The consistency regularization method is a widely used semi-supervised method that uses regularization terms constructed from unlabeled data to improve model performance. Poor-quality target predictions in regularization terms produce noisy gradient flows during training, resulting in a degradation in model performance. Recent semi-supervised methods usually filter out low-confidence target predictions to alleviate this problem, but also prevent the model from learning features from unlabeled data in low-confidence regions. Specifically, in medical imaging and other cross-domain scenarios, models are prone to producing large numbers of low-confidence predictions. To improve the quality of target predictions while utilizing unlabeled data more efficiently, we propose an uncertainty-aware semi-supervised method that incorporates the breast anatomical prior, for pectoral muscle segmentation. Our method has a typical teacher-student dual model structure, where uncertainty is used to distinguish between high- and low-confidence predictions in the teacher model output. A low-confidence prediction refinement module was designed to refine the low-confidence predictions by incorporating high-confidence predictions and a learned anatomical prior. The anatomical prior, as regularization of the target predictions, was learned from annotations and an auxiliary task. The final target predictions are a combination of high-confidence teacher predictions and refined low-confidence predictions. The proposed method was evaluated on a dataset containing 635 data points from three data centers. Compared with the baseline method, the proposed method showed an average improvement in DICE index of 1.76, an average reduction in IoU index of 3.21, and an average reduction in HD index of 5.48. The experimental results show that our method generalizes well to the test set and outperforms other methods in all evaluation metrics.

1. Introduction

Mammography is a procedure that uses low-energy X-rays to photograph the breast and is widely used for the early screening and diagnosis of breast cancer. The breast border, the nipple, and the pectoral muscle are three important anatomical markers on mammograms [1,2]. Specifically, the pectoral muscle is usually identified first in mammography-based screening and diagnosis. With the development of computer-aided detection technology, an increasing number of computer-aided detection systems have been applied in clinical practice, effectively reducing the burden of manual screening. Compared with the fatty tissue in the breast, the pectoral muscle usually has a higher intensity in mammograms, similar to that of breast masses and glandular tissue. This similarity may lead to false positives in automated scenarios, such as automatic breast mass detection and automatic breast density segmentation, where the pectoral muscle may be misidentified as the target. To avoid this, pectoral muscle removal is usually performed as a pre-task in computer-aided breast cancer detection and diagnosis algorithms [3,4].
Many automatic pectoral muscle segmentation methods have been proposed for mammography over the past few decades. Karssemeijer [5] used the Hough transform [6] to detect the pectoral boundary in breast parenchyma classification. Kwok et al. [1] used a cliff search method to improve the straight-line estimation obtained through the Hough transform and fitted it with polynomial curves. Ferrari et al. [7] designed a set of Gabor wavelet functions to enhance objects with a linear structure and then used the edge-flow method [8] to detect the pectoral boundary. Ma et al. [9] modeled the pectoral muscle segmentation problem separately using graph methods and fine-tuned the obtained pectoral muscle regions using an active contour model. In addition, many other methods based on thresholding [10,11,12,13,14,15], region-growing [16,17,18,19], clustering [20,21], graph methods [22], and boundary detection [23,24,25,26,27,28,29,30] have been proposed. Although these methods produce good results on most mammograms, they mostly rely on some simple assumptions about the intensity or shape of the pectoral muscle; thus, it is difficult to handle more complex cases. For example, threshold-based methods assume that the pectoral muscle has consistent and easily distinguishable high intensity; boundary detection methods assume that the boundary between the pectoral muscle and breast parenchyma has a cliff-like change in intensity within a small neighborhood; and, region-based methods, such as region growth, clustering, and graph methods, assume that the intensity in the pectoral muscle region is more consistent than that near its boundary. When mammograms are properly imaged, most of them can satisfy the above assumptions; however, in some mammograms, the pectoral muscle overlaps with its nearby lesions or glands, a situation in which the above assumptions cannot be validated. Incorrect breast compression and posturing during photography can also lead to changes in the intensity distribution of mammograms, in which case, the pectoral boundary is blurred or invisible [31]. In addition, most of the above methods assume that the pectoral boundary can be approximated as a straight line or simple curve, and they fit the pectoral boundary using straight-line estimation or polynomial curves. However, the shape of the pectoral muscle in mammograms is either convex, concave, or a combination of both, and polynomial curves cannot express this correctly.
In recent years, with the successful application of deep learning (DL) in semantic image segmentation, DL-based algorithms for pectoral muscle segmentation have been developed. Ma et al. [32] and Maghsoudi et al. [4] used U-Net [33] to segment the pectoral muscle. Rampun et al. [34] modified the holistically nested edge detection network [35] to detect the pectoral boundary. These DL-based methods are data-driven, and their optimization objectives are directly related to the task as loss functions. The convolutional neural networks (CNNs) used in these methods contain many learnable parameters and are automatically optimized during the training process, which avoids manual design features with oversimplified assumptions. These DL-based methods improve pectoral muscle segmentation results but still suffer from the following shortcomings:
  • The task of pectoral muscle segmentation can be split into two subtasks: detection of visible parts and estimation of invisible parts. The above methods only optimize for the final segmentation results, and the degree of optimization of the two tasks is difficult to balance during the training process. In most mammogram datasets, most images have a clear pectoral boundary, and the network tends to learn how to detect visible parts and ignore how to estimate blurred or missing parts. The performance of the model completely depends on the proportion of different types of data and the training tricks used, without addressing the problem through method design.
  • The weights of CNNs are highly correlated with the domain-relevant information of the training data, and the performance of the model may drop substantially when the training and test data are collected from different data centers.
  • Segmentation networks trained in a fully supervised manner require many pixel-level annotations. The datasets used by Ma et al. [32], Rampun et al. [34], and Maghsoudi et al. [4] contain 729, 1078, and 1100 labeled mammograms, respectively. In the field of medical image segmentation, the data annotations must be delineated by doctors specialized in the field, which is extremely costly in both time and labor.
In our previous work [36], a two-stage method for pectoral muscle segmentation was proposed to solve the first problem mentioned above using two sub-networks to solve the two subtasks separately. To solve the second and third problems, based on our previous work [36], this paper proposes a semi-supervised method for pectoral muscle segmentation in mammograms.
The key to the design of semi-supervised learning methods is the effective use of unlabeled data to improve the performance of the model. Many semi-supervised learning methods have been developed, among which the consistency regularity method is one of the most widely used. Consistency regularization is based on the manifold or smoothness assumption [37], which assumes that reasonable perturbations to the data should not cause changes in the model predictions. Accordingly, the consistency regularity term of the model can be constructed by constraining the output of the perturbed data.
Among the factors that affect the performance of the consistency regularity method, the degree of disagreement between predictions and the quality of predictions are two often-considered ones. Consistency regularization methods use the disagreement between predictions as the regularization term, and highly consistent predictions reduce the gain from the regularization term associated with unlabeled data. Meanwhile, if the prediction quality of unlabeled data is insufficient, the model suffers from consistency regularity and will struggle to learn new correct information, which is often referred to as confirmation bias [38]. To alleviate confirmation bias, many uncertainty-aware semi-supervised methods [39,40,41] simply filter out low-confidence predictions and train with only high-confidence predictions. This reduces not only the impact of confirmation bias but also the disagreement available for consistency regularization methods, and it does not make sufficient use of unlabeled data. Recent research on domain adaptation [42,43] has shown that models often produce low-confidence predictions when the data domain changes [44]. This implies that low-confidence prediction regions may contain domain-relevant information. Therefore, fully utilizing low-confidence predictions will effectively improve the generalization ability of the consistency regularization method on the test dataset.
The identification of blurred or missing boundaries is the most difficult part of pectoral boundary delineation and usually leads to low-confidence predictions in pectoral muscle segmentation. When doctors perform delineation, they consider not only the visible pectoral boundary and shape of the pectoral muscle but also the overall anatomical structure. Therefore, we propose a consistency regularization method that considers the uncertainty of pectoral muscle segmentation in mammograms. This method uses uncertainty to measure the confidence of the predictions and refines the low-confidence target predictions in unlabeled data so that the unlabeled data can be efficiently utilized. Our semi-supervised pectoral muscle segmentation method enables the model to learn domain-related information from the unlabeled data from the validation dataset during training, and improves the generalization ability of the model on the test dataset.

2. Materials and Methods

2.1. Datasets and Pre-Processing

The proposed uncertainty-aware semi-supervised method for pectoral muscle segmentation was evaluated using our in-house dataset. Mammograms were collected from three data centers using different mammography devices. The entire dataset contained 635 mammograms, with each data center providing approximately 200 mammograms. An experienced radiologist annotated the pectoral muscles of all mammograms in the dataset. Detailed information on the dataset is presented in Table 1.
To train the model effectively, all data in the dataset were pre-processed with orientation, size, and intensity normalization before training. First, all images in the dataset were resized to 448 × 320 by bilinear interpolation, which is a compromise between the hardware resources and image resolution, considering that the aspect ratios of the mammogram in all three data centers are close to 7:5, and adjusting to this ratio can reduce image distortion. Second, the intensity was normalized using a percentile normalization method, where the intensities below 10% and above 90% were set to the lowest and highest intensities, respectively, and then the overall intensity was mapped to [0, 1]. Finally, all mammograms were oriented according to the intensity distribution such that the pectoral muscle in all mammograms was located in the upper-left corner of the image. Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

2.2. Overview of Methods

The proposed consistency regularization method considers uncertainty to measure the confidence of unlabeled data predictions and improve the contribution of unlabeled data in model training by refining low-confidence predictions using prior knowledge obtained from annotations. The overall framework of this method is shown in Figure 1.
The teacher and student models in Figure 1 were derived from the mean teacher (MT) method. The MT method is a consistency regularization method proposed by Tarvainen et al. [38], with a typical teacher–student dual model structure. The student updates the weights using the back-propagation algorithm, while the weights of the teacher are updated by the exponential moving average (EMA) of the student weights. Compared with the student, the teacher changes its weights more smoothly and is expected to provide more stable predictions. Therefore, the teacher predictions can be considered pseudo-labels for the corresponding unlabeled data to guide student training. The loss function L M T of the student model in the MT method comprises two components:
L M T = L s u p M T + L c o n M T ,
where L s u p M T is the supervised loss from the labeled dataset, and L c o n M T is the consistency regularization loss between the predictions of the unlabeled data and its pseudo-labels. L s u p M T and L c o n M T are given in Equations (2) and (3), respectively.
L s u p M T = E x s ( f ( A ( x s ) , θ ) y s 2 ) ,
L c o n M T = E x u ( f ( A ( x u ) , θ ) y u 2 ) ,
where x s and x u are the samples from the labeled and unlabeled datasets, respectively, y s is the label corresponding to x s , and y u is the pseudo-label corresponding to x u . f ( · , θ ) is the segmentation network with student model weight θ . A denotes a random perturbation of the input data. In the MT model, y u is obtained from the direct prediction of the teacher model:
y u = f ( A ( x u ) , θ ) ,
where θ is the weight of the teacher network, updated by EMA.
Unlike the MT method, which uses the teacher predictions as pseudo-labels for student training, we want to obtain pseudo-labels that are more reliable than the teacher predictions to alleviate confirmation bias while making efficient use of the information from the unlabeled data. To achieve this, we must first identify which part of the teacher prediction y u is reliable and then improve the quality of the less reliable part of the predictions. Therefore, we use confidence to indicate whether a certain part of y u is reliable, where β h and β l denote the confidence indicators for the high- and low-confidence regions of y u , respectively.
One way to measure confidence is uncertainty, which describes the degree of inconsistency in predictions and can be used as an estimation of confidence. High and low uncertainty can be considered predictions with low and high confidence, respectively. Accordingly, we define
β h = 1 ( Π < δ ) ,
β l = 1 ( Π δ ) ,
where Π is the uncertainty of the teacher predictions, as output by the uncertainty estimation module in Figure 1, δ is a threshold, and 1 is the indicator function
1 ( x ) = 1 , if x 0 , 0 , otherwise .
Once the high- and low-confidence regions of the teacher predictions y u are determined, the next step is to improve the quality of the low-confidence prediction. When performing pectoral muscle delineation, doctors first look for clearly visible pectoral boundaries in the mammogram images. Second, they consider factors such as the overall shape of the pectoral muscle and anatomical structure of the breast to determine the location of blurred or invisible boundaries.
Inspired by doctors’ delineation method, we designed a low-confidence prediction refinement module to learn how to refine low-confidence predictions using the visible boundaries, shape prior of the pectoral muscle, and structure prior of the breast. The visible boundary can be approximated using high-confidence teacher predictions as follows:
y u h = y u β h ,
where ⊙ denotes the Kronecker multiplier. The contrast between the breast and air is high in mammograms; therefore, a simple breast segmentation module was designed to obtain a rough contour of the breast, encoded as the feature S b r e a s t ; more details are given in Section 2.4. Regarding the shape prior of the pectoral muscle, many studies have assumed that the shape of the pectoral boundary can be approximated by a straight line or polynomial curve, but such a simplification may not match the real pectoral boundary shape. In our study, we used a GAN to learn the shape prior of the pectoral muscle.
GAN is a widely studied generative model that has been successfully used in many image-related fields, such as face synthesis [45] and style transfer [46]. GAN can learn a mapping from a source distribution to a target distribution without using pairs of data. We can use random noise as the input and annotations of the pectoral muscle as the target to train a GAN model; however, it is difficult to utilize the shape prior in this way. To make the shape prior easy to use, we use the high-confidence predictions and the breast shape as the input, with the annotation of the pectoral muscle as the target, and we train the GAN to refine the predictions. At this point, the shape prior of the pectoral muscle can be considered implicitly encoded in the GAN model. We also take the high-confidence region indicator β h as an input feature because, as a reasonable prediction, the output of the GAN model should be aligned with the teacher predictions in the high-confidence region. The low-confidence prediction refinement module in Figure 1 is the generator of the GAN, denoted as
g = g ( y p h , β h , S b r e a s t ) .
Once the output of g is obtained, the refined low-confidence predictions can be calculated by constraining the output to the low-confidence region
y ¯ u l = g β l ,
while the target predictions are aligned with the teacher predictions in the high-confidence region
y ¯ u h = y u β h .
The final target predictions are simply the sum of Equations (10) and (11),
y ¯ u = y ¯ u l + y ¯ u h .
Our loss function for student training is
L O u r s = L s u p O u r s + L c o n O u r s ,
L s u p O u r s = E x s ( f ( A ( x s ) , θ ) y s 2 ) ,
L c o n O u r s = E x u ( f ( A ( x u ) , θ ) y ¯ u 2 ) .

2.3. Uncertainty Estimation

An important part of our method is the estimation of uncertainty, which is used to measure the confidence of the teacher predictions. Currently, many uncertainty estimation methods applied in image segmentation and image detection [39,47] are based on Monte Carlo dropout [48]. When dropout is applied, a certain percentage of the weights in the specified layer is randomly set to zero, which can be regarded as a sampling process of the model weights. Thus, a batch of different predictions with the same input can be obtained for uncertainty estimation. However, dropout is usually applied to only a few specific layers of the model, and for the layers without dropout applied, the model weights always remain constant, which may lead to over-confident predictions in uncertainty estimation.
To address this problem, instead of sampling weights from a single model, we save the model weights at different moments in the optimization process. Specifically, after the validation set loss is stabilized during training, the model weights are saved at fixed epochs, denoted as { θ i } . The batch of predictions used for uncertainty estimation can be obtained using the model with different weights as in
y u i = f ( x u , θ i ) .
After obtaining a batch of predictions for each unlabeled data point, the uncertainty can be estimated in various ways, such as through variance [36] or entropy [39,40]. Entropy measures the degree of information confusion, and higher entropy represents a prediction with higher uncertainty. For our method, we use entropy to estimate uncertainty:
Π p i x e l = i = 1 K p i log p i ,
where p = Softmax ( y u i ) , and K is the number of models giving predictions.
The uncertainty estimation in Equation (17) uses a pixel-independent method that does not consider the pectoral muscle shape. If Equation (17) is applied directly, the obtained confidence region will be highly scattered. When performing manual delineation, the clarity of the pectoral boundary is usually the dominant factor in determining whether the delineation is reliable, while pixels away from the boundary are usually ignored. It is easy to see that there are some semantic discrepancies between the uncertainty estimated in a pixel-independent manner and the basis for judging whether the manual delineation is reliable. To eliminate this discrepancy, we propose row-based uncertainty to represent the confidence of the pectoral muscle predictions.
Row-based uncertainty is a simplification of the uncertainty calculation using contour coordinates, which is based on the shape of the pectoral muscle in a mammogram. Assuming that all mammograms are properly oriented, there should be at most one intersection point with the pectoral boundary in each row of the mammogram because the pectoral muscle does not curl excessively during mammography. Given that the confidence of annotations is usually related to the clarity of the pectoral boundary, this can be approximated by the width of the high-response regions in the pixel-independent uncertainty near the boundary. Once the row-based uncertainty, denoted as Π , is obtained, the high-confidence and low-confidence regions can be determined using Equation (5) and Equation (6), respectively.
Following the above design guidelines, the calculation of the row-based uncertainty Π first requires the width of the high-response region near the boundary in Π p i x e l to be obtained. There are three problems to solve here: locating the boundary, obtaining the high-response region, and computing the width of the high-response region. We observe that the clearly visible boundary has a smaller width in the high-response region with pixel-independent uncertainty Π p i x e l . Based on this observation, we simplify the algorithm by directly determining the possible locations of the clear boundary; the rows where the clear boundaries are located belong to the high-confidence region, and the remaining rows belong to the low-confidence region.
The entire algorithm can be described as follows: First, the response of Π p i x e l is processed by the top hat morphological operation, where a larger response can be obtained for structures with smaller widths. Next, the response of the top hat is processed using thresholding to obtain smaller-width structures and suppress noise. At this point, the location of the clear boundary is clearly accessible; however, there is still noise and other tissue structures similar to the pectoral muscle, such as glandular tissue near the pectoral muscle, which may also have a high response. To eliminate these disturbances as much as possible, we choose the largest connected region as the pectoral boundary selected by the algorithm. Experiments show that such a selection strategy can meet our requirements and is sufficiently robust. Before thresholding the top hat responses, we process the acquired response maps using coherence-enhancing shock filters (CESF) [49]. This is because of the random nature of model selection in uncertainty estimation, which can lead to the random presence of low-response pixels in high-response areas. If processing is not performed, the robustness of the large connected region selection strategy is reduced. CESF is an anisotropic diffusion method proposed by Weickert [49] that determines the direction of diffusion based on the gradient of the image within a larger neighborhood. This method is commonly used for image processing with flow-like structures, such as fingerprint recovery, and is therefore also applicable to the processing of pectoralis major boundaries. The experimental results for this algorithm on sample data are shown in Figure 2.

2.4. Low-Confidence Prediction Refinement

To refine the low-confidence predictions, we use GAN to learn the shape prior of the pectoral muscle. GAN is a generative model proposed by Goodfellow [50] that contains two components: a generator and a discriminator. The generator learns how to generate target samples, while the discriminator determines whether the input samples are from the generator or the target dataset, and the two models are trained using a min-max game. At the end of the training, the data generated by the generator can follow the real data distribution, making the discriminator unable to determine the source of the input samples.
In our method, the generator of the GAN is used to implement the low-confidence prediction refinement module g in Equation (9), and the output of g is made to satisfy the shape prior of the pectoral muscle by training the GAN. The input of the low-confidence prediction refinement module g has three features: high-confidence teacher predictions y u h , high-confidence region indicator β h , and breast shape feature S b r e a s t . The acquisition of y u h and β h was described in the previous subsections, and this subsection describes how to obtain S b r e a s t .
In mammograms, there is a large intensity difference between the breast and air. After pre-processing by intensity normalization and label removal, the breast can be obtained through a sequence of morphological operations, including global thresholding, hole filling, and acquisition of the maximum connected area. Next, we must consider how to encode the segmentation result of the breast into a good feature for pectoral muscle prediction refinement. Intuitively, using the distance from the pixel to the breast boundary is a reasonable way to encode the shape of the breast, as the breast boundary and pectoral boundary are important anatomical benchmarks for localization in breast cancer diagnosis. However, factors such as the size of the breast and selection of the field of view at the time of photography can have a large impact on the distance function. In our experiments, we found that a few mammograms showed large variations in the distance function response owing to inappropriate shooting poses, such as including too much of the axilla, which interfered with the results. Therefore, we directly used the breast segmentation result as the encoding of breast shape S b r e a s t and trained the network to automatically learn the relationship between the pectoral boundary and breast.
GAN suffers from pattern collapse during training [51]. To solve this problem, the Wasserstein GAN (W-GAN) improves GAN by constraining the discriminator gradient amplitude in the loss function [51,52], thereby stabilizing the training process. In our method, we use the training method of W-GAN to train our model. The loss functions of W-GAN are
L D W G A N = E y ˜ P G [ D ( y ˜ ) ] E y P r [ D ( y ) ] L D O r i g + λ E y ^ P y ˜ [ ( y ^ D ( y ^ ) 2 1 ) 2 ] L D G r a d ,
L G W G A N = E y ˜ P G [ D ( y ˜ ) ] L G O r i g ,
where y ˜ = G ( x ) , x denotes the input to the generator, y ^ = α y ˜ + ( 1 α ) y , and P G , P r , P y ^ denote the distributions of G ( x ) , true data, and y ˜ , respectively. λ is the weight parameter for the gradient penalty term. L D O r i g , L G O r i g , L D G r a d are the generator loss and discriminator loss of the original GAN [50] and gradient penalty term of the W-GAN, respectively.
To constrain the output of g to be aligned with the teacher predictions over the high-confidence region, we add an additional constraint term to the generator loss of W-GAN, and the discriminator loss is kept consistent with W-GAN. The discriminator and generator losses for training g are
L D W G A N = E y ˜ P G [ D ( y ˜ ) ] E y P r [ D ( y ) ] L D O r i g + λ E y ^ P y ˜ [ ( y ^ D ( y ^ ) 2 1 ) 2 ] L D G r a d ,
L G W G A N = E y ˜ P G [ D ( y ˜ ) ] L G O r i g + y ˜ β h y u h 2 2 L G C o n ,
where L G C o n is the consistency loss over the high-confidence region.
Pectoral boundaries in most mammograms in our dataset are clearly visible, and training directly with high-confidence teacher predictions would make it difficult for the network to learn how to properly handle the inputs with a lower percentage of high-confidence predictions. Therefore, we randomly erase a certain percentage of the pectoral muscle annotations as input features instead of y u h . The training flow of the low-confidence prediction refinement module is shown in Figure 3.

2.5. Data Analysis

To evaluate the performance of the model, we used three metrics: Dice coefficient(Dice), Intersection over Union (IoU), and Hausdorff distance (HD). The Dice coefficient measures the area overlap between the segmentation result and the true value. The IoU measures the similarity between the predicted and ground truth regions. HD assesses the maximum surface distance, in particular, the maximum distance from one point in one set to the nearest point in the other set.
Dice is the most frequently used metric in image segmentation tasks, and it is a set similarity metric that is usually used to calculate the similarity of two samples.
D i c e = 2 | A B | A B
where A is the set of predicted values, B is the set of true values, and the numerator is the intersection between A and B.
The IoU score is a standard performance measure for object category segmentation problems. Given a set of images, the IoU measurement gives the similarity between the predicted and ground truth regions of objects present in the set of images and is defined by the following equation:
I o U = T P F P + T P + F N
where TP, FP, and FN represent true positive, false positive, and false negative counts, respectively.
The Hausdorff distance is a measure to describe the degree of similarity between two sets of points. It is a definition of the distance between two sets of points. Suppose there are two sets of sets A and B, then the Hausdorff distance between these two sets of points is defined as:
H D ( A , B ) = m a x ( h ( A , B ) , h ( B , A ) )
h ( A , B ) = m a x a A { m i n b B a b }
h ( B , A ) = m a x b B { m i n a A b a }

2.6. Implementation Details

The method was implemented using PyTorch, and all models were trained using the ADAM [53] optimizer. Before the images were fed into the networks, random augmentations were applied to the images, including random contrast adjustment and the CLAHE method. During training, the student and teacher models save their weights every five epochs, and finally, the optimal weights of the models are selected according to the validation loss.

3. Results

3.1. Quantitative Evaluation

Our method uses three-fold cross-validation, with the data from each data center serving as one fold. Only data from a single data center were used for each training, and data from the other two data centers were used as the validation and test sets, respectively. The experimental settings used in this study are described in Table 2.
We compared our method with the fully supervised and semi-supervised MT methods. In addition to DICE, IoU, and HD metrics, we adopted the overall unacceptable segmentation ratio (OUSR) to measure the performance of the three methods. “Unacceptable” segmentation was defined as a segmentation result with a DICE value less than 85% and as unacceptable as evaluated by a senior physician. The underlying segmentation networks for the three methods were all U-Net with the same architecture. A comparison of these results is presented in Table 3.
The experimental results show that our method generalizes well to the test set and outperforms other methods in all evaluation metrics. As shown in Table 3, the mean DICE exceeded 0.94 for all three groups of experiments for the fully supervised method. This suggests that a higher proportion of mammograms in our randomly sampled dataset had pectoral muscles that were easily identified, which is consistent with experience in clinical practice. These mammograms are usually taken in strict accordance with mammography standards so that a clear view of the entire pectoral boundary can be seen. However, the fully supervised method in Table 3 achieved an OUSR of 55 / 635 8 % , which indicates that there is still a certain percentage of mammograms in the dataset with unsatisfactory segmentation results. Possible reasons for this situation are insufficient training data, low percentage of mammograms with blurred pectoral boundaries, and degraded model performance caused by switching the data center of the test set.
The MT method additionally uses the data of the validation set for semi-supervised training, which provides additional constraints for the model through the consistency regularization term and alleviates the problem of insufficient data volume. At the same time, the model can obtain domain-related information from the validation set data, which increases the generalization ability of the model on the test set. Table 3 shows that the MT method reduces the overall number of unacceptable segmentations compared with the fully supervised method. However, the MT method is affected by the quality of the prediction of unlabeled data, which is greatly affected by the confirmation bias. Specifically, the MT method was better than the fully supervised method in individual indicators, but overall, the performance of the MT method in the three groups of experiments was not significantly different from that of the fully supervised method.
Our uncertainty-aware semi-supervised method is a consistency regularization method based on the MT method; therefore, it inherits the advantages of the MT method. Meanwhile, our method uses uncertainty to measure the confidence of the predictions and refines the low-confidence predictions using learned prior knowledge, thereby reducing the impact of confirmation bias. Table 3 shows that the proposed method achieved optimal results in all three groups of experiments. It can be seen that the method proposed in this study effectively improves the model generalization performance. In terms of OUSR, our method reduces this metric by more than 50% compared to fully supervised methods. It can be seen that our method performs well in improving the low-confidence prediction results.

3.2. Qualitative Evaluation

When performing pectoral muscle delineation, the doctor first delineates the part of the pectoral muscle with clearly visible boundaries and then estimates the parts that are difficult to identify in the mammogram by incorporating a priori knowledge of the anatomical structure. The degree of clarity originates from the intuition of human vision, and to draw on the doctor’s experience, we used high- and low-confidence regions as the estimation of clearly visible and blurred regions, and we calculated the high- and low-confidence regions using uncertainty. In our experiments, we found that such an estimation can match intuitive judgments. Figure 4 shows the results for some samples in the dataset using confidence regions as an estimation of the degree of clarity.

4. Discussion

Deep learning methods typically assume that the data in the training and test sets follow the same distribution; however, for applications related to medical imaging, this cannot be assumed. One reason for the variation in the data distribution for medical imaging is that imaging devices in different data centers often have different built-in photography parameters and processing algorithms. In our experiments, we collected data from different imaging devices at three data centers to simulate real-world scenarios. The training and test sets used in each of the three groups of experiments are from different data centers, which requires our model to have good generalization ability to the test set. Our proposed method improves the generalization ability of the model on the test set in the following ways:
  • Consistency regularization constraints are constructed using unlabeled data from the validation set to participate in model training. In deep learning models, it is often assumed that the training set and the dataset have the same data distribution, an assumption that is often difficult to satisfy in practice, especially because medical imaging data often contain information related to the data domain introduced by specific device parameters. If the loss function contains only information relevant to the training set, the performance of the model on the test set largely depends on its similarity to the training set data. By adding a consistency regularization term constructed from the validation set data to the loss function, the output of the model on different datasets is constrained by the consistency regularization. The results in Table 3 show that both the MT method and our proposed method reduce the OUSR in the test set in this way.
  • Leveraging domain-related information contained in the low-confidence prediction region. Recent studies on domain adaptation [42,43] show that models tend to produce low-confidence predictions when the target and source data domains differ. This suggests that low-confidence predictions may be caused by the presence of domain-related information in the target data domain that has not been learned by the model. Directly constructing consistency regularization terms using these low-confidence predictions can cause the model to suffer from confirmation bias. Meanwhile, using the strategy of discarding low-confidence predictions, as in recent studies, loses the opportunity to fully learn the domain-relevant information from the dataset. We draw on the experience of doctors when performing delineation and refine the low-confidence predictions using high-confidence predictions as well as the learned prior to enable the model to fully learn the domain-relevant information from the dataset. In the experiments, our method outperformed the other two methods in all evaluation metrics.
  • The utilization of prior knowledge reduces the dependence of target predictions on the teacher predictions and improves the accuracy and stability of model training. The GAN used in the low-confidence predictions refinement module learns the anatomical prior of the breast and pectoral muscle in mammograms, which provides regularization constraints for the target predictions used in the student model training. Therefore, fluctuations in the performance of the teacher model will not have a large impact on the target predictions. Moreover, GAN uses domain-independent semantic features as input, and its main reliance on the teacher model is the high-confidence predictions, which are usually easy to learn. Thus, the low-confidence prediction refinement module can provide regularization constraints for target predictions that do not vary by the data center.
GAN was used in our method as the model for low-confidence prediction refinement, and the loss function of the generator includes a consistency loss and adversarial loss, which can be considered the fidelity and regularity terms, respectively. The consistency loss constrained the generator’s predictions to be consistent with the high-confidence teacher predictions, while the adversarial loss constrained the generator’s predictions to conform to the shape prior given by the annotation. Since GAN’s adversarial loss can use unpaired data, each prediction of the generator is subject to regularization from the entire dataset, preventing the model from directly memorizing the annotations corresponding to the inputs.
Our method draws on doctors’ experience, takes the confident part of the pectoral muscle prediction and shape of the breast as input, and uses a generative adversarial network (GAN) to learn to use this information to refine the low-confidence predictions of the pectoral muscle. Since the dataset distribution of different hospital centers is very different, our method can effectively alleviate this problem and improve the generalization ability of the model. Compared with the baseline method, our method can better mine the latent feature structure in the low-confidence prediction data, thereby improving the segmentation ability of the model. We draw on doctors’ experience in delineation by learning the prior knowledge obtained from annotation and refining the low-confidence predictions. This allows our method to alleviate the confirmation bias problem while providing access to domain-relevant information. The experimental results demonstrate that the proposed method improves the performance of the model on the test set.
The low-confidence prediction refinement module uses the breast segmentation results as an input feature because both the breast border and pectoral muscle are important localization benchmarks and can be used as shape constraints for each other. In addition, coarse segmentation of the breast is usually a relatively easy task compared to segmentation of the pectoral muscle. This provides an idea for other medical image segmentation tasks, where additional regularization constraints can be provided for the current task using simple tasks related to the current task.

5. Conclusions

In this study, we proposed an uncertainty-aware semi-supervised segmentation method that incorporates a breast anatomical prior for pectoral muscle segmentation in mammograms. The method uses uncertainty to measure the confidence of predictions and refines the low-confidence predictions by learning task-relevant prior knowledge through a GAN. This improves the quality of the target predictions while allowing the model to learn domain-relevant features in the dataset more efficiently, thereby improving the generalization ability of the model. The experimental results show that compared with the baseline method, the proposed method has an average increase of 1.76 in the DICE index, an average decrease of 3.21 in the IoU index, and an average decrease of 5.48 in the HD index. The experimental results show that our method improves the model performance in pectoral muscle segmentation, reduces unacceptable segmentation results, and provides good generalizability.
The amount of data used in the experiment is limited, and whether the model can be extended to clinical application needs further experimental exploration. A limitation of this study is the lack of testing on a public dataset. However, publicly available datasets generally lack a gold standard for pectoral muscle delineation as well as detailed information on scan data, which is critical for model training. Another limitation of our study is that we did not quantify the low-confidence predictions. The results of low-confidence prediction were all optimized in the low-confidence prediction refinement module. No distinction was made between the confidence levels of the data predictions, so the model may miss information about the features predicted with high confidence. Therefore, distinguishing low-confidence prediction results is still a problem worthy of further study. At present, this study is only implemented based on 635 data from three hospitals. In the future, we plan to generalize it to larger datasets and to more datasets from different hospitals. At the same time, the low-confidence prediction was further quantified, and the method of using low-confidence prediction to enhance the learning ability of the model was explored. Low-confidence prediction is not only a limitation of current models, but also a clinical problem that doctors need to solve urgently. In the future, we will continue to explore new ways to leverage low-confidence prediction data.

Author Contributions

Conceptualization, T.S. and Y.L.; methodology, Y.G. and Y.T.; software, Y.G. and H.W.; validation, Y.G. and Y.T.; formal analysis, T.S. and Y.T.; investigation, Y.T.; resources, T.S. and Y.L.; writing—original draft preparation, Y.G.; writing—review and editing, Y.T. and Y.L.; visualization, Y.T.; supervision, Y.L.; project administration, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the China Department of Science and Technology under Key Grant 2023YFE0204300, in part by the R&D project of Pazhou Lab (HuangPu) under Grant 2023K0606, in part by the NSFC under Grant 82441027, Grant 62371476, in part by the Guangzhou Science and Technology bureau under Grant 2023B03J1237, in part by the Health Research Major Projects of Hunan Health Commission under grant W20241010, in part by the Guangdong Province Key Laboratory of Computational Science at the Sun Yat-sen University under grant 2020B1212060032, in part by the Key-Area Research and Development Program of Guangdong Province under Grant 2021B0101190003.

Institutional Review Board Statement

The study was conducted in accordance with the provisions of the Declaration of Helsinki and was approved by the Medical Ethics Committee of Sun Yat-sen Memorial Hospital, Sun Yat-sen University (protocol code SYSEC-KY-KS-2023-075; approval date, 1 March 2023) for studies involving humans.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patients to publish this paper.

Data Availability Statement

Data cannot be shared publicly due to privacy protection of the participants and ethical restrictions. For researchers interested in the data, requests can be made to the corresponding authors luyao23@mail.sysu.edu.cn. Researchers with questions about the model code covered in the article may also make a request to the corresponding author luyao23@mail.sysu.edu.cn.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Kwok, S.; Chandrasekhar, R.; Attikiouzel, Y.; Rickard, M. Automatic Pectoral Muscle Segmentation on Mediolateral Oblique View Mammograms. IEEE Trans. Med. Imaging 2004, 23, 1129–1140. [Google Scholar] [CrossRef]
  2. Brandt, S.S.; Karemore, G.; Karssemeijer, N.; Nielsen, M. An Anatomically Oriented Breast Coordinate System for Mammogram Analysis. IEEE Trans. Med. Imaging 2011, 30, 1841–1851. [Google Scholar] [CrossRef]
  3. Javier Perez-Benito, F.; Signol, F.; Perez-Cortes, J.C.; Fuster-Baggetto, A.; Pollan, M.; Perez-Gomez, B.; Salas-Trejo, D.; Casals, M.; Martinez, I.; LLobet, R. A Deep Learning System to Obtain the Optimal Parameters for a Threshold-Based Breast and Dense Tissue Segmentation. Comput. Methods Programs Biomed. 2020, 195, 105668. [Google Scholar] [CrossRef]
  4. Maghsoudi, O.H.; Gastounioti, A.; Scott, C.; Pantalone, L.; Wu, F.F.; Cohen, E.A.; Winham, S.; Conant, E.F.; Vachon, C.; Kontos, D. Deep-LIBRA: An Artificial-Intelligence Method for Robust Quantification of Breast Density with Independent Validation in Breast Cancer Risk Assessment. Med. Image Anal. 2021, 73, 102138. [Google Scholar] [CrossRef]
  5. Karssemeijer, N. Automated Classification of Parenchymal Patterns in Mammograms. Phys. Med. Biol. 1998, 43, 365–378. [Google Scholar] [CrossRef]
  6. Shapiro, L.; Stockman, G. Computer Vision; Prentice Hall: Saddle River, NJ, USA, 2001. [Google Scholar]
  7. Ferrari, R.; Rangayyan, R.; Desautels, J.; Borges, R.; Frere, A. Automatic Identification of the Pectoral Muscle in Mammograms. IEEE Trans. Med. Imaging 2004, 23, 232–245. [Google Scholar] [CrossRef] [PubMed]
  8. Ma, W.Y.; Manjunath, B.S. EdgeFlow: A Technique for Boundary Detection and Image Segmentation. IEEE Trans. Image Process. 2000, 9, 1375–1388. [Google Scholar]
  9. Ma, F.; Bajger, M.; Slavotinek, J.P.; Bottema, M.J. Two Graph Theory Based Methods for Identifying the Pectoral Muscle in Mammograms. Pattern Recognit. 2007, 40, 2592–2602. [Google Scholar] [CrossRef]
  10. Subashini, T.S.; Ramalingam, V.; Palanivel, S. Automated Assessment of Breast Tissue Density in Digital Mammograms. Comput. Vis. Image Underst. 2010, 114, 33–43. [Google Scholar] [CrossRef]
  11. Liu, C.C.; Tsai, C.Y.; Liu, J.; Yu, C.Y.; Yu, S.S. A Pectoral Muscle Segmentation Algorithm for Digital Mammograms Using Otsu Thresholding and Multiple Regression Analysis. Comput. Math. Appl. 2012, 64, 1100–1107. [Google Scholar] [CrossRef]
  12. Mustra, M.; Grgic, M. Robust Automatic Breast and Pectoral Muscle Segmentation from Scanned Mammograms. Signal Process. 2013, 93, 2817–2827. [Google Scholar] [CrossRef]
  13. Kurt, B.; Nabiyev, V.V.; Turhan, K. A Novel Automatic Suspicious Mass Regions Identification Using Havrda & Charvat Entropy and Otsu’s N Thresholding. Comput. Methods Programs Biomed. 2014, 114, 349–360. [Google Scholar] [CrossRef]
  14. Mahersia, H.; Boulehmi, H.; Hamrouni, K. Development of Intelligent Systems Based on Bayesian Regularization Network and Neuro-Fuzzy Models for Mass Detection in Mammograms: A Comparative Analysis. Comput. Methods Programs Biomed. 2016, 126, 46–62. [Google Scholar] [CrossRef] [PubMed]
  15. Shen, R.; Yan, K.; Xiao, F.; Chang, J.; Jiang, C.; Zhou, K. Automatic Pectoral Muscle Region Segmentation in Mammograms Using Genetic Algorithm and Morphological Selection. J. Digit. Imaging 2018, 31, 680–691. [Google Scholar] [CrossRef]
  16. Moayedi, F.; Azimifar, Z.; Boostani, R.; Katebi, S. Contourlet-Based Mammography Mass Classification Using the SVM Family. Comput. Biol. Med. 2010, 40, 373–383. [Google Scholar] [CrossRef]
  17. Maitra, I.K.; Nag, S.; Bandyopadhyay, S.K. Technique for Preprocessing of Digital Mammogram. Comput. Methods Programs Biomed. 2012, 107, 175–188. [Google Scholar] [CrossRef]
  18. Guo, Y.; Dong, M.; Yang, Z.; Gao, X.; Wang, K.; Luo, C.; Ma, Y.; Zhang, J. A New Method of Detecting Micro-Calcification Clusters in Mammograms Using Contourlet Transform and Non-Linking Simplified PCNN. Comput. Methods Programs Biomed. 2016, 130, 31–45. [Google Scholar] [CrossRef]
  19. Peng, W.; Mayorga, R.V.; Hussein, E.M.A. An Automated Confirmatory System for Analysis of Mammograms. Comput. Methods Programs Biomed. 2016, 125, 134–144. [Google Scholar] [CrossRef]
  20. Feudjio, C.K.; Klein, J.; Tiedeu, A.; Colot, O. Automatic Extraction of Pectoral Muscle in the MLO View of Mammograms. Phys. Med. Biol. 2013, 58, 8493–8515. [Google Scholar] [CrossRef]
  21. Shi, P.; Zhong, J.; Rampun, A.; Wang, H. A Hierarchical Pipeline for Breast Boundary Segmentation and Calcification Detection in Mammograms. Comput. Biol. Med. 2018, 96, 178–188. [Google Scholar] [CrossRef] [PubMed]
  22. Camilus, K.S.; Govindan, V.K.; Sathidevi, P.S. Computer-Aided Identification of the Pectoral Muscle in Digitized Mammograms. J. Digit. Imaging 2010, 23, 562–580. [Google Scholar] [CrossRef]
  23. Zhou, C.; Wei, J.; Chan, H.P.; Paramagul, C.; Hadjiiski, L.M.; Sahiner, B.; Douglas, J.A. Computerized Image Analysis: Texture-field Orientation Method for Pectoral Muscle Identification on MLO-view Mammograms. Med. Phys. 2010, 37, 2289–2299. [Google Scholar] [CrossRef] [PubMed]
  24. Tzikopoulos, S.D.; Mavroforakis, M.E.; Georgiou, H.V.; Dimitropoulos, N.; Theodoridis, S. A Fully Automated Scheme for Mammographic Segmentation and Classification Based on Breast Density and Asymmetry. Comput. Methods Programs Biomed. 2011, 102, 47–63. [Google Scholar] [CrossRef] [PubMed]
  25. Keller, B.M.; Nathan, D.L.; Wang, Y.; Zheng, Y.; Gee, J.C.; Conant, E.F.; Kontos, D. Estimation of Breast Percent Density in Raw and Processed Full Field Digital Mammography Images via Adaptive Fuzzy C-Means Clustering and Support Vector Machine Segmentation. Med. Phys. 2012, 39, 4903–4917. [Google Scholar] [CrossRef]
  26. Chakraborty, J.; Mukhopadhyay, S.; Singla, V.; Khandelwal, N.; Bhattacharyya, P. Automatic Detection of Pectoral Muscle Using Average Gradient and Shape Based Feature. J. Digit. Imaging 2012, 25, 387–399. [Google Scholar] [CrossRef]
  27. Bora, V.B.; Kothari, A.G.; Keskar, A.G. Robust Automatic Pectoral Muscle Segmentation from Mammograms Using Texture Gradient and Euclidean Distance Regression. J. Digit. Imaging 2016, 29, 115–125. [Google Scholar] [CrossRef] [PubMed]
  28. Xie, W.; Li, Y.; Ma, Y. Breast Mass Classification in Digital Mammography Based on Extreme Learning Machine. Neurocomputing 2016, 173, 930–941. [Google Scholar] [CrossRef]
  29. UrviOza; Gohel, B.; Kumar, P. Evaluation of Normalization Algorithms for Breast Mammogram Mass Segmentation. Procedia Comput. Sci. 2024, 235, 2508–2517. [Google Scholar] [CrossRef]
  30. Ghoushchi, S.J.; Ranjbarzadeh, R.; Najafabadi, S.A.; Osgooei, E.; Tirkolaee, E.B. An extended approach to the diagnosis of tumour location in breast cancer using deep learning. J. Ambient. Intell. Humaniz. Comput. 2023, 14, 8487–8497. [Google Scholar] [CrossRef]
  31. Eklund, G.; Cardenosa, G.; Parsons, W. Assessing Adequacy of Mammographic Image Quality. Radiology 1994, 190, 297–307. [Google Scholar] [CrossRef]
  32. Ma, X.; Wei, J.; Zhou, C.; Helvie, M.A.; Chan, H.P.; Hadjiiski, L.M.; Lu, Y. Automated Pectoral Muscle Identification on MLO-view Mammograms: Comparison of Deep Neural Network to Conventional Computer Vision. Med. Phys. 2019, 46, 2103–2114. [Google Scholar] [CrossRef]
  33. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—(MICCAI) 2015, 18th International Conference, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Lecture Notes in Computer Science. Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  34. Rampun, A.; Lopez-Linares, K.; Morrow, P.J.; Scotney, B.W.; Wang, H.; Garcia Ocana, I.; Maclair, G.; Zwiggelaar, R.; Gonzalez Ballester, M.A.; Macia, I. Breast Pectoral Muscle Segmentation in Mammograms Using a Modified Holistically-Nested Edge Detection Network. Med. Image Anal. 2019, 57, 1–17. [Google Scholar] [CrossRef] [PubMed]
  35. Xie, S.; Tu, Z. Holistically-Nested Edge Detection. Int. J. Comput. Vis. 2017, 125, 3–18. [Google Scholar] [CrossRef]
  36. Guo, Y.; Zhao, W.; Li, S.; Zhang, Y.; Lu, Y. Automatic Segmentation of the Pectoral Muscle Based on Boundary Identification and Shape Prediction. Phys. Med. Biol. 2020, 65, 045016. [Google Scholar] [CrossRef] [PubMed]
  37. Yang, X.; Song, Z.; King, I.; Xu, Z. A Survey on Deep Semi-supervised Learning. arXiv 2021, arXiv:2103.00550. [Google Scholar] [CrossRef]
  38. Tarvainen, A.; Valpola, H. Mean Teachers Are Better Role Models: Weight-averaged Consistency Targets Improve Semi-Supervised Deep Learning Results. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  39. Yu, L.; Wang, S.; Li, X.; Fu, C.W.; Heng, P.A. Uncertainty-Aware Self-ensembling Model for Semi-supervised 3D Left Atrium Segmentation. arXiv 2019, arXiv:1907.07034. [Google Scholar]
  40. Zhang, Y.; Liao, Q.; Jiao, R.; Zhang, J. Uncertainty-Guided Mutual Consistency Learning for Semi-Supervised Medical Image Segmentation. arXiv 2021, arXiv:2112.02508. [Google Scholar] [CrossRef]
  41. Lai, X.; Tian, Z.; Jiang, L.; Liu, S.; Zhao, H.; Wang, L.; Jia, J. Semi-Supervised Semantic Segmentation with Directional Context-aware Consistency. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR, Nashville, TN, USA, 20–25 June 2021; pp. 1205–1214. [Google Scholar] [CrossRef]
  42. Long, M.; Cao, Y.; Cao, Z.; Wang, J.; Jordan, M. Transferable Representation Learning with Deep Adaptation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 3071–3085. [Google Scholar] [CrossRef] [PubMed]
  43. Vu, T.H.; Jain, H.; Bucher, M.; Cord, M.; Perez, P. ADVENT: Adversarial Entropy Minimization for Domain Adaptation in Semantic Segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2019), Long Beach, CA, USA, 15–20 June 2019; pp. 2512–2521. [Google Scholar] [CrossRef]
  44. Meng, Y.; Zhang, H.; Zhao, Y.; Yang, X.; Qian, X.; Huang, X.; Zheng, Y. Spatial Uncertainty-Aware Semi-Supervised Crowd Counting. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 15529–15539. [Google Scholar] [CrossRef]
  45. Brock, A.; Donahue, J.; Simonyan, K. Large Scale GAN Training for High Fidelity Natural Image Synthesis. In Proceedings of the International Conference on Learning Representations, Virtual, 25–29 April 2022. [Google Scholar]
  46. Karras, T.; Laine, S.; Aila, T. A Style-Based Generator Architecture for Generative Adversarial Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2019), Los Alamitos, CA, USA, 15–20 June 2019; pp. 4396–4405. [Google Scholar] [CrossRef]
  47. Xu, A.; Wang, S.; Fan, J.; Shi, X.; Chen, Q. Dual Attention Based Uncertainty-aware Mean Teacher Model for Semi-supervised Cardiac Image Segmentation. In Proceedings of the 2021 IEEE International Conference on Progress in Informatics and Computing (PIC), Shanghai, China, 17–19 December 2021; pp. 82–86. [Google Scholar] [CrossRef]
  48. Kendall, A.; Gal, Y. What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? In Proceedings of the Advances in Neural Information Processing Systems 30 (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  49. Weickert, J. Coherence-Enhancing Shock Filters. In Pattern Recognition, Proceedings; Michaelis, B., Krell, G., Eds.; Springer: Berlin, Germany, 2003; Volume 2781, pp. 1–8. [Google Scholar]
  50. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Curran Associates, Inc.: Red Hook, NY, USA, 2014; Volume 27. [Google Scholar]
  51. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein Generative Adversarial Networks. In Proceedings of the 34th International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  52. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved Training of Wasserstein GANs. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  53. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arxiv 2017, arXiv:1412.6980. [Google Scholar] [CrossRef]
Figure 1. Overall framework of proposed method. The student model is trained by a supervised loss (blue lines) while being subjected to the consistency regularization constraint (green lines). The target predictions consist of high-confidence and low-confidence predictions. The high-confidence predictions come directly from the teacher predictions in the high-confidence region. The low-confidence predictions are refined by the low-confidence prediction refinement module, whose input consists of the high-confidence predictions, confidence region indicator, and breast segmentation result.
Figure 1. Overall framework of proposed method. The student model is trained by a supervised loss (blue lines) while being subjected to the consistency regularization constraint (green lines). The target predictions consist of high-confidence and low-confidence predictions. The high-confidence predictions come directly from the teacher predictions in the high-confidence region. The low-confidence predictions are refined by the low-confidence prediction refinement module, whose input consists of the high-confidence predictions, confidence region indicator, and breast segmentation result.
Bioengineering 12 00036 g001
Figure 2. Uncertainty estimation method. The mammogram to be segmented is shown in (A), and the fields of view in (BH) correspond to the red box in (A). (B) is the pixel-independent uncertainty response obtained by the uncertainty estimation module. Using the top hat to remove the structures with larger widths yields (C). Next, the CESF method is applied to (C) to obtain an enhanced response (E). Then, thresholding and maximum connected region acquisition are used to find a binarized boundary, as in (F). (G) is the high-confidence region indicator, and the high-confidence teacher prediction is obtained by multiplication of (G) and the teacher prediction (D).
Figure 2. Uncertainty estimation method. The mammogram to be segmented is shown in (A), and the fields of view in (BH) correspond to the red box in (A). (B) is the pixel-independent uncertainty response obtained by the uncertainty estimation module. Using the top hat to remove the structures with larger widths yields (C). Next, the CESF method is applied to (C) to obtain an enhanced response (E). Then, thresholding and maximum connected region acquisition are used to find a binarized boundary, as in (F). (G) is the high-confidence region indicator, and the high-confidence teacher prediction is obtained by multiplication of (G) and the teacher prediction (D).
Bioengineering 12 00036 g002
Figure 3. Training flow of the low-confidence prediction refinement module.
Figure 3. Training flow of the low-confidence prediction refinement module.
Bioengineering 12 00036 g003
Figure 4. Confidence region as an estimation of the degree of clarity. The images in the first row are target mammograms. The images in the second row are the pixel-independent uncertainty responses. The red and blue regions in the third row represent the high-confidence and low-confidence regions, respectively.
Figure 4. Confidence region as an estimation of the degree of clarity. The images in the first row are target mammograms. The images in the second row are the pixel-independent uncertainty responses. The red and blue regions in the third row represent the high-confidence and low-confidence regions, respectively.
Bioengineering 12 00036 g004
Table 1. Specifications of the dataset used in our study.
Table 1. Specifications of the dataset used in our study.
Data CenterNumberSizeVendor
DC1 a185 2800 × 2082 SIEMENS
DC2 b236 2812 × 2082 PLANMED NUANCE
DC3 c184 2457 × 1890 HOLOGIC
a The First Affiliated Hospital of Sun Yat-sen University; b The Sun Yat-sen Memorial Hospital of Sun Yat-sen University; c The NanFang Hospital.
Table 2. Experimental settings.
Table 2. Experimental settings.
Experiments GroupTrainng SetValidation SetTest Set
Exp1DC1DC2DC3
Exp2DC2DC3DC1
Exp3DC3DC1DC2
Table 3. Experimental results.
Table 3. Experimental results.
Method
Measure
Exp1Exp2Exp3Overall
OUSR ↓
DICE (↑)IoU (↑)HD (↓)DICEIoUHDDICEIoUHD
Supervised 95.45 91.30 16.43 94.35 89.30 18.75 93.54 87.86 21.92 55/635
MT 95.36 91.13 15.48 94.19 89.02 19.24 95.95 92.22 14.54 45/635
Ours96.1692.6013.3995.8692.0514.6296.6193.4412.6623/635
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, Y.; Guo, Y.; Wang, H.; Song, T.; Lu, Y. Uncertainty-Aware Semi-Supervised Method for Pectoral Muscle Segmentation. Bioengineering 2025, 12, 36. https://doi.org/10.3390/bioengineering12010036

AMA Style

Tang Y, Guo Y, Wang H, Song T, Lu Y. Uncertainty-Aware Semi-Supervised Method for Pectoral Muscle Segmentation. Bioengineering. 2025; 12(1):36. https://doi.org/10.3390/bioengineering12010036

Chicago/Turabian Style

Tang, Yutao, Yongze Guo, Huayu Wang, Ting Song, and Yao Lu. 2025. "Uncertainty-Aware Semi-Supervised Method for Pectoral Muscle Segmentation" Bioengineering 12, no. 1: 36. https://doi.org/10.3390/bioengineering12010036

APA Style

Tang, Y., Guo, Y., Wang, H., Song, T., & Lu, Y. (2025). Uncertainty-Aware Semi-Supervised Method for Pectoral Muscle Segmentation. Bioengineering, 12(1), 36. https://doi.org/10.3390/bioengineering12010036

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop