Next Article in Journal
Impact of Optimal Feature Selection Using Hybrid Method for a Multiclass Problem in Cross Project Defect Prediction
Next Article in Special Issue
Applied Sciences—Special Issue on Emerging Techniques in Imaging, Modelling and Visualization for Cardiovascular Diagnosis and Therapy
Previous Article in Journal
Ecosystem Services Assessment for Their Integration in the Analysis of Landslide Risk
Previous Article in Special Issue
MR Imaging and Electrophysiological Features of Doxorubicin-Induced Fibrosis: Protocol Development in a Small Preclinical Pig Study with Histological Validation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Deep Representations of Cardiac Structures for 4D Cine MRI Image Segmentation through Semi-Supervised Learning

by
S. M. Kamrul Hasan
1,* and
Cristian A. Linte
1,2
1
Center for Imaging Science, Rochester Institute of Technology, Rochester, NY 14623, USA
2
Department of Biomedical Engineering, Rochester Institute of Technology, Rochester, NY 14623, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(23), 12163; https://doi.org/10.3390/app122312163
Submission received: 7 October 2022 / Revised: 17 November 2022 / Accepted: 22 November 2022 / Published: 28 November 2022

Abstract

:
Learning good data representations for medical imaging tasks ensures the preservation of relevant information and the removal of irrelevant information from the data to improve the interpretability of the learned features. In this paper, we propose a semi-supervised model—namely, combine-all in semi-supervised learning (CqSL)—to demonstrate the power of a simple combination of a disentanglement block, variational autoencoder (VAE), generative adversarial network (GAN), and a conditioning layer-based reconstructor for performing two important tasks in medical imaging: segmentation and reconstruction. Our work is motivated by the recent progress in image segmentation using semi-supervised learning (SSL), which has shown good results with limited labeled data and large amounts of unlabeled data. A disentanglement block decomposes an input image into a domain-invariant spatial factor and a domain-specific non-spatial factor. We assume that medical images acquired using multiple scanners (different domain information) share a common spatial space but differ in non-spatial space (intensities, contrast, etc.). Hence, we utilize our spatial information to generate segmentation masks from unlabeled datasets using a generative adversarial network (GAN). Finally, to reconstruct the original image, our conditioning layer-based reconstruction block recombines spatial information with random non-spatial information sampled from the generative models. Our ablation study demonstrates the benefits of disentanglement in holding domain-invariant (spatial) as well as domain-specific (non-spatial) information with high accuracy. We further apply a structured L 2 similarity (S L 2 SIM) loss along with a mutual information minimizer (MIM) to improve the adversarially trained generative models for better reconstruction. Experimental results achieved on the STACOM 2017 ACDC cine cardiac magnetic resonance (MR) dataset suggest that our proposed (CqSL) model outperforms fully supervised and semi-supervised models, achieving an 83.2% performance accuracy even when using only 1% labeled data. We hypothesize that our proposed model has the potential to become an efficient semantic segmentation tool that may be used for domain adaptation in data-limited medical imaging scenarios, where annotations are expensive. Code, and experimental configurations will be made available publicly.

1. Introduction

1.1. Background and Problem Statement

The emerging success of deep convolutional neural networks (CNNs) has rendered them the de facto model in solving high-level computer vision tasks [1,2,3]. However, such approaches mostly rely on large amounts of annotated data for training, the acquisition of which is expensive and laborious, especially for medical imaging/diagnostic radiology data. To address the need for high performance, there has been a growing trend in using a limited amount of annotated data along with an abundance of unlabeled data in a semi-supervised learning (SSL) setting.
The recent dominant body of research that has proposed SSL methods in deep learning features various approaches, including an auxiliary loss term defined on un-annotated data (consistency regularization) [4,5], adversarial networks [6], generating pseudo-labels [7,8] based on model predictions on weakly augmented un-annotated data, self-training [9,10], adversarial learning [11] and domain adaptation [12]. Here we acknowledge their latest accomplishments in the field of domain adaptation, semi-supervised learning and interpretable representation learning by disentanglement and briefly discuss some of their yet outstanding limitations.

1.2. Ongoing Efforts and Related Work

Semi-Supervised Learning: Semi-supervised learning (SSL) [13,14] has experienced much research attention thanks to the increasing availability of large-scale unlabeled data. Semi-supervised learning aims to revamp the model performance by learning from a small portion of labeled data along with optimizing an additional unsupervised loss on a larger portion of unlabeled data, assumed to be sampled from similar distributions, depending on the type of information that needs to be captured from the unlabeled data. Commonly, the rationale of SSL is based on generative models and adversarial networks. The integration of consistency regularization in SSL has shed light on standard baselines recently. By optimizing this loss term, the model imposes several assumptions/constraints on the decision boundary to avoid high-density regions of unannotated data.
Generative adversarial networks: Moreover, generative adversarial learning can be adapted to semi-supervised learning for semantic segmentation [15,16,17] as well as by generating pseudo pixel-level predictions [18,19]. Adversarial networks use a critic to predict the pixel-level distribution of the data, which acts as an adversarial loss term with the goal to provide the generator with learnable useful visual features from the unlabeled data for medical image synthesis [20]. Nonetheless, learning high-dimensional data can be difficult. Autoencoders struggle with multi-modal data distributions, and generative models rely on computationally demanding models, which are especially difficult to train.
Mutual information estimation: Recent work on representation learning has focused on mutual information estimation [21]. As mutual information maximization has been shown to be effective at capturing the salient attributes of data, being able to disentangle these attributes is another desirable property. For example, it may be beneficial to remove data attributes that are irrelevant to a given task, such as illumination conditions in object recognition.
Disentanglement learning: Some newly introduced techniques have dedicated considerable attention to disentangle representation with generative modeling [22,23]. In disentangled representation, information is represented as a collection of (independent) factors [24], each of which corresponds to a meaningful aspect of the data [25,26]. A current line of research has argued that disentangled representations are beneficial for a variety of tasks, including (semi-)supervised learning of downstream tasks, few-shot learning [27], and exploratory medical data analysis. Additionally, these representations also make it easier for later processes to only use the relevant parts of the data as input.
Unpaired image-to-image translation: Image-to-image translation was first proposed by Isola et al. in [28] in their conditional GAN paper. Furthermore, CycleGAN [29] tackles the problem of the above paired image translation approach by introducing a cycle-consistency loss to retrieve the original images by exploiting a cycle of translation. Later work [30] improved CycleGAN from one-to-one mapping to multimodal image generation. Nevertheless, in medical applications, image synthesis without explicit anatomy design constrain may lead to volatile anatomical structures and artifacts. Moreover, these methods are not aimed at medical image segmentation.
Domain Adaptation: Domain adaptation, a form of transfer learning, encodes the distribution knowledge from a certain source domain to a different but related target domain, and thus, alleviates the domain shift discrepancy in real world applications [31]. Various methods have been proposed, including style and content-disentanglement [32], and adversary based approaches [33,34]. As described later, in this work, we disentangle the most interpretable segmentation-aware spatial (skeleton) information.
Normalization layers: Inspired by instance normalization (IN) [35], conditional batch-normalization [36] and adaptive IN (AdaIN) [37] bring significant improvement in image generation. Later on, feature-wise linear modulation (FiLM) [38] and spatially adaptive denormalization (SPADE) [39] shed additional light over other normalization layers in image synthesis. In our proposed work, we also show how we can adapt both SPADE as well as FiLM normalization as part of a residual and common decoder, respectively.
Variational autoencoder-based models: There have been several recent works involving disentangled learning with variational autoencoder (VAE) [24,40,41]. In contrast to these previous works, we will attempt to demonstrate the use of a VAE as a disentangled representation by sampling the sentiency code to separate the domain-specific information from the domain-invariant latent code.

1.3. Overview of the Proposed Method

To further address some of the shortcomings associated with existing methods, our efforts focus on learning meaningful spatial features utilizing a disentangler with a mutual information minimizer (MIM) to improve the adversarially trained generative models for improving semi-supervised segmentation and reconstruction results.
Our proposed method builds on several recent and key research findings in the fields of generative models, semi-supervised learning, and representation learning via disentanglement. We believe that the proposed framework’s reliance on as little as 1% labeled data for training, in concert with the high segmentation accuracy achieved, comparable to the fully or semi-supervised models, renders the proposed work an attractive solution for medical image segmentation, where access to vast expert-annotated data is expensive and often difficult to gain access to.
We approach this problem using a method that is based on disentangled representations and utilizes data from multiple scanners with varying intensities and contrast (Figure 1). Our method is intended to address multi-scanner unlabeled-data issues, such as intensity differences, and a lack of sufficient annotated data. Learning good data representations for medical imaging tasks ensures the preservation of relevant information and the removal of irrelevant information from the data to improve the interpretability of the learned features. Our model disentangles the input image into spatial and non-spatial space. These spatial features are represented as categorical feature maps, with each category corresponding to input pixels that are spatially similar and are from the same organ part. This semantic similarity aids in learning to be generalized the anatomical representation to any modality from different scanners. Furthermore, the non-spatial features capture the image’s global intensity information, which aids the renderer in painting the anatomy in the reconstructed image. Finally, because annotating data is time-consuming and expensive, the ability to learn this decomposition through disentanglement using a small number of labels is critical in medical image analysis.
In light of these needs, here we propose a semi-supervised (CqSL) model for learning disentangled representations that combines recent developments in semi-supervised learning–generative models and adversarial learning. We aim to factorize the representation of an image pair into two parts: a shared representation that captures the common information between images and an exclusive representation that contains the specific information of each image. Furthermore, in order to achieve representation disentanglement, we propose to minimize mutual information between shared and exclusive representations. Moreover, we use feature-wise linear modulation (FiLM) [38] to distinguish the domain-invariant information from the domain-specific information, as well as a spatially adaptive normalization (SPADE) [39]-based decoder to guide the synthesis of more texture information to restrain the posterior collapse of the VAE and spatial information.
To illustrate its adequacy, our model is applied to two of the foremost critical tasks in medical imaging—segmentation of cardiac structures and reconstruction of the original image—and both assignments are handled by the same model. Our model leverages a large amount of unannotated data from the ACDC (https://www.creatis.insa-lyon.fr/Challenge/acdc/databases.html, accessed on 2 October 2021) dataset to learn the interpretable representations through judicious choices of common factors that serve as strong prior knowledge for more complicated problems—the segmentation of cardiac structures. Figure 2 shows a simplified data view of our proposed model.

1.4. Contributions

Our proposed work makes several contributions summarized as follows:
  • We combine recent developments in disentangled representation learning with strong prior knowledge about medical imaging data that features a decomposition into “skeleton (spatial)” and “sentiency (non-spatial)”, to ensure that the spatial information is not mixed up with the non-spatial information.
  • We alter the usual cross-entropy loss to down-weigh the loss applied to well-classified samples in order to overcome the foreground–background class imbalance problem. Specifically, we exploit a novel supervised loss—the weighted-soft-background-focal (WSBF) loss, which focuses the training on a set of hard examples to ensure that this loss can differentiate between easy/hard examples.
  • We employ both qualitative and quantitative tests to evaluate the usefulness of our framework, which show that our model outperformed fully supervised methods, even when using only 1% labeled data for training.
The paper is organized as follows: Section 1 establishes the general background and motivation of the work, reviews the related literature on latest developments in the field of domain adaptation, semi-supervised learning and representation learning, and provides an overview of the proposed work; Section 2 describes our proposed methodology; Section 3 presents our quantitative and qualitative results achieved using our proposed method for both image segmentation and reconstruction, along with the associated ablation studies; Section 4 concludes the paper with a summary of our contributions and promising future research directions.

2. Methods

2.1. CqSL Model Overview

We propose a model that combines the concept of variational generative and adversarial learning, and disentangled interpretation learning in a semi-supervised learning scheme, which is suited for domain-adapted segmentation as well as reconstruction.
We define the learning task as follows: given an (unknown) data distribution p ( x , y ) over images and segmentation masks, we define a source domain having a training set, D L = { ( x i l , y i l ) } i = 1 n l with n l labeled examples, and another domain having a training set, D UL = { ( x j u l ) } j = 1 n u l with n u l unlabeled examples, which are sampled as independent, identically distributed variables from p ( x , y ) and p ( x ) distribution. Empirically, we want to minimize the target risk t ( ϕ , θ ) = m i n ϕ , θ L L ( D L , ( ϕ , θ ) ) + γ L UL ( D UL , ( ϕ , θ ) ) , where L L is the supervised loss for segmentation, L UL is the unsupervised loss defined on unlabeled images and ϕ , θ denotes the learnable parameters of the overall network.
We propose to solve the task by learning domain-specific and domain-invariant features that are discriminative of the semgentor and reconstructor. Figure 3 shows the proposed model comprised of five components—(1) disentanglement component, (2) a disentangled variational autoencoder (DVAE), (3) a mask segmentor identifier (SI), (4) a mask discriminator identifier (DI), and (5) a reconstructor R.
The disentangler D (Figure 3a) is designed to factorize the representation of an image pair into two parts: a shared spatial representation (skeleton, S K e ) that captures the common information between images and an exclusive non-spatial representation (sentiency, S e ) that contains the specific information of each image. The skeleton block S K e is a modified U-Net++ [42] type architecture (EPU-Net++) (Figure 4 and Section 2.1.1) and is responsible for capturing the domain-invariant features ( f S K ) . The sentiency block S e is a DVAE (Figure 3b) type architecture, which takes both the input image and the domain-invariant features ( f S K ) as the input to map domain-specific features ( f S E ) using the reparameterized trick [43].
The reconstruction block consists of two decoders: the SPADE-based decoder takes the ( f S E ) feature from the sentiency block and proceeds directly to the reconstructor R (Figure 3d), while the FiLM-based decoder works as another disentangler, which untangles a segmentor identifier ( S I ) (Figure 3c), used for segmentation and extracted features, which then proceed directly to the reconstructor R. The reconstructor R aims to recover the original image from both ( f S K , f S E ) . A mutual information minimizer (Figure 3a block) is applied between ( S K e and S e ) to enhance the disentanglement. A supervised trainer is trained on the labeled data to predict the segmentation mask distribution optimizing a supervised loss. An unsupervised trainer is trained on the unlabeled data, optimizing unsupervised losses (Algorithm 1 specifies the overall training procedure). Both the unsupervised and supervised trainers share the same block, as mentioned above.
Algorithm 1 CqSL mini-batch training.
  Input:
  Training set of labeled data x l , y l , c l ϵ D L
  Training set of unlabeled data x u l , size m, ϵ D UL
  Disentanglement Learned parameters: ( ϕ , θ ) , generator G; segmentor S; disentangler D; discriminator identifier DI, mutual information estimator M, and reconstructor R.
  Require:
  Shared disentangler D, Shared encoder S K d k , S e and decoder
  for  e a c h   e p o c h   do
    for e a c h   s t e p do
      Sample mini-batch from x i l ; x 1 l , , x n l l ; through D L ( x )
      Sample mini-batch from x j u l ; x 1 u l , , x n u l u l ; through D UL ( x )
      Compute model outputs for the labeled inputs
       y ^ l W ϕ , θ ( I L )
      Compute model outputs for the unlabeled inputs
       y ^ u l W ϕ , θ ( I UL )
      Calculate mutual information between the disentangled feature pair ( f s k , f s e ) with M i :
      Update the mask discriminator identifier DI along its gradient:
ϕ D I 1 | I L | i I L L D I ( x i l , y i l , y ^ i l ) +
γ 1 | I UL | i I UL L D I ( x j u l , y ^ j u l )

      Update the segmentation mask generator SI and VAE encoder along its gradient:
θ S I 1 | I L | i I L L S I ( x i l , y i l , y ^ i l ) +
θ S E 1 | I L | i I L L S e ( x i l , F ( x i l ) , z d i m l ) +
γ 1 | I UL | j I UL L G ( x j u l , y ^ j u l ) +
θ S E 1 | I UL | i I UL L S e ( x j u l , F ( x j u l ) , z d i m u l )

    end for
  end for

2.1.1. Disentanglement

Referring to Figure 3a, the disentangler block factorizes the image features into spatial (skeleton/physique) features, as well as non-spatial (sentiency) features that carry residual information. The skeleton block is a modified U-Net type architecture—EvoNorm-Projection-UNet++ (EPU-Net++) as shown in Figure 4. We attach eight different decoders at the common bottleneck layer of EPU-Net++. Each decoder captures bottleneck features from 2D cropped images and transforms them into different feature maps consisting of a number of binary channels which are then combined together to form eight most effective channels: x S T ( 0 , 1 ) ( h × w × c ) { i = 1 i = 8 f S K i } . These feature maps are responsible for capturing the domain-invariant features and contain cardiac structures (myocardium, the left and the right ventricle), effective for segmentation and some surrounding structures, effective for reconstruction (Figure 5).
We use a separate neural network for capturing the sentiency information i.e., domain-specific information. We combine the crop image and the domain-invariant features to penalize the deviation of latent features from the prior distribution employing Kullback–Leibler divergence by applying a VAE architecture (Figure 3b) with the following objective function:
L v a e = | p ( z i ) log p ( z i ) p ( z i | x i u l , f S K i ) |
A VAE learns a low dimensional latent space such that the acquired latent representations fit a prior distribution that is predetermined to be an isotropic multivariate Gaussian p ( z ) = N ( 0 , 1 ) . An encoder and a decoder make up a VAE. Given an input, the encoder guesses the Gaussian distribution’s parameters. In order to enable learning through back propagation, this distribution is then sampled using the reparameterization technique, and the resulting sample is sent through the decoder to reconstruct the input.
We use disentangled features as the prior distribution in a VAE (Equation (1)) to remove class-irrelevant features (e.g., background pixels) and ensure that domain-invariant features are well-disentangled from class-specific features, because the image-only a priori aligns the latent features to a normal distribution.

2.1.2. Mutual Information Minimizer

To better exploit the disentanglement, we add a regularization term based on mutual information (MI), denoted as M I M , which measures the “amount of information” learned from knowledge of random variable Y about the other random variable X [44]. In this paper, we adopt the mutual information neural estimator (MINE) [45], M I ( f S K , f S E ) :
1 N i = 1 N M ( α , β , θ ) log 1 N i = 1 N exp M ( α , β , θ )
where ( α , β ) are sampled from the joint distribution of ( f S K , f S E ) and β is sampled from the marginal distribution.
The mutual information can be expressed as the difference of two entropy terms M I M ( X ; Y ) = H ( X ) H ( X | Y ) ; we seek to minimize the MI between domain-invariant and domain-specific features ( f S K , f S E ) , whereas we make an assumption that the information content does not vary much between intra-domains (Figure 3a).

2.1.3. Segmentation

The mask segmentor identifier ( S I ) (Figure 3c) takes the output from the FiLM decoder f S K F as input and generates predicted segmentation mask S I ( f S K ) = y ^ l { 0 , 1 } ( H × W × L ) , where L is the number of categories (RV, LV, LV-Myo, and background) in the training dataset. We exploit a novel supervised loss, weighted soft background focal (WSBF) loss, L S I ( s e g ) L = L W S F L + L B F D for the base model, which is a combination of background focal dice loss (BFD) and weighted soft focal loss (WSFL):
L S I ( s e g ) L = α 0 + y ( α 1 α 0 ) | y y ^ | γ . w m a p . C E ( y , y ^ ) + c 2 2 y y ^ + ϵ ( y + y ^ ) + ϵ 2 y ¯ y ^ ¯ + ϵ ( y ¯ + y ^ ¯ ) + ϵ 1 γ
where α 0 and α 1 are designed to account for class imbalance and are treated as hyper-parameters, the term | y y ^ | γ is used to down-weigh examples with backgrounds, where γ varies in the range [ 1 , 3 ] . The term C E ( y , y ^ ) = y log y ^ ( 1 y ) log ( 1 y ^ ) denotes the cross-entropy loss.
On the other hand, the data with no corresponding segmentation masks are trained by minimizing the unsupervised loss via a K L divergence based on least-squares GAN [46]. However, since the least-squares loss is not sufficiently robust, we introduce a new divergence loss function by incorporating it into a Geman–McClure model [47] fashion called adversarial-Geman–McClure (adv-GM) loss between the ground truth of real mask y l and prediction on unlabeled data y u l :
L S I ( a d v - G M ) U = D I ( S I ( f S K ( x u l ) ) ) 2 + ( D I ( y ^ u l ) 1 ) 2 2 β + D I ( S I ( f S K ( x u l ) ) ) 2 + ( D I ( y ^ u l ) 1 ) 2
where β is the scale factor which varies in the range of [ 0 , 1 ] and we set β = 0.5 in our experiment.

2.1.4. Image Reconstruction

To better capture the anatomical shape and the intensity information in the synthetic image, we propose a two-branched reconstruction architecture featuring two separate decoders: one is conditioned with FiLM [38], and the other with SPADE [39] (Figure 6a) and both are then concatenated to produce a realistic image. The FiLM decoder consists of multiple FiLM layers, a gamma-beta predictor, and convolutional layers with 3 × 3 kernel and (8, 8, 8, 8, 1) channels in the stride of 1. Each convolution layer is followed by batch normalization layer along with a Leaky-ReLU layer.
To better retain the non-spatial information in the MR image, we integrate the shape knowledge into the idea of SPADE [39] and form a shape-aware normalization layer (see Figure 6). SPADE first normalizes the input feature F i n with a scale α and a shift μ learned from sampled z using an instance-normalization (InstanceNorm) layer, inspired by [38] and then denormalizes it based on a spatial representation f S K through learnable parameters γ and β . f S K is then interpolated to match the texture dimension of the sampled z from the sentiency encoder and used as a semantic mask for SPADE:
F o u t = F i n μ α × γ ( f S K ) + β ( f S K )
where F i n and F o u t denote the output feature maps. γ and β are learned from f S K by three Conv layers. Thus, the learned shape information precludes washing away the anatomical information, which encourages the image synthesis to be more accurate. The first convolution layer inside the SPADE block (Figure 6) encodes the interpolated f S K , and the other two convolution layers learn the spatial tensors γ and β . Simultaneously, an instance normalization layer is applied to the intermediate feature map, which is then modulated by the scale and shift parameters γ and β learned from sampled z to produce the output. Finally, the output of the two decoders is re-entangled in order to reconstruct an image.

2.2. Objective Functions

The training objective function consists of multiple losses for labeled and unlabeled data, each weighted by some scalar term λ :
L t o t a l = λ s e g L S I ( s e g ) L + λ a d v G M { L S I , D I ( a d v G M ) L + L S I , D I u ( a d v G M ) U } + λ v a e L v a e + λ S L 2 S I M { L S L 2 S I M L + L S L 2 S I M U } + λ M I M M I M ( f S K , f S E )
where λ t is the weight for the loss of type t. In this paper, we empirically set the weights as λ v a e = 0.01 , λ s e g = 10 , λ a d v G M = 10 , λ S L 2 S I M = 0.01 , λ M I M = 1 .

2.2.1. Segmentation Loss

Since the model is trained on both labeled and unlabeled data, the segmentation loss L s e g includes both supervised and unsupervised losses:
L s e g = L s u p + L u s u p
Supervised Loss. Our supervised cost is based on the combination of the two following functions: (1) the weighted soft focal loss, and (2) the background focal dice loss mentioned in Equation (3) ( L s u p = L S I ( s e g ) L ).
Unsupervised Loss. The discriminator identifier is adversarially trained for the labeled and unlabeled data and updated along with adversarial-Geman–McClure (adv-GM) loss L u s u p = L S I , D I ( a d v G M ) L + L S I , D I u ( a d v G M ) U . For labeled data, the adversarial loss is
L S I , D I ( a d v G M ) L = E x x i l [ D I ( S I ( f S K i ( x i l ) ) ) 2 ] + E y y i l [ ( D I ( y i l ) 1 ) 2 ] 2 β + E x x i l [ D I ( S I ( f S K i ( x i l ) ) ) 2 ] + E y y i l [ ( D I ( y i l ) 1 ) 2 ]
Similarly, for the unlabeled data, the adversarial loss is
L S I , D I u ( a d v G M ) U = E x x i u l [ D I u ( S I ( f S K i ( x i u l ) ) ) 2 ] 2 β + E x x i u l [ D I u ( S I ( f S K i ( x i u l ) ) ) 2 ] + E y y ^ i u l [ ( D I u ( y i u l ) 1 ) 2 ] + E y y ^ i u l [ ( D I u ( y i u l ) 1 ) 2 ]
VAE Loss. For the smooth texture detail of the input data, the VAE learns factorized representations to optimize a KL-divergence loss, given an image x i u l , and its decomposed skeleton feature f S K (Equation (1)).

2.2.2. Reconstruction Loss

We adopt a novel reconstruction loss as a combination of structural similarity (SSIM) and L 2 loss– S L 2 S I M in order to enforce the similarity between recovered image and original image for better learning the distribution of images.
S L 2 SIM Loss. Since the image intensities vary across imaging scanners, as a result, there are high chances that the generative model will tend to m o d e c o l l a p s e . This structural L 2 similarity (S L 2 SIM) loss provides a similarity measure between the input image and the reconstructed image based on high light-dark variance, contrast, and structural similarity. The concatenated FiLM and SPADE decoder learn the parameters to reconstruct the input image using a novel combination of structured similarity loss and L 2 loss. For labeled data, the reconstruction loss is
L S L 2 S I M L = E x i x i l [ 1 S L 2 S I M { x i l , ( F ( f S K i , f S E i ) S ( f S K i , f S E i ) ) } + α i = 1 n l | | { x i l ( F ( f S K i , f S E i ) S ( f S K i , f S E i ) ) } | | 2 2 ]
Similarly, for unlabeled data, the reconstruction loss is
L S L 2 S I M U = E x i x i u l [ 1 S L 2 S I M { x i u l , ( F ( f S K i , f S E i ) S ( f S K i , f S E i ) ) } + α i = 1 n u l | | { x i u l ( F ( f S K i , f S E i ) S ( f S K i , f S E i ) ) } | | 2 2 ]
where S L 2 SIM is the structure similarity index term and α is a regularized term.

2.3. Experiments

2.3.1. Datasets

We validate the effectiveness of CqSL on a widely adopted cardiac image segmentation challenge dataset by conducting several comparisons to other baseline models. We use the STACOM 2017 Automated Cardiac Diagnosis Challenge (ACDC) dataset (https://www.creatis.insa-lyon.fr/Challenge/acdc/databases.html, accessed on 2 October 2021), consisting of short-axis cardiac cine-MR images acquired for 100 patients (1920 labeled and 23,530 unlabeled images) divided into 5 subgroups: normal (NOR), myocardial infarction (MINF), dilated cardiomyopathy (DCM), hypertrophic cardiomyopathy (HCM), and abnormal right ventricle (ARV), available through the 2017 MICCAI-ACDC STACOM challenge [48]. The images were acquired over a 6 year period using two MRI scanners of different magnetic strengths (1.5 T and 3.0 T). The images were acquired using the SSFP sequence with spatial resolution 1.37 to 1.68 mm2/pixel and 28 to 40 frames per cardiac cycle. We split the dataset into three sets—training (70), validation (15), and test (15).

2.3.2. Implementation Details

Input: All the cine cardiac images employed slice-wise normalization in the range [ 0 , 1 ] by subtracting the mean slice intensity from each pixel intensity, then dividing it by the difference between the maximum and minimum slice intensity. All images were resampled to 1.37 mm2/pixel. Images were cropped to 192 × 192 × 1 pixels before feeding to the models. We applied data augmentation on-the-fly during training as shown in Figure 7, which includes random rotations up to 90 degrees, random zooms up to 20%, random horizontal shifts up to 20%, random horizontal and/or vertical flips, and noise addition (Figure 7).
Baselines Architecture: As the disentangled encoder in the skeletal block, we use a modified U-Net-like architecture, EPU-Net++, and as a sentiency encoder, we use VAE. As the reconstruction block, we use FiLM- and SPADE-based decoder as used in [49].
Generator–Discriminator Network: Our segmentation generator network consists of 3 convolution layers with 3 × 3 kernel and {64, 64, 1} channels in the stride of 1. Each convolution layer is followed by a batch normalization [50] layer along with a Leaky-ReLU [51] except the last layer. We use the structure similar to DCGAN [52] for the discriminator network.
EvoNorm-Projection skip connections: In our skeleton encoder, we replace the standard skip connection with a normalized-projection operation using E v o N o r m 2 D + 1 × 1 C o n v + G a u s s i a n d r o p o u t , as in Figure 4. This new normalization layer adds together two types of statistical moments–batch variance, and instance variance, both of which capture both the global and local information across images without having any explicit activation function [53]. The proposed projection operation helps in reducing the learnable weights and also allows intricate learnability of cross-channel information.
Additional Factors: The performance of semi-supervised models trained for image segmentation can be significantly impacted by the proper selection of regularizer, optimizer, and hyper-parameters. The model implemented in Keras was initialized with the He normal initializer and trained for 100 epochs with a batch size of 4. We trained all the components iteratively with the Adam optimizer with a 0.0001 learning rate to minimize the objective function. All experiments were conducted on a machine equipped with two NVIDIA RTX 2080 Ti GPU (each 11GBs memory). The detailed training procedure is presented in Algorithm 1.
Training: In our semi-supervised setup, we trained the network on varying proportions of labeled data: 1%, 10%, 20%, 30%, 50%, and 90% as a labeled set and used the rest of the data as the training unlabeled set to hold | D L | | D UL | . In Section 3, we include an ablation study to investigate the importance of adding different loss components in our model CqSL which is comprised of all the three loss functions: WSBF, MIM, Adv-GM. (Definitions are provided in Section 2.1.2 and Section 2.1.3.)
We experimented an ablation study containing four of the variants of our proposed model CqSL. The variants are described as follows: 1 CqSL, without weighted-soft focal loss (WSFL); 2 CqSL, without adversarial-Geman–McClure loss (Adv-GM); 3 CqSL, dice and cross-entropy loss only; and 4 CqSL, without mutual information minimizer loss (MIM). Here, we utilize the same backbones as the baselines with the only exceptions being different loss functions. To clarify our point, in 1 CqSL, we removed the weighted soft focal loss (WSFL) from the weighted soft background focal loss (WSBF), while keeping the background focal dice loss (BFD), mutual information minimizer loss (MIM) and adversarial-Geman–McClure (adv-GM) the same as before. In 2 CqSL, we removed our Geman–McClure version of adversarial loss, while keeping the regular adversarial loss, weighted soft background focal loss (WSBF), and mutual information minimizer loss (MIM) the same as before. Similarly, in 3 CqSL, we used D I C E + C E loss rather than using our novel weighted soft background focal loss (WSBF) while keeping the mutual information minimizer loss (MIM) and adversarial-Geman–McClure (adv-GM) the same as before. Finally, in 4 CqSL, we removed our mutual information minimizer loss (MIM) loss, while keeping the weighted soft background focal loss (WSBF), and adversarial Geman–McClure (adv-GM) the same as before. Additionally, the sentiency block, S e and the skeleton block, S K e were in place. We evaluated the performance of all four CqSL semi-supervised variants as summarized in Table 1, Table 2 and Table 3 in the Results section, and, as illustrated later, the 1 CqSL variant performed best, but for the sake of consistency, we asses and compare the performance of all four implemented variants.

2.4. Evaluation Metrics

To evaluate the performance of the semantic segmentation of cardiac structures, we use the standard metrics, including Dice score, Jaccard index, Hausdorff distance (HD), precision (Prec), and recall (Rec).
  • Dice and Jaccard Coefficients: The Dice score is used to measure the percentage of overlap between manually segmented boundaries and automatically segmented boundaries of the structures of interest. Given the set of all pixels in the image, set of foreground pixels by automated segmentation S 1 a , and the set of pixels for ground truth S 1 g , the Dice score can be compared with [ S 1 a , S 1 g ]⊆ Ω , when a vector of ground truth labels T 1 and a vector of predicted labels P 1 as
    D i c e ( T 1 , P 1 ) = 2 | T 1 P 1 | | T 1 | + | P 1 |
    The Dice score will measure the similarity between two sets, T 1 and P 1 , and | T 1 | denotes the cardinality of the set T 1 with the range of D( T 1 , P 1 ) ϵ [0, 1].
    The Jaccard index or Jaccard similarity coefficient is another metric which aids in the evaluation of the overlap in two sets of data. This index is similar to the Dice coefficient but mathematically different and typically used for different applications. For the same set of pixels in the image, Jaccard index can be written by the following expression:
    J a c c a r d ( T 1 , P 1 ) = | T 1 P 1 | | T 1 + P 1 |
  • Precision and Recall
    Precision and recall are two other metrics used to measure the segmentation quality which are sensitive to under- and over-segmentation. High values of both precision and recall indicate that the boundaries in both segmentation agree in location and level of detail. Precision and recall can be written as
    P r e c i s i o n = T P T P + F P
    R e c a l l = T P T P + F N
    where T P denotes true positive rate when a prediction-target mask pair has a score which exceeds some predefined threshold value; F P denotes the false positive rate when a predicted mask has no associated ground truth mask; and F N denotes the false negative rate when a ground truth mask has no associated predicted mask.
  • Hausdorff distance (HD): Hausdorff distance (HD) measures the maximum distance between the two surfaces. Let S A and S B be surfaces corresponding to two binary segmentation masks, A and B, respectively. The Hausdorff distance (HD) is defined as
    H D = max max p ϵ S A d ( p , S B ) , max q ϵ S B d ( q , S A )
    where d ( p , S ) = min q ϵ S d ( p , q ) is the minimum Euclidean distance of point p from the points q ϵ S.
  • Image Quality Metrics:
    PSNR: The peak signal-to-noise ratio (PSNR) is the most commonly used quality assessment technique for determining the quality of lossy image compression codec reconstruction. The signal is the original data, and the noise is the error caused by the distortion.
  • Clinical Indices: To assess the performance of the ventricles, different indices have been used in the literature [54], such as left ventricular volume (LVV), left ventricular myocardial mass (LVM), stroke volume (SV), and ejection fraction (EF). The left ventricular volume (LVV) is defined as the volume enclosed by the LV blood pool and the myocardial mass is equal to the volume of the myocardium, multiplied by the density of the myocardium:
    Myo Mass = Myo Volume ( cm 3 ) × 1.06 ( gram / cm 3 )
    Stroke volume (SV) is defined as the volume ejected during systole and is equal to the difference between the end-diastolic volume (EDV) and the end-systolic volume (ESV):
    S V = E D V E S V × 100 %
    The ejection fraction (EF) is an important cardiac parameter quantifying the cardiac output and defined as the ratio of the SV to the EDV:
    E F = S V E D V × 100 %

3. Results

3.1. Image Segmentation Assessment

We tested our CqSL model on varying proportions of labeled and unlabeled data available through the STACOM 2017 ACDC cine cardiac MRI dataset. Training and validation segmentation accuracies for three different classes (RV, LV, and LV-Myo) are shown in Figure 8 for 100 epochs. Note that the validation curves show similar trends as the training curves (Figure 8).
The CqSL experimental results were compared against a fully supervised U-Net model trained from scratch, as reported in Table 1, Table 2 and Table 3. Furthermore, to explore the effectiveness of each component in our model, we propose three different semi-supervised ablations, i.e., model I: only a GAN architecture (Figure 3c); model II: I + reconstruction (Figure 3c,d); model III: II + disentangler block (Figure 3a–d), which are also reported in Table 1, Table 2 and Table 3. The detailed comparison of our model can be seen in Table 4. The segmentation performance is evaluated both qualitatively and quantitatively. As shown in Table 1, Table 2 and Table 3, our proposed model significantly improves the segmentation performance of right ventricle (RV), left ventricle blood-pool (LV), and LV-Myocardium, respectively on varying proportions of annotated data in terms of the Dice and Jaccard indices, Hausdorff distance, precision and recall rates. Our CqSL model achieves a high dice score (±std. dev.) of 75.50 ± 10.9% for the RV, 83.21 ± 7.1% for the LV blood-pool and 77.65 ± 9.3% for the LV-Myocardium even if we use only 1% labeled data.
Figure 9 illustrates a qualitative segmentation output that compared CqSL and two others semi-supervised models, i.e., model I: only a GAN architecture (Figure 3c); model II: I + reconstruction (Figure 3c,d). For simplicity, this comparison is based on 20% unlabeled training data. As demonstrated, when only 20% of the training annotation is employed, U-Net fails completely to segment the cardiac structures from base to apex, particularly RV segmentation. As shown in the figure, the segmentation results improve with each consecutive addition of a distinct block. The GAN-only architecture performs badly, particularly during RV segmentation, whereas the addition of a reconstruction block improves performance. Finally, adding a disentangled block to the GAN and reconstruction block yielded the greatest results. Even the least performing version of our proposed CqSL model ( 4 CqSL) achieves an overall accuracy superior to the U-Net, GAN-only, as well as GAN+REC model, confirming that the proposed model is able to effectively learn correct features that ensure correct segmentation.
Figure 10 illustrates a qualitative segmentation output that compared CqSL and U-Net results with increasing proportion of unlabeled training data. For simplicity, we have shown two of our best performing models. As shown, when only 1% training annotation is used, U-Net completely fails to segment the cardiac structures. Under similar conditions, our model is still able to yield a high segmentation accuracy of LV, RV, and LV-Myocardium. When the amount of labeled data increases from 1% to 10%, the U-Net model still performs poorly, especially for RV segmentation. On the other hand, although the performance of our model improves significantly when utilizing more than 30% annotated data, its performance with even 1% labeled data is still satisfactory, comparable to that of semi-supervised models, and superior to U-Net’s performance under similar conditions.
We assessed the performance of our proposed CqSL cardiac image segmentation method against the segmentation results yielded by the well-established, fully supervised U-Net architecture [55] in light of its effectiveness across various medical image segmentation applications, as well as its extensive use as a baseline method for comparison by the participants of the ACDC cardiac image segmentation challenge. Furthermore, to explore the effectiveness of each component in our model, we experiment on three different semi-supervised ablations, i.e., model I: only a GAN architecture; model II: GAN + reconstruction; and model III: GAN + reconstruction + disentangler block (CqSL).
As shown in Figure 11, the accuracy of our CqSL models remains high when using as much as 50–90% unlabeled data, which essentially implies excellent performance with as little as as 10% annotated data. Nevertheless, both U-Net and CqSL models perform similar to each other when the amount of annotated data increases above 90%. We plot the mean accuracy for all the models in Figure 12 and confirm that under low amounts of annotated data conditions, even as low as 1%, our proposed CqSL model and all four of its semi-supervised variants ( 1 CqSL, 2 CqSL, 3 CqSL, and 4 CqSL) outperform GAN, GAN+REC, as well as U-Net models for LV, RV, and LV-Myocardium. The typical segmentation contours of complete cardiac image dataset for the mid and apical slices are shown in Figure 13.

3.2. Image Quality Assessment

Figure 14 illustrates a qualitative comparison between the original image slice and the reconstructed slices generated from our proposed approach on the ACDC dataset at the original 5 mm slice thickness. The comparison is augmented by the computed correlation coefficients (CC) and peak signal-to-noise ratio (PSNR) shown below each figure. As illustrated in Figure 14, our approach preserves the fine structural details and realistic textures while remaining visually comparable to the ground truth image. Aside from qualitative improvements, the proposed method’s CC and PSNR values also prove that the synthesized image slices preserve the fine structural details.
Table 5 shows the quantitative results of the objective quality metrics of reconstruction, indicating that the use of feature-wise linear modulation to remove domain-invariant information from the disentangled latent code guides the synthesis of more texture information. Starting with the spatial factor, we change the content of the spatial channels in Figure 15 to see how the decoder has learned a correlation between the position of each channel and different signal intensities of the skeleton parts. The sentiency factor remains constant in all of these experiments. The first two columns show the original input and the reconstruction. The third row is created by the RV spatial channels and disregarding (zeroing) the MYO and LV channel. In the fourth image, we swap the RV channels with those of LV. Finally, the fifth column is produced by considering all LV, MYO and RV channels.

3.3. Clinical Parameter Estimation

The performance of our developed segmentation method was also reflected in the computed clinical indices. These clinical indices are computed using the Simpsons method and the agreement between the ground truth and the same parameters computed using the automated segmentation results is reported using correlation statistical analysis by mapping the predicted volumes of the testing set onto the ground truth volumes of the training set. As illustrated in Table 6 the agreement between our method’s prediction and ground truth is high, characterized by a Pearson’s correlation coefficient (rho) of 0.898 (p < 0.01) for LV-EF, 0.723 for RV-EF (p < 0.1) and 0.924 (p < 0.01) for Myo-mass. There was a slight over-estimation in the RV blood-pool segmentation also reflected in the clinical parameters estimation.
Figure 16 shows a graphical comparison between the clinical parameters estimated from the cardiac features segmented via CqSL and the same homologous parameters estimated from the ground truth manual segmentations for both healthy volunteers and patients featuring various cardiac conditions. As shown, the clinical parameters estimated using our automatically segmented features show no statistically significant difference from those estimated based on the ground truth, manually segmented features.

3.4. Ablation Studies

We perform an ablation study to investigate the effect of using different loss functions in our semi-supervised setting. We demonstrate the effect of different novel loss functions used in CqSL model: WSBF, MIM, and Adv-GM by assessing the model performance when each novel loss functions is removed. Figure 17 shows a graphical representation of the results achieved on the ACDC dataset. In Figure 10, we illustrate the qualitative results on the ACDC dataset to visualize the effect of using all of the loss components. We can observe that the best results are achieved when all of the loss components are used. Specifically, without MIM, the loss curve oscillates, while without WSBF, the output images deviate drastically from the ground truth. Both the quantitative and qualitative results show that the design of CqSL improves the preservation of the subject identity and enables more accurate segmentation of cardiac structures.

4. Conclusions and Future Work

In this paper, we propose a semi-supervised learning model (CqSL) that features multiple novel loss functions, including mutual information minimization (MIM), which minimizes the mutual information between the domain-invariant as well as domain-specific features. Empirically, we show that disentanglement with mutual information can improve the performance of the segmentation accuracy, while combined with an adversarial and a reconstruction block. Our novel use of total loss function enforces the network to capture both the spatial and intensity information. Our weighted soft focal loss can minimize the class imbalance problem by applying varying weights over different classes along with a modulating term. We apply the proposed model to cardiac image segmentation tasks with varying proportion of labeled data.
Our proposed CqSL model achieves 85% accuracy, significantly outperforming other baselines. We incrementally add each component, aiming to study their effectiveness on the final results: (model I: only a GAN architecture (Figure 3c); model II: GAN + reconstruction (Figure 3c,d); model III: GAN + reconstruction + disentangled block (Figure 3a–d).
In light of consistency, all four implemented CqSL variants are evaluated and compared to the baselines, but as shown in Table 1, Table 2 and Table 3, the first variant ( 1 CqS) performs best and hence it is deemed as the most suitable and recommended CqSL framework.
The experimental results reported in this manuscript show that the proposed CqSL framework outperforms semi-supervised learning with GANs [56] as well as fully supervised-type models when using as little as even 1% labeled data and display similar performance and comparable accuracy when employing more than 50% labeled data. Unlike these, we use adversarial-Geman–McClure (adv-GM) loss to force mask generation to be spatially aligned with the image. Furthermore, we discover that the semi-supervised segmentation approach of Hung et al. [18] obtains results slightly inferior to ours. Hung et al. reported that their adversarial model achieved a 80.63% accuracy when trained on 20% labeled data using the ACDC dataset, whereas our model achieved a 81.44% accuracy under similar training conditions.
Hence, the proposed method is the first to achieve significant performance for 4D cine cardiac MRI image segmentation with very minimal annotated data, specifically 1% of the training dataset. This is a key feature of the proposed work and hence a significant contribution to the medical (cardiac, in particular) image segmentation, as access to large amounts of expert-annotated ground truth imaging data is expensive in the medical field. Nevertheless, here we demonstrate that CqSL can still yield segmentation accuracy superior to other semi-supervised methods while requiring minimal annotated data for training.

Author Contributions

Conceptualization, S.M.K.H. and C.A.L.; methodology, S.M.K.H.; software, S.M.K.H.; validation, S.M.K.H.; formal analysis, S.M.K.H. and C.A.L.; resources, S.M.K.H. and C.A.L.; writing—original draft preparation, S.M.K.H.; writing—review and editing, S.M.K.H. and C.A.L.; visualization, S.M.K.H.; supervision, C.A.L.; funding acquisition, C.A.L. All authors have read and agreed to the published version of the manuscript.

Funding

Research reported in this publication was supported by the National Institute of General Medical Sciences Award No. R35GM128877 of the National Institutes of Health, and the Office of Advanced Cyber infrastructure Award No. 1808530 of the National Science Foundation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bhowmik, A.; Gumhold, S.; Rother, C.; Brachmann, E. Reinforced feature points: Optimizing feature detection and description for a high-level task. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4948–4957. [Google Scholar]
  2. Li, S.; Wang, Z.; Liu, Z.; Tan, C.; Lin, H.; Wu, D.; Chen, Z.; Zheng, J.; Li, S.Z. Efficient Multi-order Gated Aggregation Network. arXiv 2022, arXiv:2211.03295. [Google Scholar]
  3. Ruan, J.; Xiang, S.; Xie, M.; Liu, T.; Fu, Y. MALUNet: A Multi-Attention and Light-weight UNet for Skin Lesion Segmentation. arXiv 2022, arXiv:2211.01784. [Google Scholar]
  4. Tack, J.; Yu, S.; Jeong, J.; Kim, M.; Hwang, S.J.; Shin, J. Consistency regularization for adversarial robustness. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 22 February–1 March 2022; Volume 36, pp. 8414–8422. [Google Scholar]
  5. Sajjadi, M.; Javanmardi, M.; Tasdizen, T. Regularization with stochastic transformations and perturbations for deep semi-supervised learning. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 1163–1171. [Google Scholar]
  6. Elakkiya, R.; Subramaniyaswamy, V.; Vijayakumar, V.; Mahanti, A. Cervical cancer diagnostics healthcare system using hybrid object detection adversarial networks. IEEE J. Biomed. Health Inform. 2021, 26, 1464–1471. [Google Scholar] [CrossRef] [PubMed]
  7. Hasan, S.M.K.; Linte, C. STAMP: A Self-training Student-Teacher Augmentation-Driven Meta Pseudo-Labeling Framework for 3D Cardiac MRI Image Segmentation. In Proceedings of the Annual Conference on Medical Image Understanding and Analysis, Cambridge, UK, 27–29 July 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 371–386. [Google Scholar]
  8. Sohn, K.; Berthelot, D.; Li, C.L.; Zhang, Z.; Carlini, N.; Cubuk, E.D.; Kurakin, A.; Zhang, H.; Raffel, C. Fixmatch: Simplifying semi-supervised learning with consistency and confidence. arXiv 2020, arXiv:2001.07685. [Google Scholar]
  9. Xie, Q.; Luong, M.T.; Hovy, E.; Le, Q.V. Self-training with noisy student improves imagenet classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10687–10698. [Google Scholar]
  10. Bai, W.; Oktay, O.; Sinclair, M.; Suzuki, H.; Rajchl, M.; Tarroni, G.; Glocker, B.; King, A.; Matthews, P.M.; Rueckert, D. Semi-supervised learning for network-based cardiac MR image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 10–14 September 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 253–260. [Google Scholar]
  11. Donahue, J.; Krähenbühl, P.; Darrell, T. Adversarial feature learning. arXiv 2016, arXiv:1605.09782. [Google Scholar]
  12. Saito, K.; Kim, D.; Sclaroff, S.; Darrell, T.; Saenko, K. Semi-supervised domain adaptation via minimax entropy. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8050–8058. [Google Scholar]
  13. Gomes, H.M.; Grzenda, M.; Mello, R.; Read, J.; Le Nguyen, M.H.; Bifet, A. A survey on semi-supervised learning for delayed partially labelled data streams. ACM Comput. Surv. (CSUR) 2022, 55, 1–42. [Google Scholar] [CrossRef]
  14. Hasan, S.M.K.; Linte, C.A. A Multi-Task Cross-Task Learning Architecture for Ad Hoc Uncertainty Estimation in 3D Cardiac MRI Image Segmentation. In Proceedings of the 2021 Computing in Cardiology (CinC), Brno, Czech Republic, 13–15 September 2021; Volume 48, pp. 1–4. [Google Scholar]
  15. Chan, E.R.; Lin, C.Z.; Chan, M.A.; Nagano, K.; Pan, B.; De Mello, S.; Gallo, O.; Guibas, L.J.; Tremblay, J.; Khamis, S.; et al. Efficient geometry-aware 3D generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 16123–16133. [Google Scholar]
  16. Souly, N.; Spampinato, C.; Shah, M. Semi supervised semantic segmentation using generative adversarial network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5688–5696. [Google Scholar]
  17. Chen, C.; Dou, Q.; Chen, H.; Heng, P.A. Semantic-aware generative adversarial nets for unsupervised domain adaptation in chest X-ray segmentation. In Proceedings of the International Workshop on Machine Learning in Medical Imaging, Granada, Spain, 16 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 143–151. [Google Scholar]
  18. Hung, W.C.; Tsai, Y.H.; Liou, Y.T.; Lin, Y.Y.; Yang, M.H. Adversarial learning for semi-supervised semantic segmentation. arXiv 2018, arXiv:1802.07934. [Google Scholar]
  19. Zhang, Y.; Yang, L.; Chen, J.; Fredericksen, M.; Hughes, D.P.; Chen, D.Z. Deep adversarial networks for biomedical image segmentation utilizing unannotated images. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 10–14 September 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 408–416. [Google Scholar]
  20. Chartsias, A.; Joyce, T.; Dharmakumar, R.; Tsaftaris, S.A. Adversarial image synthesis for unpaired multi-modal cardiac data. In Proceedings of the International Workshop on Simulation and Synthesis in Medical Imaging, Quebec City, QC, Canada, 10 September 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 3–13. [Google Scholar]
  21. Hjelm, R.D.; Fedorov, A.; Lavoie-Marchildon, S.; Grewal, K.; Bachman, P.; Trischler, A.; Bengio, Y. Learning deep representations by mutual information estimation and maximization. arXiv 2018, arXiv:1808.06670. [Google Scholar]
  22. Wang, Y.C.; Wang, C.Y.; Lai, S.H. Disentangled Representation with Dual-stage Feature Learning for Face Anti-spoofing. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 1955–1964. [Google Scholar]
  23. Siddharth, N.; Paige, B.; Van de Meent, J.W.; Desmaison, A.; Goodman, N.; Kohli, P.; Wood, F.; Torr, P. Learning disentangled representations with semi-supervised deep generative models. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5925–5935. [Google Scholar]
  24. Higgins, I.; Matthey, L.; Pal, A.; Burgess, C.; Glorot, X.; Botvinick, M.; Mohamed, S.; Lerchner, A. beta-vae: Learning Basic Visual Concepts with a Constrained Variational Framework. 2016. Available online: https://openreview.net/forum?id=Sy2fzU9gl (accessed on 2 October 2022).
  25. Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef] [Green Version]
  26. Lipton, Z.C. The mythos of model interpretability. Queue 2018, 16, 31–57. [Google Scholar] [CrossRef]
  27. Schölkopf, B.; Janzing, D.; Peters, J.; Sgouritsa, E.; Zhang, K.; Mooij, J. On causal and anticausal learning. arXiv 2012, arXiv:1206.6471. [Google Scholar]
  28. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  29. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  30. Huang, X.; Liu, M.Y.; Belongie, S.; Kautz, J. Multimodal unsupervised image-to-image translation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 172–189. [Google Scholar]
  31. Shen, K.; Jones, R.M.; Kumar, A.; Xie, S.M.; HaoChen, J.Z.; Ma, T.; Liang, P. Connect, not collapse: Explaining contrastive learning for unsupervised domain adaptation. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 19847–19878. [Google Scholar]
  32. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image style transfer using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 37–30 June 2016; pp. 2414–2423. [Google Scholar]
  33. Liu, A.H.; Liu, Y.C.; Yeh, Y.Y.; Wang, Y.C.F. A unified feature disentangler for multi-domain image translation and manipulation. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 2590–2599. [Google Scholar]
  34. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7167–7176. [Google Scholar]
  35. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Improved texture networks: Maximizing quality and diversity in feed-forward stylization and texture synthesis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6924–6932. [Google Scholar]
  36. Dumoulin, V.; Shlens, J.; Kudlur, M. A learned representation for artistic style. arXiv 2016, arXiv:1610.07629. [Google Scholar]
  37. Huang, X.; Belongie, S. Arbitrary style transfer in real-time with adaptive instance normalization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1501–1510. [Google Scholar]
  38. Perez, E.; Strub, F.; De Vries, H.; Dumoulin, V.; Courville, A. Film: Visual reasoning with a general conditioning layer. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  39. Park, T.; Liu, M.Y.; Wang, T.C.; Zhu, J.Y. Semantic image synthesis with spatially-adaptive normalization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 2337–2346. [Google Scholar]
  40. Marino, J. Predictive coding, variational autoencoders, and biological connections. Neural Comput. 2022, 34, 1–44. [Google Scholar] [CrossRef] [PubMed]
  41. Kim, H.; Mnih, A. Disentangling by factorising. arXiv 2018, arXiv:1802.05983. [Google Scholar]
  42. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  43. Tian, R.; Mao, Y.; Zhang, R. Learning VAE-LDA models with rounded reparameterization trick. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Virtually, 16–20 November 2020; pp. 1315–1325. [Google Scholar]
  44. Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 2172–2180. [Google Scholar]
  45. Peng, X.; Huang, Z.; Sun, X.; Saenko, K. Domain agnostic learning with disentangled representations. arXiv 2019, arXiv:1904.12347. [Google Scholar]
  46. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Paul Smolley, S. Least squares generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2794–2802. [Google Scholar]
  47. Ganan, S.; McClure, D. Bayesian Image Analysis: An Application to Single Photon Emission Tomography; American Statistical Association: Washington, DC, USA, 1985; pp. 12–18. [Google Scholar]
  48. Bernard, O.; Lalande, A.; Zotti, C.; Cervenansky, F.; Yang, X.; Heng, P.A.; Cetin, I.; Lekadir, K.; Camara, O.; Ballester, M.A.G.; et al. Deep learning techniques for automatic MRI cardiac multi-structures segmentation and diagnosis: Is the problem solved? IEEE Trans. Med. Imaging 2018, 37, 2514–2525. [Google Scholar] [CrossRef] [PubMed]
  49. Chartsias, A.; Joyce, T.; Papanastasiou, G.; Semple, S.; Williams, M.; Newby, D.E.; Dharmakumar, R.; Tsaftaris, S.A. Disentangled representation learning in cardiac image analysis. Med. Image Anal. 2019, 58, 101535. [Google Scholar] [CrossRef] [Green Version]
  50. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  51. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; Volume 30, p. 3. [Google Scholar]
  52. Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  53. Liu, H.; Brock, A.; Simonyan, K.; Le, Q.V. Evolving Normalization-Activation Layers. arXiv 2020, arXiv:2004.02967. [Google Scholar]
  54. Frangi, A.F.; Niessen, W.J.; Viergever, M.A. Three-dimensional modeling for functional analysis of cardiac images, a review. IEEE Trans. Med. Imaging 2001, 20, 2–5. [Google Scholar] [CrossRef]
  55. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  56. Luc, P.; Couprie, C.; Chintala, S.; Verbeek, J. Semantic segmentation using adversarial networks. arXiv 2016, arXiv:1611.08408. [Google Scholar]
Figure 1. Images, histograms and surface plots of two 3D cardiac images featuring all slices of two random patients from the ACDC dataset are illustrated in (a,b). From left to right: cardiac MR image in 4 dimensions, histogram plot, and surface plot.
Figure 1. Images, histograms and surface plots of two 3D cardiac images featuring all slices of two random patients from the ACDC dataset are illustrated in (a,b). From left to right: cardiac MR image in 4 dimensions, histogram plot, and surface plot.
Applsci 12 12163 g001
Figure 2. A simplified schematic overview of the proposed model.
Figure 2. A simplified schematic overview of the proposed model.
Applsci 12 12163 g002
Figure 3. Illustration of CqSL framework: Our model makes use of both labeled as well as unlabeled images. The first block (a) crops the input images to a specific dimension. Then, we disentangle the latent features of the images via a disentangled block. An input image is first encoded to a multi-channel spatial representation, S K d n = 1 , 2 8 . Then, S K d n can be fed into a segmentation network S I to generate a multi-class segmentation mask. (c) We train a generative network, which predicts semantic labels for both labeled and unlabeled data. (b) A sentiency encoder S e uses the factor S K d n and the input image to generate a latent vector z representing the imaging modality using a variational autoencoding block. (d) The decoder networks combine the two representations S K d n and z to reconstruct the input image.
Figure 3. Illustration of CqSL framework: Our model makes use of both labeled as well as unlabeled images. The first block (a) crops the input images to a specific dimension. Then, we disentangle the latent features of the images via a disentangled block. An input image is first encoded to a multi-channel spatial representation, S K d n = 1 , 2 8 . Then, S K d n can be fed into a segmentation network S I to generate a multi-class segmentation mask. (c) We train a generative network, which predicts semantic labels for both labeled and unlabeled data. (b) A sentiency encoder S e uses the factor S K d n and the input image to generate a latent vector z representing the imaging modality using a variational autoencoding block. (d) The decoder networks combine the two representations S K d n and z to reconstruct the input image.
Applsci 12 12163 g003
Figure 4. Illustration of EPU-Net++ Block: skip connections are replaced with a long projection block.
Figure 4. Illustration of EPU-Net++ Block: skip connections are replaced with a long projection block.
Applsci 12 12163 g004
Figure 5. Representative examples showing the 5 (out of 8) most semantic disentangled multi-channel binary maps of the spatial information generated from the skeleton decoder from the base to apex (top to bottom rows). Some channels indicate anatomical portions that are well-defined, such as the myocardium, left ventricle or the right ventricle, while others represent the remaining anatomy needed to characterize the input image.
Figure 5. Representative examples showing the 5 (out of 8) most semantic disentangled multi-channel binary maps of the spatial information generated from the skeleton decoder from the base to apex (top to bottom rows). Some channels indicate anatomical portions that are well-defined, such as the myocardium, left ventricle or the right ventricle, while others represent the remaining anatomy needed to characterize the input image.
Applsci 12 12163 g005
Figure 6. Detailed architecture of SPADE block: (a) shape-aware normalization block where the spatial tensors, γ and β are multiplied and added to the input features; (b) decoder block f S E S with shape-aware normalization.
Figure 6. Detailed architecture of SPADE block: (a) shape-aware normalization block where the spatial tensors, γ and β are multiplied and added to the input features; (b) decoder block f S E S with shape-aware normalization.
Applsci 12 12163 g006
Figure 7. Example images of applying data augmentation via affine transformations.
Figure 7. Example images of applying data augmentation via affine transformations.
Applsci 12 12163 g007
Figure 8. Representative accuracy curves showing the training and validation accuracy of three different classes (RV blood-pool, LV blood-pool, and LV-Myocardium).
Figure 8. Representative accuracy curves showing the training and validation accuracy of three different classes (RV blood-pool, LV blood-pool, and LV-Myocardium).
Applsci 12 12163 g008
Figure 9. Representative results showing the comparison across several best performing networks, including CqSL for the semantic segmentation of full cardiac image dataset from the base to apex showing of RV blood-pool, LV blood-pool, and LV-Myocardium on 20% labeled data in red, green, and yellow respectively.
Figure 9. Representative results showing the comparison across several best performing networks, including CqSL for the semantic segmentation of full cardiac image dataset from the base to apex showing of RV blood-pool, LV blood-pool, and LV-Myocardium on 20% labeled data in red, green, and yellow respectively.
Applsci 12 12163 g009
Figure 10. Representative results showing the semantic segmentation of RV, LV blood-pool, and LV-Myocardium on different proportion of labeled data in red, green, and yellow, respectively.
Figure 10. Representative results showing the semantic segmentation of RV, LV blood-pool, and LV-Myocardium on different proportion of labeled data in red, green, and yellow, respectively.
Applsci 12 12163 g010
Figure 11. Consistent improvement in segmentation accuracy by the proposed CqSL model over baseline semi-supervised (variants of our CqSL model: 1 CqSL, 2 CqSL, 3 CqSL, and 4 CqSL) and fully supervised models in varying proportions of labeled training data.
Figure 11. Consistent improvement in segmentation accuracy by the proposed CqSL model over baseline semi-supervised (variants of our CqSL model: 1 CqSL, 2 CqSL, 3 CqSL, and 4 CqSL) and fully supervised models in varying proportions of labeled training data.
Applsci 12 12163 g011
Figure 12. Evaluation on the robustness of CqSL in terms of mean accuracy over RV, LV, and LV-Myocardium segmentation tasks on varying amounts of labeled training samples. Note significant improvement in Dice score across all CqSL semi-supervised variants for as little as 1% unlabeled data.
Figure 12. Evaluation on the robustness of CqSL in terms of mean accuracy over RV, LV, and LV-Myocardium segmentation tasks on varying amounts of labeled training samples. Note significant improvement in Dice score across all CqSL semi-supervised variants for as little as 1% unlabeled data.
Applsci 12 12163 g012
Figure 13. Representative segmentation contours of a complete cardiac cycle for the middle and apex slices showing RV and LV blood-pool, and LV-Myocardium in green, yellow, and brown, respectively, in three different view settings (axial, sagittal, and coronal).
Figure 13. Representative segmentation contours of a complete cardiac cycle for the middle and apex slices showing RV and LV blood-pool, and LV-Myocardium in green, yellow, and brown, respectively, in three different view settings (axial, sagittal, and coronal).
Applsci 12 12163 g013
Figure 14. Qualitative comparison of the original and the reconstructed slices showing that the original images are well reconstructed by combining skeleton and sentiency information.The comparison is augmented by the computed correlation coefficients (CC) and peak signal-to-noise ratio (PSNR). The middle row illustrates the error images.
Figure 14. Qualitative comparison of the original and the reconstructed slices showing that the original images are well reconstructed by combining skeleton and sentiency information.The comparison is augmented by the computed correlation coefficients (CC) and peak signal-to-noise ratio (PSNR). The middle row illustrates the error images.
Applsci 12 12163 g014
Figure 15. Reconstructions of a sample of input images when rearranging the spatial representation’s channels. Rearranging the channels results in reconstructing only left ventricle blood-pool or only right ventricle blood-pool only or all the ventricular structures.
Figure 15. Reconstructions of a sample of input images when rearranging the spatial representation’s channels. Rearranging the channels results in reconstructing only left ventricle blood-pool or only right ventricle blood-pool only or all the ventricular structures.
Applsci 12 12163 g015
Figure 16. Graphical comparison showing no statistically significant differences between clinical parameters estimated using CqSL segmentation and same parameters estimated using the ground truth segmentation in terms of Mean (Std. Dev.) EF (mL/mL (%)) = ejection fraction, Myo-mass (in gm) = myocardial mass.
Figure 16. Graphical comparison showing no statistically significant differences between clinical parameters estimated using CqSL segmentation and same parameters estimated using the ground truth segmentation in terms of Mean (Std. Dev.) EF (mL/mL (%)) = ejection fraction, Myo-mass (in gm) = myocardial mass.
Applsci 12 12163 g016
Figure 17. Empirical analysis showing the effect of different loss functions on the 2017 STACOM ACDC dataset. The significant reduction of total loss in CqSL (in red) suggests the best performing model with best learned features.
Figure 17. Empirical analysis showing the effect of different loss functions on the 2017 STACOM ACDC dataset. The significant reduction of total loss in CqSL (in red) suggests the best performing model with best learned features.
Applsci 12 12163 g017
Table 1. Quantitative evaluation of RV blood pool segmentation results achieved using four semi-supervised variants of the proposed CqSL model in terms of mean Dice score (%) with std. dev., Jaccard index, Hausdorff distance (mm), precision (%) and recall (%) rate evaluated for varying proportions of labeled data on the ACDC dataset compared across several frameworks.
Table 1. Quantitative evaluation of RV blood pool segmentation results achieved using four semi-supervised variants of the proposed CqSL model in terms of mean Dice score (%) with std. dev., Jaccard index, Hausdorff distance (mm), precision (%) and recall (%) rate evaluated for varying proportions of labeled data on the ACDC dataset compared across several frameworks.
Right Ventricle (RV)
DiceJaccardHDPrec.Rec.
U-Net-90%80.50 ± 8.4572.03 ± 9.778.89 ± 8.4590.0994.35
U-Net-50%79.21 ± 8.4970.26 ± 10.698.90 ± 6.1285.3290.11
U-Net-30%72.32 ± 10.6066.10 ± 14.7510.19 ± 7.4379.5083.45
U-Net-20%61.29 ± 16.5955.65 ± 18.9012.88 ± 7.3267.1974.50
U-Net-10%54.90 ± 19.6646.89 ± 20.0514.58 ± 9.0360.5563.02
U-Net-1.0%39.02 ± 21.2232.10 ± 22.2215.90 ± 9.1243.0244.15
GAN-90%79.0 ± 8.1570.59 ± 10.899.55 ± 6.3585.0990.12
GAN-50%78.76 ± 8.9870.16 ± 11.189.88 ± 6.4484.3289.43
GAN-30%73.97 ± 10.8767.01 ± 13.0410.23 ± 6.9879.9384.97
GAN-20%69.92 ± 11.4563.65 ± 16.8811.66 ± 7.1479.1284.12
GAN-10%66.33 ± 13.2160.18 ± 19.2311.99 ± 7.8874.1278.34
GAN-1.0%62.43 ± 13.2356.43 ± 22.1213.43 ± 8.1169.1273.33
GAN+REC-90%78.78 ± 8.1171.13 ± 9.779.12 ± 6.4686.0990.23
GAN+REC-50%78.98 ± 8.8870.13 ± 11.139.78 ± 6.6685.1290.54
GAN+REC-30%74.83 ± 10.6768.67 ± 14.0610.01 ± 6.9880.1285.32
GAN+REC-20%71.14 ± 11.1866.65 ± 16.4411.34 ± 7.0580.2384.23
GAN+REC-10%69.24 ± 13.7863.23 ± 17.7111.80 ± 7.2375.1379.12
GAN+REC-1.0%64.19 ± 12.2259.33 ± 21.0112.91 ± 7.5470.3474.67
CqSL-90%83.0 ± 6.3377.77 ± 11.668.1 ± 6.0090.7895.12
CqSL-50%82.72 ± 8.2976.15 ± 11.08.21 ± 6.0488.4494.26
CqSL-30%81.59 ± 7.2073.27 ± 12.148.28 ± 6.1085.1992.62
CqSL-20%81.44 ± 6.1275.33 ± 11.528.56 ± 6.1183.1493.79
CqSL-10%79.21 ± 9.7671.45 ± 12.919.82 ± 6.7882.4090.93
CqSL-1.0%75.50 ± 10.8770.55 ± 12.589.87 ± 6.7280.5583.68
1 CqSL-90%81.88 ± 6.074.31 ± 11.658.5 ± 6.1590.1291.97
1 CqSL-50%82.03 ± 6.4575.22 ± 11.248.49 ± 6.1088.1193.44
1 CqSL-30%79.25 ± 8.1173.16 ± 8.148.77 ± 6.2283.6292.05
1 CqSL-20%80.21 ± 7.5473.19 ± 11.049.01 ± 6.3483.6991.05
1 CqSL-10%78.58 ± 9.2271.12 ± 11.259.48 ± 6.5782.2191.01
1 CqSL-1.0%73.90 ± 11.8868.58 ± 13.899.85 ± 6.7179.5484.54
2 CqSL-90%81.03 ± 7.1174.37 ± 11.488.74 ± 6.2588.3992.28
2 CqSL-50%80.65 ± 7.2673.36 ± 12.068.54 ± 6.2386.7893.05
2 CqSL-30%78.02 ± 9.3672.66 ± 10.559.35 ± 6.6582.8891.96
2 CqSL-20%79.55 ± 8.1073.0 ± 11.549.65 ± 6.6383.0289.15
2 CqSL-10%78.33 ± 8.9668.54 ± 12.899.77 ± 6.3480.5691.55
2 CqSL-1.0%71.21 ± 11.7663.45 ± 15.9111.82 ± 7.1276.4081.93
3 CqSL-90%81.13 ± 7.3373.04 ± 12.118.93 ± 6.3386.0290.17
3 CqSL-50%79.34 ± 8.5671.23 ± 12.879.05 ± 6.6684.3491.24
3 CqSL-30%76.77 ± 10.1172.04 ± 11.269.66 ± 6.7382.090.88
3 CqSL-20%79.01 ± 8.5871.89 ± 12.889.52 ± 6.4681.6687.56
3 CqSL-10%76.55 ± 8.2568.55 ± 13.2310.12 ± 6.8981.0288.72
3 CqSL-1.0%70.41 ± 11.8664.77 ± 15.7012.11 ± 7.2374.4480.21
4 CqSL-90%79.83 ± 8.2370.33 ± 12.669.25 ± 6.3484.5490.02
4 CqSL-50%79.02 ± 8.8872.68 ± 12.269.36 ± 6.2385.2090.22
4 CqSL-30%75.38 ± 9.7570.49 ± 12.09.52 ± 6.5480.3388.59
4 CqSL-20%75.77 ± 9.0569.88 ± 13.2210.19 ± 6.7781.0288.78
4 CqSL-10%72.24 ± 10.6566.70 ± 13.5610.55 ± 6.7579.7985.47
4 CqSL-1.0%68.97 ± 13.9063.19 ± 16.5012.88 ± 7.4372.1377.59
Table 2. Quantitative evaluation of LV blood pool segmentation results achieved using four semi-supervised variants of the proposed CqSL model in terms of mean Dice score (%) with std. dev., Jaccard index, Hausdorff distance (mm), precision (%) and recall (%) rates evaluated for varying proportions of labeled data on the ACDC dataset compared across several frameworks.
Table 2. Quantitative evaluation of LV blood pool segmentation results achieved using four semi-supervised variants of the proposed CqSL model in terms of mean Dice score (%) with std. dev., Jaccard index, Hausdorff distance (mm), precision (%) and recall (%) rates evaluated for varying proportions of labeled data on the ACDC dataset compared across several frameworks.
Left Ventricle (LV)
DiceJaccardHDPrec.Rec.
U-Net-90%88.03 ± 6.8185.09 ± 6.985.16 ± 5.9297.8898.79
U-Net-50%86.88 ± 6.0984.67 ± 5.365.29 ± 6.2097.0198.19
U-Net-30%82.98 ± 8.6680.10 ± 8.196.89 ± 6.7589.6691.05
U-Net-20%81.29 ± 8.9179.78 ± 9.028.22 ± 8.2387.5089.77
U-Net-10%79.49 ± 9.5671.29 ± 11.269.56 ± 9.8283.3386.14
U-Net-1.0%42.56 ± 19.7637.02 ± 21.4514.35 ± 10.1245.5346.17
GAN-90%86.15 ± 6.4581.23 ± 8.015.53 ± 5.0890.5792.87
GAN-50%85.34 ± 7.0381.26 ± 8.125.91 ± 6.0388.3489.43
GAN-30%84.03 ± 8.1680.22 ± 9.116.89 ± 7.0387.2388.87
GAN-20%81.90 ± 8.5979.12 ± 10.827.12 ± 7.3386.1988.12
GAN-10%81.78 ± 8.1676.67 ± 14.138.02 ± 7.5483.1587.43
GAN-1.0%75.02 ± 12.3270.22 ± 15.1210.89 ± 9.1280.2283.12
GAN+REC-90%88.06 ± 6.1181.94 ± 8.125.73 ± 5.2291.1993.35
GAN+REC-50%86.19 ± 6.8981.02 ± 8.235.76 ± 5.4390.5491.65
GAN+REC-30%85.53 ± 7.3680.34 ± 9.126.78 ± 6.3489.7690.34
GAN+REC-20%83.89 ± 8.1979.34 ± 10.226.88 ± 7.0587.1989.53
GAN+REC-10%83.29 ± 7.1677.56 ± 13.057.58 ± 8.3385.5589.02
GAN+REC-1.0%76.02 ± 11.2271.32 ± 14.2210.04 ± 9.1280.1284.43
CqSL-90%92.77 ± 4.9885.67 ± 7.314.53 ± 4.9896.1299.75
CqSL-50%92.25 ± 5.1283.98 ± 7.985.23 ± 5.0395.9197.95
CqSL-30%90.10 ± 5.8982.91 ± 8.125.93 ± 5.2393.5093.79
CqSL-20%88.98 ± 6.3381.26 ± 8.786.21 ± 5.0490.1492.90
CqSL-10%88.33 ± 6.3979.92 ± 9.216.17 ± 6.4489.3592.95
CqSL-1.0%83.21 ± 7.1277.94 ± 10.517.0 ± 5.9886.9691.36
1 CqSL-90%92.21 ± 5.1383.66 ± 7.454.88 ± 3.2195.0397.33
1 CqSL-50%91.0 ± 5.5581.61 ± 8.055.16 ± 4.0994.1296.13
1 CqSL-30%89.56 ± 5.9781.23 ± 7.895.89 ± 6.9892.2292.80
1 CqSL-20%87.28 ± 6.9180.32 ± 8.126.55 ± 5.2389.8991.0
1 CqSL-10%87.89 ± 6.4479.15 ± 9.306.05 ± 5.3389.0392.55
1 CqSL-1.0%81.78 ± 7.2275.36 ± 9.207.88 ± 5.4484.5589.17
2 CqSL-90%91.45 ± 5.8683.31 ± 7.234.90 ± 4.9095.1396.73
2 CqSL-50%90.22 ± 5.1280.78 ± 8.345.54 ± 4.5593.0296.04
2 CqSL-30%89.11 ± 5.8981.14 ± 8.105.88 ± 5.1191.1492.89
2 CqSL-20%87.02 ± 6.9881.12 ± 8.776.74 ± 5.2889.1190.58
2 CqSL-10%87.15 ± 6.9379.02 ± 8.876.44 ± 4.8788.5392.47
2 CqSL-1.0%80.80 ± 8.1275.06 ± 10.048.01 ± 6.1285.5490.20
3 CqSL-90%91.03 ± 5.5782.44 ± 7.875.32 ± 4.7795.3195.55
3 CqSL-50%89.79 ± 5.0279.15 ± 8.045.12 ± 5.1293.4495.18
3 CqSL-30%89.24 ± 6.1581.02 ± 7.955.71 ± 5.1892.2691.11
3 CqSL-20%88.19 ± 5.5380.52 ± 8.126.80 ± 5.0588.7889.10
3 CqSL-10%86.56 ± 6.1579.55 ± 8.456.56 ± 6.5487.9892.01
3 CqSL-1.0%79.58 ± 9.2573.20 ± 10.878.64 ± 7.0185.7791.05
4 CqSL-90%90.55 ± 5.8880.19 ± 8.256.55 ± 6.1293.1295.55
4 CqSL-50%89.10 ± 6.1579.01 ± 8.775.54 ± 5.8892.1193.22
4 CqSL-30%88.01 ± 6.4379.89 ± 8.005.86 ± 6.4391.5491.02
4 CqSL-20%87.78 ± 5.5380.13 ± 7.726.91 ± 5.1688.1790.56
4 CqSL-10%86.0 ± 6.3980.10 ± 8.906.92 ± 5.1285.6793.34
4 CqSL-1.0%78.13 ± 8.6674.19 ± 11.209.56 ± 8.0584.6689.10
Table 3. Quantitative evaluation of LV-Myocardium segmentation results achieved using four semi-supervised variants of the proposed CqSL model in terms of mean Dice score (%) with std. dev., Jaccard index, Hausdorff distance (mm), precision (%) and recall (%) evaluated for varying proportions of labeled data on the ACDC dataset compared to segmentation across several frameworks.
Table 3. Quantitative evaluation of LV-Myocardium segmentation results achieved using four semi-supervised variants of the proposed CqSL model in terms of mean Dice score (%) with std. dev., Jaccard index, Hausdorff distance (mm), precision (%) and recall (%) evaluated for varying proportions of labeled data on the ACDC dataset compared to segmentation across several frameworks.
LV-Myocardium (LV-Myo)
DiceJaccardHDPrec.Rec.
U-Net-90%86.93 ± 5.5684.50 ± 5.204.97 ± 3.7692.3296.54
U-Net-50%85.82 ± 6.3282.25 ± 7.665.16 ± 5.7790.1995.66
U-Net-30%77.29 ± 9.1975.49 ± 7.906.56 ± 5.6587.1189.56
U-Net-20%76.56 ± 9.1671.78 ± 16.207.69 ± 5.4583.5788.34
U-Net-10%66.23 ± 15.9060.63 ± 19.8710.10 ± 8.5559.3462.08
U-Net-1.0%29.47 ± 20.2925.39 ± 22.5013.95 ± 9.1232.2534.54
GAN-90%84.50 ± 6.1479.03 ± 9.175.89 ± 4.2388.1289.14
GAN-50%81.21 ± 7.4974.12 ± 11.775.45 ± 5.1485.5588.01
GAN-30%78.67 ± 9.6175.88 ± 12.755.19 ± 6.1584.3386.10
GAN-20%77.88 ± 9.8972.45 ± 15.916.01 ± 7.6583.3285.12
GAN-10%75.23 ± 11.1970.33 ± 17.197.87 ± 8.5576.4481.33
GAN-1.0%66.02 ± 20.1062.55 ± 20.8712.67 ± 9.7271.4376.23
GAN+REC-90%85.34 ± 6.4277.44 ± 12.135.34 ± 4.3788.4490.33
GAN+REC-50%82.33 ± 7.4975.16 ± 13.165.81 ± 4.7387.3289.10
GAN+REC-30%79.77 ± 9.2174.10 ± 14.775.91 ± 5.1286.7688.34
GAN+REC-20%78.43 ± 9.1173.32 ± 15.116.12 ± 6.1484.1287.43
GAN+REC-10%76.18 ± 11.1872.21 ± 15.807.23 ± 7.3479.4383.53
GAN+REC-1.0%67.52 ± 18.1264.22 ± 19.3312.12 ± 9.3472.4378.44
CqSL-90%89.33 ± 5.1182.03 ± 7.335.20 ± 5.1193.9896.01
CqSL-50%87.77 ± 6.1979.12 ± 9.05.88 ± 5.4393.3393.17
CqSL-30%85.89 ± 7.0777.72 ± 11.926.23 ± 6.1491.2092.25
CqSL-20%85.55 ± 7.2276.95 ± 12.96.85 ± 7.0490.0191.09
CqSL-10%84.14 ± 7.6472.76 ± 13.017.07 ± 8.0188.8490.88
CqSL-1.0%77.65 ± 9.2674.20 ± 11.8710.88 ± 8.4583.2288.10
1 CqSL-90%88.98 ± 6.0181.78 ± 7.636.11 ± 6.1094.1395.33
1 CqSL-50%86.55 ± 6.2278.31 ± 9.465.74 ± 5.3493.4194.11
1 CqSL-30%86.23 ± 7.6277.43 ± 11.896.43 ± 6.2991.8891.0
1 CqSL-20%85.10 ± 6.9876.09 ± 12.776.80 ± 6.2588.8791.09
1 CqSL-10%84.56 ± 8.0172.11 ± 13.548.13 ± 7.0389.7390.16
1 CqSL-1.0%75.54 ± 9.8973.01 ± 11.5610.05 ± 8.4380.8985.44
2 CqSL-90%88.44 ± 6.4381.03 ± 7.896.65 ± 5.2492.095.32
2 CqSL-50%86.01 ± 6.6979.28 ± 10.025.65 ± 5.2793.1992.66
2 CqSL-30%84.93 ± 8.0178.52 ± 11.616.88 ± 5.8690.4293.53
2 CqSL-20%85.33 ± 5.7377.11 ± 11.596.32 ± 7.3289.8292.38
2 CqSL-10%83.02 ± 8.3371.67 ± 14.048.71 ± 8.1087.7791.45
2 CqSL-1.0%75.0 ± 10.1072.55 ± 11.1810.20 ± 8.8881.0186.56
3 CqSL-90%87.33 ± 7.2280.73 ± 8.106.43 ± 5.5092.3194.52
3 CqSL-50%86.43 ± 6.3278.56 ± 10.225.76 ± 5.4091.3492.11
3 CqSL-30%83.10 ± 8.6678.15 ± 10.785.92 ± 6.1188.8291.63
3 CqSL-20%83.00 ± 6.0275.44 ± 13.106.65 ± 7.6390.3192.11
3 CqSL-10%82.88 ± 9.0172.00 ± 14.667.98 ± 8.3486.1190.87
3 CqSL-1.0%73.19 ± 11.5670.04 ± 12.9310.78 ± 8.5477.5083.39
4 CqSL-90%87.44 ± 7.7181.24 ± 7.456.12 ± 5.1191.3292.65
4 CqSL-50%86.01 ± 6.8176.12 ± 10.646.01 ± 6.1289.3291.88
4 CqSL-30%81.98 ± 10.0176.65 ± 11.445.32 ± 5.4487.1192.33
4 CqSL-20%84.01 ± 7.4475.15 ± 13.196.72 ± 6.4188.4391.66
4 CqSL-10%81.97 ± 10.6673.43 ± 13.786.69 ± 6.8784.7786.32
4 CqSL-1.0%71.21 ± 11.7669.25 ± 13.1611.82 ± 9.2375.4082.56
Table 4. Our proposed CqSL model achieves 84.9% accuracy, significantly outperforming other baselines. We incrementally add each component, aiming to study their effectiveness on the final results; (model I: only a GAN architecture (Figure 3c); model II: GAN + reconstruction (Figure 3c,d); model III: GAN + reconstruction + disentangled block (Figure 3a–d). ↑ denotes higher the value better the result; ↓ denotes lower the value better the result.
Table 4. Our proposed CqSL model achieves 84.9% accuracy, significantly outperforming other baselines. We incrementally add each component, aiming to study their effectiveness on the final results; (model I: only a GAN architecture (Figure 3c); model II: GAN + reconstruction (Figure 3c,d); model III: GAN + reconstruction + disentangled block (Figure 3a–d). ↑ denotes higher the value better the result; ↓ denotes lower the value better the result.
Average
ModelsDiceJaccardHDPrec.Rec.
Model I: GAN76.56 ± 9.9771.74 ± 14.548.26 ± 7.3782.87 ± 7.6685.78 ± 6.34
Model II: GAN + REC77.82 ± 9.8773.10 ± 13.928.11 ± 6.7483.84 ± 7.1287.06 ± 5.65
Model III: GAN + REC + DISENTANGLE (CqSL)84.92 ± 6.5577.85 ± 11.067.20 ± 6.0687.76 ± 5.4589.56± 5.04
Table 5. Image reconstruction assessment: correlation coefficient (CC) and PSNR comparison between reconstructed and input images based on 288 test sets.
Table 5. Image reconstruction assessment: correlation coefficient (CC) and PSNR comparison between reconstructed and input images based on 288 test sets.
Reconstruction Quality
CC (%)PSNR (dB)
n = 288n = 288
Model II: GAN + REC0.91227.32
Model III: GAN + REC + DISENTANGLE (Proposed)0.93428.89
Table 6. The correlation between the CqSL-predicted and ground truth clinical indices is significantly higher than the correlation between the U-Net-predicted and same ground truth clinical indices (☆☆ (p < 0.01), ☆ (p < 0.1)).
Table 6. The correlation between the CqSL-predicted and ground truth clinical indices is significantly higher than the correlation between the U-Net-predicted and same ground truth clinical indices (☆☆ (p < 0.01), ☆ (p < 0.1)).
Clinical Indices of Healthy Volunteers
UNetCqSL
LV EF0.4870.898 ☆☆
RV EF0.3710.723 ☆
Myo mass0.4270.924 ☆☆
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hasan, S.M.K.; Linte, C.A. Learning Deep Representations of Cardiac Structures for 4D Cine MRI Image Segmentation through Semi-Supervised Learning. Appl. Sci. 2022, 12, 12163. https://doi.org/10.3390/app122312163

AMA Style

Hasan SMK, Linte CA. Learning Deep Representations of Cardiac Structures for 4D Cine MRI Image Segmentation through Semi-Supervised Learning. Applied Sciences. 2022; 12(23):12163. https://doi.org/10.3390/app122312163

Chicago/Turabian Style

Hasan, S. M. Kamrul, and Cristian A. Linte. 2022. "Learning Deep Representations of Cardiac Structures for 4D Cine MRI Image Segmentation through Semi-Supervised Learning" Applied Sciences 12, no. 23: 12163. https://doi.org/10.3390/app122312163

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop