Next Article in Journal
The Impact of Attention Mechanisms on Speech Emotion Recognition
Previous Article in Journal
Methods of Projecting Mode Amplitude Changes on the Wavelength Axis in Order to Determine the Bending Radius on the Basis of TFBG Grating Spectra
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Brain MR Image Enhancement for Tumor Segmentation Using 3D U-Net

by
Faizad Ullah
1,2,
Shahab U. Ansari
2,
Muhammad Hanif
2,
Mohamed Arselene Ayari
3,4,
Muhammad Enamul Hoque Chowdhury
5,
Amith Abdullah Khandakar
5 and
Muhammad Salman Khan
1,6,*
1
Artificial Intelligence in Healthcare, Intelligent Information Processing Lab, National Center of Artificial Intelligence, University of Engineering and Technology, Peshawar 25120, Pakistan
2
Faculty of Computer Science and Engineering, Ghulam Ishaq Khan Institute of Engineering Sciences and Technology, Topi 23640, Pakistan
3
Technology Innovation and Engineering Education, College of Engineering, Qatar University, Doha 2713, Qatar
4
Department of Civil and Architectural Engineering, College of Engineering, Qatar University, Doha 2713, Qatar
5
Department of Electrical Engineering, College of Engineering, Qatar University, Doha 2713, Qatar
6
Department of Electrical Engineering (JC), University of Engineering and Technology, Peshawar 24241, Pakistan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(22), 7528; https://doi.org/10.3390/s21227528
Submission received: 5 October 2021 / Revised: 7 November 2021 / Accepted: 9 November 2021 / Published: 12 November 2021
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
MRI images are visually inspected by domain experts for the analysis and quantification of the tumorous tissues. Due to the large volumetric data, manual reporting on the images is subjective, cumbersome, and error prone. To address these problems, automatic image analysis tools are employed for tumor segmentation and other subsequent statistical analysis. However, prior to the tumor analysis and quantification, an important challenge lies in the pre-processing. In the present study, permutations of different pre-processing methods are comprehensively investigated. In particular, the study focused on Gibbs ringing artifact removal, bias field correction, intensity normalization, and adaptive histogram equalization (AHE). The pre-processed MRI data is then passed onto 3D U-Net for automatic segmentation of brain tumors. The segmentation results demonstrated the best performance with the combination of two techniques, i.e., Gibbs ringing artifact removal and bias-field correction. The proposed technique achieved mean dice score metrics of 0.91, 0.86, and 0.70 for the whole tumor, tumor core, and enhancing tumor, respectively. The testing mean dice scores achieved by the system are 0.90, 0.83, and 0.71 for the whole tumor, core tumor, and enhancing tumor, respectively. The novelty of this work concerns a robust pre-processing sequence for improving the segmentation accuracy of MR images. The proposed method overcame the testing dice scores of the state-of-the-art methods. The results are benchmarked with the existing techniques used in the Brain Tumor Segmentation Challenge (BraTS) 2018 challenge.

1. Introduction

Medical imaging plays an important role in disease identification and treatment planning. Generally, Magnetic Resonance Imaging (MRI) and Computed Tomography (CT) are used to monitor the disease progression and diagnosis. However, segmentation of affected regions and identification of disease is sometimes affected by the degraded image quality. Automatic segmentation of cancerous regions in brain MRI is a challenging task [1] due to the presence of noise in MRI generated at the time of acquisition and transmission [2,3], intensity inhomogeneity of MRI [4], variability of intensity ranges due to different vendor scanners, and capturing of non-brain tissues (eyes, spinal cord, and skull) in the brain MRI [5]. In literature, the skull stripping is considered as an important pre-processing step and, therefore, MRI datasets, such as the Multimodal Brain Tumor Segmentation Challenge (BraTS), are already skull stripped and co-registered to T1-contrast enhanced (T1ce) [6]. Moreover, MRI images are usually altered due to the bias field distortion, as a result the intensity of the same tissues varying across the image. In literature, N4ITK proposed by Tustison [7] is frequently employed for the correction of the bias field distortion problems [5]. Furthermore, the brain MRI images are also subjected to noise [2] up to some extent. The noisy MR images affect the subsequent analysis as it becomes difficult to distinguish between normal brain tissues and tumorous region. Hence, different de-noising techniques are suggested in the literature to obtain speckless brain MRI images [8]. Another significant pre-processing step in medical imaging is image registration [9]. Since brain MR images are acquired through different modalities or sequences, image registration is required to transform these images into a common coordinate space system.
In summary, pre-processing techniques such as skull stripping, de-noising, bias field correction, and registration are widely used to prepare brain MRI data for automatic brain tumor segmentation and analysis. Afterwards, the learning models are trained on this pre-processed data [5]. Especially, pre-processing is essential for automatic brain tumor segmentation as it directly influences the performance of deep learning models. This study is focused on investigation of the effects of different pre-processing techniques on the automatic brain MRI segmentation.

Contribution

The major contributions of this work are as follows:
  • A permutation of different pre-processing techniques for brain tumor segmentation is investigated and results are reported.
  • It is demonstrated analytically and confirmed experimentally that the Gibbs ringing artifact removal is an important pre-processing technique which improves segmentation performance in brain MRI.
  • Finally, a robust pre-processing framework is proposed for the automatic brain tumor segmentation.

2. Related Work

In practice, the acquired MRI images consist of both brain and non-brain tissue, i.e., skull, eyes, and spinal cord [5], and cannot be employed directly for the advanced processing. To refine the acquired MRI images, specifically brain MRI images, the commonly used pre-processing steps used in the literature [10,11,12,13,14] are discussed and a detailed literature review on the prevailing pre-processing techniques for MR images is presented in this section.
The presence of skull in brain MRI may affect the learning rate and computational complexity of the learning model. In addition, the presence of eyes and other non-brain tissues in MRI images of the training dataset may lead to misclassification and unexpected results [5]. Therefore, skull stripping is one of the most important pre-processing steps in the brain image analysis as the removal of the skull is expected to reduce the chances of misclassifications [5]. Based on the reported literature [15], skull stripping has a direct impact on the model performance which consequently affects the efficiency in the tumor detection, brain morphometry, and cortical surface reconstruction. Other commonly used methods for the skull stripping are ROBEX [16], BET [17], BSE and BEaST [5], ROBEX [16], and BET [17] which have been tested on the clinical datasets for skull stripping. Another technique known as LABEL [18] has been applied for pediatric brain MRI skull stripping. Moreover, BraTS dataset, one of the publicly available benchmarks for the brain tumor segmentation, is also skull stripped using skull-stripping filter for ITK [19]. Another important pre-processing step for brain image analysis is image registration. Image registration [20,21] is the process of aligning two or more images onto a common coordinate space (anatomical space) that enables images of the same subject to be geometrically aligned with different modalities, sequences, or across time. BraTS dataset MR images are co-registered to the same anatomical template as a pre-processing step. In practice, image registration has been actively applied on the clinical datasets [22,23].
Brain MRI volume might be acquired by different MRI scanners or by the same scanner with different protocols which causes the intensity inhomogeneity problem, i.e., different intensity ranges for the same tissue type. Intensity normalization is the process of scaling the intensities of the image to a reference or standard scale [15]. The method proposed by Nyúl et al. [24] is considered as one of the most widely used methods for the intensity normalization in which linear piecewise mapping of image intensities to a reference scale is employed. Other commonly used techniques in the literature for intensity normalization are data normalization using z-score (zero mean, unit variance) [14], and histogram matching [25].
Quite often, the brain MRI acquisition process is subjected to noises which affect quality of the obtained MRI images, and it becomes difficult to distinguish between the abnormal tissues from the normal ones. Some of the de-nosing methods include anisotropic diffusion filtering (ADF) [26], non-local means (NLM) [27], and independent component analysis (ICA) [28]. ADF is considered as one of the important techniques for de-noising brain MRI which preserves the edges of MRI and improves signal-to-noise ratio of an image. According to Liu et al. [15], noise reduction from the brain MRI is especially important in the context of deep learning framework as the presence of noise negatively impacts the model performance.
Another inherent characteristic of the MRI acquisition process is the magnetic field variation which may also have negative effects (artifacts) on the images [29]. Artifacts are the (undesired) features which show up in the acquired MR image, however, these features do not belong to the original object. These artifacts affect the image quality and cause hurdles in the brain MR image analysis. The artifacts which are caused due to variation in the magnetic field and coil are known as bias field distortion. The most common strategies in the literature to resolve the bias field distortion are the non-parametric non-uniform intensity normalization (N3) [30] and N4ITK [7]. Some other artifacts in MR images might be caused by the MR scanner itself (hardware) or due to the patient’s interaction with hardware [29]. The effects of such artifacts may be eliminated or corrected if the nature and underlying cause of the artifact is known. In this regard, expertise and familiarization with the scanner design and operation plays the key role. Obvious and overt artifacts could be identified and removed/minimized during data acquisition stage or at reconstruction stage if the underlying cause (associated with MR scanning) is known [29].
Kellner et al. [31] proposed a robust technique to remove ringing artifacts from MR images, in which the given image is re-interpolated based on local, subvoxel-shifts to sample the ringing pattern of the image at the zero-crossings of the oscillating sinc-function [31]. Gibbs ringing artifact which is caused by partial Fourier (PF) acquisition and zero filling interpolation in MRI data is thoroughly studied by Lee et al. [32] and a pipeline was developed for Removal of PF-induced Gibbs ringing (RPG) to remove ringing patterns of different periods by applying the conventional method twice. Deep learning based models [33,34] are also employed for Gibbs ringing artifact removal. Maksim et al. [33] proposed an extension of GAS-CNN (Gibbs-ringing Artifact Suppression Convolutional Neural Network) and called it attention-based convolutional neural network for Gibbs-ringing artifact removal. The method proposed by Yida et al. [34] was also employed to reduce Gibbs ringing artifacts in MRI using CNN. The network was trained over two types of images, i.e., with Gibbs ringing artifact images and without Gibbs ringing artifact images. Afterwards, the input images with Gibbs artifacts were refined (Gibbs-free) by the trained network [34]. However, this paper focuses on an important artifact removal technique, named Gibbs ringing artifact removal using local sub-voxel shift technique [31] in association with some other preprocessing techniques to enhance the MR image quality to improve the automatic brain tumor’s segmentation accuracy. The proposed methodology is detailed in the following section.

3. Proposed Methodology

This paper focuses on the investigation and analysis of different pre-processing techniques to enquire the impacts of pre-processing techniques separately and jointly on the segmentation performance of the deep learning model. In the first step, Gibbs ringing artifact removal technique is applied and its effects on the segmentation accuracy are studied. To the best of our knowledge, it is the first time that Gibbs ringing artifact removal is used as a pre-processing step for brain MR image analysis using a deep learning model. The detailed discussion about these pre-processing techniques is given in the following sections.

3.1. Dataset

BraTS focuses on evaluation of the state-of-the-art methods by providing 3D brain MR images along with ground truth labels annotated by expert physicians for the automatic segmentation of brain tumors [6]. We utilized the BraTS-2018 dataset for the assessment of different pre-processing sequences and identified the better performing one. BraTS data were collected from 19 institutions using various MRI scanners. The training dataset of BraTS-2018 comprises 285 cases (210 High Grade Glioblastomas (HGG) and 75 Low Grade Gliomas (LGG)). Each MR image is of size 240 × 240 × 155, with four 3D MRI modalities (Fluid Attenuated Inversion Recovery (FLAIR), T1, T1c, T2), rigidly aligned, skull stripped, and resampled to 1 × 1 × 1 mm isotropic resolution. There were 3 tumor subregions provided as annotations, such as enhancing tumor, peritumoral edema, and the necrotic and non-enhancing tumor core. These annotations were further combined into 3 nested subregions as whole tumor, tumor core, and enhancing tumor.

3.2. Utilized Pre-Processing Techniques

Different combinations of pre-processing techniques were explored for the performance improvement, and it was inferred that bias field distortion correction followed by Gibbs ringing artifact removal of brain MR images offers a robust pre-processing sequence for brain tumor segmentation. The results were benchmarked with the existing top-scoring techniques of the BraTS-2018, BraTS-2015, and BraTS-2013 challenges. To validate the proposed framework, other pre-processing techniques, i.e., intensity normalization, bias field correction, histogram equalization along with Gibbs ringing artifact removal were also employed.

3.2.1. Intensity Normalization

Brain MR images are usually collected from different institutions and hospitals for dataset development [6]. To address the intensity inhomogeneity problem, two different intensity normalization approaches were adopted in this study. The first approach is Nyul intensity normalization [24] and the second approach is Z-score calculation [14]. These techniques are often used for scaling the intensity to a reference intensity scale and hence the intensity normalization of medical imaging is performed [13,14].

3.2.2. Bias Field Correction

As discussed in Section 2, the bias field distortion affects the quality of the acquired MR images. As a result, the intensity of the same tissue may vary across the MR image [5]. In this study, N4ITK [7] approach was employed for the bias field correction which is briefly explained here. Firstly, the image formation model by N3 [30] is used, given as follows:
V x = u x f x + n x  
where V represents the observed image, u is uncorrupted image, f is bias field, and n denotes noise which is assumed to be the Gaussian and independent. After using the notation û = log u and assuming a noise free scenario the model becomes:
v ^ x = u ^ x + f ^ x
According to Tustison [7], the iterative scheme of N4ITK is given by:
u ^ n = u ^ n 1 S * u ^ n 1 E u ^     | u ^ n 1   f ^ r 2
where S* denotes the smoothing factor. According to [7], instead of convergence to the total bias field f(x), an iterative scheme is designed to converge, i.e., f ^ r n 0 and calculation of the total bias field is seen by inspecting the nested nature.
u ^ n = v ^ i = 1 n f ^ r . i
Thus, the total bias field estimate at the nth iteration is the sum of the first n residual bias fields, i.e.,
f ^ e n = i = 1 n f ^ r . i
Similar to Kamnitsas et al. [13], N4ITK [7] was employed on BraTS-2013, BraTS-2015, and BraTS-2018 datasets as a pre-processing step. N4ITK [7] is also applied to this work to overcome the bias field distortion and variations in the magnetic field. Sample images from BraTS-2018 dataset before and after bias field correction are shown in Figure 1.

3.2.3. Histogram Equalization

Noise removal and histogram equalization are important and classical image processing techniques to improve the quality of acquired image. In this study, the adaptive histogram equalization (AHE) [35] was used. Initially, 2D slices of 3D volumetric data were stacked to form a block. A histogram of each 2D block was then calculated to enhance the contrast of 3D images. The trilinear interpolation was finally employed to construct the 3d images back. Initially, the global histogram equalization technique was employed, however, it resulted in degradation of the MR image quality visually. Therefore, it is suggested to employ only AHE [35] to achieve the histogram equalized brain MR image.

3.2.4. Gibbs Ringing Artifact Removal

Gibbs ringing artifact is an artifact that degrades the MR image quality, which is also called truncation, spectral leakage, or ringing artifact. It typically appears near sharp high contrast boundaries in the MR images and is also known as Gibbs phenomenon [29]. Gibbs artifact usually occurs at the time of the conversion of MR signals into images using Fourier transform [29]. Skull–brain interface, cerebrospinal fluid (CSF), and spinal cord are more affected by Gibbs ringing artifact in MRI [36]. In this work, Gibbs ringing artifact removal was used as a pre-processing step. To remove the Gibbs artifact from MR image, the local sub-voxel shift [31] technique is employed which is given as follows:
j = F T 1 F T j x . G x + F T j y . G y
where FT {.} denotes the Fourier transform, Gx and Gy are two weighting functions, given as:
G x = 1 + cos k y   1 + cos k y   + 1 + cos k x  
G y = 1 + cos k x   1 + cos k y   + 1 + cos k x  
Using this approach, the high frequency components along the direction are enhanced and due to the normalization Gx + Gy = 1, where the artifact free images, i.e., Jx = Jy remain unchanged as given in Equation (6), which ensures that the minimal smoothing is introduced [31]. The results produced by Gibbs ringing artifact removal [31] are shown in Figure 2 below.
The top row of Figure 2 shows some sample images from the BraTS 2018 (extension of BraTS 2013 and BraTS 2015) dataset, which exhibits Gibbs ringing artifact. Each image in the top row has a corresponding image in the bottom row, where top row images show that these images are affected by Gibbs ringing artifact and the corresponding bottom row images are Gibbs ringing artifact removed images of the top row. We noticed that a good number of images from the BraTS dataset have the ringing artifact but for the sake of presentation, we added some of the sample images.

3.3. Proposed Framework

The proposed framework is depicted in Figure 3 in which the BraTS (2018, 2015, and 2013) datasets are pre-processed using various pre-processing techniques including bias field correction, intensity normalization, histogram equalization, and Gibbs ringing artifact removal. To train the deep learning model, the U-Net architecture was employed which was originally proposed by Ronneberger [37] for 2D images and further utilized by Isensee et al. [38] for 3D data (3D U-Net). The network (3D U-Net) is identical to Isensee et al. [38] except for the number of epochs, i.e., our network trained over 80 epochs with mini-batch size of 2 and training patch size is 80 × 80 × 80, however, total epochs in Isensee et al. [38] is 300 and the network was trained for a total of 300 epochs. One challenge of medical imaging data is class imbalance as in the BraTS-2017 training dataset there is 166 times as much data label-0 (background) as there is label-1 (enhancing tumor) [18]. Therefore, the multiclass dice loss function was employed to overcome the class imbalance problem as given in Equation (9) below.
L d c = 2 K k K n i n l i , k m i , k i n l i , k + i n m i , k
where l is the softmax output of the 3D U-Net and m is one hot encoding of the ground truth labels (segmentation maps). Additionally, i represents the voxel in training patch and K represents the tumor class (i.e., whole tumor, tumor core, and enhancing tumor in BraTS dataset). In Equation (9), li,k and mi,k denote the ground truths and output (softmax output) at voxel i of class k, respectively, where i belongs to I (voxel in training patch) and k belongs to K (one of the tumor classes). The fully convolutional nature of our network, i.e., 3D U-Net, allows the use of arbitrary-size images as we segment the entire patient data at once. In the first stage, several pre-processing techniques were applied in different sequences (separately and jointly). Based on experimental analysis, it was inferred that each pre-processing technique has a significant effect on the segmentation accuracy of brain MRI. However, correcting the brain MRI from bias field distortion (using (4)) followed by Gibbs ringing artifact removal gives us improved and generalized results as shown in (11). Brain MR Images were corrected from bias field using Equation (10).
I ^ n = v ^ i = 1 n f ^ r i
In (10), n is a bias field corrected MR image using N4ITK [7]. To remove the Gibbs ringing artifact from brain MR image using local sub-voxel shift [31] technique, Equation (6) was employed. In same sequence, the whole dataset (BraTS-2018) was pre-processed using (11) and network was trained.
J = F T 1 F T Î x . G x + F T Î y . G y  
In (11), FT {.} denotes the Fourier transform, Î denotes a bias field corrected MR image, Gx and Gy are weighting functions as given in (7) and (8), and J is the resultant MR image. The complete BraTS-2018 dataset was pre-processed using (10) and (11), i.e., bias field correction followed by Gibbs ringing artifact removal, which we called sequence-6, as there were 10 sequences in total which are discussed in later sections.

4. Results and Discussion

In this work, different pre-processing techniques for automatic brain tumor segmentation in brain MR images were evaluated. Experiments were performed on BraTS (2018, 2015 and 2013) datasets and 10 different sequences were evaluated in this work.

4.1. Gibbs Artifact Removed vs. Gibbs Not Removed Results

Gibbs ringing artifact removed MR images vs. original images, i.e., the images in BraTS-(2018, 2015, and 2013) were used initially for training the 3D U-Net and results are recorded in Table 1. The purpose of showing the results in Table 1 (Gibbs ranging artifact removed vs. Gibbs ringing artifact not removed) was to examine the importance of Gibbs ringing artifact removal as a pre-processing step. Hence, we can argue that Gibbs removal can be used as a pre-processing step to improve the segmentation accuracy. However, Gibbs removal technique did not improve the core and enhancing region’s accuracy on the BraTS-2013 dataset (Table 1), where we assumed that the size of BraTS-2013 dataset was less than that of BraTS-2015 and BraTS-2018 datasets, generally, the greater size inputs produce the improved results on identical networks.
To confirm and validate the importance of Gibbs ringing artifact removal as a pre-processing technique, original data results (produced by U-Net) compared with pre-processed data, i.e., Gibbs ringing artifact removed along with others pre-processing techniques results, are shown in Table 2. In Table 2, the results produced by Gibbs ringing artifact removed data for training 3D U-Net perform well and overcome the results produced by original/non-removed Gibbs ringing artifacts data. The improved dice score for each column is shown in bold, where we can clearly examine the importance of Gibbs ringing artifact removal as a pre-processing step for training on brain MR images.

4.2. Training and Testing Results

The training and testing data split were 80% and 20%, respectively. For results comparison, the standard BraTS metric was used, i.e., dice score. A total of 10 sequences were used for training and testing purposes. All sequence results, i.e., training and testing are shown in Table 3, in which serial no. 1 to 8 results are on the BraTS-2018 dataset while serial no. 9 and serial no. 10 are on the BraTS-2015 and BraTS-2013, respectively. The proposed framework was evaluated on testing data for prediction purposes, where we analyzed that the model is generalized well, as testing results of the model witnessed the generalization of the model as shown in Table 2 and Table 3.
The sequences (order of pre-processing techniques) along with training and testing dice score of each sequence are shown in Table 2, where (✔) and (✖) mean that the corresponding technique was employed to the sequence and not employed to the sequence, respectively, while (✔ *) shows that this technique was employed as pre-processing technique prior to the Gibbs ringing artifact removal technique. In Table 2, Z-score and Nyul means that these two techniques were used for intensity normalization. Intra-sequence results were then compared to analyze the effects of each pre-processing technique. In sequence-1, BraTS-2018 dataset was pre-processed by Gibbs ringing artifact removal and trained using the 3D U-Net. In sequence-2, the BraTS-2018 dataset was pre-processed by two different techniques, i.e., Gibbs ringing artifact removal followed by bias field correction (N4ITK). After that, the 3D U-Net was trained over the data and results were calculated.
In sequence-3, the BraTS-2018 dataset was again pre-processed by two different pre-processing techniques, however, this time AHE replaced the N4ITK [7] technique as a pre-processing technique and results were calculated. In sequence-4 the same data, i.e., BraTS-2018 dataset was pre-processed by three different techniques. This time all three above techniques sequentially pre-processed the data and fed to the network and results were calculated. Sequence-5 comprises two pre-processing techniques, i.e., Gibbs ringing artifact removal and Z-score calculation. In this sequence, after Gibbs ringing artifact removal and Z-score calculation the data was fed to the network to record the results. The most accurate results were produced by the proposed sequence, i.e., sequence-6 in which the Brats-2018 was initially corrected from the bias field distortion using N4ITK [7], and afterwards the Gibbs ringing artifact removal was performed (on bias field corrected data).
The dataset was fed to network and results were recorded after training. Sequence-7 comprises three pre-processing steps, i.e., N4ITK [7], Gibbs ringing artifact removal, and Z-score calculation sequentially. The network was trained on the pre-processed data and results were stored. In sequence-8, the same data, i.e., BraTS-2018 data, was pre-processed sequentially by N4ITK [7], Gibbs ringing artifact removal, and Nyul intensity normalization technique. The network was trained, and results were recorded. BraTS-2015 and BraTS-2013 datasets were also evaluated in this work. In sequence-9 the BraTS-2015 dataset was pre-processed by Gibbs ringing artifact removal and results were calculated after network training. Finally, in sequence-10 the BraTS-2013 dataset was pre-processed by Gibbs-ringing artifact removal and after that the network was trained on the pre-processed data to store the results.

4.3. Comparison with Existing Techniques

The proposed methodology (sequence 6), i.e., bias field corrected MRI data followed by Gibbs ringing artifact removal techniques [31] was compared with different published articles on BraTS datasets. The proposed method, i.e., sequence-6 outperformed the state-of-the-art methods on whole and core tumor classes segmentation. However, in enhancing tumor the model did not perform well, roughly it produced the same results as other existing techniques. The enhancing region is generally the region of tumor in brain which lies between other sharpened regions, i.e., whole tumor and core tumor. Therefore, we can assume that this region is affected little by Gibbs ringing artifact while whole region is affected more. We noticed that the training and testing results of the proposed sequence show that the model was generalized well, as the training and testing mean dice scores were close to each other. The results produced by the sequence-6 were compared with the existing techniques in Table 3, where state-of-the-art results were also compared with the existing technique.

4.4. Prediction Results

The permutations of different pre-processing techniques used in this work and their prediction results are depicted in Figure 4. As explained earlier, those 10 different permutations of pre-processing techniques were investigated. For prediction purposes, random images of FLAIR sequence were picked up from the BraTS-2018 dataset. For comparison, BraTS-2018 results were used, i.e., sequence-1 to sequence-8. To confirm the confidence of proposed sequence (sequence 6), the prediction result is shown in Figure 4. FLAIR modality along with the corresponding ground truth label for each patient was segmented via each sequence (1–8). The results show that the sequence-6 has more accurate results produced by other sequences.

5. Conclusions and Future Work

Pre-processing the input training data is an essential step for training the deep learning model for automatic brain tumor segmentation. In the present work, different pre-processing techniques, i.e., bias field correction, intensity normalization, histogram equalization, and Gibbs ringing artifact removal were employed to analyze the effect of pre-processing techniques and find out the optimal pre-processing technique (sequence-6) to train the deep network, i.e., 3D U-Net. We also investigated whether bias field correction followed by Gibbs ringing artifact removal gives us improved segmentation results.
To the best of our knowledge, using the Gibbs ringing artifact removal as a pre-processing technique to train the deep learning model was first used by this work. To confirm the confidence of Gibbs ringing artifact removal as a pre-processing technique, Gibbs ringing artifact removal was employed before bias field correction and after bias field correction. It was inferred that the results of Gibbs ringing artifact removal as a preprocessing technique are promising regarding the improvement of segmentation accuracy. Therefore, the use of Gibbs ringing artifact removal is strongly recommended as a pre-processing technique and also segmentation results are improved when Gibbs ringing artifact removal is followed by bias field correction (N4ITK).
It can be noted that Gibbs ringing artifact removal from MR images produced improved results on whole and core regions, however, this technique did not improve the the enhancing region’s accuracy on the BraTS dataset (Table 3). The enhancing region is generally the region of tumor/cancer in the brain which progresses/grows in between whole tumor and core tumor. Therefore, this region may not be affected more by Gibbs ringing artifact as the borders/regions are not sharpened in the enhancing region, because the enhancing region lies between two other sharpened regions of brain. Additionally, all cited works to date have less accuracy on enhancing regions as compared to whole and core tumors.
In this work, the Gibbs ringing artifact removal was applied to brain MR images with other pre-processing techniques to train the model. While, in future, this work can be further extended to apply the Gibbs ringing artifact removal to other parts (organs) of the body acquired with MRI modality.

Author Contributions

Conceptualization, M.H. and M.S.K.; data curation, F.U. and S.U.A.; formal analysis, F.U.; investigation, S.U.A.; methodology, F.U.; project administration, M.S.K.; resources, S.U.A.; software, F.U.; supervision, A.A.K. and M.S.K.; validation, F.U.; visualization, M.E.H.C.; writing—original draft, S.U.A., M.A.A. and M.E.H.C.; writing—review and editing, F.U., S.U.A., M.H., M.A.A., A.A.K. and M.S.K. All authors have read and agreed to the published version of the manuscript.

Funding

The authors also thank the Higher Education Commission (HEC), Pakistan, for funding this research under the Artificial Intelligence in Healthcare, IIPL, National Center of Artificial Intelligence (NCAI) and partial funding by HEC-SRGP research grant.

Institutional Review Board Statement

Not applicable as this study was carried out on a public dataset and the original study was carried out with the ethical approval from Institutional Review Board.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study when the data publishing group acquired the data.

Data Availability Statement

MICCAI BraTS 2018 data was used: https://www.med.upenn.edu/sbia/brats2018/data.html (accessed on 7 November 2021).

Acknowledgments

The authors gratefully acknowledge the support of NVIDIA Corporation with the donation of the Titan X Pascal used for this research.

Conflicts of Interest

The authors declare that they have no conflict of interest to report regarding the present study.

References

  1. Fang, L.; Wang, X.; Wang, L. Multi-modal medical image segmentation based on vector-valued active contour models. Inf. Sci. 2020, 513, 504–518. [Google Scholar] [CrossRef]
  2. Sheela, C.J.J.; Suganthi, G. An efficient denoising of impulse noise from MRI using adaptive switching modified decision based unsymmetric trimmed median filter. Biomed. Signal Process. Control 2020, 55, 101657. [Google Scholar] [CrossRef]
  3. Wang, X.; Shi, G.; Zhang, P.; Wu, J.; Li, F.; Wang, Y.; Jiang, H. High quality impulse noise removal via non uniform sampling and autoregressive modelling based super-resolution. IET Image Process. 2016, 10, 304–313. [Google Scholar] [CrossRef]
  4. Sulochana, C.H.; Blessy, S.P.S. Intensity Inhomogeneity Correction in Brain MR Images Based on Filtering Method. In Handbook of Research on Applications and Implementations of Machine Learning Techniques; IGI Global: Derry Townshi, PA, USA, 2020; pp. 96–110. [Google Scholar] [CrossRef] [Green Version]
  5. Bernal, J.; Kushibar, K.; Asfaw, D.S.; Valverde, S.; Oliver, A.; Martí, R.; Lladó, X. Deep convolutional neural networks for brain image analysis on magnetic resonance imaging: A review. Artif. Intell. Med. 2019, 95, 64–81. [Google Scholar] [CrossRef] [Green Version]
  6. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef]
  7. Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 bias correction. IEEE Trans. Med. Imaging 2010, 29, 1310–1320. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Sarkar, S.; Tripathi, P.C.; Bag, S. An Improved Non-local Means Denoising Technique for Brain MRI. In Computational Intelligence in Pattern Recognition; Springer: Singapore, 2020; pp. 765–773. [Google Scholar] [CrossRef]
  9. Ni, H.; Tan, C.; Feng, Z.; Chen, S.; Zhang, Z.; Li, W.; Guan, Y.; Gong, H.; Luo, Q.; Li, A. A robust image registration interface for large volume brain atlas. Sci. Rep. 2020, 10, 1–16. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Mlynarski, P.; Delingette, H.; Criminisi, A.; Ayache, N. 3D convolutional neural networks for tumor segmentation using long-range 2D context. Comput. Med. Imaging Graph. 2019, 73, 60–72. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Sun, L.; Zhang, S.; Luo, L. Tumor segmentation and survival prediction in glioma with deep learning. In International MICCAI Brainlesion Workshop; Springer: Cham, Switzerland, 2018; pp. 83–93. [Google Scholar] [CrossRef]
  12. Weninger, L.; Rippel, O.; Koppers, S.; Merhof, D. Segmentation of brain tumors and patient survival prediction: Methods for the brats 2018 challenge. In International MICCAI Brainlesion Workshop; Springer: Cham, Switzerland, 2018; pp. 3–12. [Google Scholar] [CrossRef]
  13. Kamnitsas, K.; Ledig, C.; Newcombe, V.F.; Simpson, J.P.; Kane, A.D.; Menon, D.K.; Rueckert, D.; Glocker, B. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017, 36, 61–78. [Google Scholar] [CrossRef]
  14. Pereira, S.; Pinto, A.; Alves, V.; Silva, C.A. Brain tumor segmentation using convolutional neural networks in MRI images. IEEE Trans. Med. Imaging 2016, 35, 1240–1251. [Google Scholar] [CrossRef]
  15. Liu, J.; Li, M.; Wang, J.; Wu, F.; Liu, T.; Pan, Y. A survey of MRI-based brain tumor segmentation methods. Tsinghua Sci. Technol. 2014, 19, 578–595. [Google Scholar] [CrossRef]
  16. Iglesias, J.E.; Liu, C.-Y.; Thompson, P.M.; Tu, Z. Robust brain extraction across datasets and comparison with publicly available methods. IEEE Trans. Med. Imaging 2011, 30, 1617–1634. [Google Scholar] [CrossRef]
  17. Smith, S.M. Fast robust automated brain extraction. Hum. Brain Mapp. 2002, 17, 143–155. [Google Scholar] [CrossRef]
  18. Shi, F.; Wang, L.; Dai, Y.; Gilmore, J.H.; Lin, W.; Shen, D. LABEL: Pediatric brain extraction using learning-based meta-algorithm. Neuroimage 2012, 62, 1975–1986. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Bauer, S.; Fejes, T.; Reyes, M. A skull-stripping filter for ITK. Insight J. 2013, 1–7. [Google Scholar] [CrossRef]
  20. Toga, A.W.; Thompson, P.M. The role of image registration in brain mapping. Image Vis. Comput. 2001, 19, 3–24. [Google Scholar] [CrossRef] [Green Version]
  21. Akkus, Z.; Galimzianova, A.; Hoogi, A.; Rubin, D.L.; Erickson, B.J. Deep learning for brain MRI segmentation: State of the art and future directions. J. Digit. Imaging 2017, 30, 449–459. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Reza, S.; Iftekharuddin, K.M. Improved brain tumor tissue segmentation using texture features. In Proceedings of the MICCAI Brain Tumor Segmentation Challenge (BraTS), Boston, MA, USA, 14–18 September 2014; pp. 27–30. [Google Scholar] [CrossRef] [Green Version]
  23. Zheng, S.; Jayasumana, S.; Romera-Paredes, B.; Vineet, V.; Su, Z.; Du, D.; Huang, C.; Torr, P.H. Conditional random fields as recurrent neural networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1529–1537. [Google Scholar] [CrossRef] [Green Version]
  24. Nyúl, L.G.; Udupa, J.K.; Zhang, X. New variants of a method of MRI scale standardization. IEEE Trans. Med. Imaging 2000, 19, 143–150. [Google Scholar] [CrossRef]
  25. Sun, X.; Shi, L.; Luo, Y.; Yang, W.; Li, H.; Liang, P.; Li, K.; Mok, V.C.; Chu, W.C.; Wang, D. Histogram-based normalization technique on human brain magnetic resonance images from different acquisitions. Biomed. Eng. Online 2015, 14, 73. [Google Scholar] [CrossRef] [Green Version]
  26. Krissian, K.; Aja-Fernández, S. Noise-driven anisotropic diffusion filtering of MRI. IEEE Trans. Image Process. 2009, 18, 2265–2274. [Google Scholar] [CrossRef]
  27. Manjón, J.V.; Carbonell-Caballero, J.; Lull, J.J.; García-Martí, G.; Martí-Bonmatí, L.; Robles, M. MRI denoising using non-local means. Med. Image Anal. 2008, 12, 514–523. [Google Scholar] [CrossRef] [PubMed]
  28. Bartlett, M.S.; Movellan, J.R.; Sejnowski, T.J. Face recognition by independent component analysis. IEEE Trans. Neural Netw. 2002, 13, 1450–1464. [Google Scholar] [CrossRef]
  29. Bellon, E.M.; Haacke, E.M.; Coleman, P.E.; Sacco, D.C.; Steiger, D.A.; Gangarosa, R.E. MR artifacts: A review. Am. J. Roentgenol. 1986, 147, 1271–1281. [Google Scholar] [CrossRef] [Green Version]
  30. Sled, J.G.; Zijdenbos, A.P.; Evans, A.C. A nonparametric method for automatic correction of intensity nonuniformity in MRI data. IEEE Trans. Med. Imaging 1998, 17, 87–97. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Kellner, E.; Dhital, B.; Kiselev, V.G.; Reisert, M. Gibbs-ringing artifact removal based on local subvoxel-shifts. Magn. Reson. Med. 2016, 76, 1574–1581. [Google Scholar] [CrossRef] [Green Version]
  32. Lee, H.H.; Novikov, D.S.; Fieremans, E. Removal of partial Fourier-induced Gibbs (RPG) ringing artifacts in MRI. Magn. Reson. Med. 2021. [Google Scholar] [CrossRef]
  33. Penkin, M.; Krylov, A.; Khvostikov, A. Attention-based Convolutional Neural Network for MRI Gibbs-ringing Artifact Suppression. In Proceedings of the 30th International Conference on Computer Graphics and Machine Vision (GraphiCon 2020), Saint Petersburg, Russia, 22–25 September 2020; p. 34. [Google Scholar] [CrossRef]
  34. Yida, W.; Yang, S.; Haibin, X.; Wenjing LBingwen, H.; Guang, Y. Reduction of Gibbs artifacts in magnetic resonance imaging based on Convolutional Neural Network. In Proceedings of the 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Shanghai, China, 14–16 October 2017; pp. 1–5. [Google Scholar] [CrossRef]
  35. Amorim, P.; Moraes, T.; Silva, J.; Pedrini, H. 3D Adaptive Histogram Equalization Method for Medical Volumes. In Proceedings of the 13th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Funchal, Portugal, 27–29 January 2018. [Google Scholar] [CrossRef]
  36. Berg, E.S. Effect of Free Water Elimination on Diffusion Indices in Healthy Volunteers and Patients with Multiple Sclerosis. Master’s Thesis, The University of Bergen, Bergen, Norway, 2018. [Google Scholar]
  37. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2015. [Google Scholar] [CrossRef] [Green Version]
  38. Isensee, F.; Kickingereder, P.; Wick, W.; Bendszus, M.; Maier-Hein, K.H. Brain tumor segmentation and radiomics survival prediction: Contribution to the brats 2017 challenge. In International MICCAI Brainlesion Workshop 2017; Springer: Cham, Switzerland, 2017. [Google Scholar] [CrossRef] [Green Version]
  39. Hussain, S.; Anwar, S.M.; Majid, M. Brain tumor segmentation using cascaded deep convolutional neural network. In Proceedings of the 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju, Korea, 11–15 July 2017; pp. 1998–2001. [Google Scholar] [CrossRef]
  40. Dong, H.; Yang, G.; Liu, F.; Mo, Y.; Guo, Y. Automatic brain tumor detection and segmentation using u-net based fully convolutional networks. In Annual Conference on Medical Image Understanding and Analysis; Springer: Cham, Switzerland, 2017; pp. 506–517. [Google Scholar] [CrossRef] [Green Version]
  41. Chen, W.; Liu, B.; Peng, S.; Sun, J.; Qiao, X. S3D-UNet: Separable 3D U-Net for brain tumor segmentation. In International MICCAI Brainlesion Workshop; Springer: Cham, Switzerland, 2019; pp. 358–368. [Google Scholar] [CrossRef]
  42. Kao, P.Y.; Ngo, T.; Zhang, A.; Chen, J.W.; Manjunath, B.S. Brain tumor segmentation and tractographic feature extraction from structural MR Images for overall survival prediction. In International MICCAI Brainlesion Workshop; Springer: Cham, Switzerland, 2018; pp. 128–141. [Google Scholar] [CrossRef] [Green Version]
  43. Feng, X.; Tustison, N.; Meyer, C. Brain tumor segmentation using an ensemble of 3d u-nets and overall survival prediction using radiomic features. In International MICCAI Brainlesion Workshop; Springer: Cham, Switzerland, 2018; pp. 279–288. [Google Scholar] [CrossRef] [Green Version]
  44. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.; Larochelle, H. Brain tumor segmentation with deep neural networks. Med. Image Anal. 2017, 38, 18–31. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Myronenko, A. 3D MRI brain tumor segmentation using autoencoder regularization. In International MICCAI Brainlesion Workshop 2019; Springer: Cham, Switzerland, 2019; pp. 311–320. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Before (top row) and after (bottom row) bias field correction. Each MR image in the top row (ac) has the corresponding bias field corrected MR image (df) in bottom row. All the images shown here are FLAIR (for fair comparison).
Figure 1. Before (top row) and after (bottom row) bias field correction. Each MR image in the top row (ac) has the corresponding bias field corrected MR image (df) in bottom row. All the images shown here are FLAIR (for fair comparison).
Sensors 21 07528 g001
Figure 2. BraTS MR images, before (top row) and after (bottom row) Gibbs ringing artifact removal. Each MR image in top row (ae) has the corresponding Gibbs ringing artifact-removed MR image (fj) in the bottom row. All the images shown here are T1ce-saggital (for fair comparison). However, other sequences, i.e., FLAIR, T1 and T2 can also be addressed for ringing artifacts.
Figure 2. BraTS MR images, before (top row) and after (bottom row) Gibbs ringing artifact removal. Each MR image in top row (ae) has the corresponding Gibbs ringing artifact-removed MR image (fj) in the bottom row. All the images shown here are T1ce-saggital (for fair comparison). However, other sequences, i.e., FLAIR, T1 and T2 can also be addressed for ringing artifacts.
Sensors 21 07528 g002
Figure 3. Proposed framework showing the (a) training and (b) testing model.
Figure 3. Proposed framework showing the (a) training and (b) testing model.
Sensors 21 07528 g003
Figure 4. Comparison of prediction (segmentation) results produced by the sequences (1–8) used in this work. Each row (ag) has MR images, ground truth labels and their corresponding segmentation results (from left to right).
Figure 4. Comparison of prediction (segmentation) results produced by the sequences (1–8) used in this work. Each row (ag) has MR images, ground truth labels and their corresponding segmentation results (from left to right).
Sensors 21 07528 g004
Table 1. Comparison of brain MRI images with (✖) Gibbs ringing artifact and without (✔) Gibbs ringing artifact using local sub voxel-shift method.
Table 1. Comparison of brain MRI images with (✖) Gibbs ringing artifact and without (✔) Gibbs ringing artifact using local sub voxel-shift method.
No.BraTS
Dataset
Gibbs Ringing
Artifact Removed
Dice Score
WholeCoreEnhancing
120180.880.820.67
220180.910.850.67
320150.850.710.57
420150.900.820.57
520130.840.820.59
620130.900.810.58
Table 2. Total sequences investigated in this work along with the training and testing dice scores. Sequence-1 to sequence-8 performed on BraTS-2018 while sequence-9 and sequence-10 performed over BaTS-2015 an BraTS-2013, respectively. * shows that this technique was employed as pre-processing technique prior to the Gibbs ringing artifact removal technique.
Table 2. Total sequences investigated in this work along with the training and testing dice scores. Sequence-1 to sequence-8 performed on BraTS-2018 while sequence-9 and sequence-10 performed over BaTS-2015 an BraTS-2013, respectively. * shows that this technique was employed as pre-processing technique prior to the Gibbs ringing artifact removal technique.
Sequence NoGibbs Artifact
Removed
Bias Field
Corrected
Intensity
Normalized
AHEMean Dice Score (Training and Testing)
WholeCoreEnhancing
TrainingTestingTrainingTestingTrainingTesting
Seq.10.910.900.850.800.670.60
Seq.20.880.880.830.780.680.57
Seq.30.860.890.750.790.730.59
Seq.40.870.810.730.660.640.53
Seq.5Z-score0.840.860.680.760.630.56
Seq.6✔ *0.910.900.860.830.700.71
Seq.7✔ *Z-score0.810.800.740.680.570.54
Seq.8✔ *Nyul0.860.880.780.800.690.56
Seq.90.900.890.820.760.570.53
Seq.100.900.850.810.720.580.52
Table 3. Comparison of the proposed methodology with existing techniques.
Table 3. Comparison of the proposed methodology with existing techniques.
S. NoReferenceBraTS
Dataset
Mean Dice Score (Training)Mean Dice Score (Testing)
WholeCoreEnhancingWholeCoreEnhancing
1[39]20130.800.670.85---
2[40]20150.860.860.65---
3[12]20180.860.810.760.840.720.62
4[38]20170.890.790.730.850.770.64
5[41]20180.890.860.680.830.780.68
6[13]20150.900.750.730.850.670.63
7[42]20180.900.810.730.870.790.74
8[43]20180.900.830.79---
9[11]20180.900.840.800.870.790.71
10[44]2013---0.880.790.73
11[23]2013---0.880.830.77
12[45]20180.910.860.820.880.810.76
13Proposed method20180.910.860.700.900.830.71
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ullah, F.; Ansari, S.U.; Hanif, M.; Ayari, M.A.; Chowdhury, M.E.H.; Khandakar, A.A.; Khan, M.S. Brain MR Image Enhancement for Tumor Segmentation Using 3D U-Net. Sensors 2021, 21, 7528. https://doi.org/10.3390/s21227528

AMA Style

Ullah F, Ansari SU, Hanif M, Ayari MA, Chowdhury MEH, Khandakar AA, Khan MS. Brain MR Image Enhancement for Tumor Segmentation Using 3D U-Net. Sensors. 2021; 21(22):7528. https://doi.org/10.3390/s21227528

Chicago/Turabian Style

Ullah, Faizad, Shahab U. Ansari, Muhammad Hanif, Mohamed Arselene Ayari, Muhammad Enamul Hoque Chowdhury, Amith Abdullah Khandakar, and Muhammad Salman Khan. 2021. "Brain MR Image Enhancement for Tumor Segmentation Using 3D U-Net" Sensors 21, no. 22: 7528. https://doi.org/10.3390/s21227528

APA Style

Ullah, F., Ansari, S. U., Hanif, M., Ayari, M. A., Chowdhury, M. E. H., Khandakar, A. A., & Khan, M. S. (2021). Brain MR Image Enhancement for Tumor Segmentation Using 3D U-Net. Sensors, 21(22), 7528. https://doi.org/10.3390/s21227528

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop