Next Article in Journal
A Twenty-Year Retrospective Analysis of Risk Assessment of Biomechanical Overload of the Upper Limbs in Multiple Occupational Settings: Comparison of Different Ergonomic Methods
Next Article in Special Issue
Adversarial Attack and Defense in Breast Cancer Deep Learning Systems
Previous Article in Journal
Liquid Metal-Based Electrode Array for Neural Signal Recording
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MultiResUNet3+: A Full-Scale Connected Multi-Residual UNet Model to Denoise Electrooculogram and Electromyogram Artifacts from Corrupted Electroencephalogram Signals

by
Md Shafayet Hossain
1,
Sakib Mahmud
2,
Amith Khandakar
2,
Nasser Al-Emadi
2,
Farhana Ahmed Chowdhury
3,
Zaid Bin Mahbub
4,*,
Mamun Bin Ibne Reaz
1,5 and
Muhammad E. H. Chowdhury
2,*
1
Department of Electrical, Electronic and Systems Engineering, Universiti Kebangsaan Malaysia, Bangi 43600, Malaysia
2
Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
3
Department of Electronics and Telecommunication Engineering, Rajshahi University of Engineering and Technology, Rajshahi 6204, Bangladesh
4
Department of Mathematics and Physics, North South University, Dhaka 1229, Bangladesh
5
Department of Electrical and Electronic Engineering, Independent University, Bashundhara, Dhaka 1229, Bangladesh
*
Authors to whom correspondence should be addressed.
Bioengineering 2023, 10(5), 579; https://doi.org/10.3390/bioengineering10050579
Submission received: 5 April 2023 / Revised: 26 April 2023 / Accepted: 4 May 2023 / Published: 10 May 2023
(This article belongs to the Special Issue Machine Learning and Signal Processing for Biomedical Applications)

Abstract

:
Electroencephalogram (EEG) signals immensely suffer from several physiological artifacts, including electrooculogram (EOG), electromyogram (EMG), and electrocardiogram (ECG) artifacts, which must be removed to ensure EEG’s usability. This paper proposes a novel one-dimensional convolutional neural network (1D-CNN), i.e., MultiResUNet3+, to denoise physiological artifacts from corrupted EEG. A publicly available dataset containing clean EEG, EOG, and EMG segments is used to generate semi-synthetic noisy EEG to train, validate and test the proposed MultiResUNet3+, along with four other 1D-CNN models (FPN, UNet, MCGUNet, LinkNet). Adopting a five-fold cross-validation technique, all five models’ performance is measured by estimating temporal and spectral percentage reduction in artifacts, temporal and spectral relative root mean squared error, and average power ratio of each of the five EEG bands to whole spectra. The proposed MultiResUNet3+ achieved the highest temporal and spectral percentage reduction of 94.82% and 92.84%, respectively, in EOG artifacts removal from EOG-contaminated EEG. Moreover, compared to the other four 1D-segmentation models, the proposed MultiResUNet3+ eliminated 83.21% of the spectral artifacts from the EMG-corrupted EEG, which is also the highest. In most situations, our proposed model performed better than the other four 1D-CNN models, evident by the computed performance evaluation metrics.

1. Introduction

The electrophysiological activity of the cerebral cortex of the human brain is represented by electroencephalogram (EEG) signals, which are non-invasively recorded at the scalp [1]. EEG is crucial for various therapeutic applications as well as neurological research. Long-duration human epileptic seizure episodes are routinely detected with EEG [2,3]. EEG is also widely used for other purposes, such as Alzheimer’s disease diagnosis [4,5], sleep stages measurement [6], assessment of cognitive workload [7], recognition of human emotion [8], establishing brain-computer interfaces (BCIs) [9], biometric systems [10], etc. EEG recordings, nevertheless, greatly suffer from a range of physiological and non-physiological artifacts, such as ocular/electrooculogram (EOG) artifacts [11], myogenic/electromyogram (EMG) noises [12], cardiac/electrocardiogram (ECG) abnormalities [13,14], power line noise and motion artifacts [15,16,17]. These noises will significantly impact the EEG data analysis outcome and may even lead to an entirely wrong diagnosis in the worst cases. To remove the physiological artifacts and maintain the neural information of EEG as much as possible, it is required to develop efficient, reliable, and robust algorithms/models.
One frequently used strategy to remove artifacts from EEG data involved converting the signal from the time domain to the frequency domain using the Fourier Transform (FT) and filtering out the spectral components corresponding to the artifacts. After discarding the noise component/s and taking the inverse FT, a noise-free signal can be obtained. EEG denoising was also carried out using filters, including the adaptive, Wiener, and Kalman filters [18,19,20]. Other signal processing-based techniques, such as the Hilbert–Huang Transformation (HHT) [21,22], empirical mode decomposition (EMD) [23,24], variational mode decomposition (VMD) [16], wavelet packet decomposition (WPD) [17], independent component analysis (ICA) [25,26], and canonical correlation analysis (CCA) [27], etc., were also utilized. These methods primarily relied on linear transformation and several presumptions, which is a severe limitation—particularly, capturing artifactual signals as the reference is an essential requirement for all types of adaptive filtering. Therefore, the denoising performance could be relatively subpar without the reference signal. The HHT-based artifact removal technique makes the false assumption that the noisy signals or components encoded in EEG have specific time properties from the other components. In [22,28], via an adaptable manner, the EEG signals were decomposed into intrinsic mode functions (IMFs), and the instantaneous frequency of the IMF was computed using HHT. The IMFs with significant deviations from other IMFs in terms of instantaneous frequency were discarded, considering them as artifact components. In [29,30], EEG signals were divided into several modes for both EMD-based and ICA-based techniques, and the noise-related components were subsequently removed using some predetermined criteria. The elimination of EMG artifacts typically involved CCA-based EEG denoising techniques, assuming that the muscle artifacts had infrequent stereotyped topographies and low autocorrelation. The muscle artifacts to be eliminated were determined by the CCA, which decomposes the EEG data into two or more uncorrelated components [15]. Some hybrid techniques like Ensemble EMD (EEMD) in combination with ICA (EEMD-ICA) [31] and EEMD-CCA [32] were investigated to remove EEG noises. However, the necessity of prior assumptions remains unresolved despite performance boosts compared to other conventional approaches. As an illustration, the value of autocorrelation thresholds in the EEMD-CCA needs to be chosen empirically through trial and error [33].
Recently, deep learning (DL) has garnered increasing interest. The effectiveness of DL-based models has significantly improved because of the growth of computing resources, the continuous development of novel network designs, and the processing capability of vast amounts of data, etc. DL has been successfully used to address various technical issues, including image processing [34,35,36] and natural language processing [37,38,39]. EEG-based motor imagery classification [40,41,42], EEG reconstruction [43,44], and EEG signal creation [45,46] are some noteworthy examples of EEG-related analysis where DL techniques have been utilized. Some established DL models, such as auto-encoder [47,48], simple and complex convolutional neural networks (CNNs) with/without residual connections [49,50], recurrent neural networks (RNNs) [50,51], and generative adversarial networks (GANs) [52], have also been used to remove artifacts from contaminated EEG. To disentangle neural signals from artifacts in the embedding space and recreate the noise-free signal, a new deep learning framework called DeepSeparator has been introduced [53]. Although significant progress has been made in the denoising of EEG through the utilization of DL-based methods, there are still some drawbacks in the studies reported earlier, such as lack of robust framework, non-usage of k-fold cross-validation methods, testing with a selective portion of the dataset, lack of model’s generalizability, the potential for further performance improvements, and the lack of appropriate performance metrics to evaluate the models.
This work is carried out with the motivation of implementing a robust and reliable DL-based model capable of removing major physiological artifacts (EOG/EMG/simultaneous EOG and EMG) from the contaminated EEG signals with improved performance. In this light, a novel one-dimensional (1D) segmentation network (MultiResUNet3+) to denoise (i) ocular/EOG, (ii) myogenic/EMG, and (iii) simultaneous EOG and EMG artifacts from contaminated EEG is proposed, and the efficacy of the MultiResUNet3+ model is compared with four other 1D-CNN models, namely (i) Feature Pyramid Network (FPN) [54], (ii) UNet [55], (iii) MCGUNet [56], and (iv) LinkNet [57]. Semi-synthetic EEG segments were generated from a publicly available dataset for training and testing DL-based 1D-CNN models [50]. The quantitative and qualitative findings from the proposed MultiResUNet3+, along with four other 1D-CNN models, provide strong evidence of the capacity of DL models to acquire knowledge of the intrinsic characteristics of artifacts mixed into EEG data, detect them, and remove/reduce them reliably and efficiently. The following are our primary contributions, in brief:
  • The proposed MultiResUNet3+ can effectively denoise EOG, EMG, and concurrent EOG and EMG artifacts from corrupted EEG waveforms.
  • We have created a diverse and representative semi-synthetic EEG dataset closely resembling real-world corrupted EEG signals. The proposed 1D-segmentation model was trained and evaluated using 5-fold cross-validation, which ensured the reliability and robustness of the proposed model.
  • We used five well-established performance metrics to comprehensively assess and compare the denoising performance of each of the five 1D-segmentation models.
  • Our developed model may be used for denoising multi-channel, actual EEG data as the model was trained with diverse artifactual data.
The remainder of this manuscript has been structured as follows: Section 2 explains the intricacies of the novel MultiResUNet3+ segmentation network’s architecture. This is trailed by a comprehensive overview of the EEG dataset employed, together with the semi-synthetic EEG data generation and normalization approaches that have been adopted in this study. Section 3 provides the experimental details, explaining the formulae of the performance metrics that are readily available for usage. The quantitative and qualitative performance of the proposed model compared with four different 1D-CNN models is provided in Section 4, and an analysis of the results is presented. Section 5 comprises a cogent discussion of the outcomes, including the limitations inherent in our study, while the paper is concluded concisely in Section 6.

2. Materials and Methods

Using the most up-to-date 1D-CNN-based segmentation networks, Figure 1 shows the end-to-end framework suggested in this study for efficiently denoising EEG signals semi-synthetically corrupted by EOG/EMG/concurrent EOG and EMG artifacts. The figure provides a comprehensive visual representation of our proposed approach and requires no further elucidation.
In the subsequent sub-sections, we detail our proposed MultiResUNet3+ segmentation network, which forms the core of our denoising framework. We also present a concise overview of the EEG dataset employed in this work and discuss the semi-synthetic data generation and normalization techniques adopted in our study, which are crucial in achieving reliable and robust denoising results.

2.1. Proposed Novel MultiResUNet3+ Model Description

Our proposed MultiResUNet3+ model for segmentation effectively combines the concepts of MultiResUNet [58] and UNet3+ [59] networks inside a single framework. The building blocks of the 1D-MultiResUNet3+ model used in this study (depth of 5) have been depicted in Figure 2.
UNet3+ consists of full-scale skip connections combining interconnections between the encoders and the decoders and intra-connections between the decoder sub-networks. Unlike UNet [55] and UNet++ [60], UNet3+ can incorporate small and larger-scale feature maps in each decoder layer, allowing it to extract fine and coarse-grained semantics in full scales. It was employed in the MultiResUNet3+ for this application to effectively minimize semantically different EOG and EMG artifacts from EEG waveforms. EMG requires a finer approach, while EOG-induced impurities are coarser. Nevertheless, UNet3+ still used direct skip connections for inter- and intra-connections, prevailing the semantic gap issue in the basic UNet architecture, which was solved by more advanced architectures, such as UNet++ and MultiResUNet. To reduce the semantic gap, MultiResUNet or similar architectures proposed replacing the skip connections with various formations of convolutional blocks. Reducing the semantic gaps in this application will help the model efficiently learn and generate EEG features, especially when EOG and EMG noises are mixed, to create a more realistic but challenging scenario. Therefore, contrary to UNet3+, our proposed MultiResUNet3+ contains full-scale Residual Paths (Figure 3) instead of direct skip connections for inter- and intra-connections. Instead of combining the encoder and decoder features straightforwardly (e.g., concatenation in UNet and addition in LinkNet [57]), the encoder features are passed through several convolutional layers with residual connections (Figure 4). Mentionable that residual connections [35] have been beneficial in several deep learning applications during their learning process. The depth of the network affects how many residual-convolutional blocks are used in the inter- and intra-ResPaths for MultiResUNet3+. For example, for the model drawn in Figure 2 with a depth of 5, the number of residual-convolutional blocks along the inter-ResPaths will be 4, 3, 2, and 1, respectively, generated from shallower to deeper layers. On the other hand, for the intra-ResPaths densely connecting the decoders, the number of residual-convolutional blocks will be 1, 2, 3, and 4, respectively, from deeper to shallower layers. More residual-convolutional blocks are placed along the ResPaths generated from a shallower layer for improved processing of the coarser features.
Here, Figure 3 represents the full-scale aggregated feature map creation process for MultiResUNet3+ for the decoder X D e 3 . Similar to MultiResUNet, the feature map from the same scale encoder layer X E n 3 is received by the decoder following a ResPath. However, contrary to MultiResUNet, a set of encoder-decoder Residual Paths delivers the semantically enhanced low-level detailed information from the smaller-scale encoder layers X E n 1 and X E n 2 through non-overlapping max-pooling operations.
On the other hand, higher-level semantic information is conveyed from the larger-scale decoder layers, such as X D e 4 and X D e 5 , through intra-ResPath connections by utilizing nearest interpolation. Now, as in Figure 2 and Figure 3, there are five total inter- and intra-Residual Paths (Figure 4) containing same-resolution feature maps waiting to be unified before reaching X D e 3 . To properly merge the shallow and deep semantic information, we perform a feature aggregation mechanism on the ResPaths by placing a Multi-Residual or MultiRes Block (Figure 5) containing W = n × d input kernels or filters of size 3 × 3 after concatenating the five feature maps. Here, n denotes the input filter number, and the depth of the model is represented by d. For example, if n is 64 and d is 5, W will be 320.
The MultiRes Block used in MultiResUNet3+ is similar to the “Module A” of the InceptionV4 network [61]. It consists of a succession of 3 × 3 sized filters being concatenated and then element-wise added by a residual connection passed through a 1 × 1 convolutional block to squeeze dimensions. This helps us to aggregate spatial features from various context sizes while avoiding expensive filters with larger kernels (e.g., 5 × 5 or 7 × 7 , primarily used in earlier Inception networks). For the MultiRes block, instead of keeping the number of filters the same, it is more efficient and less expensive to gradually increase them while keeping the total number of filters the same as if it were a single convolutional block, as explained in [58]. For instance, if the number of input filters for the MultiRes block used for feature aggregation is W = n × d = 320 , the width of the three convolution blocks inside the MultiRes block will be w / 6 = r o u n d ( 320 / 6 ) 53 , W 3 = r o u n d ( 320 / 3 ) 107 , and w / 2 = 160 , respectively. Each MultiRes block is followed by batch normalization and ReLU activation layers.

2.2. Dataset Description

A publicly available dataset, namely EEGdenoiseNet [50], was found to be suitable for this work as it contained pre-processed, well-structured physiological segments gathered from various sources, including 4514 clean EEG segments, 3400 pure EOG segments, and 5598 clean EMG segments. We have used these readily available segments to generate semi-synthetic noisy EEG, which were used to train and test the five distinct 1D-segmentation models.
Several signal-processing techniques were adopted by the EEGdenoiseNet [50] authors. They processed the clean EEG signals (collected from several sources) by applying a bandpass filter with the lower and upper cutoff frequencies of 1 and 80 Hz, removed the power line noise using a notch filter, and then down-sampled at 256 Hz. The clean EEG signals were then segmented, and 4514 clean EEG segments were generated, each containing 512 data points. The clean EOG signals (horizontal and vertical) were bandpass filtered with a passband of 0.3–10 Hz, resampled to 256 data points per second. These processed clean EOG signals were segmented, and 3400 segments were produced, where each clean EOG segment has 512 data points, the same as the EEG segments. The EMG signal was filtered using a bandpass filter with a lower cutoff frequency of 1 Hz and upper cutoff frequency of 120 Hz, and the powerline frequency component was removed using a notch filter. Then the raw EMG signals were resampled to 512 data points per second, and 5598 clean muscle/EMG segments were generated, each containing 1024 data points [50]. As EMG signals have essential features in the high-frequency band, the clean EMG signals were sampled at 512 Hz to retain those features and morphology. Figure 6 depicts one clean EEG, one clean vertical EOG, one pure horizontal EOG, and one clean EMG segment sample. Since the authors of [50] adopted all the necessary pre-processing techniques, no further pre-processing steps were undertaken in this study. Instead, the publicly available segments are directly used to produce a large and diverse semi-synthetic corrupted EEG dataset.

2.3. Semi-Synthetic Electroencephalogram Segment Generation and Normalization

For this study, we generate semi-synthetic EOG, EMG, and EOG-EMG contaminated EEG segments from the true signals in the EEGdenoiseNet dataset. By linearly mixing one clean EEG segment with one clean EOG and/or EMG segment using Equation (1), a semi-synthetic corrupted EEG segment can be produced [50]:
y = x + λ · n
Here, x denotes the ground truth or clean EEG segments, y is the generated semi-synthetic noisy EEG, and n characterizes the EOG and/or EMG artifacts. By solving Equation (2) and altering the scaling factor λ , the signal-to-noise Ratio (SNR) of the semi-synthetically generated contaminated EEG segment can easily be adjusted to different levels as follows [50]:
S N R = 10   l o g R M S ( x ) R M S ( λ · n )
For the computation of root mean square (RMS) value, Equation (3) is utilized [50]:
R M S w = 1 m i = 1 m w i 2
Here, m stands for the total number of data points of segment w , and w i stands for the i t h sample point of w . The scaling factor λ is pivotal in determining the SNR of the semi-synthetic, corrupted EEG segments. As a rule of thumb, a lower value of λ corresponds to a higher SNR, whereas a higher λ leads to a poorer SNR. The SNR of EEG solely corrupted by EOG artifacts typically falls within the range of −7 to 2 dB, as reported in previous studies [23]. On the other hand, EEG signals corrupted by EMG artifacts exhibit a comparatively wider range of SNR, typically ranging from −7 to 4 dB [62]. The availability of ground truth and noisy EEG segments allowed us to train, test and validate the deep learning models for EEG denoising. However, rather than directly feeding the ( x , y ) pairs into the 1D-CNN models, we considered the standard deviation of the corrupted EEG segments ( σ y ) and divided the clean and contaminated EEG signals by this value (Equation (4)). The result was a pair of rescaled or normalized segments ( x ^ , y ^ ), fed to the 1D-segmentation models for training, testing, and validation.
x ^ = x σ y ;   y ^ = y σ y
Since deep learning models are sensitive to variations in the magnitude of input data, without scaling or normalization, the model may be unable to make useful inferences. Normalization aids in keeping all input features of equal importance, which is especially helpful when dealing with large-scale data variations. Normalization also helps the model to converge faster during model training by minimizing the variance of the input data. Moreover, the training process can be accelerated via normalization, and the model’s learning capacity can be facilitated. Again, normalization aids in avoiding overfitting.
In this study, the EOG-contaminated EEG segments were generated by linearly mixing randomly chosen 3400 clean EEG segments (out of 4514 clean EEG segments) with the available 3400 EOG segments. This process was repeated for ten different integer SNR levels (−7 to 2 dB), producing 34,000 semi-synthetic EOG-contaminated EEG segments. The sampling frequency of both EOG and EEG segments was kept at 256 Hz. It is worthwhile to mention that as there were 4514 clean EEG segments available, we could easily produce 45,140 semi-synthetic corrupted EEG segments for ten different SNR levels by using some clean EOG more than once during linear mixing. However, this process was avoided to prevent data-leaking issues.
Similarly, EMG-contaminated EEG segments were created by combining the EEG segments with the randomly selected clean EMG segments for 10 different SNR levels. Before linearly mixing EEG and EMG segments, all the clean EEG segments were up-sampled by a factor of 2 to match the number of data points each EMG segment contains. The upsampling of EEG (from 256 to 512 Hz) did not cause any morphological change in EEG segments as the bandwidth of EEG is 1–80 Hz [63]. Eventually, we generated 45,140 EMG-contaminated EEG segments for the ten different SNR levels semi-synthetically. The simultaneous EOG and EMG contaminated EEG segments were generated by mixing the 3400 clean EEG segments (randomly taken out from 4514 clean EEG segments) with 3400 EOG and 3400 EMG (randomly taken out of 5598 EMG segments) segments for ten different SNR levels. The clean EEG and EOG segments were up-sampled at 512 Hz before linear mixing to match the sampling frequency of the EMG segments. This process generated 34,000 corrupted EEG segments for all ten different SNR levels. Figure 7a displays a picked at-random EOG-contaminated EEG segment. Figure 7b illustrates one arbitrary EMG-contaminated EEG segment, whereas Figure 7c represents one random sample of simultaneous EOG and EMG-contaminated EEG segments. The corresponding ground truth EEG segments are also superimposed in Figure 7a–c.

3. Experimental Setup

In pursuit of denoising EEG signals from physiological (EOG/EMG/concurrent EOG and EMG) artifacts, the novel MultiResUNet3+ model, alongside four other 1D-CNN models, was trained. The training involved feeding the neural networks normalized contaminated EEG segments as input (generated using Equation (4)) while providing corresponding normalized ground truth EEG segments as output. This process made it easier for the DL-based model to create a nonlinear function that translated the noisy EEG to its equivalent ground truth. The mean squared error (MSE) was employed as the loss function to generate the nonlinear mapping function. The loss function was optimized using the Adam optimizer with a learning rate of 0.0005. Eighty percent of the data was used as the training set and the remaining 20% for the test set. Ten percent of the training set data were used as the validation set. It is worth mentioning here that each EEG, EOG, and EMG segment is only used once to produce the semi-synthetic dataset. Therefore, any sort of data leaking between train, test, and validation sets is absent. All five networks were trained, validated, and tested independently using the five-fold cross-validation technique in the Google ColabPro environment using Python 3.10 framework to ensure robustness and reliability in the evaluation process. In this work, the experimentation was done with a twofold approach which is described below:

3.1. Experiment A

As mentioned earlier, there were in total 34,000 EOG-contaminated EEG segments having ten integer SNR values ranging from −7 to +2 dB, where for each SNR value, 3400 EEG segments contaminated with ocular artifacts were generated. For EMG-contaminated EEG, the total number of contaminated segments was 45,140. In contrast, for simultaneous EOG and EMG corrupted EEG, this number was 34,000 for ten different SNR levels ranging from −7 to +2 dB. For all these three types of artifacts, we trained each of the five 1D-CNN models ten times, utilizing the corrupted and corresponding ground truth segments for ten different SNR values. That is, for EOG corrupted EEG segments at SNR level −7 dB, 80% of 3400 segments (2720) were used for training the models, and the remaining 20% (680 segments) were utilized for testing. This same process was carried out separately for nine other integer SNR levels (−6 to 2 dB).
Similarly, for ten different SNR levels, EMG-contaminated EEG segments, and simultaneous EOG and EMG-contaminated EEG segments, were utilized for training and testing all five networks ten times separately. Employing a deep supervision technique as described in [64], in this experiment, we computed three established performance metrics, namely the correlation coefficient (CC) in the time domain, and temporal and spectral relative root mean squared error (RRMSE). These metrics were particularly used to evaluate the effectiveness of the five DL-based models in denoising contaminated EEG. The five-fold cross-validation technique was employed, and the metrics mentioned earlier were computed for each of the five folds to ensure robustness and reliability in the evaluation process.

3.2. Experiment B

In Experiment B, a comprehensive approach was taken to generate train and test sets so that a more robust DL-based model could be trained that would perform well across a wide range of SNR levels, which is crucial for real-world applications where the exact SNR of the input signal may vary due to factors, such as electrode placement, the patient’s movements, or equipment quality. For each of the 10 different SNR levels, 80% of the EOG-contaminated EEG segments and their corresponding ground truth EEG segments were extracted to produce a more extensive dataset (2720 pairs in each SNR level). After merging these pairs, a training dataset of 27,200 pairs was created. The remaining 20% of pairs from each of the ten distinct SNR levels were used to generate the test set (6800 pairs). Unlike in Experiment A, where models were trained and evaluated for specific SNR level segments, the models in Experiment B were trained using combined segments of all ten distinct SNR levels. This improved model generalization, making them more resilient when evaluated with noisy EEG segments with varying SNR levels. In essence, Experiment B took a holistic approach to train DL models that could perform well across the noisy EEG segments having a range of SNR levels, thereby increasing their utility in real-world scenarios where the SNR of the input signal may not be precisely known. Following a similar protocol, a total of 36,110 EMG-contaminated EEG segments and 27,200 pairs of simultaneous EOG and EMG-contaminated EEG segments were generated separately to form the train set, and the remaining 20% contaminated EEG segments were used as test set (9030 pairs of EMG-contaminated EEG segments, and 6800 pairs of simultaneous EOG and EMG contaminated EEG segments) for the segmentation models. The efficacy of the trained models was quantitatively measured using five performance metrics, i.e., the percentage reduction in artifacts in the time and frequency domain, temporal and spectral RRMSE, and the average power ratio of each of the five different EEG bands (Alpha, Beta, Gamma, Delta, Theta bands) to the whole band separately.

3.3. Performance Evaluation Metrics

The proper quantitative assessment of any DL model is paramount in determining its ability to accomplish the intended task, in our case, effectively denoising EEG signals. In this study, meticulous attention was devoted to selecting the most appropriate performance parameters for an adept evaluation of the chosen segmentation models. The efficacy of the five 1D-CNN models was quantitatively evaluated by calculating several temporal and spectral performance metrics, such as the correlation coefficient (CC), the percentage reduction in artifacts, and the relative root mean squared error (RRMSE). These metrics were critical in determining the best-performing models. The relevant equations for these measures can be found in Equations (5)–(9), as collected from [50,65],
C C t e m p o r a l = C o v ( z ^ , x ^ ) V a r ( z ^ ) V a r ( x ^ )
η = 1 1 C C t e m p o r a l a f 1 C C t e m p o r a l b f 100
γ = 1 1 C C s p e c t r a l a f 1 C C s p e c t r a l b f 100
R R M S E t e m p o r a l = R M S ( z ^ x ^ ) R M S ( x ^ )
R R M S E s p e c t r a l = R M S ( P S D ( z ^ ) P S D ( x ^ ) ) R M S ( P S D x ^ )
Here, the time domain correlation coefficient is represented by C C t e m p o r a l while covariance is denoted by C o v . The predicted EEG segments are represented by z ^ , whereas x ^ denotes the normalized ground truth EEG segments. Variance is characterized by V a r , while η represents the temporal percentage reduction of EOG/EMG/simultaneous EOG and EMG artifacts from the corrupted EEG, and γ represents the percentage reduction of EOG, EMG, or simultaneous EOG and EMG artifacts from the corrupted EEG in the frequency domain. Furthermore, the time domain correlation coefficient between the predicted and the ground truth EEG segments is denoted by C C t e m p o r a l ( a f ) , while C C t e m p o r a l ( b f ) is used to represent the time domain correlation coefficient between contaminated and ground truth EEG segments. Similarly, the frequency domain correlation coefficient between the predicted and the ground truth EEG segments is expressed by C C s p e c t r a l ( a f ) , whereas C C s p e c t r a l ( b f ) is used to represent the frequency domain correlation coefficient between the contaminated and the ground truth EEG segments. R M S is the abbreviation for root mean square, which can be computed using Equation (3), and finally, the PSD is the power spectral density computed using the Periodogram method. The Periodogram method involves calculating the Discrete FT (DFT) of the signal and then taking the square of the absolute value of the DFT to obtain the power spectral density.
We have selected these performance metrics to evaluate the 1D-CNN-based segmentation models. The reasoning for choosing these metrics is summarized as follows:
The correlation coefficient in the time domain ( C C t e m p o r a l ) measures the degree of similarity between two variables. A higher correlation coefficient between predicted and ground truth EEG would indicate that the predicted EEG is more similar to the ground truth EEG and vice-versa. Hence, calculating the correlation coefficient is a valuable tool that quantitatively measures the model’s adeptness/inability in denoising.
The metric temporal percentage reduction in artifacts ( η ) measures the proportion of EOG/EMG/simultaneous EOG and EMG artifacts removed from the noisy EEG signal in the time domain. A higher temporal percentage reduction in artifacts indicates that more of the artifact has been removed from the EEG signal, resulting in a clean and more accurate EEG signal.
In contrast, the spectral percentage reduction in EOG/EMG/simultaneous EOG and EMG artifacts ( γ ) measures the proportion of EOG/EMG artifacts removed from the EEG signal in the frequency domain. A higher spectral percentage reduction indicates that more of the artifact has been removed from the EEG signal across all frequency bands and vice-versa.
The RRMSE measured in the time domain ( R R M S E t e m p o r a l ) can provide insights into the temporal dynamics of the denoised EEG signals. A low temporal RRMSE value indicates that the predicted individual EEG segment closely approximates the corresponding ground truth EEG segment.
The RRMSE measured in the frequency domain ( R R M S E s p e c t r a l ) can help assess the ability of the predictor model to capture important spectral features of the EEG signals in the whole synthetic dataset, such as alpha, beta, gamma, delta, and theta bands. A low spectral RRMSE value indicates that the predicted EEG signal accurately captures the power distribution across different frequency bands. This is important for ensuring that the predicted EEG signal is not biased towards or against any particular frequency band, which could have unintended effects on subsequent analyses.
The average power ratio measures the power of a given frequency band against the power of the EEG signal over the whole spectrum. It is computed by dividing the power of a particular frequency band by the total power of the signal. In general, by calculating the average power ratio for each frequency band separately, we can better understand the distribution of power across the different frequency ranges, which is vital. Specifically, while estimating average power ratios for ground truth and predicted denoised signal for any frequency band, a closely matched numerical value would indicate that the signal is predicted more accurately. In contrast, the power ratio of noisy and ground truth signals should have far apart numerical values. For this reason, the performance of the 1D-CNN models for five different frequency bands of EEG is also reported in this study. Specifically, the five EEG frequency bands are delta (1–4 Hz), theta (4–8 Hz), alpha (8–13 Hz), beta (13–30 Hz), and gamma (30–80 Hz). The average power ratio of each of these bands to the entire band (1–80 Hz) is calculated separately for each predicted, corresponding ground truth, and contaminated EEG segment. The average power ratio was computed following the Periodogram method [65].

4. Results

This section details the outcomes of Experiments A and B and analyses the results concisely.

4.1. Experiment A Outcomes

For the quantitative performance evaluation of Experiment A, three performance metrics are computed, namely (i) time domain correlation coefficient, (ii) temporal RRMSE, and (iii) spectral RRMSE utilizing the ground truth EEG and the denoised EEG predicted by all five models in three different denoising scenarios, i.e., for EOG-artifacts removal, EMG-artifacts removal, and concurrent EOG and EMG artifacts removal from corrupted EEG, separately.
In Table 1, Table 2 and Table 3, the numerical values of the three performance metrics ( C C t e m p o r a l , R R M S E t e m p o r a l , and R R M S E s p e c t r a l ) are presented for the denoised EOG, EMG, and simultaneous EOG and EMG-contaminated EEG segments, respectively. On the other hand, Figure 8 illustrates the evaluation of five different 1D-CNN models in the context of EOG-contaminated EEG denoising. Specifically, the figure presents C C t e m p o r a l values obtained across ten integer SNR levels, spanning from −7 to +2 dB. Additionally, Figure 8 also displays the R R M S E t e m p o r a l and R R M S E s p e c t r a l values plotted against the same ten SNR levels. Similarly, Figure 9 and Figure 10 illustrate the same performance metrics plotted against ten different SNR levels computed for the predicted denoised EMG-contaminated EEG segments and simultaneous EOG and EMG-contaminated EEG segments.
It is apparent from Table 1 and Figure 8 that our proposed MultiResUNet3+ model outperformed all remaining four 1D-segmentation models in terms of removing EOG artifacts from noisy EEG segments. This superiority is consistently observed across all ten different SNR levels. Specifically, the MultiResUNet3+ model achieved the highest CC values, indicating a higher correlation between the predicted and ground truth EEG segments. Furthermore, the RRMSE values obtained from the MultiResUNet3+ model were the lowest in both the temporal and spectral domains, again demonstrating its superior accuracy and precision in removing EOG artifacts. Overall, these findings highlight the effectiveness and robustness of the proposed MultiResUNet3+ model in reducing EOG artifacts from EOG-contaminated EEG denoising, even under challenging conditions characterized by low SNR levels.
As can be observed from Figure 9, the five 1D-CNN models displayed similar performance in removing EMG artifacts from the EMG-contaminated EEG segments, as indicated by the almost overlapping curves of all three-performance metrics. Given the difficulty of determining the best-performing model solely based on Figure 9, Table 2 is included to provide numerical values of C C t e m p o r a l , R R M S E t e m p o r a l , and R R M S E s p e c t r a l . The tabulated data highlight the inconsistent performance of the 1D-segmentation networks in predicting EMG-artifacts-free EEG. Notably, the proposed MultiResUNet3+ and MCGUNet demonstrated relatively superior performance compared to the other three models.
During the denoising of simultaneous EOG and EMG artifacts from corrupted EEG segments, the greatest temporal correlation coefficient (a little above 0.95) is obtained from MultiResUNet3+ for the segments having the lowest noise level (SNR level of +2 dB). The smallest RRMSE in the time and frequency domain is also achieved by our proposed MultiResUNet3+ model compared to the other four 1D-CNN models while estimating the performance on test segments (refer to Table 3 and Figure 10).
Overall, for EOG/EMG/simultaneous EOG and EMG artifacts removal, the efficacy of the deep learning (DL) models is observed to be enhanced in parallel with the increment of the SNR, as anticipated. The increase of SNR level results in a proportional reduction in the noise quotient (comprising EOG, EMG, and simultaneous EOG and EMG), thereby possibly reducing the complexity of the nonlinear mapping function acquired by the DL-based segmentation models for forecasting denoised EEG segments. Ultimately, this leads to an improvement in the model’s performance in predicting noise-free EEG.

4.2. Experiment B Outcomes

All five 1D-CNN models’ objective is to estimate artifacts-free EEG segments. To determine the most effective 1D-segmentation model in reducing physiological artifacts from corrupted EEG, five different performance metrics have been computed, namely (i) Temporal percentage reduction in artifacts ( η ), (ii) spectral percentage reduction in artifacts ( γ ), (iii) RRMSE in the time domain ( R R M S E t e m p o r a l ), (iv) RRMSE in the frequency domain ( R R M S E s p e c t r a l ), and (v) the average power ratio for five distinct EEG bands (Delta, Theta, Alpha, Beta, Gamma) for ground truth, noisy, and predicted artifacts-free EEG segments. Moreover, a sample corrupted, ground truth, and predicted clean EEG segment plots are provided for visual or qualitative assessment.
Figure 11a depicts an EEG segment that has been contaminated with EOG artifacts, while Figure 11b–f illustrates EOG artifact-free EEG segments (estimated clean EEG) for all five models. Table 4 provides the average numerical values of the four performance metrics ( η ,   γ ,   R R M S E t e m p o r a l ,   a n d   R R M S E s p e c t r a l ) by the five distinct models. In contrast, Table 5 illustrates the computed average power ratios for five separate EEG bands for ground truth, noisy, and predicted EEG segments.
From Table 4, our proposed MultiResUNet3+ performed best in reducing EOG artifacts from EOG-contaminated EEG segments in the temporal and spectral domains (94.82% and 92.84%, respectively). Although the UNet model produced the lowest R R M S E t e m p o r a l (~0.11) and R R M S E s p e c t r a l (~0.12), respectively, MultiResUNet3+ was also very close. UNet produced the closest delta and alpha power ratio compared to the ground truth EEG. For theta, beta, and gamma band power ratio, our proposed MultiResUNet3+ performed best (refer to Table 5).
An example EMG-contaminated EEG segment is illustrated in Figure 12a, and to provide a qualitative overview of the five denoising 1D-CNN models, Figure 12b–f shows EMG artifacts-free EEG (predicted EEG) segments. Table 6 summarizes the four performance metrics ( η ,   γ ,   R R M S E t e m p o r a l ,   a n d   R R M S E s p e c t r a l ) obtained for the predicted EMG artifacts-free EEG segments by all the five 1D-CNN models separately, and Table 7 contains the average power ratio calculated for five different EEG bands before and after the removal of myogenic artifacts.
For EMG-corrupted EEG segments denoising, the proposed MultiResUNet3+ eliminated 83.21% of spectral EMG artifacts, whereas the UNet eliminated 89.59% of time domain EMG artifacts from the corrupted EEG segments. In terms of η , UNet performed best, and in terms of γ reduction, our proposed model performed best among all the models (refer to Table 6). Again, UNet and MultiResUNet3+ produced the lowest R R M S E t e m p o r a l (~0.22) and R R M S E s p e c t r a l (~0.19) values, respectively, when evaluated with the other four 1D CNN networks. Moreover, when comparing the power ratios computed using the MultiResUNet3+ model’s predicted segments with the ground truth EEG’s power ratios among the delta, beta, and gamma bands, the MultiResUNet3+ model comes out on top. For the theta and alpha band power ratio, LinkNet performed best (refer to Table 7).
In Figure 13a, an example segment of simultaneous EOG and EMG contaminated EEG segment is presented, and in Figure 13b–f, the denoised EEG segment (simultaneous EOG and EMG artifacts-free segment) along with the ground truth EEG is shown separately for the five 1D-CNN models. The four performance metrics for the five 1D-CNN models’ predictions on the denoised simultaneous EOG and EMG-contaminated EEG segments can be found in Table 8, whereas Table 9 provides the average power ratio between five different EEG bands before and after removing simultaneous EOG and EMG artifacts.
As observed from Table 8, the proposed MultiResUNet3+, UNet, and LinkNet models produced very close denoising performance for simultaneous EOG and EMG corrupted EEG segments. Artifacts are reduced by 89.77% in the time domain and 83.39% in the frequency domain when LinkNet and UNet models are used, respectively. When compared among the five 1D-CNN networks, the LinkNet model yielded the lowest R R M S E t e m p o r a l (~0.22) and R R M S E s p e c t r a l (~0.18) values. Finally, when compared to the ground truth EEG, the MultiResUNet3+ model excels in producing the nearest average power ratios for the delta band. However, the LinkNet model performed best in the theta, beta, and gamma band power ratio, whereas the FPN model showed superior performance in the alpha band (Table 9).

5. Discussion

The limitations of traditional single-stage and two-stage signal processing-based methods for denoising EEG signals, including low correlation coefficient, potential loss of crucial neural information, poor performance in dynamic situations, etc., are mentioned in the introduction section of this study. Although a few deep learning-based approaches have been presented for EEG denoising to remedy these drawbacks, with notable performance gains over signal processing-based methods, these DL-based approaches still have some shortcomings, among which lackings in model robustness, reliability, and generalizability are the key.
To address these limitations, five distinct DL models are utilized in this extensive study, including the proposed novel MultiResUNet3+ and four additional 1D-CNN models (FPN, UNet, MCGUNet, and LinkNet) to denoise EOG/EMG/simultaneous EOG and EMG artifacts from corrupted EEG. A publicly available dataset, namely EEGdenoiseNet, is deemed suitable for this work, as it featured pre-processed, well-structured EEG segments, including 4514 clean EEG segments, 3400 pure electrooculogram segments, and 5598 clean electromyogram segments. We semi-synthetically generated a large set of noisy EEG segments using linear mixing techniques where these clean EOG and EMG segments were used as noisy signals and clean EEG segments as ground truth to train the DL-based models. This process provided a broad range of signal types for the DL models to learn from. Furthermore, due to linear mixing, ground truth EEG segments were also available to provide a reliable reference for evaluating the performance of the models. To prevent any data leaking issues between the train and test set, the clean EEG, EOG, and EMG segments were used only once during the semi-synthetic data generation process.
Further, the quantitative measurement of five performance metrics (Experiment B) for all the five deep learning networks (FPN, UNet, MCGUNet, LinkNet, and the proposed novel MultiResUNet3+) showed that it is possible to denoise EEG signal artifacts using DL-based techniques and has a great potential to eliminate multiple artifacts simultaneously from EEG signals by using robust DL models. It should be added here that, for DL networks, high-frequency artifacts, such as EMG artifacts and simultaneous EOG and EMG artifacts, are more challenging to handle than low-frequency artifacts, such as EOG artifacts. The F-principle of neural networks can be used to explain this occurrence [66], which states that DL networks frequently learn low-frequency information at the beginning of training and high-frequency information as training iterations rise. A similar phenomenon is also observed in this study. From our extensive experiments, we have observed that removing EOG artifacts (low-frequency noise) from corrupted EEG is less challenging than denoising EMG-contaminated (high-frequency noise) EEG and simultaneous EOG and EMG-contaminated (high-frequency noise) EEG. For instance, a higher temporal percentage reduction in EOG artifacts (94.82%) is found than the temporal percentage reduction in EMG artifacts (89.59%). Similarly, a higher spectral percentage reduction in EOG artifacts (92.84%) is computed than in EMG artifacts (83.21%). Although simultaneous EOG and EMG contamination in EEG is very unlikely in a real-life scenario and challenging to remove reliably, our proposed MultiResUNet3+ provided a staggering performance—clear evidence of the efficacy and superiority of our proposed model in denoising complex artifacts as well.
Although the DL-based denoising approaches need a vast amount of ground truth EEG data during the training phase, once the model is trained, it can be utilized reliably to eradicate artifacts from unseen EEG signals corrupted with physiological artifacts. Another benefit of DL models is the ability to handle complicated artifact mixes, such as nonlinear and stationary ones. In contrast to the traditional signal processing approaches, which typically linearly attenuate artifacts, DL models can directly learn the fundamental pattern of neural activities from training data in the hidden space and then synthesize/predict the clean EEG from corrupted EEG. Therefore, DL-based techniques perform better than conventional methods in removing physiological noises from contaminated EEG.
It is indeed worth mentioning some constraints of this study. Even though the semi-synthetic EEG dataset generated in this study includes a very high number of clean EEG, EOG, and EMG segments, the lesser variability in EEG type and lack of diversity in artifacts type is a matter of concern since EEG data can be recorded when the subject is at rest or while doing various tasks. Moreover, artifacts in EEG recordings are not just limited to ocular and myogenic. Motion artifacts, one of the inevitable sources of non-physiological noise, become dominant while EEG data are captured via wearable sensors. Therefore, a more diverse and dynamic dataset curation is necessary, which will help the DL-based models to be more efficient in removing physiological and non-physiological artifacts from corrupted EEG signals.

6. Conclusions

The elimination of artifacts is a crucial aspect of EEG data analysis. In this paper, we presented a novel 1D-CNN model, i.e., MultiResUNet3+ for EEG denoising, and demonstrated its superiority over four other 1D-CNN models, namely FPN, UNet, MCGUNet, and LinkNet. Apart from proposing a novel 1D-CNN model for denoising EEG reliably, we also introduced a set of benchmark metrics to facilitate the quantitative evaluation of DL-based EEG denoising models’ adeptness. The proposed MultiResUNet3+ model reduced the EOG artifacts from EOG-corrupted EEG by 94.82% and 92.84% in the time and frequency domains, respectively. Moreover, the nearest average power ratio for the theta, beta, and gamma bands (0.5295, 0.076, 0.0197) compared with the ground truth EEG (0.5184, 0.0749, 0.0195) while removing the EOG artifact was produced by MultiResUNet3+, meaning our proposed model is capable of denoising EEG in those bands most accurately. For EMG artifacts removal from contaminated EEG, our proposed novel MultiResUNet3+ performed best by reducing 83.21% artifacts in the frequency domain and a neck-to-neck 89.33% artifacts reduction in the time domain (the best is 89.59%, produced by UNet). While denoising EMG-contaminated EEG segments, MultiResUNet3+ had the lowest spectral RRMSE (0.1893) as well. Once EMG artifacts were removed from the corrupted EEG segments, the proposed MultResUNet3+ provided the nearest average power ratio in delta, beta, and gamma bands (0.4453, 0.0658, and 0.0165, respectively) compared with the ground truth EEG (0.4421, 0.0756, and 0.0197, respectively). In denoising, simultaneous EOG and EMG from corrupted EEG, three 1D-CNN models, namely UNet, LinkNet, and novel MultiResUNet3+, performed neck-to-neck, which is evident from the computed performance evaluation metrics. Overall, the results obtained from the proposed MultiResUNet3+ clearly showed its robustness and reliability in denoising physiological artifacts (EOG/EMG/simultaneous EOG and EMG) from corrupted EEG signals. Our findings showed that DL techniques could potentially eliminate EOG/EMG artifacts from EEG data even at high noise levels. Following a similar framework, the proposed MultiResUNet3+ segmentation network may be used for multi-channel EEG noise reduction by applying the proposed model to multi-channel EEG signals separately, which should facilitate EEG-based BCI applications.

Author Contributions

Conceptualization M.S.H., S.M., A.K. and M.E.H.C.; data curation M.S.H., S.M., A.K. and F.A.C.; software F.A.C.; formal analysis M.S.H., S.M., A.K. and F.A.C.; funding acquisition M.E.H.C., M.B.I.R. and Z.B.M.; investigation M.S.H., S.M., A.K. and F.A.C.; methodology M.S.H., S.M., A.K., M.E.H.C., Z.B.M., N.A.-E. and M.B.I.R.; project administration M.E.H.C., Z.B.M., N.A.-E. and M.B.I.R.; resources M.E.H.C., Z.B.M. and N.A.-E.; supervision M.E.H.C., Z.B.M. and N.A.-E.; validation M.E.H.C., M.B.I.R. and Z.B.M. All authors equally contributed to original paper drafting and reviewing and editing the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research is financially supported by Qatar National Research Foundation (QNRF), Grant number NPRP12s-0227-190164 and Qatar University Student Grant: QUST-1-CENG-2023-795 and NSU CTRG-21-SEPS-20 grant from North South University and the ICTP through the Affiliated Centres Programme. The statements made herein are solely the responsibility of the authors. Open Access publication of this article is supported by Qatar National Library.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Zhang et al. [50] generously provided the data utilized in this study, and the semi-synthetically generated EEG dataset can be accessed by the corresponding authors of this work by making a reasonable request.

Conflicts of Interest

The authors declare that they have no known competing financial interests.

References

  1. Wolpaw, J.R.; McFarland, D.J.; Neat, G.W.; Forneris, C.A. An EEG-based brain-computer interface for cursor control. Electroencephalogr. Clin. Neurophysiol. 1991, 78, 252–259. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, Y.; Guo, Y.; Yang, P.; Chen, W.; Lo, B. Epilepsy seizure prediction on EEG using common spatial pattern and convolutional neural network. IEEE J. Biomed. Health Inform. 2019, 24, 465–474. [Google Scholar] [CrossRef]
  3. Ullah, I.; Hussain, M.; Aboalsamh, H. An automated system for epilepsy detection using EEG brain signals based on deep learning approach. Expert Syst. Appl. 2018, 107, 61–71. [Google Scholar] [CrossRef]
  4. Mammone, N.; De Salvo, S.; Bonanno, L.; Ieracitano, C.; Marino, S.; Marra, A.; Bramanti, A.; Morabito, F.C. Brain network analysis of compressive sensed high-density EEG signals in AD and MCI subjects. IEEE Trans. Industr. Inform. 2018, 15, 527–536. [Google Scholar] [CrossRef]
  5. Yang, S.; Bornot, J.M.S.; Wong-Lin, K.; Prasad, G. M/EEG-based bio-markers to predict the MCI and Alzheimer’s disease: A review from the ML perspective. IEEE Trans. Biomed. Eng. 2019, 66, 2924–2935. [Google Scholar] [CrossRef] [PubMed]
  6. Tripathy, R.; Acharya, U.R. Use of features from RR-time series and EEG signals for automated classification of sleep stages in deep neural network framework. Biocybern. Biomed. Eng. 2018, 38, 890–902. [Google Scholar] [CrossRef]
  7. Antonenko, P.; Paas, F.; Grabner, R.; Van Gog, T. Using electroencephalography to measure cognitive load. Educ. Psychol. Rev. 2010, 22, 425–438. [Google Scholar] [CrossRef]
  8. Gupta, V.; Chopda, M.D.; Pachori, R.B. Cross-subject emotion recognition using flexible analytic wavelet transform from EEG signals. IEEE Sens. J. 2018, 19, 2266–2274. [Google Scholar] [CrossRef]
  9. Xing, X.; Wang, Y.; Pei, W.; Guo, X.; Liu, Z.; Wang, F.; Ming, G.; Zhao, H.; Gui, Q.; Chen, H. A high-speed SSVEP-based BCI using dry EEG electrodes. Sci. Rep. 2018, 8, 14708. [Google Scholar] [CrossRef]
  10. Rahman, A.; Chowdhury, M.E.; Khandakar, A.; Tahir, A.M.; Ibtehaz, N.; Hossain, M.S.; Kiranyaz, S.; Malik, J.; Monawwar, H.; Kadir, M.A. Robust biometric system using session invariant multimodal EEG and keystroke dynamics by the ensemble of self-ONNs. Comput. Biol. Med. 2022, 142, 105238. [Google Scholar] [CrossRef]
  11. Flexer, A.; Bauer, H.; Pripfl, J.; Dorffner, G. Using ICA for removal of ocular artifacts in EEG recorded from blind subjects. Neural Netw. 2005, 18, 998–1005. [Google Scholar] [CrossRef] [PubMed]
  12. McMenamin, B.W.; Shackman, A.J.; Maxwell, J.S.; Bachhuber, D.R.; Koppenhaver, A.M.; Greischar, L.L.; Davidson, R.J. Validation of ICA-based myogenic artifact correction for scalp and source-localized EEG. NeuroImage 2010, 49, 2416–2432. [Google Scholar] [CrossRef] [PubMed]
  13. Jorge, J.; Bouloc, C.; Bréchet, L.; Michel, C.M.; Gruetter, R. Investigating the variability of cardiac pulse artifacts across heartbeats in simultaneous EEG-fMRI recordings: A 7T study. NeuroImage 2019, 191, 21–35. [Google Scholar] [CrossRef] [PubMed]
  14. Marino, M.; Liu, Q.; Del Castello, M.; Corsi, C.; Wenderoth, N.; Mantini, D. Heart–Brain interactions in the MR environment: Characterization of the ballistocardiogram in EEG signals collected during simultaneous fMRI. Brain Topogr. 2018, 31, 337–345. [Google Scholar] [CrossRef]
  15. Jiang, X.; Bian, G.-B.; Tian, Z. Removal of artifacts from EEG signals: A review. Sensors 2019, 19, 987. [Google Scholar] [CrossRef]
  16. Hossain, M.S.; Reaz, M.B.I.; Chowdhury, M.E.; Ali, S.H.; Bakar, A.A.A.; Kiranyaz, S.; Khandakar, A.; Alhatou, M.; Habib, R. Motion Artifacts Correction from EEG and fNIRS Signals using Novel Multiresolution Analysis. IEEE Access 2022, 10, 29760–29777. [Google Scholar] [CrossRef]
  17. Hossain, M.S.; Chowdhury, M.E.; Reaz, M.B.I.; Ali, S.H.M.; Bakar, A.A.A.; Kiranyaz, S.; Khandakar, A.; Alhatou, M.; Habib, R.; Hossain, M.M. Motion Artifacts Correction from Single-Channel EEG and fNIRS Signals Using Novel Wavelet Packet Decomposition in Combination with Canonical Correlation Analysis. Sensors 2022, 22, 3169. [Google Scholar] [CrossRef]
  18. Somers, B.; Francart, T.; Bertrand, A. A generic EEG artifact removal algorithm based on the multi-channel Wiener filter. J. Neural Eng. 2018, 15, 036007. [Google Scholar] [CrossRef]
  19. Shahabi, H.; Moghimi, S.; Zamiri-Jafarian, H. EEG eye blink artifact removal by EOG modeling and Kalman filter. In Proceedings of the 5th International Conference on Biomedical Engineering and Informatics, Chongqing, China, 16–18 October 2012; pp. 496–500. [Google Scholar]
  20. Correa, A.G.; Laciar, E.; Patiño, H.; Valentinuzzi, M. Artifact removal from EEG signals using adaptive filters in cascade. J. Phys. Conf. Ser. 2007, 90, 012081. [Google Scholar] [CrossRef]
  21. Gao, C.; Ma, L.; Li, H. An ICA/HHT hybrid approach for automatic ocular artifact correction. Intern. J. Pattern Recognit. Artif. Intell. 2015, 29, 1558001. [Google Scholar] [CrossRef]
  22. Mert, A.; Akan, A. Hilbert-Huang transform based hierarchical clustering for EEG denoising. In Proceedings of the 21st European Signal Processing Conference (EUSIPCO 2013), Marrakech, Morocco, 9–13 September 2013; pp. 1–5. [Google Scholar]
  23. Wang, G.; Teng, C.; Li, K.; Zhang, Z.; Yan, X. The removal of EOG artifacts from EEG signals using independent component analysis and multivariate empirical mode decomposition. IEEE J. Biomed. Health Inform. 2015, 20, 1301–1308. [Google Scholar] [CrossRef] [PubMed]
  24. Marino, M.; Liu, Q.; Koudelka, V.; Porcaro, C.; Hlinka, J.; Wenderoth, N.; Mantini, D. Adaptive optimal basis set for BCG artifact removal in simultaneous EEG-fMRI. Sci. Rep. 2018, 8, 8902. [Google Scholar] [CrossRef] [PubMed]
  25. Nam, H.; Yim, T.G.; Han, S.K.; Oh, J.B.; Lee, S.K. Independent component analysis of ictal EEG in medial temporal lobe epilepsy. Epilepsia 2002, 43, 160–164. [Google Scholar] [CrossRef] [PubMed]
  26. Urrestarazu, E.; Iriarte, J.; Alegre, M.; Valencia, M.; Viteri, C.; Artieda, J. Independent component analysis removing artifacts in ictal recordings. Epilepsia 2004, 45, 1071–1078. [Google Scholar] [CrossRef]
  27. De Clercq, W.; Vergult, A.; Vanrumste, B.; Van Paesschen, W.; Van Huffel, S. Canonical correlation analysis applied to remove muscle artifacts from the electroencephalogram. IEEE Trans. Biomed. Eng. 2006, 53, 2583–2587. [Google Scholar] [CrossRef]
  28. Zhang, L.; Wu, D.; Zhi, L. Method of removing noise from EEG signals based on HHT method. In Proceedings of the First International Conference on Information Science and Engineering, Nanjing, China, 26–28 December 2009; pp. 596–599. [Google Scholar]
  29. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.-C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. London. Ser. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  30. Chen, X.; Xu, X.; Liu, A.; McKeown, M.J.; Wang, Z.J. The use of multivariate EMD and CCA for denoising muscle artifacts from few-channel EEG recordings. IEEE Trans. Instrum. Meas. 2017, 67, 359–370. [Google Scholar] [CrossRef]
  31. Zeng, K.; Chen, D.; Ouyang, G.; Wang, L.; Liu, X.; Li, X. An EEMD-ICA approach to enhancing artifact rejection for noisy multivariate neural data. IEEE Trans. Neural Syst. Rehabil. Eng. 2015, 24, 630–638. [Google Scholar] [CrossRef]
  32. Chen, X.; Chen, Q.; Zhang, Y.; Wang, Z.J. A novel EEMD-CCA approach to removing muscle artifacts for pervasive EEG. IEEE Sens. J. 2018, 19, 8420–8431. [Google Scholar] [CrossRef]
  33. Sweeney, K.T.; McLoone, S.F.; Ward, T.E. The use of ensemble empirical mode decomposition with canonical correlation analysis as a novel artifact removal technique. IEEE Trans. Biomed. Eng. 2012, 60, 97–105. [Google Scholar] [CrossRef]
  34. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems 25; Curran Associates, Inc.: Red Hook, NY, USA, 2012. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  36. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  37. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1306.3781. [Google Scholar]
  38. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems 30; Curran Associates, Inc.: Red Hook, NY, USA, 2017. [Google Scholar]
  39. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to sequence learning with neural networks. In Advances in Neural Information Processing Systems 27; Curran Associates, Inc.: Red Hook, NY, USA, 2014. [Google Scholar]
  40. Sakhavi, S.; Guan, C.; Yan, S. Parallel convolutional-linear neural network for motor imagery classification. In Proceedings of the 23rd European Signal Processing Conference (EUSIPCO), Nice, France, 31 August–4 September 2015; pp. 2736–2740. [Google Scholar]
  41. Tabar, Y.R.; Halici, U. A novel deep learning approach for classification of EEG motor imagery signals. J. Neural Eng. 2016, 14, 016003. [Google Scholar] [CrossRef]
  42. Tang, Z.; Li, C.; Sun, S.J.O. Single-trial EEG classification of motor imagery using deep convolutional neural networks. Optik 2017, 130, 11–18. [Google Scholar] [CrossRef]
  43. Luo, T.-j.; Fan, Y.; Chen, L.; Guo, G.; Zhou, C. EEG signal reconstruction using a generative adversarial network with Wasserstein distance and temporal-spatial-frequency loss. Front. Neuroinform. 2020, 14, 15. [Google Scholar] [CrossRef] [PubMed]
  44. Corley, I.A.; Huang, Y. Deep EEG super-resolution: Upsampling EEG spatial resolution with generative adversarial networks. In Proceedings of the IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), Las Vegas, NV, USA, 4–7 March 2018; pp. 100–103. [Google Scholar]
  45. Hartmann, K.G.; Schirrmeister, R.T.; Ball, T. EEG-GAN: Generative adversarial networks for electroencephalographic (EEG) brain signals. arXiv 2018, arXiv:1806.01875. [Google Scholar]
  46. Fahimi, F.; Dosen, S.; Ang, K.K.; Mrachacz-Kersting, N.; Guan, C. Generative adversarial networks-based data augmentation for brain-computer interface. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4039–4051. [Google Scholar] [CrossRef]
  47. Yang, B.; Duan, K.; Fan, C.; Hu, C.; Wang, J. Automatic ocular artifacts removal in EEG using deep learning. Biomed. Signal Process. Control. 2018, 43, 148–158. [Google Scholar] [CrossRef]
  48. Leite, N.M.N.; Pereira, E.T.; Gurjao, E.C.; Veloso, L.R. Deep convolutional autoencoder for EEG noise filtering. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 2605–2612. [Google Scholar]
  49. Sun, W.; Su, Y.; Wu, X.; Wu, X. A novel end-to-end 1D-ResCNN model to remove the artifacts from EEG signals. Neurocomputing 2020, 404, 108–121. [Google Scholar] [CrossRef]
  50. Zhang, H.; Zhao, M.; Wei, C.; Mantini, D.; Li, Z.; Liu, Q. EEGdenoiseNet: A benchmark dataset for deep learning solutions of EEG denoising. J. Neural Eng. 2021, 18, 056057. [Google Scholar] [CrossRef]
  51. Zhang, H.; Wei, C.; Zhao, M.; Liu, Q.; Wu, H. A novel convolutional neural network model to remove muscle artifacts from EEG. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 1265–1269. [Google Scholar]
  52. Sawangjai, P.; Trakulruangroj, M.; Boonnag, C.; Piriyajitakonkij, M.; Tripathy, R.K.; Sudhawiyangkul, T.; Wilaiprasitporn, T. EEGANet: Removal of ocular artifact from the EEG signal using generative adversarial networks. IEEE J. Biomed. Health Inform. 2021, 26, 4913–4924. [Google Scholar] [CrossRef] [PubMed]
  53. Yu, J.; Li, C.; Lou, K.; Wei, C.; Liu, Q. Embedding decomposition for artifacts removal in EEG signals. J. Neural Eng. 2022, 19, 026052. [Google Scholar] [CrossRef] [PubMed]
  54. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  55. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  56. Asadi-Aghbolaghi, M.; Azad, R.; Fathy, M.; Escalera, S.J. Multi-level context gating of embedded collective knowledge for medical image segmentation. arXiv 2020, arXiv:2003.05056. [Google Scholar]
  57. Chaurasia, A.; Culurciello, E. Linknet: Exploiting encoder representations for efficient semantic segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  58. Ibtehaz, N.; Rahman, M.S. MultiResUNet: Rethinking the U-Net architecture for multimodal biomedical image segmentation. Neural Netw. 2020, 121, 74–87. [Google Scholar] [CrossRef] [PubMed]
  59. Huang, H.; Lin, L.; Tong, R.; Hu, H.; Zhang, Q.; Iwamoto, Y.; Han, X.; Chen, Y.-W.; Wu, J. Unet 3+: A full-scale connected unet for medical image segmentation. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1055–1059. [Google Scholar]
  60. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018; pp. 3–11. [Google Scholar]
  61. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. Proc. AAAI Conf. Artif. Intell. 2017, 31, 4278–4284. [Google Scholar] [CrossRef]
  62. Chen, X.; Peng, H.; Yu, F.; Wang, K. Independent vector analysis applied to remove muscle artifacts in EEG data. IEEE Trans. Instrum. Meas. 2017, 66, 1770–1779. [Google Scholar] [CrossRef]
  63. Weiergräber, M.; Papazoglou, A.; Broich, K.; Müller, R. Sampling rate, signal bandwidth and related pitfalls in EEG analysis. J. Neurosci. Methods 2016, 268, 53–55. [Google Scholar] [CrossRef]
  64. Zhu, Q.; Du, B.; Turkbey, B.; Choyke, P.L.; Yan, P. Deeply-supervised CNN for prostate segmentation. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 178–184. [Google Scholar]
  65. Mahmud, S.; Hossain, M.S.; Chowdhury, M.E.; Reaz, M.B.I. MLMRS-Net: Electroencephalography (EEG) motion artifacts removal using a multi-layer multi-resolution spatially pooled 1D signal reconstruction network. Neural Comput. Appl. 2023, 35, 8371–8388. [Google Scholar] [CrossRef]
  66. Xu, Z.-Q.J.; Zhang, Y.; Luo, T.; Xiao, Y.; Ma, Z. Frequency principle: Fourier analysis sheds light on deep neural networks. arXiv 2019, arXiv:1901.06523. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions, and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions, or products referred to in the content.
Figure 1. Proposed framework for removing physiological artifacts from EEG signals.
Figure 1. Proposed framework for removing physiological artifacts from EEG signals.
Bioengineering 10 00579 g001
Figure 2. The proposed MultiResUNet3+ segmentation model architecture.
Figure 2. The proposed MultiResUNet3+ segmentation model architecture.
Bioengineering 10 00579 g002
Figure 3. Full-Scale Aggregated Feature Map Creation for the third decoder Layer X D e 3 aided by Residual Paths (ResPaths in short), following the process proposed in [59].
Figure 3. Full-Scale Aggregated Feature Map Creation for the third decoder Layer X D e 3 aided by Residual Paths (ResPaths in short), following the process proposed in [59].
Bioengineering 10 00579 g003
Figure 4. Residual Path architecture adapted from [58] used for inter- and intra-connections in MultiResUNet3+ is expected to reduce the semantic gaps by replacing direct skip connections. This particular ResPath represents the one produced from X E n 1 and X D e 2 among the inter- and intra-ResPaths, respectively, in Figure 2.
Figure 4. Residual Path architecture adapted from [58] used for inter- and intra-connections in MultiResUNet3+ is expected to reduce the semantic gaps by replacing direct skip connections. This particular ResPath represents the one produced from X E n 1 and X D e 2 among the inter- and intra-ResPaths, respectively, in Figure 2.
Bioengineering 10 00579 g004
Figure 5. Structure of the Modified MultiResidual or MultiRes Block inspired from [58].
Figure 5. Structure of the Modified MultiResidual or MultiRes Block inspired from [58].
Bioengineering 10 00579 g005
Figure 6. Sample (a) clean or ground truth EEG, (b) pure vertical EOG, (c) noise-free horizontal EOG, and (d) clean EMG segment with actual amplitudes. Noticeably, EMG has a much higher magnitude compared to clean EEG and EOG segments.
Figure 6. Sample (a) clean or ground truth EEG, (b) pure vertical EOG, (c) noise-free horizontal EOG, and (d) clean EMG segment with actual amplitudes. Noticeably, EMG has a much higher magnitude compared to clean EEG and EOG segments.
Bioengineering 10 00579 g006
Figure 7. Superimposed arbitrary example segments of (a) EOG-corrupted EEG and corresponding ground truth EEG, (b) EMG-contaminated EEG and corresponding clean EEG, (c) simultaneous EOG and EMG-corrupted EEG with corresponding ground truth EEG.
Figure 7. Superimposed arbitrary example segments of (a) EOG-corrupted EEG and corresponding ground truth EEG, (b) EMG-contaminated EEG and corresponding clean EEG, (c) simultaneous EOG and EMG-corrupted EEG with corresponding ground truth EEG.
Bioengineering 10 00579 g007
Figure 8. Estimated performance metrics (CC in time domain, RRMSE in time, and frequency domain vs. SNR) of five DL models after denoising EOG-contaminated EEG.
Figure 8. Estimated performance metrics (CC in time domain, RRMSE in time, and frequency domain vs. SNR) of five DL models after denoising EOG-contaminated EEG.
Bioengineering 10 00579 g008
Figure 9. Estimated performance metrics (CC in time domain, RRMSE in time, and frequency domain vs. SNR) of five DL models after removing EMG artifacts from EMG-contaminated EEG.
Figure 9. Estimated performance metrics (CC in time domain, RRMSE in time, and frequency domain vs. SNR) of five DL models after removing EMG artifacts from EMG-contaminated EEG.
Bioengineering 10 00579 g009
Figure 10. Estimated performance metrics (CC in time domain, RRMSE in time, and frequency domain vs. SNR) of five DL models after denoising simultaneous EOG and EMG-contaminated EEG.
Figure 10. Estimated performance metrics (CC in time domain, RRMSE in time, and frequency domain vs. SNR) of five DL models after denoising simultaneous EOG and EMG-contaminated EEG.
Bioengineering 10 00579 g010
Figure 11. (a) A sample EOG-corrupted EEG segment, EOG artifacts-free EEG segment predicted by (b) FPN, (c) UNet, (d) MCGUNet, (e) LinkNet and (f) MultiResUNet3+ models plotted against the ground truth EEG.
Figure 11. (a) A sample EOG-corrupted EEG segment, EOG artifacts-free EEG segment predicted by (b) FPN, (c) UNet, (d) MCGUNet, (e) LinkNet and (f) MultiResUNet3+ models plotted against the ground truth EEG.
Bioengineering 10 00579 g011
Figure 12. (a) An example EMG-corrupted EEG segment, EMG artifacts-free EEG segment predicted by (b) FPN, (c) UNet, (d) MCGUNet, (e) LinkNet, and (f) MultiResUNet3+ networks along with the ground truth EEG.
Figure 12. (a) An example EMG-corrupted EEG segment, EMG artifacts-free EEG segment predicted by (b) FPN, (c) UNet, (d) MCGUNet, (e) LinkNet, and (f) MultiResUNet3+ networks along with the ground truth EEG.
Bioengineering 10 00579 g012
Figure 13. (a) An example simultaneous EOG and EMG-contaminated EEG segment. Simultaneous EOG and EMG-artifacts-free EEG segments created by (b) FPN, (c) UNet, (d) MCGUNet, (e) LinkNet, and (f) multiResUNet3+ networks along with the ground truth EEG.
Figure 13. (a) An example simultaneous EOG and EMG-contaminated EEG segment. Simultaneous EOG and EMG-artifacts-free EEG segments created by (b) FPN, (c) UNet, (d) MCGUNet, (e) LinkNet, and (f) multiResUNet3+ networks along with the ground truth EEG.
Bioengineering 10 00579 g013
Table 1. EOG-contaminated EEG segment denoising performance. The bold numeric values represent the best performance.
Table 1. EOG-contaminated EEG segment denoising performance. The bold numeric values represent the best performance.
Performance MetricSNR
(dB)
Model
FPNUNetMCGUNetLinkNetMultiResUNet3+
(Proposed)
CC
(Temporal)
−70.80740.82000.79130.82690.8636
−60.83200.85070.83080.85660.8865
−50.85220.86460.84590.85790.9032
−40.86950.89120.85060.87930.9152
−30.88190.89860.85970.89460.9329
−20.90040.92990.86170.92040.9347
−10.91360.93890.87100.92610.9437
00.92860.94200.87830.93880.9534
10.93220.94420.88440.94940.9633
20.94930.95540.89480.95370.9644
RRMSE
(Temporal)
−70.59030.56930.57330.55540.4694
−60.55670.51460.53780.50980.4279
−50.51840.49700.49820.51460.3954
−40.49390.43770.49420.47930.3728
−30.47390.43410.47680.44160.3193
−20.43570.33860.47850.37350.3232
−10.39490.32080.46980.36620.3021
00.35520.31340.46300.32560.2674
10.35010.30730.45840.28400.2413
20.29070.26660.41970.27440.2338
RRMSE
(Spectral)
−70.62190.58540.50660.56770.4419
−60.58270.52780.48070.53280.4234
−50.56010.52800.43720.55140.3948
−40.54710.44920.40480.53310.3827
−30.51640.47540.40220.46870.3219
−20.49020.36050.39920.39960.3382
−10.43540.34980.38490.41310.3224
00.38340.34710.36150.36210.2878
10.38970.33840.35840.31340.2707
20.31680.29920.32070.30800.2668
Table 2. EMG-contaminated EEG segment denoising performance where the bold numeric values represent the best performance.
Table 2. EMG-contaminated EEG segment denoising performance where the bold numeric values represent the best performance.
Performance MetricSNR
(dB)
Model
FPNUNetMCGUNetLinkNetMultiResUNet3+
(Proposed)
CC
(Temporal)
−70.57790.55480.58620.55520.5897
−60.64250.61980.65330.62240.6463
−50.70130.68790.70950.68640.7056
−40.75450.74470.75610.74480.7589
−30.80150.79920.80350.79730.8039
−20.84380.84180.84100.84170.8447
−10.87850.88030.87930.87930.8826
00.910.91130.90880.91060.9092
10.93360.93450.93360.93430.9343
20.95170.95290.95190.95260.9536
RRMSE
(Temporal)
−70.81360.82660.80480.82630.8049
−60.76720.78130.74860.77920.7597
−50.71190.72150.69940.72490.7087
−40.65440.66540.65360.66430.6483
−30.59630.60030.58840.60080.5924
−20.53450.53780.53550.53870.5333
−10.47750.47380.47460.47410.4695
00.41300.41020.41510.41150.4158
10.35700.35480.35710.35660.3565
20.30550.30290.29680.30340.3058
RRMSE
(Spectral)
−70.79200.81270.76700.81570.7686
−60.76220.75400.72550.75810.7087
−50.69050.68570.63890.67160.6643
−40.59970.61800.58700.61620.5926
−30.55120.54040.51670.55080.5344
−20.48430.48360.46640.48250.4770
−10.42060.42080.42830.42460.4042
00.36850.35950.35520.36170.3615
10.31150.30540.29080.30640.2971
20.25440.25270.23610.25200.2479
Table 3. Simultaneous EOG and EMG-contaminated EEG segments denoising performance. The bold numeric values represent the best performance.
Table 3. Simultaneous EOG and EMG-contaminated EEG segments denoising performance. The bold numeric values represent the best performance.
Performance MetricSNR
(dB)
Model
FPNUNetMCGUNetLinkNetMultiResUNet3+
(Proposed)
CC
(Temporal)
−70.57710.58560.66300.59870.6152
−60.63700.63550.71130.65120.6582
−50.69320.70390.74490.70520.7188
−40.75070.75650.80090.75760.8191
−30.79820.80030.83880.80100.8245
−20.84230.84280.88190.84310.8580
−10.87570.88030.88320.87800.8934
00.90250.90840.90940.90980.9228
10.93080.93250.91000.93320.9411
20.94960.95040.92770.95060.9579
RRMSE
(Temporal)
−70.81980.81220.72350.79890.7830
−60.77580.77550.68730.76380.7446
−50.72470.71890.64790.70880.6883
−40.67060.65240.56880.65320.5440
−30.61080.60170.52220.59940.5675
−20.54380.54580.44760.53700.5108
−10.48710.47360.45260.48210.4480
00.43380.42070.40540.41430.3821
10.36640.36030.41830.36100.3375
20.31590.31180.37440.31070.2867
RRMSE (Spectral)−70.82020.79030.68940.74470.7017
−60.78420.74290.63640.73990.6587
−50.72220.70880.56900.65780.5879
−40.68120.59990.48650.59690.4401
−30.61360.53410.45450.55590.4768
−20.52160.50270.36980.49720.4188
−10.47010.42870.36250.42620.3577
00.39450.37630.31960.35760.2934
10.32890.31340.33740.30480.2542
20.27160.26350.27220.26220.2142
Table 4. Measured performance metrics after denoising EOG-contaminated EEG where the bold numeric values represent the best performance.
Table 4. Measured performance metrics after denoising EOG-contaminated EEG where the bold numeric values represent the best performance.
Model η (%) γ (%) R R M S E t e m p o r a l   ± STD R R M S E s p e c t r a l   ± STD
FPN88.2681.310.23141 ± 0.072500.25518 ± 0.08122
UNet94.5991.590.11342 ± 0.038250.12123 ± 0.04411
MCGUNet73.2870.390.43087 ± 0.087750.43669 ± 0.13015
LinkNet94.4091.300.12072 ± 0.041340.13229 ± 0.04826
MultiResUNet3+ (Proposed)94.8292.840.13190 ± 0.053640.13660 ± 0.05587
Table 5. Estimated power ratios of five distinct EEG frequency bands for the ground truth, noisy, and predicted EOG-artifacts-free EEG segments. The bold numeric values represent the best performance.
Table 5. Estimated power ratios of five distinct EEG frequency bands for the ground truth, noisy, and predicted EOG-artifacts-free EEG segments. The bold numeric values represent the best performance.
Model/MethodDeltaThetaAlphaBetaGamma
FPN0.41470.54290.12780.07970.0208
UNet0.43380.52970.12300.07610.0200
MCGUNet0.39970.60570.12890.05430.0101
LinkNet0.43200.53180.12340.07620.0199
MultiResUNet3+ (Proposed)0.43370.52950.12330.07600.0197
EOG-contaminated EEG0.83010.16390.03680.02440.0069
Ground Truth EEG0.44590.51840.12060.07490.0195
Table 6. Measured performance metrics after denoising EMG-contaminated EEG. The bold numeric values represent the best performance.
Table 6. Measured performance metrics after denoising EMG-contaminated EEG. The bold numeric values represent the best performance.
Model η (in %) γ (in %) R R M S E t e m p o r a l   ± STD R R M S E s p e c t r a l   ± STD
FPN85.0675.960.31727 ± 0.111520.26395 ± 0.11695
UNet89.5980.610.22394 ± 0.077850.19231 ± 0.08107
MCGUNet87.1979.370.29214 ± 0.105200.23798 ± 0.10421
LinkNet88.9482.600.25091 ± 0.101120.19961 ± 0.10910
MultiResUNet3+ (Proposed)89.3383.210.23769 ± 0.088410.18931 ± 0.08931
Table 7. Power ratios (estimated) of five distinct EEG frequency bands for the ground truth, noisy, and predicted EMG-artifacts-free EEG. The bold numeric values represent the best performance.
Table 7. Power ratios (estimated) of five distinct EEG frequency bands for the ground truth, noisy, and predicted EMG-artifacts-free EEG. The bold numeric values represent the best performance.
Model/MethodDeltaThetaAlphaBetaGamma
FPN0.45000.55340.11150.05660.0125
UNet0.44520.54520.11500.06120.0147
MCGUNet0.43930.54360.11930.06480.0155
LinkNet0.44800.53030.11950.06560.0162
MultiResUNet3+ (Proposed)0.44530.53440.11780.06580.0165
EMG contaminated EEG0.13330.10460.06220.20790.5514
Ground Truth EEG0.44210.51630.12300.07560.0197
Table 8. Measured performance metrics after denoising simultaneous EOG and EMG-contaminated EEG. The bold numeric values represent the best performance.
Table 8. Measured performance metrics after denoising simultaneous EOG and EMG-contaminated EEG. The bold numeric values represent the best performance.
Model η (in %) γ (in %) R R M S E t e m p o r a l   ± STD R R M S E s p e c t r a l   ± STD
FPN84.1977.970.34448 ± 0.138490.27145 ± 0.15104
UNet89.6383.630.22991 ± 0.096190.18805 ± 0.10819
MCGUNet87.0380.310.29993 ± 0.109710.25156 ± 0.11219
LinkNet89.7783.390.22439 ± 0.091800.18340 ± 0.10208
MultiResUNet3+ (Proposed)89.1682.710.24891 ± 0.094450.20085 ± 0.09989
Table 9. Estimated power ratios of five distinct EEG frequency bands for the ground truth, noisy, and predicted simultaneous EOG and EMG-artifacts-free EEG segments. The bold numeric values represent the best performance.
Table 9. Estimated power ratios of five distinct EEG frequency bands for the ground truth, noisy, and predicted simultaneous EOG and EMG-artifacts-free EEG segments. The bold numeric values represent the best performance.
Model/MethodDeltaThetaAlphaBetaGamma
FPN0.45430.53320.11850.06470.0153
UNet0.45370.53040.11780.06500.0158
MCGUNet0.45430.54430.11410.05950.0143
LinkNet0.45370.52880.11750.06590.0163
MultiResUNet3+ (Proposed)0.45320.54160.11310.06080.0151
Simultaneous EOG-EMG contaminated EEG0.13550.10460.06140.01960.5503
Ground Truth EEG0.44580.51800.12080.07490.0196
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hossain, M.S.; Mahmud, S.; Khandakar, A.; Al-Emadi, N.; Chowdhury, F.A.; Mahbub, Z.B.; Reaz, M.B.I.; Chowdhury, M.E.H. MultiResUNet3+: A Full-Scale Connected Multi-Residual UNet Model to Denoise Electrooculogram and Electromyogram Artifacts from Corrupted Electroencephalogram Signals. Bioengineering 2023, 10, 579. https://doi.org/10.3390/bioengineering10050579

AMA Style

Hossain MS, Mahmud S, Khandakar A, Al-Emadi N, Chowdhury FA, Mahbub ZB, Reaz MBI, Chowdhury MEH. MultiResUNet3+: A Full-Scale Connected Multi-Residual UNet Model to Denoise Electrooculogram and Electromyogram Artifacts from Corrupted Electroencephalogram Signals. Bioengineering. 2023; 10(5):579. https://doi.org/10.3390/bioengineering10050579

Chicago/Turabian Style

Hossain, Md Shafayet, Sakib Mahmud, Amith Khandakar, Nasser Al-Emadi, Farhana Ahmed Chowdhury, Zaid Bin Mahbub, Mamun Bin Ibne Reaz, and Muhammad E. H. Chowdhury. 2023. "MultiResUNet3+: A Full-Scale Connected Multi-Residual UNet Model to Denoise Electrooculogram and Electromyogram Artifacts from Corrupted Electroencephalogram Signals" Bioengineering 10, no. 5: 579. https://doi.org/10.3390/bioengineering10050579

APA Style

Hossain, M. S., Mahmud, S., Khandakar, A., Al-Emadi, N., Chowdhury, F. A., Mahbub, Z. B., Reaz, M. B. I., & Chowdhury, M. E. H. (2023). MultiResUNet3+: A Full-Scale Connected Multi-Residual UNet Model to Denoise Electrooculogram and Electromyogram Artifacts from Corrupted Electroencephalogram Signals. Bioengineering, 10(5), 579. https://doi.org/10.3390/bioengineering10050579

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop