Next Article in Journal
Deciphering Post-Stroke Sleep Disorders: Unveiling Neurological Mechanisms in the Realm of Brain Science
Next Article in Special Issue
Movement Termination of Slow-Wave Sleep—A Potential Biomarker?
Previous Article in Journal
The Effect of Upright Stance and Vision on a Cognitive Task in Elderly Subjects and Patients with Parkinson’s Disease
Previous Article in Special Issue
Stereotactic Electroencephalogram Recordings in Temporal Lobectomy Patients Demonstrates the Predictive Value of Interictal Cross-Frequency Correlations: A Retrospective Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluation of the Relation between Ictal EEG Features and XAI Explanations

by
Sergio E. Sánchez-Hernández
,
Sulema Torres-Ramos
,
Israel Román-Godínez
and
Ricardo A. Salido-Ruiz
*
Division of Cyber-Human Interaction Technologies, University of Guadalajara (UdG), Guadalajara 44100, Mexico
*
Author to whom correspondence should be addressed.
Brain Sci. 2024, 14(4), 306; https://doi.org/10.3390/brainsci14040306
Submission received: 28 January 2024 / Revised: 20 March 2024 / Accepted: 21 March 2024 / Published: 25 March 2024

Abstract

:
Epilepsy is a neurological disease with one of the highest rates of incidence worldwide. Although EEG is a crucial tool for its diagnosis, the manual detection of epileptic seizures is time consuming. Automated methods are needed to streamline this process; although there are already several works that have achieved this, the process by which it is executed remains a black box that prevents understanding of the ways in which machine learning algorithms make their decisions. A state-of-the-art deep learning model for seizure detection and three EEG databases were chosen for this study. The developed models were trained and evaluated under different conditions (i.e., three distinct levels of overlap among the chosen EEG data windows). The classifiers with the best performance were selected, then Shapley Additive Explanations (SHAPs) and Local Interpretable Model-Agnostic Explanations (LIMEs) were employed to estimate the importance value of each EEG channel and the Spearman’s rank correlation coefficient was computed between the EEG features of epileptic signals and the importance values. The results show that the database and training conditions may affect a classifier’s performance. The most significant accuracy rates were 0.84 , 0.73 , and 0.64 for the CHB-MIT, Siena, and TUSZ EEG datasets, respectively. In addition, most EEG features displayed negligible or low correlation with the importance values. Finally, it was concluded that a correlation between the EEG features and the importance values (generated by SHAP and LIME) may have been absent even for the high-performance models.

1. Introduction

Epilepsy is a neurological disease that presents recurrent and unprovoked seizures [1]. This disease affects the physical health of the patient. It can impact mental health and significantly decrease quality of life [1,2].
According to the World Health Organization (WHO), it is estimated that around 50 million people worldwide have epilepsy, with an annual diagnosis rate of approximately five million people [3]. In the United States, the Center for Surveillance, Epidemiology, and Laboratory Services estimated that in 2010 2.3 million adults had active epilepsy, and that this number had increased to three million adults by 2015 [4]. From January to September 2019, patients undergoing assessment or treatment for epilepsy accounted for 2,998,000 consultations in the medical units of the Mexican Institute of Social Security (IMSS) [5]. In [6], through the analysis of six epilepsy studies in Mexico, the prevalence rate was found to be between 3.9 and 42.2 cases per thousand inhabitants. The prevalence varied significantly by location.
The electroencephalogram (EEG) is one of the main tools used for the diagnosis of epilepsy. Manually analyzing EEG recordings is very time consuming; therefore, using computational tools for their analysis and characterization improves the diagnostic process [7]. Several of the computational tools used to analyze EEG recordings are derived from the artificial intelligence field, including machine learning (ML) and deep learning (DL).
On the one hand, ML has been widely applied in epilepsy for seizure detection, differentiation of seizure states, and localization of seizure foci [8]. Conversely, the use of DL techniques in epilepsy has increased regarding classification and prediction tasks [9]. As mentioned by [7], these DL techniques are usually reliant on nontransparent models.
In interesting work presented by [10], the authors proposed novel representations, namely, the unigram ordinal pattern (UniOP) and bigram ordinal pattern (BiOP), to capture underlying dynamics in EEG time series for seizure detection. Their approach demonstrated high accuracy in discriminating between healthy and seizure states, outperforming existing methods. A recent work [11] employed a combination of Variable-Frequency Complex Demodulation (VFCDM) and Convolutional Neural Networks (CNN) to discriminate between health, interictal, and ictal states using electroencephalogram (EEG) data then evaluating CNN performance through leave-one-subject-out cross-validation (LOSO CV), achieving consistently high accuracy rates between healthy and epileptic states.
Unfortunately, it is difficult to explain the decisions of a nontransparent model, and, as argued in [12], explanations of the predictions are necessary to justify the reliability of the models. A solution to this problem can be found in the explainable artificial intelligence (XAI) field, which aims to produce justifications that facilitate the comprehension of a model’s functioning and of the rationale behind its decisions, allowing end users to trust the model [13].
In the state of the art, several studies have explored the utilization of XAI for the analysis of EEG signals and the detection of seizures. In [14], a deep neural network for seizure detection was designed. The model was subjected to adversarial training in order to acquire seizure representations from EEG signals. Additionally, an attention mechanism was implemented to assess the significance of individual EEG channels.
In [15], the connectivity characteristics of EEG signals were estimated, then a set of neural networks was trained to detect seizures. The activation values of the neurons in the classifiers were utilized to estimate the relevance of the characteristics. The findings indicated that the relevance values varied for each subject.
In [16], the authors discussed the application of DL algorithms for the diagnosis of epilepsy. In addition, the use of XAI to explain the model’s decisions was explored. Among the various XAI techniques evaluated, only attention pooling could extract the most significant segments from the signals. However, it was suggested that epileptographic patterns may be too complex to be captured using attention pooling.
Other XAI techniques, such as Shapley Additive Explanations (SHAP) [17], have been employed for different tasks. One such task the classification of the pre-ictal and inter-ictal phases, as described in [18], where the authors used SHAP to assess each EEG channel’s significance and demonstrated how this significance varied over time. Another task for which SHAP has been utilized was presented in [19], where the detection of epileptic seizures from time–frequency domain transformations of EEG was studied using neural networks. Here, the task for SHAP was to visually identify the frequencies that contributed most to the classification.
In [20], the authors proposed a system which uses a Bi-LSTM network for classification of normal and abnormal signals caused by epilepsy and the Layerwise Relevance Propagation (LRP) XAI method to explain the predictions of the network. The LRP method generates a relevance vector for the test input vector. The authors reported that these relevance values indicate the contribution of each datapoint of a signal, helping to classify signals into a particular class.
In another work, SHAP formed part of a methodology known as XAI4EEG developed for the detection of seizures and explanation of the model’s decisions [21]. This technique consists of extracting the time and frequency features of EEG signals for classification by two convolutional neural networks, with SHAP implemented for explanation generation.
In [22], the authors performed minor signal processing steps such as filtering, and used the discrete wavelet transform (DWT) to decompose the EEG signals and extract various eigenvalue features of the statistical time domain (STD) as linear and Fractal Dimension-based Nonlinear (FD-NL) features. Following this feature extraction step, the optimal features were identified through correlation coefficients with p-value and distance correlation analysis and classified using a Bagged Tree-Based Classifer (BTBC), followed by SHAP to provide the explanations.
Based on the understanding that DL models can identify patterns in epileptic EEG signals, are these patterns alone helpful? In addition, if we add transparency using XAI, could the explanations help to identify ictal EEG patterns?
The main objective of this work is to evaluate the utility of explanations generated by XAI techniques such as SHAP and Local Interpretable Model-Agnostic Explanations (LIME) in identifying epileptiform patterns in EEG signals. The aim is to determine whether these explanations can enhance the understanding of deep learning (DL) models and assist in identifying ictal patterns in EEG signals. To achieve this, three EEG databases and a state-of-the-art DL model are utilized to evaluate the models’ performance under different training conditions. Moreover, EEG features are computed for each channel, and the Spearman’s rank correlation coefficient between these features and the importance values generated by XAI techniques is assessed. In summary, this study aims to highlight the complexity involved in identifying ictal patterns from DL models and to explore the role of transparency provided by XAI techniques in this process.

2. Materials and Methods

This section describes the methods used for training the models, implementing XAI, and estimating the features. A general methodological overview is shown in Figure 1.

2.1. Datasets

The CHB-MIT Scalp EEG database was presented in [23] and is available in a repository [24]. It consists of surface electroencephalograms of 23 pediatric patients with epilepsy: five males (ages 3–22) and 17 females (ages 1.5–19). The patients were undergoing medication withdrawal for epilepsy surgery evaluation. One subject was recorded at an interval of 1.5 years; these recordings are considered as two different cases. There was no gender or age information for one subject. The EEG signals were sampled at a frequency of 256 Hz. The electrodes were placed following the 10/20 system. Most of the recordings contain channels of the longitudinal bipolar montage. As mentioned in [23], the dataset was mostly segmented into 1 h recordings. The recording device is not mentioned.
The Siena Scalp EEG database was presented in [25] and is available at [26]. This dataset contains surface EEG recordings from 14 adult patients: nine males (ages 36–71) and five females (ages 20–58). The recordings have a duration between 1 h and 13 h. The signals were recorded with a sampling frequency of 512 Hz. The electrodes were placed according to the 10/20 system. The channels were monopolar. EB Neuro and Natus Quantum LTM amplifiers were utilized for data acquisition. The subjects were monitored using a video scalp EEG and were asked to stay in bed most of the time.
The TUH EEG Seizure Corpus (TUSZ) was presented in [27,28]. It consists of surface EEG recordings from 315 subjects obtained over 822 sessions. Only 280 sessions contain a seizure. The gender composition is 153 males and 162 females. The dataset includes both pediatric and adult patients. Most EEG recordings have a duration between 0 and 30 min. The sampling frequency varied per patient (250 Hz, 256 Hz, 400 Hz, and 512 Hz). The channels were monopolar. The recording devices are not mentioned. Due to the breadth of this database, the recording protocols, ages, and recording durations are varied (refer to [27] for detailed information).

2.2. Data Pre-Processing

EEG recordings were discarded based on the following criteria:
  • They did not contain ictal activity;
  • The channel list was different from that of the rest of the recordings;
  • The sampling frequency was different from that of the rest of the recordings;
  • There was only a single recording for each patient.
Another consideration for this study was the use of bipolar longitudinal montages. This choice was motivated by the prevalence of databases presenting recordings in this format and the advantage of transforming monopolar montages into bipolar rather than vice versa.
Due to the large amount of data, only a subset of patients was selected from the TUSZ dataset. The number of patients, channels, and seizure types is detailed in Table 1. Finally, a Notch filter was applied to the EEG windows to remove the power line frequency. Moreover, a second-order high-pass Butterworth filter was used to remove frequencies under 0.5 Hz.

2.3. Dataset Segmentation, Training, and Testing

According to the methodology described in [29], the EEG signals were segmented into windows of 2 s. In order to ensure that both classes, ictal and non-ictal, were represented equally during training and validation, a two-step balancing process was implemented: the first step consisted of oversampling the ictal epochs with three different overlap ratios ( 80 % , 70 % , and 50 % ), while the second step involved subsampling from the larger class. While evaluating the DL models, the windows were created without overlap to avoid utilizing known data, and both classes were balanced. From the training set, 20% was used as the validation set.

2.4. Deep Learning Models

A literature search was conducted to select the DL algorithm. The search criteria were as follows:
  • Papers were included in PubMed and Clarivate;
  • Papers were published after 2015;
  • Papers were available via open access;
  • Papers described bi-class classification (ictal and non-ictal);
  • Papers described the use of raw EEG data;
  • Papers described the implementation of a deep learning model;
  • Papers described patient-specific models;
  • Papers described the use of performance metrics.
After reviewing the papers meeting the above criteria, the model presented by [30] was selected. Table 2 displays the performance and characteristics of the model. In [30], two approaches were tested: segment-based and event-based. The segment-based approach was relevant to the present research.
The model is a Siamese convolutional neural network (CNN), and is illustrated in Figure 2. The convolutional layers utilize ReLU as the activation function, while Softmax is applied to the output layer for activation. A dropout layer is appended between the fully connected layers, with a dropout rate equal to 0.25. Throughout the remainder of this document, this model is referred to as Wang_1d. We used different parameters than those used in [30]. The Adam algorithm was employed as the optimizer, using the following parameters: α = 0.001 , β 1 = 0.9 , β 2 = 0.999 , and ϵ = 1 × 10 7 . Binary cross-entropy was used as the loss function and accuracy as the evaluation metric. The number of epochs and the batch size were both 100.

2.5. Model Evaluation

The models were trained and evaluated using an intra-patient approach. A similar approach to leave-one-out cross-validation was implemented for model evaluation. A neural network was trained k times using a specific overlap rate, with k representing the number of recordings per patient. In the first iteration, k 1 recordings were utilized for training and the remaining recordings were used for evaluation.
The specificity, sensitivity, accuracy, AUC–ROC, and F1-score were calculated for each iteration. Specificity and sensitivity were used to measure the capacity of the model to detect true positives and true negatives. Accuracy and F1-score provide a general measure of performance, and are commonly computed for seizure detection models, while AUC-ROC helps to visualize the separability between classes.

2.6. Feature Computation

The EEG windows were separated into the following three sub-bands: low frequencies (<12 Hz), beta (12–25 Hz), and gamma (>25 Hz). Second-order Butterworth filters were used for band separation.
A set of 14 features that simplified the behavior of segmented bipolar EEG signals by capturing some of their key characteristics was computed; these are mentioned in Table 3. Features were computed for each sub-band separately, and primal EEG epochs were considered during computation. The total number of features was 14 f e a t u r e s × 4 b a n d s = 56 . The selection of features was based on previous research [31]. Although an exhaustive analysis was not performed, the selected features have been utilized in previous studies of epilepsy.

2.7. Explainable Artificial Intelligence

In this work, the Shapley Additive Explanations (SHAP) and Local Interpretable Model-Agnostic Explanations (LIME) techniques were chosen to assess the model’s explainability.

2.7.1. Shapley Additive Explanations

SHAP is an approach for interpreting the predictions made by ML models [17]. To evaluate an instance, each attribute is assigned a SHAP value, which indicates the relative importance of the attribute to the model’s decision-making process. The formal definition is as follows:
ϕ i ( f , x ) = z x | z | ( M | z | 1 ) ! M ! [ f x ( z ) f x ( z i ) ]
where x is the instance to be explained, f is the model, i is the feature to be evaluated, and M is the number of features. Additionally, x contains all possible perturbations of x.
In this work, we utilized the PartitionSHAP algorithm [38], which is a component of the software package introduced by the authors of [17]. This algorithm allows for the computation of importance values by evaluating a group or coalition of features. Consequently, the features of a given coalition receive the same SHAP value.
PartionSHAP computes Shapley values using a hierarchical approach that defines coalitions and returns Owen values [39]. The hierarchy depth allows the coalition size to be determined. We applied PartitionSHAP solely to the classification model exhibiting the best overall performance (refer to Section 3 for further details). Subsequently, we estimated the SHAP values for those instances correctly classified as ictal.
The depth of the PartitionSHAP hierarchy was tailored such that each coalition corresponded to a single channel of the EEG window (see Figure 3). Consequently, a single importance value was computed per channel. Two SHAP values were obtained for specific channel pairs, as the PartitionSHAP hierarchy operates through powers of 2. In such cases, the importance of these coalitions was determined as the mean value of both SHAP values.

2.7.2. Local Interpretable Model-Agnostic Explanations

LIME is an approach for explaining the predictions of any classification model by approximating it locally using an interpretable model [40]. An interpretable model should provide an understanding of both the input variables and the response.
The formal definition of LIME’s explanation is as follows:
ξ ( x ) = arg min g G L ( f , g , π x ) + Ω ( g )
where x is the instance to be explained, ξ is the instance explanation, g is a potentially interpretable model such as a linear model or decision tree, and f is the classification model. The function L measures the approximation of g to f in the locality defined by π x . The complexity of g is measured by Ω ( g ) ; this parameter is related to the complexity of the model g.
To approximate L ( f , g , π x ) , a set of instances around x is sampled. The sampling is performed by perturbing the features of x, with the nearer instances having higher weights. Later, the model g is adjusted on the basis of the perturbed dataset by considering the instance weights.
In this work, we utilized the LimeImageExplainer algorithm introduced by the authors of [40]. This method considers the neighbors’ features as superpixels. Therefore, a superpixel’s members receive the same importance value. Additionally, the algorithm uses the Euclidean distance to weight the perturbed instances and Ridge regression as the interpretable model. The adjusted regressor coefficients are considered as the importance values.
The method was adjusted to evaluate each channel as a group (see Figure 4); therefore, each channel received a unique importance value. Similar to PartitionSHAP, LIME was solely applied to the classification model exhibiting the best overall performance (refer to Section 3 for further details).

2.8. Correlation Computation

Every channel has an importance value and a corresponding set of features. For a given feature (e.g., the complexity of the beta band), we computed the Spearman’s rank correlation coefficient [41] between the feature and the importance values. This analysis was repeated for every feature and patient. During the correlation experiments, the non-ictal windows were not considered, as this research aimed to understand the behavior of the ictal stage.

2.9. Computing Hardware

The experiments were conducted on two devices with the following technical specifications: a computer equipped with an Intel Core i7 processor, 12 GB of RAM, and Ubuntu 18.04, and a server equipped with an Intel Xeon Gold processor, 125 GB of RAM, and an NVIDIA Tesla P100, Python 3.7 [42] was employed to code all experiments, along with various open-access Python libraries.

3. Results

In this section, the results of our experiments are presented. First, we present the model’s performance when trained with varying percentages of overlapping windows for each of the EEG datasets. Finally, we examine the correlation between the importance values and the EEG features.

3.1. Model Performance

This analysis aimed to visualize the variations in performance for a deep learning neural network subjected to various conditions. The varied conditions were as follows. Overlapping was applied to the training set and the EEG dataset used for training; the window overlapping on the training set functioned as a data augmentation technique to help balance the classes. In order to avoid biasing the results, the test set was not augmented.
Table 4 presents the results of the classification models. The metrics were averaged across patients. The performance on the CHB-MIT database was superior to that on the Siena and TUSZ datasets. The sensitivity and specificity were greater than 0.82 and 0.60 , respectively.
Considering that the test set is balanced and that we are modeling a bi-class problem, accuracy is a good metric for evaluating overall performance. In this sense, the highest accuracy was obtained using a 80 % overlap for each dataset: 0.84 (CHB-MIT), 0.73 (Siena), and 0.64 (TUSZ). As expected, this may imply that using more instances used during training results in better modeling. Note that none of the individual test sets were augmented. Additionally, even when the accuracy metric was calculated, sensitivity and specificity were computed as well in order to obtain a deeper view of the ictal and non-ictal predictions. It should be noted that the model’s sensitivity was always greater than its specificity, which in our case implies that the model performed better on identifying the target class (ictal). The difference between the sensitivity and specificity was considerable for the Siena dataset ( 50 % and 70 % overlap rate, respectively).
The specificity rose when the overlap was increased; on the other hand, the sensitivity tended to decrease. Generally speaking, the model’s accuracy was greater as the overlap rate increased. Contrary to the Siena and CHB-MIT datasets, on the TUSZ dataset these effects might be a consequence of the mixture of participants, particularly the mixture of adult and pediatric subjects. It is important to consider that the results in Table 4 correspond to the average of all participants per dataset. The largest F1-score was displayed when using an overlap rate of 80 % : 0.83 (CHB-MIT), 0.71 (Siena), and 0.62 (TUSZ). Unlike the accuracy, it did not follow an ascending trend. Finally, the AUC-ROC score was over 0.85 for the CHB-MIT database, while TUSZ returned the lowest score ( 0.70 , 50% overlap). The scores for the CHB-MIT and Siena datasets did not display an ascending trend.
To depict the individual prediction behavior of the models, Figure 5 displays the model performance for each EEG dataset. It includes the metrics for all patients (the number of subjects can be observed in Table 1). A 70 % overlap was used to generate this chart. The median of the CHB-MIT results was higher than that of the other two datasets; this situation applied to four metrics, specificity, accuracy, F1-score and AUC-ROC. Therefore, it can be intuited that the Wang_1d model performed best when using the CHB-MIT dataset.
Moreover, the TUSZ-related values spanned an extensive range, meaning that the model was able to achieve good performance for some patients while showing poor results for others. As displayed in Table 4, the models tended to be more sensitive than specific. As mentioned before, it is important to consider that the TUSZ dataset contains both pediatric and adult patients; thus, such heterogeneity may be reflected in the dispersion of the prediction performances.
Figure 6 shows the model accuracy when varying the overlap rate. As stated, the overlap was only applied to generate the training dataset, not the evaluation one. When using the CHB-MIT and Siena datasets, it was noted that the overall accuracy increased when the overlap increased. On the other hand, the TUSZ-related models did not show this behavior. Again, the TUSZ-related ranges were the largest among the three datasets. It is noticeable that, visually speaking, there is no effect when using different window overlaps and the accuracy is lower among all datasets.
The Friedman test [41] returned a p-value = 7.48 × 10 5 for the CHB-MIT dataset and 0.02 for the Siena dataset, indicating that increasing the overlap could increase the model’s performance. The previous statement is supported by Figure 6 and Table 4. The Friedman test applied to the TUSZ accuracy values returned a p-value = 0.54 .
Figure 7 displays the model accuracy for each patient. A 70 % overlap was used to estimate the performance. Each point denotes a patient. The variable used to perform the comparison was the seizure type. When a patient suffered from several seizure types, the most common type was considered.
Figure 7a,b displays the performance for subjects belonging to the Siena and TUSZ datasets, respectively. As per the charts, there was no clear relation between seizure type and model performance. The CHB-MIT values were not included, as the type of seizure was not explicitly mentioned for each patient.

3.2. Correlation between Importance Values and EEG Features

Considering that the greatest accuracy was achieved when using an 80 % overlap, these models were used to estimate the importance values and correlation coefficients.
Figure 8, Figure 9 and Figure 10 show the distribution of the Spearman’s rank correlation coefficients. The histograms are individually displayed for the SHAP and LIME experiments. The coefficients within the blue dashed lines can be interpreted as negligible correlations. The coefficients between the red and blue dashed lines denote a low correlation (see Table 5). It is essential to consider that the Spearman’s rank correlation coefficient measures the strength of a monotonic relationship between two variables.
The Spearman’s rank correlation coefficients estimated using the CHB-MIT dataset are displayed in Figure 8. The chart shows that there was no moderate or high correlation between the SHAP importance values and EEG features. The same was true for the LIME values. The vast majority of coefficients were near zero. A number of the experiments displayed a low positive/negative monotonic relation.
The results in Figure 9 are similar to those in Figure 8. Most coefficients fall within the blue dashed lines; these correlations are negligible. This analysis applies to both the SHAP and LIME experiments. In Figure 9a, a number of of the coefficients indicate a moderate positive correlation; these are discussed later.
Notwithstanding this, Figure 10 displays similar behavior to that in Figure 8 and Figure 9, as a more considerable number of the Spearman’s rank correlation coefficients surpassed the 0.50 and 0.70 thresholds (moderate and high correlation, respectively).
Five of the six patients were part of the TUSZ dataset. The features that had a moderate/strong monotonic relation with the XAI explanations were diverse, including STD, MAD IR, RMS, and Range. Most features were estimated for the beta band (12–25 Hz).
Although the experiments returned a moderate correlation, the model accuracy was unsatisfactory for several patients (PN05, 1027 and 6904). Therefore, a moderately strong correlation does not imply a proficient model. The only experiment returning a high correlation corresponded to patient 4456, where, remarkably, the model accuracy was 0.96 . The most common seizure type for the previous patient was gnsz.
Table 6 details the patients for whom the experiments displayed a moderate/high correlation between the importance values and EEG features. Regarding the p-value, the null hypothesis is that the samples have no ordinal correlation; the alternative hypothesis is that the correlation is nonzero.

4. Discussion

This section presents a discussion of the results and their interpretation. In addition, the limitations of the present work and future research opportunities are addressed.
A state-of-the-art DL model was evaluated in this study, specifically, the one-dimensional convolutional neural network for seizure onset detection presented in [30]. Despite aiming to select models that demonstrated high performance and rigor in their evaluation, it should be acknowledged that the search for models could have been more comprehensive.
In [30], two databases were employed, namely, CHB-MIT [23] and SWEC-ETHZ iEEG [44], with the former being surface and the latter intracranial. Furthermore, the models were trained using an intra-patient approach. The results reported in [30] using the CHB-MIT data showed mean sensitivity, specificity, and accuracy values of 0.88 , 0.99 , and 0.99 , respectively. In contrast, our study utilizing the same EEG dataset (see Table 4) reported maximum sensitivity, specificity, and accuracy of 0.84 , 0.84 , and 0.84 , respectively, with significantly lower specificity. It is essential to consider the differences in data formation between [30] and our work, such as the use of short-duration seizures, avoidance of seizure concatenation, and overlapping rates.
A brief overview of similar studies conducted using comparable methodologies is presented next. In [45], a sensitivity of 0.976 was reported for intra-patient models; this was achieved by applying a convolutional neural network in conjunction with a long short-term memory network. In [46], the authors employed inter-patient models using only the CHB-MIT database, and reported mean values of 0.90 , 0.91 , and 0.98 for sensitivity, specificity, and accuracy, respectively. In [47], the authors trained patient-specific convolutional neural networks and reported values of 0.90 and 0.98 for sensitivity and the area under the curve, respectively. In [48], the authors utilized a neural network called ScoreNet, achieving a sensitivity of 0.765 and specificity of 0.999 . Finally, an accuracy of 0.805 was obtained by [14], who trained inter-patient models using an adversarial neural network. Among the mentioned works, ref. [45,46,47,48] used the CHB-MIT database, while [14] used the TUH EEG Seizure Corpus. However, there are other works that have reported better performance in terms of accuracy, such as [11] for advanced epilepsy detection by applying VFCDM and CNN, who achieved consistently high accuracy rates between healthy and epileptic states. However, their results were obtained for the Bonn database, which contains only five patients with epilepsy. In our work exploring the importance of EEG channels in detecting epileptic seizures, we have identified that database and training conditions can affect classifier performance.
Regarding the work presented in [10], the authors selected three EEG datasets, one of which was the publicly available Bonn dataset, to demonstrate that their method outperformed several state-of-the-art methods. A second dataset was used to show the good generalization ability of their proposed method, and a third dataset to demonstrate that their method is suitable for large-scale datasets. In our work, we selected three EEG datasets to train and evaluate the models under different conditions of overlap between EEG data windows.
One of the goals of the present research was to understand how the training conditions can impact a model’s performance. First, the results showed that a model’s sensitivity and specificity can vary noticeably based on the EEG dataset. For example, the overall specificity for CHB-MIT models was more significant than for the rest of the models (see Figure 5). Accordingly, any ML model used to detect seizures should be evaluated across diverse EEG datasets.
The differences in performance can be explained by the differences between the EEG datasets, such as the EEG channels, sampling frequency, patient demographics, and epilepsy characteristics. Second, it was observed that an overlap during training impacts the model’s performance. During this research, the overlap was applied to the ictal instances to address class imbalances. In addition, the largest class (non-ictal) was subsampled to generate a balanced dataset.
Our results showed that the more significant the overlap applied to the ictal class, the higher the accuracy. It should be noted that a significant overlap implies that the training dataset size has increased. An exception is the TUSZ dataset; Table 4 shows an increase in overall accuracy for the TUSZ models, although the increase is not significant.
Our work can be compared to similar state-of-the-art works with the primary objective of automatically detecting seizures by applying machine learning methods to EEG signals and incorporating explainable artificial intelligence techniques to enhance the interpretability of the models used in seizure detection. First, we have the work presented in [20]; the authors proposed a system which uses a Bi-LSTM network to classify normal and abnormal signals caused by epilepsy, achieving an accuracy of 87.25% on the Bonn dataset. They used the Layerwise Relevance Propagation (LRP) XAI method to explain the predictions of their network. LRP generates a relevance vector containing relevance values to indicate the contribution of a signal in particular class. However, as stated by the authors, many points in the relevance vector were missed; thus, this implementation of the LRP method requires further improvement to generate more accurate results. In [22], the main difference with our work is in the feature extraction step. They reported an accuracy of 99.6% on the Bonn dataset, and their explanations were over the features rather than the signal morphology (time series). In our work, we chose a state-of-the-art deep learning model for seizure detection and three EEG databases. The developed models were trained and evaluated under different conditions and the classifiers with the best performance were selected. SHAP and LIME were then employed to estimate the importance value of each EEG channel. To measure the similarity between the explanations and the epileptic signals, we computed the Spearman’s rank correlation coefficient between the EEG features of epileptic signals (time series) and the importance values. Another important difference between these works are the data sources. In [20,22], the authors used an open access dataset published by researchers at Bonn University containing intracranial EEG recordings from a total number of ten subjects, of whom five were healthy volunteers and the other five were epilepsy patients. Compared to our work, in which three databases were tested, previous works have used a limited dataset in order to evaluate their algorithms. Furthermore, studies employing the Bonn dataset used intracranial instead of scalp EEG recordings.
The interpretability of classification models used in the medical field is crucial, as mentioned by [12,49]; although the number of XAI algorithms is significant, not all algorithms can be applied to time series. Compared to other fields, the interpretation of time series in this field is usually not intuitive, and requires domain knowledge [13,50]. In the present work, SHAP and LIME, which are both agnostic and local methods, were applied. SHAP has previously been used for EEG signals; selected use cases are displayed in Table 7.
In light of the capacity of XAI methods to identify information relevant to the classifier, it may be possible to identify discriminating ictal patterns in high-importance regions (e.g., large SHAP values would match large STD values). As XAI aims to provide transparency, understandable ictal patterns would be preferable.
In line with the above, the Spearman’s rank correlation coefficient between the XAI-generated importance values and the EEG features was computed. As previously stated, this coefficient measures the strength of a monotonic relationship.
As deduced from the results, the strength of the relationship was low and negligible even for high-performing models. The most remarkable cases are described in Table 6. It was observed that experiments resulting in a moderate correlation did not present high accuracy; therefore, a larger correlation coefficient does not imply better performance. The only exception was patient 4456 (TUSZ dataset).
To the best of the authors’ knowledge, there is no similar study in the state of the art; hence, the comparison was complex. However, the difficulties in obtaining epileptic EEG patterns have been described in previous works, e.g., [15,18]. It is relevant to mention that while several previously published works have included XAI to increase the transparency, they did not apply a pattern/explanation evaluation stage.
In addition to the above, the following limitations must be considered: the applied XAI methods, the small set of EEG features, and the chosen classification model. Although more extensive experiments could be performed to search for correlation patterns, alternative approaches must be considered. These are outlined below.
In [55], the classification of seizure and non-seizure states was performed. Random forest showed the best accuracy among the evaluated classification models. Bidirectional network graphs and the lifespan of homology classes were computed to characterize the EEG windows. SHAP was implemented to provide model transparency. The use of EEG features as model inputs provided an initial foundation for transparency, contrary to the raw EEG signals.
In [56], the classification of eight seizure types was addressed. A deep neural network was used for classification and raw EEG windows were used as model inputs. SHAP and topographic maps were applied for transparency. Notably, network activation was used to plot the topographic maps. This technique adds a spatial feature to the XAI explanations, which may indeed be helpful for medical staff.
Additionally, seizure prediction was addressed in [57]. A list of univariate linear features was estimated from EEG recordings. Support vector machine, logistic regression, and CNN were applied during classification. A diverse set of techniques was used to provide explainability, e.g., SHAP, LIME and partial dependency plots. Importantly, the explanations were evaluated by humans (data scientists and clinicians).
Finally, there are a number of limitations in the present work. First, even though a deeper literature revision was performed, modeling the three datasets using several architectures is required to measure the effect of architecture variations. Second, it would be interesting to group the information by several different conditions (for example, on the basis of seizure type, sex, and age, among others) in order to evaluate their effects on detection of ictal patterns. Third, as the use of the Spearman correlation constrains the search to monotonic relations, different nonlinear metrics should be used.

5. Conclusions

The main contribution of this work is the evaluation of the Spearman’s rank correlation coefficient between the features of EEG signals and XAI explanations to identify ictal patterns. It was observed that the EEG dataset impacted the performance of the classification models. Additionally, significant overlaps during training may increase model performance. Our results indicate a negligible and low correlation coefficient between the evaluated features and the LIME/SHAP values, although a few exceptions were observed.
For future work, it is recommended to perform a tradeoff analysis considering the model’s performance and explainability as variables. Clinicians should participate in the explainability evaluation.

Author Contributions

Conceptualization, S.E.S.-H.; methodology, S.E.S.-H., R.A.S.-R., S.T.-R. and I.R.-G.; software, S.E.S.-H.; validation, S.E.S.-H., R.A.S.-R., S.T.-R. and I.R.-G.; formal analysis, S.E.S.-H., R.A.S.-R., S.T.-R. and I.R.-G.; investigation, S.E.S.-H., R.A.S.-R., S.T.-R. and I.R.-G.; data curation, S.E.S.-H.; writing—original draft preparation, S.E.S.-H.; writing—review and editing, R.A.S.-R., S.T.-R. and I.R.-G.; visualization, S.E.S.-H., R.A.S.-R., S.T.-R. and I.R.-G.; supervision, R.A.S.-R., S.T.-R. and I.R.-G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The CHB-MIT Scalp EEG Database is available at https://physionet.org/content/chbmit/1.0.0/ (accessed on 26 January 2023). The Siena Scalp EEG Database is available at https://physionet.org/content/siena-scalp-eeg/1.0.0/ (accessed on 26 January 2023). For access to the TUH EEG Seizure Corpus, refer to [27].

Acknowledgments

We appreciate the facilities provided by the project “Identificación de patrones en registros EEG de crisis epilépticas utilizando métodos de inteligencia artificial explicable” by the Data Analysis and Supercomputing Center (CADS, for its acronym in Spanish) of the University of Guadalajara through the use of the Leo Átrox Supercomputer.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Stafstrom, C.E.; Carmant, L. Seizures and epilepsy: An overview for neuroscientists. Cold Spring Harb. Perspect. Med. 2015, 5, a022426. [Google Scholar] [CrossRef]
  2. Kammerman, S.; Wasserman, L. Seizure disorders: Part 1. Classification and diagnosis. West. J. Med. 2001, 175, 99–103. [Google Scholar] [CrossRef]
  3. WHO. Epilepsy. 2022. Available online: https://www.who.int/en/news-room/fact-sheets/detail/epilepsy (accessed on 21 November 2022).
  4. Zack, M.M.; Kobau, R. National and State Estimates of the Numbers of Adults and Children with Active Epilepsy—United States, 2015. MMWR. Morb. Mortal. Wkly. Rep. 2017, 66, 821–825. [Google Scholar] [CrossRef]
  5. IMSS. Hasta Siete de Cada 10 Derechohabientes con Epilepsia Logran el Control de su Enfermedad: IMSS. 2020. Available online: https://www.imss.gob.mx/prensa/archivo/202002/072#:~:text=Entre%20seis%20y%20siete%20de,convulsivas%20que%20caracterizan%20este%20padecimiento (accessed on 31 October 2021).
  6. Noriega-Morales, G.; Shkurovich-Bialik, P. Situación de la epilepsia en México y América Latina. An. Médi. Asoc. Médica Cent. Méd. ABC 2020, 65, 224–232. [Google Scholar] [CrossRef]
  7. Rasheed, K.; Qayyum, A.; Qadir, J.; Sivathamboo, S.; Kwan, P.; Kuhlmann, L.; O’Brien, T.; Razi, A. Machine learning for predicting epileptic seizures using EEG signals: A review. IEEE Rev. Biomed. Eng. 2020, 14, 139–155. [Google Scholar] [CrossRef] [PubMed]
  8. Siddiqui, M.K.; Morales-Menendez, R.; Huang, X.; Hussain, N. A review of epileptic seizure detection using machine learning classifiers. Brain Inform. 2020, 7, 5. [Google Scholar] [CrossRef] [PubMed]
  9. Khan, P.; Kader, M.F.; Islam, S.R.; Rahman, A.B.; Kamal, M.S.; Toha, M.U.; Kwak, K.S. Machine learning and deep learning approaches for brain disease diagnosis: Principles and recent advances. IEEE Access 2021, 9, 37622–37655. [Google Scholar] [CrossRef]
  10. Liu, Y.; Lin, Y.; Jia, Z.; Ma, Y.; Wang, J. Representation based on ordinal patterns for seizure detection in EEG signals. Comput. Biol. Med. 2020, 126, 104033. Available online: https://www.sciencedirect.com/science/article/pii/S0010482520303644 (accessed on 21 November 2023). [CrossRef]
  11. Veeranki, Y.R.; McNaboe, R.; Posada-Quintero, H.F. EEG-Based Seizure Detection Using Variable-Frequency Complex Demodulation and Convolutional Neural Networks. Signals 2023, 4, 816–835. [Google Scholar] [CrossRef]
  12. Tjoa, E.; Guan, C. A survey on explainable artificial intelligence (xai): Toward medical xai. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4793–4813. [Google Scholar] [CrossRef] [PubMed]
  13. Rojat, T.; Puget, R.; Filliat, D.; Del Ser, J.; Gelin, R.; Díaz-Rodríguez, N. Explainable artificial intelligence (xai) on timeseries data: A survey. arXiv 2021, arXiv:2104.00950. [Google Scholar]
  14. Zhang, X.; Yao, L.; Dong, M.; Liu, Z.; Zhang, Y.; Li, Y. Adversarial representation learning for robust patient-independent epileptic seizure detection. IEEE J. Biomed. Health Inform. 2020, 24, 2852–2859. [Google Scholar] [CrossRef]
  15. Mansour, M.; Khnaisser, F.; Partamian, H. An explainable model for EEG seizure detection based on connectivity features. arXiv 2020, arXiv:2009.12566. [Google Scholar]
  16. Gschwandtner, L. Deep Learning for Detecting Interictal EEG Biomarkers to Assist Differential Epilepsy Diagnosis. Master’s Thesis, Comenius University Bratislavia, Bratislava, Slovakia, 2020. [Google Scholar]
  17. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. Adv. Neural Inf. Process. Syst. 2017, arXiv:1705.0787430. [Google Scholar] [CrossRef]
  18. Dissanayake, T.; Fernando, T.; Denman, S.; Sridharan, S.; Fookes, C. Deep learning for patient-independent epileptic seizure prediction using scalp EEG signals. IEEE Sens. J. 2021, 21, 9377–9388. [Google Scholar] [CrossRef]
  19. Rashed-Al-Mahfuz, M.; Moni, M.A.; Uddin, S.; Alyami, S.A.; Summers, M.A.; Eapen, V. A deep convolutional neural network method to detect seizures and characteristic frequencies using epileptic electroencephalogram (EEG) data. IEEE J. Transl. Eng. Health Med. 2021, 9, 2000112. [Google Scholar] [CrossRef] [PubMed]
  20. Rathod, P.; Bhalodiya, J.; Naik, S. Epilepsy Detection using Bi-LSTM with Explainable Artificial Intelligence. In Proceedings of the 2022 IEEE 19th India Council International Conference (INDICON), Kochi, India, 24–26 November 2022; pp. 1–6. [Google Scholar] [CrossRef]
  21. Raab, D.; Theissler, A.; Spiliopoulou, M. XAI4EEG: Spectral and spatio-temporal explanation of deep learning-based seizure detection in EEG time series. Neural Comput. Appl. 2023, 35, 10051–10068. [Google Scholar] [CrossRef]
  22. Ahmad, I.; Yao, C.; Li, L.; Chen, Y.; Liu, Z.; Ullah, I.; Shabaz, M.; Wang, X.; Huang, K.; Li, G.; et al. An efficient feature selection and explainable classification method for EEG-based epileptic seizure detection. J. Inf. Secur. Appl. 2024, 80, 103654. [Google Scholar] [CrossRef]
  23. Shoeb, A.H. Application of Machine Learning to Epileptic Seizure Onset Detection and Treatment. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2009. [Google Scholar]
  24. Shoeb, A. CHB-MIT Scalp EEG Database. 2010. Available online: https://physionet.org/content/chbmit/1.0.0/ (accessed on 26 January 2023).
  25. Detti, P.; Vatti, G.; Zabalo Manrique de Lara, G. EEG synchronization analysis for seizure prediction: A study on data of noninvasive recordings. Processes 2020, 8, 846. [Google Scholar] [CrossRef]
  26. Detti, P. Siena Scalp EEG Database. 2020. Available online: https://physionet.org/content/siena-scalp-eeg/1.0.0/ (accessed on 26 January 2023).
  27. Shah, V.; Von Weltin, E.; Lopez, S.; McHugh, J.R.; Veloso, L.; Golmohammadi, M.; Obeid, I.; Picone, J. The temple university hospital seizure detection corpus. Front. Neuroinform. 2018, 12, 83. [Google Scholar] [CrossRef]
  28. Obeid, I.; Picone, J. The temple university hospital EEG data corpus. Front. Neurosci. 2016, 10, 196. [Google Scholar] [CrossRef]
  29. Shoeb, A.; Guttag, J. Application of Machine Learning to Epileptic Seizure Detection. In Proceedings of the 27th International Conference on International Conference on Machine Learning, Omnipress, Haifa, Israel, 21–24 June 2010; pp. 975–982. [Google Scholar]
  30. Wang, X.; Wang, X.; Liu, W.; Chang, Z.; Kärkkäinen, T.; Cong, F. One dimensional convolutional neural networks for seizure onset detection using long-term scalp and intracranial EEG. Neurocomputing 2021, 459, 212–222. [Google Scholar] [CrossRef]
  31. Sánchez-Hernández, S.E.; Salido-Ruiz, R.A.; Torres-Ramos, S.; Román-Godínez, I. Evaluation of Feature Selection Methods for Classification of Epileptic Seizure EEG Signals. Sensors 2022, 22, 3066. [Google Scholar] [CrossRef]
  32. Phinyomark, A.; Thongpanja, S.; Hu, H.; Phukpattaranont, P.; Limsakul, C. The Usefulness of Mean and Median Frequencies in Electromyography Analysis. In Computational Intelligence in Electromyography Analysis—A Perspective on Current Applications and Future Challenges; IntechOpen: London, UK, 2012; pp. 195–220. [Google Scholar] [CrossRef]
  33. Bergil, E.; Bozkurt, M.R.; Oral, C. An Evaluation of the Channel Effect on Detecting the Preictal Stage in Patients With Epilepsy. Clin. EEG Neurosci. 2020, 52, 376–385. [Google Scholar] [CrossRef]
  34. Kokoska, S.; Zwillinger, D. CRC Standard Probability and Statistics Tables and Formulae; CRC Press: Boca Raton, FL, USA, 2000. [Google Scholar]
  35. Aarabi, A.; Fazel-Rezai, R.; Aghakhani, Y. A fuzzy rule-based system for epileptic seizure detection in intracranial EEG. Clin. Neurophysiol. 2009, 120, 1648–1657. [Google Scholar] [CrossRef] [PubMed]
  36. Bruce, P.; Bruce, A. Practical Statistics for Data Scientists: 50+ Essential Concepts Using R and Python, 1st ed.; O’Reilly Media: Sebastopol, CA, USA, 2017. [Google Scholar]
  37. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed]
  38. Lundberg, S. shap.explainers.Partition. 2018. Available online: https://shap.readthedocs.io/en/latest/generated/shap.explainers.Partition.html (accessed on 21 November 2022).
  39. Owen, G. Values of games with a priori unions. In Mathematical Economics and Game Theory; Springer: Berlin/Heidelberg, Germany, 1977; pp. 76–88. [Google Scholar] [CrossRef]
  40. Ribeiro, M.; Singh, S.; Guestrin, C. “Why Should I Trust You?”: Explaining the Predictions of Any Classifier. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Demonstrations, San Diego, CA, USA, 12–17 June 2016; Association for Computational Linguistics: Stroudsburg, PA, USA, 2016; pp. 97–101. [Google Scholar] [CrossRef]
  41. Sheskin, D.J. Parametric and Nonparametric Statistical Procedures; CRC: Boca Raton, FL, USA, 2000. [Google Scholar]
  42. Python Software Foundation. Python. 2018. Available online: https://www.python.org/ (accessed on 10 June 2023).
  43. Mukaka, M.M. A guide to appropriate use of correlation coefficient in medical research. Malawi Med. J. 2012, 24, 69–71. [Google Scholar] [PubMed]
  44. Burrello, A.; Cavigelli, L.; Schindler, K.; Benini, L.; Rahimi, A. Laelaps: An energy-efficient seizure detection algorithm from long-term human iEEG recordings without false alarms. In Proceedings of the 2019 Design, Automation & Test in Europe Conference & Exhibition (DATE), Florence, Italy, 25–29 March 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 752–757. [Google Scholar] [CrossRef]
  45. Liu, X.; Richardson, A.G. Edge deep learning for neural implants: A case study of seizure detection and prediction. J. Neural Eng. 2021, 18, 046034. [Google Scholar] [CrossRef] [PubMed]
  46. Hossain, M.S.; Amin, S.U.; Alsulaiman, M.; Muhammad, G. Applying deep learning for epilepsy seizure detection and brain mapping visualization. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2019, 15, 1–17. [Google Scholar] [CrossRef]
  47. Bahr, A.; Schneider, M.; Francis, M.A.; Lehmann, H.M.; Barg, I.; Buschhoff, A.S.; Wulff, P.; Strunskus, T.; Faupel, F. Epileptic seizure detection on an ultra-low-power embedded RISC-V processor using a convolutional neural network. Biosensors 2021, 11, 203. [Google Scholar] [CrossRef] [PubMed]
  48. Boonyakitanont, P.; Lek-Uthai, A.; Songsiri, J. ScoreNet: A Neural network-based post-processing model for identifying epileptic seizure onset and offset in EEGs. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 2474–2483. [Google Scholar] [CrossRef]
  49. Qayyum, A.; Qadir, J.; Bilal, M.; Al-Fuqaha, A. Secure and robust machine learning for healthcare: A survey. IEEE Rev. Biomed. Eng. 2020, 14, 156–180. [Google Scholar] [CrossRef] [PubMed]
  50. Theissler, A.; Spinnato, F.; Schlegel, U.; Guidotti, R. Explainable AI for Time Series Classification: A review, taxonomy and research directions. IEEE Access 2022, 10, 100700–100724. [Google Scholar] [CrossRef]
  51. Cui, S.; Duan, L.; Qiao, Y.; Xiao, Y. Learning EEG synchronization patterns for epileptic seizure prediction using bag-of-wave features. J. Ambient. Intell. Humaniz. Comput. 2018, 14, 15557–15572. [Google Scholar] [CrossRef]
  52. Cho, S.; Lee, G.; Chang, W.; Choi, J. Interpretation of Deep Temporal Representations by Selective Visualization of Internally Activated Nodes. arXiv 2020, arXiv:2004.12538. [Google Scholar]
  53. Bennett, J.D.; John, S.E.; Grayden, D.B.; Burkitt, A.N. Universal neurophysiological interpretation of EEG brain-computer interfaces. In Proceedings of the 2021 9th International Winter Conference on Brain-Computer Interface (BCI), Gangwon, Republic of Korea, 22–24 February 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar] [CrossRef]
  54. Gabeff, V.; Teijeiro, T.; Zapater, M.; Cammoun, L.; Rheims, S.; Ryvlin, P.; Atienza, D. Interpreting deep learning models for epileptic seizure detection on EEG signals. Artif. Intell. Med. 2021, 117, 102084. [Google Scholar] [CrossRef]
  55. Upadhyaya, D.P.; Prantzalos, K.; Thyagaraj, S.; Shafiabadi, N.; Fernandez-BacaVaca, G.; Sivagnanam, S.; Majumdar, A.; Sahoo, S.S. Machine Learning Interpretability Methods to Characterize Brain Network Dynamics in Epilepsy. medRxiv 2023. [Google Scholar] [CrossRef]
  56. Alshaya, H.; Hussain, M. EEG-Based Classification of Epileptic Seizure Types Using Deep Network Model. Mathematics 2023, 11, 2286. [Google Scholar] [CrossRef]
  57. Pinto, M.F.; Batista, J.; Leal, A.; Lopes, F.; Oliveira, A.; Dourado, A.; Abuhaiba, S.I.; Sales, F.; Martins, P.; Teixeira, C.A. The goal of explaining black boxes in EEG seizure prediction is not to explain models’ decisions. Epilepsia Open 2023, 8, 285–297. [Google Scholar] [CrossRef]
Figure 1. Diagram of the methodology.
Figure 1. Diagram of the methodology.
Brainsci 14 00306 g001
Figure 2. Wang_1d neural network. Values in parentheses are as follows: (number of filters, kernel size/number of neurons, stride).
Figure 2. Wang_1d neural network. Values in parentheses are as follows: (number of filters, kernel size/number of neurons, stride).
Brainsci 14 00306 g002
Figure 3. The application of SHAP to an EEG epoch: (a) the ictal EEG window of patient chb01 and (b) the matrix of SHAP values.
Figure 3. The application of SHAP to an EEG epoch: (a) the ictal EEG window of patient chb01 and (b) the matrix of SHAP values.
Brainsci 14 00306 g003
Figure 4. The application of LIME to an EEG epoch: (a) the ictal EEG window of patient chb01 and (b) the matrix of LIME values.
Figure 4. The application of LIME to an EEG epoch: (a) the ictal EEG window of patient chb01 and (b) the matrix of LIME values.
Brainsci 14 00306 g004
Figure 5. Performance comparison for each EEG database. A 70 % overlap was used to estimate the metrics.
Figure 5. Performance comparison for each EEG database. A 70 % overlap was used to estimate the metrics.
Brainsci 14 00306 g005
Figure 6. Performance comparison for each overlap rate.
Figure 6. Performance comparison for each overlap rate.
Brainsci 14 00306 g006
Figure 7. Performance comparison for each patient based on the most common seizure type. Data are displayed for the (a) Siena and (b) TUSZ datasets. A 70 % overlap was used to estimate the metrics. Each circle denotes a patient.
Figure 7. Performance comparison for each patient based on the most common seizure type. Data are displayed for the (a) Siena and (b) TUSZ datasets. A 70 % overlap was used to estimate the metrics. Each circle denotes a patient.
Brainsci 14 00306 g007
Figure 8. The distribution of the Spearman´s rank correlation coefficients estimated for the CHB-MIT dataset. The range between the blue dashed lines corresponds to a negligible correlation, while the range between the blue and red dashed lines corresponds to a low correlation. (a) SHAP values; (b) LIME values.
Figure 8. The distribution of the Spearman´s rank correlation coefficients estimated for the CHB-MIT dataset. The range between the blue dashed lines corresponds to a negligible correlation, while the range between the blue and red dashed lines corresponds to a low correlation. (a) SHAP values; (b) LIME values.
Brainsci 14 00306 g008
Figure 9. The distribution of the Spearman’s rank correlation coefficients estimated for the Siena dataset. The range between the blue dashed lines corresponds to a negligible correlation, while the range between the blue and red dashed lines corresponds to a low correlation. (a) SHAP values; (b) LIME values.
Figure 9. The distribution of the Spearman’s rank correlation coefficients estimated for the Siena dataset. The range between the blue dashed lines corresponds to a negligible correlation, while the range between the blue and red dashed lines corresponds to a low correlation. (a) SHAP values; (b) LIME values.
Brainsci 14 00306 g009
Figure 10. The distribution of the Spearman’s rank correlation coefficients estimated for the TUSZ dataset. The range between the blue dashed lines corresponds to a negligible correlation, while the range between the blue and red dashed lines corresponds to a low correlation. (a) SHAP values; (b) LIME values.
Figure 10. The distribution of the Spearman’s rank correlation coefficients estimated for the TUSZ dataset. The range between the blue dashed lines corresponds to a negligible correlation, while the range between the blue and red dashed lines corresponds to a low correlation. (a) SHAP values; (b) LIME values.
Brainsci 14 00306 g010
Table 1. Details of the datasets after preprocessing.
Table 1. Details of the datasets after preprocessing.
DatabasePatientsSeizure TypesSource Montage CategoryChannels after Pre-Processing
CHB-MIT Scalp EEG Database24Not specifiedBipolarFp1-F7, F7-T7, T7-P7, P7-O1, Fp1-F3, F3-C3, C3-P3, P3-O1, FP2-F4, F4-C4, C4-P4, P4-O2, FP2-F8, F8-T8, T8-P8, P8-O2, FZ-CZ, CZ-PZ, T7-FT9, FT9-FT10, FT10-T8
Siena Scalp EEG Database10Focal onset impaired awareness (IAS), focal onset without impaired awareness (WIAS), focal to bilateral tonic–clonic (FBTC)ReferentialFp1-F7, F7-T3, T3-T5, T5-O1, Fp1-F3, F3-C3, C3-P3, P3-O1, Fp2-F4, F4-C4, C4-P4, P4-O2, Fp2-F8, F8-T4, T4-T6, T6-O2, Fz-Cz, Cz-Pz
TUH EEG Seizure Corpus21Tonic–clonic (tcsz), focal non-specific (fnsz), generalized non-specific (gnsz), absence seizure (absz), complex-partial seizure (cpsz)ReferentialFp1-F7, F7-T3, T3-T5, T5-O1, Fp1-F3, F3-C3, C3-P3, P3-O1, Fp2-F4, F4-C4, C4-P4, P4-O2, Fp2-F8, F8-T4, T4-T6, T6-O2, Fz-Cz, Cz-Pz
Table 2. Performance and characteristics of the model presented by [30].
Table 2. Performance and characteristics of the model presented by [30].
CharacteristicValue
ModelCNN
Training typePatient-specific
Accuracy (CHB-MIT) 0.99
Sensitivity (CHB-MIT) 0.88
Specificity (CHB-MIT) 0.99
Table 3. Estimated EEG features.
Table 3. Estimated EEG features.
Feature Name
Median Frequency (MedFreq) [32]
Complexity [33]
Skewness [34]
Mobility [33]
Kurtosis [34]
Interquartile Range (IR) [34]
Peak Frequency (PkFreq) [35]
Median Absolute Deviation (MAD) [36]
Root Mean Square (RMS) [34]
Sample Entropy (SampEn) [37]
Range [34]
Mean [34]
Number of Zero Crossings (ZC) [33]
Standard Deviation (STD) [34]
Table 4. Mean performance of the classification models.
Table 4. Mean performance of the classification models.
ModelOverlap (%)SensitivitySpecificityAccuracyF1-ScoreAUC–ROC
CHB-MIT
Wang_1d50 0.83 0.61 0.72 0.73 0.86
70 0.83 0.77 0.79 0.60 0.88
80 0.84 0.84 0.84 0.83 0.88
Siena
Wang_1d50 0.84 0.39 0.62 0.68 0.78
70 0.80 0.53 0.63 0.56 0.79
80 0.74 0.71 0.73 0.71 0.78
TUSZ
Wang_1d50 0.73 0.49 0.61 0.61 0.70
70 0.69 0.58 0.63 0.54 0.73
80 0.69 0.58 0.64 0.62 0.74
Table 5. Interpretation of the size of a correlation coefficient. Taken from [43].
Table 5. Interpretation of the size of a correlation coefficient. Taken from [43].
Size of CorrelationInterpretation
0.90 to 1.00 ( 0.90 to 1.00 )Very high positive (negative)
0.70 to 0.90 ( 0.70 to 0.90 )High positive (negative)
0.50 to 0.70 ( 0.50 to 0.70 )Moderate positive (negative)
0.30 to 0.50 ( 0.30 to 0.50 )Low positive (negative)
0.00 to 0.30 ( 0.00 to 0.30 )Negligible correlation
Table 6. List of patients for whom the experiments displayed a moderate/high correlation between the importance values and EEG features. The ’Top Correlation’ column indicates the three features with the largest correlation coefficients.
Table 6. List of patients for whom the experiments displayed a moderate/high correlation between the importance values and EEG features. The ’Top Correlation’ column indicates the three features with the largest correlation coefficients.
PatientDatasetAccuracyXAI MethodTop Correlation
PN05Siena 0.54 SHAPLow/STD ( 0.55 , p-value = 2.42 × 10 48 )
Low/MAD ( 0.55 , p-value = 2.05 × 10 49 )
Low/IR ( 0.55 , p-value = 2.30 × 10 49 )
1027TUSZ 0.70 SHAPFull/IR ( 0.53 , p-value = 0 )
Full/RMS ( 0.53 , p-value = 0 )
Low/Range ( 0.54 , p-value = 0 )
4456TUSZ 0.96 SHAPBeta/IR ( 0.70 , p-value = 9.31 × 10 50 )
Beta/RMS ( 0.71 , p-value = 1.58 × 10 50 )
Beta/STD ( 0.71 , p-value = 1.59 × 10 50 )
6904TUSZ 0.63 SHAPBeta/IR ( 0.66 , p-value = 0 )
Beta/MAD ( 0.66 , p-value = 0 )
Beta/RMS ( 0.66 , p-value = 0 )
LIMEBeta/MAD ( 0.53 , p-value = 0 )
Beta/RMS ( 0.53 , p-value = 0 )
Beta/STD ( 0.53 , p-value = 0 )
6563TUSZ 0.79 LIMEBeta/MAD ( 0.52 , p-value = 0 )
Beta/RMS ( 0.53 , p-value = 0 )
Beta/STD ( 0.53 , p-value = 0 )
Table 7. XAI methods applied to EEG signals.
Table 7. XAI methods applied to EEG signals.
ReferenceApplied Methods
[51]Bag of Waves
[52]Clustered Pattern of Highly Activated Period
[18]SHAP
[16]ProtoPNet, Attention Pooling, Layerwise Relevance Propagation, SHAP
[15]Activation values
[14]Attention Pooling
[19]SHAP
[53]EEG-SHAP
[54]SHAP, Maximized inputs
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sánchez-Hernández, S.E.; Torres-Ramos, S.; Román-Godínez, I.; Salido-Ruiz, R.A. Evaluation of the Relation between Ictal EEG Features and XAI Explanations. Brain Sci. 2024, 14, 306. https://doi.org/10.3390/brainsci14040306

AMA Style

Sánchez-Hernández SE, Torres-Ramos S, Román-Godínez I, Salido-Ruiz RA. Evaluation of the Relation between Ictal EEG Features and XAI Explanations. Brain Sciences. 2024; 14(4):306. https://doi.org/10.3390/brainsci14040306

Chicago/Turabian Style

Sánchez-Hernández, Sergio E., Sulema Torres-Ramos, Israel Román-Godínez, and Ricardo A. Salido-Ruiz. 2024. "Evaluation of the Relation between Ictal EEG Features and XAI Explanations" Brain Sciences 14, no. 4: 306. https://doi.org/10.3390/brainsci14040306

APA Style

Sánchez-Hernández, S. E., Torres-Ramos, S., Román-Godínez, I., & Salido-Ruiz, R. A. (2024). Evaluation of the Relation between Ictal EEG Features and XAI Explanations. Brain Sciences, 14(4), 306. https://doi.org/10.3390/brainsci14040306

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop