Next Article in Journal
Development and Earliest Validation of a Portable Device for Quantification of Hallux Extension Strength (QuHalEx)
Previous Article in Journal
In-City Rain Mapping from Commercial Microwave Links—Challenges and Opportunities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multidomain Convolution Neural Network Models for Improved Event-Related Potential Classification

1
School of Electrical, Computer, and Biomedical Engineering, Southern Illinois University, Carbondale, IL 62901, USA
2
Center of Excellence for Stress and Mental Health, VA San Diego Healthcare System, San Diego, CA 92161, USA
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(10), 4656; https://doi.org/10.3390/s23104656
Submission received: 5 April 2023 / Revised: 8 May 2023 / Accepted: 8 May 2023 / Published: 11 May 2023
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Two convolution neural network (CNN) models are introduced to accurately classify event-related potentials (ERPs) by fusing frequency, time, and spatial domain information acquired from the continuous wavelet transform (CWT) of the ERPs recorded from multiple spatially distributed channels. The multidomain models fuse the multichannel Z-scalograms and the V-scalograms, which are generated from the standard CWT scalogram by zeroing-out and by discarding the inaccurate artifact coefficients that are outside the cone of influence (COI), respectively. In the first multidomain model, the input to the CNN is generated by fusing the Z-scalograms of the multichannel ERPs into a frequency-time-spatial cuboid. The input to the CNN in the second multidomain model is formed by fusing the frequency-time vectors of the V-scalograms of the multichannel ERPs into a frequency-time-spatial matrix. Experiments are designed to demonstrate (a) customized classification of ERPs, where the multidomain models are trained and tested with the ERPs of individual subjects for brain-computer interface (BCI)-type applications, and (b) group-based ERP classification, where the models are trained on the ERPs from a group of subjects and tested on single subjects not included in the training set for applications such as brain disorder classification. Results show that both multidomain models yield high classification accuracies for single trials and small-average ERPs with a small subset of top-ranked channels, and the multidomain fusion models consistently outperform the best unichannel classifiers.

1. Introduction

The accurate classification of event-related potentials (ERPs) is crucial in brain activity research and in brain-related clinical studies, evaluations, and diagnostics. Classification of ERPs can be improved by fusing frequency, time, and spatial domain information acquired from the scalograms of the continuous wavelet transforms (CWTs) of multichannel recordings of these signals. The frequency-time information is captured in the wavelet scalograms of each channel’s ERPs, and the spatial information is contributed from the spatially distributed channels across the scalp. Convolution neural network (CNN) classifiers are ideal for inputs presented in matrix and cuboid formats [1,2,3,4,5,6,7,8] and are, therefore, a good choice for processing single-channel matrix scalograms and multichannel scalograms combined into a cuboid. Furthermore, unlike traditional classifiers, which typically require extracting “hand-engineered” features from the three domains and selecting a set of features through trial and error, CNNs can be trained to extract intertwined frequency-time-spatial features directly. The shapes and dimensions of the filters in the convolution layers as well as the number of convolution layers offer flexibility in the type of multidomain intertwining desired. Classification models designed for single subjects, which is the primary focus of this study, can take two forms: customized classifier design for brain-computer interface (BCI)-type applications and group-based design for classifying brain activity and disorders in applications related to clinical studies, evaluations, and diagnostics. Customized classifiers are designed using the individual’s own ERPs, whereas group-based classifiers are typically trained on ERPs from a group of subjects having a similar disorder but tested on the ERPs of an individual not included in the training group.
In a recent study related to the classification of unichannel (single channel) ERPs of single subjects [1], we introduced the Z-scalogram and the V-scalogram derived from the cone of influence (COI) of the standard scalogram of the CWT. The COI is a boundary that is superimposed on the wavelet scalogram to delineate the coefficients that are accurate from those that are inaccurate due to edge effects [1,9,10,11,12].
The standard S-scalogram is obtained through “same” convolutions, the Z-scalogram is obtained from the S-scalogram by zeroing-out the inaccurate coefficients outside the COI, and the V-scalogram is obtained by completely cropping out the inaccurate coefficients outside the COI. The scalogram containing only the accurate coefficients corresponds to the scalogram that would be obtained through “valid” convolutions; therefore, it is referred to as the V-scalogram. Details of the various types of convolutions involved in defining the COI can be found in [1]. Figure 1, extracted from [1], shows the S-, Z-, and V-scalograms of an ERP. The S-scalogram shown in Figure 1b is the standard Morlet wavelet transform scalogram. In the study, feature vectors derived from the three scalograms were used to design support vector machines (SVMs), random forests (RFs), k-nearest neighbor (k-NN), multilayer perceptron neural networks (MLPs), and CNNs to classify synonymous and non-synonymous ERPs. It was shown that the Z-scalogram classifiers outperformed the S-scalogram classifiers, and the V-scalogram classifiers outperformed the S- and Z-scalogram classifiers. It was also shown that the CNN classifiers yielded the best performance. However, the study focused on customized classifier design for single subjects, did not involve group-based classification, and involved only unichannel classification with no attempt to fuse information from multiple channels.
In another recent study related to the classification of multiaxial human movement, various CNN models were developed to fuse “information” from multiple multiaxial sensors at the input-level (early fusion) and the output-level (late fusion) [13]. The information in input-level fusion is the input data or features extracted from the input data, whereas information in output-level fusion is the decisions of multiple classifiers or some measures at the outputs of the multiple classifiers [13,14,15,16,17,18,19]. In the multiaxial human movement classification study, it was shown that the fusion classifiers outperformed the uniaxial classifiers. However, the multiaxial temporal signals were used directly without any form of time-frequency transformations. Furthermore, the movement signals of all individuals were mixed prior to generating the training and test sets. That is, the experiments did not involve designing customized classifiers for each individual, nor did they involve designing group-based classifiers to test the movement signals of individuals not included in the training set.
The present study builds upon these studies and improves the classification of ERPs by using CNNs which fuse frequency, time, and spatial information acquired from the COI-modified scalograms of the multichannel ERPs. Two CNN-based multidomain classification models which fuse the COI-modified scalograms of multiple channels are introduced. In the first model, referred to as the Z-CuboidNet, the input to the CNN is a frequency-time-spatial cuboid formed by fusing the Z-scalograms of the multichannel ERPs, and in the second model, referred to as the V-MatrixNet, the CNN input is a frequency-time-spatial matrix formed by fusing the V-scalogram frequency-time vectors of the multichannel ERPs. CNNs are selected for the classifier component of the multidomain models because they are ideal for inputs presented in matrix and cuboid formats. Additionally, this study focuses on the design (training and testing) of both customized and group-based ERP classifiers. To the best of our knowledge, we are not aware of any other study that has been reported on the development of multidomain classification models similar to the Z-CuboidNet and V-MatrixNet.

2. Materials and Methods

This section describes the following: (a) subsample averaging to facilitate the design of ERP classifiers, (b) formulations of the Z-CuboidNet and V-MatrixNet multidomain classifier models, (c) the formulations of the unichannel Z-MatrixNet and V-VectorNet models which are special cases of the multidomain models, (d) the ERP data set used to demonstrate the application and the performance of the classifier models, (e) cross-validation for group-based classifier design, (f) selection of subsets of channels for single subjects and the subjects combined into a group, (g) the Morlet wavelet transform, which, is the CWT used to generate the scalograms, and (h) the architectures, hyperparameters, and training options selected for the CNN models.

2.1. Subsample Averaging

An important issue related to ERP classification is the poor signal-to-noise ratio (SNR) of single trials; therefore, this issue is discussed before describing the formulations of the multidomain classification models. The SNR is improved for analysis and classification by averaging single trials from several repetitions of stimulus presentation [20,21,22]. The m -Subsample Averaging algorithm [1], which generates small-sample ERPs by repeated averaging of a small number of single trials drawn without replacement, can be employed to improve the performance over single trials. An ERP formed by averaging m single trials is referred to as an m -ERP. Subsample averaging enables the generation of a large ensemble of m -ERPs to facilitate classifier training. For consistency, single trials are referred to as 1-ERPs. The generalized formulations of the multidomain models in the following subsection assume m -ERPs. The experiments in this study are designed to evaluate the performances of the multidomain classifiers for m = 1, 2, and 4.

2.2. Multidomain Classifier Models

The most general formulations of the Z-CuboidNet and V-MatrixNet multidomain models involving multiple channels and polychotomous ERP classes are presented in this section. The channels are represented by d , d = 0,1 , , ( D 1 ) , where, D is the number of selected channels, and the polychotomous classes are represented by ω j , j = 1,2 , , Ω , where Ω is the number of classes. Using array (row-column) representations and zero-based indexing, a scalogram of an m -ERP is denoted by G f , t ;   f = 0,1 , . . . , F 1 , t = 0,1 , , T 1 , where F is the number of frequency bands and T is the duration of the ERP. Furthermore, if an m -ERP in the ensemble of each channel is indexed by q , q = 1,2 , , Q , where Q is the number of m -ERPs, the S-, Z-, and V-scalograms of the q t h   m -ERP of channel d belonging to class ω j are represented by G S , d j ( q ) f , t , G Z , d j ( q ) f , t , and G V , d j ( q ) f , t , respectively. Classification involving unichannel and dichotomous classes are special cases of the general formulations.
Without loss of generality, the CNNs in the multidomain models consist of two convolution layers followed by a max pooling layer and a fully connected network (FCN). The activation functions in the convolution layers are ReLU and are tanh in the intermediate layers of the FCN. The output of the FCN is a softmax layer which has one output for each ERP class, and each output is interpreted as the posterior probability of the input class. The CNNs are trained with the gradient descent backpropagation algorithm using the cross-entropy loss function. The formulations of the 4 models, which follow next, describe the operations in the first convolution layer of each model in detail because the key differences between the models occur in this layer. The remaining operations are described briefly to point out additional differences and for the sake of completeness. For convenience, it will be assumed that all convolutions are “same convolutions” through zero-padding so that the filtered outputs have the same dimensions as the input to the convolution layer. The equations describe the cross-correlations of the input and filters in the convolution layers because the shifting, multiplication, and summing operations performed in CNNs are cross-correlation and not convolution.

2.2.1. Z-CuboidNet

The Z-CuboidNet fuses the Z-scalograms of the D channels into a frequency( f )-time( t )-spatial( d ) cuboid given by
G Z , j q f , t , d = d = 0 ( D 1 ) G Z , d j q f , t , f = 0,1 , , F 1 ;   t = 0,1 , , T 1 ;   d = 0,1 , , ( D 1 ) .
where is the cuboid fusion operation. Note that G Z , j ( q ) f , t , d is a cuboid generated by combining the Z-scalograms of the q t h   m -ERP of the D channels. The subscript j indicates that the Z-scalograms are generated from the class ω j   m -ERPs.
The block diagram of the Z-CuboidNet is shown in Figure 2. In the first convolution layer, the cuboid input G Z , j ( q ) f , t , d is convolved with N 1 cuboid filters h [ 1 , n ] f , t , d , f = α , , 0 , , α ;   t = β , 0 , , β ;   d = 0,1 , , ( D 1 ) ;   n = 1,2 , , N 1 , and the output of the n t h filter in the first convolution layer is given by
Y Z , j q 1 , n f , t = d = 0 ( D 1 ) r = α α u = β β h 1 , n r , u , d G Z , j q f + r , t + u , d , f = 0,1 , , F 1 , t = 0,1 , , T 1 .
It is important to note that cuboid filters are selected to extract intertwined frequency, time, and spatial features. Additionally interesting to note is that the result of convolving two cuboids is a matrix. After a bias is added to the filtered outputs and passed through the nonlinear ReLU activation function, the N 1 filtered outputs are combined into a ( F × T × N 1 ) cuboid feature map. The convolutions in the second layer are also selected to be cuboid convolutions to extract more complex cross-intertwined features across the cuboid feature map. The resulting matrices are combined into a cuboid after adding biases and passing through ReLU activations. The height and width of the cuboid is shrunk by the following max pooling layer. The max-pooled cuboid is flattened and presented as the input to the FCN. The output of the FCN is represented by the posterior class probability vector P = ( P ω 1 , P ω 2 , . . , P ω Ω ) , and the test input is assigned to a class using the maximum response rule. That is, the test input is assigned to class ω i if
P ω i > P ω j   for   all   j i .

2.2.2. Z-MatrixNet

The Z-MatrixNet is a special case of the Z-CuboidNet for unichannel classification. The input cuboid reduces to a frequency-time matrix when D = 1 . A classifier is developed independently for the m -ERPs of each channel. This approach can be used to compare the performances of each channel classifier and to select the best unichannel classifier. Most importantly, the performances of the multidomain Z-CuboidNet classifier can be compared against the best unichannel Z-MatrixNet classifier. The input of channel d to the CNN is simply the Z-scalogram G Z , d j ( q ) f , t . In the first layer, the input matrix is convolved with N 1 matrix filters h [ 1 , n ] f , t , f = α , , 0 , , α ;   t = β , , 0 , , β , n = 1,2 , , N 1 . The output of filter n is given by the matrix convolution,
Y Z , d j q 1 , n f , t = r = α α u = β β h 1 , n r , u G Z , d j q f + r , t + u , f = 0,1 , , F 1 ;   t = 0,1 , , T 1 .
The matrix filters extract features that are coupled across frequency and time. Note that convolving two matrices results in a matrix. The N 1 filtered outputs are combined into a ( F × T × N 1 ) cuboid feature map after the biases are added and passed through the ReLU activation function. Cuboid convolution filters are selected in the second layer to extract more complex cross-coupled features across the cuboid feature map. The cuboid convolutions result in matrices which are combined into cuboids after adding biases and passing through ReLU activation. The cuboid is passed through the max pooling layer, flattened, and passed through the FCN. The output of the FCN of the unichannel classifier of channel d is the posterior class probability vector P d = ( P d , ω 1 , P d , ω 2 , . . , P d , ω Ω ) , and the test input is assigned to class ω i if
P d , ω i > P d , ω j   for   all   j i .

2.2.3. V-MatrixNet

The V-scalogram was introduced to overcome the inclusion of zeroed-out coefficients in the Z-scalogram. However, the resulting non-rectangular matrix presents problems for CNN classifier development because the inputs to each convolution and pooling layer are expected to be a rectangular matrix or a rectangular cuboid. The approach developed in [1] circumvents this problem by concatenating the rows of the V-scalogram, which contain only the accurately computed coefficients inside the COI, into a feature vector. The dimension of the resulting frequency-time feature vector is given by N V = f = 0 F 1 N f where N f is the duration of the frequency band f inside the COI. The feature vector extracted from the V-scalogram G V , d j ( q ) f , t of channel d is given by
G V , d j ( q ) = [ f = 0 ( F 1 ) G V , d j ( q ) f , : ] T .
where represents the row concatenation operation, T is the transpose operation, and G V , d j ( q ) f , : denotes the row f of G V , d j ( q ) f , t . The subscript j indicates that the V-scalograms are generated from the class ω j   m -ERPs.
The V-MatrixNet, illustrated in Figure 3, fuses the V-scalogram frequency-time feature vectors of the D channels into the columns of an N V × D input matrix which is given by
G V , j ( q ) f t , d = d = 0 ( D 1 ) G V , d j ( q ) , f t = 0,1 , , ( N V 1 ) ;   d = 0,1 , , ( D 1 ) .
where in this case is the operator for fusing column vectors into a matrix and f t is the column index. The index f t indicates that each element of the feature vector is a frequency-time element of the V-scalogram G V , d j ( q ) f , t .
The matrix G V , j ( q ) f t , d can be regarded as a frequency-time( f t )-spatial( d ) matrix. In the first layer, G V , j ( q ) f t , d is convolved with a matrix filter f [ 1 , n ] f t , d , f t = α , , 0 , , α ;   d = β , , 0 , , β , n = 1,2 , , N 1 ; the output of filter n is given by the matrix convolution
Y V , j q 1 , n f t , d = r = α α u = β β f 1 , n r , q G V , j q f t + r , d + u , f t = 0,1 , , ( N V 1 ) ;   d = 0,1 , , ( D 1 ) .
A bias is added to the filtered output and passed through the ReLU activation function to generate the matrix feature map. The N 1 filtered outputs are combined into a ( N v × D × N 1 ) feature cuboid. The feature cuboid is convolved with cuboid filters, and the resulting feature maps are combined into a more complex feature cuboid which is passed through the pooling layer, flattened, and passed through the FCN. The softmax layer computes the posterior class probabilities, which are stored in a probability vector, and the input is assigned to the class given by the rule in Equation (3).

2.2.4. V-VectorNet

The V-MatrixNet reduces to the V-VectorNet for unichannel classification. In order to determine the best channel classifier and to compare the performance against the multichannel V-MatrixNet fusion classifier, a V-VectorNet classifier is developed independently for the m -ERPs of each channel. In the first convolution layer, the column feature vector G V , d j ( q ) of channel d is convolved with column filters f [ 1 , n ] f t , f t = α , , 0 , , α ;   n = 1,2 , , N 1 , and the column output of the n t h filter is given by
Y V , j q 1 , n f t = r = α α f 1 , n r G V , d j q f t + r , f t = 0,1 , , N V 1 .
A bias is added to the filtered output, which is a column vector, and passed through the ReLU activation function to yield the column feature map. The N 1 column feature maps are combined into a ( N V × N 1 ) matrix that goes through the second convolution filter consisting of matrix filters. The matrix outputs are combined into a cuboid and passed through the max pooling layer. The output of the max pooled layer is flattened and fed into the FCN, which outputs the posterior class probability vector. The rule in Equation (5) is used to assign the input to a class.
In summary, the outputs of the convolution layers take on different shapes depending on the shapes of the inputs and the choice of convolution features. The desired fusion of the frequency, time, and spatial domains is controlled by the shapes and sizes of the filters. More complex discriminatory features can be extracted by making the networks deeper; that is, by increasing the number of network layers.

2.3. ERP Data Set

The EEG/ERP data used in this study to demonstrate the application and evaluate the performance of the multidomain models was the same as used in [1]. The data was downloaded from: https://eeglab.org/tutorials/10_Group_analysis/study_creation.html#description-of-the-5-subject-experiment-tutorial-data (accessed on 1 October 2022).
This data set was selected because it is compact and serves the purpose of demonstrating the application of the multidomain models for both single-subject-customized and group-based classification. Furthermore, given that the multidomain CNN models are extensions of the unichannel CNN models in [1], the performance trends can be compared to determine consistencies and/or discrepancies between the two studies. The design of the multidomain models, however, is not restricted to this particular data set or to any other data set. Details of the data relevant to this study are as follows:
  • Task: Auditory binary semantic task. Subjects distinguished between synonymous and non-synonymous word pairs.
  • Number of ERP classes: Two (synonymous, non-synonymous).
  • Number of subjects: 5.
  • Number of channels: 64.
  • Sampling rate: 200 Hz; Single trial duration: 1 s; Number of samples in single trials: 200
  • Number of single trials for each subject: 195 synonymous and 195 non-synonymous.
Complete details of the EEG data can be found on the referenced website, and the details of the m -ERPs extracted from the EEG can be found in [1].

2.4. Group-Based Cross-Validation

The goal of group-based cross-validation is to train the multidomain classifiers with the m -ERPs of a set of subjects and test the m -ERPs of the individual subjects not included in the training set. In order to do so in a systematic fashion, k -fold cross-validation is modified so that the folds are defined with respect to subjects. In this cross-validation approach, which we refer to as “ k -subject-fold cross-validation,” each fold consists of the m -ERPs of ( B / k ) subjects, where B and k are the number of subjects and folds, respectively. The classifier is trained with the m -ERPs in ( k 1) folds and validated (tested) on the m -ERPs of each subject in the left-out fold. As in regular k -fold cross-validation, the process is repeated k times so that the ERPs of all subjects are tested. The final result is obtained by averaging the classification accuracies within and across the k repetitions. The process can be repeated several times and averaged by first shuffling the order of the subjects so that the subjects fall in different folds. For the special case ( k = B ), that is, each fold contains the m -ERPs of only one subject, the procedure reduces to leave-one-subject-out cross-validation.

2.5. Channel Selection

Although fusing ERPs from multiple channels is likely to improve performance, designing multichannel ERP fusion classifiers is a challenging problem because the following issues have to be addressed:
(a)
Should all channels be used in the design? The answer is generally no because including channels that elicit ERPs that do not carry useful discriminatory information is equivalent to adding noise, and the additional dimensions lead to overfitting, which has a negative impact on performance.
(b)
If all channels are not used, how should a subset of channels be selected? Temporospatial PCA is one method that can be used to identify channels and time windows that capture effects elicited by stimuli [23,24]. Alternatively, channel selection algorithms can be applied to select a subset of useful channels [1,25,26,27].
(c)
How are the ERPs of the selected channels fused? This issue has already been addressed in the development of the Z-CuboidNet and V-MatrixNet models in Section 2.2.
In general, the answers to the questions raised above are not straightforward and are often application dependent. The generalized rank-of-rank sum channel selection strategy described in [1] was used to select the top 12 channels for each subject and across all 5 subjects. The ranking for the 5 subjects pooled into a group is obtained by summing the single subject ranks and ranking the rank-sums. The selected sets of channels are listed in Table 1, in which rank 1 is the best channel and rank 12 is the 12th best channel. Observe that the channel rankings vary across the 5 subjects.

2.6. Morlet Wavelet Transform

The analytic Morlet CWT [1,28,29] is selected in this study because it is a good choice for analyzing the oscillatory behavior of ERPs and EEGs. The analytic Morlet mother wavelet, a product of a complex exponential signal of frequency f 0 and a zero-mean Gaussian window with variance σ 2 , is given by
Ψ f 0 ( t ) = A ( e t 2 2 σ 2 ) ( e j 2 π f 0 t )
where, the constant A ensures that the wavelet energy is equal to one. The analytic wavelet coefficients are complex, and the scalogram, represented by G f , t , is a plot of the CWT amplitude | G f , t | or power | G f , t | 2 as a function of discrete frequency and discrete time. Scalograms of other CWTs, such as Morse [30] and Bump [31], can also be used for the development of the Z-CuboidNet and V-MatrixNet models if they are more suited for a given application.

2.7. CNN Architectures and Hyperparameters

The reason for selecting CNNs was explained in the Introduction. For consistency, the CNNs used in all 4 classification models had 2 convolution layers followed by a max pooling layer and an FCN. This section describes the shape, dimensions, and number of filters in the convolution layers; dimensions and strides of the max pooling filters; the number of layers in the FCN and the activations in the FCN; and the training hyperparameters.
The scalogram dimensions of each m -ERP were ( f = 108 ) × ( t = 200 ) . The top D out of the 64 channels were selected; therefore, the dimensions of the input to the Z-CuboidNet models were 108 × 200 × D . The input to Z-MatrixNet had dimensions 108 × 200 . The feature vector extracted from the V-scalogram was a ( N V = 17,056 ) -dimensional column vector (see Section 2.3); therefore, the input dimension for V-MatrixNet was 17,056 × D . The input to the V-VectorNet was a 17,056 -dimensional column vector. The number of filters in the first and second convolution layers are denoted by N 1 and N 2 , respectively. The FCN had 3 layers of neurons. The architectures, hyperparameters, and training options of the 4 classification models are summarized in Table 2.

3. Experiments and Results

This section describes the extensive set of experiments designed to evaluate the performances of the four classification models using the binary ERP data described in Section 2.3. The experimental hyperparameters were set as follows:
(a)
The top D = 4, 8, and 12 ranked channels listed in Table 1 were selected to implement the multidomain classifiers in order to demonstrate the improvements that can be expected by increasing the number of channels. The unichannel classifiers were also implemented for the top 12 channels. The unichannel classifier that gave the best result, denoted by D = 1, was used for comparisons against the multidomain classifiers.
(b)
Subsample averaging parameter m was set to 1, 2, and 4 to demonstrate single trial and small-average m -ERP classification.
(c)
The number of single trials for each class was 195. The m -subsample averaging procedure was used to generate an equal number of m = 2 and 4 m -ERPs. That is, 195 2-ERPs/class and 195 4-ERPs/class were generated.
(d)
5-fold cross-validation was used for the single-subject-customized classifier design.
(e)
Leave-one-subject-out cross-validation was used for group-based classifier design.
(f)
Cross-validation was repeated 50 times (50 runs). Each run began by initializing the models with a different set of random weights.
The number of m -ERPs tested for each subject was (195 m -ERPs/class) (2 classes) (50 runs) = 19,500. The classification accuracy, expressed as a percentage, was estimated as the number of correctly classified m -ERPs divided by 19,500. For group-based classification, the classification accuracy was estimated in the same manner because leave-one-subject-out cross-validation was used. That is, the number of m -ERPs in the cross-validation folds was the same for both cases. The classifiers were implemented using the PyTorch library.

3.1. Customized Classification Experiments

This set of experiments involved the classification of the ERPs of each individual subject using only the ERPs collected from the individual. As noted in the Introduction, the need for this approach using single trials can typically arise in the design of systems such as customized BCIs for individuals [32,33,34,35]. In order to demonstrate the improvements that can be expected by increasing the averaging parameter m , the experiments were also conducted with 2-ERPs and 4-ERPs.

3.1.1. Customized Unichannel Experiments

For each subject, the Z-MatrixNet and V-VectorNet unichannel classifiers were implemented for their 12 top-ranked channels. The complete set of results is presented in Table A1, Table A2 and Table A3 in Appendix A for m taking values 1, 2, and 4. The channels labeled 1–12 in the first column of the three tables are the ranked channels listed in Table 1. For example, the results for the channel labeled CH = 1 contain the classification accuracies of the top-ranked (rank 1) channels P3, PO4, O2, O2, and C2 of Subjects B 1 , B 2 , B 3 , B 4 , and B 5 , respectively.

3.1.2. Customized Multichannel Experiments

The Z-CuboidNet and V-MatrixNet models were implemented for the D = 4, 8, and 12 top-ranked channels listed in Table 1. The classification accuracies for each D were determined for m = 1, 2, and 4. The results are presented in Table 3. Each result in the table is the average of testing 19,500 m -ERPs. Note that m = 2 and 4 are not included for D = 12 because the multidomain models gave 100% for m = 1 across all five subjects; therefore, there was no need to increase m beyond unity. The table also contains results for D = 1 , which is the best unichannel result for each subject extracted from Table A1, Table A2 and Table A3 in Appendix A. That is, the results for D = 1 are the best Z-MatrixNet and V-VectorNet results. In order to facilitate analyses of the performance trends, the average classification accuracies of the customized classifiers of the five subjects are presented in Figure 4 as functions of D and m . That is, each bar is the average across the five customized classifiers for a given combination of D and m .

3.2. Group-Based Classification Experiments

The group-based classification experiments are different from the customized set of experiments in the previous section because these experiments involve classifying the m -ERPs of each subject using the models trained with the m -ERPs of the other subjects in the group. In this set of experiments, the leave-one-subject-out method was used to design the models. For example, the models were trained with the m -ERPs of subjects B 2 , B 3 , B 4 , and B 5 and were tested on the m -ERPs of subject B 1 . The results are summarized in Table 4 and are interpreted in the same manner as the results in Table 3. The 12 unichannel results are presented in Table A4, Table A5 and Table A6 in Appendix A. Unlike the customized case where the channels of each subject were ranked independently, the 12 top channels of the group-based classifiers have a common ranking. The average classification accuracies of the group-based classifiers of the five subjects are presented in Figure 5.

4. Discussion of Results

The four classification models can be ranked according to the classification accuracies in Table 3 and Table 4. The results in:
(a)
Table 3 show that only marginal differences exist between the performances of the customized Z-CuboidNet and the V-MatrixNet classifiers across all five subjects, all values of D , and all values of m . The Z-CuboidNet and V-MatrixNet multichannel fusion models outperform the best unichannel ( D = 1 ) Z-MatrixNet and V-VectorNet models, respectively. For a given D , the performance improves by increasing m , and for a given m , the performance improves by increasing D .
(b)
Table A1, Table A2 and Table A3 show that there is no single channel that is best across the five subjects.
(c)
Table 4 show that the performance trends of the group-based classifiers are similar to those of the customized classifiers in Table 3. That is, the Z-CuboidNet and V-MatrixNet performances differ marginally, the multichannel multidomain models outperform the best unichannel models, and the performance improves when m and D are increased.
(d)
Table A4, Table A5 and Table A6 show that there is no single channel that is best across the five subjects.
(e)
Table 4 show that the accuracies of the group-based classifiers are slightly lower than those of the corresponding customized classifiers (Table 3) for small values of D and m . The drop in accuracies can be attributed to the fact that none of the m -ERPs of a test subject are included in the group training set.
Both customized multidomain models yielded single trial classification accuracies exceeding 90% for eight channels across all five subjects. Classification accuracies of 100% were obtained for all subjects when 12 channels were used. The group-based Z-CuboidNet models yielded single trial classification accuracies exceeding 90% for eight channels and yielded 100% for twelve channels across all five subjects. The single trial ( m = 1 ) average classification accuracy of the five subjects improved from 81.06 to 100% when the number of channels ( D ) was increased from one to twelve. This improvement is dramatic. Moreover, the ability to obtain higher accuracies by increasing the number of channels is especially noteworthy because it is a clear indication that the ERPs of the spatially distributed channels carry complementary information that can be exploited to improve performance.
There is no clear winner between the Z-CuboidNet and V-MatrixNet models for both customized and group-based classification. The zeroed-out region in the input cuboid of the Z-CuboidNet is common to all ERP classes; therefore, it is likely to be ignored by the convolution filters in the process of extracting discriminatory features during training. The Z-CuboidNet offers the most direct way of fusing the scalograms. In the previous study involving unichannel customized m -ERP classification, the results for the CNN implementations of the classifiers were mixed with no clear winner between the V-scalogram and Z-scalogram classifiers. The results of the extensions of the unichannel models to multichannel models are, therefore, in agreement with the previous study. It was also established in [1] that the V- and Z-scalogram CNN classifiers outperform the V- and Z-scalogram implementations of the other classifiers (SVM, RF, k-NN, and MLP). Undoubtedly, the Z-CuboidNet and V-MatrixNet would outperform the multichannel implementations of these other classifiers. Furthermore, the other classifiers do not offer any elegant manner to fuse the V- and Z-scalograms because they require vector inputs. Vector inputs can be generated by concatenating the rows of the scalogram of each channel into a unichannel vector, followed by the concatenation of the unichannel vectors of the multiple channels into a super-sized multichannel vector. The resulting multichannel vector input will have spatial components separated by the length of the unichannel vector, which makes it difficult to design convolution filters to locally couple the spatial information.
Finally, it is important to note that high classification accuracies were obtained with no attempts to optimize the performance of the multidomain classifiers. In general, the multidomain classifiers have the potential for further improvements in performance by increasing the number of channels, increasing the network depth, tweaking the hyperparameters, and, when applicable, by increasing the averaging parameter m . Increasing the network depth, however, requires an increase in the training set size which could be a limitation in practical applications in which collecting large ERP ensembles is problematic. Additionally noteworthy is that although the primary focus was on ERP classification, the generalized formulations of the two multidomain models make them easily adaptable to other problems involving multisensor signal classification.

5. Conclusions

The Z-CuboidNet and V-MatrixNet multidomain models were introduced to improve ERP classification by fusing frequency, time, and spatial information from multichannel ERP recordings. Both CNN-based models do not require expert extraction of hand-engineered features for the input, which not only facilitates classifier design but also avoids the problems associated with selecting feature sets through trial and error. An extensive set of experiments involving customized and group-based classification of ERPs were conducted to demonstrate the application and performance of the multidomain and unichannel classifiers. The results clearly showed that the multidomain classifiers consistently outperformed the unichannel classifiers. Additionally, high classification accuracies were obtained from the multidomain models trained and tested on the single trials of single subjects, which is an especially important contribution for the design of customized classifiers for BCI-type applications in which the interface is expected to be controlled by a single stimulus presentation. It was also shown that the multidomain models trained on the ERPs from a group of subjects can accurately classify ERPs of single subjects not included in the training group, which is a notable contribution for many applications related to classifying brain activity and brain disorders. The multidomain models are highly scalable to large subject groups and to more ERP classes which are crucial requirements for practical deployment. Future efforts will focus on deploying the models for real BCI applications and large group-based brain disorder classification.

Author Contributions

Conceptualization, L.G., R.S.G. and X.C.; methodology, X.C., L.G. and R.S.G.; software, X.C.; validation, X.C., L.G. and R.S.G.; writing–original draft preparation, L.G. and X.C.; writing–review and editing, L.G., R.S.G. and X.C.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Publicly available data were used in this study. The data can be found here: https://eeglab.org/tutorials/10_Group_analysis/study_creation.html#description-of-the-5-subject-experiment-tutorial-data (accessed on 1 October 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

The customized and group-based unichannel classification accuracies of the 12 top-ranked channels are presented in this appendix. The channels, Z-MatrixNet, and V-VectorNet, are represented by CH, Z-M, and V-V, respectively. Each result in the tables is the average of testing 19,500 m -ERPs. The best result across the 12 channels is in boldface.
Table A1. Unichannel accuracies of the customized classifiers for single trials (1-ERPs).
Table A1. Unichannel accuracies of the customized classifiers for single trials (1-ERPs).
CHSubject
B 1 B 2 B 3 B 4 B 5
Z-MV-VZ-MV-VZ-MV-VZ-MV-VZ-MV-V
186.7385.7182.6583.6775.5174.1378.7877.2181.0381.39
284.9785.0381.9782.4575.8576.1177.5578.4480.8581.05
384.6983.6782.5281.2976.1976.5376.9478.2381.1480.67
483.8384.0181.2982.8575.2976.7278.3777.9781.2681.33
582.9583.8180.2780.9574.7875.5876.1978.9580.4281.12
682.9984.1379.0579.5974.1574.6377.6278.2080.3580.52
781.2281.6379.1278.4473.4073.2775.2976.3480.9380.06
881.6981.9878.9177.7674.4972.2476.0576.1680.6180.82
981.8181.1378.3178.1473.4873.0175.4276.1980.4981.32
1081.9081.6377.3478.9673.7674.9077.7075.7480.1980.26
1181.2881.5578.2078.8173.6473.3977.4876.7980.0780.21
1281.8081.1478.8578.6673.1573.2578.0778.1780.3181.57
Table A2. Unichannel accuracies of the customized classifiers for 2-ERPs.
Table A2. Unichannel accuracies of the customized classifiers for 2-ERPs.
CHSubject
B 1 B 2 B 3 B 4 B 5
Z-MV-VZ-MV-VZ-MV-VZ-MV-VZ-MV-V
192.3590.4890.4489.0883.2784.1087.3187.2087.0686.06
291.6792.2188.7390.9783.5682.9086.1787.1085.7883.61
388.9491.7790.7989.4283.2784.8586.7085.9285.6586.25
491.7092.1489.0590.7982.5883.6085.4986.4188.4583.48
590.7191.3088.5990.9182.9182.7885.4184.9584.1785.94
691.3589.1388.7591.1282.0485.4386.7988.3487.9884.11
791.3891.9789.8590.0883.6185.1485.4985.5188.1186.60
890.4192.2789.4288.7981.6683.6885.2786.2787.5885.75
989.5591.3889.6789.2882.7384.0684.7387.8185.8285.68
1091.4090.9389.8887.5981.3683.9883.2487.7185.0283.17
1191.3589.6188.6488.3580.8582.9085.7685.1183.6486.13
1289.7091.8387.4288.4981.4182.7385.7686.3083.8486.16
Table A3. Unichannel accuracies of the customized classifiers for 4-ERPs.
Table A3. Unichannel accuracies of the customized classifiers for 4-ERPs.
CHSubject
B 1 B 2 B 3 B 4 B 5
Z-MV-VZ-MV-VZ-MV-VZ-MV-VZ-MV-V
196.2395.9695.7193.3494.1191.7795.0893.7494.7994.32
295.7994.6992.5094.5392.1993.0295.4695.5594.7591.29
396.3496.2594.1292.6693.0993.3494.9894.0995.3892.17
496.2395.9493.9995.1291.6692.3493.9695.3093.6892.30
594.5294.7894.1192.8993.7491.5695.2594.8293.6292.38
694.3594.9295.2095.9392.6690.6894.4494.1493.4393.97
797.4096.0693.4393.0191.6592.6494.0492.6194.9193.81
894.5094.8391.2495.2490.1591.1894.8594.5092.1991.98
993.4593.1193.2392.9391.8590.1694.5392.3892.9391.95
1095.8994.9694.3594.5592.6492.6493.6493.8293.2091.73
1195.1696.0991.8493.4591.7590.8594.9693.9791.1192.98
1294.2295.3493.3791.9391.8691.5593.3594.2893.1593.24
Table A4. Leave-one-subject-out unichannel accuracies for 1-ERPs.
Table A4. Leave-one-subject-out unichannel accuracies for 1-ERPs.
CHSubject
B 1 B 2 B 3 B 4 B 5
Z-MV-VZ-MV-VZ-MV-VZ-MV-VZ-MV-V
Cz81.4382.8681.7082.7272.5271.4480.3480.6879.3279.25
C282.5282.6481.3681.2273.6272.5979.9378.6578.1678.91
T882.2882.5081.7982.2172.5173.1379.6679.3278.7079.01
C681.2581.8882.6482.3372.3274.8878.6579.7977.3077.91
C481.7682.1980.6181.6072.6373.1279.8379.3277.4777.84
C380.9980.8280.9580.1872.8272.6979.1278.8577.6176.91
F682.8182.3980.0681.1072.8073.3179.2379.0478.4677.89
C182.6581.9380.9480.1971.1273.7078.7278.9976.8176.26
C581.5181.9981.4581.1372.0672.5676.7678.9276.6777.41
CP182.1980.9381.0981.5372.5672.9379.8079.6676.8077.01
Fp281.6382.0981.5480.7771.9872.3978.8578.9875.7675.58
F881.5881.8681.1782.1372.1970.3078.5777.9676.3675.45
Table A5. Leave-one-subject-out unichannel accuracies for 2-ERPs.
Table A5. Leave-one-subject-out unichannel accuracies for 2-ERPs.
CHSubject
B 1 B 2 B 3 B 4 B 5
Z-MV-VZ-MV-VZ-MV-VZ-MV-VZ-MV-V
Cz86.8385.9585.7985.3581.8081.7284.7385.1384.6683.93
C285.9586.9585.3384.3679.7980.8184.5185.9783.4284.33
T885.7085.1984.4683.8779.1279.2385.4384.9583.6284.43
C686.6685.1384.3185.3181.9879.4685.3885.5583.1483.12
C486.1786.3384.3485.4681.0081.0584.0885.5082.6482.91
C386.1086.4885.3485.1680.9278.6185.1185.3182.9083.91
F686.2386.3484.4484.2580.6280.4684.4085.5582.7383.57
C185.5785.3283.8584.1781.4479.1784.6385.5082.9282.27
C586.5186.4284.1483.4281.4781.4184.8284.7783.1682.16
CP186.5186.1783.3983.6080.4481.3585.4184.6983.2983.01
Fp285.4085.1584.7383.9479.5079.5985.0084.0882.4481.88
F884.2983.9383.0783.4679.3877.8983.7683.7581.5782.08
Table A6. Leave-one-subject-out unichannel accuracies for 4-ERPs.
Table A6. Leave-one-subject-out unichannel accuracies for 4-ERPs.
CHSubject
B 1 B 2 B 3 B 4 B 5
Z-MV-VZ-MV-VZ-MV-VZ-MV-VZ-MV-V
Cz95.0295.4494.6493.6190.0690.3492.7792.2092.8293.36
C296.0597.0695.9194.4890.4387.7591.0594.8292.2791.61
T895.5296.7093.9692.6289.5887.8294.5994.8093.7693.35
C693.3696.4492.3993.5388.8388.7591.7994.7092.1993.26
C493.3293.7192.9792.6490.2189.0692.3394.7291.6993.48
C396.1193.7994.0095.0388.8289.1292.1991.5393.8092.27
F696.6293.8591.5195.4886.5787.3192.7094.4193.5192.72
C195.1493.9291.9593.4086.9887.1391.8793.3293.6593.28
C593.4494.0192.1394.7489.9588.1993.6592.0693.4392.05
CP196.3095.7093.1793.1888.4488.4290.4291.4992.7692.87
Fp294.3493.5391.3093.1287.8489.3692.9894.4792.1893.44
F894.2195.8990.8494.3386.8789.2292.9592.5891.2492.78

References

  1. Chen, X.; Gupta, R.S.; Gupta, L. Exploiting the Cone of Influence for Improving the Performance of Wavelet Transform-Based Models for ERP/EEG Classification. Brain Sci. 2023, 13, 21. [Google Scholar] [CrossRef] [PubMed]
  2. Wei, X.; Zhou, L.; Chen, Z.; Zhang, L.; Zhou, Y. Automatic seizure detection using three-dimensional CNN based on multi-channel EEG. BMC Med. Inform. Decis. Mak. 2018, 18, 71–80. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Türk, Ö.; Özerdem, M.S. Epilepsy detection by using scalogram based convolutional neural network from EEG signals. Brain Sci. 2019, 9, 115. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Lee, H.K.; Choi, Y.S. Application of continuous wavelet transform and convolutional neural network in decoding motor imagery brain-computer interface. Entropy 2019, 21, 1199. [Google Scholar] [CrossRef] [Green Version]
  5. Mao, W.; Fathurrahman, H.; Lee, Y.; Chang, T. EEG dataset classification using CNN method. J. Phys. Conf. Ser. 2020, 1456, 012017. [Google Scholar] [CrossRef] [Green Version]
  6. Mammone, N.; Ieracitano, C.; Morabito, F.C. A deep CNN approach to decode motor preparation of upper limbs from time–frequency maps of EEG signals at source level. Neural Netw. 2020, 124, 357–372. [Google Scholar] [CrossRef] [PubMed]
  7. Aslan, Z.; Akin, M. A deep learning approach in automated detection of schizophrenia using scalogram images of EEG signals. Phys. Eng. Sci. Med. 2022, 45, 83–96. [Google Scholar] [CrossRef]
  8. Kaur, A.; Shashvat, K. Implementation of convolution neural network using scalogram for identification of epileptic activity. Chaos Solitons Fractals 2022, 162, 112528. [Google Scholar] [CrossRef]
  9. Stephen, M. Singularity detection and processing with wavelets. IEEE Trans Inf. Theory 1992, 38, 617–643. [Google Scholar]
  10. Torrence, C.; Compo, G.P. A practical guide to wavelet analysis. Bull. Am. Meteorol. Soc. 1998, 79, 61–78. [Google Scholar] [CrossRef]
  11. Nobach, H.; Tropea, C.; Cordier, L.; Bonnet, J.P.; Delville, J.; Lewalle, J.; Farge, M.; Schneider, K.; Adrian, R. Review of some fundamentals of data processing. In Springer Handbooks; Springer: Berlin/Heidelberg, Germany, 2007; pp. 1337–1398. [Google Scholar]
  12. Lilly, J.M. Element analysis: A wavelet-based method for analysing time-localized events in noisy time series. Proc. R. Soc. A Math. Phys. Eng. Sci. 2017, 473, 20160776. [Google Scholar] [CrossRef] [Green Version]
  13. Amerineni, R.; Gupta, L.; Steadman, N.; Annauth, K.; Burr, C.; Wilson, S.; Barnaghi, P.; Vaidyanathan, R. Fusion Models for Generalized Classification of Multi-Axial Human Movement: Validation in Sport Performance. Sensors 2021, 21, 8409. [Google Scholar] [CrossRef]
  14. Kook, H.; Gupta, L.; Molfese, D.; Fadem, K. C Multi-stimuli multi-channel data and decision fusion strategies for dyslexia prediction using neonatal ERPs. Pattern Recognit. 2005, 38, 2174–2184. [Google Scholar] [CrossRef]
  15. Gupta, L.; Chung, B.; Srinath, M.D.; Molfese, D.L.; Kook, H. Multichannel fusion models for the parametric classification of differential brain activity. IEEE Trans. Biomed. Eng. 2005, 52, 1869–1881. [Google Scholar] [CrossRef] [PubMed]
  16. Polikar, R. Ensemble Machine Learning; Springer: Boston, MA, USA, 2012; pp. 1–34. [Google Scholar]
  17. Kota, S.; Gupta, L.; Molfese, D.; Vaidyanathan, R. Diversity-Based Selection of Polychotomous Components for Multi-Sensor Fusion Classifiers. J. Eng. Med. 2013, 227, 655–662. [Google Scholar]
  18. Kuncheva, L.I. Combining Pattern Classifiers: Methods and Algorithms; John Wiley & Sons: New York, NY, USA, 2014. [Google Scholar]
  19. Amerineni, R.; Gupta, R.; Gupta, L. Multidomain Object Classification Models Inspired by Multisensory Integration in the Brain. Brain Sci. 2019, 9, 3. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Woody, C.D. Characterization of an adaptive filter for the analysis of variable latency neuroelectric signals. Med. Biol. Eng. 1967, 5, 539–554. [Google Scholar] [CrossRef]
  21. Aunon, J.I.; McGillem, C.D.; Childers, D.G. Signal processing in evoked potential research: Averaging and modeling. Crit. Rev. Bioeng. 1981, 5, 323–367. [Google Scholar]
  22. Gupta, L.; Molfese, D.L.; Tammana, R.; Simos, P.G. Nonlinear alignment and averaging for estimating the evoked potential. IEEE Trans. Biomed. Eng. 1996, 43, 348–356. [Google Scholar] [CrossRef]
  23. Dien, J.; Spencer, K.M.; Donchin, E. Parsing the late positive complex: Mental chronometry and the ERP components that inhabit the neighborhood of the P300. Psychophysiology 2004, 41, 665–678. [Google Scholar] [CrossRef]
  24. Gupta, R.S.; Kujawa, A.; Vago, D.R. A preliminary investigation of ERP components of attentional bias in anxious adults using temporospatial principal component analysis. J. Psychophysiol. 2021, 35, 223–236. [Google Scholar] [CrossRef] [PubMed]
  25. Alotaiby, T.; Abd El-Samie, F.E.; Alshebeili, S.A.; Ahmad, I. A review of channel selection algorithms for EEG signal processing. EURASIP J. Adv. Signal Process. 2015, 2015, 66. [Google Scholar] [CrossRef] [Green Version]
  26. Baig, M.Z.; Aslam, N.; Shum, H.P. Filtering techniques for channel selection in motor imagery EEG applications: A survey. Artif. Intell. Rev. 2020, 53, 1207–1232. [Google Scholar] [CrossRef] [Green Version]
  27. Guttmann-Flury, E.; Sheng, X.; Zhu, X. Channel selection from source localization: A review of four EEG-based brain–computer interfaces paradigms. Behav. Res. Methods 2022, 2022, 1–24. [Google Scholar]
  28. Auger, F.; Flandrin, P.; Gonçalvès, P.; Lemoine, O. Time-Frequency Toolbox; CNRS: Paris, France; Rice University: Houston, TX, USA, 1996; p. 46. [Google Scholar]
  29. Cohen, M.X. A better way to define and describe Morlet wavelets for time-frequency analysis. NeuroImage 2019, 199, 81–86. [Google Scholar] [CrossRef] [PubMed]
  30. Lilly, J.M.; Olhede, S.C. Generalized Morse wavelets as a superfamily of analytic wavelets. IEEE Trans. Signal Process. 2012, 60, 6036–6041. [Google Scholar] [CrossRef] [Green Version]
  31. Ahmed, S.; Mohammad, N.; Wael, A.; Ramin, G.; Zhishen, W. Comparative Analysis of Wavelet Transform for Time-Frequency Analysis and Transient Localization in Structural Health Monitoring. Struct. Durab. Health Monit. 2021, 15, 1–22. [Google Scholar]
  32. Myrden, A.; Chau, T. A passive EEG-BCI for single-trial detection of changes in mental state. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 345–356. [Google Scholar] [CrossRef]
  33. Jochumsen, M.; Knoche, H.; Kjaer, T.W.; Dinesen, B.; Kidmose, P. EEG headset evaluation for detection of single-trial movement intention for brain-computer interfaces. Sensors 2020, 20, 2804. [Google Scholar] [CrossRef]
  34. Pham Xuan, R.; Andreessen, L.M.; Zander, T.O. Investigating the Single Trial Detectability of Cognitive Face Processing by a Passive Brain-Computer Interface. Front. Neuroergonom. 2022, 2, 39. [Google Scholar] [CrossRef]
  35. Choi, H.; Park, J.; Yang, Y.M. A Novel Quick-Response Eigenface Analysis Scheme for Brain–Computer Interfaces. Sensors 2022, 22, 5860. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (a) ERP; (b) S-scalogram; (c) Z-scalogram; (d) V-scalogram.
Figure 1. (a) ERP; (b) S-scalogram; (c) Z-scalogram; (d) V-scalogram.
Sensors 23 04656 g001
Figure 2. Block diagram of the Z-CuboidNet illustrating the formation of the fusion cuboid and the CNN which accepts the cuboid as the input.
Figure 2. Block diagram of the Z-CuboidNet illustrating the formation of the fusion cuboid and the CNN which accepts the cuboid as the input.
Sensors 23 04656 g002
Figure 3. Block Diagram of the V-MatrixNet illustrating the formation of the fusion matrix and the CNN, which accepts the matrix as the input.
Figure 3. Block Diagram of the V-MatrixNet illustrating the formation of the fusion matrix and the CNN, which accepts the matrix as the input.
Sensors 23 04656 g003
Figure 4. The average classification accuracies of the customized classifiers of the 5 subjects.
Figure 4. The average classification accuracies of the customized classifiers of the 5 subjects.
Sensors 23 04656 g004
Figure 5. The average classification accuracies of the group-based classifiers of the 5 subjects.
Figure 5. The average classification accuracies of the group-based classifiers of the 5 subjects.
Sensors 23 04656 g005
Table 1. The 12 top-ranked channels for each subject and for the 5 subjects combined into a group.
Table 1. The 12 top-ranked channels for each subject and for the 5 subjects combined into a group.
Channel Rankings
Subject123456789101112
B 1 P3P1P5PzP7POzPO2PO1PO3TP8PO5P2
B 2 PO4F3CP5PO2OzPO6TP7O1POzFp1P3P5
B 3 O2OzO1PO6PO4PO2TP7POzP7PO1TP8PO3
B 4 O2OzTP7O1F3F1F5CP5PO6Fp1PO2PO4
B 5 C2P5CzC4P3P7C1FCzC6FC1FC2TP8
GroupCzC2T8C6C4C3F6C1C5CP1Fp2F8
Table 2. Architectures, hyperparameters, and training options of the 4 classification models.
Table 2. Architectures, hyperparameters, and training options of the 4 classification models.
Z-CuboidNetZ-MatrixNetV-MatrixNetV-VectorNet
Input dimensions 108 × 200 × D 108 × 200 17,056 × D 17,056 × 1
Conv-Layer 1: Filter dimensions; number; activations 3 × 3 × D 3 × 3 3 × 3 9 × 1
N 1 = 32 N 1 = 32 N 1 = 32 N 1 = 32
ReLUReLUReLUReLU
Conv-Layer 2: Filter dimensions; number; activations 3 × 3 × 32 3 × 3 × 32 3 × 3 × 32 3 × 3
N 2 = 32 N 2 = 32 N 2 = 32 N 2 = 32
ReLUReLUReLUReLU
Max Pool: Filter dimensions; Stride 2 × 2 ; 2 2 × 2 ; 2 2 × 2 ; 2 2 × 2 ; 2
FCN256, 128, 2128, 64, 2256, 128, 2128, 64, 2
FCN activations t a n h , tanh, soft t a n h , tanh, soft t a n h , tanh, soft t a n h , tanh, soft
Number of Epochs50505050
OptimizerAdamAdamAdamAdam
Learning rate0.0010.0010.0010.001
Drop out probabilities0.150.150.150.15
Table 3. Accuracies of the customized classifiers.
Table 3. Accuracies of the customized classifiers.
Model D m Subject
B 1 B 2 B 3 B 4 B 5
Z-CuboidNet1186.7382.6576.1978.7881.26
292.3590.7983.6187.3188.45
497.4095.7194.1195.4695.38
4192.7091.5083.1687.1089.13
295.1295.2891.7894.8294.41
410010097.5698.5199.37
8110010095.0994.1890.82
210010010010098.23
4100100100100100
121100100100100100
V-MatrixNet1185.7183.6776.7278.9581.57
292.2791.1285.4388.3486.60
496.2595.9393.3495.5594.32
4192.7091.5083.1687.1089.13
295.1295.2891.7894.8294.41
410010097.5698.5199.37
8110010094.5594.9191.78
210010010010098.00
4100100100100100
121100100100100100
Table 4. Leave-one-subject-out classification accuracies.
Table 4. Leave-one-subject-out classification accuracies.
Model D m Subject
B 1 B 2 B 3 B 4 B 5
Z-CuboidNet1182.8182.6473.6280.3479.32
286.8385.7981.9885.4384.66
496.6295.9190.4394.5993.80
4191.2089.2781.7285.3085.35
293.6194.0186.1890.6491.29
497.2398.0793.3895.9796.68
8195.0695.3693.1593.0390.30
210010097.4810098.39
4100100100100100
121100100100100100
V-MatrixNet1182.8682.7274.8880.6879.25
286.9585.4681.7285.9784.43
497.0695.4890.3494.8293.48
4190.4689.4380.3983.3680.82
293.2492.8284.7389.8690.07
497.5196.6593.1395.0994.65
8188.4887.2784.5584.2480.88
210094.4598.1698.3095.17
4100100100100100
121100100100100100
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, X.; Gupta, R.S.; Gupta, L. Multidomain Convolution Neural Network Models for Improved Event-Related Potential Classification. Sensors 2023, 23, 4656. https://doi.org/10.3390/s23104656

AMA Style

Chen X, Gupta RS, Gupta L. Multidomain Convolution Neural Network Models for Improved Event-Related Potential Classification. Sensors. 2023; 23(10):4656. https://doi.org/10.3390/s23104656

Chicago/Turabian Style

Chen, Xiaoqian, Resh S. Gupta, and Lalit Gupta. 2023. "Multidomain Convolution Neural Network Models for Improved Event-Related Potential Classification" Sensors 23, no. 10: 4656. https://doi.org/10.3390/s23104656

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop