Next Article in Journal
Impact of CoAP and MQTT on NB-IoT System Performance
Next Article in Special Issue
Single-Trial Cognitive Stress Classification Using Portable Wireless Electroencephalography
Previous Article in Journal
Low-Power Highly Robust Resistance-to-Period Converter
Previous Article in Special Issue
Portable Drowsiness Detection through Use of a Prefrontal Single-Channel Electroencephalogram
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

EEG-Based Identity Authentication Framework Using Face Rapid Serial Visual Presentation with Optimized Channels

1
Key Laboratory for NeuroInformation of Ministry of Education, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu 610000, China
2
China National Digital Switching System Engineering and Technological Research Center, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(1), 6; https://doi.org/10.3390/s19010006
Submission received: 30 October 2018 / Revised: 3 December 2018 / Accepted: 10 December 2018 / Published: 20 December 2018
(This article belongs to the Special Issue EEG Electrodes)

Abstract

:
Electroencephalogram (EEG) signals, which originate from neurons in the brain, have drawn considerable interests in identity authentication. In this paper, a face image-based rapid serial visual presentation (RSVP) paradigm for identity authentication is proposed. This paradigm combines two kinds of biometric trait, face and EEG, together to evoke more specific and stable traits for authentication. The event-related potential (ERP) components induced by self-face and non-self-face (including familiar and not familiar) are investigated, and significant differences are found among different situations. On the basis of this, an authentication method based on Hierarchical Discriminant Component Analysis (HDCA) and Genetic Algorithm (GA) is proposed to build subject-specific model with optimized fewer channels. The accuracy and stability over time are evaluated to demonstrate the effectiveness and robustness of our method. The averaged authentication accuracy of 94.26% within 6 s can be achieved by our proposed method. For a 30-day averaged time interval, our method can still reach the averaged accuracy of 88.88%. Experimental results show that our proposed framework for EEG-based identity authentication is effective, robust, and stable over time.

1. Introduction

Identity authentication is an essential safety precaution in many fields, including national security, public security, e-commerce, etc. Conventional authentication systems are based on passwords that are entered with a keyboard, touch screen, or mouse. More recent approaches employ biometric mechanisms, especially approaches based on neural activities (which are currently being explored). Compared to conventional biometrics, such as faces, fingerprints or DNA, brain biometrics [1] offers interesting advantages. Brain biometrics cannot be acquired from a deceased person; furthermore, they are difficult to coerce, since threats of violence or blackmail are very likely to put significant stress on the user, and stress will strongly impact an individual’s brain activities [2]. Electroencephalogram (EEG) signals, which originate from neurons in the brain, have drawn considerable interests from researchers, due to its ease of use, portability, low cost and high temporal resolution [3]. Since an individual’s neural activity pattern is unique [4], it is impossible to imitate one’s mind [5], and circumvent the biometry field [6]; this biometric trait may change the traditional “pass-word” into “pass-thought”.
The first EEG-based identity authentication approach [7] was proposed in the late 90s, then a number of different EEG-based authentication methods were proposed. These methods can be divided into two categories: spontaneous or evoked EEGs, based on the absence or presence of a stimulus. The former includes rest eyes-open/eyes closed (REO/REC), whereas the latter involves visual stimulus presentations, mental tasks, and emotional stimuli.
In 1999, Poulos et al. [7] developed the first identity authentication system based on EEG signals. The EEG data of 4 users and 75 imposters under REC condition were collected. Auto-regressive parameters and learning vector quantization network were adopted, and the correct recognition rates of 72% and 84% were achieved, respectively. Palaniappan et al. [8] constructed a dataset of VEP signals from 20 subjects. The subjects focused on recognizing stimulus images from the Snodgrass and Vanderwart picture set [9]. The highest accuracy of 92.84% was obtained using the simplified fuzzy adaptive resonance theory. Sun et al. [10] collected the EEG signals of nine subjects while they imagined moving their right or left index finger. The researchers concluded that imagining the movements of the left index finger was more appropriate for identity identification with an accuracy of 95.6%. M. Abo-Zahhad et al. [11] proposed a novel authentication system based on the fused features of EEG and EOG. The lowest verification equal error rates (EERs) were achieved using score fusion for relaxation and VEPs with EERs of 1.3% and 1.41%, respectively, in a database of 22 subjects.
Although these previous works have achieved amazing progress in EEG-based authentication, there are some key issues that needed to be addressed for practical applications. First, current works are mainly based on event-related potentials (ERPs) detection, which usually have to average several trials data to improve the signal to noise ratio, hence the time required for authentication using those methods are relatively long; second, in most works, the neural activities are collected using 32 or more channels of EEG acquisition devices, which makes the set up and preparation inconvenient. Therefore, channel selection and optimization is essential for making the EEG-based identity authentication more feasible for use in daily life. Third, since functional brain activities are constantly changing, one clear question regarding identity authentication based on brain activities is whether it is stable over time. Hence, permanence should be considered as an important evaluation criteria.
The face is a traditional biometric trait for identity authentication [12], and existing studies in cognitive neuroscience have shown that our brain has specific regions when processing face stimulus, and the brain activity response to one’s own face is remarkedly different from the response to familiar or unfamiliar non-self-faces [13,14]. Hence, an EEG-based identity authentication framework using self and non-self-face images is proposed in this study, to combine face and EEG, two kinds of biometric traits. One interesting and meaningful related work was accomplished by Yeom et al. [15]. They used self or non-self-face images as stimuli to evoke subject-specific brain activities based on neurophysiological evidence from both EEG and fMRI. A unique subject-specific brain-wave pattern called visual self-representation was elicited by Yeom’s experimental paradigm. They obtained an average accuracy of 86.1% across 10 subjects using a non-liner support-vector machine. However, the time required for authentication was at least 31.5 s, which makes the system difficult for practical use. In addition, no real imposter was used to test the system’s performance.
To achieve rapid authentication, facial images are presented by rapid serial visual presentation (RSVP) [16] in our paradigm, which can present a large number of stimuli in a short time, eliciting strong event-related potentials (ERPs) [17]. The EEG data that reflect the neural responses to self and non-self-face stimuli are investigated. Significant differences are found under different situations. On the basis of this, an authentication method based on Hierarchical Discriminant Component Analysis (HDCA) [18,19] and Genetic Algorithm (GA) [20,21] was proposed to build subject-specific models with optimized fewer channels. As GA is an adaptive probabilistic search algorithm for global optimization inspired by the laws of natural selection and genetics, the standard parallel GA is used to search for optimized channels. Moreover, the accuracy and stability of the model over time are evaluated to demonstrate the effectiveness and robustness of our method.

2. Materials and Methods

2.1. Particpants

Forty-five subjects with a mean age of 22.4 years (standard deviation [SD] = 2.1 years) with normal or corrected to normal visual acuity participated in the experiment. None of the subjects had a history of neurological disease. All subjects gave their informed consent for inclusion before they participated in the study. The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Ethics Committee of China National Digital Switching System Engineering and Technological Research Center (NDSC-EEG-20180322-03). All subjects received monetary compensation for participating in the experiment.
The subjects were divided into two groups: Users (15 subjects), imposters (30 subjects, two for each user).

2.2. Stimuli

Stimuli consists of 195 face images (15 user face images, one for each user; 180 non-user face images, which are collected from student volunteers with their written informed consents for academic research use). All the face images present only facial information with no expression. Each grayscale face image is resized to 400 × 400 pixels. The luminance of the entire stimulus array is measured and do not differ between user face images and non-user-face images.

2.3. Experimental Procedure and Design

Stimuli are centrally presented against a black background on a 23-inch computer screen with a refresh rate of 60 Hz at a viewing distance of 80 cm. The experiment is conducted in a quiet environment without light and electromagnetic interference.
The face image-based RSVP framework for identity authentication consists of two sections: Registration and login, as shown in Figure 1.
During the registration section, subjects in the user group are regarded as legal users, and a specific face image sequence is designed for each subject in the user group. For each trial, there is one face image of the user as the self-face image and nine face images selected randomly from the non-user face image set as the non-self-face images. The duration time for each image is 300 ms. The presentation order of the self and non-self-face images in each trial is randomized to avoid the effect of subject prediction on the next stimulus. This section consists of 20 blocks, and each block consists of 10 trials (for trials in the same block, the face images are the same but in random orders; for trials in different blocks, the self-face image is the same, while the non-self-face images are different among blocks), as shown in Figure 2. A short rest comes after 10 blocks. Subjects are asked to focus on the specific face image sequence, while their EEG signals are collected synchronously and an authentication model for each specific user is generated using these signals.
During the login section, to login into a legal user’s account, subjects should watch the specific face image sequence corresponding to the user, while their EEG signals are collected.

2.4. Electrophysiological Recording

The subjects were seated individually in an electrically shielded, dimly lit room. EEG data were measured when the subjects were watching face image sequences during registration or login section. EEG data were recorded using a g.USBamp amplifier with 16 wet active electrodes positioned according to the International 10–20 system and digitized with the sample rate of 2400 Hz. The 16 channels used were as follows: Fz, Cz, P3, Pz, P4, Po7, Oz, Po8, C3, C4, F3, F4, Af7, Af8, Cp5, and Cp6. During the entire experimental task, the impedances of electrodes was kept below 5 kΩ.
The entire EEG dataset acquisition includes two sessions. In session 1, EEG datasets are obtained to generate and test the capability of the identity authentication model. In session 2, EEG datasets are acquired to test the permanence capability of the model.
During session 1, the 15 subjects in the user group are considered as the legal users; hence, specific face image-based RSVP sequences are designed for each user and shown to them. Users were told to focus on their own face images and count the number of occurrences in their minds. To generate and test the performance of the authentication model for each user, two different fraud scenarios are considered. In the first scenario (blind invasion), a subject is regarded as an imposter when he does not know the invasion account; hence, the subject will watch the specific face image sequence without knowing which face image is that of the legal user. In the second scenario (non-blind invasion), a subject is told to invade a certain user’s account. Thus, when the specific face image sequence is presented, the subject will know which face image is of the legal user and will try to invade using the same strategy of the user. For each legal user, we collected two imposters’ EEG data, one for blind invasion, the other for non-blind invasion.
During session 2, to evaluate the permanence of the model, EEG datasets are acquired for each subject in the user group, with an average time interval of 30 days from the first acquisition. The permanence test is performed on the classifiers generated from session 1.

2.5. Data Processing

2.5.1. EEG Data Preprocessing

The raw EEG data were re-referenced using REST [22,23], and were then filtered by a low-pass Chebyshev digital filter with a passband of 40 Hz and a stopband of 49 Hz for further analysis. The data were downsampled from 2400 Hz to 600 Hz by averaging four consecutive samples. Epochs of 200 ms before to 1000 ms after each user’s face image onset were extracted and were baseline corrected by subtracting the averaged amplitude of a small data segment (200 ms to 0 ms) before stimulus onset from each data point.

2.5.2. Classification with Hierarchical Discriminant Component Analysis (HDCA)

There are significant differences between the EEG signals induced by self and non-self-face stimulus, as shown in our previous studies [24,25]. Hence, it is feasible to authenticate a person’s identity by determining whether the EEG signals are induced by self-face image or not. However, the EEG data always contain a certain degree of external noise, and their spatial distribution, amplitudes and latencies may vary with subjects.
Hierarchical discriminant components analysis (HDCA) [18,19] is a classic method using both spatial and temporal information of the EEG signals for classification. In this work, HDCA is used to classify the specific EEG signals evoked by self-face and non-self-face image, which extracts both spatial and temporal features of the signals. The details of the algorithm are as follows:
(1) Spatial features extraction:
It is assumed that the spatial distribution of the EEG activities changes over time with a temporal resolution (T) of 100 ms. Hence, weight vectors, w n , i , are needed to be found for several 100 ms windows following each face image presentation. Here, Fisher linear discriminant (FLD) analysis is used to calculate the spatial coefficients, w n , i . First, each channel of the EEG signals is divided into N segments on average by the given time window (100 ms). Then, the weight for each channel is calculated in each time window to maximize the difference between EEG signals evoked by self-face and non-self-face images, using the Fisher linear discriminant analysis. Therefore, the multichannel EEG signals are compressed into a single channel signal, namely:
y n = i w n , i x i , n        i = 1 , 2 , 3 16
where i and n denotes the number of channels and EEG segments, respectively; x i , n and w n , i represent the i-th channel EEG signal in n-th segment and its weights; and y n is the desirable single channel signal.
(2) Temporal feature extraction:
The obtained single channel signal y n for each of the separated time windows are first average by time point. Then, the resulting values for each of the separate time windows y k are then combined in a weighted average to provide a final interest score Y S for each image, namely:
Y S = k v k y k
The weight v k of y k are calculated to make the self-face score higher than the non-self-face score by using the logistic regression method.

2.5.3. Channel Selection with Genetic Algorithm (GA)

Feature selection and channel optimization is an essential step for identity authentication based on EEG. Actually, models built based on optimized channels for each user may not only help improve the accuracy but also the robustness for identity authentication.
Genetic Algorithm (GA) is an adaptive probabilistic search algorithm for global optimization, inspired by the laws of natural selection and genetics. GA follows the natural evolutionary model, and starts with an initial population of individuals, which consists of a fixed length continuous or discrete strings analogous to the chromosomes in a DNA. Each individual represents a possible solution to a given optimization problem and over successive generations evolves toward a set of more optimal or fit individuals. In this study, GA is used to optimize the channels for identity authentication. Channel is regarded as a unit, which means that features from a certain channel are taken as a whole to be reserved or removed together. The details of the algorithms are as follows:
(1) Initial population generation
Initial population of individuals is generated randomly as a set of binary sequences. Each sequence is of length 16, in which each binary bit denotes if the corresponding channel is chosen or not. Bit ‘1’ means chosen, while bit ‘0’ means not chosen. The size of the initial individuals is set to be 100.
(2) Fitness evaluation
Each individual in the population is assigned a fitness value based on a fitness function, which evaluates how good a solution is to the problem. In our work, fitness is evaluated on the basis of the classification accuracy by HDCA.
(3) Judge with optimization criteria
Optimization criteria identifies if the pre-set generation is reached, or the fitness value is no longer rising. If the criteria are satisfied, then the algorithm ends; if not, a new generation of individual is produced and the algorithm goes to step 2. The evolution process consists of three basic operations: selection, crossover, and mutation. The probability of crossover operation is set to 0.85, while the probability of mutation operation is set to 0.1.

3. Results

3.1. Face Image Evoked ERPs Anaylses

ERP analyses were used to illustrate the variability of the neural responses induced by the face image sequences. EEG data were processed and analyzed using EEGLAB. Artifact free data were epoched—200 to 1000 ms around the users’ face images. Hence, for each subject, there are 200 available trials. For each subject in the user group, the EEG data are induced by self-face images, while for their two corresponding imposters, the EEG data are induced by non-self-face images.
In order to assess brain activities evoked by self-face and non-self-face images, averaging across all trials in each condition were done. Figure 3 shows the grand ERPs evoked by self-face and non-self-face in two scenarios at electrode P3. The red dotted line shows the averaged ERP across all trials and users, when their own face images were presented. The black dotted and the blue solid lines are the averaged ERP evoked by non-self-face in two different scenarios. The blue solid line shows the averaged ERPs across all trials and imposters who do not know the users. The black dotted line shows the averaged ERPs across all trials and imposters who are familiar to the users. This figure shows that the ERPs (N170, N250, P3a, and P3b) evoked by the same face image are different under the three conditions. The amplitude of N250 for the users are little larger than for the imposters. The difference of P3a and P3b among the three conditions is the largest. The amplitude of P3a and P3b for the users are much larger than those for the imposters, especially for the imposters in Scenario 1.
In addition, the topographic maps under the three conditions are compared, as shown in Figure 4. Significant difference can be found for self-face and non-self-face in Scenario 1 at P3a and P3b. Although the two maps of self-face and non-self-face in Scenario 2 are similar at P3a and P3b, their strengths are different. More strong signals are evoked for users when their own face images are presented.

3.2. Identity Authenication Performance

There were 45 subjects in total, divided into two groups (user group and imposter group). In the user group, there were 15 subjects. For each user, there were two corresponding imposters (one had the knowledge of who he intends to invade, the other did not). For these three subjects, artifact free and baseline corrected data were epoched 0 to 1000 ms around the user’s face image onset. There were 200 trials data for each subject, which is 16 × 600 × 200 (16 channels, 600 time points, 200 trials). To obtain a better signal-to-noise ratio, the data was averaged across each two adjacent trials. Hence, for each subject, the sample data is 16 × 600 × 100 (16 channels, 600 time points, 100 trials). There were 45 samples, which included 15 samples for self-face images and 30 samples for non-self-face images.
Classification accuracy (ACC), false acceptance rate (FAR), and false rejection rate (FRR) was used to evaluate the performance of the system for each user, which are defined as follows:
ACC = number   of   correctly   authenticated   samples total   number   of   test   samples
FAR = number   of   falsely   accepted   samples total   number   of   imposter   test   samples
FRR = number   of   falsely   rejected   samples total   number   of   user   test   samples
Authentication performance is explored by performing two classification tests. In the first test, a 5-fold cross validation is used to determine the authentication performance for all classification methods applied in this study. The 45 samples collected in session 1 are used in this test. The second test is the permanence test: for each user, the EEG data collected in session 2 are tested on the classifiers generated from session 1. Thus, only classification accuracy (ACC) is adopted to evaluate the performance.
The classification results by HDCA and GA optimized HDCA (GA-HDCA) are shown in Table 1. It can be seen from the results that the GA-optimized HDCA algorithm had a better performance in ACC, FAR and FRR than the traditional HDCA algorithm. The average classification accuracy (ACC) is improved from 89.16% to 94.26%, whereas the false acceptance rate (FAR) and the false rejection rate (FRR) decrease from 10.97% to 6.27%, and 10.77% to 5.26%, respectively.
In order to test the permanence of the proposed GA-HDCA algorithm, a second EEG data acquisition session was conducted for each user with a 30-day average time interval. For each user, the EEG data collected in the second session was tested on the classifiers generated from the first session. Table 2 shows the permanence test results for each user using HDCA and GA-HDCA. For a 30-day average time interval, the model generated by the proposed GA-HDCA algorithm can still achieve an averaged accuracy of 88.88%, better than HDCA. The results demonstrate the stability of our method, which is essential for practical applications of the EEG-based identity authentication system.

3.3. Optimized Channels for Authentication

Channel optimization for identity authentication mainly serves two purposes. One is to enhance practicality. Using fewer channels in EEG-based identity authentication system will greatly reduce preparation time and make the system more convenient. The other is to improve authentication performance. Since there are significant differences among individuals’ EEG data, even under the same condition, selecting specific channels for each user will not only improve classification accuracy, but also make the model specific for the user and more robust to resist invasion. Table 3 shows the optimized channels selected by the proposed GA-HDCA for the 15 users. “0” denotes that the corresponding channel is not chosen, while “1” denotes that the corresponding channel has been chosen. It can be seen that the selected channels are different among users. The selected times of each channel are also calculated. The most selected electrodes are “P4”, “Pz”, and “P3”, where the “P4” electrode is selected 14 times out of 15 users.

4. Discussion

Evoking significant and stable individual specific features is essential for identity authentication. The face is a traditional biometric trait that can well represent a person’s identity, and EEG signals can reflect a person’s brain activities. Hence, a face image-based RSVP paradigm for identity authentication is proposed, which combines the face and EEG, the two kinds of biometric traits. Experimental results have shown that EEG signals are different when a person watches his own face and another person’s face. Although the self-face image and non-self-face image can both evoke N170, N250, and P300 components, their amplitude, latency and spatial distribution are different between the two categories, especially for the P300 component. Even for an imposter who is familiar with the user he intends to invade, the amplitude and the latency of the P300 component evoked by the non-self-face is lower and longer than that of the self-face. The lower amplitude may indicate that non-self-face is less sensitive for a person than self-face, while the longer latency may reflect more complicated processing when dealing with non-self-face than self-face.
Compared to the meaningful related work accomplished by Yeom et al. [15], in our study, RSVP paradigm is induced into an authentication framework, which enables us to present a stimulus in a very short time (300 ms). Averaged authentication accuracy of 94.26% within 6 s can be achieved in our study. Both real-time capability and accuracy of the system are significantly improved, compared to Yeom’s work.
On the basis of HDCA, a GA-optimized HDCA (GA-HDCA) method is proposed to build subject-specific models for classifying user and imposter, using the differences between self-face and non-self-face evoked EEG signals. The GA-HDCA tries to find the global optimization solution for selecting the best set of channels in classifying the two category of EEG signals. Results show that the solution is subject-specific. After channel optimization, the overall authentication performance is improved in ACC, FAR and FRR, which means that the optimization can not only reduce redundant information and increase accuracy, but it also makes the model subject-specific to resist invasion. Figure 5 shows the positions of the most related electrodes for our proposed authentication framework based on face image-RSVP. These six electrodes are the most selected ones by GA-HDCA across the 15 users. This result is consistent with our previous study [24,25], where “Cz”, ”Pz”, “P3”, and “P4” were also found to be important for classification. Moreover, these selected electrodes are exactly distributed in the brain regions where the EEG signals evoked by self-face and non-self-face images are most different, especially for the P300 component.
A comparison of our method with previous related works is provided in Table 4. The superiority of our proposed method can be seen from the performance comparison. For our proposed method, the average accuracy of 94.36% is the highest, whereas the FAR of 6.27% and FRR of 5.26 are the second lowest. Moreover, the stability of our method is also tested using data collected with an averaged interval of 30 days. It can still achieve a promising accuracy of 88.88%. However, since the proposed method is attempting to find the global optimized solution, for each individual, the time required for training the model with GA-HDCA is approximately 40 min, 10 times longer than that of HDCA. But once the model has been built, it is ready for use. Thus, this time requirement will only affect the registration, not the login section. Hence, this does not affect the proposed method for practical use. Furthermore, in order to make our system applicable to real-time scenarios, the authentication model for each user is trained using data of two adjacent trials. This means that during the login section, the model only needs the data of two trials to determine whether the person is a legal user or an imposter. Since the duration time for each trial is 3 s, the time required for one authentication is only 6 s in our system, which shows better real-time performance than previous studies.

5. Conclusions

In this paper, a face image-based RSVP paradigm is designed for identity authentication, which combines face and EEG, the two kinds of biometric traits, to evoke more specific and stable traits for authentication. Significant differences are found for the ERP components and topographic maps induced by self-face and non-self-face (familiar and not familiar). On the basis of this, an authentication method based on HDCA and GA is proposed to build subject-specific models with optimized fewer channels. The averaged authentication accuracy of 94.26% in 6 s with averaged 6 channels is achieved by our proposed method. For a 30-day averaged time interval, our method can still reach the averaged accuracy of 88.88%. Hence, the experimental results show the effectiveness, robustness, and stability of the proposed framework for EEG-based identity authentication. In future work, the experiment will be repeated with a longer interval time to further explore the system’s stability. Portable EEG acquisition equipment, such as the Emotiv EPOC headset, will be used to improve system practicability. Moreover, the open set ability of our system will be tested with random people who are not used in the training phase.

Author Contributions

Conceptualization, Y.Z. and D.Y.; Methodology, Y.Z., Q.W. and B.Y.; Software, Q.W.; Validation, K.Y., L.T. and Y.Z.; Investigation, K.Y. and Y.Z.; Resources, J.S.; Data Curation, J.S.; Writing-Original Draft Preparation, Y.Z.; Writing-Review & Editing, Y.Z.; Visualization, L.T.; Supervision, D.Y.; Project Administration, B.Y.; Funding Acquisition, B.Y.

Funding

This research was funded by the National Key R&D Program of China under grant 2017YFB1002502, the National Natural Science Foundation of China (No. 61701089, No. 61601518 and No. 61372172) and the National Defense Science and Technology Innovation Zone Project.

Acknowledgments

The authors would like to thank all participants who provide rich data support for our approach.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Campisi, P.; La Rocca, D. Brain waves for automatic biometric-based user recognition. IEEE Trans. Inf. Forensics Secur. 2014, 9, 782–800. [Google Scholar] [CrossRef]
  2. Cavanagh, J.F.; Allen, J.J. Multiple aspects of the stress response under social evaluative threat: An electrophysiological investigation. Psychoneuroendocrinology 2008, 33, 41–53. [Google Scholar] [CrossRef]
  3. Abo-Zahhad, M.; Ahmed, S.M.; Abbas, S.N. State-of-the-art methods and future perspectives for personal recognition based on electroencephalogram signals. IET Biom. 2015, 4, 179–190. [Google Scholar] [CrossRef]
  4. Palaniappan, R.; Mandic, D.P. Biometrics from brain electrical activity: A machine learning approach. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 738–742. [Google Scholar] [CrossRef] [PubMed]
  5. Bassett, D.S.; Gazzaniga, M.S. Understanding complexity in the human brain. Trends Cogn. Sci. 2011, 15, 200–209. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Pozo-Banos, M.D.; Alonso, J.B.; Ticay-Rivas, J.R.; Travieso, C.M. Electroencephalogram subject identification: A review. Expert Syst. Appl. 2014, 41, 6537–6554. [Google Scholar] [CrossRef]
  7. Poulos, M.; Rangoussi, M.; Chrissikopoulos, V.; Evangelou, A. Person identification based on parametric processing of the EEG. In Proceedings of the 6th IEEE International Conference on Electronics, Circuits and Systems, Pafos, Cyprus, 5–8 September 1999; Volume 281, pp. 283–286. [Google Scholar]
  8. Ravi, K.V.R.; Palaniappan, R. Leave-one-out authentication of persons using 40 hz EEG oscillations. In Proceedings of the International Conference on Computer as a Tool, Belgrade, Serbia, 21–24 November 2005; pp. 1386–1389. [Google Scholar]
  9. Snodgrass, J.G.; Vanderwart, M. A standardized set of 260 pictures: Norms for name agreement, image agreement, familiarity, and visual complexity. J. Exp. Psychol. Hum. Learn. Mem. 1980, 6, 174–215. [Google Scholar] [CrossRef]
  10. Shiliang, S. Multitask learning for EEG-based biometrics. In Proceedings of the 2008 19th International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008; pp. 1–4. [Google Scholar]
  11. Abo-Zahhad, M.; Ahmed, S.M.; Abbas, S.N. A new multi-level approach to EEG based human authentication using eye blinking. Pattern Recognit. Lett. 2016, 82, 216–225. [Google Scholar] [CrossRef]
  12. Weng, R.; Lu, J.; Tan, Y.P. Robust point set matching for partial face recognition. IEEE Trans. Image Process. 2016, 25, 1163–1176. [Google Scholar] [CrossRef]
  13. Miyakoshi, M.; Kanayama, N.; Iidaka, T.; Ohira, H. EEG evidence of face-specific visual self-representation. Neuroimage 2010, 50, 1666–1675. [Google Scholar] [CrossRef]
  14. Sugiura, M.; Sassa, Y.; Jeong, H.; Horie, K.; Sato, S.; Kawashima, R. Face-specific and domain-general characteristics of cortical responses during self-recognition. Neuroimage 2008, 42, 414–422. [Google Scholar] [CrossRef] [PubMed]
  15. Yeom, S.K.; Suk, H.I.; Lee, S.W. Person authentication from neural activity of face-specific visual self-representation. Pattern Recognit. 2013, 46, 1159–1169. [Google Scholar] [CrossRef]
  16. Potter, M.C. Rapid serial visual presentation (RSVP): A method for studying language processing. New Methods Read. Compr. Res. 1984, 118, 91–118. [Google Scholar]
  17. Acqualagna, L.; Blankertz, B. Gaze-independent BCI-spelling using rapid serial visual presentation (RSVP). Clin. Neurophysiol. 2013, 124, 901–908. [Google Scholar] [CrossRef] [PubMed]
  18. Gerson, A.D.; Parra, L.C.; Sajda, P. Cortically coupled computer vision for rapid image search. IEEE Trans. Neural Syst. Rehabil. Eng. 2006, 14, 174–179. [Google Scholar] [CrossRef] [PubMed]
  19. Pohlmeyer, E.A.; Wang, J.; Jangraw, D.C.; Lou, B.; Chang, S.F.; Sajda, P. Closing the loop in cortically-coupled computer vision: A brain-computer interface for searching image databases. J. Neural Eng. 2011, 8, 036025. [Google Scholar] [CrossRef] [PubMed]
  20. Goldberg, D.E. Genetic Algorithms in Search, Optimization, and Machine Learning; Addison-Wesley Pub. Co.: Boston, MA, USA, 1989. [Google Scholar]
  21. Yang, J.; Honavar, V. Feature subset selection using a genetic algorithm. IEEE Comput. Soc. 2002, 13, 44–49. [Google Scholar]
  22. Yao, D. A method to standardize a reference of scalp EEG recordings to a point at infinity. Physiol. Meas. 2001, 22, 693–711. [Google Scholar] [CrossRef]
  23. Zhai, Y.; Yao, D. A study on the reference electrode standardization technique for a realistic head model. Comput. Methods Progr. 2004, 76, 229–238. [Google Scholar] [CrossRef]
  24. Wu, Q.J.; Zeng, Y.; Zhang, C.; Tong, L.; Yan, B. An EEG-based person authentication system with open-set capability combining eye blinking signals. Sensor 2018, 18, 335. [Google Scholar] [CrossRef]
  25. Wu, Q.J.; Yan, B.; Zeng, Y.; Zhang, C.; Tong, L. Anti-deception: Reliable EEG-based biometrics with real-time capability from the neural response of face rapid serial visual prestation. Biomed. Eng. Online 2018, 17, 55. [Google Scholar] [CrossRef] [PubMed]
  26. Armstrong, B.C.; Ruiz-Blondet, M.V.; Khalifian, N.; Kurtz, K.J.; Jin, Z.; Laszlo, S. Brainprint: Assessing the uniqueness, collectability, and permanence of a novel method for ERP biometrics. Neurocomputing 2015, 166, 59–67. [Google Scholar] [CrossRef]
  27. Marcel, S.; Millan, J.D.R. Person Authentication Using Brainwaves (EEG) and Maximum A Posteriori Model Adaptation. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 743–752. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Miyamoto, C.; Baba, S.; Nakanishi, I. Biometric person authentication using new spectral features of electroencephalogram (EEG). In Proceedings of the 2008 International Symposium on Intelligent Signal Processing and Communications Systems, Bangkok, Thailand, 8–11 February 2009; pp. 1–4. [Google Scholar]
  29. Mu, Z.; Hu, J.; Min, J. EEG-Based Person Authentication Using a Fuzzy Entropy-Related Approach with Two Electrodes. Entropy 2016, 18, 432. [Google Scholar] [CrossRef]
Figure 1. The face image-based RSVP framework for identity authentication.
Figure 1. The face image-based RSVP framework for identity authentication.
Sensors 19 00006 g001
Figure 2. The self and non-self-face RSVP paradigm for identity authentication.
Figure 2. The self and non-self-face RSVP paradigm for identity authentication.
Sensors 19 00006 g002
Figure 3. The grand ERPs evoked by self-face and non-self-face at electrode P4.
Figure 3. The grand ERPs evoked by self-face and non-self-face at electrode P4.
Sensors 19 00006 g003
Figure 4. The topographic maps of EEG data for self-face and non-self-face.
Figure 4. The topographic maps of EEG data for self-face and non-self-face.
Sensors 19 00006 g004
Figure 5. The most related electrodes for face image-based RSVP paradigm.
Figure 5. The most related electrodes for face image-based RSVP paradigm.
Sensors 19 00006 g005
Table 1. Authentication performances with HDCA and GA-HDCA.
Table 1. Authentication performances with HDCA and GA-HDCA.
UserACC (%)FAR (%)FRR (%)
HDCAGA-HDCAHDCAGA-HDCAHDCAGA-HDCA
182.592.821.59.513.55.0
292.395.005.52.510.07.5
386.392.013.011.014.55.0
493.897.505.53.57.01.5
586.893.313.06.013.57.5
689.096.312.03.510.04.0
785.893.011.55.517.08.5
891.097.312.55.05.50.5
991.594.86.04.011.06.5
1085.093.316.58.013.55.5
1193.097.06.01.58.04.5
1292.892.86.09.08.55.5
1391.595.89.54.07.54.5
1485.892.019.514.09.02.0
1590.391.06.57.013.011.0
Mean (std)89.16 (3.52)94.26 (2.12)10.97 (5.22)6.27 (3.47)10.77 (3.27)5.26 (2.74)
Table 2. The classification accuracy of each user with a 30-day averaged time interval.
Table 2. The classification accuracy of each user with a 30-day averaged time interval.
UserACC (%)ACC (%)
HDCAGA-HDCA
180.286.9
272.980.7
388.585.1
489.595.6
586.985.8
684.992.2
786.883.1
891.994.0
978.880.4
1084.689.9
1194.093.0
1288.191.4
1382.286.2
1494.196.6
1589.792.3
Mean (std)86.21(5.83)88.88(5.25)
Table 3. The optimized channels for all the 15 users.
Table 3. The optimized channels for all the 15 users.
UserChannelsNumber of Selected Channels
FzCzP3PzP4Po7OzPo8C3C4F3F4Af7Af8Cp5Cp6
100111001000000004
200011010010000015
300111001000000004
400111001000000105
501101001000000105
600111000000000003
701010000100000014
800111000110000117
901001001010000015
1001111010100000107
1101001001010101006
1211101111000010019
1300011011110000118
1401111010100000017
15001110111011011111
Table 4. Performance comparison with previous related works.
Table 4. Performance comparison with previous related works.
AuthorStimulus TypeTime Required (s)Imposter ScenariosStability TestACC (%)FAR (%)FRR (%)
Armstrong et al. [26]Text readingNANoneYes89NANA
Yeom et al. [15]Self-or non-self-face images31.5~41NoneNone86.113.913.9
Marcel et al. [27]]Motor imagery15NoneNone80.714.424.3
Miyamoto et al. [28]Resting state60NoneNone79.021.021.0
Mu et al. [29]Self- and non-self-photos6.5NoneNone87.35.55.6
Wu et al. [25]Face RSVP62 scenariosYes91.469.237.85
Proposed methodFace RSVP62 scenariosYes94.266.275.26

Share and Cite

MDPI and ACS Style

Zeng, Y.; Wu, Q.; Yang, K.; Tong, L.; Yan, B.; Shu, J.; Yao, D. EEG-Based Identity Authentication Framework Using Face Rapid Serial Visual Presentation with Optimized Channels. Sensors 2019, 19, 6. https://doi.org/10.3390/s19010006

AMA Style

Zeng Y, Wu Q, Yang K, Tong L, Yan B, Shu J, Yao D. EEG-Based Identity Authentication Framework Using Face Rapid Serial Visual Presentation with Optimized Channels. Sensors. 2019; 19(1):6. https://doi.org/10.3390/s19010006

Chicago/Turabian Style

Zeng, Ying, Qunjian Wu, Kai Yang, Li Tong, Bin Yan, Jun Shu, and Dezhong Yao. 2019. "EEG-Based Identity Authentication Framework Using Face Rapid Serial Visual Presentation with Optimized Channels" Sensors 19, no. 1: 6. https://doi.org/10.3390/s19010006

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop