Next Article in Journal
Diversity of Hemodynamic Reactive Profiles across Persons—Psychosocial Implications for Personalized Medicine
Previous Article in Journal
Intraoperative Anterior Segment Optical Coherence Tomography in the Management of Cataract Surgery: State of the Art
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Full Informational Spectral Analysis for Auditory Steady-State Responses in Human Brain Using the Combination of Canonical Correlation Analysis and Holo-Hilbert Spectral Analysis

1
Department of Electrical Engineering, National Central University, Taoyuan 320, Taiwan
2
Department of Rehabilitation, Taoyuan General Hospital, Taoyuan 330, Taiwan
3
CWLab International, Thousand Oaks, CA 91320, USA
4
The Pavlov Institute of Physiology, Russian Academy of Sciences, 199034 St. Petersburg, Russia
5
Division of Cardiovascular Surgery, Taipei Veterans General Hospital, Taipei 112, Taiwan
6
Department of Surgery, School of Medicine, Taipei Medical University, Taipei 106, Taiwan
*
Authors to whom correspondence should be addressed.
J. Clin. Med. 2022, 11(13), 3868; https://doi.org/10.3390/jcm11133868
Submission received: 6 April 2022 / Revised: 19 June 2022 / Accepted: 24 June 2022 / Published: 4 July 2022

Abstract

:
Auditory steady-state response (ASSR) is a translational biomarker for several neurological and psychiatric disorders, such as hearing loss, schizophrenia, bipolar disorder, autism, etc. The ASSR is sinusoidal electroencephalography (EEG)/magnetoencephalography (MEG) responses induced by periodically presented auditory stimuli. Traditional frequency analysis assumes ASSR is a stationary response, which can be analyzed using linear analysis approaches, such as Fourier analysis or Wavelet. However, recent studies have reported that the human steady-state responses are dynamic and can be modulated by the subject’s attention, wakefulness state, mental load, and mental fatigue. The amplitude modulations on the measured oscillatory responses can result in the spectral broadening or frequency splitting on the Fourier spectrum, owing to the trigonometric product-to-sum formula. Accordingly, in this study, we analyzed the human ASSR by the combination of canonical correlation analysis (CCA) and Holo-Hilbert spectral analysis (HHSA). The CCA was used to extract ASSR-related signal features, and the HHSA was used to decompose the extracted ASSR responses into amplitude modulation (AM) components and frequency modulation (FM) components, in which the FM frequency represents the fast-changing intra-mode frequency and the AM frequency represents the slow-changing inter-mode frequency. In this paper, we aimed to study the AM and FM spectra of ASSR responses in a 37 Hz steady-state auditory stimulation. Twenty-five healthy subjects were recruited for this study, and each subject was requested to participate in two auditory stimulation sessions, including one right-ear and one left-ear monaural steady-state auditory stimulation. With the HHSA, both the 37 Hz (fundamental frequency) and the 74 Hz (first harmonic frequency) auditory responses were successfully extracted. Examining the AM spectra, the 37 Hz and the 74 Hz auditory responses were modulated by distinct AM spectra, each with at least three composite frequencies. In contrast to the results of traditional Fourier spectra, frequency splitting was seen at 37 Hz, and a spectral peak was obscured at 74 Hz in Fourier spectra. The proposed method effectively corrects the frequency splitting problem resulting from time-varying amplitude changes. Our results have validated the HHSA as a useful tool for steady-state response (SSR) studies so that the misleading or wrong interpretation caused by amplitude modulation in the traditional Fourier spectrum can be avoided.

1. Introduction

Steady-state response (SSR) elicited by periodic sensory stimuli is a synchronized and phase-locked type of neural activity in the human brain. The SSR has been extensively used to investigate the electrophysiological responses underlying different neural networks [1]. Different stimulus parameters, such as stimulus frequency [2], duty cycle [3], intensity contrast [4], phase synchrony [5], attention modulation [6], etc., have been studied to characterize their functional capacities and neural mechanisms. Classical physiological models assume SSR as a stereotypical time-invariant neural response, in which background fluctuations are linearly added to the elicited SSR. A time-averaging technique is usually chosen to suppress the signal fluctuations that are uncorrelated to the given periodic stimulus. However, recent studies indicate SSR contains nonlinear neural responses, which could be modulated by the subject’s attention [7], wakefulness state [8,9], mental load [10,11,12,13], and mental fatigue [14]. Traditional frequency analysis methods, such as Fourier transform or wavelet transform, assume the measured time series are the linear combinations of vectors from a predetermined basis, in which the modulation among different signal scales is not considered. Nevertheless, it has been demonstrated that neural adaption can alter the strategy of neural coding and cause modulation in the phase, latency, or amplitude in evoked potentials [15]. Several studies have shown neural coding plays an important role in the modulation of oscillatory activities within cortical neuron populations [16]. The neural coding describes that the information transmitted from sensory perception to the central nervous system can be encoded by varying the rate and timing of action potentials (spikes), depending on the properties of the perceived stimulus and the subject’s physical states [17]. For example, Zhong et al. (2013) investigated the effects of noise-induced hearing loss on the coding of an envelope structure in the central auditory system, and they observed increased auditory responses at the applied modulation frequency in people with cochlear hearing loss [18]. Nasir et al. (2013) tested the plasticity in sensory systems using somatosensory evoked potentials (SEPs) and observed that sensorimotor adaption induced alternation in the neural coding of somatosensory stimuli [19]. Several neural coding hypotheses have been proposed to account for the processing of neural activity patterns in our brain and nervous systems, such as rate coding, temporal coding, population coding, correlation coding, independent spike coding, position coding, sparse coding, etc. [20]. Among the neural coding hypotheses, rate coding and temporal coding are the two neural coding strategies that were studied the most in previous literature. The rate coding model, sometimes called the frequency coding model, hypothesizes that the information of a signal is contained in the firing rate of neural spikes, while the temporal coding model assumes the information of interneuron transmission is contained in the precise timing of spikes and inter-spike intervals. Aghababaiyan et al. (2019) studied the capacity and error probability analysis of neuro-spike communication using a numerical simulation approach and concluded temporal coding has a higher efficiency than spike rate coding in terms of achievable data rate [21]. The neural coding could also be influenced by a subject’s physiological states, such as attention [22] and emotion states [23], in order to achieve better efficiency of information transmission [24]. Since the exertion of a modulation frequency on a sinusoidal signal will result in a frequency shifting away from its oscillatory frequency according to the trigonometric product-to-sum formula, neural modulation, induced by changes in internal statuses or external environments [25], could result in deviation in the detected oscillatory frequency of human SSR. The development of a frequency analysis method, which takes both the oscillatory and modulatory frequencies into account, is necessary to study the subtle dynamics in the measured SSR.
Auditory steady-state responses (ASSR) elicited by rapid auditory stimuli are a kind of SSR commonly used in clinics. ASSR is a sinusoidal-like response consisting of evoked responses from the central auditory pathway and auditory cortex, which allows the measurement of hearing response at a specified frequency. The modulation technique for ASSR stimulus can be either amplitude modulation [26,27,28] or frequency modulation [29]. It has been demonstrated that ASSR is a useful translational biomarker for several neurological and psychiatric disorders [30,31,32,33]. The ASSR is also an objective tool for estimating hearing sensitivity in individuals, especially for hearing loss [34,35], hearing assessment [36,37], and evaluation of aural rehabilitation [38]. Khuwaja et al. (2015) measured the ASSR responses during different stages. They obtained a 100% classification rate for sleep stage classification using a neural network classifier [39]. Swanepoel (2011) compared ASSR and brain stem response (ABR) in infants and concluded ASSR is a more reliable and useful tool for diagnosing hearing loss in infants [40]. Sugiyama et al. (2021) reviewed the ASSR processing in psychiatric disorders, including schizophrenia, bipolar disorder, and autism. The ASSR amplitude is suppressed within the gamma band (~40 Hz), which indicates the imbalance between GABAergic and N-methyl-D-aspartate (NMDA) receptor-medicated neurotransmission [41]. Yokota and Naruse (2015) found that the phase coherence of ASSR can reflect the amount of cognitive workload in an N-back task [42]. Kallenberg et al. (2007) implemented an ASSR-based brain–computer interface (BCI) by giving two tones with distinct modulation frequencies. Subjects were requested to pay attention to one out of the two tones to achieve the auditory BCI control [43].
Most ASSR studies utilized tone bursts modulated at 40 Hz, in which the test frequencies are often chosen at 500, 1000, 2000, and 4000 Hz with a 100% modulation depth [38]. Physicians adjust the intensities of ASSR auditory stimuli from high to low to objectively determine the subject’s hearing threshold. In most of the aforementioned studies, the ASSR were detected using ensemble average or Fourier-based spectral analysis. Both the techniques assume the auditory-induced responses are stationary across the whole measurement data. Temporal modulation in the recorded auditory response is usually neglected, which could result in the information loss for studying ASSR neural mechanisms.
In order to preserve the dynamic information of ASSR, several signal processing approaches have been proposed to suppress ASSR-unrelated noise in the EEG recordings. Biesmans et al. (2015) filtered the EEG signals within a narrow band and designed a spatial filter to enhance the neural activities by optimizing the SNR of recorded ASSR [44]. Janani et al. (2018) utilized independent component analysis (ICA) to suppress ASSR-unrelated noise by removing artifact-contaminated components. They proposed a correlation-based algorithm for the component selection process [45]. Wang et al. (2015) applied empirical mode decomposition (EMD) to extract ASSR signals in normal subjects and tinnitus patients [46]. They constructed a spatial map for each intrinsic mode function (IMF), and the ASSR components were identified by comparing the spatial map distribution with a pre-defined spatial template. Bin et al. (2009) [47] applied canonical correlation analysis (CCA) to enhance the steady-state visual evoked potential (SSVEP)-associated frequency information, and they used it as the control signal for brain–computer interface (BCI) control. The aforementioned methods utilized spatial filter or blind source separation (BSS) to achieve noise removal in EEG signals. However, these methods did not analyze the effect of time changes in amplitudes, in which the temporal modulation can lead to frequency splitting or spectrum broadening (trigonometric product-to-sum formula). The frequency shift caused by amplitude modulation will result in ambiguity in clinical ASSR diagnoses.
One spectral analysis tool, the Holo-spectral analysis (HHSA) proposed by Huang et al. (2016) [48], was proposed to analyze both the amplitude modulation (AM) and the carrier frequency (frequency modulation, FM) information in a measured signal. The HHSA was constructed based on the structure of two-layer empirical mode decompositions (EMD), in which the EMD is a signal decomposition method used to extract meaningful oscillation features into intrinsic mode functions (IMF) by means of recursively applying a sifting process [49]. In HHSA, the first-layer EMD decomposes the input signal into frequency modulation (FM) components. In the second layer of HHSA, the signal envelopes of the IMFs obtained from the first-layer EMD are further decomposed into amplitude modulation (AM) components. The HHSA enables a two-dimensional AM-FM representation of the input signal, in which the FM frequency represents the fast-changing intra-mode frequency, and the AM frequency represents the slow-changing inter-mode frequency. In this study, we propose a method by combining the CCA and HHSA for the spectral analysis of ASSR. The CCA was applied as a signal preprocessing approach by selecting ASSR-related canonical vectors. The ASSR-related canonical vectors were chosen for reconstructing noise-suppressed channel signals. The source activities at the left auditory cortex (LAC) and right auditory cortexes (RAC) were estimated, and the HHSA was applied to give the AM-FM representation of the measured ASSR. In our analysis, we adopted CCA as preprocessing for noise removal and applied the HHSA for AM-FM analysis. The study results might shed light on exploring the subtle information processing underlying the auditory neuronal circuitry.

2. Materials and Methods

2.1. Auditory Stimulation

Our steady-state auditory stimuli utilized a 1 kHz sinusoidal wave modulated at 37 Hz with 100% modulation depth [50]. Monaural auditory stimuli were presented to the left and right ears of each participant in separate sessions. Each session was three minutes in duration, and triggers were generated at the beginning of each second for subsequent signal processing. The intensities of auditory stimuli were set to 70 dB sound pressure level (SPL), generated by an analog-to-digital conversion card (D/A) conversion card (NI USB-6259, National Instrument, Austin, TX, USA), which was precisely controlled by the LabView software (National Instruments, Austin, TX, USA).

2.2. Subjects and Tasks

Twenty-five normal subjects (14 males and 11 females, all right-handed subjects; mean ages = 31 ± 4.2 years) were recruited to participate in this experiment. Subjects were requested to sit in a comfortable armchair in a dimly illuminated electromagnetic shielded room. All participants were requested to participate in two auditory stimulation sessions, including one right-ear and one left-ear monaural steady-state auditory stimulations (one for the right ear and one for the left ear) for three minutes. A three-minute empty-room measurement was also recorded for each participant to monitor the background noise. All participants gave informed consent, and the study was approved by the Ethics Committee of the Institutional Review Board (IRB), Tao-Yuan General Hospital, Taiwan. The measurements were noninvasive, and the subjects were free to withdraw at any time.

2.3. Electroencephalography Recordings

The electroencephalography (EEG) signals were recorded using a 32-channel whole-head EEG system (band-pass, 0.05 Hz–250 Hz; digitized at 1 kHz; QuickAmp, Brain Products Co., Munich, Germany). A 46 Hz–65 Hz bandstop filter (3rd-order Butterworth IIR filter) was also applied to filter out electricity noises, in which the 46 and 65 Hz cut-off frequencies were chosen by setting 9 Hz frequency margins for the ASSR-related 37 Hz (fundamental stimulation frequency) and 74 Hz (the first harmonic frequency) frequencies, respectively. Electrodes were placed in accordance with the international 10–20 channel placement. The exact position of the head with respect to the EEG electrodes was determined by measuring magnetic signals from four head position indicator (HPI) coils placed on the scalp. Coil positions were identified with a three-dimensional digitizer (Fastrack system, Polhemus, Colchester, CT, USA) with respect to three predetermined landmarks (naison and bilateral preauricular points) on the scalp, and these data were used to superimpose EEG source signals on individual MRI images, obtained with a 3.0 T Bruker MedSpec S300 system (Bruker, Kalsrube, Germany). The anatomical image was acquired using a high-resolution T1-weighted, 3D gradient-echo pulse sequence (MDEFT: Modified Driven Equilibrium Fourier Transform; TR/TE/TI = 88.1 ms/4.12 ms/650 ms, 128 × 128 × 128 matrix, FOV = 250 mm). The whole-head EEG signals recorded in monaural steady-state auditory stimulations were stored on a hard disk for subsequent off-line CCA and HHSA processing.

2.4. Canonical Correlation Analysis and Selection Pertinent

Since the EEG channels in temporal areas are close to the human auditory cortex, the channels from left or right temporal areas (the F7, FC5, T7, and CP5 in the left auditory area; the F8, FC6, T8, and CP6 in the right auditory area) were chosen as channels-of-interest (COI) for ASSR extractions. We considered the ASSR responses generated from the left auditory area might not be synchronized with the ASSR responses obtained from the right auditory area. The ASSR responses from the left auditory area and right auditory area were extracted by setting the COI at left and right auditory areas separately. Given an M-channel and N time points for EEG measurement XM×N (M = 4, including the F7, FC5, T7, and CP5 in the left auditory area, or the F8, FC6, T8, and CP6 in the right auditory area) and a set of reference signals YK×N (K = 4 in this paper, containing the sinusoidal signals with fundamental ASSR frequency and its first harmonic frequency) with N time points (N = 180,000), the canonical correlation analysis (CCA) tries to find vectors a ( a M ) and b ( b K ) to maximize the correlation ρ = c o r r ( a T X , b T Y ) , in which the u = aTX and v = bTY are defined as one pair of canonical vectors. The CCA seeks to find z (z = min(M,K)) pairs of canonical vectors, and the ith pair of canonical vectors are uncorrelated with the jth pair of canonical vectors (i.e., u i u j and v i v j , for ij). In this study, we took the fundamental frequency and the first harmonic frequency of steady-state auditory stimulus into consideration so that the reference signals Y can be represented as follows:
Y = [ sin ( 2 π × f 0 × t ) cos ( 2 π × f 0 × t ) sin ( 2 π × 2 × f 0 × t ) cos ( 2 π × 2 × f 0 × t ) ] ,
where f0 (f0 = 37 Hz) is the ASSR stimulus frequency and t is the time index.
The correlation value between the ith pair of canonical vectors is:
ρ i = a i T XY b i a i T XX a i b i T YY b i ,
where ρ i is the correlation value between ui and vi, ui = aiTX and vi = biTY are the ith pair of canonical vectors, XX and YY are the auto-correlation matrixes of X and Y, XY is the cross-covariance matrix between X and Y.
The ui (i = 1, …, 4) constructs a signal space obtained from the original EEG recordings around the left and right auditory area so that the canonical vectors are the vector directions which multichannel EEG samples correlate with most. The vi (i = 1, …, 4) are the signal spaces constructed from the reference signals, which are not used in this study because we want to reconstruct the ASSR responses based on the signal features extracted from original recordings, rather than the signal space constructed from the reference signals. Further arranging of the vectors for creating each pair of canonical vectors into matrix A, the four canonical vectors can be obtained by the multiplication of X and A, represented as:
U 4 × N = ( A 4 × 4 ) T ( X 4 × N ) = [ U ( 1 ) U ( 2 ) U ( 3 ) U ( 4 ) ] ,
where U contains the four 1 × N canonical vectors, A = [ a 1 a 2 a 3 a 4 ] contains the matrix to create the four 1 × N canonical vectors, and U(1), U(2), U(3), U(4) are the canonical vectors generated from the EEG recordings in the first, second, third, and fourth pairs of canonical vectors, respectively.
Figure 1 shows the spatial maps and Fourier spectra of the four canonical vectors. The spatial map was created to represent the topographic distribution (correlation) of each canonical vector over the whole-head EEG channels. For the 32-channel whole-head EEG recording E32×N, the correlation of each canonical vector on all EEG channels can be calculated by:
M ( i ) 32 × 1 = A B S { ( E 32 × N ) ( U ( i ) 1 × N / U ( i ) 1 × N ) T } ,
where E32×N contains the 32-channel EEG recording samples with N time points, the U(i) is the ith canonical vector obtained by setting COIs surrounding the left or right auditory area, ‖ ‖ is the function of the L2-norm, ABS{·} is the function of absolute value, and M(i) contains the absolute values of the correlation coefficients between the ith canonical vector and E32×N. In CCA, the relation between each canonical vector and the reference signals is examined by performing a statistical significance analysis. The canonical vectors with p < 0.01 were identified as ASSR-related canonical vectors and were subjected to the reconstruction of noise-suppressed ASSR responses. Figure 1 demonstrates the CCA processing for extracting ASSR responses in subject 5. The left panel (Figure 1A) shows the channel plot of the recorded EEG signals. The middle-upper panel (Figure 1B) and the middle-lower panel (Figure 1C) show the spatial maps and the Fourier spectra of canonical vectors obtained by setting COIs located around left and right temporal areas. In Figure 1B, the canonical vectors U(1) and U(2), whose p values are smaller than 0.01 (p < 0.01) and had clear 37 Hz spectral peaks with their spatial maps concentrated on the right temporal area, were chosen as ASSR-related canonical vectors of the right hemisphere. Similarly, in Figure 1C, the U(1), U(2) and U(3) (p < 0.01) obtained from left COIs, which presented clear 37 Hz spectral peaks with spatial maps concentrated on the left temporal area, were identified as ASSR-related canonical vectors of the left hemisphere. The EEG signals were then projected to the canonical vectors obtained from left and right ROIs, and the signals were reconstructed to suppress ASSR-unrelated noises. The data reconstruction procedure can be represented by the following equation:
E 32 × N r e c o n = { S elected   i   from   canonical   vectors of   left   ROI [ ( E 32 × N ) ( U L ( i ) 1 × N / U L ( i ) 1 × N ) T ] U L ( i ) } + { S elected   i   from   canonical   vectors of   right   ROI [ ( E 32 × N ) ( U R ( i ) 1 × N / U R ( i ) 1 × N ) T ] U R ( i ) } ,
where Erecon is the reconstructed ASSR-related EEG recordings, E is the original EEG recording, and UL and UR are the canonical vectors obtained by setting COIs close to the left and right temporal areas. The ASSR-related EEG signals were reconstructed by examining ASSR-related canonical vectors (p < 0.01) and the original EEG recordings were projected to these ASSR-related canonical vectors to suppress task-unrelated components. It can be noticed that the spatial maps generated from ASSR-related canonical vectors on the left and right hemispheres had the characteristic of unilateral distributions, which indicated the low inter-hemispheric correlation of ASSR responses. Accordingly, in this study, we first calculated the ASSR responses in the left and right hemispheres separately and then summated them together to obtain the whole-head ASSR responses (Figure 1D) for the following ASSR source estimations.

2.5. Analysis of ASSR Source Activities Using Minimum Norm Estimation

To estimate the source activities in the left and right auditory cortexes (LAC and RAC), minimum norm estimation (MNE) with a realistic model generated from an individual’s magnetic resonance image (MRI) (BrainStorm software, University of South California, Los Angeles, CA, USA; http://neuroimage.usc.edu/brainstorm, accessed on 11 December 2021) was adopted [51]. Only the nodes on the cortical surface with source amplitudes survived statistical significance (p < 0.01) among the total surface nodes that were rendered on an individual’s anatomical MRI (see Figure 2). Figure 2A,B shows the source estimation results obtained from the reconstructed EEG signals shown in Figure 1D. The auditory-induced source activities were estimated by means of minimum norm estimation (MNE) (BrainStorm software, University of South California, Los Angeles, USA; http://neuroimage.usc.edu/brainstorm, accessed on 11 December 2021) [51], with a realistic head model generated from an individual’s magnetic resonance image (MRI) using brainVISA software (SHFJ, Orsay, France; http://brainvisa.info/, accessed on 11 December 2021) [52]. The estimated neural sources were overlaid on anatomical MRI, and only those cortical surface nodes with source amplitudes that survived statistical significance (p < 0.05) among total surface nodes were rendered on MRI. The source activities in RAC and LAC were plotted in Figure 2C,D, respectively. Figure 2E,F shows the power density spectra (Welch’s power spectrum [53]; signal window 1000 ms; overlapping 500 ms) obtained from Figure 2C,D, respectively. The spectral peaks at the fundamental frequency (37 Hz) and the first harmonic frequency (74 Hz) are marked by dashed lines.

2.6. Full Informational Spectral Analysis of Source Activities Using Holo-Hilbert Spectral Analysis (HHSA)

In this study, we adopted the Holo-Hilbert spectral analysis (HHSA) [48] to analyze the spectral information of the ASSR source activities estimated by MNE. The HHSA utilizes a two-layer empirical mode decomposition (EMD) architecture to extract the frequency-modulated (FM) information and the amplitude-modulated (AM) information. In the HHSA, the first EMD was applied to decompose the oscillatory signal into a set of intra-wave frequency-modulated IMFs, denoted as IMFFM. The second EMD was applied to the upper envelope of each IMFFM to obtain amplitude-modulated IMFs, denoted as IMFAM. The EMD was proposed by Huang et al. (1998) [49]. It attempts to sequentially decompose a signal into the sum of a finite number of intrinsic mode functions (IMFs) by iteratively conducting a sifting process, representing coarse-to-fine information of the recordings. Each IMF is a simple oscillatory signal whose amplitudes and frequencies are allowed to be varied with time so that it is beneficial to present local characteristics of nonstationary signals. The IMF is decomposed with the following definitions: (1) the number of local extrema (including local maxima and local minima) and the number of zero-crossings must either equal or differ at most by one, and (2) the mean value of the envelope defined by the local maxima and the envelope defined by the local minima are zeros.
After applying the first EMD, the ASSR source activities x(t) can be represented as the summation of IMFFMs, represented as:
x ( t ) = j = 1 K F j ( t ) ,
where F j ( t ) is the jth IMFFM.
Each IMFFM has its characteristic carrier frequency with its own time-varying modulation function and can be expressed as follows:
F j ( t ) = Re { a j ( t ) e i φ j ( t ) } ,
where Re { } is the function for the real-valued signal, a j ( t ) is the real-valued modulation function for the jth IMFFM, φ j ( t ) = t ω j ( τ ) d τ is the accumulated phase from the initial state to time t, and ω j ( τ ) is the instantaneous frequency at time τ.
By applying the second EMD to the envelope modulation function a j ( t ) of jth IMFFM, the a j ( t ) can be expressed as the summation of IMFAMs:
a j ( t ) = Re { m = 1 M a j , m ( t ) e i ψ j , m ( t ) } ,
where a j , m ( t ) is the mth IMFAM for jth IMFFM, ψ j , m ( t ) = t Ω j , m ( τ ) d τ is the accumulated phase, and Ω j , m ( τ ) is the instantaneous frequency at time τ for the mth IMFAM.
The ω j ( t ) and Ω j , m ( t ) are the instantaneous frequencies of the jth IMFFM in the first EMD and the mth IMFAM in the second EMD, respectively. The instantaneous frequencies ( ω j ( t ) , Ω j , m ( t ) ) were incorporated with the magnitude of envelop function | a j , m ( t ) | across all time points into vectors and organized as a three-dimensional AM-FM matrix ( ω j , Ω j , m , | a j , m | ) . The AM-FM matrices obtained from all IMFs were integrated together to achieve the Holo-Hilbert spectrum (HHS).
The signal processing steps for the HHSA processing are listed as follows:
  • Perform the first EMD to decompose the ASSR source activities into IMFFMs, and calculate instantaneous frequencies ω.
  • Take the absolute values of the IMFFMs.
  • Generate upper and lower envelopes for the absolute values of the IMFFMs by connecting the local extrema using spline interpolation.
  • Perform the second EMD to the upper envelopes of the IMFFMs to obtain IMFAMs and calculate instantaneous frequencies Ω for IMFAMs.
  • Arrange instantaneous amplitudes | a j , m ( t ) | and instantaneous frequencies ωj(t) and Ωj,m(t) into vectors to form a three-dimensional AM-FM matrix.
  • Construct the Holo-Hilbert Spectrum (HHS) by integrating the three-dimensional AM-FM matrices from all IMFFMs and IMFAMs.
Figure 3A,B demonstrates the HHS plots of the LAC and RAC source activities shown in Figure 2C,D, respectively. The lower panels and the left panels in Figure 3A,B are the FM and AM spectra, which were calculated by accumulating the values on HHS across all AM and FM frequencies, respectively. It can be observed that the FM spectra (lower panels) have clear spectral peaks at the ASSR fundamental frequency and its first harmonic frequency (marked by blue and green dashed lines). For the AM spectra (left panels), since we are interested in the responses at stimulus-related frequencies, the AM spectra at 37 Hz (the fundamental frequency) and 74 Hz (the first harmonic frequency) are plotted and marked in blue and green colors, respectively. The AM spectra at 37 Hz have spectral peaks around 2, 4.6 and 9.2 Hz, in which the 9.2 Hz could be the harmonic frequency of 4.6 Hz. In contrast, the AM spectra at 74 Hz have spectral peaks around 5.6 and 13 Hz. Since the temporal changes in amplitudes can result in frequency splitting or spectral broadening on traditional Fourier spectra, the HHSA interprets the amplitude changes in AM spectrum on HHS and avoids the frequency splitting problem on its FM spectrum.

3. Results

In this study, we studied the FM and AM spectra in monaural stimulations using HHSA. The auditory-induced neural activities on the left and right auditory cortexes (LAC and RAC) were obtained by means of using the minimum norm estimation (MNE) technique, and then the HHSA was applied to analyze the FM and AM frequency responses. Figure 4 shows the cross-subject average of FM spectra over the 25 subjects during left and right ear stimulations. The 37 Hz amplitudes of left monaural stimulation in LAC and RAC were 2.19 ± 0.07 μv/Hz vs. 2.25 ± 0.1 μv/Hz, and the 37 Hz amplitudes of right monaural stimulation in LAC and RAC were 2.58 ± 0.06 μv/Hz vs. 2.49 ± 0.09 μv/Hz. The 37 Hz auditory responses in right-ear stimulation were all higher than the auditory responses in left-ear stimulation (p < 0.01; Wilcoxon signed rank test), and the auditory responses on the contralateral side were higher than the responses on the ipsilateral side (p < 0.05; Wilcoxon signed rank test). For the amplitudes at 74 Hz (first harmonic frequency), the amplitudes of left monaural stimulation in LAC and RAC were 1.02 ± 0.10 μv/Hz vs. 0.99 ± 0.13 μv/Hz, and the 74 Hz amplitudes of the right monaural stimulation in LAC and RAC were 1.19 ± 0.10 μv/Hz vs. 1.14 ± 0.09 μv/Hz. The amplitudes in right-ear stimulation were higher than the amplitudes in left-ear stimulation (p < 0.01; Wilcoxon signed rank test). The 74 Hz amplitude of RAC was higher than the amplitude of LAC in left ear stimulation (p < 0.05; Wilcoxon signed rank test).
The noteworthy benefit of HHSA is its capability to analyze AM spectrum for a measured signal. Figure 5 and Figure 6 demonstrate the AM spectra of the auditory responses at 37 and 74 Hz, respectively. The cross-subject averages of AM spectra at 37 Hz during left and right monaural stimulations are shown in Figure 5A,B, respectively. Both the left and right monaural stimulations presented three clear spectral peaks at 2.5, 5 and 9 Hz. The amplitudes of 2.5, 5 and 9 Hz of LAC in left-ear stimulations were 0.115, 0.118 and 0.083 μv/Hz, respectively, and the amplitudes of 2.5, 5 and 9 Hz of LAC in right-ear stimulations were 0.123, 0.124 and 0.084 μv/Hz, respectively. For the AM spectra of 37 Hz in right-ear stimulations, the amplitudes of 2.5, 5 and 9 Hz in LAC were 0.134, 0.143 and 0.100 μv/Hz, respectively, and the amplitudes of 2.5, 5 and 9 Hz in RAC were 0.132, 0.138 μv/Hz and 0.098 μv/Hz, respectively. The 2.5, 5 and 9 Hz were the three main peaks at the 37 Hz FM frequency. Since the amplitudes at 2.5 and 5 Hz had similar amplitude levels and the third peak (9 Hz) was not located at the second harmonic frequency of 2.5 Hz, the 5 and 9 Hz were not likely the harmonic frequency peaks of the 2.5 Hz component in each AM spectrum. The three modulation frequencies might be generated from distinct neural mechanisms. Further studies are needed to check the physiological meanings of these peaks on the steady-state auditory responses.
Examining the AM spectra of the 74 Hz auditory responses, three spectral peaks at 3 Hz, 6 Hz and one another spectral peak around 12 Hz were observed in the left-ear and right-ear stimulations. Similar to the three spectral peaks of AM spectra at 37 Hz (Figure 5), three spectral peaks at 3, 6 and 12 Hz were observed. The frequencies of the third spectral peaks in left-ear stimulations were 1 Hz higher than the third spectral peaks in right-ear stimulations. The amplitudes of the 3, 6 and 13 Hz in left-ear stimulations were 0.047, 0.037 and 0.026 μv/Hz in LAC, and the amplitudes were 0.050, 0.033 and 0.021 μv/Hz in RAC, respectively. For the AM spectra of 74 Hz in right-ear stimulations, the amplitudes of 3, 6 and 12 Hz were 0.055, 0.043 and 0.03 μv/Hz in LAC, and the amplitudes were 0.053, 0.039 and 0.027 μv/Hz in RAC, respectively. The 6 and 12 Hz could be the first and second harmonic frequencies of 3 Hz in the AM spectrum of 74 Hz auditory responses.
The HHSA enables scientists to observe the amplitude modulations in the measured signals. With the AM-FM representation of HHSA, the AM spectrum perfectly explains the effects of amplitude changes on each FM frequency, and the HHSA arranges the FM and AM components in a two-dimensional HHS representation (see Figure 3). In Figure 7, the Fourier spectra (Welch’s power spectrum [53]; signal window 1000 ms; overlapping 500 ms) (marked in blue) of source activities in LAC and RAC were overlapped with the FM spectra shown in Figure 4 (marked in red). For the 37 Hz fundamental frequency, it can be observed that the Fourier spectrum in LAC during the left-ear stimulation (the left panel in Figure 7A) had two additional spectral side peaks, located at 33.33 and 40.74 Hz, which were 3.67 and 3.74 Hz away from the 37 Hz stimulation frequency, respectively. The spectral side peaks were not seen in the FM spectrum obtained from HHSA. For the spectral analysis of source activities in RAC (the right panel in Figure 7A), the Fourier spectra had two spectral side peaks at 31.48 and 42.59 Hz in the left-ear stimulation, which were 5.52 and 5.59 Hz away from the 37 Hz stimulation frequency, respectively. Compared to the FM spectrum of left-ear stimulation in RAC, the spectral side peaks were also not seen. It validates the benefit of the AM-FM analysis using HHSA. The HHSA expands the traditional Fourier spectrum into a two-dimension HHS representation so that the effects of AM components are not confused with FM components. Accordingly, the frequency splitting (trigonometric product-to-sum formula) caused by amplitude modulation can be corrected. A similar observation was found in the right-ear stimulation (Figure 7B). The Fourier spectrum in LAC had a 40.74 Hz spectral side peak, and the Fourier spectrum in RAC had 34 and 40.74 Hz spectral side peaks. In contrast to the 74 Hz component on Fourier spectra, no 74 Hz spectral peak was found. The missing 74 Hz harmonic frequency could be blurred owing to the frequency splitting or spectral broadening caused by the effects of amplitude modulation. In addition, an 80.5 spectral side peak was observed in all FM spectra, which might require further studies to understand its physiological meaning.

4. Discussion

In the present study, the ASSR signals in human brain were analyzed using the combination of CCA and HHSA. The CCA was adopted to extract ASSR-related signal features, and the selected canonical vectors were used to reconstruct noise-suppressed auditory responses. The HHSA was applied to analyze the AM-FM spectral information by calculating the instantaneous frequencies (IF) in local time slots. However, the IF requires the calculation of the first derivative in signal phases, which is sensitive to noise when the phase difference between samples is too large. Accordingly, the adoption of CCA in the preprocessing step helps the signal frequencies to be confined around our interested ASSR frequencies so that the phase fluctuations can be suppressed. Since ASSR is a sinusoidal-like wave with frequencies located at the stimulation frequency and its harmonics, the adoption of CCA can effectively enhance the ASSR-related features by setting the sinusoidal waves at fundamental and harmonic frequencies. The CCA aims to find a linear transformation to maximally correlate the extracted EEG features with the preset sinusoidal vectors. The canonical vectors in U (see Equation (3)) constructed a feature space. The multichannel EEG samples were projected to these canonical vectors for the extraction of ASSR-related signals. After calculating the weights of these canonical vectors on each EEG channel, the noise-suppressed EEG data can then be reconstructed by the summation of multiplying the weights of ASSR-related canonical vectors with their time series (see Equation (5)).
In our study, the canonical vectors with a statistical significance level of p < 0.01 were identified as ASSR-related canonical vectors for reconstructing ASSR-related signals. Compared to other CCA papers [54], the task-related canonical vectors were determined by defining a fixed threshold level on the correlation values. Since EEG signals are weak and susceptible to environmental noises, e.g., electromagnetic interference and motion artifacts, the use of a fixed correlation threshold for component selection might be risky owing to the contamination of nonstationary noise in varied measurement environments. For example, the correlation values of the canonical vectors extracted from right auditory ROI were 0.032, 0.031, 0.004, and 0.003 (see Figure 1B), respectively, and the correlation values of the canonical vectors extracted from left auditory ROI were 0.033, 0.017, 0.010, and 0.003 (see Figure 1C), respectively. For the canonical vectors in Figure 1C, the correlation values of the second and third canonical vectors (0.017 and 0.010) had a gap lower than the correlation value of the first canonical vector (0.033) and a gap higher than the correlation value of the fourth canonical vector (0.003). This resulted in the difficulty of finding a clear threshold for selecting the ASSR-related canonical vectors since the correlation values are changed depending on the noise patterns and noise magnitudes exerted on the ASSR responses. Determining the correlation threshold based on unsupervised clustering techniques, such as k-means, fuzzy C-means, Gaussian mixture model, etc. [55], could still be heuristic because the cluster number and grouping strategy should be defined by users in those unsupervised cluster classifiers. Setting wrong parameters in the use of clustering methods could include task-unrelated components and result in feature loss or noise engagement. Therefore, in order to find pertinent canonical vectors for reconstructing ASSR signals, we created a spatial map for each canonical vector by projecting the canonical vector on the measured signal of each EEG channel. We found the canonical vectors with statistical significance levels had clear spatial distribution concentrated around the left or right auditory ROI, which demonstrated the source activities of these canonical vectors originated from the left or right auditory cortexes. The canonical vectors with significant p values were thence chosen for reconstructing ASSR signals.
Because EEG is more sensitive to the radial component of neuroelectric activity, the measured EEG amplitude is decreased owing to the projection angle between the dipole orientation and the radial direction of the electrode surface [56]. Therefore, current sources with rotating orientations can result in amplitude changes in the electric fields picked up by the EEG sensor [56]. In this study, the HHSA was applied to the neural activities in the auditory cortex to avoid the influence of source orientations. The source activities were first estimated using minimum norm estimation (MNE) [51], and the neural activities at LAC and RAC were used as input signals for HHSA analysis in order to avoid the influence of rotating source caused amplitude changes on the sensor level (EEG sensors).
One noteworthy advantage of HHSA is the extending of the traditional spectral analysis from FM to AM-FM representation [57,58,59]. By unfolding the amplitude modulation and carrier frequency into AM and FM axes, the two-dimension frequency expression resolves the frequency splitting problem caused by the trigonometric product-to-sum formula. It has been reported that different physiological states can affect the modulation of rhythmic source activities. Because the human brain is a dynamic system, neurophysiological activities can be modulated by attention [6,7,10,60], emotional state [61,62,63], autonomic regulation [64,65,66], brain resting-state network [67,68,69], etc. In our study, we found the ASSRs at 37 and 74 Hz were modulated by AM waves with composite frequency components. For the 37 Hz auditory responses, the AM spectra had three spectral peaks (see Figure 5), located around 2.5, 5 and 9 Hz, and a broad frequency modulation range with −3 dB bandwidth about 20 Hz (±10 Hz around 37 Hz). Similarly, the AM spectrum of the 74 Hz auditory responses presented three peaks (see Figure 6), located around 3, 6 and 12 Hz, and a broad modulation frequency range with −3 dB bandwidth of about 14 Hz (±7 Hz around 74 Hz). As shown in Figure 7, the spectral side peaks were observed around 37 Hz in Fourier spectra, which could be the splitting frequencies caused by the product-to-sum formula. The amplitude modulation results in spurious spectral peaks or broadened spectrum, which could result in misleading or wrong interpretation of the measured signal. In Figure 7, the first harmonic frequency (74 Hz) was missing in all the Fourier spectra and restored in the FM spectra obtained from HHSA, which demonstrated the effectiveness of HHSA in avoiding modulation-caused spectral broadening.

5. Conclusions

The present study proposed a CCA and HHSA combination method for ASSR analyses. With the benefit of HHSA in AM and FM frequency analysis, the information on AM and FM spectra can be obtained. Because the frequency analysis in HHSA is calculated based on the IF technique, which is sensitive to noise, we adopted CCA as a preprocessing tool to avoid fluctuations in the estimated instantaneous frequencies. With the HHSA, we have demonstrated the existence of modulation frequencies in the measured ASSR responses. Both the 37 and 74 Hz auditory responses were modulated by AM spectra with at least three composite frequencies. In contrast to the frequency splitting in traditional Fourier spectra, the HHSA interprets the time-varying amplitude changes by AM frequencies on HHS. The proposed method effectively corrects the frequency splitting caused by modulation effects (product-to-sum formula), which avoids a misleading or wrong interpretation of the spectral analysis in the measured steady-state responses.

Author Contributions

Conceptualization, P.-L.L. and H.-H.C.; methodology, P.-L.L. and T.-M.L.; validation, W.-K.L., N.N.C. and Y.E.S.; formal analysis, P.-L.L. and T.-M.L.; investigation, H.-T.H. and W.-K.L.; resources, H.-H.C.; writing—original draft preparation, P.-L.L. and H.-H.C.; writing—review and editing, N.N.C. and Y.E.S.; project administration, P.-L.L.; funding acquisition, P.-L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Ministry of Science and Technology under Grant Number: 108-2634-F-008-003 through Pervasive Artificial Intelligence Research (PAIR) Labs, Taiwan. This study was also funded by the National Central University, Ministry of Science and Technology 111-2217-E-008-001-MY3, 111-2811-E-008-002-MY3, 110-2221-E-008-095-MY3, 110-2923-E-008-002-MY3, 109-2221-E-008-074), and Taoyuan Hospital Intramural project.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board (Ethics Committee of Institutional Review Board (IRB), Tao-Yuan General Hospital, Taiwan).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Acknowledgments

The program of Holo-Hilbert spectral analysis (HHSA) was provided by Norden E. Huang in the Research Center for Adaptive Data Analysis, National Central University, Taiwan.

Conflicts of Interest

The authors declare no conflict of interest. Narisa Chu is the co-founder of the company, CWLab International, USA. The authors declare that the company, CWLab International, had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Brenner, C.A.; Krishnan, G.P.; Vohs, J.L.; Ahn, W.-Y.; Hetrick, W.P.; Morzorati, S.L.; O’Donnell, B.F. Steady state responses: Electrophysiological assessment of sensory function in schizophrenia. Schizophr. Bull. 2009, 35, 1065–1077. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Vialatte, F.-B.; Maurice, M.; Dauwels, J.; Cichocki, A. Steady-state visually evoked potentials: Focus on essential paradigms and future perspectives. Prog. Neurobiol. 2010, 90, 418–438. [Google Scholar] [CrossRef] [PubMed]
  3. Lee, P.-L.; Yeh, C.-L.; Cheng, J.Y.-S.; Yang, C.-Y.; Lan, G.-Y. An SSVEP-based BCI using high duty-cycle visual flicker. IEEE Trans. Biomed. Eng. 2011, 58, 3350–3359. [Google Scholar] [CrossRef] [PubMed]
  4. Vergeer, M.; Mesik, J.; Baek, Y.; Wilmerding, K.; Engel, S.A. Orientation-selective contrast adaptation measured with SSVEP. J. Vis. 2018, 18, 2. [Google Scholar] [CrossRef] [Green Version]
  5. Zhu, D.; Molina, G.G.; Mihajlović, V.; Aarts, R.M. Phase synchrony analysis for SSVEP-based BCIs. In Proceedings of the 2010 2nd International Conference on Computer Engineering and Technology, Chengdu, China, 16–18 April 2010; IEEE: Piscataway Township, NJ, USA, 2010; Volume 2, pp. V2-329–V322-333. [Google Scholar]
  6. Kelly, S.P.; Lalor, E.C.; Reilly, R.B.; Foxe, J.J. Visual spatial attention tracking using high-density SSVEP data for independent brain-computer communication. IEEE Trans. Neural Syst. Rehabil. Eng. 2005, 13, 172–178. [Google Scholar] [CrossRef]
  7. Ross, B.; Picton, T.; Herdman, A.; Pantev, C. The effect of attention on the auditory steady-state response. Neurol. Clin. Neurophysiol. 2004, 2004, 22. [Google Scholar]
  8. Sharon, O.; Nir, Y. Attenuated fast steady-state visual evoked potentials during human sleep. Cereb. Cortex 2018, 28, 1297–1311. [Google Scholar] [CrossRef]
  9. Xu, P.; Tian, C.; Zhang, Y.; Jing, W.; Wang, Z.; Liu, T.; Hu, J.; Tian, Y.; Xia, Y.; Yao, D. Cortical network properties revealed by SSVEP in anesthetized rats. Sci. Rep. 2013, 3, 2496. [Google Scholar] [CrossRef] [Green Version]
  10. Palaniappan, R.; Mouli, S.; Balli, T.; McLoughlin, I. On the mental fatigue analysis of SSVEP entrainment. In Proceedings of the 2018 2nd International Conference on BioSignal Analysis, Processing and Systems (ICBAPS), Kuching, Malaysia, 24–26 July 2018; IEEE: Piscataway Township, NJ, USA, 2018; pp. 19–24. [Google Scholar]
  11. Xie, J.; Xu, G.; Luo, A.; Li, M.; Zhang, S.; Han, C.; Yan, W. The role of visual noise in influencing mental load and fatigue in a steady-state motion visual evoked potential-based brain-computer interface. Sensors 2017, 17, 1873. [Google Scholar] [CrossRef]
  12. Xie, J.; Xu, G.; Wang, J.; Li, M.; Han, C.; Jia, Y. Effects of mental load and fatigue on steady-state evoked potential based brain computer interface tasks: A comparison of periodic flickering and motion-reversal based visual attention. PLoS ONE 2016, 11, e0163426. [Google Scholar] [CrossRef]
  13. Lee, P.-L.; Lee, W.-K.; Chang, H.-H.; Hsu, H.-T.; Yeh, T.-K. Real-Time Cognitive Load Measurement in Classroom Environment using a Dry-Electrode EEG system. Int. J. Clin. Med. Bioeng. 2021, 1, 46–52. [Google Scholar] [CrossRef]
  14. Cao, T.; Wan, F.; Wong, C.M.; da Cruz, J.N.; Hu, Y. Objective evaluation of fatigue by EEG spectral analysis in steady-state visual evoked potential-based brain-computer interfaces. Biomed. Eng. Online 2014, 13, 1–13. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Lopour, B.A.; Tavassoli, A.; Fried, I.; Ringach, D.L. Coding of information in the phase of local field potentials within human medial temporal lobe. Neuron 2013, 79, 594–606. [Google Scholar] [CrossRef] [Green Version]
  16. Kisley, M.A.; Cornwell, Z.M. Gamma and beta neural activity evoked during a sensory gating paradigm: Effects of auditory, somatosensory and cross-modal stimulation. Clin. Neurophysiol. 2006, 117, 2549–2563. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Borst, A.; Theunissen, F.E. Information theory and neural coding. Nat. Neurosci. 1999, 2, 947–957. [Google Scholar] [CrossRef]
  18. Zhong, Z.; Henry, K.S.; Heinz, M.G. Sensorineural hearing loss amplifies neural coding of envelope information in the central auditory system of chinchillas. Hear. Res. 2014, 309, 55–62. [Google Scholar] [CrossRef] [Green Version]
  19. Nasir, S.M.; Darainy, M.; Ostry, D.J. Sensorimotor adaptation changes the neural coding of somatosensory stimuli. J. Neurophysiol. 2013, 109, 2077–2085. [Google Scholar] [CrossRef] [Green Version]
  20. Gerstner, W.; Kistler, W.M. Spiking Neuron Models: Single Neurons, Populations, Plasticity; Cambridge University Press: Cambridge, UK, 2002. [Google Scholar]
  21. Aghababaiyan, K.; Shah-Mansouri, V.; Maham, B. Capacity and error probability analysis of neuro-spike communication exploiting temporal modulation. IEEE Trans. Commun. 2019, 68, 2078–2089. [Google Scholar] [CrossRef]
  22. Borji, A.; Itti, L. Optimal attentional modulation of a neural population. Front. Comput. Neurosci. 2014, 8, 34. [Google Scholar] [CrossRef] [Green Version]
  23. Wager, T.D.; Davidson, M.L.; Hughes, B.L.; Lindquist, M.A.; Ochsner, K.N. Neural mechanisms of emotion regulation: Evidence for two independent prefrontal-subcortical pathways. Neuron 2008, 59, 1037. [Google Scholar] [CrossRef] [Green Version]
  24. Pfister, J.-P.; Toyoizumi, T.; Barber, D.; Gerstner, W. Optimal spike-timing-dependent plasticity for precise action potential firing in supervised learning. Neural Comput. 2006, 18, 1318–1348. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Yokota, Y.; Tanaka, S.; Miyamoto, A.; Naruse, Y. Estimation of human workload from the auditory steady-state response recorded via a wearable electroencephalography system during walking. Front. Hum. Neurosci. 2017, 11, 314. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Boettcher, F.A.; Poth, E.A.; Mills, J.H.; Dubno, J.R. The amplitude-modulation following response in young and aged human subjects. Hear. Res. 2001, 153, 32–42. [Google Scholar] [CrossRef]
  27. Burton, M.J.; Cohen, L.T.; Rickards, F.W.; McNally, K.I.; Clark, G.M. Steady-state evoked potentials to amplitude modulated tones in the monkey. Acta Oto-Laryngol. 1992, 112, 745–751. [Google Scholar] [CrossRef] [PubMed]
  28. Rees, A.; Green, G.G.; Kay, R.H. Steady-state evoked responses to sinusoidally amplitude-modulated sounds recorded in man. Hear. Res. 1986, 23, 123–133. [Google Scholar] [CrossRef]
  29. John, M.S.; Dimitrijevic, A.; van Roon, P.; Picton, T.W. Multiple auditory steady-state responses to AM and FM stimuli. Audiol. Neurotol. 2001, 6, 12–27. [Google Scholar] [CrossRef]
  30. Kim, S.; Jang, S.-K.; Kim, D.-W.; Shim, M.; Kim, Y.-W.; Im, C.-H.; Lee, S.-H. Cortical volume and 40-Hz auditory-steady-state responses in patients with schizophrenia and healthy controls. NeuroImage Clin. 2019, 22, 101732. [Google Scholar] [CrossRef]
  31. Rass, O.; Krishnan, G.; Brenner, C.A.; Hetrick, W.P.; Merrill, C.C.; Shekhar, A.; O’Donnell, B.F. Auditory steady state response in bipolar disorder: Relation to clinical state, cognitive performance, medication status, and substance disorders. Bipolar Disord. 2010, 12, 793–803. [Google Scholar] [CrossRef] [Green Version]
  32. Vanvooren, S.; Poelmans, H.; Hofmann, M.; Luts, H.; Dandache, S.; Ghesquière, P.; Wouters, J. Auditory temporal processing in dyslexia: An ASSR study in young children. Int. J. Psychophysiol. 2012, 85, 358. [Google Scholar] [CrossRef] [Green Version]
  33. De Siati, R.D.; Rosenzweig, F.; Gersdorff, G.; Gregoire, A.; Rombaux, P.; Deggouj, N. Auditory neuropathy spectrum disorders: From diagnosis to treatment: Literature review and case reports. J. Clin. Med. 2020, 9, 1074. [Google Scholar] [CrossRef]
  34. Ishida, I.M.; Stapells, D.R. Multiple-ASSR Interactions in Adults with Sensorineural Hearing Loss. Int. J. Otolaryngol. 2012, 2012, 802715. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Van Maanen, A.; Stapells, D.R. Multiple-ASSR thresholds in infants and young children with hearing loss. J. Am. Acad. Audiol. 2010, 21, 535–545. [Google Scholar] [CrossRef] [PubMed]
  36. Lins, O.G.; Picton, T.W.; Boucher, B.L.; Durieux-Smith, A.; Champagne, S.C.; Moran, L.M.; Perez-Abalo, M.C.; Martin, V.; Savio, G. Frequency-specific audiometry using steady-state responses. Ear Hear. 1996, 17, 81–96. [Google Scholar] [CrossRef] [PubMed]
  37. Picton, T.W.; Dimitrijevic, A.; Perez-Abalo, M.C.; Van Roon, P. Estimating audiometric thresholds using auditory steady-state responses. J. Am. Acad. Audiol. 2005, 16, 140–156. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Beck, D.L.; Speidel, D.P.; Petrak, M. Auditory steady-state response (ASSR): A beginner’s guide. Hear. Rev. 2007, 14, 34–37. [Google Scholar]
  39. Khuwaja, G.A.; Haghighi, S.J.; Hatzinakos, D. 40-Hz ASSR fusion classification system for observing sleep patterns. EURASIP J. Bioinform. Syst. Biol. 2015, 2015, 2. [Google Scholar] [CrossRef] [Green Version]
  40. Swanepoel, D.; Hugo, R.; Roode, R. Auditory steady-state responses for children with severe to profound hearing loss. Arch. Otolaryngol.-Head Neck Surg. 2004, 130, 531–535. [Google Scholar] [CrossRef] [Green Version]
  41. Sugiyama, S.; Ohi, K.; Kuramitsu, A.; Takai, K.; Muto, Y.; Taniguchi, T.; Kinukawa, T.; Takeuchi, N.; Motomura, E.; Nishihara, M. The auditory steady-state response: Electrophysiological index for sensory processing dysfunction in psychiatric disorders. Front. Psychiatry 2021, 12, 644541. [Google Scholar] [CrossRef]
  42. Yokota, Y.; Naruse, Y. Phase coherence of auditory steady-state response reflects the amount of cognitive workload in a modified N-back task. Neurosci. Res. 2015, 100, 39–45. [Google Scholar] [CrossRef] [Green Version]
  43. Kallenberg, M.; Desain, P.; Gielen, S. Auditory selective attention as a method for a brain computer interface. Nijmegen CNS 2006, 2, 1–16. [Google Scholar]
  44. Biesmans, W.; Bertrand, A.; Wouters, J.; Moonen, M. Optimal spatial filtering for auditory steady-state response detection using high-density EEG. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April; IEEE: Piscataway Township, NJ, USA, 2015; pp. 857–861. [Google Scholar]
  45. Janani, A.S.; Grummett, T.S.; Lewis, T.W.; Fitzgibbon, S.P.; Whitham, E.M.; DelosAngeles, D.; Bakhshayesh, H.; Willoughby, J.O.; Pope, K.J. Improved artefact removal from EEG using Canonical Correlation Analysis and spectral slope. J. Neurosci. Methods 2018, 298, 1–15. [Google Scholar] [CrossRef] [PubMed]
  46. Wang, K.-W.; Chang, H.-H.; Hsu, C.-C.; Chen, K.-C.; Hsieh, J.-C.; Li, L.P.-H.; Lee, P.-L.; Shiao, A.-S. Extractions of steady-state auditory evoked fields in normal subjects and tinnitus patients using complementary ensemble empirical mode decomposition. BioMedical Eng. OnLine 2015, 14, 1–18. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Bin, G.; Gao, X.; Yan, Z.; Hong, B.; Gao, S. An online multi-channel SSVEP-based brain–computer interface using a canonical correlation analysis method. J. Neural Eng. 2009, 6, 046002. [Google Scholar] [CrossRef]
  48. Huang, N.E.; Hu, K.; Yang, A.C.; Chang, H.-C.; Jia, D.; Liang, W.-K.; Yeh, J.R.; Kao, C.-L.; Juan, C.-H.; Peng, C.K. On Holo-Hilbert spectral analysis: A full informational spectral representation for nonlinear and non-stationary data. Philos. Trans. R. Soc. A: Math. Phys. Eng. Sci. 2016, 374, 20150206. [Google Scholar] [CrossRef] [PubMed]
  49. Huang, N.E.; Shen, Z.; Long, S.R.; Wu, M.C.; Shih, H.H.; Zheng, Q.; Yen, N.-C.; Tung, C.C.; Liu, H.H. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis. Proc. R. Soc. London. Ser. A Math. Phys. Eng. Sci. 1998, 454, 903–995. [Google Scholar] [CrossRef]
  50. Schlee, W.; Weisz, N.; Bertrand, O.; Hartmann, T.; Elbert, T. Using auditory steady state responses to outline the functional connectivity in the tinnitus brain. PLoS ONE 2008, 3, e3720. [Google Scholar] [CrossRef]
  51. Tadel, F.; Baillet, S.; Mosher, J.C.; Pantazis, D.; Leahy, R.M. Brainstorm: A user-friendly application for MEG/EEG analysis. Comput. Intell. Neurosci. 2011, 2011, 879716. [Google Scholar] [CrossRef]
  52. Cointepas, Y.; Mangin, J.-F.; Garnero, L.; Poline, J.-B.; Benali, H. BrainVISA: Software platform for visualization and analysis of multi-modality brain data. Neuroimage 2001, 13, 98. [Google Scholar] [CrossRef]
  53. Solomon, O.M., Jr. PSD Computations Using Welch’s Method. [Power Spectral Density (PSD)]; Sandia National Labs: Albuquerque, NM, USA, 1991. [Google Scholar]
  54. von Lühmann, A.; Li, X.; Müller, K.-R.; Boas, D.A.; Yücel, M.A. Improved physiological noise regression in fNIRS: A multimodal extension of the general linear model using temporally embedded canonical correlation analysis. NeuroImage 2020, 208, 116472. [Google Scholar] [CrossRef]
  55. Saxena, A.; Prasad, M.; Gupta, A.; Bharill, N.; Patel, O.P.; Tiwari, A.; Er, M.J.; Ding, W.; Lin, C.-T. A review of clustering techniques and developments. Neurocomputing 2017, 267, 664–681. [Google Scholar] [CrossRef] [Green Version]
  56. Mosher, J.C.; Leahy, R.M. Recursive MUSIC: A framework for EEG and MEG source localization. IEEE Trans. Biomed. Eng. 1998, 45, 1342–1354. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Chang, H.-H.; Hsu, C.-C.; Chen, C.-Y.; Lee, W.-K.; Hsu, H.-T.; Shyu, K.-K.; Yeh, J.-R.; Lin, P.-J.; Lee, P.-L. A method for respiration rate detection in wrist PPG signal using Holo-Hilbert spectrum. IEEE Sens. J. 2018, 18, 7560–7569. [Google Scholar] [CrossRef]
  58. Hsu, H.-T.; Lee, W.-K.; Shyu, K.-K.; Yeh, T.-K.; Chang, C.-Y.; Lee, P.-L. Analyses of EEG oscillatory activities during slow and fast repetitive movements using Holo-Hilbert spectral analysis. IEEE Trans. Neural Syst. Rehabil. Eng. 2018, 26, 1659–1668. [Google Scholar] [CrossRef] [PubMed]
  59. Lee, T.-M.; Lee, P.-L.; Lee, I.-H.; Lee, W.-K.; Wu, T.-Y.; Hsu, H.-T.; Yeh, C.-L.; Lin, P.-J.; Shyu, K.-K. Study of heart-rate variability in a video task using Holo-Hilbert spectral analysis. Biomed. Signal Processing Control. 2022, 71, 103229. [Google Scholar] [CrossRef]
  60. Griskova-Bulanova, I.; Ruksenas, O.; Dapsys, K.; Maciulis, V.; Arnfred, S.M. Distraction task rather than focal attention modulates gamma activity associated with auditory steady-state responses (ASSRs). Clin. Neurophysiol. 2011, 122, 1541–1548. [Google Scholar] [CrossRef]
  61. Heo, J.; Baek, H.J.; Hong, S.; Chang, M.H.; Lee, J.S.; Park, K.S. Music and natural sounds in an auditory steady-state response based brain–computer interface to increase user acceptance. Comput. Biol. Med. 2017, 84, 45–52. [Google Scholar] [CrossRef]
  62. Vohs, J.L.; Andrew Chambers, R.; Krishnan, G.P.; O’Donnell, B.F.; Berg, S.; Morzorati, S.L. GABAergic modulation of the 40 Hz auditory steady-state response in a rat model of schizophrenia. Int. J. Neuropsychopharmacol. 2010, 13, 487–497. [Google Scholar] [CrossRef] [Green Version]
  63. Wang, Y.; Li, Z.; Tian, Z.; Wang, X.; Li, Y.; Qin, L. Emotional arousal modifies auditory steady state response in the auditory cortex and prefrontal cortex of rats. Stress 2019, 22, 492–500. [Google Scholar] [CrossRef]
  64. Anderssen, S.; Nicolaisen, R.; Gabrielsen, G. Autonomic response to auditory stimulation. Acta Paediatr. 1993, 82, 913–918. [Google Scholar] [CrossRef]
  65. Dimberg, U. Facial electromyographic reactions and autonomic activity to auditory stimuli. Biol. Psychol. 1990, 31, 137–147. [Google Scholar] [CrossRef]
  66. Valenti, V.E.; Guida, H.L.; Frizzo, A.C.; Cardoso, A.C.; Vanderlei, L.C.M.; de Abreu, L.C. Auditory stimulation and cardiac autonomic regulation. Clinics 2012, 67, 955–958. [Google Scholar] [CrossRef]
  67. Drewes, J.; Muschter, E.; Zhu, W.; Melcher, D. Individual resting-state alpha peak frequency and within-trial changes in alpha peak frequency both predict visual dual-pulse segregation performance. Cereb. Cortex 2022, 32, bhac026. [Google Scholar] [CrossRef] [PubMed]
  68. Noda, Y.; Takano, M.; Hayano, M.; Li, X.; Wada, M.; Nakajima, S.; Mimura, M.; Kondo, S.; Tsubota, K. Photobiological Neuromodulation of Resting-State EEG and Steady-State Visual-Evoked Potentials by 40 Hz Violet Light Optical Stimulation in Healthy Individuals. J. Pers. Med. 2021, 11, 557. [Google Scholar] [CrossRef] [PubMed]
  69. Zou, Q.; Long, X.; Zuo, X.; Yan, C.; Zhu, C.; Yang, Y.; Liu, D.; He, Y.; Zang, Y. Functional connectivity between the thalamus and visual cortex under eyes closed and eyes open conditions: A resting-state fMRI study. Hum. Brain Mapp. 2009, 30, 3066–3078. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. The demonstration of CCA for ASSR extraction in subject5: (A) The left panel shows the channel plot of the recorded EEG signals. (B) The spatial maps and Fourier spectra of canonical vectors by setting the COI located in the vicinity of the right auditory cortex. (C) The spatial maps and Fourier spectra of canonical vectors by setting the COI located in the vicinity of the left auditory cortex. (D) The channel plot of the ASSR signals were reconstructed from ASSR-related canonical vectors.
Figure 1. The demonstration of CCA for ASSR extraction in subject5: (A) The left panel shows the channel plot of the recorded EEG signals. (B) The spatial maps and Fourier spectra of canonical vectors by setting the COI located in the vicinity of the right auditory cortex. (C) The spatial maps and Fourier spectra of canonical vectors by setting the COI located in the vicinity of the left auditory cortex. (D) The channel plot of the ASSR signals were reconstructed from ASSR-related canonical vectors.
Jcm 11 03868 g001
Figure 2. The source activities of ASSRs estimated by MNE in RAC and LAC: (A) The neural sources of ASSR in RAC; (B) the neural sources of ASSR in LAC; (C) the source activities in RAC; (D) the source activities in LAC; (E) the Fourier spectrum of the source activities in RAC; (F) the Fourier spectrum of the source activities in LAC.
Figure 2. The source activities of ASSRs estimated by MNE in RAC and LAC: (A) The neural sources of ASSR in RAC; (B) the neural sources of ASSR in LAC; (C) the source activities in RAC; (D) the source activities in LAC; (E) the Fourier spectrum of the source activities in RAC; (F) the Fourier spectrum of the source activities in LAC.
Jcm 11 03868 g002
Figure 3. The HHS plots of the source activities in auditory cortexes: (A) The HHS plot of the source activities in LAC. The lower panel shows the FM spectrum, and the left panel shows the AM spectrum. (B) The HHS plot of the source activities in RAC. The lower panel shows the FM spectrum, and the left panel shows the AM spectrum.
Figure 3. The HHS plots of the source activities in auditory cortexes: (A) The HHS plot of the source activities in LAC. The lower panel shows the FM spectrum, and the left panel shows the AM spectrum. (B) The HHS plot of the source activities in RAC. The lower panel shows the FM spectrum, and the left panel shows the AM spectrum.
Jcm 11 03868 g003
Figure 4. The FM spectra averaged across the twenty-five subjects in our study: (A) The cross-subject average of FM spectrum in LAC and RAC during left-ear auditory stimulations; (B) the cross-subject average of FM spectrum in LAC and RAC during right-ear auditory stimulations.
Figure 4. The FM spectra averaged across the twenty-five subjects in our study: (A) The cross-subject average of FM spectrum in LAC and RAC during left-ear auditory stimulations; (B) the cross-subject average of FM spectrum in LAC and RAC during right-ear auditory stimulations.
Jcm 11 03868 g004
Figure 5. The AM spectra of the 37 Hz FM frequency averaged across the twenty-five subjects in our study: (A) the cross-subject average of the AM spectrum at 37 Hz in LAC and RAC during left-ear auditory stimulations; (B) the cross-subject average of the AM spectrum at 37 Hz in LAC and RAC during right-ear auditory stimulations.
Figure 5. The AM spectra of the 37 Hz FM frequency averaged across the twenty-five subjects in our study: (A) the cross-subject average of the AM spectrum at 37 Hz in LAC and RAC during left-ear auditory stimulations; (B) the cross-subject average of the AM spectrum at 37 Hz in LAC and RAC during right-ear auditory stimulations.
Jcm 11 03868 g005
Figure 6. The AM spectra of the 74 Hz FM frequency averaged across the twenty-five subjects in our study: (A) the cross-subject average of the AM spectrum at 74 Hz in LAC and RAC during left-ear auditory stimulations; (B) the cross-subject average of the AM spectrum at 74 Hz in LAC and RAC during right-ear auditory stimulations.
Figure 6. The AM spectra of the 74 Hz FM frequency averaged across the twenty-five subjects in our study: (A) the cross-subject average of the AM spectrum at 74 Hz in LAC and RAC during left-ear auditory stimulations; (B) the cross-subject average of the AM spectrum at 74 Hz in LAC and RAC during right-ear auditory stimulations.
Jcm 11 03868 g006
Figure 7. The comparison of the Fourier spectra (marked in blue) and the FM spectra (marked in red) obtained from left-ear and right-ear auditory stimulations: (A) the cross-subject averaged spectra obtained from Fourier analysis and the HHSA during left-ear stimulations; (B) the cross-subject averaged spectra obtained from Fourier analysis and the HHSA during right-ear stimulations.
Figure 7. The comparison of the Fourier spectra (marked in blue) and the FM spectra (marked in red) obtained from left-ear and right-ear auditory stimulations: (A) the cross-subject averaged spectra obtained from Fourier analysis and the HHSA during left-ear stimulations; (B) the cross-subject averaged spectra obtained from Fourier analysis and the HHSA during right-ear stimulations.
Jcm 11 03868 g007
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, P.-L.; Lee, T.-M.; Lee, W.-K.; Chu, N.N.; Shelepin, Y.E.; Hsu, H.-T.; Chang, H.-H. The Full Informational Spectral Analysis for Auditory Steady-State Responses in Human Brain Using the Combination of Canonical Correlation Analysis and Holo-Hilbert Spectral Analysis. J. Clin. Med. 2022, 11, 3868. https://doi.org/10.3390/jcm11133868

AMA Style

Lee P-L, Lee T-M, Lee W-K, Chu NN, Shelepin YE, Hsu H-T, Chang H-H. The Full Informational Spectral Analysis for Auditory Steady-State Responses in Human Brain Using the Combination of Canonical Correlation Analysis and Holo-Hilbert Spectral Analysis. Journal of Clinical Medicine. 2022; 11(13):3868. https://doi.org/10.3390/jcm11133868

Chicago/Turabian Style

Lee, Po-Lei, Te-Min Lee, Wei-Keung Lee, Narisa Nan Chu, Yuri E. Shelepin, Hao-Teng Hsu, and Hsiao-Huang Chang. 2022. "The Full Informational Spectral Analysis for Auditory Steady-State Responses in Human Brain Using the Combination of Canonical Correlation Analysis and Holo-Hilbert Spectral Analysis" Journal of Clinical Medicine 11, no. 13: 3868. https://doi.org/10.3390/jcm11133868

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop