Next Article in Journal
Precision Optical Metrology and Smart Sensing
Previous Article in Journal
Needs Companion: A Novel Approach to Continuous User Needs Sensing Using Virtual Agents and Large Language Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Resting-State Electroencephalogram Depression Diagnosis Based on Traditional Machine Learning and Deep Learning: A Comparative Analysis

Heilongjiang Province Key Laboratory of Laser Spectroscopy Technology and Application, Harbin University of Science and Technology, Harbin 150080, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(21), 6815; https://doi.org/10.3390/s24216815
Submission received: 17 September 2024 / Revised: 6 October 2024 / Accepted: 16 October 2024 / Published: 23 October 2024
(This article belongs to the Section Biomedical Sensors)

Abstract

:
The global prevalence of Major Depressive Disorder (MDD) is increasing at an alarming rate, underscoring the urgent need for timely and accurate diagnoses to facilitate effective interventions and treatments. Electroencephalography remains a widely used neuroimaging technique in psychiatry, due to its non-invasive nature and cost-effectiveness. With the rise of computational psychiatry, the integration of EEG with artificial intelligence has yielded remarkable results in diagnosing depression. This review offers a comparative analysis of two predominant methodologies in research: traditional machine learning and deep learning methods. Furthermore, this review addresses key challenges in current research and suggests potential solutions. These insights aim to enhance diagnostic accuracy for depression and also foster further development in the area of computational psychiatry.

1. Introduction

Depression, a prevalent mental disorder, is characterized by a persistent low mood, anhedonia, or diminished interest in activities, and in severe cases, may lead to suicide [1]. According to the World Health Organization (WHO), an estimated 280 million people worldwide are affected by depression [2,3,4]. Consequently, depression is prioritized in the World Health Organization’s Mental Health Gap Action Programme [5]. Clinically, the diagnosis of a depressive episode typically requires a duration of at least two weeks, accompanied by significant distress or impairment in social functioning. The primary diagnostic criteria for depression are detailed in the Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition (DSM-5) or Fourth Edition (DSM-IV) [6,7]. Diagnoses typically rely on psychiatric evaluations and psychometric questionnaires, including the Hamilton Depression Scale (HAM-D) and the Self-Rating Depression Scale [8,9,10,11,12,13]. However, this approach relies heavily on patient cooperation and clinician expertise, which complicates the diagnostic process and may delay early intervention for patients [14,15]. Depression disrupts neural communication and neurogenesis, impairing the function of specific brain regions and leading to fluctuating patterns of brain activity [16,17]. Research shows that individuals with depression exhibit dynamic disturbances in the frontal limbic network and abnormal inter-cerebral functional connectivity [18,19], which is indicative of altered complex neuronal interactions. Consequently, numerous clinicians and researchers recognize that identifying objective physiological indicators for the direct diagnosis of depression could substantially enhance both diagnostic accuracy and treatment outcomes.
Electroencephalography (EEG) is a non-invasive neuroimaging technique offering a high temporal resolution and easy accessibility [20,21]. EEG accurately reflects brain network activity, offers objective neural markers, and remains free from clinical bias and subjective self-assessment. It has also proved valuable for capturing information about physiological changes in the brain associated with depression [22,23,24]. Compared to magnetic resonance imaging (MRI), EEG is more cost-effective and suitable for frequent testing, thereby offering significant advantages in diagnosing depression [25,26]. Compared to EEG during task execution, resting-state EEG (rsEEG) not only minimizes interference from visual scenes, instructions, and task performance variations influenced by subjects’ cognitive levels, gender, age, and interest [27], but it also effectively captures the intrinsic brain activity. The resting state allows subjects to focus their attention inward, thus facilitating self-related brain activity [28,29]. Furthermore, a study comparing resting-state and task-state EEG demonstrates that classification accuracy is higher in the resting state than in the task state [30]. With advances in computational psychiatry [31,32,33], the integration of rsEEG with artificial intelligence produces remarkable results in diagnosing depression [34]. In contemporary research on depression diagnosis, computer-aided diagnostic models primarily fall into two categories: traditional machine learning (TML) models and deep learning (DL) models. Traditional machine learning models rely on feature engineering to process and analyze training data, using algorithms to make decisions. Conversely, deep learning models can handle large datasets and autonomously learn complex feature representations.
Currently, there are only a few review articles that address the application of EEG in diagnosing depression. Numerous studies highlight the impact of feature extraction and the selection of classification models on diagnostic outcomes; however, few directly compare traditional machine learning methods with deep learning approaches. We compare and analyze common elements such as sample size, data acquisition, preprocessing, feature extraction, feature selection, classification methods, and validation techniques across these two types of studies. Furthermore, we examine the primary challenges in the current research and propose desirable practices. In all tables presented in this review, a dash (“-”) indicates that relevant data are unavailable in the reference.

2. Methods

2.1. Search Strategy

The use of deep learning methods for diagnosing depression has seen a rapid increase since 2018. Consequently, our study reviews diagnostic research published from 2018 to 2024. The search database used in this review is Web of Science. For the search of the literature, we use the following subject terms: (“Machine learning” OR “Deep learning”) AND (“EEG” OR “Electroencephalography”) AND (“Depression” OR “MDD”). A total of 371 papers are retrieved from the search.

2.2. Inclusion Criteria

Initially, the titles and abstracts of all papers are reviewed to exclude duplicates, withdrawn studies, and conference proceedings. Subsequently, the screening process continues with the following criteria:
  • The studies primarily focus on diagnosing depression;
  • The sample comprises both patients diagnosed with MDD and healthy controls (HC);
  • The EEG data consist exclusively of resting-state recordings;
  • Depression is diagnosed using either TML or DL;
  • The study involves tasks related to EEG data acquisition, EEG signal processing, classification, and validation.
The final sample for the comparative analysis comprises 49 relevant articles. Each article employs various signal processing, classification, and validation methods, essential for analyzing and comparing the performance of TML and DL methods.

3. Results

Forty-nine studies employing rsEEG for depression diagnosis, published between 2018 and 2024, are systematically reviewed. This includes 22 studies employing TML methods and 27 studies utilizing DL approaches. Two of these studies employ both methods [35,36]; they are classified based on the method that achieves the highest accuracy rate in the analysis. Section 3.1, Section 3.2, Section 3.3, Section 3.4, Section 3.5, Section 3.6 and Section 3.7 provide a detailed account of the sample size, data acquisition, and preprocessing approaches used in these studies, along with the methods applied for feature extraction and selection, classification, and validation. Figure 1 illustrates the general framework for diagnosing depression using TML and DL methods. The figure clearly shows that the raw EEG obtained from the EEG cap must first undergo preprocessing, as indicated by the elliptical area in the preprocessing box grid, which highlights the four most commonly utilized methods. Subsequently, two approaches are presented: on the left, the TML method and on the right, the DL method. In TML methods, feature extraction and selection are two distinct steps, with the most frequently used methods detailed in these modules. If a deep learning approach is employed, the feature extraction step includes two strategies: manual feature extraction integrated with a DL model, as depicted in the ellipse area, and automatic feature extraction, with schematic diagrams of the two predominant structures, convolutional neural network and long short-term memory, systematically arranged from top to bottom within the box frame. In the classification and validation module, each segment’s size in the sector diagram corresponds to the frequency of use of each method.

3.1. Sample Size

3.1.1. Studies Based on TML Methods

In 2012, Ahmadlou and his team conducted the first study on diagnosing depression using traditional machine learning with rsEEG, involving a limited sample size of 24 subjects [37]. In 2018, Cai et al. [38] expanded their study’s sample size to include 213 participants, comprising 92 individuals with depression and 121 controls without depression. Subsequently, Mumtaz et al. [39] published a study including 34 with MDD and 30 age-matched HC. Wan et al. [40] used two separate samples: one consists of 35 participants (23 with MDD and 12 HC), and the other comprises 30 participants (15 with MDD and 15 HC). In 2021, Wu et al. [41] significantly increased their sample size by collecting rsEEG data from 400 participants, evenly divided into 200 with MDD and 200 HC, across four different hospitals. In contrast, in 2022, Avots et al. [42] conducted a study featuring a notably limited sample size of 20 participants, consisting of 10 individuals with MDD and 10 HC. In the same year, Li et al. [43] analyzed data from 92 cases. Two studies by Soni et al. [44,45] used multiple datasets, each with sample sizes ranging from 30 to 55 and incorporating data from the study by Seal et al. [46], which includes 15 patients with MDD and 18 HC. According to the WHO, depression proves more prevalent in women than in men [3]. In 2023, Shim et al. [47] analyzed EEG data exclusively from female participants, consisting of 49 patients with MDD and 49 HC. This study is unique, as it exclusively includes female subjects. In the same year, two other studies included samples of 32 participants (19 with MDD and 13 HC) and 80 participants (40 with MDD and 40 HC), respectively [48,49]. Table 1 outlines the fundamental experimental setup for EEG data acquisition in studies employing both TML and DL methodologies. It features four columns, detailing from left to right: sample size, frequency (in Hz), number of electrodes, and the corresponding study. Table 2 provides the web address of the publicly available dataset used in the depression diagnostic study, which is accessible by clicking on the link. Additionally, numerous researchers have utilized these publicly available sample data in their studies, as detailed in Table 2. The dataset from Mumtaz et al. [39,50,51], which includes 34 patients with MDD and 30 HC, is frequently used by researchers for analysis [52,53,54,55,56,57,58]. Nassib et al. [59] used a public dataset (42 with MDD + 42 HC) published by Cavanagh’s team [60]. The Multi-modal Open Dataset for Mental-disorder Analysis (MODMA), provided by Cai’s team [61] and comprising two subsets—one with 24 patients with MDD and 29 HC, and another with 26 with MDD and 29 HC—significantly contributes to depression detection research conducted by various researchers [44,45,62,63].

3.1.2. Studies Based on DL Methods

In 2018, Acharya et al. [64] published a study on depression detection, collecting EEG signals from 30 participants, comprising 15 with MDD and 15 HC. Subsequently, Mao et al. [81] and Ay et al. [66] published studies with sample sizes of 34 participants (17 with MDD and 17 HC) and 30 participants (15 with MDD and 15 HC), respectively. In 2020, Wan’s team [72] and Duan et al. [35] chose the same data source for their experiments, though their sample sizes differ: Wan et al. included 35 participants (12 with MDD and 23 HC), while Duan et al. analyzed data from 32 participants (16 with MDD and 16 HC). In 2021, Seal et al. [65] analyzed a sample of 33 participants, including 15 with MDD and 18 HC. The samples employed by the researchers mentioned above are relatively small, with none exceeding 35 participants. Subsequently, Khan et al. [79] increased the sample size to 60 participants for depression diagnosis research, dividing them equally between patients with MDD and HC. In 2022, Yan et al. analyzed a dataset of 80 participants, evenly split between 40 with MDD and 40 HC [69,70]. In 2023, Zhang et al. [75] published a study including 53 participants, consisting of 24 with MDD and 29 HC. Subsequently, Xu et al. [73] further increased the sample size to 75 participants, including 41 with MDD and 34 HC, in their analysis. To better validate the proposed methods, an increasing number of researchers are opting to use multiple datasets. In these studies, the sample sizes generally hover around 105 participants, with one study exceeding 200 participants (refer to Table 2). The use of an entirely different set of EEG data renders the results significantly more compelling. Researchers extensively use publicly available datasets in studies employing deep learning frameworks. As indicated in Table 2, the publicly available dataset from Mumtaz’s team [39,50,51,89] is utilized by more researchers than any other, followed by the MODMA dataset from Cai’s team [61]. Furthermore, the datasets from Cavanagh’s team [60], the EDRA dataset shared by Yang et al. [90], and the PRED + CT dataset are also popular among various researchers.

3.1.3. Analysis

As indicated in Table 1, only four studies feature sample sizes exceeding 200, with one study comprising a sample of 400. Nine studies include an equal number of participants in the depression and control groups. Publicly available datasets are used more frequently in research utilizing deep learning frameworks, rather than in research employing traditional machine learning methods (refer to Table 2). Additionally, variations in the samples used across different studies were observed. Table 3 offers a detailed summary of the unpublished datasets used in depression diagnostic studies, including the male-to-female ratio, subjects’ mean age or age range, diagnostic criteria for MDD, data collection sources, and associated studies. As shown in Table 3, the diagnostic methods used in various studies to ascertain the status of MDD patients vary. These methods include the DSM-IV, HAM-D, International Classification of Diseases (ICD), Emotional State Questionnaire (EST-Q), Mini-international Neuropsychiatric Interview (MINI), and Patient Health Questionnaire-9 (PHQ-9), among others [11,91,92,93,94,95,96,97]. Similarly, the state of the participant during the experiment—whether their eyes are open, closed, or both—can significantly influence the EEG signals captured. Additional parameters influencing brain activity include age, sex, the time of day the experiment is conducted, and prior physical activity. Painkillers, antidepressants, or any other medications that significantly alter brain activity can influence the outcomes of the experiments. Similarly, studies vary in terms of participants’ medication use: some require all participants to have ceased any medication use at least six weeks before the experiment; others enroll participants who are receiving medication for the first time; and some include participants who are already on medication. Some studies fail to report participants’ medication use. When selecting subjects, it is advisable to include individuals who are at least 18 years old and have a minimum education level of middle school. Subjects with MDD must meet both DSM-IV and criteria, whereas healthy controls must have no history of mental disorders. Common exclusion criteria for participation include the following: (1) severe physical disabilities that preclude completion of the experiment; (2) diagnoses of other mental or neurological disorders; (3) histories of brain injuries that have resulted in coma; and (4) consumption of alcohol, nicotine, or caffeine prior to the experiment.

3.2. Data Acquisition

The process of data acquisition represents the initial stage of EEG research, and data collection methods vary across studies. Next, the included studies are compared based on parameters such as the number of electrodes, sampling rate, and other relevant factors. Initially, Ahmadlou’s team [37] use 19 electrodes to record rsEEG signals from both depressed patients and healthy controls with their eyes closed for three minutes, and the sampling frequency is 256 Hz. Numerous subsequent investigations adopt identical electrode quantities and sampling frequencies, with only slight deviations observed in the sampling duration. As EEG acquisition techniques evolve, a growing number of researchers enhance both channel count and sampling rates to acquire more granular data on rsEEG signals [41,47,49,59]. Yan et al. [69,70] utilize sampling rates of up to 1000 Hz to record EEG signals in their experiments.
Research has shown that effective classification tasks can be accomplished using fewer electrodes [68]. This reduction conserves not only time and computational resources, but it also decreases susceptibility to noise. Fewer electrodes mean a simpler process, which will facilitate the expansion of EEG datasets in this field. Given the strong correlation between the frontal lobe and emotional processes [99], Cai et al. [38] employ a three-electrode frontal system to collect 90 s of data, using the sampling frequency of 250 Hz. Sakib et al. [48] record data from only 14 channels using a wireless EEG headset. Some research teams employ multiple sampling rates for various signal processing methods during experiments to enhance data reliability [40,42].
Additionally, the placement of the electrodes is of equal importance. The 10–20 system for electrode placement remains the standard method prescribed by the International Society for Electroencephalography, and it is widely utilized in EEG studies [100,101,102]. Most researchers adhere to the 10–20 international standard system for electrode placement during data acquisition, as illustrated in Figure 2. Table 4 summarizes the electrodes placed in different regions of the brain according to the 10–20 international standard system. The various brain regions utilized for EEG acquisition are depicted in Figure 3 [103]. An increasing number of researchers are focusing on EEG signals from specific brain regions. Acharya et al. [64] and Ay et al. [66] analyze EEG recordings from four electrodes, located in the left and right hemispheres of the brain. During data collection, Wan et al. [72] select several representative brain regions in the prefrontal cortex (PFC), frontal cortex, and parietal cortex—regions known for their strong association with depression—for EEG collection. To investigate inter-hemispheric asymmetry in patients with MDD, Duan et al. [35] utilize 28 pairs of electrodes to conduct experiments across five brain regions. The EEG signals used in the studies conducted by Wu’s team [74] and Xu’s team [73] are primarily collected from the frontal regions of the brain.

Analysis

The frequency of EEG signals typically lies below 100 Hz. According to Nyquist’s theorem, when the sampling frequency is more than twice the highest frequency of the signal, the sampled signal can retain the complete information of the raw signal. An examination of the depression diagnostic studies listed in Table 1 reveals that nearly all of them maintain a sampling rate exceeding 200 Hz. However, the indiscriminate use of a high sampling rate to garner more information may not be advisable. Processing data at high sampling rates demands increased computational resources, thereby substantially extending training times, particularly for deep learning models. Increased data volumes significantly raise storage needs and preprocessing times, potentially causing bottlenecks in both feature extraction and training phases. Additionally, higher-dimensional data from increased channel counts may increase the risk of model overfitting. Given that multi-channel data often contain substantial redundant information, models are likely to capture noise and irrelevant features rather than the intended signal patterns. Handling high-dimensional data often requires more complex model architectures, thereby increasing both the complexity of model training and the difficulty of tuning. In practice, researchers must balance sampling rates, channel counts, and model complexity to optimize the performance within the constraints of the available resources.

3.3. Preprocessing

Noise and artifacts must be filtered from EEG data prior to analysis [104]. EEG data inherently exhibit instability, weakness, and high susceptibility to external interference. Errors in experimental setups, environmental noise, and artifacts from other biological signals adversely impact the overall signal. Preprocessing constitutes a key component of EEG signal processing, significantly enhancing the signal-to-noise ratio and establishing a reliable foundation for further analyses and interpretation [105]. This article reviews a range of preprocessing methods employed in the included studies, including filtering, independent component analysis (ICA), data segmentation, Z-score and min–max normalization, principal component analysis (PCA), re-referencing, soft-thresholding algorithms, sphere spline, FastICA, adaptive autoregression (AAR) models, adaptive noise cancellation (ANC), pseudo-subspace reconstruction (PSR), discrete and multilevel discrete wavelet transforms (DWT and MDWT), fast Fourier transform (FFT), among others. Table 5 summarizes the preprocessing methods used in studies that employ both TML and DL approaches. The left column enumerates the various preprocessing methods, while the right column provides details of the studies that implemented these methods. Figure 4 graphically presents the frequency distributions of various preprocessing methods, with the numbers after the colon (“:”) denoting the usage frequency of each method. Furthermore, the thickness of each ribbon strip corresponds proportionally to its frequency.
As indicated in Table 5 and Figure 4, filtering is the most commonly employed preprocessing method. In all included studies on depression diagnosis, filtering was employed 71 times: 36 times in studies employing traditional machine learning and 35 times in those utilizing deep learning approaches. Various filter types employ specific window sizes and cutoff frequencies. Considering that the power line frequency in China is 50 Hz, researchers frequently use 50 Hz notch filters to reduce power line interference in recorded EEG signals. Bandpass, high-pass, and low-pass, each characterized by varying cutoff frequencies, are frequently employed as well. Kalman filters, adaptive filters, smoothing filters, and Hanning filters remain in the exploratory phase of application in this field. ICA is also a frequently employed method, utilized eight times in studies based on TML and nine times in those based on DL approaches. In contrast, FastICA was employed only twice in studies using deep learning approaches, as shown in Figure 4. This usage demonstrates FastICA’s faster convergence compared to traditional ICA and its suitability for deep learning tasks. ICA and FastICA are widely employed to isolate and remove disturbances attributable to eye and muscle movements [106,107,108,109]. Decomposing the EEG signal into its constituent components allows for the isolation and removal of artifacts by identifying and excluding these components [110]. A notable drawback of this artifact removal technique is its potential to alter the inherent signal dynamics. Research indicates that minimal preprocessing of physiological signals retains more of their original information [111]. To minimize signal alteration, some researchers engage EEG specialists who visually identify artifact-free segments for subsequent analysis [42].
Typically, the acquisition of EEG data requires special equipment and professional operation. Consequently, EEG data samples from depressed patients are often limited, especially in medical research, where acquiring large volumes of labeled data for depression diagnosis is challenging. Additionally, EEG data are characterized by nonlinearity, time variability, and sensitivity to noise, and must be processed prior to training deep learning models to avoid overfitting. Given the high temporal resolution of EEG signals, many researchers have employed data augmentation methods to improve the utility of limited data [112,113,114]. As illustrated in Figure 4, the method has been utilized in both TML and DL research contexts. Data segmentation, a technique frequently used in data augmentation, involves dividing a long EEG signal into multiple short time windows, thereby significantly increasing the amount of data available [115,116]. Segmentation improves the model’s focus on local features of the time series, thereby enhancing its performance. Researchers must consider the length of the segmentation window; a window that is too short risks losing global information, while one that is too long may not significantly increase the sample size. Some researchers have segmented EEG into specific time intervals using data cropping [81,117], a technique that enables researchers to concentrate on the electrical activity of the brain during certain key time windows. However, it is essential that data cropping ensures the selected time segments are sufficiently informative; otherwise, non-informative or noisy segments may be inadvertently included.
Z-score normalization is widely used to standardize EEG data across various channels or time periods to a consistent scale, thus eliminating discrepancies arising from varying amplitudes or baseline deviations [118,119]. Normalization improves the visibility of abnormal values within EEG signals, thereby facilitating the identification of atypical brain activity features. However, this method exhibits high sensitivity to outliers, such as noise from motion artifacts or electrode detachment. These outliers can substantially affect the calculation of means and standard deviations, consequently impacting the standardized results. Additionally, several studies have utilized specialized software and libraries to mitigate various types of noise, including Brain Electric Source Analysis (BESA), EEGLAB, Curry 7, Open MEEG toolbox, Brainstorm, ICLabel and MARA plugins, and MNE-Python.

Analysis

The preprocessing methods outlined above can be divided into two primary categories: automatic and semi-automatic or manual. In automatic preprocessing methods, tools such as EEGLAB and BESA rapidly process large volumes of EEG data, minimizing human intervention with techniques such as filtering, ICA, and normalization. This approach is especially well-suited for large-scale data analysis. Automated methods ensure consistency in preprocessing steps. For example, the use of uniform filters and standardization algorithms produces reproducible results across experiments, thereby enhancing the study’s reliability. FastICA and ANC effectively extract and remove artifacts and enhance signal-to-noise ratios, thereby improving signal quality. Typically, these methods can rapidly identify and eliminate artifacts caused by eye movements or other biological signals. However, automatic preprocessing can alter the intrinsic dynamics of the signal, especially when techniques such as ICA are employed [106]. Incorrect component identification may lead to the loss of crucial signal features. Among semi-automatic or manual preprocessing methods, the manual approach allows researchers to tailor the preprocessing strategy to specific experimental requirements. Expert visual inspection ensures the accuracy and integrity of the signal and minimizes potential distortions. Expert intervention facilitates more efficient identification and handling of outliers, thereby preventing the omission of important signals that may be overlooked by automated processing. This approach generally preserves more information and minimizes alterations to the signal. Although semi-automatic or manual methods can enhance data quality, insufficient sample sizes may lead to both the training and test sets containing the same individuals, thus impairing the model’s generalization capabilities. In conclusion, automatic preprocessing methods provide significant advantages in terms of efficiency and consistency, making them ideal for large-scale data analyses. However, they may also alter the signal characteristics. Conversely, semi-automatic or manual methods offer enhanced flexibility and improved control over signal quality, thereby allowing for the retention of more useful information.
Although there are many similarities in EEG signal preprocessing between TML and DL approaches, notable differences in method selection and data processing strategies remain. In terms of method selection, deep learning predominantly employs sophisticated, automated techniques like ASR and ANC to effectively manage high-dimensional and complex data [120]. In contrast, traditional machine learning methods primarily employ classical signal processing techniques like PCA and Kalman filters, which concentrate on the statistical properties of signals and linear transformations. In terms of data processing methods, the transformation of EEG signals into images for convolutional neural network (CNN) analysis represents a novel approach, which is seldom employed in TML. This method underscores the capabilities of DL to effectively manage unstructured data, such as images and videos. In contrast, traditional machine learning prioritizes specific filtering techniques and models, such as Kalman filters and adaptive prediction filters, which are tailored for precise signal modeling and prediction. These differences highlight the unique needs and strengths of each approach in addressing data complexity and ensuring adaptability.
In summary, deep learning methods increasingly depend on automation and advanced techniques during preprocessing to enhance data quality and improve model performance. Traditional machine learning methods, known for their conciseness and efficiency, are particularly well-suited for environments with limited computing resources. These methods concentrate on fundamental signal characteristics and employ straightforward transformations during the preprocessing phase.

3.4. Feature Extraction

3.4.1. Studies Based on TML Methods

Although EEG signals encompass a broad spectrum of information, the specific aspects related to depression remain unclear. Therefore, the signal must be optimized through the extraction of features that accurately represent the relevant information [121,122]. Feature extraction entails generating features by discerning hidden patterns within the input signal. In the reviewed literature, a range of methods have been employed to extract and select diverse types of features for diagnosing depression. Table 6 summarizes the various categories of features extracted from studies that employ traditional machine learning methods. As shown in Table 6, the investigated features are classified into five categories: time-domain, frequency-domain, nonlinear, connectivity, and generated features. Various methods have been utilized to extract and select diverse features for the diagnosis of depression, including FFT, DWT, MDWT, Neighborhood Component Analysis (NCA), Fourier–Bessel Series Expansion (FBSE), histogram-based feature generation, parametric modeling with autoregressive models, and detrended fluctuation analysis (DFA), among others.
Several researchers have introduced methods for automatic feature extraction. Several researchers have developed a graph where the nodes represent subjects within the dataset, employing the Node2vec algorithm to compute feature representations as node embeddings [44,45]. These node embeddings serve as valuable features readily applicable to classification algorithms. Aydemir’s team [54] pioneers the development of a molecular shape-based feature generator, termed the “melamine pattern”. In the approach, the melamine pattern is innovatively integrated with the DWT to generate texture features, also known as spatial domain features, significantly enhancing the classification accuracy. Tasci et al. [62] decompose the original signal using the Daubechies Four (DB4) parent wavelet function. They introduce an innovative Twin Pascal Triangle Lattice Pattern (TPTLP) approach to extract texture features from EEG signals.

3.4.2. Studies Based on DL Methods

TML methods depend on the manual design and selection of features, requiring the expertise and experience of domain specialists. Numerous researchers have attempted to remove this limitation by applying DL models to the field. During data processing, features are progressively extracted and integrated across multiple layers of the model, ultimately leading to a final decision or prediction based on these automatically extracted features [105]. However, deep learning models are often described as “black-box“, making their internal feature representations difficult to interpret. Furthermore, raw EEG signals may include information extraneous to depression, potentially impairing model performance if used directly as input. As a result, some researchers strive to integrate manually extracted features into deep learning models. In EEG signal processing, brain waves are typically categorized into four primary frequency bands: δ, θ, α, and β [123,124,125]. Each frequency band is associated with specific brain states and activities. Several researchers have focused their research on these frequency bands to enhance model performance through the extraction of structural features and power spectral density. Inspired by advances in deep learning for computer vision, several researchers have transformed EEG into image form, capitalizing on the strengths of DL in image processing to enhance EEG signal analysis. However, this approach could result in potential information loss, heightened processing complexity, and an increased risk of overfitting. Furthermore, given that the brain functional connectivity analysis reveals the interactions and synergistic activities among various brain regions [126,127,128], several researchers have combined brain functional connectivity matrices with feature fusion techniques. This approach increases the comprehensiveness of raw EEG signal information and provides substantial noise immunity. Furthermore, several researchers have utilized a range of techniques for EEG signal analysis, including Partial Directed Coherence (PDC), and the integration of temporal and spatial features, as well as both distance-based and non-distance-based projection methods. Table 7 summarizes the diverse feature extraction methods employed in combination with deep learning techniques.

3.4.3. Analysis

An analysis of Table 6 shows that the features extracted in studies employing traditional machine learning methods are primarily time-domain features, with skewness and kurtosis being the most frequently utilized, followed by the mean. Among the nonlinear features, entropy calculation is a widely used technique, with approximate entropy and Shannon entropy also being frequently employed by researchers. Regarding the use of entropy, the EEG diagnosis of depression using DL methods has not yet been explored. Conversely, band power is the most frequently utilized feature in the frequency domain. Both approaches emphasize the critical role of functional connectivity in elucidating the brain’s functional networks. The application of molecular structure maps and lattices in EEG-based depression diagnosis represents an exceptionally innovative approach, providing significant potential for future exploration. As detailed in Table 7, studies employing deep learning methods frequently combine diverse feature extraction techniques with various deep learning models, including CNN, DAN, LSTM, GNN, and MLP.
Features extracted through traditional machine learning methods are typically low-level and directly derived from the original signal. Deep learning methods can extract high-level features across multiple layers, producing more abstract and complex representations that significantly improve model performance. Traditional machine learning features are crafted for straightforward interpretation, while deep learning models exhibit greater complexity and reduced interpretability. In their experiments, researchers often employ visualization techniques, including activation maps and feature maps, to aid interpretation. In summary, while deep learning offers robust capabilities for automatic feature extraction and learning in EEG signal processing, its complexity and the significant computational resources it demands must also be considered.

3.5. Feature Selection

Feature selection reduces feature dimension to remove redundant or irrelevant features. The extracted features may be redundant or irrelevant. Utilizing all of these as inputs to the classifier could potentially decrease its accuracy. To improve classifier accuracy, key features must be selectively identified and utilized in the analysis [129]. Table 8 presents a summary of the feature selection methods employed in EEG-based depression diagnosis studies. According to Table 8, Neighborhood Component Analysis (NCA) emerges as the most frequently employed method. The primary objective of Neighborhood Component Analysis (NCA) is to enhance classification performance through the development of a feature transformation matrix. This matrix aims to bring samples from the same category closer together, and to separate samples from different categories within the transformed feature space. However, NCA exhibits significant computational complexity, demonstrates sensitivity to noise and outliers, and is optimally suited for analyzing linear feature relationships. Additionally, methods such as the genetic algorithm (GA), ReliefF, Node2vec, and others are commonly utilized.
Certain DL models, including CNN, LSTM, and Recurrent Neural Networks (RNNs), automatically filter and prioritize input features through their inherent structure and training process. This method effectively highlights useful features while minimizing the impact of redundant or irrelevant ones. Additionally, certain deep learning models employ regularization techniques, such as L1 and L2 regularization, Dropout, and Batch Normalization, to mitigate overfitting and indirectly facilitate feature selection. Visualization tools, such as Grad-CAM and t-SNE, facilitate researchers’ interpretation of model features and regions of interest, thereby enhancing feature selection and analysis.
In summary, deep learning-based methods eliminate the necessity for manual feature selection and excel at processing large-scale, high-dimensional, and complex data. However, the feature selection process in deep learning models is frequently viewed as a black-box, complicating efforts to interpret the decision-making logic within the model. Feature selection continues to be a critical area of research, substantially affecting classification performance and playing a crucial role in subsequent data analyses.

3.6. Classification Methods

3.6.1. Studies Based on TML Methods

Table 9 summarizes the performance of each algorithm in studies of depression diagnosis that use traditional machine learning methods. From left to right, the table lists the algorithm used, the validation strategy (which details the proportion of the dataset allocated to training, testing, and validation), the top-performing classifier, the accuracy given as a percentage, and the associated study. The choice of classifiers significantly influences the performance of the models, with various studies employing distinct classification methods. Currently, a diverse array of traditional machine learning algorithms is systematically applied to the diagnosis of depression and categorized into seven primary groups:
  • Support Vector Machine-Based Classifiers: this category includes the Support Vector Machine (SVM) with various kernels and Least Squares-SVM (LS-SVM).
  • Tree-Based Structures: features methods such as Decision Tree (DT), Best-First Tree (BF-Tree), and Coarse Tree.
  • K-nearest neighbor (KNN) and variants: encompasses KNN-GA and E-KNN.
  • Ensemble Learning Methods: comprises Bagging, Extreme Gradient Boosting (XGBoost), GentleBoost (GB), RusBoost (RB), Random Forest (RF), and Adaptive Boosting (AdaBoost).
  • Probabilistic models: includes well-known methods such as Naïve Bayesian (NB) and Logistic Regression (LR).
  • Discriminant Analysis-Based Classifiers: features Linear Discriminant Analysis (LDA).
  • Other methods: includes approaches such as K-means Singular Value Decomposition (K-SVD) and Linear Regression.
Figure 5 illustrates the frequency distribution of depression diagnosis algorithms that utilize traditional machine learning, with the numbers following the colon (“:”) indicating the frequency of usage for each method.

Analysis

This review mainly analyzes the classification accuracy of various studies. Classification accuracy represents the percentage of all participants who were correctly classified as either depressed or in the control group. As illustrated in Figure 5, KNN and SVM were the classifiers most frequently used in the included studies on depression diagnosis, with KNN appearing 15 times and SVM 18 times. As shown in Table 9, KNN and SVM emerged as the top-performing classifiers in several studies. On non-public datasets, the KNN classification demonstrated a strong performance, with accuracies ranging from 76.83% to 98.43%. On public datasets, the KNN classification yielded superior results, surpassing other classifiers such as BF-Tree and K-SVD in studies using the MODMA dataset [61] and the Mumtaz team’s dataset [39,50,51], achieving peak classification accuracies of 100% and 99.11%, respectively. This reflects the model’s exceptionally high performance under specific conditions; however, this does not rule out the possibility of sample leakage, which could artificially inflate the classification accuracy. SVM also exhibits an outstanding performance, with accuracy rates ranging from 83.67% to 98%. Probabilistic models like LR, as well as LDA, have also been explored somewhat in this area; however, none have emerged as the top-performing classifiers in any of the studies reviewed. Furthermore, ensemble methods have exhibited high accuracies, underscoring the advantages of integrated learning approaches for complex tasks, achieving a maximum accuracy rate of 99.11%. As shown in Figure 5 and Table 9, among tree-based classifiers, Decision Tree is not only the most frequently used algorithm, but it also ranks as the highest-performing classifier in several studies, with accuracies reaching up to 95%.
In summary, the accuracies reported in the referenced studies varied from 76.83% to 100%. For traditional machine learning methods, the classification accuracy depends more on the selected features than on the choice of classifier [130]. Many studies focus on examining the impact of diverse feature extraction methods on model performance. Some studies aim to reduce system complexity by minimizing the number of channels used, while still preserving accuracy. Although many studies report respectable classification accuracies, they often involve a small number of participants, which may limit the model’s generalizability and its applicability in clinical application.

3.6.2. Studies Based on DL Methods

Table 10 summarizes the performance of various algorithms used in DL-based depression diagnosis studies. From left to right, the table lists the algorithms employed, the validation strategy (which details the percentage of the dataset allocated to training, testing, and validation), the top-performing classifier, the accuracy presented as a percentage, and the associated study. Deep learning algorithms for diagnosing depression in the reviewed studies are categorized into four primary groups:
  • CNN and variants, including CNN, 2D-CNN, CNN-LSTM, Lightweight-CNN, AchCNN, and EEGNet;
  • Graph Convolutional Network (GCN) and variants, such as GCN, MGGCN, GICN, and AMGCN;
  • Recurrent Neural Network (RNN) and variants, including RNN, 4D-CRNN, GRU, and GC-GRU;
  • Inception-based models and variants, notably Inception, InceptionTime, InceptionNet, and Tsception.
Among these, numerous hybrid models are identified, including DeprNet [65], AchLSTM [66], AchCNN [64], T-LSTM [131], H-KNN1 [38], H-KNN2 [132], S-EMD [133], S-SVM [134], H-DBN [135], EEGNet [136], DeepConvNet [137], ShallowConvNet [137], LSDD-EEGNet [70], HybridEEGNet [72], SynEEGNet [72], RegEEGNet [72], MRCNN-RSE [73], MRCNN-LSTM [73], DeepCoral [138], VGG16 [139], AlexNet [139], ResNet50 [139], MV-SDGC-RAFFNet [84], TS-SEFFNet [140], 4D-CRNN [141], DBGC-ATFFNet [142], ASTGCN [143,144], DepHNN [145], TSUnet-CC [87], DiffMDD [88], 1D-CNN-Transformer [66,145], CWT-1D-CNN [146], CWP-2D-CNN [146], GRU-Conv [147], and GC-GRU [148]. Figure 6 illustrates the frequency distribution of DL-based algorithms used in diagnosing depression, with the numbers following the colon (“:”) indicating the usage frequency of each method.

Analysis

As shown in Figure 6, CNN and its variants constitute the most commonly employed deep learning method in the research. An analysis of Table 10 shows that most studies utilizing CNN and its variants (e.g., 2D-CNN, CNN-LSTM) exhibit an excellent performance. In studies using publicly available datasets from Mumtaz’s team [39,50,51,89] and Cavanagh’s team [60], CNN variants (2D-CNN and Lightweight-CNN) surpassed other classifiers in attaining the highest classification accuracy. Among these, the 2D-CNN is particularly notable, with Khan’s team [80] achieving a 100% classification accuracy in the study, which highlights its exceptional capability in EEG signal processing. The accuracy of Lightweight-CNN has reached 99.87% and reduced computational resource consumption while preserving high accuracy, rendering it adaptable for a wide range of applications. The CNN-LSTM model merges the feature extraction capabilities of CNN with the time series processing advantages of LSTM, attaining an accuracy of up to 99.9%. LSTM variants excel in processing EEG signals but prove relatively ineffective when used independently. Hybrid models, including MV-SDGC-RAFFNet and TSUnet-CC, which integrate the strengths of multiple models, achieve accuracies of 99.19% and 99.22%, respectively, highlighting their potential for future hybrid model development. The application of GCN and its variants in EEG signal processing has shown excellent accuracy (see Table 10). In studies utilizing the public dataset MODMA [61], MGGCN delivered the highest performance, achieving an accuracy of 99.69% [76]. The application of GCN and its variants in EEG signal processing merits additional exploration. MLP delivered the highest performance in the study utilizing the public dataset EDRA.
In summary, deep learning methods consistently outperform traditional machine learning techniques in classification accuracy, demonstrating substantial potential for EEG-based automatic depression diagnosis. However, the black-box-like properties of DL models complicate the interpretation of their decision-making processes, potentially reducing their trustworthiness in clinical applications. Studies utilizing various publicly available datasets have yielded diverse optimal classifiers, indicating that the existing findings lack sufficient generalizability. Gathering more heterogeneous data and performing additional experimental comparisons will help address this issue.

3.7. Validation Methods

To detect overfitting and ensure model generalization, it is essential to evaluate both traditional and deep learning models using suitable validation techniques. Figure 7 depicts the frequencies of various validation methods used in depression diagnostic studies, with the numbers above the bars showing the frequency of each method. As depicted in Figure 7, among the included depression diagnostic studies, the n-fold cross-validation (CV) is the most frequently employed method by researchers, succeeded by the LOOCV. The n-fold CV [149,150,151], typically configured for 5 or 10 folds, is the approach most commonly employed by researchers (refer to Table 9 and Table 10). This method is the most commonly utilized validation technique, and numerous models in this study exhibit a strong performance when assessed using this approach. Specifically, across n iterations, each sample is included n − 1 times in the training set and once in the validation set. This method effectively utilizes the entire dataset and mitigates the issue of small sample sizes prevalent in many studies. However, n-fold CV presents certain drawbacks, including significant computational costs attributed to multiple training and evaluation cycles. Additionally, if the dataset size is not divisible by n, this can lead to varying fold sizes, potentially impacting the accuracy of the evaluation results. To address this issue, several studies have chosen to partition the dataset on a per-subject basis. However, this validation method entails the risk that certain sample types may exclusively appear in either the training or the testing set, especially in cases of imbalanced datasets. Wan’s team [40] implemented a no-putback sampling strategy in their study, ensuring consistent sample sizes for both the depressed and normal control groups, thereby addressing the issue of model performance degradation due to sample imbalance.
To optimize information utilization from each sample, several studies have implemented LOOCV [152] and its variants, including LOSOCV and LOPOCV [47,58,62]. LOPOCV and LOSOCV are well-suited for datasets with substantial individual differences, while LOOCV is more appropriate for datasets with small sample sizes. Additionally, one study [54] implements the hold-out validation strategy [153], which entails randomly partitioning the dataset into two mutually exclusive subsets: one designated for training and the other for testing. If the test set is too small, the evaluation results might be unstable and lack accuracy; conversely, an overly large test set can undermine the realism of model training. Furthermore, hold-out validation can introduce bias from the data partitioning process, particularly when significant differences exist between the sample proportions in the training and testing sets. Maintaining data independence is crucial when employing these methods to preserve the accuracy and reliability of model assessments. However, owing to limited sample sizes in many studies, the test and training sets frequently lack independence, potentially resulting in data leakage. Wu’s team [41] mitigates this issue by implementing 5-fold CV and validating the model with an independent test set, thus minimizing the risk of data leakage. A research team [63] has used the nested CV strategy [154], effectively reducing data leakage by separating parameter selection from performance evaluation, thus providing a more robust assessment of the model’s generalization capability. However, this strategy incorporates two layers of cross-validation, which may complicate its implementation.
When selecting a validation strategy, researchers should base their decisions on factors including dataset size, computational resources, model complexity, and evaluation requirements. For large-scale datasets constrained by limited computational resources, hold-out CV provides a rapid and straightforward method. For medium-sized datasets, k-fold CV, such as 5-fold or 10-fold, ensures a more stable evaluation. When dealing with small datasets and ample computational resources, LOOCV optimizes data utilization. For complex models and hyperparameter tuning, nested CV is the preferred method, providing low-bias model evaluations and optimized generalization error estimations. In scenarios involving unbalanced datasets, stratified k-fold or stratified nested CVs may be utilized to ensure consistent sample proportions across categories. Future research should emphasize the utilization of large datasets and independent samples wherever feasible.

4. Discussion

Researchers in existing studies commonly utilize two types of methods: TML and DL. Numerous studies report substantial accuracy in diagnosing depression using rsEEG. Given the distinctions between the two research methods, we opted to investigate the commonalities among these studies.
To begin with, accurately identifying depression is the first challenge that researchers must address. Episodes of MDD often coexist with other disorders, such as bipolar disorder and anxiety disorders [155]. The clinical manifestations of major depression and bipolar disorder are challenging to distinguish [156,157,158]; moreover, most individuals with bipolar disorder also experience phases of major depression [159,160]. Recent research has revealed that resting-state connectivity biomarkers are capable of defining neurophysiological subtypes of depression [161]. Future research may explore solutions to this issue through EEG-based, data-driven approaches [162].
Various studies may utilize differing diagnostic tools and criteria, leading to heterogeneity in EEG parameters across subjects. Consequently, identifying reliable biomarkers for MDD based on EEG signals is crucial [163,164,165]. Several researchers have investigated the feasibility of employing wavelet coherence between regions of DMN as a biomarker [80]. Moreover, research has shown that both effective connectivity and functional connectivity among various brain regions are essential for accurately distinguishing between depressed patients and healthy participants [79,86,166,167].
Currently, numerous studies employing TML and DL methods with EEG data face the pervasive issue of insufficient sample sizes. Among the studies included in the comparative analysis, only four featured a sample size exceeding 200 cases. Most studies exhibited small to moderate sample sizes and relied on limited sample sources. A large sample size is typically not derived from a single location; rather, data are gathered from multiple sources, including hospitals, clinics, neuroscience research institutes, clinical trials, research programs, and universities. Furthermore, the range of subjects’ occupations could be broadened to include IT workers, doctors, video bloggers, business managers, laborers, and sales personnel. Recently, advancements in portable EEG recording devices have streamlined the collection of EEG data [168,169]. For instance, wireless EEG caps such as the Epoch, ENOBIO Neuroelectrics, and iMotions not only offer portability and comfort, but they also facilitate real-time data transmission. The enhancement of public databases and the development of standardized datasets have offered essential support for research in depression diagnostics. Mumtaz’s team [39,50,51,89], Cavanagh’s team [60], Cai’s team [61], Yang’s team [90], and researchers including Wang et al. [98], and Kang et al. [77] have publicly shared their datasets or code, thereby facilitating knowledge sharing. Many researchers have expanded upon these contributions. We encourage researchers in this field to share their data and code. Doing so not only promotes scientific transparency and reproducibility, enhances research credibility, and facilitates the verification and reproduction of results by others, but it also fosters cooperation and communication, thereby advancing technological innovation in this area [170].
The limited availability of shared datasets in this field can be partly attributed to the complexity of EEG signal acquisition. In clinical practice, EEG signals are typically acquired using traditional wet electrodes [171]. However, the requirement to apply a conductive medium, such as conductive paste, between the electrodes and the scalp complicates the procedure and creates discomfort for the subject. Dry and semi-dry electrodes offer a more efficient alternative for EEG signal acquisition, especially in situations where professionals are unavailable [172]. Currently, newly developed semi-dry gel electrodes offer greater comfort compared to wet electrodes, and they address the issue of rigid dry electrodes not adhering stably to the skin [173,174]. This advancement simplifies the EEG collection process and enhances the multifunctionality of the technology across a range of scenarios. The widespread adoption of gel electrodes could significantly enhance EEG signal acquisition.
KNN and SVM are the classifiers most frequently utilized in studies employing traditional machine learning methods. Meanwhile, studies that utilize deep learning methods frequently employ various types of CNNs and their variants. These models have demonstrated a robust performance in diagnosing depression, achieving exceptionally high accuracy rates. To enhance the generalization capabilities of the models, researchers frequently employ n-fold CV and LOOCV methods for evaluation. However, most studies were not tested with independent samples, and some lacked the explicit reporting of validation information, casting doubt on the reliability of the model results. Future studies should ensure the complete independence of training and test sets and should transparently report their distributions to affirm the reliability and clinical applicability of the model results. Generalization ability denotes a model’s capacity to predict and process new data beyond the training dataset. Consequently, external validation is essential. Among the included studies on depression diagnosis, none were found to have developed a model on one dataset (its own dataset) and validated it on another (a public dataset). Implementing such a process would offer a robust test of the model’s generalizability. Only a few existing studies have considered using multiple datasets, and they mostly rely on commonly used public datasets (e.g., MODMA [61]), potentially affecting model robustness. The subsequent phase entails the development of models on proprietary datasets, followed by their validation on datasets that are publicly accessible. Furthermore, the model must be interpretable [175]. With highly interpretable models, researchers can determine which EEG features, including activity in specific frequency bands, most significantly influence diagnostic outcomes. This understanding facilitates the more precise identification of depression biomarkers, offering essential insights for advancing neuroscience research. Enhancing model interpretability presents a substantial challenge for deep learning-based approaches. Researchers must meticulously select algorithms based on their intended applications. The development of visualization tools and interpretative algorithms can aid clinicians in comprehending and applying the results derived from studies employing CNN architectures. Moving forward, it is imperative to intensify comparative studies among various models to identify optimal solutions that improve current clinical practices and provide more effective treatment strategies for patients with depression.
Presently, methods that diagnose depression via a single EEG signal still encounter significant limitations regarding accuracy and clinical applicability. Multimodal approaches to EEG are generally more effective than unimodal methods, as computer vision techniques can extract features more efficiently by leveraging multiple models, rather than relying on a single model [132,176,177,178,179]. Future research should investigate the integration of multimodal data, incorporating a variety of sources including speech, facial expressions, text, images, eye tracking, RNA-Seq, functional magnetic resonance imaging (fMRI), and other modalities to enhance the diagnostic accuracy and generalizability [18,164,180,181,182,183,184,185,186,187,188,189,190,191].
Classifying the subjects as those with MDD or HC represents only the preliminary step in depression research employing EEG signals. The development of predictive models for depression treatment outcomes and relapse is crucial for preventing ineffective treatment strategies [192,193,194,195,196,197,198]. Furthermore, the use of IoT technology in developing personalized medical models, which take into account individual differences and clinical characteristics, offers considerable potential for providing precise diagnoses and tailored treatment plans for patients with depression [199,200].
As this research advances, computational intelligence-based diagnostic systems for depression are anticipated to exert a growing influence on clinical applications, offering more reliable and intelligent diagnostic and treatment solutions. Simultaneously, these technological advances are also expected to yield valuable insights into the diagnosis and research of other mental disorders.

5. Conclusions

The diagnosis of depression remains a significant and challenging issue in contemporary medical research. Diagnostic methods traditionally depend on clinical assessments and self-reports; however, these approaches often face challenges stemming from subjectivity and clinical bias. In recent years, the advent of computational psychiatry has catalyzed a growing number of studies that integrate electroencephalography with artificial intelligence, substantially enhancing the diagnosis of depression. In this review, we conduct a comparative analysis of 49 studies that utilize rsEEG in conjunction with both TML and DL techniques to diagnose depression.
EEG is capable of detecting specific brain physiological changes related to depression, offering a crucial objective foundation for depression diagnosis. TML methods are extensively employed in the diagnosis of depression, notably demonstrating significant advancements in feature extraction. Although feature extraction is complex and time-intensive, traditional machine learning offers superior interpretability in analyzing diagnostic results compared to deep learning methods. Commonly used traditional machine learning algorithms, including KNN and SVM, have demonstrated an excellent classification performance in numerous studies. Recently, research on the application of DL methods for diagnosing depression has expanded significantly. Deep learning demonstrates significant potential in processing raw EEG signals due to its capability for automatic feature extraction. CNN and its variants represent the most commonly utilized deep learning architectures. By integrating manually extracted features with these models, researchers further enhance the accuracy of depression diagnosis. However, deep learning-based studies still face limitations due to the scarcity of large-scale EEG datasets, and the challenge of model interpretability persists. Furthermore, validating these models using independent datasets is crucial to ensure their generalization capabilities. Going forward, researchers will focus more on multimodal data fusion methods. Currently, the integration of artificial intelligence and EEG in depression research offers new opportunities for precision medicine and preventive strategies. We anticipate a growing number of individuals benefiting from this advancement.

Author Contributions

Conceptualization, H.L. and J.F.; methodology, H.L., J.F. and J.Z.; formal analysis, H.L., J.F. and X.Z.; data curation, J.F. and J.Z.; writing—original draft preparation, H.L., J.F. and J.Z.; writing—review and editing, H.L., J.F., X.Z., W.P. and Y.L.; and supervision, H.L., J.F., W.P. and Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Heilongjiang Province, grant number [LH2022F028].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lang, U.E.; Borgwardt, S. Molecular mechanisms of depression: Perspectives on new treatment strategies. Cell. Physiol. Biochem. 2013, 31, 761–777. [Google Scholar] [CrossRef] [PubMed]
  2. Woody, C.A.; Ferrari, A.J.; Siskind, D.J.; Whiteford, H.A.; Harris, M.G. A systematic review and meta-regression of the prevalence and incidence of perinatal depression. J. Affect. Disord. 2017, 219, 86–92. [Google Scholar] [CrossRef] [PubMed]
  3. World Health Organization. Depression and Other Common Mental Disorders: Global Health Estimates; World Health Organization: Geneva, Switzerland, 2017. [Google Scholar]
  4. Bueno-Notivol, J.; Gracia-Garcia, P.; Olaya, B.; Lasheras, I.; Lopez-Anton, R.; Santabarbara, J. Prevalence of depression during the COVID-19 outbreak: A meta-analysis of community-based studies. Int. J. Clin. Health Psychol. 2021, 21, 100196. [Google Scholar] [CrossRef] [PubMed]
  5. Keynejad, R.; Spagnolo, J.; Thornicroft, G. WHO mental health gap action programme (mhGAP) intervention guide: Updated systematic review on evidence and impact. BMJ Ment. Health 2021, 24, 124–130. [Google Scholar] [CrossRef] [PubMed]
  6. Chmielewski, M.; Clark, L.A.; Bagby, R.M.; Watson, D. Method Matters: Understanding Diagnostic Reliability in DSM-IV and DSM-5. J. Abnorm. Psychol. 2015, 124, 764–769. [Google Scholar] [CrossRef]
  7. Bell, C.C. DSM-IV: Diagnostic and statistical manual of mental disorders. JAMA 1994, 272, 828–829. [Google Scholar] [CrossRef]
  8. Dev, A.; Roy, N.; Islam, M.K.; Biswas, C.; Ahmed, H.U.; Amin, M.A.; Sarker, F.; Vaidyanathan, R.; Mamun, K.A. Exploration of EEG-Based Depression Biomarkers Identification Techniques and Their Applications: A Systematic Review. IEEE Access 2022, 10, 16756–16781. [Google Scholar] [CrossRef]
  9. Rush, A.J.; Trivedi, M.H.; Wisniewski, S.R.; Nierenberg, A.A.; Stewart, J.W.; Warden, D.; Niederehe, G.; Thase, M.E.; Lavori, P.W.; Lebowitz, B.D.; et al. Acute and longer-term outcomes in depressed outpatients requiring one or several treatment steps: A STAR*D report. Am. J. Psychiatry 2006, 163, 1905–1917. [Google Scholar] [CrossRef]
  10. Bagby, R.M.; Ryder, A.G.; Schuller, D.R.; Marshall, M.B. The Hamilton depression rating scale: Has the gold standard become a lead weight? Am. J. Psychiatry 2004, 161, 2163–2177. [Google Scholar] [CrossRef]
  11. Hamilton, M. A rating scale for depression. J. Neurol. Neurosurg. Psychiatry 1960, 23, 56. [Google Scholar] [CrossRef]
  12. Zigmond, A.S.; Snaith, R.P. The hospital anxiety and depression scale. Acta Psychiatr. Scand. 1983, 67, 361–370. [Google Scholar] [CrossRef]
  13. Rush, A.J.; Trivedi, M.H.; Ibrahim, H.M.; Carmody, T.J.; Arnow, B.; Klein, D.N.; Markowitz, J.C.; Ninan, P.T.; Kornstein, S.; Manber, R. The 16-Item Quick Inventory of Depressive Symptomatology (QIDS), clinician rating (QIDS-C), and self-report (QIDS-SR): A psychometric evaluation in patients with chronic major depression. Biol. Psychiatry 2003, 54, 573–583. [Google Scholar] [CrossRef] [PubMed]
  14. Pichot, P. Self-report inventories in the study of depression. In New Results in Depression Research; Springer: Berlin/Heidelberg, Germany, 1986; pp. 53–58. [Google Scholar]
  15. Ben-Porath, Y.S. Assessing personality and psychopathology with self-report inventories. Handb. Psychol. Assess. Psychol. 2003, 10, 553–577. [Google Scholar]
  16. Pampouchidou, A.; Simos, P.G.; Marias, K.; Meriaudeau, F.; Yang, F.; Pediaditis, M.; Tsiknakis, M. Automatic Assessment of Depression Based on Visual Cues: A Systematic Review. IEEE Trans. Affect. Comput. 2019, 10, 445–470. [Google Scholar] [CrossRef]
  17. George, M.S.; Ketter, T.A.; Post, R.M. Prefrontal cortex dysfunction in clinical depression. Depression 1994, 2, 59–72. [Google Scholar] [CrossRef]
  18. de Kwaasteniet, B.; Ruhe, E.; Caan, M.; Rive, M.; Olabarriaga, S.; Groefsema, M.; Heesink, L.; van Wingen, G.; Denys, D. Relation Between Structural and Functional Connectivity in Major Depressive Disorder. Biol. Psychiatry 2013, 74, 40–47. [Google Scholar] [CrossRef]
  19. Du, L.; Liu, H.; Du, W.; Chao, F.; Zhang, L.; Wang, K.; Huang, C.; Gao, Y.; Tang, Y. Stimulated left DLPFC-nucleus accumbens functional connectivity predicts the anti-depression and anti-anxiety effects of rTMS for depression. Transl. Psychiatry 2018, 7, 3. [Google Scholar] [CrossRef]
  20. Michel, C.M.; Brunet, D. EEG Source Imaging: A Practical Review of the Analysis Steps. Front. Neurol. 2019, 10, 325. [Google Scholar] [CrossRef]
  21. Radenković, M.Č.; Lopez, V.L. Machine Learning Approaches for Detecting the Depression from Resting-State Electroencephalogram (EEG): A Review Study. arXiv 2019, arXiv:1909.03115. [Google Scholar]
  22. Jollans, L.; Whelan, R. The Clinical Added Value of Imaging: A Perspective From Outcome Prediction. Biol. Psychiatry. Cogn. Neurosci. Neuroimaging 2016, 1, 423–432. [Google Scholar] [CrossRef]
  23. Acharya, U.R.; Sudarshan, V.K.; Adeli, H.; Santhosh, J.; Koh, J.E.W.; Adeli, A. Computer-Aided Diagnosis of Depression Using EEG Signals. Eur. Neurol. 2015, 73, 329–336. [Google Scholar] [CrossRef] [PubMed]
  24. Pold, T.; Päeske, L.; Hinrikus, H.; Lass, J.; Bachmann, M. Temporal stability and correlation of EEG markers and depression questionnaires scores in healthy people. Sci. Rep. 2023, 13, 21996. [Google Scholar] [CrossRef]
  25. Crosson, B.; Ford, A.; McGregor, K.M.; Meinzer, M.; Cheshkov, S.; Li, X.; Walker-Batson, D.; Briggs, R.W. Functional imaging and related techniques: An introduction for rehabilitation researchers. J. Rehabil. Res. Dev. 2010, 47, vii–xxxiv. [Google Scholar] [CrossRef] [PubMed]
  26. Babiloni, C.; Pizzella, V.; Del Gratta, C.; Ferretti, A.; Romani, G.L. Fundamentals of electroencefalography, magnetoencefalography, and functional magnetic resonance imaging. Int. Rev. Neurobiol. 2009, 86, 67–80. [Google Scholar] [PubMed]
  27. Fingelkurts, A.A.; Fingelkurts, A.A. Altered Structure of Dynamic Electroencephalogram Oscillatory Pattern in Major Depression. Biol. Psychiatry 2015, 77, 1050–1060. [Google Scholar] [CrossRef]
  28. Baars, B.J.; Ramsoy, T.Z.; Laureys, S. Brain, conscious experience and the observing self. Trends Neurosci. 2003, 26, 671–675. [Google Scholar] [CrossRef] [PubMed]
  29. Koenig, T.; Prichep, L.; Lehmann, D.; Sosa, P.V.; Braeker, E.; Kleinlogel, H.; Isenhart, R.; John, E.R. Millisecond by millisecond, year by year: Normative EEG microstates and developmental stages. NeuroImage 2002, 16, 41–48. [Google Scholar] [CrossRef]
  30. Kaushik, P.; Yang, H.; Roy, P.P.; van Vugt, M. Comparing resting state and task-based EEG using machine learning to predict vulnerability to depression in a non-clinical population. Sci. Rep. 2023, 13, 7467. [Google Scholar] [CrossRef]
  31. Kato, A.; Kunisato, Y.; Katahira, K.; Okimura, T.; Yamashita, Y. Computational Psychiatry Research Map (CPSYMAP): A new database for visualizing research papers. Front. Psychiatry 2020, 11, 578706. [Google Scholar] [CrossRef]
  32. Maia, T.V. Introduction to the series on computational psychiatry. Clin. Psychol. Sci. 2015, 3, 374–377. [Google Scholar] [CrossRef]
  33. Huys, Q.J.M.; Maia, T.V.; Frank, M.J. Computational psychiatry as a bridge from neuroscience to clinical applications. Nat. Neurosci. 2016, 19, 404–413. [Google Scholar] [CrossRef] [PubMed]
  34. Tran, B.X.; McIntyre, R.S.; Latkin, C.A.; Phan, H.T.; Vu, G.T.; Nguyen, H.L.T.; Gwee, K.K.; Ho, C.S.; Ho, R.C. The current research landscape on the artificial intelligence application in the management of depressive disorders: A bibliometric analysis. Int. J. Environ. Res. Public Health 2019, 16, 2150. [Google Scholar] [CrossRef] [PubMed]
  35. Duan, L.; Duan, H.; Qiao, Y.; Sha, S.; Qi, S.; Zhang, X.; Huang, J.; Huang, X.; Wang, C. Machine learning approaches for MDD detection and emotion decoding using EEG signals. Front. Hum. Neurosci. 2020, 14, 284. [Google Scholar] [CrossRef] [PubMed]
  36. Khadidos, A.O.; Alyoubi, K.H.; Mahato, S.; Khadidos, A.O.; Mohanty, S.N. Machine Learning and Electroencephalogram Signal based Diagnosis of Dipression. Neurosci. Lett. 2023, 809, 137313. [Google Scholar] [CrossRef]
  37. Ahmadlou, M.; Adeli, H.; Adeli, A. Fractality analysis of frontal brain in major depressive disorder. Int. J. Psychophysiol. 2012, 85, 206–211. [Google Scholar] [CrossRef]
  38. Cai, H.; Han, J.; Chen, Y.; Sha, X.; Wang, Z.; Hu, B.; Yang, J.; Feng, L.; Ding, Z.; Chen, Y. A pervasive approach to EEG-based depression detection. Complexity 2018, 2018, 5238028. [Google Scholar] [CrossRef]
  39. Mumtaz, W.; Ali, S.S.A.; Yasin, M.A.M.; Malik, A.S. A machine learning framework involving EEG-based functional connectivity to diagnose major depressive disorder (MDD). Med. Biol. Eng. Comput. 2018, 56, 233–246. [Google Scholar] [CrossRef]
  40. Wan, Z.; Zhang, H.; Huang, J.; Zhou, H.; Yang, J.; Zhong, N. Single-channel EEG-based machine learning method for prescreening major depressive disorder. Int. J. Inf. Technol. Decis. Mak. 2019, 18, 1579–1603. [Google Scholar] [CrossRef]
  41. Wu, C.-T.; Huang, H.-C.; Huang, S.; Chen, I.-M.; Liao, S.-C.; Chen, C.-K.; Lin, C.; Lee, S.-H.; Chen, M.-H.; Tsai, C.-F. Resting-state EEG signal for major depressive disorder detection: A systematic validation on a large and diverse dataset. Biosensors 2021, 11, 499. [Google Scholar] [CrossRef]
  42. Avots, E.; Jermakovs, K.; Bachmann, M.; Päeske, L.; Ozcinar, C.; Anbarjafari, G. Ensemble approach for detection of depression using EEG features. Entropy 2022, 24, 211. [Google Scholar] [CrossRef]
  43. Li, Y.; Shen, Y.; Fan, X.; Huang, X.; Yu, H.; Zhao, G.; Ma, W. A novel EEG-based major depressive disorder detection framework with two-stage feature selection. BMC Med. Inf. Decis. Mak. 2022, 22, 209. [Google Scholar] [CrossRef] [PubMed]
  44. Soni, S.; Seal, A.; Yazidi, A.; Krejcar, O. Graphical representation learning-based approach for automatic classification of electroencephalogram signals in depression. Comput. Biol. Med. 2022, 145, 105420. [Google Scholar] [CrossRef]
  45. Soni, S.; Seal, A.; Mohanty, S.K.; Sakurai, K. Electroencephalography signals-based sparse networks integration using a fuzzy ensemble technique for depression detection. Biomed. Signal Process. Control 2023, 85, 104873. [Google Scholar] [CrossRef]
  46. Seal, A.; Bajpai, R.; Karnati, M.; Agnihotri, J.; Yazidi, A.; Herrera-Viedma, E.; Krejcar, O. Benchmarks for machine learning in depression discrimination using electroencephalography signals. Appl. Intell. 2023, 53, 12666–12683. [Google Scholar] [CrossRef]
  47. Shim, M.; Hwang, H.-J.; Lee, S.-H. Toward practical machine-learning-based diagnosis for drug-naïve women with major depressive disorder using EEG channel reduction approach. J. Affect. Disord. 2023, 338, 199–206. [Google Scholar] [CrossRef]
  48. Sakib, N.; Islam, M.K.; Faruk, T. Machine Learning Model for Computer-Aided Depression Screening among Young Adults Using Wireless EEG Headset. Comput. Intell. Neurosci. 2023, 2023, 1701429. [Google Scholar] [CrossRef]
  49. Huang, Y.; Yi, Y.; Chen, Q.; Li, H.; Feng, S.; Zhou, S.; Zhang, Z.; Liu, C.; Li, J.; Lu, Q. Analysis of EEG features and study of automatic classification in first-episode and drug-naïve patients with major depressive disorder. BMC Psychiatry 2023, 23, 832. [Google Scholar] [CrossRef]
  50. Mumtaz, W.; Xia, L.; Mohd Yasin, M.A.; Azhar Ali, S.S.; Malik, A.S. A wavelet-based technique to predict treatment outcome for Major Depressive Disorder. PLoS ONE 2017, 12, e0171409. [Google Scholar] [CrossRef]
  51. Mumtaz, W. MDD Patients and Healthy Controls EEG Data (New); Dataset; Figshare: London, UK, 2016. [Google Scholar]
  52. Movahed, R.A.; Jahromi, G.P.; Shahyad, S.; Meftahi, G.H. A major depressive disorder classification framework based on EEG signals using statistical, spectral, wavelet, functional connectivity, and nonlinear analysis. J. Neurosci. Methods 2021, 358, 109209. [Google Scholar] [CrossRef]
  53. Movahed, R.A.; Jahromi, G.P.; Shahyad, S.; Meftahi, G.H. A major depressive disorder diagnosis approach based on EEG signals using dictionary learning and functional connectivity features. Phys. Eng. Sci. Med. 2022, 45, 705–719. [Google Scholar] [CrossRef]
  54. Aydemir, E.; Tuncer, T.; Dogan, S.; Gururajan, R.; Acharya, U.R. Automated major depressive disorder detection using melamine pattern with EEG signals. Appl. Intell. 2021, 51, 6449–6466. [Google Scholar] [CrossRef]
  55. Saeedi, M.; Saeedi, A.; Maghsoudi, A. Major depressive disorder assessment via enhanced k-nearest neighbor method and EEG signals. Phys. Eng. Sci. Med. 2020, 43, 1007–1018. [Google Scholar] [CrossRef]
  56. Mahato, S.; Paul, S. Analysis of region of interest (RoI) of brain for detection of depression using EEG signal. Multimed. Tools Appl. 2024, 83, 763–786. [Google Scholar] [CrossRef]
  57. Yang, J.; Zhang, Z.; Fu, Z.; Li, B.; Xiong, P.; Liu, X. Cross-subject classification of depression by using multiparadigm EEG feature fusion. Comput. Methods Programs Biomed. 2023, 233, 107360. [Google Scholar] [CrossRef] [PubMed]
  58. Mohammed, H.; Diykh, M. Improving EEG major depression disorder classification using FBSE coupled with domain adaptation method based machine learning algorithms. Biomed. Signal Process. Control 2023, 85, 104923. [Google Scholar] [CrossRef]
  59. Nassibi, A.; Papavassiliou, C.; Atashzar, S.F. Depression diagnosis using machine intelligence based on spatiospectrotemporal analysis of multi-channel EEG. Med. Biol. Eng. Comput. 2022, 60, 3187–3202. [Google Scholar] [CrossRef]
  60. Cavanagh, J.F.; Bismark, A.W.; Frank, M.J.; Allen, J.J. Multiple dissociations between comorbid depression and anxiety on reward and punishment processing: Evidence from computationally informed EEG. Comput. Psychiatry 2019, 3, 1. [Google Scholar] [CrossRef]
  61. Cai, H.; Gao, Y.; Sun, S.; Li, N.; Tian, F.; Xiao, H.; Li, J.; Yang, Z.; Li, X.; Zhao, Q. Modma dataset: A multi-modal open dataset for mental-disorder analysis. arXiv 2020, arXiv:2002.09283. [Google Scholar]
  62. Tasci, G.; Loh, H.W.; Barua, P.D.; Baygin, M.; Tasci, B.; Dogan, S.; Tuncer, T.; Palmer, E.E.; Tan, R.-S.; Acharya, U.R. Automated accurate detection of depression using twin Pascal’s triangles lattice pattern with EEG Signals. Knowl.-Based Syst. 2023, 260, 110190. [Google Scholar] [CrossRef]
  63. Khan, S.; Umar Saeed, S.M.; Frnda, J.; Arsalan, A.; Amin, R.; Gantassi, R.; Noorani, S.H. A machine learning based depression screening framework using temporal domain features of the electroencephalography signals. PLoS ONE 2024, 19, e0299127. [Google Scholar] [CrossRef]
  64. Acharya, U.R.; Oh, S.L.; Hagiwara, Y.; Tan, J.H.; Adeli, H.; Subha, D.P. Automated EEG-based screening of depression using deep convolutional neural network. Comput. Methods Programs Biomed. 2018, 161, 103–113. [Google Scholar] [CrossRef] [PubMed]
  65. Seal, A.; Bajpai, R.; Agnihotri, J.; Yazidi, A.; Herrera-Viedma, E.; Krejcar, O. DeprNet: A deep convolution neural network framework for detecting depression using EEG. IEEE Trans. Instrum. Meas. 2021, 70, 1–13. [Google Scholar] [CrossRef]
  66. Ay, B.; Yildirim, O.; Talo, M.; Baloglu, U.B.; Aydin, G.; Puthankattil, S.D.; Acharya, U.R. Automated depression detection using deep representation and sequence learning with EEG signals. J. Med. Syst. 2019, 43, 205. [Google Scholar] [CrossRef] [PubMed]
  67. Zhang, J.; Xu, B.; Yin, H. Depression screening using hybrid neural network. Multimed. Tools Appl. 2023, 82, 26955–26970. [Google Scholar] [CrossRef]
  68. Rafiei, A.; Zahedifar, R.; Sitaula, C.; Marzbanrad, F. Automated detection of major depressive disorder with EEG signals: A time series classification using deep learning. IEEE Access 2022, 10, 73804–73817. [Google Scholar] [CrossRef]
  69. Yan, D.-D.; Zhao, L.-L.; Song, X.-W.; Zang, X.-H.; Yang, L.-C. Automated detection of clinical depression based on convolution neural network model. Biomed. Eng./Biomed. Tech. 2022, 67, 131–142. [Google Scholar] [CrossRef]
  70. Song, X.; Yan, D.; Zhao, L.; Yang, L. LSDD-EEGNet: An efficient end-to-end framework for EEG-based depression detection. Biomed. Signal Process. Control 2022, 75, 103612. [Google Scholar] [CrossRef]
  71. Xia, M.; Zhang, Y.; Wu, Y.; Wang, X. An end-to-end deep learning model for EEG-based major depressive disorder classification. IEEE Access 2023, 11, 41337–41347. [Google Scholar] [CrossRef]
  72. Wan, Z.; Huang, J.; Zhang, H.; Zhou, H.; Yang, J.; Zhong, N. HybridEEGNet: A convolutional neural network for EEG feature learning and depression discrimination. IEEE Access 2020, 8, 30332–30342. [Google Scholar] [CrossRef]
  73. Xu, Y.; Zhong, H.; Ying, S.; Liu, W.; Chen, G.; Luo, X.; Li, G. Depressive disorder recognition based on frontal EEG signals and deep learning. Sensors 2023, 23, 8639. [Google Scholar] [CrossRef]
  74. Wu, W.; Ma, L.; Lian, B.; Cai, W.; Zhao, X. Few-Electrode EEG from the Wearable Devices Using Domain Adaptation for Depression Detection. Biosensors 2022, 12, 1087. [Google Scholar] [CrossRef] [PubMed]
  75. Zhang, B.; Wei, D.; Yan, G.; Li, X.; Su, Y.; Cai, H. Spatial–Temporal EEG Fusion Based on Neural Network for Major Depressive Disorder Detection. Interdiscip. Sci. Comput. Life Sci. 2023, 15, 542–559. [Google Scholar] [CrossRef] [PubMed]
  76. Sun, X.; Xu, Y.; Zhao, Y.; Zheng, X.; Zheng, Y.; Cui, L. Multi-Granularity Graph Convolution Network for Major Depressive Disorder Recognition. IEEE Trans. Neural Syst. Rehabil. Eng. 2023, 32, 559–569. [Google Scholar] [CrossRef] [PubMed]
  77. Kang, M.; Kwon, H.; Park, J.-H.; Kang, S.; Lee, Y. Deep-asymmetry: Asymmetry matrix image for deep learning method in pre-screening depression. Sensors 2020, 20, 6526. [Google Scholar] [CrossRef] [PubMed]
  78. Sharma, G.; Joshi, A.M.; Gupta, R.; Cenkeramaddi, L.R. DepCap: A smart healthcare framework for EEG based depression detection using time-frequency response and deep neural network. IEEE Access 2023, 11, 52327–52338. [Google Scholar] [CrossRef]
  79. Khan, D.M.; Yahya, N.; Kamel, N.; Faye, I. Automated diagnosis of major depressive disorder using brain effective connectivity and 3D convolutional neural network. IEEE Access 2021, 9, 8835–8846. [Google Scholar] [CrossRef]
  80. Khan, D.M.; Masroor, K.; Jailani, M.F.M.; Yahya, N.; Yusoff, M.Z.; Khan, S.M. Development of wavelet coherence EEG as a biomarker for diagnosis of major depressive disorder. IEEE Sens. J. 2022, 22, 4315–4325. [Google Scholar] [CrossRef]
  81. Mao, W.; Zhu, J.; Li, X.; Zhang, X.; Sun, S. Resting State EEG Based Depression Recognition Research Using Deep Learning Method. In Proceedings of the International Conference on Brain Informatics (BI), Arlington, TX, USA, 7–9 December 2018; pp. 329–338. [Google Scholar]
  82. Yang, L.; Wang, Y.; Zhu, X.; Yang, X.; Zheng, C. A gated temporal-separable attention network for EEG-based depression recognition. Comput. Biol. Med. 2023, 157, 106782. [Google Scholar] [CrossRef]
  83. Zhang, S.; Wang, H.; Zheng, Z.; Liu, T.; Li, W.; Zhang, Z.; Sun, Y. Multi-View Graph Contrastive Learning via Adaptive Channel Optimization for Depression Detection in EEG Signals. Int. J. Neural Syst. 2023, 33, 2350055. [Google Scholar] [CrossRef]
  84. Cui, W.; Sun, M.; Dong, Q.; Guo, Y.; Liao, X.-F.; Li, Y. A multiview sparse dynamic graph convolution-based region-attention feature fusion network for major depressive disorder detection. IEEE Trans. Comput. Social Syst. 2023, 11, 2691–2702. [Google Scholar] [CrossRef]
  85. Wang, H.-G.; Meng, Q.-H.; Jin, L.-C.; Hou, H.-R. AMGCN-L: An adaptive multi-time-window graph convolutional network with long-short-term memory for depression detection. J. Neural Eng. 2023, 20, 056038. [Google Scholar] [CrossRef] [PubMed]
  86. Li, L.; Wang, X.; Li, J.; Zhao, Y. An EEG-based marker of functional connectivity: Detection of major depressive disorder. Cognit. Neurodyn. 2024, 18, 1671–1687. [Google Scholar] [CrossRef] [PubMed]
  87. Yang, C.; Sun, Z.; Zhang, F.; Shu, H.; Li, J.; Xiang, W. TSUnet-CC: Temporal Spectrogram Unet embedding Cross Channel-wise attention mechanism for MDD identification. In Proceedings of the 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Sydney, Australia, 24–27 July 2023; pp. 1–4. [Google Scholar]
  88. Wang, Y.; Zhao, S.; Jiang, H.; Li, S.; Luo, B.; Li, T.; Pan, G. DiffMDD: A Diffusion-based Deep Learning Framework for MDD Diagnosis Using EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2024, 32, 728–738. [Google Scholar] [CrossRef] [PubMed]
  89. Mumtaz, W.; Qayyum, A. A deep learning framework for automatic diagnosis of unipolar depression. Int. J. Med. Inf. 2019, 132, 103983. [Google Scholar] [CrossRef] [PubMed]
  90. Yang, L.; Wei, X.; Liu, F.; Zhu, X.; Zhou, F. Automatic feature learning model combining functional connectivity network and graph regularization for depression detection. Biomed. Signal Process. Control 2023, 82, 104520. [Google Scholar] [CrossRef]
  91. First, M.B.; Gibbon, M. The structured clinical interview for DSM-IV axis I disorders (SCID-I) and the structured clinical interview for DSM-IV axis II disorders (SCID-II). In Comprehensive Handbook of Psychological Assessment, Vol. 2. Personality Assessment; John Wiley & Sons, Inc.: Oxford, UK, 2004. [Google Scholar]
  92. Rohan, K.J.; Rough, J.N.; Evans, M.; Ho, S.-Y.; Meyerhoff, J.; Roberts, L.M.; Vacek, P.M. A protocol for the Hamilton Rating Scale for Depression: Item scoring rules, Rater training, and outcome accuracy with data on its application in a clinical trial. J. Affect. Disord. 2016, 200, 111–118. [Google Scholar] [CrossRef]
  93. Treede, R.-D.; Rief, W.; Barke, A.; Aziz, Q.; Bennett, M.I.; Benoliel, R.; Cohen, M.; Evers, S.; Finnerup, N.B.; First, M.B. Chronic pain as a symptom or a disease: The IASP Classification of Chronic Pain for the International Classification of Diseases (ICD-11). pain 2019, 160, 19–27. [Google Scholar] [CrossRef]
  94. Williams, L.S.; Brizendine, E.J.; Plue, L.; Bakas, T.; Tu, W.; Hendrie, H.; Kroenke, K. Performance of the PHQ-9 as a screening tool for depression after stroke. stroke 2005, 36, 635–638. [Google Scholar] [CrossRef]
  95. Lecrubier, Y.; Sheehan, D.V.; Weiller, E.; Amorim, P.; Bonora, I.; Sheehan, K.H.; Janavs, J.; Dunbar, G.C. The Mini International Neuropsychiatric Interview (MINI). A short diagnostic structured interview: Reliability and validity according to the CIDI. Eur. Psychiatry 1997, 12, 224–231. [Google Scholar] [CrossRef]
  96. Gilbody, S.; Sheldon, T.; House, A. Screening and case-finding instruments for depression: A meta-analysis. Can. Med. Assoc. J. 2008, 178, 997–1003. [Google Scholar] [CrossRef]
  97. Ren, Y.; Yang, H.; Browning, C.; Thomas, S.; Liu, M. Performance of Screening Tools in Detecting Major Depressive Disorder among Patients with Coronary Heart Disease: A Systematic Review. Med. Sci. Monit. 2015, 21, 646–653. [Google Scholar] [CrossRef] [PubMed]
  98. Wang, B.; Kang, Y.; Huo, D.; Feng, G.; Zhang, J.; Li, J. EEG diagnosis of depression based on multi-channel data fusion and clipping augmentation and convolutional neural network. Front. Physiol. 2022, 13, 1029298. [Google Scholar] [CrossRef] [PubMed]
  99. Harmon-Jones, E. Contributions from research on anger and cognitive dissonance to understanding the motivational functions of asymmetrical frontal brain activity. Biol. Psychol. 2004, 67, 51–76. [Google Scholar] [CrossRef]
  100. Acharya, J.N.; Hani, A.J.; Cheek, J.; Thirumala, P.; Tsuchida, T.N. American Clinical Neurophysiology Society Guideline 2: Guidelines for Standard Electrode Position Nomenclature. Neurodiagnostic J. 2016, 56, 245–252. [Google Scholar] [CrossRef] [PubMed]
  101. Jurcak, V.; Tsuzuki, D.; Dan, I. 10/20, 10/10, and 10/5 systems revisited: Their validity as relative head-surface-based positioning systems. Neuroimage 2007, 34, 1600–1611. [Google Scholar] [CrossRef]
  102. Tsuzuki, D.; Watanabe, H.; Dan, I.; Taga, G. MinR 10/20 system: Quantitative and reproducible cranial landmark setting method for MRI based on minimum initial reference points. J. Neurosci. Methods 2016, 264, 86–93. [Google Scholar] [CrossRef]
  103. Valiulis, V. The Effect of Transcranial Magnetic Stimulation on Brain Bioelectrical Activity. Ph.D. Thesis, Vilnius University, Vilniaus, Lithuania, 2014. [Google Scholar]
  104. Li, X.; Zhang, X.; Zhu, J.; Mao, W.; Sun, S.; Wang, Z.; Xia, C.; Hu, B. Depression recognition using machine learning methods with different feature generation strategies. Artif. Intell. Med. 2019, 99, 101696. [Google Scholar] [CrossRef]
  105. Saeidi, M.; Karwowski, W.; Farahani, F.V.; Fiok, K.; Taiar, R.; Hancock, P.A.; Al-Juaid, A. Neural Decoding of EEG Signals with Machine Learning: A Systematic Review. Brain Sci. 2021, 11, 1525. [Google Scholar] [CrossRef]
  106. Pontifex, M.B.; Gwizdala, K.L.; Parks, A.C.; Billinger, M.; Brunner, C. Variability of ICA decomposition may impact EEG signals when used to remove eyeblink artifacts. Psychophysiology 2017, 54, 386–398. [Google Scholar] [CrossRef]
  107. Xue, Z.; Li, J.; Li, S.; Wan, B. Using ICA to remove eye blink and power line artifacts in EEG. In Proceedings of the 1st International Conference on Innovative Computing, Information and Control (ICICIC 2006), Beijing, China, 30 August–1 September 2006; pp. 107–110. [Google Scholar]
  108. Van, L.D.; Wu, D.Y.; Chen, C.S. Energy-Efficient FastICA Implementation for Biomedical Signal Separation. IEEE Trans. Neural Netw. 2011, 22, 1809–1822. [Google Scholar] [CrossRef]
  109. Mishra, P.; Singla, S.K. Artifact Removal from Biosignal using Fixed Point ICA Algorithm for Pre-processing in Biometric Recognition. Meas. Sci. Rev. 2013, 13, 7–11. [Google Scholar] [CrossRef]
  110. Ksibi, A.; Zakariah, M.; Menzli, L.J.; Saidani, O.; Almuqren, L.; Hanafieh, R.A.M. Electroencephalography-based depression detection using multiple machine learning techniques. Diagnostics 2023, 13, 1779. [Google Scholar] [CrossRef]
  111. Goldberger, A.L.; Peng, C.K.; Lipsitz, L.A. What is physiologic complexity and how does it change with aging and disease? Neurobiol. Aging 2002, 23, 23–26. [Google Scholar] [CrossRef]
  112. Wang, F.; Zhong, S.H.; Peng, J.F.; Jiang, J.M.; Liu, Y. Data Augmentation for EEG-Based Emotion Recognition with Deep Convolutional Neural Networks. In Proceedings of the 24th International Conference on MultiMedia Modeling (MMM), Chulalongkorn University, Bangkok, Thailand, 5–7 February 2018; pp. 82–93. [Google Scholar]
  113. Rommel, C.; Paillard, J.; Moreau, T.; Gramfort, A. Data augmentation for learning predictive models on EEG: A systematic comparison. J. Neural Eng. 2022, 19, 066020. [Google Scholar] [CrossRef]
  114. Zhang, Z.; Liu, Y.; Zhong, S.H. GANSER: A Self-Supervised Data Augmentation Framework for EEG-Based Emotion Recognition. IEEE Trans. Affect. Comput. 2023, 14, 2048–2063. [Google Scholar] [CrossRef]
  115. Nguyen, P.; Nguyen, T.A.; Zeng, Y. Segmentation of design protocol using EEG. Artif. Intell. Eng. Des. Anal. Manuf. 2019, 33, 11–23. [Google Scholar] [CrossRef]
  116. Asensio-Cubero, J.; Gan, J.Q.; Palaniappan, R. Extracting optimal tempo-spatial features using local discriminant bases and common spatial patterns for brain computer interfacing. Biomed. Signal Process. Control 2013, 8, 772–778. [Google Scholar] [CrossRef]
  117. Song, J.X.; Zhai, Q.; Wang, C.; Liu, J.Z. EEGGAN-Net: Enhancing EEG signal classification through data augmentation. Front. Hum. Neurosci. 2024, 18, 1430086. [Google Scholar] [CrossRef] [PubMed]
  118. Polat, K.; Nour, M. Epileptic Seizure Detection Based on New Hybrid Models with Electroencephalogram Signals. Irbm 2020, 41, 331–353. [Google Scholar] [CrossRef]
  119. Shoeibi, A.; Sadeghi, D.; Moridian, P.; Ghassemi, N.; Heras, J.; Alizadehsani, R.; Khadem, A.; Kong, Y.A.; Nahavandi, S.; Zhang, Y.D.; et al. Automatic Diagnosis of Schizophrenia in EEG Signals Using CNN-LSTM Models. Front. Neuroinf. 2021, 15, 777977. [Google Scholar] [CrossRef]
  120. Robbins, K.A.; Touryan, J.; Mullen, T.; Kothe, C.; Bigdely-Shamlo, N. How Sensitive Are EEG Results to Preprocessing Methods: A Benchmarking Study. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 1081–1090. [Google Scholar] [CrossRef] [PubMed]
  121. Singh, A.K.; Krishnan, S. Trends in EEG signal feature extraction applications. Front. Artif. Intell. 2022, 5, 1072801. [Google Scholar] [CrossRef] [PubMed]
  122. Uktveris, T.; Jusas, V. Comparison of Feature Extraction Methods for EEG BCI Classification. In Proceedings of the 21st International Conference on Information and Software Technologies (ICIST), Druskininkai, Lithuania, 15–16 October 2015; pp. 81–92. [Google Scholar]
  123. Debener, S.; Beauducel, A.; Nessler, D.; Brocke, B.; Heilemann, H.; Kayser, J. Is resting anterior EEG alpha asymmetry a trait marker for depression? Findings for healthy adults and clinically depressed patients. Neuropsychobiology 2000, 41, 31–37. [Google Scholar] [CrossRef]
  124. Grin-Yatsenko, V.A.; Baas, I.; Ponomarev, V.A.; Kropotov, J.D. EEG Power Spectra at Early Stages of Depressive Disorders. J. Clin. Neurophysiol. 2009, 26, 401–406. [Google Scholar] [CrossRef] [PubMed]
  125. Gold, C.; Fachner, J.; Erkkilä, J. Validity and reliability of electroencephalographic frontal alpha asymmetry and frontal midline theta as biomarkers for depression. Scand. J. Psychol. 2013, 54, 118–126. [Google Scholar] [CrossRef] [PubMed]
  126. Mohammadi, Y.; Moradi, M.H. Prediction of Depression Severity Scores Based on Functional Connectivity and Complexity of the EEG Signal. Clin. EEG Neurosci. 2021, 52, 52–60. [Google Scholar] [CrossRef]
  127. Miljevic, A.; Bailey, N.W.; Murphy, O.W.; Perera, M.P.N.; Fitzgerald, P.B. Alterations in EEG functional connectivity in individuals with depression: A systematic review. J. Affect. Disord. 2023, 328, 287–302. [Google Scholar] [CrossRef]
  128. Fingelkurts, A.A.; Fingelkurts, A.A.; Rytsälä, H.; Suominen, K.; Isometsä, E.; Kähkönen, S. Impaired functional connectivity at EEG alpha and theta frequency bands in major depression. Hum. Brain Mapp. 2007, 28, 247–261. [Google Scholar] [CrossRef]
  129. Li, X.W.; Hu, B.; Sun, S.T.; Cai, H.S. EEG-based mild depressive detection using feature selection methods and classifiers. Comput. Methods Programs Biomed. 2016, 136, 151–161. [Google Scholar] [CrossRef]
  130. Čukić, M.; Stokić, M.; Simić, S.; Pokrajac, D. The successful discrimination of depression from EEG could be attributed to proper feature extraction and not to a particular classification method. Cognit. Neurodyn. 2020, 14, 443–455. [Google Scholar] [CrossRef]
  131. Thoduparambil, P.P.; Dominic, A.; Varghese, S.M. EEG-based deep learning model for the automatic detection of clinical depression. Phys. Eng. Sci. Med. 2020, 43, 1349–1360. [Google Scholar] [CrossRef] [PubMed]
  132. Cai, H.; Qu, Z.; Li, Z.; Zhang, Y.; Hu, X.; Hu, B. Feature-level fusion approaches based on multimodal EEG data for depression recognition. Inf. Fusion 2020, 59, 127–138. [Google Scholar] [CrossRef]
  133. Shen, J.; Zhang, X.; Wang, G.; Ding, Z.; Hu, B. An improved empirical mode decomposition of electroencephalogram signals for depression detection. IEEE Trans. Affect. Comput. 2019, 13, 262–271. [Google Scholar] [CrossRef]
  134. Sharma, M.; Patel, S.; Acharya, U.R. Automated detection of abnormal EEG signals using localized wavelet filter banks. Pattern Recognit. Lett. 2020, 133, 188–194. [Google Scholar] [CrossRef]
  135. Cai, H.; Sha, X.; Han, X.; Wei, S.; Hu, B. Pervasive EEG diagnosis of depression using Deep Belief Network with three-electrodes EEG collector. In Proceedings of the 2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Shenzhen, China, 15–18 December 2016; pp. 1239–1246. [Google Scholar]
  136. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain–computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef]
  137. Schirrmeister, R.T.; Springenberg, J.T.; Fiederer, L.D.J.; Glasstetter, M.; Eggensperger, K.; Tangermann, M.; Hutter, F.; Burgard, W.; Ball, T. Deep learning with convolutional neural networks for EEG decoding and visualization. Hum. Brain Mapp. 2017, 38, 5391–5420. [Google Scholar] [CrossRef]
  138. Sun, B.; Saenko, K. Deep coral: Correlation alignment for deep domain adaptation. In Proceedings of the Computer Vision–ECCV 2016 Workshops, Amsterdam, The Netherlands, 8–10 and 15–16 October 2016; pp. 443–450. [Google Scholar]
  139. Alom, M.Z.; Taha, T.M.; Yakopcic, C.; Westberg, S.; Sidike, P.; Nasrin, M.S.; Van Esesn, B.C.; Awwal, A.A.S.; Asari, V.K. The history began from alexnet: A comprehensive survey on deep learning approaches. arXiv 2018, arXiv:1803.01164. [Google Scholar]
  140. Li, Y.; Guo, L.; Liu, Y.; Liu, J.; Meng, F. A temporal-spectral-based squeeze-and-excitation feature fusion network for motor imagery EEG decoding. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 1534–1545. [Google Scholar] [CrossRef]
  141. Shen, F.; Dai, G.; Lin, G.; Zhang, J.; Kong, W.; Zeng, H. EEG-based emotion recognition using 4D convolutional recurrent neural network. Cognit. Neurodyn. 2020, 14, 815–828. [Google Scholar] [CrossRef]
  142. Sun, M.; Cui, W.; Yu, S.; Han, H.; Hu, B.; Li, Y. A dual-branch dynamic graph convolution based adaptive transformer feature fusion network for EEG emotion recognition. IEEE Trans. Affect. Comput. 2022, 13, 2218–2228. [Google Scholar] [CrossRef]
  143. Li, X.; Qian, B.; Wei, J.; Li, A.; Liu, X.; Zheng, Q. Classify EEG and reveal latent graph structure with spatio-temporal graph convolutional neural network. In Proceedings of the 2019 IEEE International Conference on Data Mining (ICDM), Beijing, China, 8–11 November 2019; pp. 389–398. [Google Scholar]
  144. Jia, Z.; Lin, Y.; Wang, J.; Zhou, R.; Ning, X.; He, Y.; Zhao, Y. GraphSleepNet: Adaptive spatial-temporal graph convolutional networks for sleep stage classification. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, Yokohama, Japan, 11–17 July 2020; pp. 1324–1330. [Google Scholar]
  145. Sharma, G.; Parashar, A.; Joshi, A.M. DepHNN: A novel hybrid neural network for electroencephalogram (EEG)-based screening of depression. Biomed. Signal Process. Control 2021, 66, 102393. [Google Scholar] [CrossRef]
  146. Shahabi, M.S.; Shalbaf, A.; Maghsoudi, A. Prediction of drug response in major depressive disorder using ensemble of transfer learning with convolutional neural network based on EEG. Biocybern. Biomed. Eng. 2021, 41, 946–959. [Google Scholar] [CrossRef]
  147. Xu, G.; Guo, W.; Wang, Y. Subject-independent EEG emotion recognition with hybrid spatio-temporal GRU-Conv architecture. Med. Biol. Eng. Comput. 2023, 61, 61–73. [Google Scholar] [CrossRef] [PubMed]
  148. Luo, G.; An, P.; Li, Y.; Hong, R.; Chen, S.; Rao, H.; Chen, W. Exploring adaptive graph topologies and temporal graph networks for EEG-based depression detection. IEEE Trans. Neural Syst. Rehabil. Eng. 2023, 31, 3947–3957. [Google Scholar] [CrossRef]
  149. Wong, T.T.; Yeh, P.Y. Reliable Accuracy Estimates from k-Fold Cross Validation. IEEE Trans. Knowl. Data Eng. 2020, 32, 1586–1594. [Google Scholar] [CrossRef]
  150. Jung, Y. Multiple predicting K-fold cross-validation for model selection. J. Nonparam. Stat. 2018, 30, 197–215. [Google Scholar] [CrossRef]
  151. Fushiki, T. Estimation of prediction error by using K-fold cross-validation. Stat. Comput. 2011, 21, 137–146. [Google Scholar] [CrossRef]
  152. Pang, Y.; Wang, Y.T.; Lai, X.A.; Zhang, S.; Liang, P.W.; Song, X.G. Enhanced Kriging leave-one-out cross-validation in improving model estimation and optimization. Comput. Methods Appl. Mech. Eng. 2023, 414, 116194. [Google Scholar] [CrossRef]
  153. Yadav, S.; Shukla, S. Analysis of k-fold cross-validation over hold-out validation on colossal datasets for quality classification. In Proceedings of the IEEE 6th International Advance Computing Conference (IACC), Sagi RamaKrishnam Raju Engn Coll, Bhimavaram, India, 27–28 February 2016; pp. 78–83. [Google Scholar]
  154. Wainer, J.; Cawley, G. Nested cross-validation when selecting classifiers is overzealous for most practical applications. Expert Syst. Appl. 2021, 182, 115222. [Google Scholar] [CrossRef]
  155. Smith, D.J.; Craddock, N. Unipolar and bipolar depression: Different or the same? Br. J. Psychiatry 2011, 199, 272–274. [Google Scholar] [CrossRef]
  156. Liu, Y.; Gu, H.Y.; Zhu, J.; Niu, Y.M.; Zhang, C.; Guo, G.L. Identification of Hub Genes and Key Pathways Associated With Bipolar Disorder Based on Weighted Gene Co-expression Network Analysis. Front. Physiol. 2019, 10, 1081. [Google Scholar] [CrossRef] [PubMed]
  157. Han, K.-M.; De Berardis, D.; Fornaro, M.; Kim, Y.-K. Differentiating between bipolar and unipolar depression in functional and structural MRI studies. Prog. Neuro-Psychopharmacol. Biol. Psychiatry 2019, 91, 20–27. [Google Scholar] [CrossRef] [PubMed]
  158. McIntyre, R.S.; Calabrese, J.R. Bipolar depression: The clinical characteristics and unmet needs of a complex disorder. Curr. Med. Res. Opin. 2019, 35, 1993–2005. [Google Scholar] [CrossRef] [PubMed]
  159. Rolin, D.; Whelan, J.; Montano, C.B. Is it depression or is it bipolar depression? J. Am. Assoc. Nurse Pract. 2020, 32, 703–713. [Google Scholar] [CrossRef] [PubMed]
  160. Hosang, G.M.; Korszun, A.; Jones, L.; Jones, I.; McGuffin, P.; Farmer, A.E. Life-event specificity: Bipolar disorder compared with unipolar depression. Br. J. Psychiatry 2012, 201, 458–465. [Google Scholar] [CrossRef] [PubMed]
  161. Drysdale, A.T.; Grosenick, L.; Downar, J.; Dunlop, K.; Mansouri, F.; Meng, Y.; Fetcho, R.N.; Zebley, B.; Oathes, D.J.; Etkin, A.; et al. Resting-state connectivity biomarkers define neurophysiological subtypes of depression. Nat. Med. 2017, 23, 28–38. [Google Scholar] [CrossRef]
  162. Beijers, L.; Wardenaar, K.J.; van Loo, H.M.; Schoevers, R.A. Data-driven biological subtypes of depression: Systematic review of biological approaches to depression subtyping. Mol. Psychiatry 2019, 24, 888–900. [Google Scholar] [CrossRef]
  163. Cukic, M.; Stokic, M.; Radenkovic, S.; Ljubisavljevic, M.; Simic, S.; Savic, D. Nonlinear analysis of EEG complexity in episode and remission phase of recurrent depression. Int. J. Methods Psychiatr. Res. 2020, 29, e1816. [Google Scholar] [CrossRef]
  164. Espinola, C.W.; Gomes, J.C.; Pereira, J.M.S.; dos Santos, W.P. Detection of major depressive disorder using vocal acoustic analysis and machine learning—An exploratory study. Res. Biomed. Eng. 2021, 37, 53–64. [Google Scholar] [CrossRef]
  165. Gillan, C.M.; Daw, N.D. Taking Psychiatry Research Online. Neuron 2016, 91, 19–23. [Google Scholar] [CrossRef] [PubMed]
  166. Zhang, Y.; Wu, W.; Toll, R.T.; Naparstek, S.; Maron-Katz, A.; Watts, M.; Gordon, J.; Jeong, J.; Astolfi, L.; Shpigel, E.; et al. Identification of psychiatric disorder subtypes from functional connectivity patterns in resting-state electroencephalography. Nat. Biomed. Eng. 2021, 5, 309–323. [Google Scholar] [CrossRef] [PubMed]
  167. Peng, H.; Xia, C.; Wang, Z.H.; Zhu, J.; Zhang, X.; Sun, S.T.; Li, J.X.; Huo, X.N.; Li, X.W. Multivariate Pattern Analysis of EEG-Based Functional Connectivity: A Study on the Identification of Depression. IEEE Access 2019, 7, 92630–92641. [Google Scholar] [CrossRef]
  168. Krigolson, O.E.; Williams, C.C.; Norton, A.; Hassall, C.D.; Colino, F.L. Choosing MUSE: Validation of a Low-Cost, Portable EEG System for ERP Research. Front. Neurosci. 2017, 11, 109. [Google Scholar] [CrossRef] [PubMed]
  169. Neumann, T.; Baum, A.K.; Baum, U.; Deike, R.; Feistner, H.; Scholz, M.; Hinrichs, H.; Robra, B.P. Assessment of the technical usability and efficacy of a new portable dry-electrode EEG recorder: First results of the HOMEONE study. Clin. Neurophysiol. 2019, 130, 2076–2087. [Google Scholar] [CrossRef]
  170. Perrino, T.; Howe, G.; Sperling, A.; Beardslee, W.; Sandler, I.; Shern, D.; Pantin, H.; Kaupert, S.; Cano, N.; Cruden, G.; et al. Advancing Science Through Collaborative Data Sharing and Synthesis. Perspect. Psychol. Sci. 2013, 8, 433–444. [Google Scholar] [CrossRef]
  171. Hinrichs, H.; Scholz, M.; Baum, A.K.; Kam, J.W.Y.; Knight, R.T.; Heinze, H.J. Comparison between a wireless dry electrode EEG system with a conventional wired wet electrode EEG system for clinical applications. Sci. Rep. 2020, 10, 5218. [Google Scholar] [CrossRef]
  172. Yuan, H.W.; Li, Y.; Yang, J.J.; Li, H.J.; Yang, Q.Y.; Guo, C.P.; Zhu, S.M.; Shu, X.K. State of the Art of Non-Invasive Electrode Materials for Brain-Computer Interface. Micromachines 2021, 12, 1521. [Google Scholar] [CrossRef] [PubMed]
  173. Shen, G.C.; Gao, K.P.; Zhao, N.; Wang, Z.Z.; Jiang, C.P.; Liu, J.Q. A Fully Flexible Hydrogel Electrode for Daily EEG Monitoring. IEEE Sens. J. 2022, 22, 12522–12529. [Google Scholar] [CrossRef]
  174. Li, G.L.; Wu, J.T.; Xia, Y.H.; He, Q.G.; Jin, H.G. Review of semi-dry electrodes for EEG recording. J. Neural Eng. 2020, 17, 051004. [Google Scholar] [CrossRef]
  175. Mi, J.X.; Li, A.D.; Zhou, L.F. Review Study of Interpretation Methods for Future Interpretable Machine Learning. IEEE Access 2020, 8, 191969–191985. [Google Scholar] [CrossRef]
  176. Ding, X.F.; Yue, X.X.; Zheng, R.; Bi, C.; Li, D.; Yao, G.Z. Classifying major depression patients and healthy controls using EEG, eye tracking and galvanic skin response data. J. Affect. Disord. 2019, 251, 156–161. [Google Scholar] [CrossRef] [PubMed]
  177. Zhu, J.; Wang, Y.; La, R.; Zhan, J.; Niu, J.; Zeng, S.; Hu, X. Multimodal Mild Depression Recognition Based on EEG-EM Synchronization Acauisition Network. IEEE Access 2019, 7, 28196–28210. [Google Scholar] [CrossRef]
  178. Zhang, Z.; Lin, W.; Liu, M.; Mahmoud, M. Multimodal deep learning framework for mental disorder recognition. In Proceedings of the 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), Buenos Aires, Argentina, 16–20 November 2020; pp. 344–350. [Google Scholar]
  179. Muzammel, M.; Salam, H.; Othmani, A. End-to-end multimodal clinical depression recognition using deep neural networks: A comparative analysis. Comput. Methods Programs Biomed. 2021, 211, 106433. [Google Scholar] [CrossRef] [PubMed]
  180. Vazquez-Romero, A.; Gallardo-Antolin, A. Automatic Detection of Depression in Speech Using Ensemble Convolutional Neural Networks. Entropy 2020, 22, 688. [Google Scholar] [CrossRef] [PubMed]
  181. Song, S.; Jaiswal, S.; Shen, L.; Valstar, M. Spectral Representation of Behaviour Primitives for Depression Analysis. IEEE Trans. Affect. Comput. 2022, 13, 829–844. [Google Scholar] [CrossRef]
  182. Yang, L.; Sahli, H.; Xia, X.; Pei, E.; Oveneke, M.C.; Jiang, D. Hybrid depression classification and estimation from audio video and text information. In Proceedings of the 7th Annual Workshop on Audio/Visual Emotion Challenge, New York, NY, USA, 23 October 2017; pp. 45–51. [Google Scholar]
  183. Alhanai, T.; Ghassemi, M.; Glass, J.; Int Speech Commun, A. Detecting Depression with Audio/Text Sequence Modeling of Interviews. In Proceedings of the 19th Annual Conference of the International-Speech-Communication-Association (INTERSPEECH 2018), Hyderabad, India, 2–6 September 2018; pp. 1716–1720. [Google Scholar]
  184. Yang, L.; Jiang, D.; Xia, X.; Pei, E.; Oveneke, M.C.; Sahli, H. Multimodal measurement of depression using deep learning models. In Proceedings of the 7th Annual Workshop on Audio/Visual Emotion Challenge, Mountain View, CA, USA, 23–27 October 2017; pp. 53–59. [Google Scholar]
  185. Liu, L.; Peng, D.; Zheng, W.-L.; Lu, B.-L. Objective Depression Detection Using EEG and Eye Movement Signals Induced by Oil Paintings. In Proceedings of the 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Sydney, Australia, 24–27 July 2023; pp. 1–4. [Google Scholar] [CrossRef]
  186. Zhu, J.; Wang, Z.; Gong, T.; Zeng, S.; Li, X.; Hu, B.; Li, J.; Sun, S.; Zhang, L. An improved classification model for depression detection using EEG and eye tracking data. IEEE Trans. Nanobiosci. 2020, 19, 527–537. [Google Scholar] [CrossRef]
  187. Verma, P.; Shakya, M. Machine learning model for predicting Major Depressive Disorder using RNA-Seq data: Optimization of classification approach. Cognit. Neurodyn. 2022, 16, 443–453. [Google Scholar] [CrossRef]
  188. Yu, Q.; Wang, R.; Liu, J.; Hu, L.; Chen, M.; Liu, Z. GNN-Based Depression Recognition Using Spatio-Temporal Information: A fNIRS Study. IEEE J. Biomed. Health. Inf. 2022, 26, 4925–4935. [Google Scholar] [CrossRef]
  189. Wang, G.; Wu, N.; Tao, Y.; Lee, W.H.; Cao, Z.; Yan, X.; Wang, G. The Diagnosis of Major Depressive Disorder Through Wearable fNIRS by Using Wavelet Transform and Parallel-CNN Feature Fusion. IEEE Trans. Instrum. Meas. 2023, 72, 1–11. [Google Scholar] [CrossRef]
  190. Mahayossanunt, Y.; Nupairoj, N.; Hemrungrojn, S.; Vateekul, P. Explainable Depression Detection Based on Facial Expression Using LSTM on Attentional Intermediate Feature Fusion with Label Smoothing. Sensors 2023, 23, 9402. [Google Scholar] [CrossRef]
  191. Zhang, J.; Yu, Y.; Barra, V.; Ruan, X.; Chen, Y.; Cai, B. Feasibility study on using house-tree-person drawings for automatic analysis of depression. Comput. Methods Biomech. Biomed. Eng. 2024, 27, 1129–1140. [Google Scholar] [CrossRef] [PubMed]
  192. Kubo, K.; Sakurai, H.; Tani, H.; Watanabe, K.; Mimura, M.; Uchida, H. Predicting relapse from the time to remission during the acute treatment of depression: A re-analysis of the STAR*D data. J. Affect. Disord. 2023, 320, 710–715. [Google Scholar] [CrossRef] [PubMed]
  193. Flint, A.J.; Bingham, K.S.; Alexopoulos, G.S.; Marino, P.; Mulsant, B.H.; Neufeld, N.H.; Rothschild, A.J.; Voineskos, A.N.; Whyte, E.M.; Meyers, B.S.; et al. Predictors of relapse of psychotic depression: Findings from the STOP-PD II randomized clinical trial. J. Psychiatr. Res. 2023, 157, 285–290. [Google Scholar] [CrossRef]
  194. Othmani, A.; Zeghina, A.-O.; Muzammel, M. A Model of Normality Inspired Deep Learning Framework for Depression Relapse Prediction Using Audiovisual Data. Comput. Methods Programs Biomed. 2022, 226, 107132. [Google Scholar] [CrossRef]
  195. Muzammel, M.; Othmani, A.; Mukherjee, H.; Salam, H. Identification of signs of depression relapse using audio-visual cues: A preliminary study. In Proceedings of the 2021 IEEE 34th International Symposium on Computer-Based Medical Systems (CBMS), Aveiro, Portugal, 7–9 June 2021; pp. 62–67. [Google Scholar]
  196. Othmani, A.; Zeghina, A.O. A multimodal computer-aided diagnostic system for depression relapse prediction using audiovisual cues: A proof of concept. Healthc. Anal. 2022, 2, 100090. [Google Scholar] [CrossRef]
  197. Wu, W.; Zhang, Y.; Jiang, J.; Lucas, M.V.; Fonzo, G.A.; Rolle, C.E.; Cooper, C.; Chin-Fatt, C.; Krepel, N.; Cornelssen, C.A.; et al. An electroencephalographic signature predicts antidepressant response in major depression. Nat. Biotechnol. 2020, 38, 439–447. [Google Scholar] [CrossRef]
  198. Watts, D.; Pulice, R.F.; Reilly, J.; Brunoni, A.R.; Kapczinski, F.; Passos, I.C. Predicting treatment response using EEG in major depressive disorder: A machine-learning meta-analysis. Transl. Psychiatry 2022, 12, 332. [Google Scholar] [CrossRef]
  199. Beyrouthy, T.; Mostafa, N.; Roshdy, A.; Karar, A.S.; Alkork, S.; Mourtzis, D.; Niu, Q.; Yang, X. Review of EEG-Based Biometrics in 5G-IoT: Current Trends and Future Prospects. Appl. Sci. 2024, 14, 534. [Google Scholar] [CrossRef]
  200. Gutierrez, L.J.; Rabbani, K.; Ajayi, O.J.; Gebresilassie, S.K.; Rafferty, J.; Castro, L.A.; Banos, O. Internet of Things for Mental Health: Open Issues in Data Acquisition, Self-Organization, Service Level Agreement, and Identity Management. Int. J. Environ. Res. Public Health 2021, 18, 1327. [Google Scholar] [CrossRef]
Figure 1. General framework for depression diagnosis based on TML and DL methods.
Figure 1. General framework for depression diagnosis based on TML and DL methods.
Sensors 24 06815 g001
Figure 2. The 10–20 system electrode position diagram, taken from Ref. [68].
Figure 2. The 10–20 system electrode position diagram, taken from Ref. [68].
Sensors 24 06815 g002
Figure 3. Schematic diagram of brain regions.
Figure 3. Schematic diagram of brain regions.
Sensors 24 06815 g003
Figure 4. Frequency of different preprocessing methods used in studies of depression diagnosis.
Figure 4. Frequency of different preprocessing methods used in studies of depression diagnosis.
Sensors 24 06815 g004
Figure 5. Frequency distribution of TML-based algorithms for depression diagnosis.
Figure 5. Frequency distribution of TML-based algorithms for depression diagnosis.
Sensors 24 06815 g005
Figure 6. Frequency distribution of deep learning-based algorithms for depression diagnosis.
Figure 6. Frequency distribution of deep learning-based algorithms for depression diagnosis.
Sensors 24 06815 g006
Figure 7. Frequency distribution of different validation methods in depression diagnostic studies.
Figure 7. Frequency distribution of different validation methods in depression diagnostic studies.
Sensors 24 06815 g007
Table 1. Basic experimental setup for EEG data acquisition in diagnostic studies of depression.
Table 1. Basic experimental setup for EEG data acquisition in diagnostic studies of depression.
Researches Based on TML Methods
Sample Size (MDD + HC)Frequency (Hz)ElectrodesStudy
92 + 1212503[38]
34 + 4025019[39]
23 + 12, 15 + 15500, 5122, 1[40]
200 + 20050026[41]
10 + 10200, 40018[42]
92-19[43]
23 + 33, 24 + 29, 15 + 18250, 256, 5123, 128, 19[44]
23 + 33, 24 + 29, 15 + 18, 32 + 30250, 250, 512, 2563, 128, 19, 19[45]
15 + 1825622[46]
49 + 49100064[47]
19 + 1312814[48]
40 + 40100032[49]
34 + 3025619[52]
34 + 3025619[53]
34 + 3025619[54]
34 + 3025019[55]
30 + 3025619[56]
24 + 2525619[57]
34 + 3025019[58]
42 + 425002[59]
24 + 29250128[62]
26 + 29-3[63]
Research Based on DL Methods
Sample Size (MDD + HC)Frequency (Hz)ElectrodesStudy
34 + 3025619[36]
15 + 152564[64]
15 + 1825619[65]
15 + 152564[66]
24 + 24250128[67]
34 + 3025619[68]
40 + 4010003[69]
40 + 4010003[70]
30 + 2825619[71]
23 + 125006[72]
41 + 342506[73]
16 + 16-56[35]
26 + 292503[74]
24 + 2912564[75]
23 + 25, 23 + 25, 40 + 67250, 256, 500 128, 19, 66[76]
34 + 3025619[77]
34 + 3025619[78]
30 + 3025619[79]
30 + 3025619[80]
17 + 17-128[81]
26 + 24, 24 + 29500, 25062, 128[82]
24 + 29, 26 + 24250, 500128, 62[83]
24 + 29, 34 + 30250, 256128, 19[84]
20 + 32, 24 + 29500, 25066, 128[85]
27 + 28, 22 + 22256, 25619, 19[86]
28 + 2425619[87]
34 + 30, 23 + 19256, 25619, 19[88]
Table 2. Publicly available datasets used in diagnostic studies of depression.
Table 2. Publicly available datasets used in diagnostic studies of depression.
Public DatasetsNetwork SourceStudy
TMLDL
Mumtaz et al. [39,50,51,89]https://figshare.com/articles/dataset/EEG_Data_New/4244171, (accessed on 15 October 2024) [52,53,54,55,56,57,58][36,67,68,71,74,76,77,78,79,80,82,83,84,85,86,87,88]
MODMA [61]http://modma.lzu.edu.cn/data/index/, (accessed on 15 October 2024) [44,45,62,63][67,74,76,82,83,84,85]
Cavanahg et al. [60]https://openneuro.org/datasets/ds003478/versions/1.1.0, (accessed on 15 October 2024) [59][76,86]
EDRA [82]https://github.com/EllieYLJ/, (accessed on 15 October 2024) [82][82,83]
PRED + CThttp://predict.cs.unm.edu/downloads.php, (accessed on 15 October 2024)-[85,88]
Table 3. Detailed information on non-public datasets used in diagnostic studies of depression.
Table 3. Detailed information on non-public datasets used in diagnostic studies of depression.
Studies Based on TML Methods
Male/FemaleAge Mean or Age RangeDiagnostic Criteria for MDDSourceStudy
--MMSE-[38]
MDD: 11/12 HC: 6/6, MDD: 9/6 HC: 11/4MDD: 29.3 HC: 26.4, MDD: 34 HC: 33.6DSM-IV HAM-D MINIBeijing Anding Hospital, China[40]
MDD: 58/142; HC: 58/142MDD: 53.88 HC: 52.23DSM-5 MININTUH TVGH CGMHKL CGMHLnK[41]
6/1424–60HAM-D EST-Q-[42]
--HAM-DShenzhen Traditional Chinese Hospital[43]
MDD: 11/4 HC: 16/2MDD: 21.86 HC: 21PHQ-9-[46]
0/98MDD: 45.39 HC: 48.18DSM-5 HAM-DThe Psychiatry Department of Inje University Ilsan Paik Hospital[47]
16/16MDD: 21.6 HC: 21.3PHQ-9Independent University, Bangladesh[48]
MDD: 14/22 HC: 14/17MDD: 22.7 HC: 25.2MINI HAM-DThe Affiliated Brain Hospital of Guangzhou Medical University[49]
Studies based on DL methods
Male/FemaleAge mean or age rangeDiagnostic criteria for MDDSourceStudy
-20–50Clinical diagnosisThe Psychiatry Department, Medical College, Calicut, Kerala, India[64]
--PHQ-9-[65,66]
MDD: 15/25 HC: 17/23MDD: 45.5 HC: 44.9ICD HAM-DThe Second Affiliated Hospital of Jining Medical College[69,70]
MDD: 11/12 HC: 6/6MDD: 29.2 HC: 26.4-Beijing Anding Hospital, China[72]
MDD: 10/31 HC: 11/23MDD: 45.22 HC: 40.18HAM-DThe designated hospital for psychosis and the local community[98]
MDD: 7/9 HC: 7/9MDD: 31.0 HC: 26.1DSM-IV HAMDBeijing Anding Hospital, China[35]
MDD: 13/11 HC: 20/9MDD: 30.88 HC: 31.45PHQ-9 LES GAD SSRS PSQISecond Hospital of Lanzhou University in Gansu, China, and Lanzhou city of Gansu Province using posters[75]
MDD: 11/6 HC: 13/4MDD: 33.35 HC: 30.29MINI PHQ-9Lanzhou University Second Hospital[81]
MMSE: Mini Mental State Examination, NTUH: National Taiwan University Hospital, TVGH: Taipei Veterans General Hospital, CGMHKL: Chang Gung Memorial Hospital KeeLung, CGMHLnK: Chang Gung Memorial Hospital LinKou, LES: Life Event Scale, GAD: Generalized Anxiety Disorder Scale, SSRS: Social Support Research Scale, and PSQI: Pittsburgh Sleep Quality Index.
Table 4. Electrodes placed in different brain regions in the 10–20 international standard system.
Table 4. Electrodes placed in different brain regions in the 10–20 international standard system.
Brain RegionElectrodes
FrontalFp1, Fp2, Fz, F3, F4, F7, F8
TemporalT3, T4, T5, T6
ParietalP3, P4, Pz
OccipitalO1, O2, Oz
CentralC3, C4, Cz
Reference points electrodesA1, A2
Table 5. Preprocessing methods used in depression diagnostic studies.
Table 5. Preprocessing methods used in depression diagnostic studies.
Researches Based on TML MethodsStudy
Finite Impulse Response (FIR) filter[38,57,59]
DWT[38,40,54,55]
Multiscale DWT, Kalman filter, Adaptive Predictor Filter (APF), Adaptive Auto Regressive (AAR) model[38]
Multiple source modeling technique[39]
Notch filter[44,45,46,49,53,55,63]
Z-score normalization/Z-score standardization[40,43]
Data segmentation[40,42,48,52,53,56,58,63]
Adaptive filter, soft-thresholding algorithm[40]
Bandpass filter[41,42,47,49,55,56]
ICA[41,46,49,52,53,56,57,59]
Low-pass filter[44,45,46,52,53,57,58]
Normalization, Demean[46]
PCA[47]
IIR Butterworth filter, Spherical spline[48]
Rereference[49,56]
High-pass filter[44,45,46,52,53,57,58,59]
Multilevel discrete wavelet transform (MDWT)[62]
Smoothing filter[53]
Research based on DL methodsStudy
Bandpass filter[36,68,71,73,75,76,78,79,80,85,86,87]
Notch filter[36,64,65,66,67,68,69,70,71,78,79,80,86]
ICA[35,36,65,73,77,78,83,85,87]
Z-score normalization[64,71,72,76,78]
High-pass filter[65,88]
Low-pass filter[65,84,88]
Adaptive noise cancellation (ANC)[67,75]
Artifact subspace reconstruction (ASR)[68]
Wavelet threshold filter[69,70]
Finite Impulse Response (FIR) filter[35,70]
Rereference[71,76,87]
Downsampling[73]
Baseline correction[73]
Data segmentation[73,77,78,82]
FFT[35]
Three Channel Data Merge Chart, Synthesis by RGB[74]
DWT[75]
FastICA[76,81]
Wavelet packet transformation (WPT)[76]
Min–max normalization[77]
Hanning filter[81]
Data cropping[81,82]
Signal slicing[82]
Baseline removal[84]
Detrend[84]
Table 6. Different classes of features extracted in studies based on traditional machine learning methods.
Table 6. Different classes of features extracted in studies based on traditional machine learning methods.
Time-domain features
Feature extraction methodStudy
Peak[38]
Skewness[38,40,48,52,58,59,63]
Kurtosis[38,40,48,52,58,59]
Mean/average[40,46,52,58,63]
Standard deviation (SD)[40]
Maximum, minimum[40,52,58]
Median[40,46,58]
Variance[38,48,58,59]
25th and 75th percentile[40]
Mode[46,58]
Mean cube, standard deviation, first difference, normalized first difference, second difference, normalized second difference, mobility, complexity, and Pearson’s coefficient of skewness[46]
Range, first quartile, third quartile[58]
Hjorth parameters (activity, mobility, and complexity)[38,48,52,59]
Standard deviation[58,63]
Relative median, root mean square[59]
Maximum and minimum amplitude, peak-to-peak signal value, peak-to-peak time, mean of absolute values of first and second difference[63]
Frequency-domain features
Feature extraction methodStudy
Relative centroid frequency, absolute centroid frequency relative power, and absolute power[38]
Power spectrum, 8-level DWT, and the square summation of the detailed coefficients[40]
Band power (BP)[41,46,52,55,56,59]
Coherence[41]
Alpha power variability, spectral asymmetry index[42]
β/α ratio[43]
Global relative powers, relative fronto-central powers, alpha asymmetry 1, alpha asymmetry 2[46]
Power spectral density (PSD)[47]
Power spectrum[49]
Interhemispheric asymmetry, DWT[52]
Wavelet packet decomposition[55]
Relative band power[42,59]
Mean frequency, median frequency, PSD, maximum-PSD, singular value[59]
Energy[63]
Nonlinear features
Feature extraction methodStudy
Kolmogorov entropy[38,52]
Shannon entropy[38,46,48,52,58,59,63]
Approximate entropy[40,52,55,56,58]
Wavelet entropy[40]
Higuchi’s Fractal Dimension (HFD)[41,42]
C0-complexity[38,40,52]
Katz’s Fractal Dimension (KFD)[41]
Lempel–Ziv complexity (LZC)[42,57]
Power spectral entropy[38,40]
Detrended fluctuation analysis (DFA)[42,52,56]
Log energy entropy[48]
Higuchi, correlation dimension, Lyapunov exponent[52]
Correlation dimension (CD)[38,56]
Sample entropy[55,56]
Dispersion entropy, zero crossing rate[58]
Connectivity features
Feature extraction methodStudy
Synchronization likelihood (SL)[39,52,53]
Node embeddings[44,45]
Functional connectivity[47,49]
Network indices[47]
Feature generation
Feature extraction methodStudy
Melamine pattern and DWT-based multileveled feature generation[54]
TPTLP-based feature extraction[62]
Table 7. Different feature extraction methods combined with deep learning methods.
Table 7. Different feature extraction methods combined with deep learning methods.
Feature Extraction MethodsArchitecturesStudy
Combining the structural features and connectivity featuresCNN[35]
Convert EEG signals into imagesDAN[74]
Fusion of temporal and spatial domain EEG features3 layer LSTM + 4 layer CNN[75]
Learning topological features of changes between functional brain regions and brain salience patterns on multi-granular functional neural networks using graph neural networksGNN[76]
Converting asymmetric features of EEG to matrix images2D-CNN[77]
Generating Spectrogram Images from EEG signals using Short-Time Fourier transformCNN + LSTM[78]
Estimating effective connectivity within the brain’s default mode network (DMN) using the PDC algorithm3D-CNN[79]
Estimation of wavelet coherence (WCOH) between DMN regions of the brain using EEG signals2D-CNN[80]
Using the projection method to construct EEG signalsCNN[81]
Power spectral density (PSD) features were extracted for four bands: δ, θ, α, and βGTSAN[82]
Build two similarity metric views MLP[83]
A MultiView (MV) Feature ExtractorMV-SDGC-RAFFNet[84]
Automatic generation of brain functional connectivity included in the adjacency matrix and congregating of spatio-temporal featuresAMGCN-L[85]
Constructing brain functional connectivity matrices using a fusion feature called P-MSWCLightweight-CNN[86]
Using multi-scale saliency-encoded spectrogramTSUnet-CC[87]
Designing a forward diffusion noise training module to extract noise-independent featuresDiffMDD[88]
DAN: Deep Adaptation Network, LSTM: long short-term memory, GNN: graph neural network, GTSAN: Gated Temporal-Separable Attention Network, MLP: Multi-Layer Perceptron, MV-SDGC-RAFFNet: Multiview Sparse Dynamic Graph Convolution-based Region-Attention Feature Fusion Network, AMGCN-L: Adaptive Multi-time-window Graph Convolutional Network with long-short-term memory, TSUnet-CC: Temporal Spectrogram Unet embedding Cross Channel-wise attention mechanism, and DiffMDD: a diffusion-based deep learning framework for MDD diagnosis using EEG.
Table 8. Feature selection methods used in EEG-based depression diagnostic studies.
Table 8. Feature selection methods used in EEG-based depression diagnostic studies.
Feature Selection/Dimensionality Reduction MethodStudy
Minimal redundancy maximal relevance (MRMR)[38,56]
Rank-based feature selection method[39]
GA[40,55]
Feature averaging strategy across epochs[41]
Univariate feature ranking using F-tests[42]
ReliefF[42,56]
A two-stage feature selection method named PAR[43]
Node2vec algorithm[44,45]
Analysis of Variance (ANOVA) test and correlation analysis[46]
Fisher score-based feature selection method[47]
SVM-RFE, LASSO-LR, and PCA[49]
Sequential backward feature selection (SBFS)[52]
Synchronization likelihood method[53]
Neighborhood Component Analysis (NCA)[54,59,62]
Maximum Likelihood Function (MLE)[56]
Linear combination and concatenation[57]
Student’s t-test, Wilcoxon test[58]
Correlation-based feature selection (CFS)[63]
SVM-RFE: Support Vector Machine-Recursive Feature Elimination and LASSO-LR: Least Absolute Shrinkage and Selection Operator-Logistic Regression.
Table 9. Performance of various algorithms in depression diagnosis studies based on TML.
Table 9. Performance of various algorithms in depression diagnosis studies based on TML.
AlgorithmsValidation Methods (Training/Testing/Validation)Top-Performing ClassifierAccuracy (%)Study
KNN, SVM, CT, ANN10-fold CVKNN76.83[38]
LR, NB, SVM10-fold CV (90%/10%/-)SVM98[39]
KNN, KNN-GA, RF, RF-GA, LDA, LDA-GA, CART, CART-GALOPOCVKNN-GA94[40]
KNN, LDA, SVM, CK-SVM5-fold CV (70%/30%/-)CK-SVM84.16[41]
SVM, LDA, NB, KNN, DT10-fold CVDT95[42]
LR, SVM, LNR5-fold CVSVM98.95[43]
KNN, SVM, LR, LDA, XGBoost, DT10-fold CV (90%/-/10%)KNN93.3[44]
KNN, SVM, LR, LDA, XGBoost, DT10-fold CV (90%/-/10%)KNN96[45]
KNN, NB, DT, MLP, SVM, XGBoost, RF10-fold CV (53.6%/33%/13.4%)XGBoost87[46]
SVMLOOCVSVM83.67[47]
SVM (Linear, Quadratic, Cubic, Gaussian radial basis), KNN (Fine, Medium, Coarse, Cosine, Cubic, Weighted)5-fold CV (70%/30%/-)Fine KNN98.43[48]
DT, SVM, GBDT, NB, KNN10-fold CV (90%/10%/-)KNN88.2[49]
SVM (Linear, Radial basis function), RF, LR, DT, GB, NB, RB10-fold CV (90%/10%/-)RBF-SVM99[52]
LC-KSVD, CLC-KSVD10-fold CV (90%/10%/-)LC-KSVD99[53]
Weighted KNN, SVM (Quadratic)Hold-out CV (80%/20%/-)Weighted KNN99.11[54]
E-KNN, SVM, MLP10-fold CV (90%/-/10%)E-KNN98.44[55]
SVM, DT, NB, LDA, LR, Bagging10-fold CVSVM95.23[56]
SVM, KNN, DT10-fold CV (90%/10%/-)SVM94.03[57]
SVM, LS-SVM, KNN, DT, RF, Gradient Boosting, (Bagged, Boosted, Stacked) Ensemble10-fold CV, LOOCV (90%/10%/-)Stacked Ensemble99.11[58]
SVM (Coarse Gaussian, Cubic, Fine Gaussian, Linear, Quadratic, Medium Gaussian), KNN (Coarse, Cosine, Cubic, Fine, Ensemble Subspace, Medium, Weighted), (Coarse, Fine, Ensemble Bagged, Ensemble Boosted) Tree, (Linear, Ensemble Subspace) Discriminant, (Trilayered, Narrow, Wide, Bilayered) Neural Network, Kernel Naïve Bayes (KNB), Logistic Regression Kernel10-fold CV (75%/25%/-)KNB91.8[59]
KNNLOOCV, 10-fold CVKNN100[62]
BF-Tree, KNN, AdaBoost10-fold CV, Nested CV (90%/10%/-)BF-Tree96.36[63]
CT: Classification Tree, ANN: Artificial Neural Network, NB: Naïve Bayesian, KNN-GA: K-nearest neighbor-genetic algorithm, RF-GA: Random Forest-genetic algorithm, LDA-GA: Linear Discriminant Analysis-genetic algorithm, CART: Classification and Regression Tree, CART-GA: Classification and Regression Tree-genetic algorithm, CK-SVM: Conformal Kernel-Support Vector Machine, LNR: Linear Regression, GBDT: gradient-boosting Decision Tree, RBF-SVM: Radial Basis Function kernel-Support Vector Machine, LC-KSVD: Label Consistent K-SVD, CLC-KSVD: Correlation-based Label Consistent K-SVD, E-KNN: enhanced K-nearest neighbor, LOPOCV: Leave-One-Participant-Out Cross-Validation, and LOOCV: Leave-One-Out Cross-Validation.
Table 10. Performance of various algorithms in depression diagnosis studies based on DL.
Table 10. Performance of various algorithms in depression diagnosis studies based on DL.
AlgorithmsValidation Methods (Training/Testing/Validation)Top-Performing ClassifierAccuracy (%)Study
CNN, RF, LSTM, Bi-LSTM, XGBoost, RNN, GRU, Gradient Boosting10-fold CV (70%/30%/-)CNN98.13[36]
CNN10-fold CV (90%/10%/-)CNN95.49[64]
DeprNet, AchLSTM, AchCNN, T-LSTM, H-KNN1, H-KNN2, S-EMD, S-SVM, H-DBN10-fold CV (53.3%/33.3%/13.4%)DeprNet99.37[65]
CNN-LSTM10-fold CV (70%/15%/15%)CNN-LSTM99.12[66]
2DCNN-LSTM, SVM, KNN, DTLOOCV2DCNN-LSTM95.1[67]
InceptionTime10-fold CV (20.9%/66.9%/12.2%)InceptionTime91.67[68]
EEGNet, DeepConvNet, ShallowConvNet(70%/10%/20%)EEGNet94.27[69]
LSDD-EEGNet, RF, XGBoost, LightGBM, KNN, EEGNet, DeepConvNet(70%/30%/-)LSDD-EEGNet94.69[70]
CNN, CNN-LSTM, DeprNet, Two-branch CNNLOSOCV, 10-fold CVTwo-branch CNN98.45[71]
HybridEEGNet, SynEEGNet, RegEEGNet, DeepConvNet, AchCNN, EEGNet10-fold CVHybridEEGNet79.08[72]
MRCNN-RSE, MRCNN-LSTM, EEGNet, DeprNet, 1DCNN-LSTM, 2DCNN-LSTM5-fold CVMRCNN-RSE98.48[73]
CNN, SVM, KNN10-fold CVCNN94.13[35]
DAN, DANN, DeepCoral(87.5%/12.5%/-), (50%/50%/-)DAN77[74]
LDA, SVM, 2DCNN-LSTM, Dictionary Learning10-fold CV, LOOCV2DCNN-LSTM96.33[75]
MGGCN, SVM, GICN5-fold CV (70%/30%/-)MGGCN99.69[76]
2D-CNN5-fold CV (80%/20%/-)2D-CNN98.85[77]
VGG16, AlexNet, Inception, ResNet50, CNN, CNN-GRU, CNN-LSTM10-fold CVCNN-LSTM99.9[78]
3D-CNN10-fold CV, 15-fold CV (50%/50%/-)3D-CNN95.65[79]
2D-CNN10-fold CV (50%/50%/-)2D-CNN100[80]
TCN, LSTM, CNN8-fold CVCNN77.20[81]
GTSAN, SVM, KNN, LSTM, MLP, GRU, TCN(80%/20%/-)GTSAN98.33[82]
MLP, SVM, LR, GCN, GAT10-fold CV (80%/10%/10%)MLP99.19[83]
MV-SDGC-RAFFNet, EEGNet, Shallow ConvNet, Deep ConvNet, TS-SEFFNet, 4D-CRNN, DBGC-ATFFNet10-fold CVMV-SDGC-RAFFNet99.19[84]
SVM, KNN, FCN, AMGCN, LSTM, CNN-LSTM, ASTGCN, AMGCN-L, GICN, DepHNN, LSDD_EEGnet10/5-fold CVAMGCN-L90.57[85]
Lightweight-CNN, LDA, LR, SVM, RF, AdaBoost(70%/30%/-)Lightweight-CNN99.87[86]
TSUnet-CCLOSOCVTSUnet-CC99.22[87]
DiffMDD, LR, SVM, XGBoost, 1D-CNN-Transformer, CWT-1D-CNN, CWP2D-CNN, EEGNet, InceptionNet, TSception, GRU-Conv, DeprNet, GC-GRU10-fold CVDiffMDD94.0[88]
Bi-LSTM: Bidirectional long short-term memory, GRU: Gate Recurrent Unit, TCN: Temporal Convolutional Networks, LightGBM: Light Gradient Boosting Machine, DANN: Domain Adversarial neural network, MGGCN: Multi-Granularity Graph Convolution Network, GICN: Graph Input layer attention Convolutional Network, GAT: Graph Attention Network, FCN: Fully connected network, AMGCN: Adaptive Multi-time-window Graph Convolutional Network, and LOSOCV: Leave-One-Subject-Out Cross-Validation.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, H.; Fang, J.; Zhang, J.; Zhang, X.; Piao, W.; Liu, Y. Resting-State Electroencephalogram Depression Diagnosis Based on Traditional Machine Learning and Deep Learning: A Comparative Analysis. Sensors 2024, 24, 6815. https://doi.org/10.3390/s24216815

AMA Style

Lin H, Fang J, Zhang J, Zhang X, Piao W, Liu Y. Resting-State Electroencephalogram Depression Diagnosis Based on Traditional Machine Learning and Deep Learning: A Comparative Analysis. Sensors. 2024; 24(21):6815. https://doi.org/10.3390/s24216815

Chicago/Turabian Style

Lin, Haijun, Jing Fang, Junpeng Zhang, Xuhui Zhang, Weiying Piao, and Yukun Liu. 2024. "Resting-State Electroencephalogram Depression Diagnosis Based on Traditional Machine Learning and Deep Learning: A Comparative Analysis" Sensors 24, no. 21: 6815. https://doi.org/10.3390/s24216815

APA Style

Lin, H., Fang, J., Zhang, J., Zhang, X., Piao, W., & Liu, Y. (2024). Resting-State Electroencephalogram Depression Diagnosis Based on Traditional Machine Learning and Deep Learning: A Comparative Analysis. Sensors, 24(21), 6815. https://doi.org/10.3390/s24216815

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop