Next Article in Journal
What Does Soluble C-Type Lectin-like Receptor 2 (sCLEC-2) × D-Dimer/Platelet (PLT) (sCLEC-2 × D-Dimer/PLT) Mean for Coagulation/Fibrinolysis Conditions? Comment on Yamamoto et al. Super Formula for Diagnosing Disseminated Intravascular Coagulation Using Soluble C-Type Lectin-like Receptor 2. Diagnostics 2023, 13, 2299
Previous Article in Journal
Musculoskeletal Radiology Education: A National Survey by the Italian College of Musculoskeletal Radiology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Silo, Privacy-Preserving, and Lightweight Federated Multimodal System for the Identification of Major Depressive Disorder Using Audio and Electroencephalogram

1
Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura 140417, Punjab, India
2
Department of Computer Science and Engineering, School of Engineering and Technology, Central University of Haryana, Mahendergarh 123031, Haryana, India
3
Department of Electronics and Communication Engineering, University Institute of Engineering and Technology, Kurukshetra University, Kurukshetra 136119, Haryana, India
4
Department of Computer Science, Jyotiba Phule Government College, Radaur, Yamunanagar 135133, Haryana, India
5
Department of Computer Applications, Sir Chottu Ram Institute of Engineering & Technology, Ch. Charan Singh University, Meerut 250001, Uttar Pradesh, India
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(1), 43; https://doi.org/10.3390/diagnostics14010043
Submission received: 16 October 2023 / Revised: 12 December 2023 / Accepted: 19 December 2023 / Published: 25 December 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
In this day and age, depression is still one of the biggest problems in the world. If left untreated, it can lead to suicidal thoughts and attempts. There is a need for proper diagnoses of Major Depressive Disorder (MDD) and evaluation of the early stages to stop the side effects. Early detection is critical to identify a variety of serious conditions. In order to provide safe and effective protection to MDD patients, it is crucial to automate diagnoses and make decision-making tools widely available. Although there are various classification systems for the diagnosis of MDD, no reliable, secure method that meets these requirements has been established to date. In this paper, a federated deep learning-based multimodal system for MDD classification using electroencephalography (EEG) and audio datasets is presented while meeting data privacy requirements. The performance of the federated learning (FL) model was tested on independent and identically distributed (IID) and non-IID data. The study began by extracting features from several pre-trained models and ultimately decided to use bidirectional short-term memory (Bi-LSTM) as the base model, as it had the highest validation accuracy of 91% compared to a convolutional neural network and LSTM with 85% and 89% validation accuracy on audio data, respectively. The Bi-LSTM model also achieved a validation accuracy of 98.9% for EEG data. The FL method was then used to perform experiments on IID and non-IID datasets. The FL-based multimodal model achieved an exceptional training and validation accuracy of 99.9% when trained and evaluated on both IID and non-IIID datasets. These results show that the FL multimodal system performs almost as well as the Bi-LSTM multimodal system and emphasize its suitability for processing IID and non-IIID data. Several clients were found to perform better than conventional pre-trained models in a multimodal framework for federated learning using EEG and audio datasets. The proposed framework stands out from other classification techniques for MDD due to its special features, such as multimodality and data privacy for edge machines with limited resources. Due to these additional features, the framework concept is the most suitable alternative approach for the early classification of MDD patients.

1. Introduction

Mental disorders are a major global health problem as they contribute significantly to the global burden of disease and have a major impact on people’s social and economic well-being. According to the World Health Organization (WHO), an estimated 264 million women and men of all ages suffer from a mental disorder, which means that up to 27% of the world’s population is affected by mental health problems at some point in their lives [1]. According to current WHO reports from 2023, an estimated 5% of adults worldwide suffer from depression [2]. Major depressive disorder (MDD) is a mental illness that has a significant impact on general behavior, emotions, and cognitive abilities [3]. In addition, it has been difficult to transfer diagnostic descriptions from clinically established nosology to various biological indices. We need new methods to accurately diagnose depression, including biomarkers [4,5,6,7,8]. According to the literature, depressed people release fewer neurotransmitters and have lower concentrations of synaptic receptors compared to healthy people [9,10].
Electroencephalography (EEG) is the conventional method used to detect brain activities which helps us in further diagnosis of brain diseases [11]. On the other hand, audio recordings are an unconventional method that can be used to recognize the frequency of valence and arousal in a person’s voice. Machine learning (ML) and deep learning (DL) are used to create a centralized model for different fields such as healthcare, agriculture, smart industry, intelligent management, and traffic and environmental monitoring systems [12]. Centralized models perform better because they have access to more data. Merging model parameters and training on the device only increases training time in federated learning (FL) in terms of model communication and aggregation [13,14,15]. For platforms focused on real-time and continuous monitoring of mental health problems, including depression, these trade-offs may be significant. Despite the fact that FL [16,17] has been used for a variety of healthcare services, there is a dearth of research that has utilized FL for privacy-preserving multimodal analysis to assess MDD.
Lightweight FL describes the establishment of technologies and approaches that reduce the amount of computational and communication resources required to build MDD models on mobile devices or distributed platforms [18]. It aims to make federated learning efficient and flexible in contexts with minimal resources [19]. It could be used in many different domains that require fast analysis of large amounts of distributed data [20]. In the field of FL, there is still a lack of studies on MDD classification. These challenges need to be addressed for FL to be used successfully in the real world for people with MDD. FL could, therefore, provide a private and secure system for a sensitive disease such as MDD, as many hospitals do not share their data for privacy reasons [21].
Major depressive disorder is a severe form of depression. However, there are many conventional methods for identifying MDD that are far less than optimal. AI-trained models have helped to develop more accurate and robust models for detecting MDD. Due to the sensitivity of medical data, it is difficult to share all patient information on a centralized device. Therefore, in this work, we have developed a privacy-preserving federated learning-based multimodal [22] system after studying the literature and research on MDD detection. Thus, in this work, a privacy-preserving multimodal FL system for MDD classification is developed using audio and EEG data. By having a local model at each node, the efficiency of a federated learning-based node is not significantly affected when connectivity is interrupted. It also reduces the communication overhead, which is the main drawback of algorithms based on deep learning or machine learning.
Researchers have developed a range of neural network-based machine learning and deep learning models to differentiate between MDD and non-MDD individuals. With developments in artificial intelligence, these models may be able to identify MDD subjects reliably and with high accuracy. Cai et al. [10] used four machine learning methods, Support Vector Machine (SVM), K-Nearest Neighbour (KNN), Classification Tree (CT), and Artificial Neural Network (ANN), to classify depression based on EEG signals and achieved a significant accuracy of 79.27% with the KNN algorithm. Orabi et al. [6] processed many deep learning algorithms to access the results of “X, formerly Twitter” data for depression detection and found that the Convolutional Neural Network (CNN) based model performed well compared to the Recurrent Neural Network (RNN) based model. Several techniques for recognizing various mental disorders are described in Table 1. However, finding a balance between data accuracy and privacy is a major problem. While the technology may detect diseases with great accuracy, there may be privacy issues.
Lam et al. [5] tested a transformer model and a 1D-CNN model for identifying depressed and non-depressed individuals using audio and text data, which performed well compared to other state-of-the-art depression detection systems. Zogan et al. [24] proposed a hybrid deep learning model using Tweeter data for detecting depressed individuals and achieved a good prediction with an F1 score of 89%. Mousavian et al. [32] developed depression detection based on resting-state MRI and structured MRI images using a machine learning model.
With the development of social media, people started sharing their emotions and feelings by posting their messages on different platforms like Facebook, “X, formerly Twitter”, WhatsApp, etc. Therefore, to diagnose and control depression, several optimized machine and deep learning algorithms were developed using social media data [6,24,29]. Many resources restrict access to other users, so data must be shared in a centralized location for privacy and sensitivity. The proposed study on the establishment of DL algorithms in a federated learning setting to improve data privacy has been emphasized in Table 1, Table 2 and Table 3.
The research gap identified in Table 1, Table 2 and Table 3 shows that considerable accuracy has been achieved in previous work but at the cost of privacy and resource utilization. Multimodal frameworks are expected to provide more precise and accurate results as the unique characteristics of each trained DL framework are considered across different MDD datasets [42,43,44]. Therefore, in the proposed framework, a multimodal system is developed using audio and EEG data. A major limitation of DL algorithms is the confidentiality of patient mental health data. Therefore, decentralized training with federated learning helps to protect the privacy of the data. The main limitation in centralized training of DL algorithms is the high communication cost required to transfer MDD-related data between the client and the server. The datasets in centralized DL are not IID. For non-ID datasets, independence means that all sample values are independent and not identically distributed. FL works with both IID and non-ID datasets. The aims of this research were as follows:
  • To develop and analyze baseline EEG and audio-based multimodal systems for the classification of MDD.
  • To implement and analyze privacy-preserved FL multimodal systems for the classification of MDD using EEG and audio databases.
  • To analyze the impact of identical and non-identical multimodal Cross-Silo databases on the FL-based MDD classification system.
The following sections of this work are structured as follows. Section 2 describes the associated work done in the suggested study using comparative analysis tables and research gap data. Section 3 mathematically defines the materials and methods used in the proposed study, as well as their subsections which include the proposed FL framework, DL methods, data collection details, multimodal framework, and methodology used. The outcomes of the base DL models and the suggested FL multimodal employing EEG and audio data are presented extensively in Section 4. Finally, Section 5 represents the proposed study’s conclusion.

2. Materials and Methods

The following section provides examples of the procedures and methods used in the proposed work. In this study, a federated deep learning framework is presented to solve privacy issues related to depression classification. The framework uses a decentralized client-server architecture to ensure the confidentiality of all identically and non-identically distributed (IID and non-IID) client data. This research was conducted to classify MDD patients and healthy control subjects. For this purpose, we first collected an online dataset containing MDD data from EEG and audio recordings of many subjects. Subsequently, these data were preprocessed for further training and finally analyzed using our proposed framework. The results are presented below in Figure 1.
The privacy-preserving Multimodal Federated Learning (MFL) model was developed to achieve a robust model with the security of raw data by building a model on local devices. The audio and EEG data from MDD and healthy control subjects were collected in an online dataset. Subsequently, the data were pre-processed and divided into IID and non-IID datasets, which were then forwarded to the implementation phase. First, the DL algorithms Bidirectional Long-Term Memory (Bi-LSTM), Long Short-Term Memory (LSTM), and Convolutional Neural Network (CNN) were applied separately to audio and EEG datasets. In addition, the DL algorithm Bi-LSTM was implemented to create a multimodal system from audio and EEG datasets. Finally, the multimodal FL framework was developed to create a privacy-sensitive system for MDD subjects.

2.1. Proposed Multimodal Federated Learning Framework for MDD Classification

In general, collaborative learning can be applied to the simultaneous processing of large amounts of data on independently generated datasets with data processing nodes that have approximately similar computational power. Data sharing may also require collaborative learning with significant communication capacity and high data security. Federated learning is a collaborative machine learning architecture that enables data evaluation, such as neural network (NN) training, directly on the data storage device. Only the result parameters, such as changed weights of the NN, are imported into the neural network to create an aggregated analysis framework in FL. Therefore, compared to regular distributed systems, FL systems do not collect data in a centralized data warehouse.
The proposed FL framework for MDD classification is shown in Figure 2. For federated machine learning or deep learning, the proposed framework involves a client-server architecture, and the given tasks need to be performed on the client and server sides with separate programming codes. N clients were connected to a server in a federated cluster, each with an individual set of programming languages. In the current study, the proposed federated learning framework for MDD identification, including a client-server architecture, was implemented using Python programming libraries led by Keras and TensorFlow.
The server, the communication setting, and the clients are the three essential elements of a FL architecture. The formal FL explanation includes N clients (K1, K2, … KN) who have their datasets (D1, D2, … DN), so the full dataset is represented as DN = D1 ∪ D2 ∪…∪ DN. All shared datasets DN’s are pooled together in a standard decentralized training method to develop a deep learning framework. In the FL, each client’s data Di is trained independently, and each trained model collaborates without disclosing any client information due to other clients Ki or the server S. Further, the FL averaging structure is discussed below in Algorithm 1.
Algorithm 1. Federated Averaging
K—number of clients from 1 to n
B—minimum batch size
E—number of epochs
F—fractions of clients
Server function—FedAvg
Initialize global weights w
for round i = 1, 2…do
   M ← max (F.K, 1)
   V t ← (random sets of M clients)
  for client k € V t do parallel
  ClientUpdate(k, w t )
    w t + 1 Σ K k = 1 W k t + 1
  end
end

Client function—ClientUpdate(k, w )
B ← Split data into batches
For each local epoch i from1 to E do
   Update client w
Return w to server

2.2. Deep Learning Methods

The DL techniques can help in the diagnosis of mental health issues using data from social media, clinical health records, and mobile device sensors. The centralized DL methods are widely used to identify and classify depression, and many researchers have developed robust models. Due to their ability to easily work on big data problems, in this work, we have used LSTM, CNN, and Bi-LSTM for preparing our base models.
In the CNN algorithm, a convolutional layer loops across the width and length in two dimensions. When the data inputs are Mel Frequency Cepstral Coefficients, whose dimensions represent duration and bandwidth, respectively, this does not apply. By addressing Mel Frequency Cepstral Coefficients (MFCC) in the depression detection problem, one-dimensional convolution is more appropriate than two-dimensional convolution. Convolution across the frequency axis is applied in the suggested model. As a result, the model is able to provide features that represent a brief temporal association. The configuration of a 1D CNN model requires forward and backward propagation layers; this is referred to as Multi-Layer Perceptron (MLP). In CNN layers, the forward propagation layer is presented as follows: X k l = b k l + i = 1 n l 1   c o n v 1 D ( w i k l 1 , s i l 1 ) where X k l is the input, b k l is the bias factor at k t h the neuron of layer l , w i k l 1 is the kernel, and s i l 1 is the output of i t h neuron. So, the output can be expressed as Y k l = f ( X k l ) .
Long Short-Term Memory (LSTM) is specially used for sequential data like time series, waves, speech, text, and so on. LSTM is designed to overcome the long-term dependencies of Recurrent Neural Networks and provides more accurate predictions by remembering the information for a long time. It includes three gates; input gate, output gate, and forget gate. All of these gates generate values as 0 or 1, where 0 denotes that the gate is blocked, and no information is passing and 1 denotes the opposite. In LSTM, the representation of the Forgot Gate, R, is R t = σ W f . C t 1 , h t 1 , a t b f . Input gate, P, is defined as P t = σ W i . C t 1 , h t 1 , a t b i , and Output gate, T, is described as T t = σ W o . C t 1 , h t 1 , a t b o .
Bidirectional LSTM is referred to as an updated version of LSTM because it works for information transmission in both directions, forward and backward at the same time. In this way, it helps to better predict the output value. The model can learn from both the present and the future time steps thanks to BiLSTM. We used BiLSTM to demonstrate the depression identification and detection system because we wanted to extract the properties of EEG and sound data. Therefore, it calculates the output value jointly as D f = σ W f x n , h n 1 1 , h n 1 2 + b f .

2.3. Multimodal Architecture

A data representation that incorporates information from various sources is known as a multimodal representation. Multimodality refers to obtaining most of the details of data by using multiple techniques, so that the deep learning algorithm may be trained effectively to produce superior outputs [45]. As each modality operates differently, merging them is difficult, and the beginning of each modality requires a different pattern. Stacking numerous diverse datasets may result in improved performance compared to individual models. A growing amount of multimodal data is being transferred through internet connections as an outcome of the expansion of the network and a wide range of smart gadgets in the past few years. Therefore, a larger variety of multimodal scenarios for applications are arising [46].
The recent achievement of large frameworks and their multimodal variations, in particular, highlights AI’s potential in multimodal core systems. The joint representations are methods that combine unimodal representations into a multimodal environment. Joint representations are employed in tasks that contain multimodal information during both the training and the testing processes [45]. A concatenation of specific modality attributes is the most basic example of a combined representation. The multimodal ‘ m ’ is described mathematically as Y m = f ( Y 1 , Y 2 , , Y n ) .

2.4. Data Acquisition

In this experiment, to create a multimodal system, we collected 3-channel EEG and audio recordings from the MODMA dataset [47]. EEG recordings include data from 55 participants; here, 26 depressed and 29 non-depressed participants were recorded. The data was recorded in a resting state (participants’ eyes were closed) for 90 s in a room with no noise [48]. A pervasive EEG collection device with three electrodes is used for data collection from the prefrontal lobe of participants because this part of the human brain relates to emotions and psychiatric conditions [49].
Audio recordings include data from 52 Chinese subjects (23 depressed and 29 healthy participants). These data were collected from the Second Affiliated Hospital of Lanzhou University [50]. In this dataset, each subject had 29 recordings, which were classified as positive, neutral, and negative emotional stimuli. High-quality equipment was used to collect speech data. All the participants from healthy and depressed groups were aged 18 to 55.
For data preprocessing, the Mel Frequency Cepstral Coefficients technique was implemented. Due to its efficiency in identifying the change of low-frequency data and focusing on people’s perceptions, the MFCC feature is the most commonly utilized audio feature in speech-related activities and signals like EEG waves. To avoid its large variance affecting the training, all MFCC factors were normalized. Thus, after preprocessing, 161 features were extracted using MFCC for both audio and EEG signals. Results are discussed in Table 4.

2.5. IIDs and Non-IIDs

IID and non-IID data in FL represent various data distribution trends among the involved clients or systems. Every client or system in the federated learning configuration has an equivalent distribution of IID data available to it. The average distribution of the data is similar for all clients and is independent. However, non-IID denotes that data distribution among clients is neither identical nor independent. The data could be unbalanced, have distinct feature visualizations, or have multiple category distributions, among other statistical characteristics. The non-IID dataset was created by randomly distributing the original dataset among clients. Therefore, in non-IID, data is distributed in different patterns among clients. When a federated learning algorithm is combined with a deep learning algorithm, it effectively works on diverse datasets by involving relatively reliable clients as shown in Figure 3.

3. Results

In this section, the results of the DL algorithms LSTM, Bi-LSTM, and CNN are briefly discussed by comparing their performance in the classification of Major Depressive Disorder. By comparing the training and validation accuracy, the best-trained model was identified. Another FL-based deep multimodal model was trained for the classification of MDD. However, different Python libraries such as Pandas, Tensor-Flow, NumPy, Sklearn, and others were used for the preprocessing, categorization, scaling, testing, and training of the data.

3.1. Deep Learning for Audio Dataset

The DL algorithms are applied to the audio dataset after preprocessing the data to see the results in different situations with the best performance. In datasets related to depression, there is a significant data imbalance that brings a preference for non-depressed individuals in categorization. To reduce the prediction bias, sampling methods must be used to equalize the size of the depressed and non-depressed classes.
In addition, multiple signals from a single individual may emphasize very personal characteristics. Therefore, it is important to standardize the length of each person’s responses. MFCC is a commonly used voice descriptor for depression detection of depression. Thus, with the help of MFCC, 161 columns of features were extracted from the audio dataset. Subsequently, the DL models LSTM, CNN, and Bi-LSTM were used to classify depressed and non-depressed subjects.
Table 5 and Figure 5 displays the outcomes of comparing various DL algorithms working with audio features. Therefore, validation parameters clearly show that Bi-LSTM performed better upon audio features compared to the CNN and LSTM models. The Bi-LSTM model attained the highest validation accuracy, precision, and recall, with a score of 91% for each of these outcomes. On the other hand, LSTM and CNN models achieved validation accuracy of 89% and 85%, respectively. In Figure 4, the confusion matrix clearly shows that the Bi-LSTM model identified MDD and normal subjects more correctly than CNN and LSTM models using the audio dataset.
Figure 4. Confusion matrix of DL models with the audio dataset.
Figure 4. Confusion matrix of DL models with the audio dataset.
Diagnostics 14 00043 g004
Figure 5. Training and Validation results using DL models for audio data.
Figure 5. Training and Validation results using DL models for audio data.
Diagnostics 14 00043 g005aDiagnostics 14 00043 g005b

3.2. Deep Learning for EEG Dataset

EEG data is firstly preprocessed and 3-channel data is converted into 161 columns and 20,134 rows for the application of the DL model. EEG signals are one-dimensional data. For their classification, a one-dimensional model was developed using CNN, LSTM, and Bi-LSTM algorithms. However, the EEG data was collected from the subject’s prefrontal lobe using a 3-channel device that detects brain activities.
The deep learning models gave better results with EEG data, as shown in Table 6. The training and validation results show that the Bi-LSTM and LSTM models performed approximately equally upon the EEG dataset. In Figure 6, confusion matrices present the number of identified MDD and normal patients with the help of different DL models using the EEG dataset.
The results in Table 6 demonstrate that the validation accuracy, precision, and recall using the Bi-LSTM model is the highest with a score of 98.95% for each parameter. The LSTM model took second place with a score of 98.93% for each of the three parameters; the validation accuracy, precision, and recall. However, the CNN model has also achieved good validation results with 96.92% accuracy, precision, and recall. Figure 7, shows the results of the DL models increasing in value with every epoch from 0 to 100.

3.3. Deep Learning Using Multimodal Audio and EEG Datasets

Multimodal learning is the process of linking information from different sources. The fundamental challenge when using DL algorithms is to determine the best architecture for a target. A strong model that provides high accuracy in EEG identification may not perform as well on voice recognition tests. The aim of this work is to present a DL model for MDD classification to improve the accuracy of the classification problem. The DL algorithm Bi-LSTM was developed to train a multimodal model based on audio and EEG data.
For this purpose, the audio and EEG datasets are first preprocessed, and then the extracted features are collected in 161 rows for each dataset. To train a multimodal system, we need equal parameters. Thus, the concatenation layer merges both models in the form of a multimodal model.
Table 7 shows the results of the multimodal DL model with audio and EEG data. The validation results clearly show that it is a robust model with an accuracy, precision, and recall of 100% in each case. In Figure 8, we can see that the accuracy increases from the first epoch, and the differences between the training and validation accuracy are very small. This model could, therefore, be used in real-life situations.

3.4. Federated Learning Multimodal Using Audio and EEG Datasets

The FL model was used together with the DL model to increase the confidentiality of the data by developing our model at the client’s site. The weights of the central model are initialized randomly in the initial phase of the FL procedure. After initialization, the central server repeatedly contacts the clients at each communication cycle until the frame converges. There are mainly two FL settings for data partitioning: (1) the IID setting, in which the data is distributed independently and identically among the clients, and (2) the non-ID setting, in which each client has data from different distributions. The communication process, therefore, becomes slower as the size of the NN increases. However, in this framework, the communication parameters are not applied. Further, FL could be used for heterogeneous data as in non-IID environments. Non-IID data could affect the accuracy and precision of the model, but this largely depends on the type of data the clients have. If the clients have data from healthy control subjects, this would result in negative samples, whereas in contrast, the model would recognize a positive number of samples for depressed control subjects. This model performed better in both IID and non-IID situations, demonstrating the stability of the proposed MFL model.
As we can see from Figure 9 and Table 8, under the IID setting, the outcomes are slightly less than the central DL multimodal system because of the decentralized training of data at local locations. As such, the training and validation accuracies for all IID data are 99.94% and 99.93%, respectively. Further, the evaluation matrix’s result for validation precision is 99.9% and recall is 99.94%. Thus, a privacy-preserved FL model could be used for MDD diagnosis.
Figure 10 and Table 9 show the performance of the FL multimodal system in non-IID settings. The results are slightly higher in non-IID settings than in IID settings. As such, the training accuracy, precision, and recall for non-IID settings are 99.99% for each. Further, we obtained 0 loss for training and validation.

4. Discussion

The performance of the proposed MFL model for detecting depression was compared with the pre-trained models in Table 1, Table 2 and Table 3. The proposed system performed better with more privacy by identifying depression with maximum accuracy and minimum loss using audio and EEG datasets. First, a base model with LSTM, Bi-LSTM, and CNN algorithms was trained individually on EEG and audio datasets. In both data sets, Bi-LSTM outperformed the other algorithms. Subsequently, a multimodal model was created from EEG and audio data sets by applying Bi-LSTM. The DL multi-model achieved 100% accuracy and precision in the validation phase. Finally, the FL algorithm was applied to the multimodal in combination with Bi-LSTM after splitting the EEG and audio datasets into four clients for IID and non-IID scenarios. The system captured the audio and EEG features more accurately with both IID and non-IID datasets. The proposed system achieved 99.9% accuracy with 0 losses in both IID and non-IIID scenarios. Thus, it is clear that the proposed system is consistent with training parameters with a wide set of clients.

5. Conclusions and Future Trends

In today’s world, it is crucial to classify MDD individuals while protecting privacy by using multimodal DL. In this study, the use of privacy-preserving federated deep learning for MDD classification was investigated. First, pre-trained deep learning models such as CNN, LSTM, and Bi-LSTM were used and evaluated. Compared to the other algorithms discussed, Bi-LSTM obtained a validation accuracy of 91% for audio data and 98.95% for EEG data. Consider the best performing model, i.e., Bi-LSTM, as the baseline model for multimodal deep learning. The multimodal DL model achieved a score of 100% validation accuracy, precision, and recall. In addition, a federated learning-based multimodal privacy-preserving model was implemented with Bi-LSTM for individuals with MDD using audio and EEG data. The proposed model could protect the high-risk and sensitive mental health data by developing it on the client’s side. Moreover, the concept of multimodality increases the robustness and reliability of the model by performing it with multiple data sources and types. For this experiment, we used the multimodal MODMA dataset. As a result, the proposed system achieved a training accuracy of 99.99% with 0 losses and a validation accuracy of 99.97% for IID and non-IID settings. Based on the results, it was found that the proposed system outperforms the federated multimodal deep learning system in terms of resource utilization and privacy while achieving classification results very similar to those of the baseline deep learning-based multimodal system. Most importantly, communication costs are lower in the FL model as the trained model is shared between client and server rather than the entire dataset. In future studies, this model could be applied to different types of datasets such as image data, video data, text data, etc.

Author Contributions

Conceptualization C.G., V.K. and N.G.; methodology C.G., V.K. and K.S.; software C.G., V.K., N.G. and K.S.; validation C.G., V.K., N.G., K.S., R.B. and S.K.; formal analysis C.G., V.K., R.B., S.K. and R.R.; investigation C.G., V.K., R.B., R.R. and S.K.; writing—original draft preparation, C.G., V.K., N.G. and R.R.; writing—review and editing, C.G., V.K., N.G. and R.R.; project administration, V.K.; funding, N.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

The data presented in this study are openly available in https://modma.lzu.edu.cn/data/index/ (accessed on 18 December 2023) at [47].

Conflicts of Interest

The authors declare that they have no conflicts of interest to report pertaining to the present study.

References

  1. Yasin, S.; Hussain, S.A.; Aslan, S.; Raza, I.; Muzammel, M.; Othmani, A. EEG Based Major Depressive Disorder and Bipolar Disorder Detection Using Neural Networks: A Review. Comput. Methods Programs Biomed. 2021, 202, 106007. [Google Scholar] [CrossRef] [PubMed]
  2. World Health Organization. Depressive Disorder (Depression). 2023. Available online: https://www.who.int/news-room/fact-sheets/detail/depression (accessed on 21 November 2023).
  3. Uyulan, C.; Ergüzel, T.T.; Unubol, H.; Cebi, M.; Sayar, G.H.; Nezhad Asad, M.; Tarhan, N. Major Depressive Disorder Classification Based on Different Convolutional Neural Network Models: Deep Learning Approach. Clin. EEG Neurosci. 2021, 52, 38–51. [Google Scholar] [CrossRef] [PubMed]
  4. de Aguiar Neto, F.S.; Rosa, J.L.G. Depression Biomarkers Using Non-Invasive EEG: A Review. Neurosci. Biobehav. Rev. 2019, 105, 83–93. [Google Scholar] [CrossRef] [PubMed]
  5. Lam, G.; Dongyan, H.; Lin, W. Context-Aware Deep Learning for Multi-Modal Depression Detection. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 1–5. [Google Scholar]
  6. Orabi, A.H.; Buddhitha, P.; Orabi, M.H.; Inkpen, D. Deep Learning for Depression Detection of Twitter Users. In Proceedings of the Fifth Workshop on Computational Linguistics and Clinical Psychology: From Keyboard to Clinic, New Orleans, LA, USA, 5 June 2018; Loveys, K., Niederhoffer, K., Prud’hommeaux, E., Resnik, R., Resnik, P., Eds.; Association for Computational Linguistics: New Orleans, LA, USA, 2018; pp. 88–97. [Google Scholar]
  7. Ansari, L.; Ji, S.; Chen, Q.; Cambria, E. Ensemble Hybrid Learning Methods for Automated Depression Detection. IEEE Trans. Comput. Soc. Syst. 2023, 10, 211–219. [Google Scholar] [CrossRef]
  8. Xu, R.; Hua, A.; Wang, A.; Lin, Y. Early-Depression-Detection-for-Young-Adults-with-a-Psychiatric-and-Ai-Interdisciplinary-Multimodal-Framework. Int. J. Psychol. Behav. Sci. 2021, 15, 219–225. [Google Scholar]
  9. Barberis, A.; Petrini, E.M.; Mozrzymas, J.W. Impact of synaptic neurotransmitter concentration time course on the kinetics and pharmacological modulation of inhibitory synaptic currents. Front. Cell. Neurosci. 2011, 5, 6. [Google Scholar] [CrossRef]
  10. Cai, H.; Han, J.; Chen, Y.; Sha, X.; Wang, Z.; Hu, B.; Yang, J.; Feng, L.; Ding, Z.; Chen, Y.; et al. A Pervasive Approach to EEG-Based Depression Detection. Complexity 2018, 2018, 5238028. [Google Scholar] [CrossRef]
  11. Gupta, C.; Khullar, V. Contemporary Intelligent Technologies for Electroencephalogram-Based Brain Computer Interface. In Proceedings of the 2022 10th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions), ICRITO 2022, Noida, India, 13–14 October 2022. [Google Scholar]
  12. Tsouvalas, V.; Ozcelebi, T.; Meratnia, N. Privacy-preserving Speech Emotion Recognition through Semi-Supervised Federated Learning. In Proceedings of the 2022 IEEE International Conference on Pervasive Computing and Communications Workshops and Other Affiliated Events (PerCom Workshops), Pisa, Italy, 21–25 March 2022; pp. 359–364. [Google Scholar] [CrossRef]
  13. Fan, Z.; Su, J.; Gao, K.; Peng, L.; Qin, J.; Shen, H.; Hu, D.; Zeng, L.-L. Federated Learning on Structural Brain MRI Scans for the Diagnostic Classification of Major Depression. Biol. Psychiatry 2021, 89, S183. [Google Scholar] [CrossRef]
  14. Luo, J.; Wu, S. FedSLD: Federated Learning with Shared Label Distribution for Medical Image Classification. In Proceedings of the 2022 IEEE 19th International Symposium on Biomedical Imaging (ISBI), Kolkata, India, 28–31 March 2022; pp. 1–5. [Google Scholar]
  15. Ain, Q.u.; Khan, M.A.; Yaqoob, M.M.; Khattak, U.F.; Sajid, Z.; Khan, M.I.; Al-Rasheed, A. Privacy-Aware Collaborative Learning for Skin Cancer Prediction. Diagnostics 2023, 13, 2264. [Google Scholar] [CrossRef]
  16. Agbley, B.L.Y.; Li, J.; Hossin, M.A.; Nneji, G.U.; Jackson, J.; Monday, H.N.; James, E.C. Federated Learning-Based Detection of Invasive Carcinoma of No Special Type with Histopathological Images. Diagnostics 2022, 12, 1669. [Google Scholar] [CrossRef]
  17. Subashchandrabose, U.; John, R.; Anbazhagu, U.V.; Venkatesan, V.K.; Thyluru Ramakrishna, M. Ensemble Federated Learning Approach for Diagnostics of Multi-Order Lung Cancer. Diagnostics 2023, 13, 3053. [Google Scholar] [CrossRef] [PubMed]
  18. Ju, C.; Gao, D.; Mane, R.; Tan, B.; Liu, Y.; Guan, C. Federated Transfer Learning for EEG Signal Classification. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS, Montreal, QC, Canada, 20–24 July 2020; pp. 3040–3045. [Google Scholar]
  19. Khullar, V.; Singh, H.P. Privacy Protected Internet of Unmanned Aerial Vehicles for Disastrous Site Identification. Concurr. Comput. 2022, 34, e7040. [Google Scholar] [CrossRef]
  20. Wadhwa, S.; Saluja, K.; Gupta, S.; Gupta, D. Blockchain Based Federated Learning Approach for Detection of COVID-19 Using Io MT. SSRN Electron. J. 2022, 25, 760–764. [Google Scholar] [CrossRef]
  21. Sharma, S.; Guleria, K. A Comprehensive Review on Federated Learning Based Models for Healthcare Applications. Artif. Intell. Med. 2023, 146, 102691. [Google Scholar] [CrossRef] [PubMed]
  22. Sheu, R.K.; Chen, L.C.; Wu, C.L.; Pardeshi, M.S.; Pai, K.C.; Huang, C.C.; Chen, C.Y.; Chen, W.C. Multi-Modal Data Analysis for Pneumonia Status Prediction Using Deep Learning (MDA-PSP). Diagnostics 2022, 12, 1706. [Google Scholar] [CrossRef] [PubMed]
  23. Othmani, A.; Kadoch, D.; Bentounes, K.; Rejaibi, E.; Alfred, R.; Hadid, A. Towards Robust Deep Neural Networks for Affect and Depression Recognition from Speech. In Pattern Recognition: Proceedings of the ICPR International Workshops and Challenges, ICPR 2021, Virtual, 10–15 January 2021; Springer International Publishing: Cham, Switzerland, 2021; pp. 5–19. [Google Scholar]
  24. Zogan, H.; Razzak, I.; Wang, X.; Jameel, S.; Xu, G. Explainable Depression Detection with Multi-Modalities Using a Hybrid Deep Learning Model on Social Media. World Wide Web 2022, 25, 281–304. [Google Scholar] [CrossRef] [PubMed]
  25. Diwakar, A.; Kaur, T.; Ralekar, C.; Gandhi, T.K. Deep Learning Identifies Brain Cognitive Load via EEG Signals. In Proceedings of the 2020 IEEE 17th India Council International Conference, INDICON 2020, New Delhi, India, 10–13 December 2020. [Google Scholar]
  26. Saeedi, A.; Saeedi, M.; Maghsoudi, A.; Shalbaf, A. Major Depressive Disorder Diagnosis Based on Effective Connectivity in EEG Signals: A Convolutional Neural Network and Long Short-Term Memory Approach. Cogn. Neurodyn 2021, 15, 239–252. [Google Scholar] [CrossRef]
  27. Amanat, A.; Rizwan, M.; Javed, A.R.; Abdelhaq, M.; Alsaqour, R.; Pandya, S.; Uddin, M. Deep Learning for Depression Detection from Textual Data. Electronics 2022, 11, 676. [Google Scholar] [CrossRef]
  28. Escorcia-Gutierrez, J.; Beleño, K.; Jimenez-Cabas, J.; Elhoseny, M.; Dahman Alshehri, M.; Selim, M.M. An Automated Deep Learning Enabled Brain Signal Classification for Epileptic Seizure Detection on Complex Measurement Systems. Measurement 2022, 196, 111226. [Google Scholar] [CrossRef]
  29. Bhat, P.; Anuse, A.; Kute, R.; Bhadade, R.S.; Purnaye, P. Mental Health Analyzer for Depression Detection Based on Textual Analysis. J. Adv. Inf. Technol. 2022, 13, 67–77. [Google Scholar] [CrossRef]
  30. Adarsh, V.; Arun Kumar, P.; Lavanya, V.; Gangadharan, G.R. Fair and Explainable Depression Detection in Social Media. Inf. Process. Manag. 2023, 60, 103168. [Google Scholar] [CrossRef]
  31. Tutun, S.; Johnson, M.E.; Ahmed, A.; Albizri, A.; Irgil, S.; Yesilkaya, I.; Ucar, E.N.; Sengun, T.; Harfouche, A. An AI-Based Decision Support System for Predicting Mental Health Disorders. Inf. Syst. Front. 2023, 25, 1261–1276. [Google Scholar] [CrossRef] [PubMed]
  32. Mousavian, M.; Chen, J.; Traylor, Z.; Greening, S. Depression Detection from SMRI and Rs-FMRI Images Using Machine Learning. J. Intell. Inf. Syst. 2021, 57, 395–418. [Google Scholar] [CrossRef]
  33. Shen, G.; Jia, J.; Nie, L.; Feng, F.; Zhang, C.; Hu, T.; Chua, T.-S.; Zhu, W. Depression Detection via Harvesting Social Media: A Multimodal Dictionary Learning Solution. In Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 3838–3844. [Google Scholar]
  34. Li, B.; Zhu, J.; Wang, C. Depression Severity Prediction by Multi-Model Fusion. In Proceedings of the HEALTHINFO 2018, the Third International Conference on Informatics and Assistive Technologies for Health-Care, Medical Support and Wellbeing, Nice, France, 14 October 2018; pp. 19–24. [Google Scholar]
  35. Wang, Y.; Wang, Z.; Li, C.; Zhang, Y.; Wang, H. A Multimodal Feature Fusion-Based Method for Individual Depression Detection on Sina Weibo. In Proceedings of the 2020 IEEE 39th International Performance Computing and Communications Conference, IPCCC 2020, Austin, TX, USA, 6–8 November 2020; pp. 1–8. [Google Scholar]
  36. Pessanha, F.; Kaya, H.; Akdag Salah, A.A.; Salah, A.A. Towards Using Breathing Features for Multimodal Estimation of Depression Severity. In Proceedings of the 2022 International Conference on Multimodal Interaction, Bengaluru, India, 7–11 November 2022; pp. 128–138. [Google Scholar]
  37. Zhou, L.; Liu, Z.; Shangguan, Z.; Yuan, X.; Li, Y.; Hu, B. TAMFN: Time-Aware Attention Multimodal Fusion Network for Depression Detection. IEEE Trans. Neural Syst. Rehabil. Eng. 2023, 31, 669–679. [Google Scholar] [CrossRef] [PubMed]
  38. Xu, X.; Peng, H.; Bhuiyan, M.Z.A.; Hao, Z.; Liu, L.; Sun, L.; He, L. Privacy-Preserving Federated Depression Detection from Multisource Mobile Health Data. IEEE Trans. Ind. Inform. 2022, 18, 4788–4797. [Google Scholar] [CrossRef]
  39. Bn, S.; Abdullah, S. Privacy Sensitive Speech Analysis Using Federated Learning to Assess Depression. In Proceedings of the ICASSP 2022—2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 6272–6276. [Google Scholar]
  40. Cui, Y.; Li, Z.; Liu, L.; Zhang, J.; Liu, J. Privacy-Preserving Speech-Based Depression Diagnosis via Federated Learning. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS, Glasgow, UK, 11–15 July 2022; pp. 1371–1374. [Google Scholar]
  41. Pranto, M.A.M.; Al Asad, N. A Comprehensive Model to Monitor Mental Health Based on Federated Learning and Deep Learning. In Proceedings of the 2021 IEEE International Conference on Signal Processing, Information, Communication and Systems, SPICSCON 2021, Dhaka, Bangladesh, 3–4 December 2021; pp. 18–21. [Google Scholar]
  42. Asad, N.A.; Pranto, A.M.; Afreen, S.; Islam, M. Depression Detection by Analyzing Social Media Posts of User. In Proceedings of the 2019 IEEE International Conference on Signal Processing, Information, Communication & Systems (SPICSCON), Dhaka, Bangladesh, 28–30 November 2019; pp. 13–17. [Google Scholar]
  43. Alsagri, H.S.; Ykhlef, M. Machine Learning-Based Approach for Depression Detection in Twitter Using Content and Activity Features. IEICE Trans. Inf. Syst. 2020, E103D, 1825–1832. [Google Scholar] [CrossRef]
  44. Prakash, A.; Agarwal, K.; Shekhar, S.; Mutreja, T.; Chakraborty, P.S. An Ensemble Learning Approach for the Detection of Depression and Mental Illness over Twitter Data. In Proceedings of the 2021 8th International Conference on Computing for Sustainable Global Development, INDIACom 2021, New Delhi, India, 17–19 March 2021; pp. 565–570. [Google Scholar]
  45. Baltrušaitis, T.; Ahuja, C.; Morency, L.-P. Multimodal Machine Learning: A Survey and Taxonomy. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 41, 423–443. [Google Scholar] [CrossRef]
  46. Guo, W.; Wang, J.; Wang, S. Deep Multimodal Representation Learning: A Survey. IEEE Access 2019, 7, 63373–63394. [Google Scholar] [CrossRef]
  47. Cai, H.; Gao, Y.; Sun, S.; Li, N.; Xiao, H.; Li, J.; Yang, Z.; Li, X.; Zhao, Q.; Liu, Z.; et al. MODMA Dataset: A Multi-Modal Open Dataset for Mental-Disorder Analysis. arXiv 2020, arXiv:2002.09283. [Google Scholar]
  48. Shi, Q.; Liu, A.; Chen, R.; Shen, J.; Zhao, Q.; Hu, B. Depression Detection Using Resting State Three-Channel EEG Signal. arXiv 2020, arXiv:2002.09175. [Google Scholar]
  49. Li, Y.; Shen, Y.; Fan, X.; Huang, X.; Yu, H.; Zhao, G.; Ma, W. A Novel EEG-Based Major Depressive Disorder Detection Framework with Two-Stage Feature Selection. BMC Med. Inform. Decis. Mak. 2022, 22, 209. [Google Scholar] [CrossRef] [PubMed]
  50. Liu, Z.; Wang, D.; Zhang, L.; Hu, B. A Novel Decision Tree for Depression Recognition in Speech. arXiv 2020, arXiv:2002.12759. [Google Scholar]
Figure 1. Proposed methodology workflow.
Figure 1. Proposed methodology workflow.
Diagnostics 14 00043 g001
Figure 2. Multimodal Federated Learning based MDD detection using EEG and audio data.
Figure 2. Multimodal Federated Learning based MDD detection using EEG and audio data.
Diagnostics 14 00043 g002
Figure 3. IID and non-IID data distribution of the EEG and audio dataset.
Figure 3. IID and non-IID data distribution of the EEG and audio dataset.
Diagnostics 14 00043 g003
Figure 6. Confusion matrix of DL models with the EEG dataset.
Figure 6. Confusion matrix of DL models with the EEG dataset.
Diagnostics 14 00043 g006
Figure 7. Training and Validation results using DL models for EEG data.
Figure 7. Training and Validation results using DL models for EEG data.
Diagnostics 14 00043 g007aDiagnostics 14 00043 g007b
Figure 8. Training and validation results for multimodal DL using audio and EEG data.
Figure 8. Training and validation results for multimodal DL using audio and EEG data.
Diagnostics 14 00043 g008
Figure 9. FL multimodal training and validation results using IID data.
Figure 9. FL multimodal training and validation results using IID data.
Diagnostics 14 00043 g009aDiagnostics 14 00043 g009b
Figure 10. FL multimodal training and validation results using non-IID data.
Figure 10. FL multimodal training and validation results using non-IID data.
Diagnostics 14 00043 g010
Table 1. Comparative analysis of previous studies’ results.
Table 1. Comparative analysis of previous studies’ results.
ReferencesData TypeMethod AppliedMulti-ModalityData Privacy Parameters
[23]Speech dataDeep Neural Network ArchitectureNoNo 96.7%
[24]X, formerly Twitter dataHybrid DL ModelNo No F1 score 89%
[25]EEG data1D CNN ModelNo No Accuracy 90.5%
[26]EEG dataDL ModelsNo No Accuracy 99.24%
[27]Text dataDeep Learning ModelNo No Accuracy 99%
[28]EEG dataDeep Belief Network (DBN) ModelNoNo Accuracy 83.16%, 86.09% under binary and multiple classes
[29]X, formerly Twitter dataDL ModelsNo No Got high accuracy of 98% with CNN
[30]Reddit dataEL ModelNo No Accuracy 98.05%
[31]Questionnaire dataAI-based Decision Support SystemNo No Accuracy 89%
Table 2. Depression detection using a Multimodal System.
Table 2. Depression detection using a Multimodal System.
ReferencesFeatures ExtractedMethod Applied Data Privacy Parameters
[33]Social networks, visual, emotional, and user profileA Multimodal Dictionary LearningNo F1-measure is 85%
[34]Audio, video, and text data Multi-model Fusion No Root Mean Square Error is 5.98
[35]Text, picture, and behaviour Multimodal Feature Fusion Network No F1-score is 0.9685
[36]Audio and text Cross Dataset Prediction NoRoot Mean Square Error is 5.62 using Lenear Regression model
[37]Audio and imageTime-Aware Attention-based
Multimodal Fusion Depression Detection Network
No F1-score is 0.75
Table 3. FL-based depression detection.
Table 3. FL-based depression detection.
ReferencesData TypeMethod Applied Multi-ModalityData Privacy Parameters
[38]Multi-source mobile health dataFL ModelNo Yes Accuracy 85.13%
[39]Speech dataFL Method No Yes Accuracy 87%
[40]Audio dataFL Model No Yes Accuracy with IID 86.3% and with non-IID 85.3%
[41]Text dataFL FrameworkNo Yes Accuracy 93.46%
Table 4. The details of participants from the audio and 3-channel EEG dataset.
Table 4. The details of participants from the audio and 3-channel EEG dataset.
Subject TypeAge
(in Years)
Gender
MaleFemale
Has Depression16–561511
No Depression18–551910
Table 5. Result table of DL algorithms for audio data.
Table 5. Result table of DL algorithms for audio data.
ParametersBi LSTMCNNLSTM
Accuracy99.0833399.6666799.33333
Val Accuracy918589
Precision99.0833399.6666799.33333
Val Precision918589
Recall99.0833399.6666799.33333
Val Recall918589
Loss0.0323760.0129260.015737
Val Loss0.3490780.4395540.347897
Table 6. Result table of DL algorithms for EEG data.
Table 6. Result table of DL algorithms for EEG data.
ParametersBiLSTMCNNLSTM
Accuracy99.2859898.9941799.19285
Val Accuracy98.9570496.9207898.93221
Precision99.2859898.9941799.19285
Val Precision98.9570496.9207898.93221
Recall99.2859898.9941799.19285
Val Recall98.9570496.9207898.93221
Loss0.0193030.026460.023499
Val Loss0.0439290.097950.047406
Table 7. Result table of DL multimodal system.
Table 7. Result table of DL multimodal system.
ParametersDL Multimodal
Accuracy99.99411
Validation Accuracy100
Precision99.97054
Validation Precision100
Recall99.95877
Validation Recall100
Loss0.00047
Validation Loss0.000117
Table 8. Result table of the FL multimodal system for IID settings.
Table 8. Result table of the FL multimodal system for IID settings.
IID Settings
ParametersFL Multimodal
Accuracy99.94
Validation Accuracy99.93
Precision99.97
Validation Precision99.9
Recall99.88
Validation Recall99.94
Loss0
Validation Loss0
Table 9. Result table of the FL multimodal system for non-IID settings.
Table 9. Result table of the FL multimodal system for non-IID settings.
Non-IID Settings
ParametersFL Multimodal
Accuracy99.99
Validation Accuracy99.97
Precision99.99
Validation Precision99.96
Recall99.99
Validation Recall99.95
Loss0
Validation Loss0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gupta, C.; Khullar, V.; Goyal, N.; Saini, K.; Baniwal, R.; Kumar, S.; Rastogi, R. Cross-Silo, Privacy-Preserving, and Lightweight Federated Multimodal System for the Identification of Major Depressive Disorder Using Audio and Electroencephalogram. Diagnostics 2024, 14, 43. https://doi.org/10.3390/diagnostics14010043

AMA Style

Gupta C, Khullar V, Goyal N, Saini K, Baniwal R, Kumar S, Rastogi R. Cross-Silo, Privacy-Preserving, and Lightweight Federated Multimodal System for the Identification of Major Depressive Disorder Using Audio and Electroencephalogram. Diagnostics. 2024; 14(1):43. https://doi.org/10.3390/diagnostics14010043

Chicago/Turabian Style

Gupta, Chetna, Vikas Khullar, Nitin Goyal, Kirti Saini, Ritu Baniwal, Sushil Kumar, and Rashi Rastogi. 2024. "Cross-Silo, Privacy-Preserving, and Lightweight Federated Multimodal System for the Identification of Major Depressive Disorder Using Audio and Electroencephalogram" Diagnostics 14, no. 1: 43. https://doi.org/10.3390/diagnostics14010043

APA Style

Gupta, C., Khullar, V., Goyal, N., Saini, K., Baniwal, R., Kumar, S., & Rastogi, R. (2024). Cross-Silo, Privacy-Preserving, and Lightweight Federated Multimodal System for the Identification of Major Depressive Disorder Using Audio and Electroencephalogram. Diagnostics, 14(1), 43. https://doi.org/10.3390/diagnostics14010043

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop