Next Article in Journal
Immunological Interactions between Intestinal Helminth Infections and Tuberculosis
Previous Article in Journal
PR3-ANCAs Detected by Third-Generation ELISA Predicts Severe Disease and Poor Survival in Primary Sclerosing Cholangitis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Scoping Review on the Multimodal Classification of Depression and Experimental Study on Existing Multimodal Models

1
Faculty of Electrical Engineering and Computer Science, The University of Maribor, 2000 Maribor, Slovenia
2
Department of Psychology, Faculty of Arts, The University of Maribor, 2000 Maribor, Slovenia
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(11), 2683; https://doi.org/10.3390/diagnostics12112683
Submission received: 31 August 2022 / Revised: 27 October 2022 / Accepted: 31 October 2022 / Published: 3 November 2022
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Depression is a prevalent comorbidity in patients with severe physical disorders, such as cancer, stroke, and coronary diseases. Although it can significantly impact the course of the primary disease, the signs of depression are often underestimated and overlooked. The aim of this paper was to review algorithms for the automatic, uniform, and multimodal classification of signs of depression from human conversations and to evaluate their accuracy. For the scoping review, the PRISMA guidelines for scoping reviews were followed. In the scoping review, the search yielded 1095 papers, out of which 20 papers (8.26%) included more than two modalities, and 3 of those papers provided codes. Within the scope of this review, supported vector machine (SVM), random forest (RF), and long short-term memory network (LSTM; with gated and non-gated recurrent units) models, as well as different combinations of features, were identified as the most widely researched techniques. We tested the models using the DAIC-WOZ dataset (original training dataset) and using the SymptomMedia dataset to further assess their reliability and dependency on the nature of the training datasets. The best performance was obtained by the LSTM with gated recurrent units (F1-score of 0.64 for the DAIC-WOZ dataset). However, with a drop to an F1-score of 0.56 for the SymptomMedia dataset, the method also appears to be the most data-dependent.

1. Introduction

Depression is a global public health concern with an alarming rise in prevalence [1]. In fact, the global prevalence of depression was estimated at 28% in 2021 [2]. Furthermore, depression is one of the most common comorbidities in patients with severe physical disorders [3] (e.g., cancer, cardiovascular, neurological, and inflammatory disorders). Comorbid depression has a significant negative impact on a patient’s quality of life (QoL) [4,5,6] and is associated with poor adherence to medical treatments. Additionally, it involves higher rates of healthcare expenditure [3,7].
Although depression is common in comorbidity and multimorbidity patterns, it is often underestimated and greatly undertreated [8], even in higher-income countries [9]. Since depression is often associated with other clinical conditions such as anxiety and pain syndromes, the signs are often overlooked during daily medical care by professionals without specific training in mental health [10]. There are also structural barriers related to the availability, affordability, time constraints, and attitudinal factors of patients, e.g., non-treatment-seeking behavior [11], which significantly contribute to late diagnosis.
Smartphones and ‘mHealth’ apps can overcome several of the barriers mentioned above There is growing evidence that remote sign and symptom monitoring of various diseases and disorders can significantly improve health outcomes [12]. The ‘mHealth’ apps represent the foundation of digital biomarkers such as risk, diagnostic, monitoring, prognostic, predictive, or response markers [13]. Since measurements can be performed in real time, as a momentary assessment, these biomarkers can be less subjected to recall bias than retrospective questionnaires completed at the clinic [14]. The latter are also prone to reporting bias; for instance, self-reports of negative mood states experienced in the past tend to be exaggerated in a negative direction [15]. Moreover, social desirability and adherence can further distort the quality of the information provided by using self-report questionnaires [16]. Hence, psychiatry and other relevant fields may improve their screening processes and care by complementing subjective indicators (such as validated clinical questionnaires) with objective measures (such as those gathered via ‘mHealth’ apps) [17].
There is growing interest regarding the exploitation of observable cues [18], such as language use, speech, and facial expressions, in the screening and prediction of mental health problems [19]. It is a well-known fact that a variety of observable cues are affected by being in a depressed state [20,21]. Among non-verbal cues, facial muscle movements [22], slower pupil dilation [23], changes in vocal tract dynamics [24], and various other cues are characteristic of depression [22]. Since these cues are generated spontaneously, the use of AI-supported classification can contribute to improving the objectivity of the assessment and monitoring of the signs of depression [22] and other psychiatric disorders [25].
The main motivation of this scoping review was to identify and evaluate the existing algorithms for the classification of signs of depression from multiple observable cues, simultaneously expressed using the three conversational modalities [26]—language (i.e., verbal), speech (i.e., vocal), and facial expressions (i.e., visual)—to complement data gathered with self-report questionnaires. The work within this paper is driven to address the following research questions: (Q1) “What artificial models for depression classification analyse all three groups of observable cues?” (Q2) “How do the multimodal algorithms perform compared to unimodal and bimodal implementations?” (Q3) “How, if at all, does the use of multiple multimodalities mitigate the performance degradation in unfavorable datasets?” While the findings of our review represent an important contribution to the literature in the broadest sense, they will also be applied in digital interventions implemented in the multicentric single-case experimental prospective study of the PERSIST project [27]. The overall objective of PERSIST is to leverage the potential of big data and advanced data analytics, which can improve cancer survivors’ treatment and follow-up decision-making and engagement. PERSIST is particularly concerned with patient-reported outcomes, patient satisfaction scoring, identifying high-risk patients and giving them the appropriate treatment, and supporting patients who are at high risk of reoccurrence. Depression classification algorithms, the central topic of this review, will be used in PERSIST to derive complementary information about the mental health of patients and to offer personalized feedback by analyzing the videos of diary recordings, which will be recorded by patients via smartphone. We build our research on the assumption that depression is likely expressed through all three modalities of communication continuously and simultaneously [18]. Hence, the prevailing unimodal and bimodal approaches to the AI-based recognition of signs of depression may be inadequate [28] Depression has a complex structure in general due to having a wide array of potential symptoms that can manifest slightly differently for each individual, depending on their current and past personal psychosocial context. Thus, the fusion of multimodal features is needed to create a discriminative and explanatory approach able to deal with the complex nature of depression. However, to our knowledge, the existing reviews have focused on biochemical studies [29], mental health applications and monitoring systems [8,30,31], and single observable cues or bimodal approaches [22,25,32]. Furthermore, the existing studies have revealed relatively high accuracy [31,33]. However, they should only be considered as proof-of-concept studies, without addressing any specific issues such as decision errors, data bias and transparency, and other ethical concerns. The data have the main role as far as biases in the process are concerned. Decision errors (especially false negatives) and data bias are some of the main issues of machine learning algorithms. From a clinical perspective, those biases could lead to improper clinical decisions and may cause severe consequences. The main sources of unwanted biases are either choosing the wrong subjects for the training dataset or the wrong factors for training. In any case, both situations may cause fatal errors and challenges in medical situations [34].
The outline of this scoping review is as follows. The six-stage methodological framework is introduced in Section 2. The illustration and analyses of the experimental results are presented in Section 3. Finally, the key findings, challenges, and limitations are explained in Section 4.

2. Materials and Methods

2.1. Overview

In the preparation of the study, we followed the six-stage methodological framework for scoping reviews outlined by Arksey and O’Malley [35], as follows: identifying the research questions; searching for relevant studies; selecting the relevant studies; charting the data; collating, summarizing, and reporting the results; consulting the study findings. PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses—Extension for Scoping Reviews) [36] guidelines were followed to guarantee that the scoping review was systematic, transparent, and complete.
Beyond the scoping review framework, we applied the identified algorithms (with the available code) to two depression datasets to evaluate the accuracy of the AI-based processing of multimodal features for the classification of depression. The performance of the algorithms was evaluated using standard metrics.

2.2. Research Questions

Our scoping review was guided by these research questions derived from the main aim of the review:
  • To what extent do the papers include the different types of modalities for depression classification?
  • What types of algorithms currently exist for depression classification for observable cues and what is the performance of those algorithms?
  • What is the effect of data bias on the results of existing algorithms and the contributions of the multimodal approach to decrease the bias?

2.3. Search Strategy

Four large and commonly used databases, Web of Science (WOS), IEEE Xplore, SCOPUS, and ACM Digital Library, were used to identify the relevant papers. After a preliminary search in all databases, which helped us to refine the search strategy and ensure that the databases provided adequate coverage of the research topic, we conducted the main search between 14 October and 21 October 2021.
Our search strategy combined terms related to depression (depression, depressed), cues (cue, feature, indicator, marker, sign, signal, symbol, pattern, style, clue, manifestation, expression), features (“feature extraction”, speech, text, audio, video, face, visual), emotion (emotion, “emotion recognition”), algorithms (“explainable AI”, “deep learning”, “machine learning”, supervised, unsupervised), and data fusion (“feature fusion”, “data fusion”, “decision fusion”) (Box 1).
Box 1. The exact search string for Scopus.
TITLE-ABS-KEY (“Depression”) AND TITLE-ABS-KEY (“Feature extraction” OR speech OR text OR audio OR video OR face OR visual) AND TITLE-ABS-KEY (cue OR feature OR indicator OR marker OR sign OR signal OR symbol OR pattern OR style OR clue OR manifestation OR expression OR emotion OR “emotion recognition” OR “Feature fusion” OR “data fusion” OR “decision fusion”) AND TITLE-ABS-KEY (“Explainable AI” OR “Deep learning” OR “machine learning” OR “supervised” OR “unsupervised”) AND (LIMIT-TO (LANGUAGE, “English”)) AND (LIMIT-TO (DOCTYPE, “ar”) OR LIMIT-TO (DOCTYPE, “cp”) OR LIMIT-TO (DOCTYPE, “ch”))
The inclusion criteria (being available in English and being published, i.e., pre-prints and other unpublished papers were not considered) were derived from the research questions guiding this scoping review and were set a priori. The exclusion criteria included non-technical papers including reviews, studies that included human participants who suffered from other disorders that affect behavior and displays of emotions (i.e., dementia, Parkinson’s disease, autism, schizophrenia, Alzheimer’s disease, neurological disorder, stroke, Huntington’s disease, paralysis, autism, multiple sclerosis, cerebral palsy, Down syndrome), studies that focused on psychological disorders other than depression (e.g., anxiety, insomnia), and studies that offered classifications based on other observable features of depression (i.e., ECG, EEG, fMRI). Additionally, for the qualitative synthesis or experimental part of the paper, studies that included algorithms not operating over all three channels of communication simultaneously were excluded.

2.4. Study Selection

All citations identified in the electronic databases were exported to Excel spreadsheets. The database searches yielded a total of 1082 papers, and 13 additional papers were identified through other sources (Figure 1). After duplicates were removed, 775 titles and abstracts were screened in three stages. In the first stage, the authors (I.M. and U.A.) individually excluded highly irrelevant papers. In the second stage, all authors (U.A., I.M., U.S., and N.P.) independently reviewed the titles and abstracts of the remaining citations and settled disagreements through discussion. After this step, 245 papers underwent a thorough full-text review independently by two authors (I.M. and U.A.). Disagreements in this phase were settled through discussion and adjudication by the remaining authors (U.S. and N.P.). In the end, 20 papers fulfilled the pre-determined criteria and were included in the scoping review. In the 20 papers, we identified four different approaches that processed the three communication channels simultaneously and corresponding pre-trained models that were found on public repositories (i.e., GitHub). The four approaches were implemented and evaluated by three authors (I.M., U.A., and U.S.) in experiments using DAIC-WOZ [37] and Media datasets [38] to assess the data bias.

2.5. Data Collection and Charting

We developed a spreadsheet to determine variables to be extracted from the reviewed papers. Two authors (I.M. and U.A.) then extracted the following data from each paper, as follows: (1) authors; (2) year of publication; (3) type of paper; (4) targeted diseases; (5) exclusion criteria; (6) modality (e.g., text, audio, visual cues); (7) AI/ML technique (e.g., CNN, LSTM, RF, SVM). During this process, the results were categorized into the following categories: (1) unimodal; (2) bimodal; (3) multimodal. The processes for data extraction and the categorization of results were refined and updated in an iterative process as the papers were reviewed. The resulting chart was used for the analysis.

2.6. Collating, Summarizing, and Reporting Results

We did not follow a well-structured meta-analytic approach in comparing and summarizing the results given the aim and nature of the scoping review. Since the aim of a scoping review includes mapping the existing findings on a topic, providing a descriptive overview, and (re)evaluating the algorithms on a proprietary dataset [39], the results were analyzed by I.M. and U.A. using a thematic analysis [40] for the multimodal category. To maintain rigor in collating and summarizing the results, this process was reviewed by U.S. and N.P.

2.7. Consultation Exercises

The review protocol of the study was developed with the help of the expertise from psychological (U.S., N.P.), technological (I.M., U.A.), and methodological (I.M., U.S., N.P.) fields. Relevant inclusion and exclusion criteria (e.g., specific conditions or comorbidities that could impact the expression of depression) were determined by consulting the healthcare professionals in the PERSIST and HosmartAI project.

3. Results

3.1. Study Characteristics of Scoping Review

Out of 242 eligible papers indicated after screening phase, 179 papers described unimodal algorithms (Figure 1), processing only text (n = 78), only acoustic cues (n = 66), or only visual cues (n = 35). 38 papers were found processing two communication channels simultaneously (i.e., bimodal algorithms); most of them considered acoustic and visual cues (n = 19), followed by text and acoustic cues (n = 8), acoustic and physiological cues (n = 3), visual and physiological cues (n = 3), text and visual cues (n = 2), text and physiological cues (n = 2), and different combinations of two modalities (n = 1). Twenty-five papers were found to process all three modalities, out of which 5 papers processed audio and visual cues alongside physiological biomarkers and 20 papers simultaneously processed language, speech, and visual cues (Figure 2). Since we were interested in algorithms capable of analyzing symptoms of depression from videos of diary recordings, we excluded those processing physiological biomarkers and those that did not process all of the modalities. The final 20 selected papers are summarized in Table 1.

3.1.1. Data and Input Pre-Processing

Different modalities require different types of or approaches to input pre-processing. The most widely used techniques for data pre-processing are the linear discriminant analysis (LDA), synthetic minority oversampling technique (SMOTE), principal component analysis (PCA), and hidden Markov models (HMM) [60]. The LDA is used to remove redundant features by transforming them from a spatial space onto a lower-dimensional space. Using the LDA, the dataset can retain the most important features and can achieve higher class separability [61]. The PCA is an alternative technique to reduce the dimensionality of the datasets [62,63]. The PCA rotates the features in a way that possibly correlated features are transformed into a set of linearly uncorrelated features; that is, when orthogonal rotation is applied [64]. The SMOTE is used to generate a balanced class distribution by using an underrepresented dataset with synthetic patterns [65]. Most existing datasets, including DAIC-WOZ, are unbalanced, with depressive samples significantly being underrepresented [66]. HMMs are probabilistic models that are used to explain or derive the probabilistic characteristics of any random process. Using HMMs, the input data can be modelled as a series of outputs generated by multiple internal states [67]. As an alternative, evolutionary algorithms [64] and unsupervised autoencoders [68] were deployed to further optimize the datasets.

3.1.2. Feature Extraction

Feature selection represents a step in the process in which a subset of variables (features) is selected to most accurately predict the targeted variable. Among the more popular approaches are the clustering (e.g., K-means or its variations), particle swarm optimization (PSO), RELIEFF, multivariate processing, and Boruta algorithm approaches [69]. The K-best approach is a K-nearest neighbor-based clustering feature selection technique. It is non-parametric and univariate in nature [69]. It is used to select K-best features from the feature set using an univariate statistical test. Particle swarm optimization is a computational technique that is used for optimizing non-linear functions, clustering, and feature selection [70]. PSO-based clustering and wrapper-based PSO were, for instance, proposed to improve the discrimination ability of the extracted speech features and to enhance the accuracy of the speaker-independent multiclass emotion recognition process by selecting only discriminative features [71]. RELIEFF [72] is one of the most successful filtering feature selection methods. In the algorithm, the quality of a feature is based on its capacity to differentiate samples that are near to each other. Minimum redundancy and maximum relevance (mRMR) is a multivariate processing approach where features are ranked based on their relevance to the target variable. The feature with the maximum relevance and minimum redundancy gets the highest rank in the mRMR process [73]. The Boruta algorithm is build based on a random forest classifier [74]. It works based on an iterative methodology to remove irrelevant attributes.

3.1.3. AI/ML Techniques in the Reviewed Studies

CNN: In this review, it was found that the CNN technique is the most preferred AI/ML technique for the classification of depression (in 7 studies). CNN is one of the feed-forward NNs that can be used for many different fields (clustering, classification, regression, prediction, optimization, etc.). The CNN architecture is successfully applied in different ways to many speech recognition, emotion, and depression classification applications because of having powerful properties for capturing mixed or multimodal non-linear relationships among variables. Within our reviewed papers, [41,43] used a CNN as a depression estimation module for hand-crafted features, [46] used deep neural networks with attention fusion networks for the estimation of PHQ-8 scores, [50] proposed a hybrid deep learning model for different types of inputs and fused the outputs in a fusion network, and [58] used a causal CNN structure to model long sequences of patient video recordings.
SVM: The SVM technique is one of the most popular methods for depression classification due to having well-fitted binary issues of high dimensionality and suitability for non-linear kernel function settings [75]. Within our reviewed papers, [41] used an SVM as a classifier for psychoanalytic symptom presence identification, [28,45,53] built a depression detection model on an SVM as one of the regression models, and [57] investigated the different fusion types with SVMs.
RF: The RF technique obtains a result by using the collection of outputs of multiple decision trees. The main advantages of the RF technique are its easy usage and adaptation flexibility for both classification and regression issues. Within our reviewed papers, [28,41] used the RF technique for depression classification modeling, [42] translated features into predictive scores via RF, and [56] predicted the PHQ-8 scores an using RF model.
LSTM: LSTMs are used mainly for retaining information for longer periods for both classification and regression, but especially for sequence prediction issues. In a bidirectional LSTM (BiLSTM), each training sequence is applied in both directions (forward and backward) to separate recurrent nets. If you are using LSTM with multimodal features, the contribution of text features can be diminished via audio and visual modalities. To overcome this problem, a gating mechanism is used to control the contribution of each modality to the last decision (for more details, see Section 3.2.3). Within our reviewed papers, [43] utilized an LSTM by processing long-term dependencies, recognizing context information, and dealing with recent information; [44,51,52] used a BiLSTM for depression recognition; [59] utilized an LSTM to predict PHQ-8 scores; and [52] applied a gating mechanism to their network.
DT: The decision tree is one of the most popular and powerful algorithms for both classification and regression. Its structure is similar to a tree’s structure in that the branches show the results of the test and the leaves show the labels. Within our reviewed papers, [49] used predicted PHQ-8 scores and participant characteristics to construct decision tree, while [55] used a decision tree for multimodal fused features to classify the presence of depression.
MV-R: Multivariate regression is mainly used when the model has more than one response and one explanatory variable. Multivariate regression is used as a supervised machine learning algorithm and can be applied to different scientific disciplines. Within our reviewed papers, [48] used the Gaussian staircase model to generalize the distributions to be used in the binary classification of depression via multivariate regression.

3.1.4. Early or Late Fusion

Fusion involves joining information from at least two or more different modalities such as audio, text, and video into a consistent representation to predict the assessment result where there are complementary expressions such as depression. Fusing different modalities increases the chance of obtaining higher precision by overcoming the shortages of individual modalities or interactions of multimodal features. All automatic depression classification models have five common stages, namely the input, pre-processing, feature extraction, machine learning, and output stages. Fusion can be applied either at the feature level (early fusion) or at the decision level (late fusion). In early fusion (feature level), multimodal features are involved in a merged feature set to be fed into the machine learning algorithms. Early fusion has different opportunities to be applied before or after the dimensionality reduction step in the feature extraction stage [76,77,78]. In late fusion (decision level), the results of different machine learning algorithms are combined and processed with a decision fusion algorithm to obtain the final decision.
In our scoping review, 9 papers (45%) applied late fusion, 6 papers (30%) applied early fusion, and 5 papers (25%) applied hybrid fusion (early+late) in their assessments. Early fusion was applied mainly via feature selection via an importance factor, continuing with only the most informative ones and concatenating them into one vector (fusion) for the classification process. As a different strategy, [51] used an attention fusion network and [56] used the bidirectional LSTM method for the fusion of the features. Late fusion was implemented with different approaches such as the winner-takes-all strategy (taking the decision of the single modality with the highest confidence score) [42], ensemble prediction (where depression is classified if one of the modalities predicts the subject as being depressed) [43], and the temporal framework using a simple feed-forward NN [44] and two hierarchical BiLSTM designs for the integration of all features [51].

3.2. Experimental Study

Following the scoping review, we carried out complementary experiments on two different depression datasets: DAIC-WOZ [37] and SymptomMedia [38]. In this section, we describe the datasets, the list of the multimodal features, and the feature extraction process. Then, we explain the implementation of AI models with performance metrics (F1-score, precision, and recall) and error metrics (MSE, RMSE, and MAE). Finally, we conclude the section with the results of the depression classification process.

3.2.1. Datasets

DAIC-WOZ Database: This database is part of a larger clinical interviews corpus, the Distress Analysis Interview Corpus (DAIC), which is used for different types of psychological disorders, such as anxiety, depression, and post-traumatic stress disorder. A virtual agent (called Ellie and controlled by a person in a different place) was used in communicating with the patients to identify both verbal and non-verbal signs of psychological disorders [79]. There are 189 sessions and the training, development, and testing sets are pre-defined. Those sets include information about patient IDs, PHQ-8 scores, binary labels for PHQ-8 >= 10, and gender, and each session involves the transcript of the interaction, an audio file, audio features, and facial features.
SymptomMedia Dataset: This dataset contains guided film case studies about mental health (of diagnoses within the DSM-5 and ICD 10 classification systems), which is mainly used for continuing education for medical students, doctors, psychologists, counsellors, and other health care professionals. It contains more than 600 films of actors (actual patients were not used in the production of these films). The durations of the film range from 4 to 14 min. The recordings include only the appearance of the patient, while the interviewer is hidden. Moreover, a transcript is available for each film (for both interviewer and patient).
For our study, we selected 14 SymptomMedia recordings of individuals (i.e., professional actors) simulating different diagnoses of depression and 14 recordings of individuals simulating other mental health-related diagnoses that have a low probability of co-occurring together with major depression compared to other disorders [80,81] (Table 2). In total, 28 videos were selected to ensure an evenly balanced test set.

3.2.2. Multimodal Feature Extraction

To evaluate the AI models, the following features were extracted from the SymptomMedia films (Table 3).
The feature extraction was applied only for the SymptomMedia dataset because the DAIC-WOZ dataset provides the extracted features within the dataset. For the SymptomMedia dataset, the audio file was extracted from the video via the Ffmpeg multimedia framework [82] as a first step. Then, the audio features were extracted with the COVAREP (Collaborative Voice Analysis Repository) tool [83] using MATLAB [84] software; the visual features were extracted using the OpenFace toolkit [85] and the Word2Vec model [86] was used to transform the words in the transcription file into a 300-feature vector representation (sentences x words x 300 features) for text feature extraction (Figure 3).

3.2.3. AI Models and Evaluation Metrics

Based on Box 1, we selected SVM + RF with SVM late fusion, SVM + RF with RF late fusion, LSTM without gating, and LSTM with gating as models to be further investigated. To evaluate the models, we adapted the implementation proposed in [52] and related the code available via the GitHub repository [87]. In the following sections, we outline the implementation of each of the models.
SVM and RF: The SVM tries to predict the final output as an optimization problem. In that problem, it explores the best hyperplanes that maximize the size of the margin between two groups. By using the kernel function (K), an algorithm can easily find the most suitable hyperplane in a transformed feature space. In our case, the decision function for the test (x) has the following equation:
g x = i α i y i K x i , x b                                
where g(x) is the decision function, α is the learned weight constant, y is the class label, K is the kernel function used for training (xi) and testing (x) the sample, and b is the learned threshold value.
On the other hand, the random forest (RF) ensembles the decision trees using the bagging method (a combination of learning models). In this way, the RF gives more accurate and stable predictions. There are mainly four steps for an RF:
  • Take n number of random records from the dataset;
  • Construct decision trees for each sample;
  • Generate an output for each tree;
  • Apply majority voting or averaging for the final output.
Here, we wanted to show the different effects of SVM and RF on late fusion. Each modality was classified separately using SVM and RF models and then the late fusions were performed by applying another SVM or RF model to predict the values (Figure 4).
LSTMs with or without gating: The gating mechanism is used for learning the effects of visual and auditory modalities on lexical information. This mechanism is important for determining the noise quantity while combining the three modalities with different effects to obtain the final result. The main motivation for using the gating mechanism was to overcome the undermining problem with the learned representation of text using other modalities. By using a gating mechanism, the contributions of visual and auditory modalities to the final result can be controlled [52].
The model includes feed-forward highway layers [88] for use with the gating mechanism for audio and video features. Those layers have two non-linear transforms (the ‘carry gate (Cg)’ to measure the transformation degree and ‘transform gate (Tg)’ to determine the amount of information to move forward).
The definitions of the gates are as follows:
T r a n s f o r m   G a t e :   T g   = σ W T g I t + b T g      
where W is the weight matrix and b is the bias vector:
C a r r y   G a t e :   C g = 1 T g
The input vectors (It) are controlled by those two gates at each layer (feed-forward layer, H), as follows [52]:
y = T g   .   H + C g   .   I t
The audio and video features were applied to those layers separately and then the outputs were concatenated with the text features. The final output of the model was obtained by feeding the concatenated features to the LSTM model. The model architecture consists of 3 feed-forward highway layers for gating, dense layers (sigmoid activation function) for dimensionality reductions, a concatenation layer, and an LSTM with 128 hidden nodes (the learning rate is 0.0001 and EarlyStopping call-back is used) (Figure 5).
F1-score, recall, precision, MSE, RMSE, MAE: We used the standard metrics for the evaluation of the model performance [89]. The precision gives the ratio of the actual positive cases to all predicted positive cases and tries to answer the question, ‘How many retrieved items are relevant?’ A high precision value is expected, and this indicates that the model has a low false positive rate. The recall gives the ratio of the actual positive cases to all cases with an answer ‘yes’ and tries to answer the question, ‘How many relevant items are retrieved?’ Again, it is expected that the recall value should be above 0.5. Another metric, the F1-score, shows the weighted average of the precision and recall metrics. Using this metric, all false positives and false negatives are considered. The F1-metric becomes a more useful measure than the accuracy when there is an uneven class distribution in the model.
P r e c i s i o n = T r u e   P o s i t i v e s T r u e   P o s i t i v e s + F a l s e   P o s i t i v e s      
R e c a l l = T r u e   P o s i t i v e s T r u e   P o s i t i v e s + F a l s e   N e g a t i v e s    
F 1   S c o r e =   2 R e c a l l P r e c i s i o n R e c a l l + P r e c i s i o n
Other metrics, namely RMSE, MSE, and MAE, are defined in Equations (8)–(10). The MAE gives the mean value of the absolute values of each prediction error; the MSE mainly calculates the sum of the relative square error of each sample, while the RMSE is the square root of the MSE. While the RMSE and MSE show the standard deviation of the prediction errors, the MAE measures the average difference between predicted and actual depression values.
R M S E = 1 n i = 1 n P i O i 2              
M S E = 1 n i = 1 n P i O i 2          
M A E = 1 n i = 1 n P i O i      
For n number of interviews, Pi and Oi denote the predicted and actual depression values, respectively.

3.2.4. Experimental Results

In this section, we carried out experiments on both the DAIC-WOZ and SymptomMedia datasets. We trained the models with the DAIC-WOZ dataset’s training+development sets and performed a depression classification for the DAIC-WOZ dataset’s testing set and the SymptomMedia dataset (Table 4). While evaluating the algorithms on DAIC-WOZ and SymptomMedia datasets, we extracted the same features from both. The features were those used in the original algorithms and are listed in Table 3. For the evaluation of all models (SVM+RF with SVM late fusion, SVM + RF with RF late fusion, LSTM without gating, and LSTM with gating), we provided the following tables and figures for the use of both bimodal and multimodal features with performance and error metrics.
Table 5 reports the performance of the SVM+RF with SVM late fusion model for the use of bimodal and multimodal features for both the DAIC-WOZ and SymptomMedia datasets. Although the use of multimodal features gave better results (F1-score: 0.46; MSE: 0.54) against the T + A and T + V bimodal features for the DAIC-WOZ dataset, it showed better performance (F1-score: 0.4; MSE: 0.57) against only the A + V bimodal features for the SymptomMedia dataset. The best recall value for SVM+RF with the SVM late fusion model was obtained by the A + V (0.55) and T + A (0.46) bimodal features for DAIC-WOZ and SymptomMedia datasets, respectively. However, the T+A+V multimodal features gave the highest precision value, 0.61, for the DAIC-WOZ dataset, and again the T + A bimodal features gave the highest precision value, 0.46, for the SymptomMedia dataset. In general, the A + V bimodal features showed a better performance for the DAIC-WOZ dataset and the T + A bimodal features showed a better performance for the SymptomMedia dataset.
Similar results for the comparison between the use of bimodal and multimodal features were found for the SVM + RF with RF late fusion (Table 6). The A + V bimodal features showed the best performance (F1-score: 0.56; MSE: 0.45) for the DAIC-WOZ dataset and the T + A bimodal features showed the best performance (F1-score: 0.46; MSE: 0.53) for the SymptomMedia dataset. The best recall and precision values for the SVM + RF with the RF late fusion model were obtained by the A + V (recall: 0.45; precision: 0.61) and T + A (recall: 0.46; precision: 0.46) bimodal features for the DAIC-WOZ and SymptomMedia datasets, respectively. In general, the A + V bimodal features showed better performance for the DAIC-WOZ dataset and the T + A bimodal features showed better performance for the SymptomMedia dataset, as did the SVM+RF with SVM late fusion model.
Table 7 reports the performance of the LSTM without gating model for the use of bimodal and multimodal features for both the DAIC-WOZ and SymptomMedia datasets. In general, this model gave better results than both the SVM + RF with SVM late fusion and SVM + RF with RF late fusion models. All modalities gave similar F1-scores, but the lowest MSE score (0.29) was given by the T+V bimodal features for the DAIC-WOZ dataset. For the SymptomMedia dataset, the use of multimodal features and the T + A bimodal features gave the same values for the F1-score (0.56) and all modalities gave similar MSE scores. For the DAIC-WOZ dataset, all modalities performed similarly according to the recall and precision values. However, for the SymptomMedia dataset, the T + A bimodal features and T + A + V multimodal features performed better, with the same F1-score (0.56), precision (0.58), and recall (0.57) values.
The best results for the use of multimodal features (F1-score: 0.64; MSE: 0.32) were obtained with LSTM with gating model for the DAIC-WOZ dataset (Table 8). Additionally, similar performance results (F1-score: 0.48; MSE: 0.34) were obtained with the use of multimodal features according to the use of bimodal features for the SymptomMedia dataset. Again, the T + A + V multimodal features performed well, with the best recall (0.66) and precision (0.94) values for the DAIC-WOZ dataset, while the T + A + V multimodal features performed the same with other modalities, with good recall (0.5) and precision (0.5) values for SymptomMedia dataset.
The highest F1-score (0.64) was obtained for the use of multimodal features with the LSTM with gating model for the DAIC-WOZ dataset, while the highest F1-score (0.56) was obtained for the use of multimodal features and T + A bimodal features with the LSTM without gating model for the SymptomMedia dataset. For the error metrics, the lowest MSE score (0.27) was obtained for the T + A bimodal features with the LSTM without gating model for the DAIC-WOZ dataset, and the lowest MSE score (0.34) was obtained for the use of multimodal features with the LSTM with gating model for the SymptomMedia dataset.

4. Discussion

4.1. Key Findings

The aim of this study was two-fold: first, to review the extent to which papers exploit different types of modalities in the implementation of depression classifications together with the types of AI models currently existing and the percentages of multimodal (text, audio, and video) data usage in the AI models in the literature (i.e., research question Q1); second, to explore the effects of multimodality and data bias with an experimental study as complementary work to the scoping review and to explain the need for explainable AI models (i.e., research questions Q2 and Q3).
Although there are some scoping reviews in the literature about the use of AI for mental health (including depression) [90,91] and depression recognition [92,93], to our knowledge, our scoping review is the first review investigating the different types of modalities used for depression classification with different AI models. Beyond our scoping review, we also conducted an experimental study and described its results. As such, our study has a unique position in the literature. The existing scoping reviews about depression recognition are mainly focused on using only one modality (facial [22], speech [25]) or using different observation techniques (rather than audio, text, and video) [94].
To address research question Q1, we identified 242 research papers or studies in total, of which only 20 (8.26%) analyzed all three groups of observable cues during depression classification. Most of the identified studies (179, 73.96%) in the literature still used a unimodal architecture that could detect only one aspect of depression. On the other hand, 38 (15.7%) studies used a bimodal architecture and 5 (2.06%) studies used a multimodal approach with physiological features.
Another important finding from the scoping review was related to the features that are used in multimodal classifications. The majority of the algorithms explore similar features, as AI algorithms for multimodal depression classification studies need to have a large volume dataset to obtain more accurate and less biased results. Powerful and accurate AI algorithms need more parameters and variables to feed their models, which means more data. Combining more data into your model helps in two ways: first, it provides more raw variables to be used as features, and second, it allows for new derived variables to be made by combining existing ones [95]. On the other hand, including all features in the models that we can find is not the aim of AI studies. The best possible feature set should be selected for the best result, but this discussion is beyond the scope of our review. All reviewed studies except two of them used the same dataset (DAIC-WOZ) for training and testing. This dataset has only 189 sessions. On the other hand, the variety of AI/ML techniques should be increased to increase the accuracy and decrease the error rate.
Another issue with multimodal depression classification studies is not having enough open-source codes for researchers, app developers, and product teams. Open source in an AI algorithm means that it is available for both commercial and non-commercial usage under certain licenses. Open source is the key element for innovation through collaboration. Open-source AI systems may include datasets, algorithms, and user interfaces. According to the latest technology report [96], nowadays 48% of enterprises are using open-source codes, and this number will increase to 65% within two years. Within our reviewed studies, only 3 (15%) studies provided their codes via GitHub. This may be one of the main reasons for the smaller number of multimodal depression classification studies in the literature.
A similar issue occurs regarding the types of reviewed studies. Only 7 (35%) studies were published in scientific journals, while the rest were published in conference proceedings. This result shows that there are still opportunities and open areas to contribute scientifically to multimodal depression classification.
To address research questions Q2 and Q3, we carried out experiments on the most widely used multimodal architectures, i.e., LSTMs, SVMs, and RFs. In the experimental study, we performed the depression classification process with four different models (SVM, RF, LSTM with gating, and LSTM without gating models) and four different modalities (T + A, T + V, A + V, and T + A + V).
The key finding of the experimental study was that there was an obvious benefit of analyzing multiple modalities during depression classification. The results clearly showed that multimodal approaches have the potential to outperform uni- and bimodal approaches. Overall, to answer research question Q2, the results showed that all multimodal architectures give superior or at least similar results to bimodal models (Table 6, Table 7 and Table 8).
It is easily observed that the negative effect of the small volume of datasets (data bias) is apparent and valid for all models. This is the main reason behind having lower performance rates and higher error rates. We provide a comparison table about the reported performance values of uni-, bi-, and multimodalities to emphasize the contribution of each modality to the depression classification process. According to the reported results, multimodal feature sets provide the best performance against the use of uni- or bimodal features in all studies (Table 9). To answer research question Q3, the results clearly show that the accuracy of the algorithms, especially those operating on uni- or bimodal assumptions, tends to degrade when introduced to recordings outside the nature of the training dataset. A more detailed analysis of the impact of using multiple modalities, especially in unfavorable datasets, is represented in the following section.

4.2. Challenges

4.2.1. Decision Errors (Type I and Type II)

Misinterpretations or mistakes always exist within scientific studies because of their nature. The complete elimination of those decision errors is impossible, but controlling them is possible. There are two types of decision errors: false positives (type I errors), which show the decision as true while the actual case is false; and false negatives (type II errors), which show the decision as false when the actual case as true. To control and optimize these errors, we must understand the roles of false negatives and false positives specifically within our cases; while false negatives are preferable in some cases (e.g., criminal counts [97]), those errors can be very dangerous in different cases (e.g., clinical decisions [98], like in our case). If a patient with depression is not classified as depressed (false negative), then that patient may not get the proper clinical treatment and the mental disorder may get worse because of decision errors. False-positive cases for automatic depression classification have a less detrimental effect on people who have no depression at all and can be easily verified by other standardized tools such as questionnaires [99].
The F1-score is a more useful metric than other metrics because it considers the false negatives and is more meaningful when you have a non-symmetric dataset (Section 3.2.3.). On the other hand, the accuracy score gives the best performance when the burdens of the false negatives and false positives are similar. However, false negatives are more important than false positives concerning the consequences of different types of decisions in our case.
From the performance and error metrics tables (Table 3, Table 4, Table 5 and Table 6), the favorability of the datasets, variability between datasets, and variability among models can be compared according to the variability of the F1-scores. The A + T bimodal features were the only features that showed improvements for all four models between datasets. The A+V bimodal features showed the most variation (more data bias) in a negative direction, with the results varying depending on the favorability of the dataset. Our testing dataset, the SymptomMedia dataset, which is a high-quality dataset with high expressivity of the features, might be favorable for the bimodal features for SVM and RF approaches (higher F1-scores), and this shows that if the features are not so distinct within the dataset (the SymptomMedia dataset has higher F1-scores than the DAIC-WOZ dataset), the F1-score will significantly decrease, which is not reliable at all. On the other hand, the multimodal features showed minimal F1-score variations for all four models, which means that the usage of multimodal features gives more resilient and reliable results. When we compare the models, the SVM and RF models generated less variability than the other models.

4.2.2. Data Bias

The data bias is one of the important challenges for machine learning models. These models lack visibility because of having a ‘black box’ structure. If we are trying to solve a problem in the healthcare sector and want to use AI, the paramount factor to consider is the data itself. The data are the main and only source of bias, which tends to be influenced by multiple factors such as the patient population and the correlation between parameters [58]. Training without considering those factors results in the misrepresentation and misunderstanding of the phenomenon. The other source of data bias is not having enough data in the health domain, as in other fields such as computer vision and data mining. Even in our scoping review, the sizes of the datasets were not sufficient (DAIC-WOZ, 189 subjects; BDC WB, 51 subjects; AViD-C, 84 subjects). We selected those studies because they contained only multimodal depression features. Besides those features, other features can be gathered and added to the model from different devices such as EEG, ECG, and fMRI devices. If we take into account all of those features, then those may lead to another problem called ‘the curse of dimensionality’ [100]. In Section 3.1.1 and Section 3.1.2 methods, the LDA, PCA, clustering, and RELIEFF approaches are highlighted as practical solutions to at least partially overcome the challenge of dimensionality and to generate optimal sets of features.
Beyond those issues, another problem that causes bias with the DAIC-WOZ dataset is having binary labels based on PHQ-8 scores from the participants’ self-reports. These self-reports may depend on social, subjective, and other kinds of biases, and may affect the annotation accuracy [101]. Thus, taking the DAIC-WOZ dataset as a ground truth may not be the most accurate way to measure the depression status. As a result, those issues can be counted as additional potential reasons for the low performance values.
One of the ways to overcome these challenges is to use explainable and interpretable models. Explainable AI models provide transparency, fairness, and reliability in the process and results. The mitigation of bias in the model includes removing the effect of the bias. There are different techniques for mitigation that can be applied in each phase of the model (pre-processing, in-processing, and post-processing). For example, while the reweighing technique [102] involves changing the weights of the features before the training in the pre-processing phase, the adversarial debiasing technique [103] involves applying an optimization process using a GAN during the training process. On the other hand, the techniques used after the training phase have less pronounced effects. For example, the equalized odds [104] technique involves adjusting the thresholds to obtain smaller differences between the true and false results. In summary, it is shown that explainable AI techniques can be used for bias detection and that there are some useful tools that can be used for this, such as a fairness report [105], SHAP [106], and CERTIFAI [107].

4.3. Limitations of the Scoping Review

While this scoping review provides comprehensive insight into the cues of depression, which are observable from text, audio, and video sources, and contributes to the literature by presenting an additional experimental study to show the importance and effects of the use of multimodal features for depression classification, it does not provide a complete picture of the current state of the research on depression classification due to certain limitations. First, our scoping review did not include other observable features of depression from different devices (i.e., ECG, EEG, fMRI). Although those features may also contribute to the classification of depression, we may again be faced some problems related to the existing dataset and feature dimensionality. Second, we focused only on papers written in English. The studies in other languages may include additional information about the classification of depression. Third, we tried to show which method tends to be more resilient to data collected outside the training datasets in the experimental part of our study. Although the SymptomMedia recordings are not exactly categorized as real-world data, the data are of a different nature, and from the psychological perspective represent a high-quality simulation of real-world data. Fourth, as specified in the Discussion, one of the main issues of ML in the health domain is the volume of available datasets. In our experiments, we trained our models with 142 cases and tested them with 75 cases. Thus, our results, just like those of the reviewed studies, need to be validated with larger datasets. In the future, if new depression datasets emerge and are available as open-source data, we will try to increase the sizes of both the training and testing sets.

5. Future Work

This scoping review illustrates the state of the existing studies, their limitations, and the future directions for multimodal depression classification. More studies on multimodal depression classification should be carried out in the future to show the different aspects of the observable cues of depression. To increase the accuracy rates of the classification algorithms, high-volume depression datasets should be constructed and made available for researchers. With the advances in omics, genetics, and epigenetics, these datasets have become valuable and affordable sources of real-world data in multiple studies of depression classification [29,108]. With a stable and more distinct representation of depressive disorders, these data could significantly increase the discriminatory power of the AI algorithms. Although less accessible at the moment, future studies should analyze the correlation between genetic or epigenetic and observable features of depression. Furthermore, there is a lack of explainable artificial intelligence algorithms, which would significantly improve the interpretability of the results and expedite their integration into clinical routine. Additionally, explainable AI or at least hybrid model algorithms may offer a way to mitigate the data quantity requirements of deep models and to exploit the natural relationships (or interplay) between the modalities, as observed by experts in real-world manifestations of depression. The future research should focus more on the multimodal nature of human interactions and the explainability of AI decisions, as multimodal techniques are, according to our findings, more resilient to data bias. Finally, further clinical studies should be carried out to evaluate the performance of the algorithms in real-world settings, with real patients, and to compare the algorithms against standardized instruments.

Author Contributions

Conceptualization, U.A., U.S., N.P. and I.M.; methodology, I.M. and U.S.; software, U.A. and I.M.; validation, U.S. and N.P.; investigation, U.A., U.S., N.P. and I.M.; data curation, U.A. and I.M.; funding acquisition, I.M.; visualization, U.A.; writing—original draft preparation, I.M. and U.A.; writing—review and editing, U.A., U.S., N.P. and I.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work has received funding from the European Union Horizon 2020 research and innovation program, project PERSIST, supported under grant agreement no. 875406; project HosmartAI, supported under grant agreement no. 101016834; and Slovenian Research Agency (ARRS) project ‘Empowerment of ageing individuals: Self-regulatory mechanisms and support of digital technology in achieving higher quality of life’ (grant agreement No. J5-3120). The content of this paper does not reflect the official opinions of the funders or any other institution. The responsibility for the information and views expressed herein lies entirely with the authors.

Data Availability Statement

The dataset used and analyzed during the current study is available from the corresponding author on reasonable request. The DAIC-WOZ database can be obtained from the original authors (https://dcapswoz.ict.usc.edu/daic-woz-database-download/ accessed on 6 April 2022). The SymptomMedia dataset can be obtained from the original authors (https://symptommedia.com/ accessed on 31 March 2022). Please refer to the websites for more information.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

List of Abbreviations

AIArtificial Intelligence
AUsAction Units
AVEC2016Audio Visual Emotion Challenge
AViD-CAViD-Corpus
BDC-WBBipolar Disorder Corpus and Well-Being
BiLSTMBidirectional Long Short-Term Memory
CNNConvolutional Neural Network
DAIC-WOZDistress Analysis Interview Corpus Wizard-of-Oz
DSM 5The Diagnostic and Statistical Manual of Mental Disorders, Fifth Edition
DTDecision Tree
ECGElectrocardiogram
EEGElectroencephalography
fMRIFunctional Magnetic Resonance Imaging
ICD 10The International Classification of Diseases Tenth Edition
LSTMLong Short-Term Memory
MAEMean Absolute Error
MLMachine Learning
MSEMean Square Error
MV-RMultivariate Regression
NNNeural Network
PHQ-8The Eight-Item Patient Health Questionnaire Depression Scale
PRISMA-ScR(Preferred Reporting Items for Systematic Reviews and Meta-Analyses—Extension for Scoping Reviews)
QoLQuality of Life
RFRandom Forest
RMSERoot Mean Square Error
SVMSupport Vector Machines
TPR/TNRTrue Positive Ratio/True Negative Ratio
WOSWeb of Science

References

  1. Pfefferbaum, B.; North, C.S. Mental Health and the COVID-19 Pandemic. N. Engl. J. Med. 2020, 383, 510–512. [Google Scholar] [CrossRef] [PubMed]
  2. Nochaiwong, S.; Ruengorn, C.; Thavorn, K.; Hutton, B.; Awiphan, R.; Phosuya, C.; Ruanta, Y.; Wongpakaran, N.; Wongpakaran, T. Global prevalence of mental health issues among the general population during the coronavirus disease-2019 pandemic: A systematic review and meta-analysis. Sci. Rep. 2021, 11, 10173. [Google Scholar] [CrossRef] [PubMed]
  3. Gold, S.M.; Köhler-Forsberg, O.; Moss-Morris, R.; Mehnert, A.; Miranda, J.J.; Bullinger, M.; Steptoe, A.; Whooley, M.A.; Otte, C. Comorbid depression in medical diseases. Nat. Rev. Dis. Primers 2020, 6, 69. [Google Scholar] [CrossRef] [PubMed]
  4. Jenkins, P.E.; Ducker, I.; Gooding, R.; James, M.; Rutter-Eley, E. Anxiety and depression in a sample of UK college students: A study of prevalence, comorbidity, and quality of life. J. Am. Coll. Health 2021, 69, 813–819. [Google Scholar] [CrossRef] [PubMed]
  5. Sbolli, M.; Fiuzat, M.; Cani, D.; O’Connor, C.M. Depression and heart failure: The lonely comorbidity. Eur. J. Heart Fail. 2020, 22, 2007–2017. [Google Scholar] [CrossRef]
  6. Park, J.; Rodriguez, J.L.; O’Brien, K.M.; Nichols, H.B.; Hodgson, M.E.; Weinberg, C.R.; Sandler, D.P. Health-related quality of life outcomes among breast cancer survivors. Cancer 2021, 127, 1114–1125. [Google Scholar] [CrossRef]
  7. Kang, H.J.; Kim, S.Y.; Bae, K.Y.; Kim, S.W.; Shin, I.S.; Yoon, J.S.; Kim, J.M. Comorbidity of depression with physical disorders: Research and clinical implications. Chonnam Med. J. 2015, 51, 8–18. [Google Scholar] [CrossRef] [Green Version]
  8. Weisel, K.K.; Fuhrmann, L.M.; Berking, M.; Baumeister, H.; Cuijpers, P.; Ebert, D.D. Standalone smartphone apps for mental health—A systematic review and meta-analysis. NPJ Digit. Med. 2019, 2, 118. [Google Scholar] [CrossRef] [Green Version]
  9. Thornicroft, G.; Chatterji, S.; Evans-Lacko, S.; Gruber, M.; Sampson, N.; Aguilar-Gaxiola, S.; Al-Hamzawi, A.; Alonso, J.; Andrade, L.; Borges, G.; et al. Undertreatment of people with major depressive disorder in 21 countries. Br. J. Psychiatry 2017, 210, 119–124. [Google Scholar] [CrossRef] [Green Version]
  10. Wang, J.; Wu, X.; Lai, W.; Long, E.; Zhang, X.; Li, W.; Zhu, Y.; Chen, C.; Zhong, X.; Liu, Z.; et al. Prevalence of depression and depressive symptoms among outpatients: A systematic review and meta-analysis. BMJ Open 2017, 7, e017173. [Google Scholar] [CrossRef]
  11. Mojtabai, R.; Olfson, M.; Sampson, N.A.; Jin, R.; Druss, B.; Wang, P.S.; Wells, K.B.; Pincus, H.A.; Kessler, R.C. Barriers to mental health treatment: Results from the National Comorbidity Survey Replication. Psychol. Med. 2011, 41, 1751–1761. [Google Scholar] [CrossRef] [Green Version]
  12. Jim, H.S.; Hoogland, A.I.; Brownstein, N.C.; Barata, A.; Dicker, A.P.; Knoop, H.; Gonzalez, B.D.; Perkins, R.; Rollison, D.; Gilbert, S.M.; et al. Innovations in research and clinical care using patient-generated health data. CA Cancer J. Clin. 2020, 70, 182–199. [Google Scholar] [CrossRef]
  13. Coravos, A.; Khozin, S.; Mandl, K.D. Developing and adopting safe and effective digital biomarkers to improve patient outcomes. NPJ Digit. Med. 2019, 2, 14. [Google Scholar] [CrossRef] [Green Version]
  14. Phillips, K.M.; Faul, L.A.; Small, B.J.; Jacobsen, P.B.; Apte, S.M.; Jim, H.S.L. Comparing the Retrospective Reports of Fatigue Using the Fatigue Symptom Index with Daily Diary Ratings in Women Receiving Chemotherapy for Gynecologic Cancer. J. Pain Symptom Manag. 2013, 46, 282–288. [Google Scholar] [CrossRef]
  15. Sato, H.; Kawahara, J. Selective bias in retrospective self-reports of negative mood states. Anxiety Stress Coping 2011, 24, 359–367. [Google Scholar] [CrossRef]
  16. Robinson, M.A.; Boies, K. On the quest for quality self-report data: HEXACO and indicators of careless responding. Can. J. Behav. Sci. Rev. Can. Des Sci. Du Comport. 2021, 53, 377–380. [Google Scholar] [CrossRef]
  17. Hsin, H.; Fromer, M.; Peterson, B.; Walter, C.; Fleck, M.; Campbell, A.; Varghese, P.; Califf, R. Transforming Psychiatry into Data-Driven Medicine with Digital Measurement Tools. NPJ Digit. Med. 2018, 1, 37. [Google Scholar] [CrossRef]
  18. Smrke, U.; Mlakar, I.; Lin, S.; Musil, B.; Plohl, N. Language, Speech, and Facial Expression Features for Artificial Intelligence–Based Detection of Cancer Survivors’ Depression: Scoping Meta-Review. JMIR Ment. Health 2021, 8, e30439. [Google Scholar] [CrossRef]
  19. Liang, Y.; Zheng, X.; Zeng, D.D. A survey on big data-driven digital phenotyping of mental health. Inf. Fusion 2019, 52, 290–307. [Google Scholar] [CrossRef]
  20. Ellgring, H. Non-Verbal Communication in Depression; Cambridge University Press: Cambridge, UK, 2008. [Google Scholar]
  21. Waxer, P.H. Therapist training in nonverbal communication I: Nonverbal cues for depression. J. Clin. Psychol. 1974, 30, 215–218. [Google Scholar] [CrossRef]
  22. Pampouchidou, A.; Simos, P.G.; Marias, K.; Meriaudeau, F.; Yang, F.; Pediaditis, M.; Tsiknakis, M. Automatic Assessment of Depression Based on Visual Cues: A Systematic Review. IEEE Trans. Affect. Comput. 2019, 10, 445–470. [Google Scholar] [CrossRef]
  23. Silk, J.S.; Dahl, R.E.; Ryan, N.D.; Forbes, E.E.; Axelson, D.A.; Birmaher, B.; Siegle, G.J. Pupillary Reactivity to Emotional Information in Child and Adolescent Depression: Links to Clinical and Ecological Measures. AJP 2007, 164, 1873–1880. [Google Scholar] [CrossRef] [PubMed]
  24. Cummins, N.; Scherer, S.; Krajewski, J.; Schnieder, S.; Epps, J.; Quatieri, T.F. A review of depression and suicide risk assessment using speech analysis. Speech Commun. 2015, 71, 10–49. [Google Scholar] [CrossRef]
  25. Low, D.M.; Bentley, K.H.; Ghosh, S.S. Automated assessment of psychiatric disorders using speech: A systematic review. Laryngoscope Investig. Otolaryngol. 2020, 5, 96–116. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Selting, M. 38. Verbal, vocal, and visual practices in conversational interaction. In Handbücher zur Sprach- und Kommunikationswissenschaft/Handbooks of Linguistics and Communication Science (HSK) 38/1; Müller, C., Cienki, A., Fricke, E., Ladewig, S., McNeill, D., Tessendorf, S., Eds.; DE GRUYTER: Berlin, Germany, 2013; pp. 589–609. [Google Scholar] [CrossRef]
  27. Mlakar, I.; Lin, S.; Aleksandraviča, I.; Arcimoviča, K.; Eglītis, J.; Leja, M.; Salgado Barreira, Á.; Gómez, J.G.; Salgado, M.; Mata, J.G.; et al. Patients-centered SurvivorShIp care plan after Cancer treatments based on Big Data and Artificial Intelligence technologies (PERSIST): A multicenter study protocol to evaluate efficacy of digital tools supporting cancer survivors. BMC Med. Inform. Decis. Mak. 2021, 21, 243. [Google Scholar] [CrossRef]
  28. Guohou, S.; Lina, Z.; Dongsong, Z. What reveals about depression level? The role of multimodal features at the level of interview questions. Inf. Manag. 2020, 57, 103349. [Google Scholar] [CrossRef]
  29. Beijers, L.; Wardenaar, K.J.; van Loo, H.M.; Schoevers, R.A. Data-driven biological subtypes of depression: Systematic review of biological approaches to depression subtyping. Mol. Psychiatry 2019, 24, 888–900. [Google Scholar] [CrossRef]
  30. Garcia-Ceja, E.; Riegler, M.; Nordgreen, T.; Jakobsen, P.; Oedegaard, K.J.; Tørresen, J. Mental health monitoring with multimodal sensing and machine learning: A survey. Pervasive Mob. Comput. 2018, 51, 1–26. [Google Scholar] [CrossRef]
  31. Graham, S.; Depp, C.; Lee, E.E.; Nebeker, C.; Tu, X.; Kim, H.C.; Jeste, D.V. Artificial Intelligence for Mental Health and Mental Illnesses: An Overview. Curr. Psychiatry Rep. 2019, 21, 116. [Google Scholar] [CrossRef]
  32. Kim, J.; Uddin, Z.A.; Lee, Y.; Nasri, F.; Gill, H.; Subramanieapillai, M.; Lee, R.; Udovica, A.; Phan, L.; Lui, L.; et al. A Systematic review of the validity of screening depression through Facebook, Twitter, Instagram, and Snapchat. J. Affect. Disord. 2021, 286, 360–369. [Google Scholar] [CrossRef]
  33. El-Sherif, D.M.; Abouzid, M.; Elzarif, M.T.; Ahmed, A.A.; Albakri, A.; Alshehri, M.M. Telehealth and Artificial Intelligence Insights into Healthcare during the COVID-19 Pandemic. Healthcare 2022, 10, 385. [Google Scholar] [CrossRef]
  34. Celi, L.A.; Cellini, J.; Charpignon, M.L.; Dee, E.C.; Dernoncourt, F.; Eber, R.; Mitchell, W.G.; Moukheiber, L.; Schirmer, J.; Situ, J.; et al. Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review. PLOS Digit. Health 2022, 1, e0000022. [Google Scholar] [CrossRef]
  35. Arksey, H.; O’Malley, L. Scoping studies: Towards a methodological framework. Int. J. Soc. Res. Methodol. 2005, 8, 19–32. [Google Scholar] [CrossRef] [Green Version]
  36. Tricco, A.C.; Lillie, E.; Zarin, W.; O’Brien, K.K.; Colquhoun, H.; Levac, D.; Moher, D.; Peters, M.D.; Horsley, T.; Weeks, L.; et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): Checklist and Explanation. Ann. Int. Med. 2018, 169, 467–473. [Google Scholar] [CrossRef] [Green Version]
  37. Gratch, J.; Artstein, R.; Lucas, G.; Stratou, G.; Scherer, S.; Nazarian, A.; Wood, R.; Boberg, J.; DeVault, D.; Marsella, S.; et al. The Distress Analysis Interview Corpus of human and computer interviews. In Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC’14), Reykjavik, Iceland, 26–31 May 2014; pp. 3123–3128. [Google Scholar]
  38. Symptom Media Films. Available online: https://symptommedia.com/film-library/ (accessed on 20 December 2021).
  39. Pham, M.T.; Rajić, A.; Greig, J.D.; Sargeant, J.M.; Papadopoulos, A.; McEwen, S.A. A scoping review of scoping reviews: Advancing the approach and enhancing the consistency. Res. Syn. Meth. 2014, 5, 371–385. [Google Scholar] [CrossRef] [Green Version]
  40. Braun, V.; Clarke, V. Using thematic analysis in psychology. Qual. Res. Psychol. 2006, 3, 77–101. [Google Scholar] [CrossRef] [Green Version]
  41. Yang, L.; Jiang, D.; Sahli, H. Integrating Deep and Shallow Models for Multi-Modal Depression Analysis—Hybrid Architectures. IEEE Trans. Affect. Comput. 2021, 12, 239–253. [Google Scholar] [CrossRef]
  42. Samareh, A.; Jin, Y.; Wang, Z.; Chang, X.; Huang, S. Detect depression from communication: How computer vision, signal processing, and sentiment analysis join forces. IISE Trans. Healthc. Syst. Eng. 2018, 8, 196–208. [Google Scholar] [CrossRef]
  43. Prabhu, S.; Mittal, H.; Varagani, R.; Jha, S.; Singh, S. Harnessing emotions for depression detection. Pattern Anal. Applic. 2022, 25, 537–547. [Google Scholar] [CrossRef]
  44. Ceccarelli, F.; Mahmoud, M. Multimodal temporal machine learning for Bipolar Disorder and Depression Recognition. Pattern Anal. Applic. 2022, 25, 493–504. [Google Scholar] [CrossRef]
  45. Dham, S.; Sharma, A.; Dhall, A. Depression Scale Recognition from Audio, Visual and Text Analysis. arXiv 2017, arXiv:1709.05865. [Google Scholar]
  46. Qureshi, S.A.; Saha, S.; Hasanuzzaman, M.; Dias, G. Multitask Representation Learning for Multimodal Estimation of Depression Level. IEEE Intell. Syst. 2019, 34, 45–52. [Google Scholar] [CrossRef]
  47. Ray, A.; Kumar, S.; Reddy, R.; Mukherjee, P.; Garg, R. Multi-level Attention network using text, audio and video for Depression Prediction. In Proceedings of the 9th International on Audio/Visual Emotion Challenge and Workshop, Nice, France, 21–25 October 2019. [Google Scholar] [CrossRef]
  48. Williamson, J.R.; Godoy, E.; Cha, M.; Schwarzentruber, A.; Khorrami, P.; Gwon, Y.; Kung, H.T.; Dagli, C.; Quatieri, T.F. Detecting Depression using Vocal, Facial and Semantic Communication Cues. In Proceedings of the 6th International Workshop on Audio/Visual Emotion Challenge, Amsterdam, The Netherlands, 16 October 2016; pp. 11–18. [Google Scholar] [CrossRef]
  49. Yang, L.; Jiang, D.; He, L.; Pei, E.; Oveneke, M.C.; Sahli, H. Decision Tree Based Depression Classification from Audio Video and Language Information. In Proceedings of the 6th International Workshop on Audio/Visual Emotion Challenge, Amsterdam, The Netherlands, 16 October 2016; pp. 89–96. [Google Scholar] [CrossRef]
  50. Yang, L.; Jiang, D.; Xia, X.; Pei, E.; Oveneke, M.C.; Sahli, H. Multimodal Measurement of Depression Using Deep Learning Models. In Proceedings of the 7th Annual Workshop on Audio/Visual Emotion Challenge, Mountain View, CA, USA, 23–27 October 2017; pp. 53–59. [Google Scholar] [CrossRef]
  51. Yin, S.; Liang, C.; Ding, H.; Wang, S. A Multi-Modal Hierarchical Recurrent Neural Network for Depression Detection. In Proceedings of the 9th International on Audio/Visual Emotion Challenge and Workshop—AVEC ’19, Nice, France, 21–25 October 2019; pp. 65–71. [Google Scholar] [CrossRef] [Green Version]
  52. Rohanian, M.; Hough, J.; Purver, M. Detecting Depression with Word-Level Multimodal Fusion. Proc. Interspeech 2019, 1443–1447. [Google Scholar] [CrossRef] [Green Version]
  53. Gong, Y.; Poellabauer, C. Topic Modeling Based Multi-modal Depression Detection. In Proceedings of the 7th Annual Workshop on Audio/Visual Emotion Challenge, Mountain View, CA, USA, 23–27 October 2017; pp. 69–76. [Google Scholar] [CrossRef] [Green Version]
  54. Gupta, R.; Malandrakis, N.; Xiao, B.; Guha, T.; Van Segbroeck, M.; Black, M.; Potamianos, A.; Narayanan, S. Multimodal Prediction of Affective Dimensions and Depression in Human-Computer Interactions. In Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge—AVEC ’14, Orlando, FL, USA, 7 November 2014; pp. 33–40. [Google Scholar] [CrossRef]
  55. Pampouchidou, A.; Simantiraki, O.; Fazlollahi, A.; Pediaditis, M.; Manousos, D.; Roniotis, A.; Giannakakis, G.; Meriaudeau, F.; Simos, P.; Marias, K.; et al. Depression Assessment by Fusing High and Low Level Features from Audio, Video, and Text. In Proceedings of the 6th International Workshop on Audio/Visual Emotion Challenge, Amsterdam, The Netherlands, 16 October 2016; pp. 27–34. [Google Scholar] [CrossRef]
  56. Sun, B.; Zhang, Y.; He, J.; Yu, L.; Xu, Q.; Li, D.; Wang, Z. A Random Forest Regression Method with Selected-Text Feature for Depression Assessment. In Proceedings of the 7th Annual Workshop on Audio/Visual Emotion Challenge, Mountain View, CA, USA, 23–27 October 2017; pp. 61–68. [Google Scholar] [CrossRef]
  57. Morales, M.; Scherer, S.; Levitan, R. A Linguistically-Informed Fusion Approach for Multimodal Depression Detection. In Proceedings of the Fifth Workshop on Computational Linguistics and Clinical Psychology: From Keyboard to Clinic, New Orleans, LA, USA, 5 June 2018; pp. 13–24. [Google Scholar] [CrossRef]
  58. Haque, A.; Guo, M.; Miner, A.S.; Fei-Fei, L. Measuring Depression Symptom Severity from Spoken Language and 3D Facial Expressions. arXiv 2018, arXiv:1811.08592. [Google Scholar]
  59. Oureshi, S.A.; Dias, G.; Saha, S.; Hasanuzzaman, M. Gender-Aware Estimation of Depression Severity Level in a Multimodal Setting. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–23 July 2021; pp. 1–8. [Google Scholar] [CrossRef]
  60. Aleem, S.; Huda, N.U.; Amin, R.; Khalid, S.; Alshamrani, S.S.; Alshehri, A. Machine Learning Algorithms for Depression: Diagnosis, Insights, and Research Directions. Electronics 2022, 11, 1111. [Google Scholar] [CrossRef]
  61. Sugiyama, M. Dimensionality reduction of multimodal labeled data by local fisher discriminant analysis. J. Mach. Learn. Res. 2007, 8, 1027–1061. [Google Scholar]
  62. Amanat, A.; Rizwan, M.; Javed, A.R.; Abdelhaq, M.; Alsaqour, R.; Pandya, S.; Uddin, M. Deep Learning for Depression Detection from Textual Data. Electronics 2022, 11, 676. [Google Scholar] [CrossRef]
  63. Dunteman, G. Principal Components Analysis; SAGE Publications, Inc.: Newbury Park, CA, USA, 1989. [Google Scholar] [CrossRef]
  64. Iliou, T.; Konstantopoulou, G.; Ntekouli, M.; Lymperopoulou, C.; Assimakopoulos, K.; Galiatsatos, D.; Anastassopoulos, G. ILIOU machine learning preprocessing method for depression type prediction. Evol. Syst. 2019, 10, 29–39. [Google Scholar] [CrossRef]
  65. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic Minority Over-sampling Technique. Jair 2002, 16, 321–357. [Google Scholar] [CrossRef]
  66. Cho, S.-E.; Geem, Z.W.; Na, K.-S. Predicting Depression in Community Dwellers Using a Machine Learning Algorithm. Diagnostics 2021, 11, 1429. [Google Scholar] [CrossRef]
  67. Bengio, Y.; Frasconi, P. An input output HMM architecture. Adv. Neural Inf. Process. Syst. 1994, 7, 427–434. [Google Scholar]
  68. Mian, T.S. An Unsupervised Neural Network Feature Selection and 1D Convolution Neural Network Classification for Screening of Parkinsonism. Diagnostics 2022, 12, 1796. [Google Scholar] [CrossRef]
  69. Zulfiker, M.S.; Kabir, N.; Biswas, A.A.; Nazneen, T.; Uddin, M.S. An in-depth analysis of machine learning approaches to predict depression. Curr. Res. Behav. Sci. 2021, 2, 100044. [Google Scholar] [CrossRef]
  70. Xue, B.; Zhang, M.; Browne, W.N. Particle swarm optimisation for feature selection in classification: Novel initialisation and updating mechanisms. Appl. Soft Comput. 2014, 18, 261–276. [Google Scholar] [CrossRef]
  71. Muthusamy, H.; Polat, K.; Yaacob, S. Particle Swarm Optimization Based Feature Enhancement and Feature Selection for Improved Emotion Recognition in Speech and Glottal Signals. PLoS ONE 2015, 10, e0120344. [Google Scholar] [CrossRef]
  72. Robnik-Šikonja, M.; Kononenko, I. Theoretical and empirical analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef] [Green Version]
  73. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  74. Kursa, M.B.; Rudnicki, W.R. Feature Selection with the Boruta Package. J. Stat. Soft. 2010, 36, 1–13. [Google Scholar] [CrossRef] [Green Version]
  75. Bartlett, M.S.; Littlewort, G.; Frank, M.; Lainscsek, C.; Fasel, I.; Movellan, J. Recognizing Facial Expression: Machine Learning and Application to Spontaneous Behavior. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; pp. 568–573. [Google Scholar] [CrossRef]
  76. Scherer, S.; Stratou, G.; Morency, L.-P. Audiovisual behavior descriptors for depression assessment. In Proceedings of the 15th ACM on International conference on multimodal interaction—ICMI ’13, Sydney, Australia, 13–17 November 2013; pp. 135–140. [Google Scholar] [CrossRef]
  77. Cummins, N.; Joshi, J.; Dhall, A.; Sethu, V.; Goecke, R.; Epps, J. Diagnosis of depression by behavioural signals: A multimodal approach. In Proceedings of the 3rd ACM international workshop on Audio/visual emotion challenge, Barcelona, Spain, 21 October 2013; pp. 11–20. [Google Scholar] [CrossRef]
  78. Kaya, H.; Çilli, F.; Salah, A.A. Ensemble CCA for Continuous Emotion Prediction. In Proceedings of the 4th International Workshop on Audio/Visual Emotion Challenge—AVEC ’14, Orlando, Florida, USA, 7 November 2014; pp. 19–26. [Google Scholar] [CrossRef]
  79. DeVault, D.; Artstein, R.; Benn, G.; Dey, T.; Fast, E.; Gainer, A.; Georgila, K.; Gratch, J.; Hartholt, A.; Lhommet, M.; et al. SimSensei kiosk: A virtual human interviewer for healthcare decision support. In Proceedings of the 13th International Conference on Autonomous Agents and Multiagent Systems (AAMAS’14), Paris, France, 5–9 May 2014. [Google Scholar]
  80. Friborg, O.; Martinsen, E.W.; Martinussen, M.; Kaiser, S.; Øvergård, K.T.; Rosenvinge, J.H. Comorbidity of personality disorders in mood disorders: A meta-analytic review of 122 studies from 1988 to 2010. J. Affect. Disord. 2014, 152–154, 1–11. [Google Scholar] [CrossRef]
  81. Plana-Ripoll, O.; Pedersen, C.B.; Holtz, Y.; Benros, M.E.; Dalsgaard, S.; De Jonge, P.; Fan, C.C.; Degenhardt, L.; Ganna, A.; Greve, A.N.; et al. Exploring Comorbidity Within Mental Disorders Among a Danish National Population. JAMA Psychiatry 2019, 76, 259. [Google Scholar] [CrossRef] [Green Version]
  82. ‘ffmpeg’. Available online: http://ffmpeg.org/about.html (accessed on 13 November 2021).
  83. Degottex, G.; Kane, J.; Drugman, T.; Raitio, T.; Scherer, S. COVAREP—A collaborative voice analysis repository for speech technologies. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 960–964. [Google Scholar] [CrossRef]
  84. ‘mathworks’. Available online: https://www.mathworks.com/ (accessed on 5 November 2021).
  85. Baltrusaitis, T.; Robinson, P.; Morency, L.-P. OpenFace: An open source facial behavior analysis toolkit. In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–10. [Google Scholar] [CrossRef] [Green Version]
  86. ‘word2vec’. Available online: https://code.google.com/archive/p/word2vec/ (accessed on 15 March 2022).
  87. Depression-Detection-Through-Multi-Modal-Data. Available online: https://github.com/notmanan/Depression-Detection-Through-Multi-Modal-Data (accessed on 17 May 2022).
  88. Srivastava, R.K.; Greff, K.; Schmidhuber, J. Highway Networks’. arXiv 2015, arXiv:1505.00387. [Google Scholar]
  89. Vujovic, Ž.Ð. Classification Model Evaluation Metrics. IJACSA 2021, 12, 599–606. [Google Scholar] [CrossRef]
  90. Thieme, A.; Belgrave, D.; Doherty, G. Machine Learning in Mental Health: A Systematic Review of the HCI Literature to Support the Development of Effective and Implementable ML Systems. ACM Trans. Comput.-Hum. Interact. 2020, 27, 1–53. [Google Scholar] [CrossRef]
  91. Shatte, A.B.R.; Hutchinson, D.M.; Teague, S.J. Machine learning in mental health: A scoping review of methods and applications. Psychol. Med. 2019, 49, 1426–1448. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  92. Pigoni, A.; Delvecchio, G.; Madonna, D.; Bressi, C.; Soares, J.; Brambilla, P. Can Machine Learning help us in dealing with treatment resistant depression? A review. J. Affect. Disord. 2019, 259, 21–26. [Google Scholar] [CrossRef] [PubMed]
  93. Tokgöz, P.; Hrynyschyn, R.; Hafner, J.; Schönfeld, S.; Dockweiler, C. Digital Health Interventions in Prevention, Relapse, and Therapy of Mild and Moderate Depression: Scoping Review. JMIR Ment. Health 2021, 8, e26268. [Google Scholar] [CrossRef]
  94. Liu, G.-D.; Li, Y.-C.; Zhang, W.; Zhang, L. A Brief Review of Artificial Intelligence Applications and Algorithms for Psychiatric Disorders. Engineering 2020, 6, 462–467. [Google Scholar] [CrossRef]
  95. Han, J.; Kamber, M. Data Mining: Concepts and Techniques, 3rd ed.; Elsevier: Burlington, MA, USA, 2012. [Google Scholar]
  96. Cormier, P. The State of Enterprise Open Source: A Red Hat Report. Feb. 2022. Available online: https://www.redhat.com/en/resources/state-of-enterprise-open-source-report-2022 (accessed on 17 May 2022).
  97. Berk, R.; Heidari, H.; Jabbari, S.; Kearns, M.; Roth, A. Fairness in Criminal Justice Risk Assessments: The State of the Art. Sociol. Methods Res. 2021, 50, 3–44. [Google Scholar] [CrossRef]
  98. Burt, T.; Button, K.; Thom, H.; Noveck, R.; Munafò, M. The Burden of the “False-Negatives” in Clinical Development: Analyses of Current and Alternative Scenarios and Corrective Measures. Clin. Transl. Sci. 2017, 10, 470–479. [Google Scholar] [CrossRef] [Green Version]
  99. Levis, B.; Benedetti, A.; Thombs, B.D. Accuracy of Patient Health Questionnaire-9 (PHQ-9) for screening to detect major depression: Individual participant data meta-analysis. BMJ 2019, 365, l1476. [Google Scholar] [CrossRef] [Green Version]
  100. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning; Springer New York: New York, NY, USA, 2009. [Google Scholar] [CrossRef]
  101. Alghowinem, S.; Goecke, R.; Wagner, M.; Epps, J.; Hyett, M.; Parker, G.; Breakspear, M. Multimodal Depression Detection: Fusion Analysis of Paralinguistic, Head Pose and Eye Gaze Behaviors. IEEE Trans. Affect. Comput. 2018, 9, 478–490. [Google Scholar] [CrossRef]
  102. Kamiran, F.; Calders, T. Data preprocessing techniques for classification without discrimination. Knowl. Inf. Syst. 2012, 33, 1–33. [Google Scholar] [CrossRef] [Green Version]
  103. Zhang, B.H.; Lemoine, B.; Mitchell, M. Mitigating Unwanted Biases with Adversarial Learning. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, New Orleans, LA, USA, 2–3 February 2018; pp. 335–340. [Google Scholar] [CrossRef] [Green Version]
  104. Ravindran, S.; Pai, D.; Narang, A.; Srivastava, V. Multiple Attribute Fairness: Application to Fraud Detection. arXiv 2022, arXiv:2207.14355. [Google Scholar]
  105. Ahn, Y.; Lin, Y.-R. FairSight: Visual Analytics for Fairness in Decision Making. IEEE Trans. Visual. Comput. Graph. 2019, 26, 1086–1095. [Google Scholar] [CrossRef] [Green Version]
  106. Lundberg, S.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. arxiv 2017, arXiv:1705.07874. [Google Scholar]
  107. Sharma, S.; Henderson, J.; Ghosh, J. CERTIFAI: Counterfactual Explanations for Robustness, Transparency, Interpretability, and Fairness of Artificial Intelligence models. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, New York, NY, USA, 7–8 February 2020; pp. 166–172. [Google Scholar] [CrossRef]
  108. McIntosh, A.M.; Sullivan, P.F.; Lewis, C.M. Uncovering the Genetic Architecture of Major Depression. Neuron 2019, 102, 91–103. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the search and study selection process.
Figure 1. Flowchart of the search and study selection process.
Diagnostics 12 02683 g001
Figure 2. The number of papers per observable cue after the initial screening process.
Figure 2. The number of papers per observable cue after the initial screening process.
Diagnostics 12 02683 g002
Figure 3. Feature extraction pipeline.
Figure 3. Feature extraction pipeline.
Diagnostics 12 02683 g003
Figure 4. SVM and RF model architecture.
Figure 4. SVM and RF model architecture.
Diagnostics 12 02683 g004
Figure 5. The LSTM model architecture.
Figure 5. The LSTM model architecture.
Diagnostics 12 02683 g005
Table 1. The final set of 20 papers that passed the eligibility criteria and were reviewed in detail.
Table 1. The final set of 20 papers that passed the eligibility criteria and were reviewed in detail.
NoReferenceAI/ML TechniqueReported Performance ValueDatasetFeature
Fusion
Github CodeType
1Yang et al., 2021 [41]CNN, SVM, RFF1: 0.74DAIC-WOZEarly/LateNoJournal
2Samareh et al., 2018 [42]RFRMSE: 4.35DAIC-WOZLateNoJournal
3Prabhu t al., 2021 [43]CNN, LSTMF1: 0.75DAIC-WOZLateNoJournal
4Ceccarelli and Mahmoud, 2021 [44]LSTMF1: 0.87BDC-WBLateYesJournal
5Shan et al., 2020 [28]SVR and RFRMSE: 4.96DAIC-WOZEarlyNoJournal
6Dham et al., 2017 [45]SVM and NNRMSE: 5.35DAIC-WOZLateNoJournal
7Qureshi et al., 2019 [46]CNNF1: 0.48DAIC-WOZEarlyNoJournal
8Ray et al., 2019 [47]LSTMRMSE: 4.28DAIC-WOZLateNoProceeding
9Williamson et al., 2016 [48]MV-RF1: 0.81DAIC-WOZLateNoProceeding
10Yang et al., 2016 [49]DTF1: 0.72DAIC-WOZEarlyNoProceeding
11Yang et al., 2017 [50]CNNRMSE: 4.65DAIC-WOZLateNoProceeding
12Yin et al., 2019 [51]LSTMRMSE: 5.50DAIC-WOZEarlyNoProceeding
13Rohanian et al., 2019 [52]LSTMF1: 0.81DAIC-WOZEarly/LateYesProceeding
14Gong, 2018 [53] SVM and RFF1: 0.6DAIC-WOZEarlyNoProceeding
15Gupta et al., 2014 [54]LRCorr. Coef: 0.52AViD-CEarly/LateNoProceeding
16Pampouchidou et al., 2016 [55]DTF1: 0.66DAIC-WOZEarly/LateNoProceeding
17Sun et al., 2017 [56]RFRMSE: 4.7DAIC-WOZLateNoProceeding
18Morales et al., 2018 [57]SVMF1: 0.49DAIC-WOZEarly/LateNoProceeding
19Haque et al., 2018 [58]CNNSens: 83.3%; Spec: 82.6%DAIC-WOZEarlyYesProceeding
20Qureshi et al., 2021 [59]LSTMRMSE: 3.49DAIC-WOZLateNoProceeding
Table 2. The selected SymptomMedia films.
Table 2. The selected SymptomMedia films.
No.Recording TitleDepression Status
1Depression Assessment A-1D
2Depression Assessment A-2D
3Depression Assessment A-3D
4Major Depressive Disorder A-1D
5Major Depressive Disorder Recurrent Episode Severe Part 1D
6Major Depressive Disorder Recurrent Episode Severe Part 2D
7Major Depressive Disorder Recurrent Episode Severe Part 3D
8Major Depressive Disorder Recurrent Episode Severe Part 4D
9Major Depressive Disorder with Anxious DistressD
10Major Depressive Disorder with Melancholic FeaturesD
11Major Depressive Disorder with Peripartum OnsetD
12Patients with Depression A-1D
13Major Depressive Disorder with Seasonal PatternD
14Major Depressive DisorderD
15Narcissistic Personality DisorderND
16Antisocial Personality Disorder Version 1ND
17Histrionic Personality Disorder Version 2ND
18Stimulant Use DisorderND
19Schizoid Personality DisorderND
20Schizotypal Personality DisorderND
21Alcohol Use DisorderND
22Anorexia Nervosa Binge-Eating PurgingND
23Antisocial Personality Disorder Version 2ND
24Avoidant Personality DisorderND
25Binge Eating DisorderND
26Bulimia NervosaND
27Dependent Personality DisorderND
28Histrionic Personality Disorder Version 1ND
Notes: D = diagnosis of depression; ND = no diagnosis of depression.
Table 3. The details of the extracted features.
Table 3. The details of the extracted features.
ModalityCategoryFeaturesDescriptions
AudioProsodicF0Fundamental frequency created by the vibrations in the vocal cords
VUVBinary decision of voicing—unvoicing part
EnergyAmplitude variation of speech signals over time (volume or intensity)
DurationAmount of time to build vowels, words, and similar constructs
SpectralMFCCMel Frequency Cepstral Coefficients: Short-term power spectrum of the speech signal.
HMPDM0-24Harmonic Model and Phase Distortion mean
HMPDD0-12Harmonic Model and Phase Distortion deviations
LPCCLinear Prediction Cepstral Coefficients
LFPCLog-Frequency Power Coefficients
GFCCGammatone Frequency Cepstral Coefficients
FormantsFrequencies of the acoustic resonance of the vocal tract
Voice qualityNAQNormalized amplitude quotient
QOQQuasi-open quotient
H1H2Differentiated glottal source spectrum
PSPParabolic spectral parameter
MDQMaxima dispersion quotient
peakSlopeSpectral tilt/slope of wavelet responses
RdShape parameter of the Liljencrants–Fant model of the glottal pulse dynamics
jittervariability of fundamental frequency between successive vibratory cycles
shimmervariable of the amplitude between successive vibratory cycles
Harmonics to Noise Ratiorelative level of noise in the frequency spectrum of vowels
Teager Energy Operatorused to detect stress in speech
Video/ImageAction UnitsAU01- 02/04- 06/09- 10/12/14- 15/17/20/25- 26_rRegression values of Inner brow raiser, Outer brow raiser, Brow lower, Upper lid raiser, Cheek raiser, Nose wrinkle, Upper lip raiser, Lip corner puller, Dimpler, Lip corner depressor, Chin raiser, Lip stretched, Lip part, Jaw drop
AU04/12/15/23/28/45_cBinary outputs of Brow lower, Lip corner puller, Lip corner depressor, Lip tightener, Lip suck, Blink
Eye gazex_0, y_0, z_0, x_1, y_1, z_1Vectors of both eyes in world coordinate space
x_h0, y_h0, z_h0, x_h1, y_h1, z_h1Vectors of both eyes in head coordinate space
Head posepose_confidence/successpose_confidence/success
Tx, Ty, TzVectors of head in world position coordinate space
Rx, Ry, RzVectors of head in head rotation coordinate space
Facial Landmarks68 2D points on face.The points are in pixel coordinates.
68 3D points on face.The points are in millimetres in the world coordinate space
TextLexicalWord2Vec300-feature vector representation (sentences × words × 300 features)
Table 4. The sizes of the datasets.
Table 4. The sizes of the datasets.
DAIC-WOZSymptomMedia
Testing Set4728
Development Set35-
Training Set107-
Table 5. The performance and error metrics for the SVM + RF with SVM late fusion (T + A: text and audio; T + V: text and video; A + V: audio and video; T + A + V: text, audio, and video).
Table 5. The performance and error metrics for the SVM + RF with SVM late fusion (T + A: text and audio; T + V: text and video; A + V: audio and video; T + A + V: text, audio, and video).
DAIC-WOZ DB DatasetSymptomMedia Dataset
ModalityF1-ScoreRecallPrecisionMSERMSEMAEF1-ScoreRecallPrecisionMSERMSEMAE
T + A0.380.390.550.610.780.610.460.460.460.530.730.53
T + V0.420.410.550.590.760.590.420.430.420.570.750.57
A + V0.560.550.60.450.670.450.370.390.380.60.770.6
T + A + V0.460.450.610.540.730.540.40.430.410.570.750.57
Table 6. The performance and error metrics for the SVM + RF with RF late fusion (T + A: text and audio; T + V: text and video; A + V: audio and video; T + A + V: text, audio, and video).
Table 6. The performance and error metrics for the SVM + RF with RF late fusion (T + A: text and audio; T + V: text and video; A + V: audio and video; T + A + V: text, audio, and video).
DAIC-WOZ DB DatasetSymptomMedia Dataset
ModalityF1-ScoreRecallPrecisionMSERMSEMAEF1-ScoreRecallPrecisionMSERMSEMAE
T + A0.380.390.550.610.780.610.460.460.460.530.730.53
T + V0.420.410.550.590.760.590.420.430.420.570.750.57
A + V0.460.450.610.540.730.540.30.360.290.640.80.64
T + A + V0.440.430.590.560.750.560.40.430.410.570.750.57
Table 7. Performance and error metrics for the LSTM without gating (T + A: text and audio; T + V: text and video; T + A + V: text, audio, and video).
Table 7. Performance and error metrics for the LSTM without gating (T + A: text and audio; T + V: text and video; T + A + V: text, audio, and video).
DAIC-WOZ DB DatasetSymptomMedia Dataset
ModalityF1-ScoreRecallPrecisionMSERMSEMAEF1-ScoreRecallPrecisionMSERMSEMAE
T + A0.530.570.510.270.520.480.560.570.580.370.610.47
T + V0.520.550.50.290.540.490.450.460.460.360.60.48
T + A + V0.530.550.520.350.590.510.560.570.580.370.60.47
Table 8. The performance and error metrics for the LSTM with gating (T + A: text and audio; T + V: text and video; T + A + V: text, audio, and video).
Table 8. The performance and error metrics for the LSTM with gating (T + A: text and audio; T + V: text and video; T + A + V: text, audio, and video).
DAIC-WOZ DB DatasetSymptomMedia Dataset
ModalityF1-ScoreRecallPrecisionMSERMSEMAEF1-ScoreRecallPrecisionMSERMSEMAE
T + A0.450.430.570.330.580.480.480.50.50.350.590.44
T + V0.420.410.550.350.590.490.50.50.50.380.610.47
T + A + V0.640.660.640.320.570.450.480.50.50.340.580.44
Table 9. Comparison table for the reported performance values for all modalities (T: text; A: audio; V: video).
Table 9. Comparison table for the reported performance values for all modalities (T: text; A: audio; V: video).
StudyReported Performance Values for Uni-/Bi-/Multimodalities
Shan et al., 2020 [28]TT + AT + VT + A + V
RMSE: 5.58, MAE: 4.41RMSE: 5.3(SVR), MAE: 3.94RMSE: 5.44, MAE: 4.24RMSE: 4.96, MAE: 3.84
Dham et al., 2017 [45] T + AVT + A + V
RMSE: 5.98, MAE: 4.37RMSE: 5.37, MAE: 4.65RMSE: 5.35, MAE: 4.37
Qureshi et al., 2019 [46]TAVT + A + V
RMSE: 4.7, MAE: 3.81RMSE: 6.55, MAE: 5.67RMSE: 6.28, MAE: 5.03RMSE: 4.24, MAE: 3.29
Ray et al., 2019 [47]A + VT + AT + VT + A + V
RMSE: 5.38RMSE: 4.37RMSE: 4.64RMSE: 4.28
Williamson et al., 2016 [48]TAVT + A + V
RMSE: 4.46, MAE: 3.34RMSE: 6.38, MAE: 5.32RMSE: 6.45, MAE: 5.33RMSE: 5.31, MAE: 4.18
Samareh et al., 2018 [42]TAVT + A + V
RMSE: 5.86, MAE: 4.88RMSE: 5.89, MAE: 5.18RMSE: 5.65, MAE: 4.87RMSE: 4.78, MAE: 4.05
Pampouchidou et al., 2016 [55]TAVT + A + V
F1: 0.59 (0.87)F1: 0.5 (0.9)F1: 0.62 (0.91)
Haque et al., 2018 [58]TAVT + A + V
TPR: 57.5, TNR: 60.3TPR: 53.9, TNR: 54.6TPR: 61, TNR: 59.7TPR: 74.2, TNR: 76.1
Qureshi et al., 2021 [59]TAVT + A + V
MSE: 24.02, MAE: 4.09MSE: 41.96, MAE: 5.19MSE: 44.29, MAE: 5.23MSE:22.25, MAE: 3.49
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Arioz, U.; Smrke, U.; Plohl, N.; Mlakar, I. Scoping Review on the Multimodal Classification of Depression and Experimental Study on Existing Multimodal Models. Diagnostics 2022, 12, 2683. https://doi.org/10.3390/diagnostics12112683

AMA Style

Arioz U, Smrke U, Plohl N, Mlakar I. Scoping Review on the Multimodal Classification of Depression and Experimental Study on Existing Multimodal Models. Diagnostics. 2022; 12(11):2683. https://doi.org/10.3390/diagnostics12112683

Chicago/Turabian Style

Arioz, Umut, Urška Smrke, Nejc Plohl, and Izidor Mlakar. 2022. "Scoping Review on the Multimodal Classification of Depression and Experimental Study on Existing Multimodal Models" Diagnostics 12, no. 11: 2683. https://doi.org/10.3390/diagnostics12112683

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop