Skip to Content
You are currently on the new version of our website. Access the old version .
DiagnosticsDiagnostics
  • Article
  • Open Access

26 February 2024

Multitask Learning for Mental Health: Depression, Anxiety, Stress (DAS) Using Wearables

and
Computer Engineering Department, Boğaziçi University, 34342 İstanbul, Türkiye
*
Authors to whom correspondence should be addressed.

Abstract

This study investigates the prediction of mental well-being factors—depression, stress, and anxiety—using the NetHealth dataset from college students. The research addresses four key questions, exploring the impact of digital biomarkers on these factors, their alignment with conventional psychology literature, the time-based performance of applied methods, and potential enhancements through multitask learning. The findings reveal modality rankings aligned with psychology literature, validated against paper-based studies. Improved predictions are noted with temporal considerations, and further enhanced by multitasking. Mental health multitask prediction results show aligned baseline and multitask performances, with notable enhancements using temporal aspects, particularly with the random forest (RF) classifier. Multitask learning improves outcomes for depression and stress but not anxiety using RF and XGBoost.

1. Introduction

Smartwatches and fitness trackers are examples of wearable devices that have become popular with their ability to continuously monitor various physiological signals and record real-time temporal trends. The abundance of temporal data has intriguing prospects for tailored, context-aware applications, providing many insights into human activities, health monitoring, and behavior analysis.
One particular application area focuses on monitoring and predicting individuals’ mental health, encompassing their emotional and psychological states. Mental health is crucial in determining an individual’s sense of contentment, happiness, and overall fulfillment in life.
Mental health is a multifaceted aspect of human well-being, encompassing dimensions such as depression, anxiety, and stress. The World Health Organization (WHO) indicated that, in 2019, around 301 million people suffered from anxiety, and 280 million people lived with depression https://www.who.int/observatories/global-observatory-on-health-research-and-development/analyses-and-syntheses/mental-health/global-strategic-direction (accessed on 17 January 2024). These two factors are known to be related to mental disorders and may also be referred to as mental health conditions https://www.who.int/news-room/fact-sheets/detail/mental-disorders (accessed on 17 January 2024). Stress is another factor affecting mental and physiological problems https://www.apa.org/topics/stress/health (accessed on 17 January 2024).
To measure these three important factors for mental health, psychologists developed assessment techniques based on questionnaires. One of the most well-known scales was developed in 1995 by Lovibond and Lovibond, and it is called the Depression, Anxiety, and Stress Scale (DASS) [1]. The main contribution of this new scale is to be a combined version of separate questionnaires for each factor. Traditional approaches to mental health assessment often rely on subjective self-reports and clinical interviews, which may not capture the full spectrum of these conditions. The accurate assessment and differentiation of these constructs have been long-standing challenges in the field. In recent years, there has been a growing interest in leveraging wearable devices to monitor and predict mental health, responding to the increasing awareness of the importance of mental well-being in modern society [2].
The sensors integrated into wearable devices provide non-invasive means of collecting continuous data on individuals’ physiological and behavioral patterns. This invaluable resource enables us to explore the intricate connections between mental health and wearable-derived digital biomarkers. The most adapted devices are smartphones and smartwatches. These devices have various sensors, such as accelerometer, heart rate, electro-dermal activity (EDA), etc. Their computational capacity enables them to collect continuous data and even run machine learning models on them to extract meaningful, high-level, human-interpretable information.
Traditional machine learning algorithms, such as logistic regression, decision trees, and ensemble methods, have been extensively used to classify well-being and mental health using data collected from wearables and questionnaires [3,4]. However, many studies often overlook the temporal aspect of the data, not considering how past data impacts future mental health levels.
Integrating time-based prediction techniques with wearable devices provides a unique opportunity to capture and analyze temporal patterns, facilitating early detection and timely interventions to improve mental well-being. Deep learning architectures, including recurrent neural networks (RNNs), particularly long short-term memory (LSTMs), have emerged as powerful tools for addressing time-based prediction problems.
In this paper, we utilize the NetHealth dataset [5] and specifically focus on the prediction of students’ mental health, collected from undergraduate students at Notre Dame University spanning from Fall 2015 to Spring 2019, constituting eight waves corresponding to each semester. The dataset encompassed data from approximately 700 students during the 2015–2017 period and 300 students during the 2017–2019 period, reflecting variations in participation rates over time. This rich dataset comprised information from various modalities, including social networks, physical activity, sleep data collected from Fitbit wearable devices, and ground truth data sourced from questionnaires.
When there is a set of related but not identical tasks { T i } i = 1 m for i ranging from 1 to m, multitask learning endeavors to enhance the learning of a model for task T i by harnessing the knowledge embedded in the other tasks, either individually or as a subset of the entire set [6]. In this paper, we delve into the domain of multitask learning (MTL). This powerful machine learning paradigm enables the simultaneous modeling of multiple mental health dimensions, including depression, anxiety, and stress, in our context. By jointly training models for these tasks, we aim to enhance predictive accuracy and achieve a more comprehensive understanding of mental well-being. In our study, where different tasks in multitask supervised learning share the training data, MTL can be viewed as a generalization of multi-output regression.
Our investigation encompasses a wide spectrum of predictive techniques. We delve into deep learning methodology, specifically LSTM networks, to capture temporal dependencies in mental health data. Additionally, we incorporate conventional prediction algorithms such as random forest (RF) and XGBoost to provide a holistic view of mental health assessment. In addition, we consider their performances in the time-based (lagged) version of data. By leveraging wearable technology, MTL, and advanced predictive techniques, we aim to enhance our understanding of the complex interplay between depression, anxiety, and stress.
Our study seeks to address the following research questions:
  • RQ1: What are the underlying biomarkers of mental well-being, considering factors such as stress, anxiety, and depression?
  • RQ2: Are these biomarkers in line with conventional psychological studies?
  • RQ3: Do time-based versions of conventional algorithms lead to improved prediction performance in the context of mental health?
  • RQ4: Do further methodologies such as MTL on top of time-based models improve the final performance?
This paper is organized as follows: Section 2 provides insights into the motivation of the DASS and sheds light on the existing literature related to mental health monitoring with wearables and, most importantly, what has been done in the literature about mental health via multitask learning and considering the time aspect. In Section 3, we delve into the details of the motivation of methodology and details of the methods, such as the dataset, the missing data handling process, and the preprocessing steps undertaken before the analysis, along with the details regarding the algorithms. Section 4 presents the results of our analyses, including related biomarkers to the target variables and a wide range of performance of applied algorithms. In Section 5, we present the model performance results based on various types, from conventional to time-based and multitask. Finally, we conclude the paper with Section 6, offering interpretations, discussions, comparisons to related studies, and insights into the future of mental health prediction.

3. Methodology

This section provides a detailed overview of our methodology, driven by the motivation to apply time-based and multitask models. We provide dataset details and highlight used modalities with target variables to provide baseline information about the aim of this study. We explain the imputation methodology since we include more than one semester’s data by imputing the missing values in the other semester’s data. We also explain data preprocessing steps before the model training and provide details about the used models.

3.1. Motivation for Methodologies

3.1.1. Time-Based Modeling

Time-based modeling in mental health utilizes temporal data and techniques for profound insights, prediction, and management of conditions over time [15]. It is particularly valuable in the context of longitudinal studies, where data are collected over extended periods, enabling the exploration of how mental health issues evolve over time and the identification of influential factors. Time-series analysis, focusing on fluctuations in symptoms, mood, and behaviors, aids in pattern recognition and identifying triggers. Predictive models leverage historical data to forecast future mental health outcomes, potentially paving the way for proactive interventions. These models also facilitate the determination of optimal intervention timings, a crucial aspect of effective mental healthcare. Wearable technology, capable of monitoring physiological and behavioral markers, can be integrated into these models for real-time monitoring and feedback [16].

3.1.2. Multitask Learning

Multitask learning (MTL) aims to train models for multiple similar tasks simultaneously [17]. Generally, it has been used to reduce overfit and improve model accuracy [18]. In the context of mental health, instead of training separate models for each mental health aspect (depression, anxiety, stress), one can use MTL to build a single model that predicts all three simultaneously. This approach recognizes that different aspects of mental health, such as depression, anxiety, and stress, are interconnected and can benefit from joint modeling, considering their interdependencies. Additionally, it can help identify shared risk factors or protective factors across different mental health conditions, contributing to a deeper understanding of the complexities of mental health. MTL can potentially enhance mental health assessment, intervention, and support by leveraging the synergy between multiple mental health-related tasks and domains.

3.2. Dataset

In this study, we utilize the NetHealth dataset http://sites.nd.edu/nethealth/ (accessed on 27 January 2024) [5], which was gathered from undergraduate students at Notre Dame (ND) University between the Fall of 2015 and the Spring of 2019. This dataset covers eight distinct periods (waves), each corresponding to a semester. Initially, it included data from approximately 700 students from 2015 to 2017, but this number dropped to 300 during the 2017–2019 period due to decreased participation.
The dataset encompasses various types of information, including social network data, physical activity data collected from Fitbit wearable devices (such as step counts, active minutes, heart rate, and sleep metrics), the administrative records of courses and grades from the ND Registrar’s Office, a calendar displaying weekly class schedules and breaks, and self-reported survey data covering aspects like physical and mental health, social-psychological states, personal preferences, self-reported behaviors, demographic information, and background traits (Figure 1).
Figure 1. NetHealth dataset and its components.
It is important to note that the data collection process complied with Institutional Review Board (IRB) protocols, with the informed consent of all participating individuals. However, some data are not publicly shared due to privacy concerns.
In this study, we specifically focused on distinct sub-datasets within the collected data, including wearable data containing fitness and sleep measurements and self-reported survey data. Additionally, we combined all semester’s information and imputed the missing values for expanded data, which will be explained in the data preprocessing steps (Section 3.3). Since a large number of survey questions are collected, we are focusing on only some selected ones in this study’s context. For clarity in navigating the parameter space, we provide detailed feature information in Table 1. Our feature selection aligns with our previous study [19], except for the grade dataset, as it falls outside the scope of the current study. Similarly, mental health features are excluded, as they serve as target values. It is essential to note that, unlike Table 1 in the previous study [19], the features in the current analysis are not categorized by semester number since we utilize information spanning all semesters.
Table 1. Details of the parameters in the dataset.

3.2.1. Sensing Streams

Within the scope of our study, we have directed our attention to specific sub-datasets within the broader dataset. These sub-datasets encompass wearable data, capturing essential fitness and sleep measurements through Fitbit devices. Furthermore, we delved into the self-reported survey data, which offer a multifaceted view of students’ well-being. We did not utilize other data information, such as social network and communication data, since they require different methodologies, such as graph neural networks, to obtain meaningful results. We will consider them in our future studies.

3.2.2. Predictors

Our research focuses on the assessment and prediction of three mental health indicators—stress, anxiety, and depression—similarly to the DASS [20]. To unravel the interplays between these mental health conditions and an array of predictor variables, we consider a diverse set of factors, encompassing physical activity levels, sleep patterns, demographic characteristics, and self-reported behaviors.

3.2.3. Target Values

Even though we focus on DAS parameters, our dataset does not explicitly contain DASS values collected in a single questionnaire, as performed for DASS-21 or DASS-42; instead, we have separately collected values for depression, anxiety, and stress. For simplicity, we will use the term ‘DAS’ to represent all three target variables. In the following, we describe the target values and their characteristics specific to the current dataset.
  • Depression
    Depression is a pervasive mental health condition marked by persistent and deep-seated feelings of sadness, hopelessness, and diminished interest or pleasure in once-enjoyable activities, leading to substantial impairment in daily functioning and overall quality of life [1,21]. NetHealth contains two different survey responses for depression such as ‘CES-D’ https://www.apa.org/pi/about/publications/caregivers/practice-settings/assessment/tools/depression-scale (accessed on 17 January 2024) and ‘BDI’ https://www.apa.org/pi/about/publications/caregivers/practice-settings/assessment/tools/beck-depression (accessed on 17 January 2024). Additionally, we have continuous scores ranging from 0 to 60 for CES-D and from 0 to 63 for BDI, with higher scores indicating greater depressive symptoms. For CES-D, there is also a grouped version ranging from 0 to 1, representing not depressed and depressed. For BDI, the grouped version ranges from 0 to 3, representing minimal, mild, moderate, and severe levels of depression. Throughout the study, we employed the BDI with its corresponding classification levels.
  • Anxiety
    Anxiety is a prolonged state of excessive worry and fear about future events, characterized by heightened alertness and unease; while it can be a normal response to stress, overwhelming and impairing forms may indicate an anxiety disorder [1,22]. Similar to depression values, NetHealth also contains two different survey responses for anxiety: ‘STAI’ https://www.apa.org/pi/about/publications/caregivers/practice-settings/assessment/tools/trait-state (accessed on 17 January 2024) and ‘BAI’ https://en.wikipedia.org/wiki/Beck_Depression_Inventory (accessed on 17 January 2024). We have continuous scores ranging from 20 to 80 for STAI and from 0 to 63 for BAI, where higher scores indicate greater anxiety. STAI has a grouped version ranging from 0 to 1, representing not anxious and anxious, respectively. On the other hand, BAI ranges from 0 to 2, representing low, moderate, and severe anxiety levels. In the scope of this study, we utilized the BAI with its corresponding levels.
  • Stress
    Stress is a physiological and psychological response to perceived threats, encompassing external and internal stressors, triggering the body’s “fight or flight” response; while acute stress can be adaptive, chronic stress, persisting over time, can negatively impact both physical and mental health, contributing to conditions like anxiety and depression [1,23]. In NetHealth, we have only ‘PSS’ https://www.das.nh.gov/wellness/docs/percieved%20stress%20scale.pdf (accessed on 17 January 2024). We are only provided with its Likert-scale version ranging from 0 to 4, corresponding to never through very often. Given the absence of alternative options for the stress target, we utilized a class-based scale for the entire study. We made choices regarding the survey and value types for other target variables (raw continuous value or classes) by aligning them with the response type available for stress. This decision was made to ensure coherence among target variables despite the differing class scale ranges.

3.3. Missing Data

In the context of our study, spanning 8 semesters of data, we encountered challenges as certain questions were not consistently addressed or answered across all semesters. Despite having rich data over the semesters, the absence of commonalities prompted the need for a comprehensive approach to handling missing data.
To manage missing data effectively, we adopted the Multiple Imputation by Chained Equations (MICE) method [24]. MICE provides a systematic framework for filling in data gaps by generating educated estimates based on the available data. The process involves iteratively imputing missing values for each variable using conditional distributions. In simpler terms, MICE breaks down the imputation process into a series of steps, where each variable’s missing values are estimated based on the observed values of other variables. This iterative process continues until convergence is achieved, resulting in multiple complete datasets.
Upon examination, we identified that missing data from wearables were relatively minimal. However, substantial amounts of missing values were observed in survey data. For certain variables, such as gender and related survey questions, which were not consistently posed every semester, we addressed the issue by replicating a single collected response across all missing columns for each individual throughout different semesters. On the other hand, we employed MICE to fill in the missing values for missing values related to aspects like sleep, health, and mental health features, ensuring a more holistic dataset for our analysis. At the end of this process, we ended up with 310,967 rows and 74 columns.

3.4. Data Preprocessing for Time-Based Methods

To encounter time aspect and temporal dependencies within the data, we employed a time-lagged version of the data where each data point is associated with the previous data point. Even though LSTM has its own time dependency, in our previous study on another dataset [10], we found that LSTM on time-lagged data further improved model performance.
Consider a mental health parameter, like depression, anxiety, or stress (DAS) at a given time ‘t’. To be able to create a lagged version of DAS, we present the DAS parameter as follows:
D A S ( t ) = [ D A S ( t n ) , D A S ( t ( n 1 ) ) , , D A S ( t 1 ) , D A S ( t ) ]
Here, DAS(t) represents one of the mental health parameters (e.g., depression level) at a time ‘t’, ranging from, for instance, 0 to 4 for severity. D A S ( t n ) , D A S ( t ( n 1 ) ) , …, D A S ( t 1 ) , D A S ( t ) represent the individual measurements of the mental health parameter at different time points within a specified time window. The ‘n’ in this context signifies the size of the time window, indicating the number of past measurements considered when assessing the mental health parameter.
In our study, we employed various look-back (n) sizes, such as 1, 7, 15, and 30, a semester to consider daily, weekly, fortnightly, monthly, and semesterly patterns in mental health data.

3.5. Models and Evaluation Metrics

We employed traditional RF, XGBoost, LSTM, and their lagged versions as prediction/regression models. Since one of our goals is to show the effect of MTL, we employed multitask versions of baselines (RF, XGBoost, LSTM) and their lagged versions. Once we found the best-performing look-back and look-up combination with the baseline’s time-lagged version, we employed the same combination, which is 1-day look-up and 15-day look-back, for the time-lagged version of the multitask.
Since all of our tasks are regression-based for MTL, we employed multi-output-based multitask learning. Multitask learning and multi-output learning differ primarily because output variables in multi-output learning often share the same training data or features. In contrast, separate tasks may be learned in MTL on different training sets or features. Multitask supervised learning becomes multi-label learning or multi-output regression when multiple tasks share the training set [6,25]. Instead of running algorithms over the same data for different task predictions, by employing MTL, we aim to run each method once to measure their performance on different tasks.
The RF algorithm is employed due to its performances in the literature [10,26]; XGBoost is employed as it is indicated to perform better compared to deep-learning techniques and even better once combined with deep-learning compared to the alone version [9,27]; and LSTM is employed as it inherently considers the time aspects of data. The hyper-parameters for each algorithm are provided in Table A1.
We also have their lagged versions for 1 , 7 , 15 , 30 , 90 , 180 days look-back to consider the contribution of past data points and 1 , 7 , 15 , 30 days lookup to understand up to how many days we can reach reasonable errors and which one’s performance is better.
We used MAE as an error metric since we did not encounter a sign difference between the target and predicted variable (an example of this visualization can be seen in Figure 2). It belongs to the multitask-LSTM-anxiety case’s visualization. The X axis represents the indexes corresponding to the data points in the testing dataset. Furthermore, to prevent over-fitting, we split the dataset into 5, 2, and 2 semesters for train, validation, and testing, respectively.
Figure 2. Visualization of multitask LSTM prediction results for anxiety over test data.
Please note that while the target variables in the original dataset were discrete, the MICE application resulted in obtaining continuous values for missing target entries. Considering our expertise, especially in the context of psychological questionnaires, we refrained from categorizing these values into specific classes. Instead, we chose to utilize models that align with the nature of the dataset, prioritizing prediction over classification.
Additionally, due to the same consideration of utilizing prediction models, there are instances where the predicted values fall outside the range of the original target values. Despite this, we opted not to remove such instances. Our rationale is rooted in the nature of our predictive modeling approach, where the goal is to estimate numerical values without imposing constraints based on predefined classes.

5. Conventional, Multitask, Time-Based, and Multitask Time-Based DAS Predictions

We are dealing with three target variables: depression on a scale of 0 to 3, anxiety on a scale of 0 to 2, and stress on a scale of 0 to 4. We aim to predict these target values using RF, XGBoost, and LSTM. Additionally, we aim to empirically assess the impact of a lagged dataset on performance and investigate whether multitasking alone or in combination with the lagged version influences outcomes (RQ3, RQ4).
The comprehensive outcomes are outlined in Table 2, with additional insights into time-based performances accessible in Table 3. Despite their minimal impact on the ultimate performance enhancement, these details afford a broad perspective on the consequences of diverse look-back and look-up combinations.
Table 2. Model prediction performances for DAS, presented in terms of MAE, where bold indicates the best performance for each type.
Table 3. Further results of the time-based predictions.
The baseline and MTL performances are closely aligned, with MTL showing no significant improvement. RF and XGBoost results also exhibit a minimal difference in both scenarios. Notably, incorporating the temporal aspect has led to substantial performance enhancements, with the highest outcome achieved using RF with a 15-day look-back and 1-day lookup. While Table 2 outlines the time-based results for a 1-day lookup, Table 3 encompasses various combinations of look-backs and look-ups. In addition, when considering a longer look-back period in the time-based method, such as comparing 15 days to 1 month or 1 month to 1 semester, despite the increased computational and time burden, we did not observe any performance improvement. In the pursuit to refine time-based performances, we implemented the multitask version on top of time-based methods. This strategy yielded improved outcomes for depression and stress target variables, although it did not enhance the anxiety target performances. RF and XGBoost performances remained closely aligned. An interesting observation was the consistently poorer performances with LSTM across all cases compared to RF and XGBoost. However, the difference is not too high. Since LSTM intrinsically handles the time aspect of data and we observe that there is an improvement when we consider lagged data, this might be due to the usage of a simple LSTM architecture with only one layer.
A closer look at Table 3 reveals a general trend of increasing error values (indicating worse performance) from top to bottom as we compare each algorithm with its counterpart in different combinations. However, this trend is not consistently observed in all cases. For instance, the performance of LSTM in the 7-day look-up and 1-semester look-back combination is improved when employing the 15-day look-up and 15-day look-back combination. Therefore, a consistent pattern that applies to all cases and algorithms cannot be conclusively determined. Interestingly, longer look-back periods did not result in performance improvements. Implementing MTL on time-based methods improved outcomes for depression and stress but did not enhance anxiety target performances. LSTM consistently exhibited poorer performances across all cases compared to RF and XGBoost. Analyzing Table 3 reveals an overall trend of increasing error values from top to bottom, which is inconsistent across all cases and algorithms. For example, the LSTM performance improved in the 7-day look-up and 1-semester look-back combination when employing the 15-day look-up and 15-day look-back combination, suggesting a lack of a universal pattern.
To the best of our knowledge, our study is the first to incorporate the time-based aspect of data among the studies utilizing NetHealth. Unlike studies solely focusing on one target variable, we selected three related variables appropriately termed DAS in the literature. Furthermore, in addition to the time-based improvement, we utilized MTL. Since all three tasks involve regression, we approached multitasking as a multi-output regression problem. This approach allowed us to fairly compare algorithm performances only while concentrating on their type changes rather than their architectures.

6. Conclusions and Future Work

We focused on predicting mental well-being factors, including depression, stress, and anxiety, using the NetHealth dataset from college students. Addressing four research questions, we explored digital biomarkers’ impact on these factors (RQ1), their alignment with conventional psychology literature (RQ2), the time-based performance of applied methods (RQ3), and since the target variables are correlated, we wanted to explore further performance improvements while implementing specific methods designed for this purpose such as MTL (RQ4).
Our findings revealed interesting modality rankings for each target variable, aligning with conventional psychology literature for each parameter. Additionally, we validated our results by comparing them with paper-based studies that lacked wearable device usage. We ended up with no parameter coming from wearable on the top 20 variables. That means their contributions to the final model performance are lower than the presented survey parameters in the SHAP figures. Device-measured modalities aligned with traditional studies, and we observed improved prediction performances when considering the temporal aspect of the data. We further improved by employing the multitasking technique while considering the time aspect of data. The updated analysis of mental health multitask results indicates that baseline and MTL performances align closely, with no significant improvement observed in the multitask scenario. RF and XGBoost results exhibit minimal differences, and substantial performance enhancements are achieved by incorporating the temporal aspect, particularly with RF using a 15-day look-back and 1-day lookup.
There are some limitations to this study. We observed that “egoid” is among the most important parameters, indicating that we may achieve further high performances by personalization. However, we did not encounter personalization in the context of the current study. We are planning to consider it in future studies.

Author Contributions

Conceptualization, B.S.; Methodology, B.S.; Validation, Ö.D.İ.; Formal analysis, B.S.; Investigation, B.S.; Writing—original draft, B.S.; Writing—review and editing, B.S., Ö.D.İ.; Supervision, Ö.D.İ. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Boğaziçi University Research Fund, project number: 19301p.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

NetHealth dataset is available at https://sites.nd.edu/nethealth/ (accessed on 27 January 2024).

Acknowledgments

Tübitak Bideb 2211-A academic reward is gratefully acknowledged. We want to thank all the contributors of the datasets used in this study for providing the data for research purposes.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Hyper-parameters for different algorithms predicting each mental-health parameter.
Table A1. Hyper-parameters for different algorithms predicting each mental-health parameter.
AlgorithmHyperparameters
RF-Depression(‘max depth’, 5), (‘min samples leaf’, 1), (‘min samples split’, 2), (‘n estimators’, 10)
RF-Anxiety(‘max depth’, 5), (‘min samples leaf’, 2), (‘min samples split’, 3), (‘n estimators’, 6)
RF-Stress(‘max depth’, 4), (‘min samples leaf’, 1), (‘min samples split’, 10), (‘n estimators’, 5)
XGBoost-Depression(‘alpha’, 1), (‘colsample bytree’, 0.7333635223098999), (‘gamma’, 1), (‘learning rate’, 0.058336569586572316),
(‘max depth’, 8), (‘min child weight’, 3), (‘n estimators’, 82), (‘subsample’, 0.641372976280757)
XGBoost-Anxiety(‘alpha’, 0), (‘colsample bytree’, 0.510066209080432), (‘gamma’, 0), (‘learning rate’, 0.04359820016477877),
(‘max depth’, 7), (‘min child weight’, 8), (‘n estimators’, 70), (‘subsample’, 0.9080823751699505)
XGBoost-Stress(‘alpha’, 0), (‘colsample bytree’, 1.0), (‘gamma’, 0), (‘learning rate’, 0.2993), (‘max depth’, 3), (‘min child weight’, 6),
(‘n estimators’, 157), (‘subsample’, 0.5)
LSTM25 layered LSTM with epochs 10, batch size 72, Adam optimizer with learning rate 0.001, hyperbolic tangent as an activation function,
MAE as a loss function
RF-MTL(‘max depth’, 6), (‘min samples leaf’, 1), (‘min samples split’, 6), (‘n estimators’, 5)
XGBoost-MTL(‘alpha’, 0), (‘colsample bytree’, 0.8530567994101286), (‘gamma’, 1), (‘learning rate’, 0.07023334252979543), (‘max depth’, 4),
(‘min child weight’, 7), (‘n estimators’, 124), (‘subsample’, 0.9930831348301362)
LSTM-MTL32 layered LSTM with epochs 10, batch size 8, Adam optimizer with learning rate 0.001, hyperbolic tangent as an activation function,
MAE as a loss function

References

  1. Lovibond, P.F.; Lovibond, S.H. The structure of negative emotional states: Comparison of the Depression Anxiety Stress Scales (DASS) with the Beck Depression and Anxiety Inventories. Behav. Res. Ther. 1995, 33, 335–343. [Google Scholar] [CrossRef]
  2. Gomes, N.; Pato, M.; Lourenço, A.R.; Datia, N. A Survey on Wearable Sensors for Mental Health Monitoring. Sensors 2023, 23, 1330. [Google Scholar] [CrossRef] [PubMed]
  3. Graham, S.; Depp, C.; Lee, E.E.; Nebeker, C.; Tu, X.; Kim, H.C.; Jeste, D.V. Artificial intelligence for mental health and mental illnesses: An overview. Curr. Psychiatry Rep. 2019, 21, 116. [Google Scholar] [CrossRef] [PubMed]
  4. Bond, R.R.; Mulvenna, M.D.; Potts, C.; O’Neill, S.; Ennis, E.; Torous, J. Digital transformation of mental health services. NPJ Ment. Health Res. 2023, 2, 13. [Google Scholar] [CrossRef]
  5. Purta, R.; Mattingly, S.; Song, L.; Lizardo, O.; Hachen, D.; Poellabauer, C.; Striegel, A. Experiences measuring sleep and physical activity patterns across a large college cohort with fitbits. In Proceedings of the 2016 ACM International Symposium on Wearable Computers, Heidelberg, Germany, 12–16 September 2016; pp. 28–35. [Google Scholar]
  6. Zhang, Y.; Yang, Q. An overview of multi-task learning. Natl. Sci. Rev. 2018, 5, 30–43. [Google Scholar] [CrossRef]
  7. Chung, J.; Teo, J. Mental health prediction using machine learning: Taxonomy, applications, and challenges. Appl. Comput. Intell. Soft Comput. 2022, 2022, 9970363. [Google Scholar] [CrossRef]
  8. Coutts, L.V.; Plans, D.; Brown, A.W.; Collomosse, J. Deep learning with wearable based heart rate variability for prediction of mental and general health. J. Biomed. Inform. 2020, 112, 103610. [Google Scholar] [CrossRef] [PubMed]
  9. Garriga, R.; Mas, J.; Abraha, S.; Nolan, J.; Harrison, O.; Tadros, G.; Matic, A. Machine learning model to predict mental health crises from electronic health records. Nat. Med. 2022, 28, 1240–1248. [Google Scholar] [CrossRef]
  10. Saylam, B.; İncel, Ö.D. Quantifying Digital Biomarkers for Well-Being: Stress, Anxiety, Positive and Negative Affect via Wearable Devices and Their Time-Based Predictions. Sensors 2023, 23, 8987. [Google Scholar] [CrossRef]
  11. Tateyama, N.; Fukui, R.; Warisawa, S. Mood Prediction Based on Calendar Events Using Multitask Learning. IEEE Access 2022, 10, 79747–79759. [Google Scholar] [CrossRef]
  12. Taylor, S.; Jaques, N.; Nosakhare, E.; Sano, A.; Picard, R. Personalized multitask learning for predicting tomorrow’s mood, stress, and health. IEEE Trans. Affect. Comput. 2017, 11, 200–213. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, K.; Wang, C.; Gu, Y.; Sarsenbayeva, Z.; Tag, B.; Dingler, T.; Wadley, G.; Goncalves, J. Behavioral and physiological signals-based deep multimodal approach for mobile emotion recognition. IEEE Trans. Affect. Comput. 2021, 14, 1082–1097. [Google Scholar] [CrossRef]
  14. la Barrera, U.D.; Flavia, A.; Monserrat, C.; Inmaculada, M.C.; José-Antonio, G.G. Using ecological momentary assessment and machine learning techniques to predict depressive symptoms in emerging adults. Psychiatry Res. 2024, 332, 115710. [Google Scholar] [CrossRef] [PubMed]
  15. Kaushik, S.; Choudhury, A.; Sheron, P.K.; Dasgupta, N.; Natarajan, S.; Pickett, L.A.; Dutt, V. AI in healthcare: Time-series forecasting using statistical, neural, and ensemble architectures. Front. Big Data 2020, 3, 4. [Google Scholar] [CrossRef] [PubMed]
  16. Seng, K.P.; Ang, L.M.; Peter, E.; Mmonyi, A. Machine Learning and AI Technologies for Smart Wearables. Electronics 2023, 12, 1509. [Google Scholar] [CrossRef]
  17. Evgeniou, T.; Pontil, M. Regularized multi–task learning. In Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 22–25 August 2004; pp. 109–117. [Google Scholar]
  18. Ruder, S. An overview of multi-task learning in deep neural networks. arXiv 2017, arXiv:1706.05098. [Google Scholar]
  19. Saylam, B.; Ekmekci, E.Y.; Altunoğlu, E.; Incel, O.D. Academic Performance Relation with Behavioral Trends and Personal Characteristics: Wearable Device Perspective. In Proceedings of the Information Society Multiconference, Ljubljana, Slovenia, 10–14 October 2022; pp. 35–39. [Google Scholar]
  20. Akin, A.; Çetin, B. The Depression Anxiety and Stress Scale (DASS): The study of validity and reliability. Kuram Uygulamada Egit. Bilim. 2007, 7, 260. [Google Scholar]
  21. Conejero, I.; Olié, E.; Calati, R.; Ducasse, D.; Courtet, P. Psychological pain, depression, and suicide: Recent evidences and future directions. Curr. Psychiatry Rep. 2018, 20, 1–9. [Google Scholar] [CrossRef]
  22. Abend, R.; Bajaj, M.A.; Coppersmith, D.D.; Kircanski, K.; Haller, S.P.; Cardinale, E.M.; Salum, G.A.; Wiers, R.W.; Salemink, E.; Pettit, J.W.; et al. A computational network perspective on pediatric anxiety symptoms. Psychol. Med. 2021, 51, 1752–1762. [Google Scholar] [CrossRef]
  23. Fink, G. Stress: Definition and history. Stress Sci. Neuroendocrinol. 2010, 3, 3–14. [Google Scholar]
  24. Raghunathan, T.E.; Lepkowski, J.M.; Van Hoewyk, J.; Solenberger, P. A multivariate technique for multiply imputing missing values using a sequence of regression models. Surv. Methodol. 2001, 27, 85–96. [Google Scholar]
  25. Xu, D.; Shi, Y.; Tsang, I.W.; Ong, Y.S.; Gong, C.; Shen, X. Survey on multi-output learning. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 2409–2429. [Google Scholar] [CrossRef]
  26. Taelman, J.; Vandeput, S.; Spaepen, A.; Van Huffel, S. Influence of mental stress on heart rate and heart rate variability. In Proceedings of the 4th European Conference of the International Federation for Medical and Biological Engineering: ECIFMBE 2008, Antwerp, Belgium, 23–27 November 2008; Springer: Berlin/Heidelberg, Germany, 2009; pp. 1366–1369. [Google Scholar]
  27. Shwartz-Ziv, R.; Armon, A. Tabular data: Deep learning is not all you need. Inf. Fusion 2022, 81, 84–90. [Google Scholar] [CrossRef]
  28. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. Adv. Neural Inf. Process. Syst. 2017, 30, 1–10. [Google Scholar]
  29. Shekhawat, N.; Jain, K. Relationship of Self-Regulation with Mental Health Among Boy and Girl Adolescents. Int. J. Indian Psychol. 2022, 10, 463–467. [Google Scholar]
  30. Scott, A.J.; Webb, T.L.; Martyn-St James, M.; Rowse, G.; Weich, S. Improving sleep quality leads to better mental health: A meta-analysis of randomised controlled trials. Sleep Med. Rev. 2021, 60, 101556. [Google Scholar] [CrossRef] [PubMed]
  31. Billah, M.A.; Akhtar, S.; Khan, M.N. Loneliness and trust issues reshape mental stress of expatriates during early COVID-19: A structural equation modelling approach. BMC Psychol. 2023, 11, 140. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.