Next Article in Journal
Robust Self-Adaptation Fall-Detection System Based on Camera Height
Next Article in Special Issue
Detecting Moments of Stress from Measurements of Wearable Physiological Sensors
Previous Article in Journal
Laser Safety Calculations for Imaging Sensors
Previous Article in Special Issue
Predicting Depression, Anxiety, and Stress Levels from Videos Using the Facial Action Coding System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Empirical Study Comparing Unobtrusive Physiological Sensors for Stress Detection in Computer Work

1
Department of Informatics, University of California, Irvine, CA 92617, USA
2
Computational Physiology Laboratory, University of Houston, Houston, TX 77004, USA
3
Perception, Sensing, and Instrumentation Laboratory, Texas AM University, College Station, TX 77843, USA
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(17), 3766; https://doi.org/10.3390/s19173766
Submission received: 21 July 2019 / Revised: 15 August 2019 / Accepted: 28 August 2019 / Published: 30 August 2019
(This article belongs to the Special Issue Sensors for Affective Computing and Sentiment Analysis)

Abstract

:
Several unobtrusive sensors have been tested in studies to capture physiological reactions to stress in workplace settings. Lab studies tend to focus on assessing sensors during a specific computer task, while in situ studies tend to offer a generalized view of sensors’ efficacy for workplace stress monitoring, without discriminating different tasks. Given the variation in workplace computer activities, this study investigates the efficacy of unobtrusive sensors for stress measurement across a variety of tasks. We present a comparison of five physiological measurements obtained in a lab experiment, where participants completed six different computer tasks, while we measured their stress levels using a chest-band (ECG, respiration), a wristband (PPG and EDA), and an emerging thermal imaging method (perinasal perspiration). We found that thermal imaging can detect increased stress for most participants across all tasks, while wrist and chest sensors were less generalizable across tasks and participants. We summarize the costs and benefits of each sensor stream, and show how some computer use scenarios present usability and reliability challenges for stress monitoring with certain physiological sensors. We provide recommendations for researchers and system builders for measuring stress with physiological sensors during workplace computer use.

1. Introduction

Many individuals spend an increasingly significant proportion of their day at a computer, especially those in information work. Some workplace computer tasks are known to be associated with stress, such as answering emails [1,2] and presenting to a remote audience [3,4]. Besides cognitively demanding tasks, workplace stressors include time pressure [5], social pressure [6], interruptions [7] and anticipatory stress from upcoming deadlines [8,9]. Excessive exposure to workplace stress has direct effects on health and quality of life, as it can lead to burnout, diminished productivity, and several health problems including cardiovascular disease and impaired immunity functions [10,11,12]. Thus, capturing stress levels in the workplace is vital for improving our understanding of real-life stress and the factors surrounding it. Measuring stress unobtrusively and in real time at the workplace can enable affective computing applications that incorporate user’s stress and new forms of context-aware interactions [13,14]. Mental health professionals and organizational psychologists can also benefit from stress monitoring at the workplace, to better understand stress and associated factors, and to deliver interventions.
To capture stress in the workplace, several methods have been tested. Self-reported stress, also referred to as ‘perceived stress’ [15], is often considered a ground truth of stress. Several instruments have been developed, such as the Perceived Stress Scale [16], the Daily Stress Inventory [17], and one-item surveys deployed through experience sampling [18]. Although self-report instruments are commonly used in the literature, they have several limitations for stress monitoring at the workplace. Self-reports are subjective and are affected by memory and emotion expression biases. They can also be disruptive as they require the full cognitive attention of the user, and do not allow continuous stress measurement. Advances in sensor technologies embedded in wearable devices have motivated researchers to investigate the usability of unobtrusive and wearable sensors for stress measurement in the workplace [19], especially during computer use. As stress produces several physiological reactions, capturing physiological signals with sensors provides the potential to measure stress objectively, unobtrusively, and in real time.
In this paper, we review recent research on stress measurement with physiological sensors in workplace and computer use settings. We identify a gap in the literature as most studies focus on specific high-stress short-duration computer tasks to induce stress [20,21,22], which might not be representative of those in real workplace settings and can overlook issues and challenges related to stress measurement with physiological sensors during different computer activities. Workplace computer use includes activities that vary in the level of cognitive or emotional stress they could induce, the physical motions and dexterity they require, the user posture, and their duration, all which can potentially affect sensor performance. An empirical study to examine the usability and reliability of a set of unobtrusive sensors across a spectrum of computer activities is lacking.
This paper addresses the following research question: what sensor modality functions best to measure stress across computer tasks? To answer this research question, we compare the use of different sensor modalities across varied computer tasks, investigating the usability, reliability, and problems with each type of sensor. We report our results based on testing the sensors in a simulated office environment. The contribution of this work is as follows:
  • A review of the literature on stress measurement with sensors in the workplace and laboratory studies examining computer use. Unlike reviews focusing on the results of these studies, we focus on the methods and present a summary of sensors used, tasks performed, ground truth measures, and other dependent stress variables, the number of subjects, and the duration of measurement.
  • An empirical comparison of the usability and reliability of a set of sensor channels for stress response measurement during computer use, including an emerging non-contact method using thermal imaging.
  • Identifying challenges for some sensors, specific to certain common computer tasks, which limit the efficacy of these sensors for continuous stress monitoring in situ in a workplace setting.
  • Recommendations for researchers and system builders interested in stress measurement with unobtrusive physiological sensors during computer use.
To the best of our knowledge, this is the first study to include this collection of physiological sensor streams (heart-rate from ECG and PPG sensors, breathing rate, skin conductance and thermal imaging) which are collected simultaneously for several computer tasks, and the first study to compare thermal imaging against other wearable physiological sensors as a stress measurement technique in computer tasks (for previous studies using thermal imaging for stress detection in other contexts, see [22,23,24,25,26]).

2. Background

2.1. Physiological Stress Reactions

Stress is the result of appraising a situation as having demands exceeding resources [27] (e.g., time, mental resources, money, etc.). When the body experiences stress, several physiological events occur driven by two branches of the Autonomic Nervous System, which is responsible for regulating bodily functions. The first branch is the Sympathetic Nervous System, which drives the body’s resources to respond to a challenge or a threat; sympathetic activation leads to an increased heart-rate and respiration rate, and tense muscles, a reaction that is known as the ‘fight-or-flight’ response [28]. During the fight-or-flight response, systems that are not essential to immediate survival, such as the digestive system, the reproductive system, and the immune system are suppressed. This process is complemented by the Parasympathetic Nervous System, which brings the body back to a state of relaxation. In non-stressful settings, these two systems work in coordination to achieve homeostasis, the condition where internal functions remain stable and balanced. In a stressful setting, the autonomic nervous systems are unbalanced. Prolonged imbalance in these two systems leads to long-term health problems. In the short term, acute stress is associated with negative feelings such as anger, frustration and fear [29], as well as lack of motivation, impaired decision making and decreased productivity [10,30].
The gold-standard in measuring stress is measuring the level of cortisol, known as the stress hormone. Cortisol measurements are used in clinical and lab studies (see [31] for a review), but they are unsuitable for workplace settings as they require collecting saliva or blood samples. As an alternative, continuous, unobtrusive stress tracking through other physiological measures has been explored.
In a survey of affective computing for stress detection, Greene et al. [32] list several measures of stress manifested in the human body that can be measured with existing technologies and that have been used in affective computing applications. They divide these bodily measures of stress into physical and physiological measures. Physical measures include facial expressions, eye activity, and body gestures, whereas physiological measures include cortisol level, brain activity, muscle activity, heart activity, skin response, respiratory activity and blood activity [32]. In this paper, we focus on physiological measures that can be obtained through unobtrusive wearable sensors.

2.2. Unobtrusive Physiological Sensing

For seamless and continuous stress monitoring in the workplace, stress measurement tools should not interfere with the user’s work or create additional stress and burden. Advances in wearable sensors and algorithms to analyze physiological signals enable continuous unobtrusive sensing of stress-related processes, which cannot be achieved with traditional self-report surveys. With recent wearable sensors, unobtrusive continuous monitoring of workplace stress is possible through capturing and analyzing physiological reactions to stress, such as changes in heart activity, skin conductance, and breathing. Sensors that capture these physiological signals can be embedded in comfortable wearable devices such as wristbands and chest bands. Besides wearable sensors, researchers have recently explored non-contact alternatives to approximate physiological signals. Digital cameras have been used to approximate physiological attributes based on analyzing facial expressions and subtle variations in skin tone (e.g., [33,34]). One of the most recent non-contact methods for unobtrusively measuring physiological signals associated with stress is thermal imaging. Thermal imaging provides a heatmap of the area of interest (usually the face) and highlights changes in surface temperature that are associated with stress, such as warming of the forehead [22] or perspiration in the perinasal area [24]. Thermal imaging for stress detection has been successfully validated in several contexts such as surgical training, security monitoring duty and office space [22,23,24,25,26].

2.3. Stress Monitoring in the Workplace: Literature Review

In this section, we review studies of stress monitoring in the workplace. The reviewed studies include research in workplace settings or computer use contexts. To narrow the scope of the review, we consider studies that use physiological signals to detect stress, and exclude studies focused on physical, facial and behavioral signals of stress (e.g., [35,36,37,38,39,40,41]). Studies approximating physiological measures with motion-based sensors such as accelerometers and gyroscopes (e.g., [42,43,44]) are also beyond the scope of this review.
Seventeen publications in areas spanning human–computer interaction, ubiquitous computing, biomedical informatics, user modeling, multimodal interaction, and affective computing from the years 2006–2017 were included in this review. Table 1, Table 2 and Table 3 summarize the reviewed studies based on the sensors and physiological signals, the computer task/stressor involved, the dependent variable (i.e., the stress measure), number of subjects, duration of physiological measurement and whether it is a lab or field study.
Most of the reviewed studies are controlled lab studies where subjects perform a task on the computer while wearing sensors to capture stress. The reviewed studies used computer tasks that simulate workplace computer use scenarios that might lead to stress. The tasks include computerized versions of validated stress-inducing tasks such as problem solving, solving puzzles, memory tasks, cognitive tasks, and mental arithmetic. Some tasks are validated stressors (such as the Stroop Color-Word test) while other tasks had additional stressors introduced (such as time pressure or social stress) to create the desired effect. For most studies, sample size ranged from 10 to 35 subjects, but varied in terms of unit of analysis (i.e., hours, sessions). A direct comparison of the results of all the above studies is not possible due to their differences in stress definitions, study design, sensors used, features extracted, and analysis methods.
The most common experimental setting in the reviewed studies was comparing a condition where stress was induced (e.g., by performing a stressful task or introducing social stressors), against another condition where no stress was induced. This approach results in binary classification models where data points are classified into either stress or rest. This classification is an oversimplification of workplace stress, as employees are seldom at rest (i.e., doing nothing). Some studies tried to address this limitation by increasing the number of classes (e.g., ‘relaxed’, ‘concentrated’, and ‘stressed’ in [50]) or replacing the ‘rest’ condition with non-stressful computer work (i.e., ‘low cognitive load’ vs. ‘stress’ in [51]). Other than predicting the stress condition, studies have also considered self-reports as ground truth, and used physiological signals as predictive variables (e.g., [20,49]). Finally, a stress measure that has been used, which captures more variation in stress, is departure from the baseline physiological measure, where stress is said to be detected if the physiological signal during the task is higher than the subject’s baseline measure (e.g., [21,22]).
While many studies measure stress during standardized computerized tasks (such as the Stroop Color-Word test) as a proxy for workplace computer use, Koldijk et al. [47,48] present a dataset of physiological measures during email interruption and time pressure as simulated workplace stressors, validated by self-reports of mental load. Using this dataset, Sriramprakash et al. [53] were able to build a model discriminating a neutral condition from the email interruption and time pressure condition using heart-rate and skin conductance measures. More work exploring workplace computer use scenarios beyond standardized computerized stressors is needed to account for the variation in workplace activities and the possible challenges for real-time stress monitoring during those activities, which is what we present in this study.
While the reviewed studies help advance unobtrusive stress measurement in the workplace, deploying these systems in real-life work scenarios requires a more nuanced understanding of the costs and benefits involved, and their robustness across different computer use scenarios in the workplace. We present a study where several sensor streams measuring stress were collected simultaneously during the performance of several computer tasks commonly performed at the workplace.

3. Methods

3.1. Experiment Design

As a part of a larger study on workplace stress, we simulated a workplace scenario where subjects conducted several tasks on the computer. The experiment consisted of several phases (Figure 1) starting with obtaining consent, filling out demographic and psychometric surveys, and equipment setup, followed by four minutes of rest to obtain baseline physiological measures for each subject. After the resting baseline period, subjects were asked to write an essay about a given topic for five minutes. Next, half of the subjects took the Stroop Color-Word test (CWT), while the other half watched a calm video. The Stroop Color-Word test is a validated stressor where the subject is shown a word designating a color, and the subject has to choose the color of the font of the word, rather than the color the word is designating. Next, subjects were asked to complete a 50-min dual task (DT) that consisted of writing an essay while responding to emails delivered under two conditions that represented high or low degrees of interruption. In the high interruption condition (multitasking), subjects were frequently interrupted by emails which they had to respond to as they arrived, simulating a multitasking computer work condition. In the low interruption condition (monotasking), subjects received the same number of emails but in a batch (i.e., all together) and had dedicated time to reply to them before returning to work on the essay. Finally, subjects were asked to present their essays in front of a virtual audience through video conferencing, a common real-life workplace task (In the original study, half of the participants were told in advance that they have to present in order to create anticipatory stress, while the other half were not told so. We found no difference in the physiological measures or self-reported stress of the two groups, hence, they are grouped together in this study.). More details about the experiment design, including additional measures collected but not used in this study, can be found in [8]. The experimental session, including preparation, setup and all tasks, lasted about 90 min. The experiment’s data, software systems, and data curation scripts along with the generated text and videos are publicly available at https://osf.io/zd2tn/.
The experiment was designed to minimize experimenter interaction to avoid additional stress caused by the experimenter’s presence. Software was designed so that each experimental phase could progress to the next, with minimal experimenter interaction. The experimenter sat in the same room and was separated from the subject by a partition. The experiment took place in a typical office booth with a desktop computer and a 24-inch display monitor. A webcam was placed on top of the monitor screen to video record the sessions so that it could be used to diagnose abnormal sensor readings.

3.2. Participants

Recruitment took place across three university campuses in the U.S. west and southwest through emails and flyers calling for participation. Participants had to be at least 18 years of age, have done all their schooling in English, and have at least a high school education to be eligible for this study. We recruited 96 participants, out of which 33 were excluded due to technical errors that caused their data not to be recorded. The remaining 63 participants (45 females and 18 males) were aged between 18 to 54 years, with a mean of 23.75 years (SD = 8.76 years). All participants signed informed consent and the study was approved by the institutional review boards of the participating universities. Two participants were removed as they had an abnormal resting baseline sensor reading. Two participants withdrew before or during the presentation session. Thus, data from 61 participants are included in this study, with the presentation session having 59 participants.
The majority (79%) of the participants were undergraduate students, 8% were graduate students, and 10% were employees. All participants rated themselves as fluent in English (score of 4+ on a 7-point scale) and the majority (93.4%) said they use email often in their daily work and life (score of 4+ on a 7-point scale), which makes them suitable for our study of computer work tasks involving writing essays and handling emails.

3.3. Sensors and Measures

We used well-validated sensors that provide measurements of the physiological changes that accompany stress. The following sensors and measures were used:
  • Zephyr™BioHarness 3.0 Chest Strap with a BioModule: provides an ECG sensor for heart-rate (chest.HR) monitoring and an internal breathing sensor (BR). Subjects were instructed to wear the chest-band under their clothes for direct skin contact.
  • Empatica E4 wristband: provides a PPG sensor for heart-rate monitoring (wrist.HR) and an EDA sensor. Participants wore the device on the wrist of their non-dominant hand.
  • Thermal camera, Tau 640 longwave infrared (LWIR) camera (FLIR Commercial Systems, Goleta, CA, USA): Captures thermal images of subjects’ faces, from which perinasal perspiration (PP) is extracted using an algorithm by [24,58]. The thermal camera was placed under the participant’s computer monitor.

3.4. Data Normalization

There is no consensus in the literature on whether and how physiological features should be normalized by accounting for the baseline (i.e., tonic) or average level. Some studies show support for normalizing physiological features (e.g., [59]) while others found that models perform better with non-relative features (i.e., non-normalized) [51] or a mix of relative and non-relative features [60]. In our analysis we choose to normalize the obtained physiological signals by subtracting baseline values for each participant (i.e., physiological signals at rest) to account for individual differences and capture stress as a departure from the baseline level.

4. Results

4.1. Capturing Stress

We denote stress as a departure from the baseline physiological level for each participant. Equation (1) shows how stress was calculated for each participant (P), during each session (S), for each sensor stream (E). The stress in a given session is given by deducting the mean of physiological signal i during the resting baseline session for participant j from the mean of physiological signal i during session k for participant j, where i { P P , B R , c h e s t . H R , w r i s t . H R , E D A } , k { e s s a y w r i t i n g , C W T , r e l a x i n g v i d e o , m o n o t a s k i n g , m u l t i t a s k i n g , p r e s e n t a t i o n } , j { 1 , , 63 } .
Δ E i ( P j , S k ) = E i ¯ ( P j , S k ) E i ¯ ( P j , S r e s t i n g b a s e l i n e )
We compare the set of sensor signals by their accuracy in capturing changes in stress during the different tasks by (1) using a t-test to determine whether there is a statistically significant difference between each session and the baseline; and (2) comparing the percentage of subjects for whom stress was captured by each sensor in each session.

4.1.1. Essay Writing Session

During the first essay writing session, PP, chest.HR and BR showed a higher average than the baseline session ( p 0.001 for each). Signals from the wrist sensors (i.e., EDA and wrist.HR) did not show a statistically significant difference from baseline. Table 4 shows the ratio of subjects with a mean difference greater than 0 (i.e., higher than baseline) for the writing task. PP picked up the increased stress for 90% of the subjects, exceeding other signals which might not be as sensitive or generalizable across subjects. Since this task involves typing and moving the wrists, a potential explanation for the poor signal of EDA and wrist.HR for many subjects could be the influence of motion artifacts on the signal obtained from electrodes and PPG sensor on the wrist, and potential friction or detachment of the sensors which can cause sudden peaks or drops in the signal.

4.1.2. Color-Word Test

This task is a validated stressor, and its computerized version is used in many simulated workplace stress studies (e.g., [52,55]) as a proxy for cognitively demanding computer tasks at the workplace. Ideally, all measures should show a significant increase in stress. However, in our data, only two measures detected a higher level of stress (Table 5). PP and BR showed a statistically significant difference from baseline ( p 0.01 ), chest.HR showed a trend of an increase ( p 0.06 ), while wrist sensors (EDA and wrist.HR) did not show a statistically significant difference. Although the overall EDA average across participants is not significantly higher than the baseline, EDA was higher during this session for 83% of the subjects, which might indicate that a few outliers affected the overall average. This task involved using the mouse with the dominant hand, so we do not expect that the wrist sensor placed on the non-dominant hand was affected by motion artifacts.

4.1.3. Watching a Relaxing Video

We expected that watching a relaxing video would not generate any increase in stress compared to the baseline for any of the sensor streams. We found that PP captured a small increase in stress, while chest.HR captured a decrease in stress. The increase in PP was the smallest across all the sessions, which aligns with our expectation. Similarly, no other session generated as much decrease in chest.HR. These results for PP and chest.HR were to a large extent generalizable across subjects, with 79% of subjects showing a small increase in PP, and 76% of subjects showing a decrease in chest.HR. For other sensor streams, the difference from baseline was not statistically significant (Table 6).

4.1.4. Dual Task—Monotasking and Multitasking

When subjects work on two tasks, whether monotasking or constantly switching between tasks, their stress level is expected to increase [61]. In our data, we found that PP and BR captured an overall statistically significant increase during the dual task ( p 0.001 ), while chest.HR, wrist.HR, and EDA did not. In terms of percentage of subjects showing an increase in stress, PP captured the increase for most subjects (Table 7 and Table 8). Given that this task includes typing, typically with both hands, we expect that wrist sensors had motion artifacts that might have affected the signal.

4.1.5. Presenting to a Virtual Audience

Giving a presentation is a validated stressor in previous studies [3,4,62]. We expected that all sensor streams would show a significant increase in stress during this session. Our results (Table 9) show that only PP and chest.HR have an overall statistically significant increase in stress ( p 0.0001 ) with 96% and 82% of subjects showing higher stress than baseline level for PP and chest.HR, respectively. For both PP and chest.HR, the average increase in stress during the presentation session is the highest across all sessions, which is aligned with our expectations. For BR, we expect that the signal would be affected by speech respiratory patterns [63], which might explain the non-discriminant signal. Wrist sensors did not capture an overall statistically significant difference from baseline, although a higher stress level was captured for 74% and 53% of subjects for EDA and wrist.HR, respectively.
Overall, PP with the thermal camera detected an increase in stress for all computer tasks, with the highest increase being during the presentation session, and the lowest increase being during the video watching session, as expected. HR from the chest-worn sensor detected an increased stress only during essay writing and presentation sessions. BR showed an increase in all stressful sessions except presentation, where speech respiratory patterns interfere with the signal. Signals from the wrist-worn sensor (i.e., EDA and wrist.HR) did not capture increased stress for any session overall, although increased stress was captured for many individual subjects (50% of subjects in some sessions).

4.2. Sample Participant Data

Besides the statistical tests to determine whether stress was detected, we visually inspected each participant’s data for each sensor stream to identify patterns and potential issues. Below we provide examples of the patterns and issues identified, which explain and visualize the results of the previous section on capturing stress.

4.2.1. EDA

For most subjects, the EDA level was close to zero, with small amplitude dynamics (i.e., small difference between the signal’s extreme values). Typically, a signal with small amplitude dynamics and no significant phasic activity (i.e., no abrupt peaks) makes it challenging to capture physiological changes associated with a task, especially in the absence of a discrete stimuli. However, we found that the tonic EDA level (i.e., the slowly increasing smooth pattern) differed among sessions for some participants. Figure 2 shows two examples where the baseline level is the lowest, and the presentation session has the highest EDA, as expected. It is important to note that the tonic EDA level can be naturally increasing over time, which contributed to finding significant differences among the sessions. However, as can be seen in the second chart in Figure 2, the order of the increase in EDA level does not always follow the chronological order of the sessions (i.e., the CWT session has higher tonic EDA than the DT, although CWT preceded DT). This figure also shows the effect of typing on the quality of the signal, as it shows more noise in the essay writing session and in certain bouts during the dual task session, compared to sessions that did not require typing. The severity of signal disturbance due to typing differed among subjects, with some subjects having significantly more noise than others.
For 18 participants, no significant signal was detected (i.e., EDA 0.02 for all sessions) and hence the stress level in different tasks could not be distinguished.

4.2.2. Heart-Rate (Wrist PPG Sensor)

The wrist-worn HR sensor provided the average heart-rate with a sampling frequency of 1 Hz. The HR signal is filtered by the device to remove motion artifacts. It is expected that HR increases with stress. For most participants, it could not be established that the resting baseline is the lowest HR across the different tasks, as can be seen in Figure 3. However, a higher HR was detected for some participants during the presentation session, suggesting that HR from a PPG wrist sensor can capture strong stress reactions but is non-discriminant for lower stress reactions. Thus, the wrist sensor did not provide an HR signal that can capture stress in different computer task scenarios.

4.2.3. Breathing Rate

As can be seen in the examples in Figure 4, the breathing rate signal degrades in the presentation session, even if it shows a higher BR than baseline for some subjects.

4.2.4. Heart-Rate (Chest ECG Sensor)

For most participants (83%), chest.HR detected the expected increased stress during the presentation session. Figure 5 shows an example of a participant with a clear difference in HR during different sessions. For participants where no significant difference was detected, inspecting the data showed some high frequency responses that made the signals from different sessions overlap (Figure 6), making capturing stress responses difficult.

4.2.5. Perinasal Perspiration

For most participants, the PP signal shows smooth patterns with clear distinctions among the different sessions. Figure 7 and Figure 8 show examples from a participant who received the color-word test as the third task, and another participant who received the relaxing video as the third task in the experiment, to eliminate the potential confound of potentially naturally increasing PP over time. As can be seen in Figure 7 and Figure 8, the relaxing video is the closest to the baseline PP level, while the color-word test is closest to the presentation PP level, as expected.

4.3. Missing Data

The thermal camera is the only non-wearable sensor in our experiment. While wearable sensors are attached to the skin and provide continuous readings, the thermal camera’s continuous reading is dependent on having the participant in a relatively still position facing the camera. Therefore, we investigate gaps in the recorded PP readings across different tasks to assess the suitability of using thermal imaging in different computer use contexts.
As can be seen in Figure 9 and Table 10, the thermal camera captured perinasal perspiration continuously with fewer gaps in tasks where subjects were sitting still, looking straight with minimal head movement. The session with the least missing data percentage is the resting baseline, with less than 10% missing data for each individual subject (average 0.3% for all subjects).
The virtual presentation session had the highest percentage of missing data, reaching more than 50% for some subjects. After revisiting the captured thermal video, we noticed that subjects were moving more than other sessions, which causes the perinasal area tracker to be lost. However, the average percentage of missing data is only 11%, with half the participants having less than 2% missing data. In a previous study by Hernandez [45], physiological measures obtained in an in situ study had an average of 0% missing data for wrist EDA, and 8%, 20% and 39% missing data for chest sensors’ HR, HRV, and BR, respectively. Our results with thermal imaging outperform chest sensors in terms of providing continuous readings and minimizing missing data. However, an in situ study spanning several workdays is needed to make a direct comparison with previous studies.
Some instances of missing data with thermal imaging were successfully avoided or recovered by having the experimenter re-select the perinasal area during the experiment, or by post-hoc analysis of the recorded thermal videos. Missing data also occurred when the experimenters failed to focus the camera on the participant at the beginning of the experiment, or anytime when the camera’s focus was lost during the experiment.
Overall, the average ratio of missing PP data is low (Table 10) with presentation being the session with the highest average ratio of missing data (11%) and the remaining sessions having between zero and 6% missing data, with at least 50% of subjects having less than 2% missing data in each session.

5. Discussion

Given the variety of computer tasks conducted at the workplace, our analysis showed that some sensors do not perform accurately to capture stress during certain tasks. For wrist-worn sensors, several reasons could cause the failure to capture stress. Sensor readings are prone to different types of sensor artifacts. For example, sensor electrodes can move, detach from the skin, or change in pressure on the skin, all which can affect the sensor signals, especially in dexterous tasks such as typing. In addition, we used a wrist sensor with dry electrodes, which depend on sweat for conductance. Thus, for calm sedentary users in an air-conditioned lab, the EDA signal might require a length of time of skin contact with the electrodes for the signal to appear. Previous studies have reported that detection of small EDA responses with wrist sensors is problematic [64], which might also explain why EDA did not detect the mild stress from computer tasks in our study. Palmar EDA (EDA obtained from the palm, or palm side of fingers) have shown better results for classifying calm and distress in sedentary settings in previous studies [65], but can be uncomfortable to wear during some computer activities. Finally, some subjects naturally do not produce adequate EDA signal in at least one wrist [66].
Many studies on unobtrusively capturing workplace stress with physiological sensors focus on specific high-stress computerized tasks. With a similar rationale as our study with common office tasks, McDuff et al. [34] considered more realistic everyday computer activities that require cognitive processing and dexterity. Their selected computer activities could introduce motion artifacts that can negatively influence the quality of the physiological readings and introduce physiological changes associated with body motions. Among other findings, McDuff et al. [34] report that HR and BR alone were not very discriminative indicators of cognitive stress, although their previous work showed BR to be significantly different during cognitive tasks compared to rest periods. Therefore, they suggest that BR might be dependent on the type of task and thus less generalizable. Another study with common office computer tasks (i.e., email interruptions) also reported low accuracy for predicting stress with HR and EDA [47]. Our findings are consistent with previous studies on common workplace computer tasks, showing that BR and HR for capturing stress are task-dependent.
While chest-worn sensors can provide an accurate reading for HR and BR, several considerations must be taken into account to ensure acquiring a good signal and reducing noise. For example, posture is important to avoid an abnormal signal. HR signals from the chest-worn sensor can drop to zero if the sensor disconnects due to crouching. HR signal can also be abnormally high due to sensor friction with the skin producing strong high frequency responses. Ramos et al. [60] reported that they instructed participants to refrain from leaning against the back of the chair to avoid signal noise introduced into the BR readings from the chest-worn sensor when the device was pressed against other objects, which makes wearing the sensor during real-life work contexts uncomfortable. Lastly, BR as a measure of stress is not accurate when the subject is talking, which restricts some workplace scenarios for using this signal to detect stress. These limitations introduce a usability problem with a cost-benefit tradeoff, where producing a good signal might require uncomfortable posture and restricted activities.
Additional filtering for noise reduction can partially address artifact-contaminated signals. However, since the focus of this study is to highlight the issues for different sensor streams during several common computer tasks, we did not pursue developing algorithms for further denoising. Previous work has investigated approaches to process artifact-contaminated data. For example, Hernandez [45] used a motion-sensor to detect ‘still’ moments in daily activities to opportunistically measure HR and respiration within the detected still motion time. Another approach by Alamudun et al. [67] suggest a preprocessing technique to remove the effects of factors interfering with physiological signals (e.g., posture or physical activity). They used a method called orthogonal signal correction, which attempts to remove any source of variance that is orthogonal to the dependent variable of stress level. Another method they used is linear discriminant correction, which models the source of noise (i.e., posture or physical activity) and removes it from the physiological signals. Their methods improved stress prediction from physiological data from an accuracy of 53.5% to 76.3%. However, it is unclear whether these approaches that have been developed for physical activities such as walking can successfully address motion artifacts from crouching or finer-grained activities such as typing.
Considering all methods, we found that perinasal perspiration with a thermal camera is the most generalizable method to capture stress across different tasks, as it can capture even slight changes and is robust against subject movement during computer tasks, providing reliable and continuous measurement with minimal missing data.
Our findings reveal that stress measurement in workplace environments, though important to do, is challenging, and relying on a single modality has many limitations. Previous studies have provided support for multimodal stress measurement given that physiological, personality, gender, sensor location and subject posture affect the selection of the best features to predict stress [8,45,50,68,69]. We extend those findings to show that the performed tasks also affect the choice of the best sensor signals. It may, however, be impractical to use multiple types of sensors. Therefore, thermal imaging appears to offer the most benefit in terms of usability and signal validity and reliability in the context of sedentary computer work.
In terms of usability, all sensors used in the study are unobtrusive and do not interfere with people’s ability to perform computer tasks. Sensors using electrodes (i.e., EDA and ECG sensors) can be uncomfortable for long-term use, as the electrodes become sticky after prolonged contact with the skin. This problem is avoided with non-contact thermal imaging. For data collection and analysis, all devices used come with software that collects and processes raw signals in real time, which is useful for human–computer interaction researchers who want to use these sensors in lab or in situ studies. Thermal imaging has the additional advantage of having the thermal video, which allows for revisiting the video to investigate abnormalities and re-extract features. In terms of cost, all devices have low costs during use and the main cost is the upfront cost of the device.

5.1. Scientific Contribution

The main contribution of our work lies in the breadth of sensor comparisons we used and the context in which they took place. A few other studies in affective computing have conducted sensor comparisons (e.g., [32,70,71]). However, our study is the first to compare thermal imaging and wearable sensors, capturing multiple physiological variables from different parts of the body with different measurement techniques. The breadth of sensors investigated positions this study as a reference for researchers and practitioners (see Section 5.3).
Another distinct and important contribution is the context of this study. Most previous studies of empirical comparisons of sensors take place in a context of either using a highly restricted experimental task in the laboratory (e.g., the Stroop Color-Word test) or as observations in the wild. In the case of restricted, standardized tasks conducted in the laboratory, experimenters have control over confounding factors, but ecological validity is compromised, which raises questions about the relevance of the findings for real-world applications. Sensor measurements done with an experimental task in an abstract lab environment may lack important characteristics that are associated with office tasks, such as time pressure or semantic context. In the case of field studies, ecological validity is high, but confounding factors are hard to control, which affect the robustness of sensor comparisons. The context of this current study aimed for ecological validity with multiple common computer tasks, instead of using an abstract laboratory task. Hence, while we controlled for confounding factors, the computer tasks we used are generalizable to real-world office tasks, which makes the sensor comparisons more relevant for use in the workplace. We used a variety of office tasks that were complex, and common in the information workplace, such as answering email and giving presentations.
Lastly, as a result of our sensor comparisons, our empirical study showed thermal imaging to be a robust stress measurement technique that is suitable for workplace and computer use settings, as it is less affected by confounding variables that introduce noise to other wearable sensor streams. Moreover, physiological sensing with thermal imaging has a capacity for correction, because it is not a one-dimensional temporal signal, but a derivative signal from imagery, which can be improved with better extraction processes or algorithms, even years after its original capture. This finding and the empirical testing of thermal imaging in a realistic context advances affective sensing methods and has implications for researchers and system builders.

5.2. Limitations

Our analysis investigated five common sensor streams. However, there are more physiological signals that can be unobtrusively monitored to measure stress that were not covered in our study. For example, heart-rate variability (HRV), blood volume pulse (BVP) and skin temperature (ST) can be extracted from sensors embedded in wearables [72]. Future work can compare HRV, BVP, and ST with other physiological signals during different workplace computer tasks.
Finally, despite having simulated a workplace environment which allowed us to investigate specific computer tasks, deploying sensors in real-life contexts can have additional challenges that cannot be modeled in lab settings. In the lab setting, careful instrumentation and real-time inspection of the sensor streams ensured high-quality signals. While our study discussed some challenges that are likely to occur in real-life settings, in situ studies can uncover additional validity and usability challenges for unobtrusive stress monitoring in the wild.

5.3. Insights for Researchers and System Builders

Our work has insights and implications for researchers and system builders, which could be synopsized as follows:
1. Controlled experiments are necessary to study cause and effect by isolating nuisance factors. This, however, does not imply that experimentation needs to be void of realism. In studies of stressful computer-based tasks, researchers relied for too long on standardized treatments alone, such as the Stroop Color-Word test (CWT), to investigate phenomena of interest. Such standardized treatments need to be accompanied by carefully designed realistic tasks (e.g., report writing interrupted by emails in the present study) if the goal is to generalize to real-world applications. Importantly, as the sensing results demonstrated in our study, the stress responses generated by standardized treatments often underestimate the stress responses generated by controlled realistic tasks, and thus potentially by real tasks in the wild as well.
2. All unobtrusive physiological sensors—wearable and imaging—are affected by motion artifacts. The advantage of imaging (thermal imaging in this case), however, is that the physiological signals are extracted algorithmically from video streams. Hence, one can visually identify the cause of noise (e.g., head turn) in the original source and compensate for it, either by removing the specific signal segment or by applying an algorithmic correction. In wearable sensor signals, this is more difficult, because there is no primary source of information (i.e., a 3D matrix) out of which these signals are extracted. The 1D temporal signal is all that the wearable sensor provides, and thus identification of motion artifacts is purely conjectural.

6. Conclusions

We have empirically compared five physiological signals that are known to be associated with stress. Across six computer tasks, perinasal perspiration captured through thermal imaging was the most generalizable as it captured even small changes in all tasks and for most participants. Heart-rate and breathing rate from chest-worn sensors captured changes in stress for some tasks, while heart-rate and EDA from wrist-worn sensors did not capture significant changes in stress, overall. We highlighted the effect of movement during typing tasks, and the effect of speaking during the presentation task. These findings advance our understanding of the complexity of computationally modeling workplace stress. With its breadth of sensor comparisons and realistic context, our study addressed a gap in the affective sensing literature. Our study is a step towards effective unobtrusive monitoring of stress in the workplace taking into consideration the various tasks and the challenges they introduce for stress monitoring.

Author Contributions

Conceptualization, F.A. and G.M.; Data curation, I.P.; Formal analysis, F.A.; Funding acquisition, G.M., I.P. and R.G.-O.; Visualization, F.A.; Writing—original draft, F.A.; Writing—review editing, G.M., I.P. and R.G.-O.

Funding

This material is based upon work supported by the National Science Foundation under grants #1704889, #1704682, and #1704636.

Acknowledgments

The authors would like to thank the research assistants who administered the experiments and curated physiological data.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
BCSTThe Berg Card Sorting Task
BRBreathing Rate
BVPBlood Volume Pulse
CWTStroop Color-Word test
DTDual Task (email and essay)
ECGElectrocardiogram
EDAElectrodermal Activity
HRHeart-Rate
HRVHeart-Rate Variability
MISTThe Montreal Imaging Stress Task
PDPupil Diameter
PPPerinasal Perspiration
PPGPhotoplethysmogram
sEMGSurface Electromyogram
STSkin Temperature
STAIState-Trait Anxiety Inventory
sVRIStress-Induced Vascular Response Index

References

  1. Kushlev, K.; Dunn, E.W. Checking email less frequently reduces stress. Comput. Hum. Behav. 2015, 43, 220–228. [Google Scholar] [CrossRef]
  2. Mark, G.; Iqbal, S.T.; Czerwinski, M.; Johns, P.; Sano, A.; Lutchyn, Y. Email duration, batching and self-interruption: Patterns of email use on productivity and stress. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, San Jose, CA, USA, 7–12 May 2016; pp. 1717–1728. [Google Scholar]
  3. Wallergård, M.; Jönsson, P.; Johansson, G.; Karlson, B. A virtual reality version of the Trier Social Stress Test: A pilot study. Presence 2011, 20, 325–336. [Google Scholar] [CrossRef]
  4. Kelly, O.; Matheson, K.; Martinez, A.; Merali, Z.; Anisman, H. Psychosocial stress evoked by a virtual audience: Relation to neuroendocrine activity. CyberPsychol. Behav. 2007, 10, 655–662. [Google Scholar] [CrossRef] [PubMed]
  5. Michie, S. Causes and management of stress at work. Occup. Environ. Med. 2002, 59, 67–72. [Google Scholar] [CrossRef] [PubMed]
  6. Beehr, T.A. Psychological Stress in the Workplace (Psychology Revivals); Routledge: London, UK, 2014. [Google Scholar]
  7. Mark, G.; Gudith, D.; Klocke, U. The Cost of Interrupted Work: More Speed and Stress. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’08), Florence, Italy, 5–10 April 2008; ACM: New York, NY, USA, 2008; pp. 107–110. [Google Scholar] [CrossRef]
  8. Akbar, F.; Bayraktaroglu, A.E.; Buddharaju, P.; Da Cunha Silva, D.R.; Gao, G.; Grover, T.; Gutierrez-Osuna, R.; Jones, N.C.; Mark, G.; Pavlidis, I.; et al. Email Makes You Sweat: Examining Email Interruptions and Stress Using Thermal Imaging. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, UK, 4–9 May 2019; p. 668. [Google Scholar]
  9. O’Connor, P.; Nguyen, J.; Anglim, J. Effectively coping with task stress: A study of the validity of the Trait Emotional Intelligence Questionnaire–Short Form (TEIQue–SF). J. Personal. Assess. 2017, 99, 304–314. [Google Scholar] [CrossRef] [PubMed]
  10. Colligan, T.W.; Higgins, E.M. Workplace stress: Etiology and consequences. J. Workplace Behav. Health 2006, 21, 89–97. [Google Scholar] [CrossRef]
  11. Kivimäki, M.; Virtanen, M.; Elovainio, M.; Kouvonen, A.; Väänänen, A.; Vahtera, J. Work stress in the etiology of coronary heart disease—A meta-analysis. Scand. J. Work. Environ. Health 2006, 32, 431–442. [Google Scholar] [CrossRef]
  12. Melamed, S.; Shirom, A.; Toker, S.; Berliner, S.; Shapira, I. Burnout and risk of cardiovascular disease: Evidence, possible causal paths, and promising research directions. Psychol. Bull. 2006, 132, 327. [Google Scholar] [CrossRef]
  13. Picard, R.W.; Vyzas, E.; Healey, J. Toward machine emotional intelligence: Analysis of affective physiological state. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 10, 1175–1191. [Google Scholar] [CrossRef]
  14. Dey, A.K. Understanding and using context. Pers. Ubiquitous Comput. 2001, 5, 4–7. [Google Scholar] [CrossRef]
  15. Cohen, S.; Kamarck, T.; Mermelstein, R. A global measure of perceived stress. J. Health Soc. Behav. 1983, 24, 385–396. [Google Scholar] [CrossRef] [PubMed]
  16. Cohen, S.; Kamarck, T.; Mermelstein, R. Perceived stress scale. Measuring Stress: A Guide for Health and Social Scientists; Mind Garden, Inc.: Menlo Park, CA, USA, 1994; pp. 235–283. [Google Scholar]
  17. Brantley, P.J.; Waggoner, C.D.; Jones, G.N.; Rappaport, N.B. A daily stress inventory: Development, reliability, and validity. J. Behav. Med. 1987, 10, 61–73. [Google Scholar] [CrossRef] [PubMed]
  18. Hernandez, J.; McDuff, D.; Infante, C.; Maes, P.; Quigley, K.; Picard, R. Wearable ESM: Differences in the experience sampling method across wearable devices. In Proceedings of the 18th International Conference on Human-Computer Interaction with Mobile Devices and Services, Florence, Italy, 6–9 September 2016; pp. 195–205. [Google Scholar]
  19. Alberdi, A.; Aztiria, A.; Basarab, A. Towards an automatic early stress recognition system for office environments based on multimodal measurements: A review. J. Biomed. Inform. 2016, 59, 49–75. [Google Scholar] [CrossRef] [PubMed]
  20. Gjoreski, M.; Luštrek, M.; Gams, M.; Gjoreski, H. Monitoring stress with a wrist device using context. J. Biomed. Inform. 2017, 73, 159–170. [Google Scholar] [CrossRef] [PubMed]
  21. Lyu, Y.; Luo, X.; Zhou, J.; Yu, C.; Miao, C.; Wang, T.; Shi, Y.; Kameyama, K.I. Measuring photoplethysmogram-based stress-induced vascular response index to assess cognitive load and stress. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, Seoul, Korea, 18–23 April 2015; pp. 857–866. [Google Scholar]
  22. Levine, J.A.; Pavlidis, I.T.; MacBride, L.; Zhu, Z.; Tsiamyrtzis, P. Description and clinical studies of a device for the instantaneous detection of office-place stress. Work 2009, 34, 359–364. [Google Scholar]
  23. Puri, C.; Olson, L.; Pavlidis, I.; Levine, J.; Starren, J. StressCam: Non-contact measurement of users’ emotional states through thermal imaging. In Proceedings of the Extended Abstracts on Human Factors in Computing Systems (CHI’05), Portland, OR, USA, 2–7 April 2005; pp. 1725–1728. [Google Scholar]
  24. Shastri, D.; Papadakis, M.; Tsiamyrtzis, P.; Bass, B.; Pavlidis, I. Perinasal imaging of physiological stress and its affective potential. IEEE Trans. Affect. Comput. 2012, 3, 366–378. [Google Scholar] [CrossRef]
  25. Pavlidis, I.; Tsiamyrtzis, P.; Shastri, D.; Wesley, A.; Zhou, Y.; Lindner, P.; Buddharaju, P.; Joseph, R.; Mandapati, A.; Dunkin, B.; et al. Fast by nature-how stress patterns define human experience and performance in dexterous tasks. Sci. Rep. 2012, 2, 305. [Google Scholar] [CrossRef]
  26. Pavlidis, I.; Dcosta, M.; Taamneh, S.; Manser, M.; Ferris, T.; Wunderlich, R.; Akleman, E.; Tsiamyrtzis, P. Dissecting driver behaviors under cognitive, emotional, sensorimotor, and mixed stressors. Sci. Rep. 2016, 6, 25651. [Google Scholar] [CrossRef]
  27. Lazarus, R.S.; Folkman, S. Stress, Appraisal, and Coping; Springer Publishing Company: New York, NY, USA, 1984. [Google Scholar]
  28. Quick, J.C.; Spielberger, C.D. Walter Bradford cannon: Pioneer of stress research. Int. J. Stress Manag. 1994, 1, 141–143. [Google Scholar] [CrossRef]
  29. Lazarus, R.S. From psychological stress to the emotions: A history of changing outlooks. Annu. Rev. Psychol. 1993, 44, 1–22. [Google Scholar] [CrossRef]
  30. Cohen, R.A. Yerkes–Dodson Law. In Encyclopedia of Clinical Neuropsychology; Springer: Berlin, Germany, 2011; pp. 2737–2738. [Google Scholar]
  31. Kirschbaum, C.; Hellhammer, D.H. Salivary cortisol in psychobiological research: An overview. Neuropsychobiology 1989, 22, 150–169. [Google Scholar] [CrossRef] [PubMed]
  32. Greene, S.; Thapliyal, H.; Caban-Holt, A. A survey of affective computing for stress detection: Evaluating technologies in stress detection for better health. IEEE Consum. Electron. Mag. 2016, 5, 44–56. [Google Scholar] [CrossRef]
  33. Poh, M.Z.; McDuff, D.J.; Picard, R.W. Advancements in noncontact, multiparameter physiological measurements using a webcam. IEEE Trans. Biomed. Eng. 2011, 58, 7–11. [Google Scholar] [CrossRef]
  34. McDuff, D.J.; Hernandez, J.; Gontarek, S.; Picard, R. Cogcam: Contact-free measurement of cognitive stress during computer tasks with a digital camera. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, San Jose, CA, USA, 7–12 May 2016; pp. 4000–4004. [Google Scholar]
  35. Lu, H.; Frauendorfer, D.; Rabbi, M.; Mast, M.S.; Chittaranjan, G.T.; Campbell, A.T.; Gatica-Perez, D.; Choudhury, T. Stresssense: Detecting stress in unconstrained acoustic environments using smartphones. In Proceedings of the 2012 ACM Conference on Ubiquitous Computing, Pittsburgh, PA, USA, 5–8 September 2012; pp. 351–360. [Google Scholar]
  36. Aigrain, J.; Dubuisson, S.; Detyniecki, M.; Chetouani, M. Person-specific behavioural features for automatic stress detection. In Proceedings of the 2015 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG), Ljubljana, Slovenia, 4–8 May 2015; Volume 3, pp. 1–6. [Google Scholar]
  37. Bogomolov, A.; Lepri, B.; Ferron, M.; Pianesi, F.; Pentland, A.S. Daily Stress Recognition from Mobile Phone Data, Weather Conditions and Individual Traits. In Proceedings of the 22nd ACM International Conference on Multimedia (MM ’14), Orlando, FL, USA, 3–7 November 2014; ACM: New York, NY, USA, 2014; pp. 477–486. [Google Scholar] [CrossRef] [Green Version]
  38. Epp, C.; Lippold, M.; Mandryk, R.L. Identifying Emotional States Using Keystroke Dynamics. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI ’11), Vancouver, BC, Canada, 7–12 May 2011; ACM: New York, NY, USA, 2011; pp. 715–724. [Google Scholar] [CrossRef]
  39. Hernandez, J.; Paredes, P.; Roseway, A.; Czerwinski, M. Under pressure: Sensing stress of computer users. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Toronto, ON, Canada, 26 April–1 May 2014; pp. 51–60. [Google Scholar]
  40. McDuff, D.; Karlson, A.; Kapoor, A.; Roseway, A.; Czerwinski, M. AffectAura: An intelligent system for emotional memory. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Austin, TX, USA, 5–10 May 2012; pp. 849–858. [Google Scholar]
  41. Sano, A.; Picard, R.W. Stress recognition using wearable sensors and mobile phones. In Proceedings of the 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction, Geneva, Switzerland, 2–5 September 2013; pp. 671–676. [Google Scholar]
  42. Wang, E.J.; Zhu, J.; Jain, M.; Lee, T.J.; Saba, E.; Nachman, L.; Patel, S.N. Seismo: Blood pressure monitoring using built-in smartphone accelerometer and camera. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, Montreal, QC, Canada, 21–26 April 2018; p. 425. [Google Scholar]
  43. Hernandez, J.; McDuff, D.J.; Picard, R.W. Biophone: Physiology monitoring from peripheral smartphone motions. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; pp. 7180–7183. [Google Scholar]
  44. Hernandez, J.; McDuff, D.; Picard, R.W. Biowatch: Estimation of heart and breathing rates from wrist motions. In Proceedings of the 2015 9th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth), Istanbul, Turkey, 20–23 May 2015; pp. 169–176. [Google Scholar]
  45. Hernandez Rivera, J. Towards Wearable Stress Measurement. Ph.D. Thesis, Massachusetts Institute of Technology, Cambridge, MA, USA, 2015. [Google Scholar]
  46. Kocielnik, R.; Sidorova, N.; Maggi, F.M.; Ouwerkerk, M.; Westerink, J.H. Smart technologies for long-term stress monitoring at work. In Proceedings of the 26th IEEE International Symposium on Computer-Based Medical Systems, Porto, Portugal, 20–22 June 2013; pp. 53–58. [Google Scholar]
  47. Koldijk, S.; Sappelli, M.; Neerincx, M.; Kraaij, W. Unobtrusive monitoring of knowledge workers for stress self-regulation. In Proceedings of the International Conference on User Modeling, Adaptation, and Personalization, Rome, Italy, 10–14 June 2013; pp. 335–337. [Google Scholar]
  48. Koldijk, S.; Sappelli, M.; Verberne, S.; Neerincx, M.A.; Kraaij, W. The SWELL knowledge work dataset for stress and user modeling research. In Proceedings of the 16th international conference on multimodal interaction, Istanbul, Turkey, 12–16 November 2014; pp. 291–298. [Google Scholar]
  49. Muaremi, A.; Arnrich, B.; Tröster, G. Towards measuring stress with smartphones and wearable devices during workday and sleep. BioNanoScience 2013, 3, 172–183. [Google Scholar] [CrossRef] [PubMed]
  50. Nakashima, Y.; Kim, J.; Flutura, S.; Seiderer, A.; André, E. Stress recognition in daily work. In Proceedings of the International Symposium on Pervasive Computing Paradigms for Mental Health, Milan, Italy, 24–25 September 2015; pp. 23–33. [Google Scholar]
  51. Setz, C.; Arnrich, B.; Schumm, J.; La Marca, R.; Tröster, G.; Ehlert, U. Discriminating stress from cognitive load using a wearable EDA device. IEEE Trans. Inf. Technol. Biomed. 2009, 14, 410–417. [Google Scholar] [CrossRef] [PubMed]
  52. Smets, E.; Casale, P.; Großekathöfer, U.; Lamichhane, B.; De Raedt, W.; Bogaerts, K.; Van Diest, I.; Van Hoof, C. Comparison of machine learning techniques for psychophysiological stress detection. In Proceedings of the International Symposium on Pervasive Computing Paradigms for Mental Health, Milan, Italy, 24–25 September 2015; pp. 13–22. [Google Scholar]
  53. Sriramprakash, S.; Prasanna, V.D.; Murthy, O.R. Stress detection in working people. Procedia Comput. Sci. 2017, 115, 359–366. [Google Scholar] [CrossRef]
  54. Wijsman, J.; Grundlehner, B.; Liu, H.; Penders, J.; Hermens, H. Wearable physiological sensors reflect mental stress state in office-like situations. In Proceedings of the 2013 Humaine Association Conference on Affective Computing and Intelligent Interaction, Geneva, Switzerland, 2–5 September 2013; pp. 600–605. [Google Scholar]
  55. Zhai, J.; Barreto, A. Stress detection in computer users based on digital signal processing of noninvasive physiological variables. In Proceedings of the 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, New York, NY, USA, 30 August–3 September 2006; pp. 1355–1358. [Google Scholar]
  56. Zhai, J.; Barreto, A. Stress Recognition Using Non-invasive Technology. In Proceedings of the Nineteenth International Florida Artificial Intelligence Research Society Conference, Melbourne Beach, FL, USA, 11–13 May 2006; pp. 395–401. [Google Scholar]
  57. Barreto, A.; Zhai, J.; Adjouadi, M. Non-intrusive physiological monitoring for automated stress detection in human-computer interaction. In Proceedings of the International Workshop on Human-Computer Interaction, Rio de Janeiro, Brazil, 20 October 2007; pp. 29–38. [Google Scholar]
  58. Zhou, Y.; Tsiamyrtzis, P.; Pavlidis, I.T. Tissue tracking in thermo-physiological imagery through spatio-temporal smoothing. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, London, UK, 20–24 September 2009; pp. 1092–1099. [Google Scholar]
  59. Vizer, L.M.; Zhou, L.; Sears, A. Automated stress detection using keystroke and linguistic features: An exploratory study. Int. J. Hum.-Comput. Stud. 2009, 67, 870–886. [Google Scholar] [CrossRef]
  60. Ramos, J.; Hong, J.H.; Dey, A.K. Stress Recognition-A Step Outside the Lab. In Proceedings of the Conference on Physiological Computing Systems (PhyCS 2014), Lisbon, Portugal, 7–9 January 2014; pp. 107–118. [Google Scholar]
  61. Mark, G. Multitasking in the digital age. Synth. Lect. Hum.-Centered Inform. 2015, 8, 1–113. [Google Scholar] [CrossRef]
  62. Kirschbaum, C.; Pirke, K.M.; Hellhammer, D.H. The ‘Trier Social Stress Test’—A tool for investigating psychobiological stress responses in a laboratory setting. Neuropsychobiology 1993, 28, 76–81. [Google Scholar] [CrossRef]
  63. Conrad, B.; Schönle, P. Speech and respiration. Arch. Psychiatr. Nervenkrankh. 1979, 226, 251–268. [Google Scholar] [CrossRef]
  64. Tsiamyrtzis, P.; Dcosta, M.; Shastri, D.; Prasad, E.; Pavlidis, I. Delineating the operational envelope of mobile and conventional EDA sensing on key body locations. In Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems, San Jose, CA, USA, 7–12 May 2016; pp. 5665–5674. [Google Scholar]
  65. Zangróniz, R.; Martínez-Rodrigo, A.; Pastor, J.M.; López, M.T.; Fernández-Caballero, A. Electrodermal activity sensor for classification of calm/distress condition. Sensors 2017, 17, 2324. [Google Scholar] [CrossRef]
  66. Picard, R.W.; Fedor, S.; Ayzenberg, Y. Multiple arousal theory and daily-life electrodermal activity asymmetry. Emot. Rev. 2016, 8, 62–75. [Google Scholar] [CrossRef]
  67. Alamudun, F.; Choi, J.; Gutierrez-Osuna, R.; Khan, H.; Ahmed, B. Removal of subject-dependent and activity-dependent variation in physiological measures of stress. In Proceedings of the 2012 6th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth) and Workshops, San Diego, CA, USA, 21–24 May 2012; pp. 115–122. [Google Scholar]
  68. Mundy-Castle, A.; McKiever, B. The psychophysiological significance of the galvanic skin response. J. Exp. Psychol. 1953, 46, 15. [Google Scholar] [CrossRef] [PubMed]
  69. Mozos, O.M.; Sandulescu, V.; Andrews, S.; Ellis, D.; Bellotto, N.; Dobrescu, R.; Ferrandez, J.M. Stress detection using wearable physiological and sociometric sensors. Int. J. Neural Syst. 2017, 27, 1650041. [Google Scholar] [CrossRef] [PubMed]
  70. Kushki, A.; Fairley, J.; Merja, S.; King, G.; Chau, T. Comparison of blood volume pulse and skin conductance responses to mental and affective stimuli at different anatomical sites. Physiol. Meas. 2011, 32, 1529. [Google Scholar] [CrossRef] [PubMed]
  71. Betti, S.; Lova, R.M.; Rovini, E.; Acerbi, G.; Santarelli, L.; Cabiati, M.; Del Ry, S.; Cavallo, F. Evaluation of an integrated system of wearable physiological sensors for stress monitoring in working environments by using biological markers. IEEE Trans. Biomed. Eng. 2017, 65, 1748–1758. [Google Scholar] [PubMed]
  72. Zangróniz, R.; Martínez-Rodrigo, A.; López, M.T.; Pastor, J.M.; Fernández-Caballero, A. Estimation of mental distress from photoplethysmography. Appl. Sci. 2018, 8, 69. [Google Scholar] [CrossRef]
Figure 1. Experiment phases (CWT: Stroop Color-Word test).
Figure 1. Experiment phases (CWT: Stroop Color-Word test).
Sensors 19 03766 g001
Figure 2. EDA signal during the five sessions for two participants. The x-axis is cut at 400 s, thus only showing the first 400 s of the DT.
Figure 2. EDA signal during the five sessions for two participants. The x-axis is cut at 400 s, thus only showing the first 400 s of the DT.
Sensors 19 03766 g002
Figure 3. Wrist.HR signal during the five sessions for two participants. The x-axis is cut at 400 s, thus only showing the first 400 s of the DT.
Figure 3. Wrist.HR signal during the five sessions for two participants. The x-axis is cut at 400 s, thus only showing the first 400 s of the DT.
Sensors 19 03766 g003
Figure 4. An example of a participant’s BR data showing degrading signal in the presentation session. The x-axis is cut at 400 s, thus only showing the first 400 s of the DT.
Figure 4. An example of a participant’s BR data showing degrading signal in the presentation session. The x-axis is cut at 400 s, thus only showing the first 400 s of the DT.
Sensors 19 03766 g004
Figure 5. An example of a participant’s chest.HR data where increased stress is captured during stressful tasks.
Figure 5. An example of a participant’s chest.HR data where increased stress is captured during stressful tasks.
Sensors 19 03766 g005
Figure 6. An example of a participant’s data with overlapping high frequency responses in the baseline session, likely due to sensor friction and detachment from the skin.
Figure 6. An example of a participant’s data with overlapping high frequency responses in the baseline session, likely due to sensor friction and detachment from the skin.
Sensors 19 03766 g006
Figure 7. Example of a participant with PP signal that captures increased stress in stressful sessions. This participant took the color-word test and received emails in batches in DT (monotasking). This example also shows instances of missing data during the presentation session.
Figure 7. Example of a participant with PP signal that captures increased stress in stressful sessions. This participant took the color-word test and received emails in batches in DT (monotasking). This example also shows instances of missing data during the presentation session.
Sensors 19 03766 g007
Figure 8. Example of a participant with the PP signal that captures increased stress in stressful sessions. This participant watched the relaxing video and received emails continually in DT (multitasking).
Figure 8. Example of a participant with the PP signal that captures increased stress in stressful sessions. This participant watched the relaxing video and received emails continually in DT (multitasking).
Sensors 19 03766 g008
Figure 9. Boxplot of the ratios of missing PP data for all participants per session.
Figure 9. Boxplot of the ratios of missing PP data for all participants per session.
Sensors 19 03766 g009
Table 1. Sensors and Signals of the reviewed studies.
Table 1. Sensors and Signals of the reviewed studies.
PublicationSensor: Signal
[20]Wrist sensor: PPG, EDA, ST
[45]Chest sensors: HR, HRV, BR; Wrist sensors: EDA, ST
[46]Wrist sensor: EDA, ST, acceleration
[47]Chest sensors: HR, HRV; Finger sensor: EDA; Cameras; Kinect 3D
[48]Chest sensors: HR, HRV; Finger sensor: EDA; Cameras; Kinect 3D
[22]Thermal imaging of the corrugator
[21]PPG: sVRI, blood pressure; ECG: HRV
[34]Digital camera: HR, BR and HRV
[49]Smartphones: audio, physical activity, social interaction; Chest belts: HRV
[50]Pressure sensor; eye-tracker; fingertip sensor: EDA, BVP, HR
[51]Hand sensor: EDA
[52]Necklace sensor: ECG; Fingertip sensor: EDA and ST; Chest sensor: BR.
[53]Chest sensors: HR, HRV; Finger sensor: EDA
[54]Chest belt: ECG and respiration; Hand sensor: EDA; Shoulder electrodes: sEMG
[55,56,57]Hand sensor: BVP, EDA, ST; Eye-tracker: PD.
This workWristband: PPG (HR) and EDA; chest-band: ECG (HR), BR; Thermal camera: PP
Abbreviations: PPG: Photoplethysmogram, EDA: Electrodermal Activity, ST: Skin Temperature, HR: Heart-Rate, HRV: Heart-Rate Variability, BR: Breathing Rate, sVRI: Stress-Induced Vascular Response Index, ECG: Electrocardiogram, BVP: Blood Volume Pulse, sEMG: Surface Electromyogram, PD: Pupil Diameter, PP: Perinasal Perspiration.
Table 2. Computer tasks/Stressors of the reviewed studies.
Table 2. Computer tasks/Stressors of the reviewed studies.
PublicationComputer Task/Stressor
[20]MIST
[45]Unconstrained work environment
[46]Unconstrained work environment
[47]Writing reports with email interruptions and time pressure
[48]Writing reports with email interruptions and time pressure
[22]CWT and mental arithmetic
[21]Arithmetic problems
[34]Cognitive tasks: ball control task and BCST
[49]Unconstrained environment—in and outside of work
[50]CWT and information pick up task
[51]MIST
[52]CWT; talking about stressful experiences; math test
[53]Writing reports with email interruptions and time pressure
[54]Problem solving, puzzle, and memory task, done under time pressure, social pressure, and distracting noise
[55,56,57]CWT
This workCWT, relaxing video, multitasking, monotasking, essay writing, online presentation
Abbreviations: MIST: The Montreal Imaging Stress Task (mental arithmetic under time and evaluation pressure), CWT: Stroop Color-Word test, BCST: The Berg Card Sorting Task.
Table 3. Summary of reviewed studies.
Table 3. Summary of reviewed studies.
PublicationDependent/Output Variable# SubjectsDuration of measurementsControlled
[20]STAI-YLab: 21, Field: 5Total: 1564 min (lab), 1327 h (field)Partially
[45]Self-report155 daysNo
[46]EDA level104 weeksNo
[47]Self-report253 hYes
[48]Self-report253 hYes
[22]Difference from baseline1112 minYes
[21]Physiological measures4050 minYes
[34]Stress condition1010 minYes
[49]Self-report354 monthsNo
[50]Stress condition1021 minYes
[51]Stress condition334 hYes
[52]Stress condition2020 minYes
[53]Stress condition253 hYes
[54]Stress condition3040 minYes
[55,56,57]Stress condition3210 minYes
This workDifference from baseline6190 minYes
Abbreviations: STAI: State-Trait Anxiety Inventory. Controlled: Whether data is collected in a controlled lab experiment.
Table 4. Results for the essay writing session. Delta: the difference between the session and the Resting Baseline.
Table 4. Results for the essay writing session. Delta: the difference between the session and the Resting Baseline.
SignalMean Deltastdt (Delta ≠ 0)p (Delta ≠ 0)% of Delta  0
PP0.0005390.0007145.899090
chest.HR2.6414.9773.7890.00178
BR3.4213.5427.296086
EDA−0.0060.597−0.0860.93269
wrist.HR−2.54613.22−1.1710.24951
Table 5. Results for the CWT session. Delta: the difference between the session and the Resting Baseline.
Table 5. Results for the CWT session. Delta: the difference between the session and the Resting Baseline.
SignalMean Deltastdt (Delta ≠ 0)p (Delta ≠ 0)% of Delta  0
PP0.0004320.0007083.4010.00271
HR2.1924015.5361.980.05968
BR2.7242933.1274.6920.00179
EDA0.0980850.3791.0980.28783
wrist.HR0.3612256.090.2520.80450
Table 6. Results for the video session. Delta: the difference between the session and the Resting Baseline.
Table 6. Results for the video session. Delta: the difference between the session and the Resting Baseline.
SignalMean Deltastdt (Delta ≠ 0)p (Delta ≠ 0)% of Delta  0
PP0.0004040.0009372.3220.02879
HR−2.7035.304−2.5480.01824
BR0.5572.9910.9850.33354
EDA−0.140.534−1.1150.28172
wrist.HR−1.3899.002−0.6730.5132
Table 7. Results for the monotasking session. Delta: the difference between the session and the Resting Baseline.
Table 7. Results for the monotasking session. Delta: the difference between the session and the Resting Baseline.
SignalMean Deltastdt (Delta ≠ 0)p (Delta ≠ 0)% of Delta 0
PP0.0007770.0008824.6570.00179
chest.HR−0.9944.567−1.110.27835
BR1.192.32.5340.01979
EDA−0.0170.133−0.5450.59367
wrist.HR−3.0879.877−1.3260.20233
Table 8. Results for the multitasking session. Delta: the difference between the session and the Resting Baseline.
Table 8. Results for the multitasking session. Delta: the difference between the session and the Resting Baseline.
SignalMean Deltastdt (Delta ≠ 0)p (Delta ≠ 0)% of Delta 0
PP0.0010.0015.603085
chest.HR−0.557.934−0.3470.73252
BR2.3574.4482.9020.00767
EDA−0.6062.62−0.9820.3478
wrist.HR−2.511.756−0.9270.36647
Table 9. Results for the presentation session. Delta: the difference between the session and the Resting Baseline.
Table 9. Results for the presentation session. Delta: the difference between the session and the Resting Baseline.
SignalMean Deltastdt (Delta ≠ 0)p (Delta ≠ 0)% of Delta  0
PP0.0021460.00147611.167097
chest.HR9.83212.4875.34083
BR−0.574.226−10.32244
EDA−0.4842.329−1.2280.22874
wrist.HR1.47913.6040.6520.51853
Table 10. Mean and median of the ratio of missing PP data per session.
Table 10. Mean and median of the ratio of missing PP data per session.
SessionMeanMedian
Essay writing0.0240.000
CWT0.0270.007
Presentation0.1090.019
Resting Baseline0.0030.000
Monotasking0.0360.014
Multitasking0.0460.014
Calming Video0.0640.002

Share and Cite

MDPI and ACS Style

Akbar, F.; Mark, G.; Pavlidis, I.; Gutierrez-Osuna, R. An Empirical Study Comparing Unobtrusive Physiological Sensors for Stress Detection in Computer Work. Sensors 2019, 19, 3766. https://doi.org/10.3390/s19173766

AMA Style

Akbar F, Mark G, Pavlidis I, Gutierrez-Osuna R. An Empirical Study Comparing Unobtrusive Physiological Sensors for Stress Detection in Computer Work. Sensors. 2019; 19(17):3766. https://doi.org/10.3390/s19173766

Chicago/Turabian Style

Akbar, Fatema, Gloria Mark, Ioannis Pavlidis, and Ricardo Gutierrez-Osuna. 2019. "An Empirical Study Comparing Unobtrusive Physiological Sensors for Stress Detection in Computer Work" Sensors 19, no. 17: 3766. https://doi.org/10.3390/s19173766

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop