Next Article in Journal
Scan Aided Dental Arch Width Prediction via Internationally Recognized Formulas and Indices in a Sample of Kurdish Population/Iraq
Next Article in Special Issue
Analysis of the Effectiveness of Metaheuristic Methods on Bayesian Optimization in the Classification of Visual Field Defects
Previous Article in Journal
Elastography of the Liver in Wilson’s Disease
Previous Article in Special Issue
Improving the Trustworthiness of Interactive Visualization Tools for Healthcare Data through a Medical Fuzzy Expert System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-World Driver Stress Recognition and Diagnosis Based on Multimodal Deep Learning and Fuzzy EDAS Approaches

1
Department of Electronics, University of Peshawar, Peshawar 25120, Pakistan
2
Department of Computer Science, Iqra National University, Peshawar 25000, Pakistan
3
Department of Software Engineering, University of Malakand, Dir Lower, Chakdara 23050, Pakistan
4
Department of Computer Science, King Khalid University, Abha 61421, Saudi Arabia
5
Department of Mechanical Engineering, University of Engineering & Technology, Peshawar 25120, Pakistan
6
Department of Computer Science, HITEC University, Taxila 47080, Pakistan
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(11), 1897; https://doi.org/10.3390/diagnostics13111897
Submission received: 6 March 2023 / Revised: 4 April 2023 / Accepted: 13 May 2023 / Published: 29 May 2023
(This article belongs to the Special Issue Medical Data Processing and Analysis—2nd Edition)

Abstract

:
Mental stress is known as a prime factor in road crashes. The devastation of these crashes often results in damage to humans, vehicles, and infrastructure. Likewise, persistent mental stress could lead to the development of mental, cardiovascular, and abdominal disorders. Preceding research in this domain mostly focuses on feature engineering and conventional machine learning approaches. These approaches recognize different levels of stress based on handcrafted features extracted from various modalities including physiological, physical, and contextual data. Acquiring good quality features from these modalities using feature engineering is often a difficult job. Recent developments in the form of deep learning (DL) algorithms have relieved feature engineering by automatically extracting and learning resilient features. This paper proposes different CNN and CNN-LSTSM-based fusion models using physiological signals (SRAD dataset) and multimodal data (AffectiveROAD dataset) for the driver’s two and three stress levels. The fuzzy EDAS (evaluation based on distance from average solution) approach is used to evaluate the performance of the proposed models based on different classification metrics (accuracy, recall, precision, F-score, and specificity). Fuzzy EDAS performance estimation shows that the proposed CNN and hybrid CNN-LSTM models achieved the first ranks based on the fusion of BH, E4-Left (E4-L), and E4-Right (E4-R). Results showed the significance of multimodal data for designing an accurate and trustworthy stress recognition diagnosing model for real-world driving conditions. The proposed model can also be used for the diagnosis of the stress level of a subject during other daily life activities.

1. Introduction

Successful driving activities always require both mental and physical skills [1,2,3]. Acute stress reduces the driver’s ability to fix hazardous situations, which causes significant damage to humans and vehicles every year [4,5,6,7,8]. Dangerous driving situations are triggered due to human errors, individual factors, and ambiance conditions [9]. According to the National Motor Vehicle Crash Causation Survey (NMVCCS) in the United States (US), human errors caused 94% of crashes alone, while vehicle defects, ambiance conditions, and other factors collectively caused 6% of crashes during 2005–2007 [10]. Human errors are linked to the driver’s perceptual conditions, so a complete understanding of these conditions is crucial for preventing traffic accidents.
To detect and diagnose drivers’ different stress levels, physiological, physical, and contextual information are widely utilized [11]. Moreover, different traditional machine learning models based on handcrafted feature extraction methods are utilized for the classification of stress. Extracting the best features using these approaches is always a challenging task, as the quality of extracted features has a significant effect on the classification performance [12]. These approaches are laborious, ad hoc, less robust to noise, and need thorough skill [13]. To come through these challenges, deep learning models have been utilized to automatically produce complex nonlinear features reliably [14,15,16]. In addition to automatic feature extraction from raw data, these models offer noise robustness and better classification accuracy [17,18,19]. Different deep learning algorithms are used in recent research, e.g., CNN, RNN, DNN, and LSTM.
The models proposed in the current work are based on 1D CNN and hybrid 1D CNN-LSTM networks. The proposed models are separately trained using multiple physiological signals (SRAD) and multimodal data (AffectiveROAD) including physiological signals and other information about the vehicle, driver, and ambiance. Multimodal fusion of data based on deep learning approaches can be used to develop a precise driver stress level recognition model with improved performance and reliability.
Contributions of this research study include: (1) proposing 1D CNN and hybrid 1D CNN-LSTM-based real-world driver stress level recognition models using fused physiological signals (SRAD dataset) and fused multimodal data (AffectiveROAD dataset) and (2) ranking the assessment of the proposed models for the two and three levels of stress based on the fuzzy EDAS approach.
The organization of this research article is given below. Analysis of the existing stress recognition models is presented in Section 2. The proposed methodology is elaborated in Section 3 in terms of datasets, data pre-processing, architectures of the proposed CNN and hybrid CNN-LSTM models, and the fuzzy EDAS approach. Performance evaluation of the proposed models is conducted in Section 4. A fuzzy EDAS-based rank estimation of the proposed models for the driver’s two and three levels of stress is also presented in this section. Section 5 gives a detailed assessment of the proposed and existing stress recognition schemes. Finally, Section 6 concludes the paper and gives future directions to further explore this research area.

2. Related Work

This section provides a review of the existing work in the driver’s stress analysis domain and underscores the current contribution. Several driver stress level recognition schemes exist in the literature based on simulated and real-world driving environments. These schemes can be broadly categorized as conventional machine learning or deep learning models.
Several machine learning approaches have been proposed for real-world driver mental stress recognition based on different physiological signals. Dalmeida and Masala [20], Vargas-Lopez et al. [21], Khowaja et al. [8], Lopez-Martinez et al. [22], Haouij et al. [23], Chen et al. [4], Ghaderi et al. [24], Zhang et al. [25], and Healey and Picard [26] propose conventional machine models based on physiological signals obtained from the PhysioNet SRAD public database [27]. Unlike the previous studies, Rigas et al. [28] presented a real-world binary stress recognition model based on multimodal data, including physical and contextual data, in addition to physiological signals. On the other hand, Zontone et al. [29], Bianco et al. [30], Lee et al. [31], Lanatà et al. [32], and Gao et al. [33] proposed conventional machine learning models for driver stress recognition based on simulated driving situations. Lanatà et al. [32] and Lee et al. [31] presented driver stress recognition models based on multimodal data. Contrary to previous studies, Šalkevicius et al. [34], Rodríguez-Arce et al. [35], Can et al. [36], Al abdi et al. [37], Betti et al. [38], Siramprakas et al. [39], de Vries et al. [40], and Sun et al. [41] proposed stress recognition models during controlled, lab, semi-lab, and physical (such as sitting, standing, and walking) environments. Recent development in deep learning and machine learning models have shown good results in various applied domains that can be applied in driver stress detection [42,43].
All the mentioned studies are based on feature engineering techniques, and various conventional machine learning algorithms were employed to classify levels of stress. However, handcrafted features are less robust to noise and subjective changes, and need a considerable amount of time and hard work [8,13,19,34,35,44]. Moreover, capturing the features’ sequential nature is difficult due to the absence of explicit features and high dimensionality despite using complex feature selection methods. Likewise, the dependence of the model on past observations would make it impractical to process all the information due to the growing complexity. The feature-level multimodal fusion models proposed by Chen et al. [4], Healey and Picard [26], Haouij et al. [23], Lee et al. [31], Bianco et al. [30], Sun et al. [41], and Can et al. [36] mainly concentrate on pattern learning in individual signals instead of multiple simultaneous signals [18]. Thus, these models are inappropriate to obtain the nonlinear correlation across multiple signals appearing simultaneously. Various linear and non-linear methods employed in these conventional machine learning models have not been able to perform the vigorous investigation of such manifold time series signals [19].
To address the issues faced by conventional machine learning models, deep learning methods have been introduced. Deep learning models are developed based on signal preprocessing (noise filtering), designing a particular deep neural network based on the area of interest, network training, and model testing. Deep learning models learn and classify raw data using multilayer deep neural networks [45]. The last fully connected (FC) layers are utilized to obtain the final output. Contrary to feature engineering techniques used in conventional machine learning approaches, deep learning models automatically produce steady features [14,15]. Moreover, deep learning models are more robust to noise and achieve improved classification accuracy [19]. Different deep learning algorithms are used in recent research, e.g., the recurrent neural network (RNN), deep aeural network (DNN), LSTM, and CNN. Rastgoo et al. [11], Zhang et al. [46], Kanjo et al. [17], Lim and Yang [47], Yan et al. [48], Hajinoroozi et al. [49], and Lee et al. [50] presented different deep learning models to identify different driver states. Rastgoo et al. [11], Kanjo et al. [17], Lim and Yang [47], and Yan et al. [48] proposed deep learning models based on multimodal data. On the other hand, the models proposed by Hajinoroozi et al. [49] and Lee et al. [50] are based on physiological signals only. The stress recognition model proposed by Zhang et al. [46] is based on facial images only. Apart from driving scenarios, Masood and Alghamdi [51], Cho et al. [52], Seo et al. [53], Hwang et al. [54], and He et al. [55] proposed stress recognition models based on deep learning techniques and physiological signals in academic, workplace, and lab settings. Most of these studies including [46,49,50,52,53,54,55,56] are based on two levels of stress only. Moreover, the schemes presented by [46,50,52,55,56] are based on images. Likewise, the schemes proposed by [49,52,53,54,55,56] are either based on physiological signals or a single modality. On the other hand, the model proposed by [11] is based on multimodal data collected during simulated driving.
The models proposed in this study are based on the fusion of multimodal data collected during real-world driving (SRAD and AffectiveROAD datasets). Moreover, these models are based on 1D CNN and 1D CNN-LSTM networks to detect driver’s two (stressed and relaxed) and three levels (low, medium, and high). The fuzzy EDAS approach is also used to find the performance ranks of the proposed models based on different classification metrics.

3. Materials and Methods

The proposed unimodal and fusion models for real-world driver stress level recognition are based on physiological signals and deep learning approaches, such as CNN and hybrid CNN-LSTM. The proposed models are implemented in the latest MATLAB 2022a platform. The proposed stress recognition models are based on the fusion of ECG, HR, HGSR, FGSR, EMG, and RESP signals collected from the PhysioNet SRAD database, and breathing rate (BR), GSR, BVP, HR, TEMP, ACCEL, posture, and activity data are collected from AffectiveROAD database. Data input mechanisms used in this research are based on raw signals. These raw signals are preprocessed to obtain cleaned signals.

3.1. SRAD Dataset

The ECG, HR, GSR, EMG, and RESP signals analyzed in the current work belong to the SRAD PhysioNet public database [57]. Experiments were performed while driving a customized Volvo S70 series station wagon. Five different sensors were used to acquire physiological signals from the nine drivers during twenty-four drives. The sensors were connected to an embedded computer through an analog-to-digital converter (ADC). The ECG sensor was placed using a modified lead II configuration to decrease the motion artifacts. The EMG sensor was positioned on the shoulder near the trapezius muscle to record the emotional stress. Two GSR sensors were located on the driver’s sole and palm of the left foot and hand. Expansion of the chest cavity was used to measure the RESP signals through an elastic Hall effect sensor fastened around the diaphragm.
All drives comprise rest, highway, and city driving phases on a specific route 31 km in length in Boston, US. These rest, highway, and city driving phases are assumed to trigger low, medium, and high levels of stress, respectively. Initially, the drivers are informed about the travel plan and compliance with certain guidelines regarding the speed limits and tuning out the radio. To avoid rush hours, all drives were performed in the midmorning and afternoon. Two rest intervals of 15 min in the parking area were added at the start and end of each drive to collect the driver’s low-stress baseline. Due to stop-and-go traffic in the city area, drivers usually observe high-stress situations. After passing the toll booth, the city road then turns into the highway. Uninterrupted highway driving normally indicates medium-stress conditions. The trip completes after returning to the starting position using the same highway and city routes. The total length of all drives varies from 50 to 90 min, including two 15 min rest intervals.
The dataset contains information about 17 drives, but some drives have incomplete signals and markers. These incomplete drives are removed from the experiments. Figure 1, Figure 2, Figure 3, Figure 4 and Figure 5 separately show the ECG, HR, GSR, EMG, and RESP waveforms for the three levels of stress. The figures show that all five signals have distinct waveforms for the three levels of stress.

3.2. AffectiveROAD Database

Experiments were performed using wireless sensors networked together inside different cars to collect physiological signals and additional information about the vehicle, driver, and ambiance. The Zephyr Bio-harness (BH) chest strap was placed on the driver’s chest to collect HR, breathing rate (BR), posture, and activity information. Two Empatica E4-Left (E4-L) and E4-Right (E4-R) wearable devices were mounted on the driver’s left and right arms to capture GSR, BVP, inter-beat interval (IBI), HR, TEMP, and ACCEL data. The Intel Edison developer kit-based environmental platform was placed in the car’s rear seat for collecting luminosity, temperature, pressure, and humidity information. A sound meter and microphone were used to obtain sound amplitude and audio signal. Two cameras were placed on the windshield of the car to record inside and outside events. A real-time continuous subjective metric was prepared by an experimenter during each drive to monitor the driver’s stress level. The stress metric along with two video recordings were then used by the drivers to correct and validate the experimenter’s ratings.
All drives comprise rest, highway, and city driving phases on a fixed route 31 km in length in the Grand Tunis area. Fourteen driving experiments were performed by 10 experienced drivers with valid driver’s licenses. Each drive included two 15 min rest periods at the start and end of the session. The whole experiment normally took about 86 min to travel through the zone, city1, highway, and city2, and then travel back in the opposite direction to reach the starting point. The rest, highway, and city drives were supposed to yield low, medium, and high levels of stress, respectively.

3.3. Pre-Processing

Physiological data are normally derived from the human body in the form of low-amplitude signals with different frequency ranges. These signals are mostly polluted by different noises and artifacts. To model the driver’s stress levels accurately, it is necessary to preprocess the ECG, HR, HGSR, FGSR, EMG, and RESP signals first.
ECG signals normally contain different unwanted components including baseline wander, powerline interference (PLI), and high-frequency EMG noise [58]. Moreover, the PLI adds 50–60 Hz noise components in ECG signals [59]. Likewise, high-frequency EMG noise components caused by muscle contractions contaminate the ECG signals [58]. HR signals are commonly derived from ECG signals, so they inherit some noise and artifacts from ECG signals. To remove the baseline wander and other artifacts form ECG signals, a band-pass Butterworth filter (5–15 Hz) was used to eliminate the baseline wander. Similarly, a finite impulse response (FIR), Notch filter (59–61 Hz), and FIR band-pass filter (1.5–150 Hz) were used for noise removal. The min–max normalization approach is then utilized to remove the subject-specific baseline and motion artifacts.
A GSR signal is an effective stress measure that is comparatively less susceptible to noise [60]. Yet, the authors of [61] used a low-pass filter (4 Hz) and a Gaussian filter for denoising the GSR signal. These filters are used in this study too to obtain cleaned GSR signals. The signals are also normalized to the maximum value.
The EMG signal is contaminated by several unwanted signals including motion artifacts, PLI, capacitive effects, and ECG artifact signals. In this work, a band-pass Butterworth filter (0.5–500 Hz) is used to remove the low- and high-frequency noises in the EMG signals. Likewise, PLI is eliminated using a 60 Hz Notch filter. The min–max normalization is performed to remove the subject-specific baseline and motion artifacts. EMG signals in the SRAD dataset were initially collected at a lower sampling frequency of 495 Hz. Although, the EMG signal contains information up to 450 Hz. As per the Nyquist theorem, at least a 900 Hz sampling frequency is required for the EMG signals.
The RESP signal is normally polluted by different undesirable signals including baseline wander, PLI, and motion artifacts. To remove high-frequency noise and baseline signal from the RESP signal, we applied Butterworth high-pass (0.05 Hz) and low-pass (0.70 Hz) filters, respectively.

3.4. 1D CNN Models

CNN-based models were originally developed to learn the internal representation of 2D images and then classify them into certain output classes. The same approach can be utilized for automatic feature learning and classification of time series sequenced data [62]. A 1D CNN uses several filters to perform 1D convolution (Conv1D) operations for constructing feature maps from such data. These networks can better match the 1D characteristic of different physiological signals. Increasing the convolutional layers can help CNN models to gradually extract unique and vigorous higher-level features. The 1D CNN models used in this research are based on the signal fusion of the SRAD and AffectiveROAD datasets for both two-stress and three-stress classes. Thus, all signals in the SRAD dataset are combinedly trained using the 1D CNN model. The AffectiveROAD dataset consists of multimodal data collected using BH, E4-L, and E4-R devices. So, different 1D CNN models are trained using the BH, E4-L, E4-R, E4-(Left+Right) (E4-(L+R)), and BH+E4-(L+R) datasets. A sliding window approach is used to convert each cleaned signal into equal size segments. These segments are then fed to a 1D CNN as new training data. The CNN-based driver stress recognition performs both feature learning and classification tasks.
A 1D CNN architecture is defined using multiple Conv1D blocks each containing convolution, ReLU, and layer normalization (LN) layers. One-dimensional CNN architectures based on SRAD, E4-L, E4-R, E4-(L+R), BH, and BH+E4-(L+R) datasets are shown in Table 1. The convolution layer utilizes trainable filters (kernels) to convolve the low-level features of each segment or the previous layer’s output to produce a feature map. The number of filters in each convolutional block is set differently depending on the dataset. Causal padding is used in all convolutional layers to produce outputs with the same length. It pads the layer’s input with zeros to predict the values of early time steps in the frame. The convolutional layer is followed by the ReLU layer, which is based on a piecewise linear function. This function returns output for positive inputs and is zero otherwise, thus alleviating the vanishing gradient problem [63]. Moreover, the function adds nonlinearity to the model to learn complex patterns in the data. A GAP layer is added after the four convolutional blocks to produce a single vector output. This layer finds the average output of each feature map generated by the convolutional layers and provides a substitute for the flattening block. The last three layers including FC, softmax, and classification layers perform the classification task. The vector output of the GAP layer is fed to the FC layer, which is also known as the hidden layer. The FC layer is used to map the output classes to a vector of probabilities. The output of the FC layer is utilized by the softmax layer to perform the final classification decision by allocating probabilities to low, medium, and high classes of stress. The final classification layer uses a cross-entropy loss function to evaluate the performance of the classification model. An increase in cross-entropy loss reflects the divergence of the predicted probability from the actual label and vice versa. The classification layer assumes the number of classes from the FC and softmax layers.

Network’s Training

Before starting the training process, several parameters need to be settled. These pa-rameters include the training algorithm, mini-batch size, validation frequency, initial learning rate, and maximum epochs. Parameter settings for different 1D CNN models are shown in Table 1.
A training algorithm is used to reduce the loss function of a learning model iterative-ly based on a training dataset. Adaptive moment estimation (Adam) is used as a training algorithm. It combines the benefits of RMSProp and AdaGrad by calculating the individ-ual adaptive learning rates based on the parameters estimated for the first and second moments of gradients. The mini-batch represents a subset of segments used in a single training iteration. Min-batch size is set to a small value to ensure the uniform distribution and utilization of the full dataset during a single epoch. The validation frequency repre-sents the training iterations between evaluations of validation metrics, while training iter-ation is a single step performed by an optimization algorithm to reduce the loss function for a mini-batch. The network’s validation frequency is set to 10. The epoch represents the maximum iterations completed by the optimization algorithm to reduce the loss function for the entire dataset. All datasets are divided into 80% for training data and 20% for vali-dation data.

3.5. Hybrid 1D CNN-LSTM Models

The LSTM is a particular type of RNN developed by Hochreiter and Schmidhuber [64]. It is useful to discover and remember long sequences of data efficiently. Generally, the LSTM is a chain of repeating cells of neural networks, such as a RNN, but both have different cell structures. The RNN’s cell consists of a single neural network based on tangent hyperbolic function, while LSTM’s cell has four interacting neural network layers based on sigmoid functions and pointwise multiplication operations. The LSTM has several cells connected to each other horizontally. Information can be added or removed from the cell state using four different gates. Each LSTM cell consists of an input gate, cell state gate, forget gate, and output gate. The forget gate is based on the sigmoid function, which determines which information needs to be forgotten from the cell state. The information is removed if the gate generates zero output and it is retained if the gate produces one output. The cell state gate determines the cell state based on the new information. First, the input gate based on the sigmoid function determines the values to be updated. Next, a vector is created for the new candidate values by the tangent hyperbolic activation function. The cell state is updated by combining the results of the two functions. To generate the output of a cell, the output gate first applies the sigmoid function to the part of a cell state. Next, a tangent hyperbolic function is applied to the cell state and the resulting value is multiplied by the output of the sigmoid function.
The hybrid CNN-LSTM model utilizes both 1D CNN and LSTM networks to classify sequenced data. In such a model, the CNN is used as a front end to extract features from physiological data followed by the LSTM layers to perform learning and classification tasks. The hybrid CNN-LSTM model has a similar architecture to the CNN model with additional LSTM cells after the FC layers. The architecture of the CNN model is already discussed in the previous section. The hybrid 1D CNN-LSTM architectures based on the SRAD, BH, E4-L, E4-R, E4-(L+R), and BH+E4-(L+R) datasets are shown in Table 2. Moreover, parameter settings for the proposed models are also shown in the same table.

3.6. Fuzzy EDAS Approach

The fuzzy EDAS approach is used to evaluate the performance of the proposed real-world driver stress level detection models based on different modalities. This approach performs the rank estimation of the proposed models in terms of accuracy, recall, precision, F-score, and specificity. Fuzzy EDAS is an eight-step process where each step performs some sort of calculations, which in turn is used by the coming steps, as elaborated below:
Step 1: First, the “solution of the average value ( ψ ) ” is calculated for all matrices, as shown mathematically in the equation below:
( ψ ) = [ ψ β ] 1 × δ
where:
ψ β = i = 1 x X α β x
The aggregate solution of Equations (1) and (2) can be found as the average value ( ψ ) against every criterion’s estimated quantity for each performance metric.
Step 2: The positive distances from the average ( P I ) of each signal for the driver’s each stress level is calculated using the following equation:
P I = [ P I α β ] δ × δ
The ( P I ) α β in Equation (3) is the positive distance of β th model from the average value for the α th parameter. It can be found using either of two ways. If β th criterion is more favorable, then it is calculated using the equation below:
( P I ) α β = M a x i m u m ( 0 , ( A V β X α β ) ) A V β
On the other hand, if the β th criterion is not favorable, it is calculated by the following equation:
( P I ) α β = M a x i m u m ( 0 , ( X α β A V β ) ) A V β
Step 3: The negative distances from the average ( N I ) of each signal for the driver’s stress level is calculated using the following equation:
( N I ) = [ ( N I ) α β ] δ × δ
The ( N I ) α β in Equation (6) is the negative distance of the β th model from the average value for the α th parameter. It can be found using either of two ways. If the β th criterion is more favorable, then it is calculated using the equation below:
( N I ) α β = M a x i m u m ( 0 , ( A V β X α β ) ) A V β
On the other hand, if the β th criterion is not favorable, it is calculated by the following equation:
( N I ) α β = M a x i m u m ( 0 , ( X α β A V β ) ) A V β
Step 4: The weighted sum of ( P I ) α β is calculated using the following equation:
( S P I ) α = β = 1 x y β ( P I ) α β
The aggregate ( P I ) is estimated for each signal evaluated using the proposed model for each stress level.
Step 5: The weighted sum of ( N I ) α β is calculated using the following equation:
( S N I ) α = β = 1 x y β ( N I ) α β
The aggregate ( N I ) is estimated for each signal evaluated using the proposed model for each stress level.
Step 6: The normalized values of ( S P I ) α and ( S N I ) α of each signal for the driver’s stress level are found using the following two equations:
N ( S P I ) α = ( S P I ) α m a x i m u m α ( ( S P I ) α )
N ( S N I ) α = 1 ( S N I ) α m a x i m u m α ( S N I ) α
Step 7: The appraisal score ( λ ) of each signal for the driver’s stress level is calculated using the equation given below:
λ α = 1 2 ( N ( S P I ) α N ( S P I ) α )
The appraisal score ( λ α ) lies are given as 0 λ α 1 .
Step 8: Each signal for the driver’s stress level is ranked according to the decreasing values of the appraisal score ( λ α ) . Thus, the signal with the lowest appraisal score ( λ α ) for a particular stress level has the highest performance among the other signals.

4. Results

The SRAD and AffectiveROAD datasets are randomly distributed into two groups, with 85% and 15% for training and validation, respectively. Results are acquired for the 1D CNN and 1D CNN-LSTM models trained using the SRAD, BH, E4-L, E4-R, and BH+E4-(L+R) datasets. A performance assessment of the proposed driver stress recognition models for the low, medium, and high classes of stress is carried out using different classification metrics. These performance metrics include accuracy (ACC), recall (RCL), precision (PRC), F-score (F1), and specificity (SPC).

4.1. Models’ Evaluation for the Two-Stress Class

Results of the proposed driver stress recognition models for the two-stress class are shown in Table 3. These results are based on the training data obtained from the SRAD and AffectiveROAD datasets for real-world driving. The training graphs of the proposed CNN models are shown in Figure 1, Figure 2, Figure 3, Figure 4, Figure 5 and Figure 6. Similarly, the training graphs of the proposed hybrid CNN-LSTM models are shown in Figure 7, Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12. Results show that the BH+E4-(L+R)-based CNN model outperformed other models based on the SRAD, Bio BH, E4-L, E4-R, and E4-(L+R) datasets by 2.9%, 6.5%, 9.1%, 7.3%, and 3.25%, respectively, with an overall validation accuracy of 95.6% for the two-stress class. The proposed BH+E4-(L+R)-based hybrid CNN-LSTM model outperformed other models based on SRAD, BH, E4-L, E4-R, and E4-(L+R) datasets by 4.79%, 1.1%, 7.76%, 5.94%, and 1.94%, respectively, with an overall validation accuracy of 96.59% for the two-stress class.
Confusion matrices of the proposed CNN and hybrid CNN-LSTM models are shown in Figure 13 and Figure 14. In Figure 13f, 214 relaxed instances are predicted correctly, while 5 relaxed instances are incorrectly predicted as stressed by the model. Thus, the total correct prediction for the relaxed class is 97.7%. Similarly, for the stressed class, 291 out of 309 instances are correctly predicted, which amounts to a total accuracy of 94.2% for the stressed class. In Figure 14f, 233 relaxed instances are predicted correctly, while 15 relaxed instances are incorrectly predicted as stressed by the model. Thus, the total correct prediction for the relaxed class is 94%. Similarly, for the stressed class, 277 out of 280 instances are correctly predicted, which amounts to a total accuracy of 98.9% for the stressed class.

4.2. Models’ Evaluation for the Three-Stress Class

Results of the proposed driver stress recognition models for the three-stress class are shown in Table 4. These results are based on the training data obtained for the SRAD and AffectiveROAD datasets for real-world driving. The training graphs of the proposed CNN models are shown in Figure 15, Figure 16, Figure 17, Figure 18, Figure 19 and Figure 20. Similarly, the training graphs of the proposed hybrid CNN-LSTM models are shown in Figure 21, Figure 22, Figure 23, Figure 24, Figure 25 and Figure 26. Results show that the proposed CNN model based on the BH+E4-(L+R) datasets outperform the other models based on the SRAD, BH, E4-L, E4-R, and E4-(L+R) datasets significantly by 6.16%, 6.76%, 9.16%, 8.87%, and 1.72%, respectively, with an overall validation accuracy of 85.66%. Similarly, the proposed hybrid CNN-LSTM model based on the BH+E4-(L+R) datasets outperform the other models based on the SRAD, BH, E4-L, E4-R, and E4-(L+R) datasets significantly by 2.15%, 0.15%, 11.22%, 5.89%, and 3.82%, respectively, with an overall validation accuracy of 87.95%.
Confusion matrices for the proposed CNN and hybrid CNN-LSTM models are shown in Figure 27 and Figure 28. In Figure 27f, 190 low instances are predicted correctly, while 3 and 6 low instances are incorrectly predicted as medium and high by the CNN model. So, the total correct prediction for the high-stress class is 95.5%. Likewise, for the medium- and high-stress classes, 34 out of 50 and 224 out of 274 instances were correctly predicted, which amounts to total accuracies of 68% and 81.8% for medium- and high-stress classes, respectively. Similarly, in Figure 28f, 214 low instances are predicted correctly, while 3 and 8 low instances are incorrectly predicted as medium and high by the CNN-LSTM model. Therefore, the total correct prediction for the low-stress class is 95.1%. Similarly, for the medium- and high-stress classes, 47 out of 61 and 199 out of 237 instances were correctly predicted, which amounts to total accuracies of 77% and 84% for the medium- and high-stress classes, respectively.

4.3. Rank-Based Performance Evaluation

The eight-step fuzzy EDAS procedure [65] defined in Section 3.6 is utilized here to evaluate the ranks of the SRAD, BH, E4-L, E4-R, E4-(L+R), and BH+E4-(L+R)-based CNN and hybrid CNN-LSTM models for the two-stress and three-stress classes. This procedure is separately followed for each the driver’s stress level. The classification metrics calculated in Table 3 and Table 4 are regarded as a criterion for the proposed CNN and hybrid CNN-LSTM driver stress level classification models for the two-stress and three-stress classes.

4.3.1. Rank Estimation of the CNN Models for Two Levels of Stress

A rank estimation of the CNN models for the two-stress class (relaxed state) is performed in a series of steps. The results of each step are shown in Table 5, Table 6, Table 7, Table 8, Table 9 and Table 10. The first step determines the cross-efficient values ( ψ β ) using Equations (1) and (2), as shown in Table 5. In the next two steps, the positive distance ( P I ) and negative distance ( N I ) are separately determined based by Equations (5) and (8), as given in Table 6 and Table 7. In the fourth and fifth steps, the weighted sum of ( P I ) and ( N I ) are separately calculated with the help of Equations (9) and (10), as shown in Table 8 and Table 9. The sixth step normalizes the weighted sums ( S P I ) α and ( S N I ) α independently to obtain the aggregate scores of the models based on Equations (11) and (12), as indicated in Table 10. Finally, the appraisal score ( λ α ) is determined based on the aggregate scores N ( S P I ) α and N ( S N I ) α in the seventh step with the help of Equation (13), as given in Table 10. The eighth step uses the appraisal scores ( λ α ) to determine the ranks of the proposed CNN models based on the BH, E4-L, E4-R, E4-(L+R), and BH+E4-(L+R) datasets. The model with the lowest appraisal score ( λ α ) has the highest performance among the candidate models. Table 10 shows that the proposed BH+E4-(L+R), E4-L, E4-R, SRAD, E4-(L+R), and BH-based CNN models achieved first, second, third, fourth, fifth, and fifth positions for the relaxed state. Likewise, the same eight-step procedure is utilized for the stressed state, and the resulting ranks of each CNN model are given in Table 11. Table 11 shows that the proposed BH+E4-(L+R), SRAD, E4-L, E4-R, E4-(L+R), and BH-based CNN models achieved first, second, third, fourth, fifth, and fifth positions for the stressed state.

4.3.2. Rank Estimation of the CNN-LSTM Models for Two Levels of Stress

For the rank estimation of the SRAD, BH, E4-L, E4-R, E4-(L+R), and BH+E4-(L+R)-based hybrid CNN-LSTM models, the same eight-step procedure is utilized for the relaxed state and stressed state, and the resulting ranks are given in Table 12 and Table 13, respectively. Table 12 shows that the proposed BH+E4-(L+R), BH, E4-L, E4-R, E4-(L+R), and SRAD-based CNN-LSTM models achieved first, second, third, third, fourth, and fifth positions for the relaxed state. Similarly, Table 13 shows that the proposed fused BH+E4-(L+R), BH, E4-L, E4-R, SRAD, and E4-(L+R)-based CNN-LSTM models achieved first, second, third, third, fourth, and fifth positions for the stressed state.

4.3.3. Rank Estimation of the CNN Models for Three Levels of Stress

The eight-step fuzzy EDAS procedure is also utilized for the three levels of stress (low, medium, and high) and the resulting ranks of the SRAD, BH, E4-L, E4-R, E4-(L+R), and BH+E4-(L+R)-based CNN models are shown in Table 14, Table 15 and Table 16, respectively. For the low-stress level, BH+E4-(L+R), SRAD, E4-(L+R), E4-L, E4-R, and BH-based CNN models achieved first, second, third, fourth, fifth, and fifth positions, as shown in Table 14. For the medium-stress level, E4-R, E4-(L+R), BH+E4-(L+R), E4-L, SRAD and BH-based CNN models achieved first, second, third, fourth, fifth, and fifth positions, as shown in Table 15. Likewise, Table 16 shows that the proposed BH+E4-(L+R), SRAD, E4-L, E4-R, BH, and E4-(L+R)-based CNN models achieved first, second, third, fourth, fifth, and fifth positions for the high-stress level.

4.3.4. Rank Estimation of the CNN-LSTM Models for Three Levels of Stress

For the rank estimation of the SRAD, BH, E4-L, E4-R, E4-(L+R), and BH+E4-(L+R)-based hybrid CNN-LSTM models, the same eight-step procedure is utilized for low-stress, medium-stress, and high-stress, and the resulting ranks are given in Table 17, Table 18 and Table 19, respectively. Table 17 shows that the proposed BH+E4-(L+R), BH, E4-L, E4-R, SRAD, and E4-(L+R)-based CNN-LSTM models achieved first, second, third, third, fourth, and fifth positions for the low-stress level. Moreover, Table 18 shows that the proposed, SRAD, E4-R, E4-L, BH, and BH+E4-(L+R)-based CNN-LSTM models achieved first, second, third, third, fourth, and fifth positions for the medium-stress level. For the high-stress level, BH, BH+E4-(L+R), SRAD, E4-L, E4-R, and E4-(L+R)-based CNN models achieved first, second, third, third, fourth, and fifth positions, as shown in Table 19.

4.4. Comparison of the Proposed 1D CNN and 1D CNN-LSTM Models

Comparisons of the proposed 1D CNN and 1D CNN-LSTM models for two and three levels of stress based on training time, accuracy, and fuzzy EDAS ranking are shown in Table 20, Table 21, Table 22 and Table 23. Execution environments for all proposed models are based on a single CPU. The fuzzy EDAS approach performs a comprehensive rank estimation of the proposed models in terms of accuracy, recall, precision, F-score, and specificity. The model with the lowest appraisal score ( λ α ) has the highest performance among the candidate models.
A comparison shows that there is a tradeoff between the training time and performance of various models, with the exception of the SRAD dataset. For example, Table 20 shows that the proposed BH+E4-(L+R)-based 1D CNN model achieved the best EDAS ranks at the cost of a maximum training time of 72 min and 42 s. On the other hand, the BH-based 1D CNN model has the worst EDAS rank with the lowest training time of 3 min and 15 s. The performance of the SRAD-based model is somehow in between with the highest computational cost of 277 min and 29 s. Similarly, Table 21 reveals that the proposed BH+E4-(L+R)-based 1D CNN-LSTM model secured the top EDAS ranks at the cost of a maximum training time of 72 min and 42 s. Moreover, the BH-based 1D CNN model has the worst EDAS rank with the lowest training time of 3 min and 15 s. However, the SRAD-based model achieved average performance, with a maximum computational cost of 167 min and 56 s.
Table 22 shows the comparison of the 1D CNN model for the three-stress class. As usual, the proposed BH+E4-(L+R) secured the best EDAS ranks at a maximum cost of 72 min and 51 s. On the other hand, the BH-based model has the worst EDAS rank for a minimum computational cost of 3 min and 45 s. However, the SRAD-based model has average performance with the highest training time of 279 min and 6 s. Similarly, Table 23 reveals that the proposed SRAD-based 1D CNN-LSTM model secured the top EDAS rank at the cost of a maximum training time of 171 min and 31 s. The BH-based 1D CNN-LSTM model achieved an average EDAS rank with the lowest training time of 2 min and 2 s. However, the E4-(L+R)-based model has the worst EDAS rank despite the high training time of 71 min and 9 s. The proposed models have a high training time due to the usage of a single CPU. Utilizing GPUs may reduce the training time of the proposed algorithms.

5. Discussion

The proposed CNN and hybrid CNN-LSTM models are analyzed using the SRAD and AffectiveROAD datasets in the previous section. AffectiveROAD is based on the BH, Empatica E4-L, Empatica E4-R, and E4-(L+R) datasets. Both BH and Empatica E4 datasets are individually and combinedly used to train the proposed models for the two-stress and three-stress classes. It is evident from the previous tables that the models trained on multimodal data (AffectiveROAD) achieved the maximum performance compared to SRAD data. This shows the importance of physiological, physical, and contextual information in the domain of stress recognition. Moreover, it is also clear that the hybrid 1D CNN-LSTM models achieved better performance than the 1D CNN models. The fuzzy EDAS procedure also shows that the proposed CNN and hybrid CNN-LSTM models achieved the first rank based on the fused AffectiveROAD BH+E4-(L+R) datasets. The achieved performance of the proposed real-world driver stress recognition models is greatly enhanced compared to the existing schemes. A comparison of the proposed driver stress recognition models with the existing schemes is shown in Table 24. It is clear from the table that the proposed models achieved the highest performance for both two and three levels of stress compared to the existing schemes. Rastgoo et al. [11] achieved a higher performance than the proposed models for the three-stress class but their study was based on simulated driving conditions, while the proposed models were based on real-world driving conditions.

6. Conclusions

This paper concludes that in addition to physiological signals, other information regarding the driver, vehicle, and ambiance has an important role in designing reliable and accurate driver stress recognition systems. It is also evident that the hybrid CNN-LSTM models have better performance than the CNN models. Moreover, the fusion of the AffectiveROAD datasets (BH, E4-L, and E4-R) achieved the best performance with the least computational cost compared to the SRAD dataset. This is due to several factors including the utilization of multimodal information (physiological signals and information regarding the driver, vehicle, and environment), quality of the hardware and software tools used for capturing the data, and accurate sampling of the signals. Thus, hybrid deep learning models and multimodal data have a key role in designing an accurate and reliable stress recognition model for real-world driving conditions.
The fusion models based on 1D CNN and hybrid 1D CNN-LSTM produced promising results, but these models may be further improved by utilizing more complex CNN and LSTM architectures. Moreover, a joint CNN-LSTM architecture may be used to further improve stress level recognition. The current study is based on the driver’s stress level, thus in the future, these models may be utilized for drowsiness, cognitive workload, activity, fatigue, and feeling recognition. The AffectiveROAD and PhysioNet SRAD datasets used in this study are based on real-world driving conditions. Such datasets are usually contaminated by different noises and artifacts. Enhanced pre-processing techniques can further improve the performance of the models. Future work may also include stress recognition by training the proposed models using physiological signals acquired using non-contact sensors and smart watches.

Author Contributions

Methodology, M.A. (Muhammad Amin); Software, K.U.; Investigation, M.A. (Muhammad Asif); Resources, K.U. and A.M.; Data curation, M.A. (Muhammad Amin) and M.A.K.; Writing—original draft, M.A. (Muhammad Amin); Writing—review & editing, K.U. and M.A. (Muhammad Asif); Visualization, A.M.; Supervision, K.U. and M.A.K.; Project administration, H.S. and A.M.; Funding acquisition, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University KSA for funding this work through Small Group: RGP.1/368/43.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data will be available on request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Evans, G.W.; Carrère, S. Traffic Congestion, Perceived Control, and Psychophysiological Stress among Urban Bus Drivers. J. Appl. Psychol. 1991, 76, 658–663. [Google Scholar] [CrossRef]
  2. Hanzlíková, I. Professional drivers: The sources of occupational stress. In Proceedings of the Young Researchers Seminar, The Hague, The Netherlands, 2005. [Google Scholar]
  3. de Naurois, C.J.; Bourdin, C.; Stratulat, A.; Diaz, E.; Vercher, J.L. Detection and prediction of driver drowsiness using artificial neural network models. Accid. Anal. Prev. 2019, 126, 95–104. [Google Scholar] [CrossRef]
  4. Chen, L.L.; Zhao, Y.; Ye, P.F.; Zhang, J.; Zou, J.Z. Detecting driving stress in physiological signals based on multimodal feature analysis and kernel classifiers. Expert Syst. Appl. 2017, 85, 279–291. [Google Scholar] [CrossRef]
  5. Manseer, M.; Riener, A. Evaluation of driver stress while transiting road tunnels. In Proceedings of the 6th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, Seattle, WA, USA, 17–19 September 2014; Association for Computing Machinery (ACM): New York, NY, USA, 2014; pp. 1–6. [Google Scholar]
  6. Benlagha, N.; Charfeddine, L. Risk factors of road accident severity and the development of a new system for prevention: New insights from China. Accid. Anal. Prev. 2020, 136, 105411. [Google Scholar] [CrossRef] [PubMed]
  7. American Psychological Association. Stress in America: The State of our Nation. Stress in America Survey; American Psychological Association: Washington, DC, USA, 2017. [Google Scholar]
  8. Khowaja, S.A.; Prabono, A.G.; Setiawan, F.; Yahya, B.N.; Lee, S.L. Toward soft real-time stress detection using wrist-worn devices for human workspaces. Soft Comput. 2021, 25, 2793–2820. [Google Scholar] [CrossRef]
  9. Shinar, D.; Compton, R. Aggressive driving: An observational study of driver, vehicle, and situational variables. Accid. Anal. Prev. 2004, 36, 429–437. [Google Scholar] [CrossRef]
  10. Singh, S. Critical reasons for crashes investigated in the National Motor Vehicle Crash Causation Survey. Natl. Highw. Traffic Saf. Adm. 2015, 1–2. [Google Scholar]
  11. Rastgoo, M.N.; Nakisa, B.; Maire, F.; Rakotonirainy, A.; Chandran, V. Automatic driver stress level classification using multimodal deep learning. Expert Syst. Appl. 2019, 138, 112793. [Google Scholar] [CrossRef]
  12. Nakisa, B.; Rastgoo, M.N.; Tjondronegoro, D.; Chandran, V. Evolutionary computation algorithms for feature selection of EEG-based emotion recognition using mobile sensors. Expert Syst. Appl. 2018, 93, 143–155. [Google Scholar] [CrossRef]
  13. Pourbabaee, B.; Roshtkhari, M.J.; Khorasani, K. Deep Convolutional Neural Networks and Learning ECG Features for Screening Paroxysmal Atrial Fibrillation Patients. IEEE Trans. Syst. Man Cybern. Syst. 2018, 48, 2095–2104. [Google Scholar] [CrossRef]
  14. Liu, Y.; Chen, X.; Peng, H.; Wang, Z. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  15. Zheng, Y.; Liu, Q.; Chen, E.; Ge, Y.; Zhao, J.L. Exploiting multi-channels deep convolutional neural networks for multivariate time series classification. Front. Comput. Sci. 2016, 10, 96–112. [Google Scholar] [CrossRef]
  16. Zheng, Y.; Liu, Q.; Chen, E.; Ge, Y.; Zhao, J.L. Time series classification using multi-channels deep convolutional neural networks. In Proceedings of the Lecture Notes in Computer Science (Including Subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), Macau, China, 16–18 June 2014; Springer: Berlin/Heidelberg, Germany, 2014; Volume 8485, pp. 298–310. [Google Scholar]
  17. Kanjo, E.; Younis, E.M.G.; Ang, C.S. Deep learning analysis of mobile physiological, environmental and location sensor data for emotion detection. Inf. Fusion 2019, 49, 46–56. [Google Scholar] [CrossRef]
  18. Ngiam, J.; Khosla, A.; Kim, M.; Nam, J.; Lee, H.; Ng, A.Y. Multimodal deep learning. In Proceedings of the 28th International Conference on Machine Learning, ICML 2011, Bellevue, WA, USA, 28 June–2 July 2011; pp. 689–696. [Google Scholar]
  19. Deng, L.; Yu, D. Deep learning: Methods and applications. Found. Trends Signal Process. 2013, 7, 197–387. [Google Scholar] [CrossRef]
  20. Dalmeida, K.M.; Masala, G.L. Hrv features as viable physiological markers for stress detection using wearable devices. Sensors 2021, 21, 2873. [Google Scholar] [CrossRef] [PubMed]
  21. Vargas-Lopez, O.; Perez-Ramirez, C.A.; Valtierra-Rodriguez, M.; Yanez-Borjas, J.J.; Amezquita-Sanchez, J.P. An explainable machine learning approach based on statistical indexes and svm for stress detection in automobile drivers using electromyographic signals. Sensors 2021, 21, 3155. [Google Scholar] [CrossRef]
  22. Lopez-Martinez, D.; El-Haouij, N.; Picard, R. Detection of Real-World Driving-Induced Affective State Using Physiological Signals and Multi-View Multi-Task Machine Learning. In Proceedings of the 2019 8th International Conference on Affective Computing and Intelligent Interaction Workshops and Demos, ACIIW 2019, Cambridge, UK, 3–6 September 2019; pp. 356–361. [Google Scholar]
  23. El Haouij, N.; Poggi, J.M.; Ghozi, R.; Sevestre-Ghalila, S.; Jaïdane, M. Random forest-based approach for physiological functional variable selection for driver’s stress level classification. Stat. Methods Appl. 2019, 28, 157–185. [Google Scholar] [CrossRef]
  24. Ghaderi, A.; Frounchi, J.; Farnam, A. Machine learning-based signal processing using physiological signals for stress detection. In Proceedings of the 2015 22nd Iranian Conference on Biomedical Engineering, ICBME 2015, Tehran, Iran, 25–27 November 2016; pp. 93–98. [Google Scholar]
  25. Zhang, L.; Tamminedi, T.; Ganguli, A.; Yosiphon, G.; Yadegar, J. Hierarchical multiple sensor fusion using structurally learned Bayesian network. In Proceedings of the Proceedings—Wireless Health 2010, WH’10, San Diego, CA, USA, 5–7 October 2010; pp. 174–183. [Google Scholar]
  26. Healey, J.A.; Picard, R.W. Detecting stress during real-world driving tasks using physiological sensors. IEEE Trans. Intell. Transp. Syst. 2005, 6, 156–166. [Google Scholar] [CrossRef]
  27. PhysioNet. Stress Recognition in Automobile Drivers, Version: 1.0.0; PhysioNet (MIT Laboratory for Computational Physiology US): Cambridge, MA, USA, 2008. [Google Scholar] [CrossRef]
  28. Rigas, G.; Goletsis, Y.; Bougia, P.; Fotiadis, D.I. Towards driver’s state recognition on real driving conditions. Int. J. Veh. Technol. 2011, 2011, 617210. [Google Scholar] [CrossRef]
  29. Zontone, P.; Affanni, A.; Bernardini, R.; Piras, A.; Rinaldo, R.; Formaggia, F.; Minen, D.; Minen, M.; Savorgnan, C. Car Driver’s Sympathetic Reaction Detection through Electrodermal Activity and Electrocardiogram Measurements. IEEE Trans. Biomed. Eng. 2020, 67, 3413–3424. [Google Scholar] [CrossRef]
  30. Bianco, S.; Napoletano, P.; Schettini, R. Multimodal car driver stress recognition. In Proceedings of the PervasiveHealth: Pervasive Computing Technologies for Healthcare, Trento, Italy, 20–23 May 2019; pp. 302–307. [Google Scholar]
  31. Lee, D.S.; Chong, T.W.; Lee, B.G. Stress Events Detection of Driver by Wearable Glove System. IEEE Sens. J. 2017, 17, 194–204. [Google Scholar] [CrossRef]
  32. Lanatà, A.; Valenza, G.; Greco, A.; Gentili, C.; Bartolozzi, R.; Bucchi, F.; Frendo, F.; Scilingo, E.P. How the Autonomic nervous system and driving style change with incremental stressing conditions during simulated driving. IEEE Trans. Intell. Transp. Syst. 2015, 16, 1505–1517. [Google Scholar] [CrossRef]
  33. Gao, H.; Yuce, A.; Thiran, J.P. Detecting emotional stress from facial expressions for driving safety. In Proceedings of the 2014 IEEE International Conference on Image Processing, ICIP 2014, Paris, France, 27–30 October 2014; pp. 5961–5965. [Google Scholar]
  34. Šalkevicius, J.; Damaševičius, R.; Maskeliunas, R.; Laukienė, I. Anxiety level recognition for virtual reality therapy system using physiological signals. Electronics 2019, 8, 1039. [Google Scholar] [CrossRef]
  35. Rodríguez-Arce, J.; Lara-Flores, L.; Portillo-Rodríguez, O.; Martínez-Méndez, R. Towards an anxiety and stress recognition system for academic environments based on physiological features. Comput. Methods Programs Biomed. 2020, 190, 105408. [Google Scholar] [CrossRef]
  36. Can, Y.S.; Chalabianloo, N.; Ekiz, D.; Ersoy, C. Continuous stress detection using wearable sensors in real life: Algorithmic programming contest case study. Sensors 2019, 19, 1849. [Google Scholar] [CrossRef]
  37. Al abdi, R.M.; Alhitary, A.E.; Abdul Hay, E.W.; Al-bashir, A.K. Objective detection of chronic stress using physiological parameters. Med. Biol. Eng. Comput. 2018, 56, 2273–2286. [Google Scholar] [CrossRef] [PubMed]
  38. Betti, S.; Lova, R.M.; Rovini, E.; Acerbi, G.; Santarelli, L.; Cabiati, M.; Del Ry, S.; Cavallo, F. Evaluation of an integrated system of wearable physiological sensors for stress monitoring in working environments by using biological markers. IEEE Trans. Biomed. Eng. 2018, 65, 1748–1758. [Google Scholar] [CrossRef] [PubMed]
  39. Sriramprakash, S.; Prasanna, V.D.; Murthy, O.V.R. Stress Detection in Working People. In Proceedings of the Procedia Computer Science, Cochin, India, 22–24 August 2017; Volume 115, pp. 359–366. [Google Scholar]
  40. de Vries, G.J.J.; Pauws, S.C.; Biehl, M. Insightful stress detection from physiology modalities using Learning Vector Quantization. Neurocomputing 2015, 151, 873–882. [Google Scholar] [CrossRef]
  41. Sun, F.T.; Kuo, C.; Cheng, H.T.; Buthpitiya, S.; Collins, P.; Griss, M. Activity-aware mental stress detection using physiological sensors. In Proceedings of the Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering, LNICST, Santa Clara, CA, USA, 25–28 October 2012; Volume 76, pp. 282–301. [Google Scholar]
  42. Anceschi, E.; Bonifazi, G.; De Donato, M.C.; Corradini, E.; Ursino, D.; Virgili, L. Savemenow.ai: A machine learning based wearable device for fall detection in a workplace. Stud. Comput. Intell. 2021, 911, 493–514. [Google Scholar] [CrossRef]
  43. Bonifazi, G.; Corradini, E.; Ursino, D.; Virgili, L. Defining user spectra to classify Ethereum users based on their behavior. J. Big Data 2022, 9, 37. [Google Scholar] [CrossRef]
  44. Zalabarria, U.; Irigoyen, E.; Martinez, R.; Larrea, M.; Salazar-Ramirez, A. A Low-Cost, Portable Solution for Stress and Relaxation Estimation Based on a Real-Time Fuzzy Algorithm. IEEE Access 2020, 8, 74118–74128. [Google Scholar] [CrossRef]
  45. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  46. Zhang, J.; Mei, X.; Liu, H.; Yuan, S.; Qian, T. Detecting negative emotional stress based on facial expression in real time. In Proceedings of the 2019 IEEE 4th International Conference on Signal and Image Processing, ICSIP 2019, Wuxi, China, 19–21 July 2019; pp. 430–434. [Google Scholar]
  47. Lim, S.; Yang, J.H. Driver state estimation by convolutional neural network using multimodal sensor data. Electron. Lett. 2016, 52, 1495–1497. [Google Scholar] [CrossRef]
  48. Yan, S.; Teng, Y.; Smith, J.S.; Zhang, B. Driver behavior recognition based on deep convolutional neural networks. In Proceedings of the 2016 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery, ICNC-FSKD 2016, Changsha, China, 13–15 August 2016; pp. 636–641. [Google Scholar]
  49. Hajinoroozi, M.; Mao, Z.; Jung, T.P.; Lin, C.T.; Huang, Y. EEG-based prediction of driver’s cognitive performance by deep convolutional neural network. Signal Process. Image Commun. 2016, 47, 549–555. [Google Scholar] [CrossRef]
  50. Lee, J.; Lee, H.; Shin, M. Driving stress detection using multimodal convolutional neural networks with nonlinear representation of short-term physiological signals. Sensors 2021, 21, 2381. [Google Scholar] [CrossRef] [PubMed]
  51. Masood, K.; Alghamdi, M.A. Modeling Mental Stress Using a Deep Learning Framework. IEEE Access 2019, 7, 68446–68454. [Google Scholar] [CrossRef]
  52. Cho, Y.; Bianchi-Berthouze, N.; Julier, S.J. DeepBreath: Deep learning of breathing patterns for automatic stress recognition using low-cost thermal imaging in unconstrained settings. In Proceedings of the 2017 7th International Conference on Affective Computing and Intelligent Interaction, ACII 2017, San Antonio, TX, USA, 23–26 October 2017; Volume 2018, pp. 456–463. [Google Scholar]
  53. Seo, W.; Kim, N.; Kim, S.; Lee, C.; Park, S.M. Deep ECG-respiration network (DeepER net) for recognizing mental stress. Sensors 2019, 19, 3021. [Google Scholar] [CrossRef]
  54. Hwang, B.; You, J.; Vaessen, T.; Myin-Germeys, I.; Park, C.; Zhang, B.T. Deep ECGNet: An optimal deep learning framework for monitoring mental stress using ultra short-term ECG signals. Telemed. E-Health 2018, 24, 753–772. [Google Scholar] [CrossRef]
  55. He, J.; Li, K.; Liao, X.; Zhang, P.; Jiang, N. Real-Time Detection of Acute Cognitive Stress Using a Convolutional Neural Network from Electrocardiographic Signal. IEEE Access 2019, 7, 42710–42717. [Google Scholar] [CrossRef]
  56. Martínez-Rodrigo, A.; García-Martínez, B.; Huerta, Á.; Alcaraz, R. Detection of negative stress through spectral features of electroencephalographic recordings and a convolutional neural network. Sensors 2021, 21, 3050. [Google Scholar] [CrossRef]
  57. Healey, J.A.; Picard, R.W. Stress Recognition in Automobile Drivers. IEEE Trans. Intell. Transp. Syst. 2005. [Google Scholar]
  58. Clifford, G.D.; Azuaje, F.; Mcsharry, P. ECG statistics, noise, artifacts, and missing data. Adv. Methods Tools ECG Data Anal. 2006, 6, 18. [Google Scholar]
  59. Sörnmo, L.; Laguna, P. Bioelectrical Signal Processing in Cardiac and Neurological Applications; Academic Press: Cambridge, MA, USA, 2005; ISBN 9780124375529. [Google Scholar]
  60. Rahman, J.S.; Gedeon, T.; Caldwell, S.; Jones, R.; Jin, Z. Towards Effective Music Therapy for Mental Health Care Using Machine Learning Tools: Human Affective Reasoning and Music Genres. J. Artif. Intell. Soft Comput. Res. 2021, 11, 5–20. [Google Scholar] [CrossRef]
  61. Sánchez-Reolid, R.; López de la Rosa, F.; López, M.T.; Fernández-Caballero, A. One-dimensional convolutional neural networks for low/high arousal classification from electrodermal activity. Biomed. Signal Process. Control 2022, 71, 103203. [Google Scholar] [CrossRef]
  62. Ordóñez, F.J.; Roggen, D. Deep convolutional and LSTM recurrent neural networks for multimodal wearable activity recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef]
  63. Zhao, G.; Zhang, Z.; Guan, H.; Tang, P.; Wang, J. Rethinking ReLU to Train Better CNNs. In Proceedings of the Proceedings—International Conference on Pattern Recognition, Beijing, China, 20–24 August 2018; Volume 2018, pp. 603–608. [Google Scholar]
  64. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  65. Amin, M.; Ullah, K.; Asif, M.; Waheed, A.; Haq, S.U.; Zareei, M.; Biswal, R.R. ECG-Based Driver’s Stress Detection Using Deep Transfer Learning and Fuzzy Logic Approaches. IEEE Access 2022, 10, 29788–29809. [Google Scholar] [CrossRef]
Figure 1. Training graph of the CNN model for the two-stress class based on the SRAD dataset.
Figure 1. Training graph of the CNN model for the two-stress class based on the SRAD dataset.
Diagnostics 13 01897 g001
Figure 2. Training graph of the CNN model for the two-stress class based on the BH dataset.
Figure 2. Training graph of the CNN model for the two-stress class based on the BH dataset.
Diagnostics 13 01897 g002
Figure 3. Training graph of the CNN model for the two-stress class based on the E4-L dataset.
Figure 3. Training graph of the CNN model for the two-stress class based on the E4-L dataset.
Diagnostics 13 01897 g003
Figure 4. Training graph of the CNN model for the two-stress class based on the E4-R dataset.
Figure 4. Training graph of the CNN model for the two-stress class based on the E4-R dataset.
Diagnostics 13 01897 g004
Figure 5. Training graph of the CNN model for the two-stress class based on the E4-(L+R) datasets.
Figure 5. Training graph of the CNN model for the two-stress class based on the E4-(L+R) datasets.
Diagnostics 13 01897 g005
Figure 6. Training graph of the CNN model for the two-stress class based on the BH+E4-(L+R) datasets.
Figure 6. Training graph of the CNN model for the two-stress class based on the BH+E4-(L+R) datasets.
Diagnostics 13 01897 g006
Figure 7. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the SRAD dataset.
Figure 7. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the SRAD dataset.
Diagnostics 13 01897 g007
Figure 8. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the BH dataset.
Figure 8. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the BH dataset.
Diagnostics 13 01897 g008
Figure 9. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the E4-L dataset.
Figure 9. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the E4-L dataset.
Diagnostics 13 01897 g009
Figure 10. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the E4-R dataset.
Figure 10. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the E4-R dataset.
Diagnostics 13 01897 g010
Figure 11. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the E4-(L+R) datasets.
Figure 11. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the E4-(L+R) datasets.
Diagnostics 13 01897 g011
Figure 12. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the BH+E4-(L+R) datasets.
Figure 12. Training graph of the hybrid CNN-LSTM model for the two-stress class based on the BH+E4-(L+R) datasets.
Diagnostics 13 01897 g012
Figure 13. Confusion matrices of the CNN models for the two-stress class based on: (a) the SRAD dataset; (b) BH dataset; (c) E4-L dataset; (d) E4-R dataset; (e) E4-(L+R) datasets; (f) BH+E4-(L+R) datasets.
Figure 13. Confusion matrices of the CNN models for the two-stress class based on: (a) the SRAD dataset; (b) BH dataset; (c) E4-L dataset; (d) E4-R dataset; (e) E4-(L+R) datasets; (f) BH+E4-(L+R) datasets.
Diagnostics 13 01897 g013
Figure 14. Confusion matrices of the hybrid CNN-LSTM models for the two-stress class based on: (a) the SRAD dataset; (b) BH dataset; (c) E4-L dataset; (d) E4-R dataset; (e) E4-(L+R) datasets; (f) BH+E4-(L+R) datasets.
Figure 14. Confusion matrices of the hybrid CNN-LSTM models for the two-stress class based on: (a) the SRAD dataset; (b) BH dataset; (c) E4-L dataset; (d) E4-R dataset; (e) E4-(L+R) datasets; (f) BH+E4-(L+R) datasets.
Diagnostics 13 01897 g014
Figure 15. Training graph of the CNN model for the three-stress class based on the SRAD dataset.
Figure 15. Training graph of the CNN model for the three-stress class based on the SRAD dataset.
Diagnostics 13 01897 g015
Figure 16. Training graph of the CNN model for the three-stress class based on the BH dataset.
Figure 16. Training graph of the CNN model for the three-stress class based on the BH dataset.
Diagnostics 13 01897 g016
Figure 17. Training graph of the CNN model for the three-stress class based on the E4-L dataset.
Figure 17. Training graph of the CNN model for the three-stress class based on the E4-L dataset.
Diagnostics 13 01897 g017
Figure 18. Training graph of the CNN model for the three-stress class based on the E4-R dataset.
Figure 18. Training graph of the CNN model for the three-stress class based on the E4-R dataset.
Diagnostics 13 01897 g018
Figure 19. Training graph of the CNN model for the three-stress class based on the E4-(L+R) datasets.
Figure 19. Training graph of the CNN model for the three-stress class based on the E4-(L+R) datasets.
Diagnostics 13 01897 g019
Figure 20. Training graph of the CNN model for the three-stress class based on the BH+E4-(L+R) datasets.
Figure 20. Training graph of the CNN model for the three-stress class based on the BH+E4-(L+R) datasets.
Diagnostics 13 01897 g020
Figure 21. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the SRAD dataset.
Figure 21. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the SRAD dataset.
Diagnostics 13 01897 g021
Figure 22. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the BH dataset.
Figure 22. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the BH dataset.
Diagnostics 13 01897 g022
Figure 23. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the E4-L dataset.
Figure 23. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the E4-L dataset.
Diagnostics 13 01897 g023
Figure 24. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the E4-R dataset.
Figure 24. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the E4-R dataset.
Diagnostics 13 01897 g024
Figure 25. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the E4-(L+R) datasets.
Figure 25. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the E4-(L+R) datasets.
Diagnostics 13 01897 g025
Figure 26. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the BH+E4-(L+R) datasets.
Figure 26. Training graph of the hybrid CNN-LSTM model for the three-stress class based on the BH+E4-(L+R) datasets.
Diagnostics 13 01897 g026
Figure 27. Confusion matrices of the CNN models for the three-stress class based on: (a) the SRAD dataset; (b) BH dataset; (c) E4-L dataset; (d) E4-R dataset; (e) E4-(L+R) datasets; (f) BH+E4-(L+R) datasets.
Figure 27. Confusion matrices of the CNN models for the three-stress class based on: (a) the SRAD dataset; (b) BH dataset; (c) E4-L dataset; (d) E4-R dataset; (e) E4-(L+R) datasets; (f) BH+E4-(L+R) datasets.
Diagnostics 13 01897 g027
Figure 28. Confusion matrices of the hybrid CNN-LSTM models for the three-stress class based on: (a) the SRAD dataset; (b) BH dataset; (c) E4-L dataset; (d) E4-R dataset; (e) E4-(L+R) datasets; (f) BH+E4-(L+R) datasets.
Figure 28. Confusion matrices of the hybrid CNN-LSTM models for the three-stress class based on: (a) the SRAD dataset; (b) BH dataset; (c) E4-L dataset; (d) E4-R dataset; (e) E4-(L+R) datasets; (f) BH+E4-(L+R) datasets.
Diagnostics 13 01897 g028
Table 1. One-dimensional CNN architectures and parameter settings.
Table 1. One-dimensional CNN architectures and parameter settings.
Dataset Feature Learning (1D CNN)Training Options (1D CNN)
Filter SizePaddingLayers
Optimization
Algorithm
(Feature Learning/
Classification)
No. of Epochs
(Feature Learning/
Classification)
Mini-Batch size
(Feature Learning/
Classification)
Validation
Frequency
Training/
Validation Set
SRAD03Causal1st Bloch: Conv1D (Filters: 8), ReLU, LN
2nd: Bloch: Conv1D (Filters: 32), ReLU, LN
3rd Block: Conv1D (Filters: 64), ReLU, LN
4th Block: Conv1D (Filters: 128), ReLU, LN
Conv1D (Filters: 8)
GAP; FC; Softmax; Classification
Adam20301080:20
E4-L30
20
20
10
E4-R30
10
20
20
E4-(L+R)Sum of the layers of E4-L and E4-R30 + 30
10
20 + 20
20
BH1st Bloch: Conv1D (Filters: 128), ReLU, LN
2nd: Bloch: Conv1D (Filters: 64), ReLU, LN
3rd Block: Conv1D (Filters: 32), ReLU, LN
GAP; Softmax; FC; Classification

80
10

20
20
BH+E4-(L+R)Sum of the layers of BH, E4-L, and E4-R80 + 30 + 30
10
20 + 20 + 20
20
Table 2. One-dimensional CNN-LSTM architectures and parameter settings.
Table 2. One-dimensional CNN-LSTM architectures and parameter settings.
Dataset Features Learning (1D CNN)Training Options (1D CNN-LSTM)Training/
Validation Set
Filter SizePaddingLayersHidden LayersDropoutOptimization
Algorithm
(Feature Learning/
Classification)
Epochs (Feature
Learning/
Classification)
Mini-Batch Size
(Feature Learning/
Classification)
Validation Frequency
SRAD03Causal1st Bloch: Conv1D (Filters: 8), ReLU, LN
2nd: Bloch: Conv1D (Filters: 32), ReLU, LN
3rd Block: Conv1D (Filters: 64), ReLU, LN
4th Block: Conv1D (Filters: 128), ReLU, LN
Conv1D (8 Filters)
GAP; FC; Softmax; Classification
2500.4Adam20
20
301080:20
E4-L3000.530
10
20
20
E4-R2000.530
10
20
20
E4-(L+R)Sum of the layers of E4-L and E4-R2000.530 + 30
10
20 + 20
20
BH1st Bloch: Conv1D (Filters: 128), ReLU, LN
2nd: Bloch: Conv1D (Filters: 64), ReLU, LN
3rd Block: Conv1D (Filters: 32), ReLU, LN
GAP; Softmax; FC; Classification

200

0.5

80
20

20
20
BH+E4-(L+R)Sum of the layers of BH, E4-L, and E4-R2000.580 + 30 + 30
15
20 + 20 + 20
20
Table 3. Performance analysis of the proposed fusion models for the two-stress class.
Table 3. Performance analysis of the proposed fusion models for the two-stress class.
Deep
Learning Model
DatasetDriver’s Stress LevelPerformance Measure
ACCRCLPRCF1SPCOverall
ACC
1D CNNSRADRelaxed0.92710.87740.91180.89420.954192.7%
Stressed0.92710.95410.93500.94440.8774
BHRelaxed0.89090.92780.82670.87430.865489.1%
Stressed0.89090.86540.94540.90360.9278
E4-LRelaxed0.86530.80720.93940.86830.936386.53%
Stressed0.86530.93630.79890.86220.8073
E4-RRelaxed0.88250.86360.86930.86640.897488.3%
Stressed0.88250.89740.89290.89510.8636
E4-(L+R)Relaxed0.92350.97760.86170.9160.883392.35%
Stressed0.92350.88330.98150.92980.9776
BH+E4-(L+R)Relaxed0.95640.97720.92240.9490.941795.6%
Stressed0.95640.94170.98310.9620.9772
1D CNN-LSTMSRADRelaxed0.91800.90410.87420.88890.926191.8%
Stressed0.91810.92610.94440.93520.9041
BHRelaxed0.95450.98570.90790.94520.934095.5%
Stressed0.95450.93400.99000.96120.9857
E4-LRelaxed0.88820.90070.850.87460.878888.83%
Stressed0.88820.87880.92060.89920.9007
E4-RRelaxed0.90650.90990.86490.88680.904190.65%
Stressed0.90650.90410.93710.92030.9099
E4-(L+R)Relaxed0.94650.95550.93280.9440.938494.65%
Stressed0.94650.93840.95930.94870.9555
BH+E4-(L+R)Relaxed0.96590.93950.98730.96280.989396.59%
Stressed0.96590.98930.94860.96850.9395
Table 4. Performance analysis of the proposed fusion models for the three-stress class.
Table 4. Performance analysis of the proposed fusion models for the three-stress class.
Deep Learning ModelDatasetDriver’s Stress LevelPerformance Measure
ACCRCLPRCF1SPCOverall
ACC
1D CNNSRADLow0.93380.85960.96070.90740.978779.5%
Medium0.83770.59210.71430.64750.9203
High0.83770.86610.7080.77910.7895
BHHigh0.86590.84610.89590.87030.888378.9%
Medium0.88860.62160.39650.48420.9131
Low0.82270.74560.78260.76360.8708
E4-LLow0.89110.85140.92550.88690.93176.5%
Medium0.85670.66670.11320.19350.8618
High0.78220.67880.82960.74670.875
E4-RLow0.89970.86870.90850.88820.925976.79%
Medium0.84810.4930.67310.56910.9388
High0.7880.79660.65280.71760.7835
E4-(L+R)Low0.93120.91590.92410.92000.942883.94%
Medium0.8930.76000.46340.57580.907
High0.85470.78540.8940.83620.9167
BH+E4-(L+R)Low0.95410.95480.92680.94060.953785.66%
Medium0.89290.68000.45950.54840.9154
High0.86620.81750.9180.86490.9197
1D CNN-LSTMSRADLow0.94540.93880.91390.92620.949285.6%
Medium0.90320.80330.64470.71530.9210
High0.86350.81030.89770.85170.9135
BHLow0.95740.92450.98000.95150.984587.8%
Medium0.92040.91300.44680.60000.9210
High0.87780.82940.90970.86770.9231
E4-LLow0.85960.84150.85710.84920.875776.22%
Medium0.87680.65520.36540.46910.8969
High0.7880.69870.80150.74660.8601
E4-RLow0.90270.90210.88330.89260.903182.06%
Medium0.90460.620.50.55360.9346
High0.8340.78240.84230.81130.8772
E4-(L+R)Low0.92350.93950.88210.90990.912384.13%
Medium0.90820.78570.45830.57890.9189
High0.85090.77070.92340.84020.9338
BH+E4-(L+R)Low0.95220.95110.93860.94480.95387.95%
Medium0.91780.77050.61840.68610.9372
High0.88910.83970.90870.87280.9301
Table 5. Cross-efficient values of the CNN models for the relaxed state.
Table 5. Cross-efficient values of the CNN models for the relaxed state.
DatasetPerformance Measure (Driver’s Relaxed State)
ACCRCLPRCF1SPC
SRAD0.92710.87740.91180.89420.9541
BH0.89090.92780.82670.87430.8654
E4-L0.86530.80720.93940.86830.9363
E4-R0.88250.86360.86930.86640.8974
E4-(L+R)0.92350.97760.86170.9160.8833
BH+E4-(L+R)0.95640.97720.92240.9490.9417
ψ β 0.90760.90510.88860.89470.9130
Table 6. Analysis results of the average ( P I ) of the CNN models for the relaxed state.
Table 6. Analysis results of the average ( P I ) of the CNN models for the relaxed state.
DatasetPerformance Measure (Driver’s Relaxed State)
ACCRCLPRCF1SPC
SRAD0.00000.03060.00000.00060.0000
BH0.01840.00000.06960.02280.0522
E4-L0.04660.10820.00000.02950.0000
E4-R0.02770.04590.02170.02490.0171
E4-(L+R)0.00000.00000.03020.00000.0326
BH+E4-(L+R)0.00000.00000.00000.00000.0000
Table 7. Analysis results of the average ( N I ) of the CNN models for the relaxed state.
Table 7. Analysis results of the average ( N I ) of the CNN models for the relaxed state.
DatasetPerformance Measure (Driver’s Relaxed State)
ACCRCLPRCF1SPC
SRAD0.00000.00000.02620.00000.0450
BH0.00000.02500.00000.00000.0000
E4-L0.00000.00000.05720.00000.0255
E4-R0.00000.00000.00000.00000.0000
E4-(L+R)0.01750.08010.00000.02380.0000
BH+E4-(L+R)0.05370.07960.01910.06070.0314
Table 8. Analysis results of the aggregate ( P I ) of the CNN models for the relaxed state.
Table 8. Analysis results of the aggregate ( P I ) of the CNN models for the relaxed state.
Weight of Criteria0.41760.28500.145330.08440.0676
DatasetPerformance Measure (Driver’s Relaxed State)
ACCRCLPRCF1SPC ( S P I ) α
SRAD0.00000.00870.00000.00050.00000.0088
BH0.00770.00000.01000.00190.00350.0233
E4-L0.00000.00000.00000.00000.00000.0000
E4-R0.01160.00000.00310.00270.00120.0185
E4-(L+R)0.00000.00000.00440.00000.00220.0066
BH+E4-(L+R)0.00000.00000.00000.00000.00000.0000
Table 9. Analysis results of the aggregate ( N I ) of the CNN models for the relaxed state.
Table 9. Analysis results of the aggregate ( N I ) of the CNN models for the relaxed state.
Weight of Criteria0.41760.28500.14530.08440.0676
DatasetPerformance Measure (Driver’s Relaxed State)
ACCRCLPRCF1SPC ( S N I ) α
SRAD0.00900.00000.00380.00000.00300.0158
BH0.00000.00710.00000.00000.00000.0071
E4-L0.00000.00000.00830.00000.00170.0100
E4-R0.00000.00000.00000.00000.00000.0000
E4-(L+R)0.00730.02280.00000.00200.00000.0321
BH+E4-(L+R)0.02240.02270.00550.00510.00210.0579
Table 10. Analysis results of the CNN models for the relaxed state.
Table 10. Analysis results of the CNN models for the relaxed state.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.00880.01580.37740.72700.55224
BH0.02330.00711.00000.87680.93846
E4-L0.00000.01000.00000.82670.41332
E4-R0.00000.00000.00001.00000.50003
E4-(L+R)0.01850.03210.79680.44520.62105
BH+E4-(L+R)0.00660.05790.28350.00000.14171
Table 11. Analysis results of the CNN models for the stressed state.
Table 11. Analysis results of the CNN models for the stressed state.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.00210.02630.08740.47910.28322
BH0.02370.00521.00000.89600.94806
E4-L0.00000.00730.00000.85620.42803
E4-R0.00000.00000.00001.00000.50004
E4-(L+R)0.02130.02320.89850.54010.71935
BH+E4-(L+R)0.00930.05050.39130.00000.19561
Table 12. Analysis results of the CNN-LSTM models for the relaxed state.
Table 12. Analysis results of the CNN-LSTM models for the relaxed state.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.02140.00001.00001.00001.00005
BH0.00000.03110.00000.23280.11642
E4-L0.00000.00000.00001.00000.50003
E4-R0.00000.00000.00001.00000.50003
E4-(L+R)0.02120.02250.98870.44520.71694
BH+E4-(L+R)0.00000.04050.00000.00000.00001
Table 13. Analysis results of the CNN-LSTM models for the stressed state.
Table 13. Analysis results of the CNN-LSTM models for the stressed state.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.00930.00000.58751.00000.79384
BH0.00000.02470.00000.34960.17482
E4-L0.00000.00000.00001.00000.50003
E4-R0.00000.00000.00001.00000.50003
E4-(L+R)0.01580.01451.00000.61950.80975
BH+E4-(L+R)0.00000.03800.00000.00000.00001
Table 14. Analysis results of the CNN models for the low-stress level.
Table 14. Analysis results of the CNN models for the low-stress level.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.00750.01900.16960.59960.38462
BH0.04410.00001.00001.00001.00006
E4-L0.00000.00030.00000.99370.49684
E4-R0.00000.00000.00001.00000.50005
E4-(L+R)0.01040.02140.23570.55040.39313
BH+E4-(L+R)0.00000.04760.00000.00000.00001
Table 15. Analysis results of the CNN models for the medium-stress level.
Table 15. Analysis results of the CNN models for the medium-stress level.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.03480.10061.00000.00000.50005
BH0.03220.00940.92500.90610.91556
E4-L0.00000.01400.00000.82400.41204
E4-R0.00000.07610.00000.24380.12191
E4-(L+R)0.01030.07930.29560.21170.25372
BH+E4-(L+R)0.00220.03920.06380.61010.33703
Table 16. Analysis results of the CNN models for the high-stress level.
Table 16. Analysis results of the CNN models for the high-stress level.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.02240.03710.38350.46360.42362
BH0.01940.00090.33260.98680.65975
E4-L0.00000.00710.00000.89740.44873
E4-R0.00000.00540.00000.92120.46064
E4-(L+R)0.05840.04391.00000.36480.68246
BH+E4-(L+R)0.00000.06910.00000.00000.00001
Table 17. Analysis results of the CNN-LSTM models for the low-stress level.
Table 17. Analysis results of the CNN-LSTM models for the low-stress level.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.00000.02040.00000.44670.22334
BH0.00000.03680.00000.00000.00002
E4-L0.00000.00000.00001.00000.50003
E4-R0.00000.00000.00001.00000.50003
E4-(L+R)0.01730.00721.00000.80340.90175
BH+E4-(L+R)0.00580.03320.33650.09810.21731
Table 18. Analysis results of the CNN-LSTM models for the medium-stress level.
Table 18. Analysis results of the CNN-LSTM models for the medium-stress level.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.00090.07320.05590.00000.02801
BH0.01700.06531.00000.10720.55364
E4-L0.00000.00000.00001.00000.50003
E4-R0.00000.00090.00000.98700.49352
E4-(L+R)0.00850.01180.49740.83830.66786
BH+E4-(L+R)0.01680.05610.98910.23270.61095
Table 19. Analysis results of the CNN-LSTM models for the high-stress level.
Table 19. Analysis results of the CNN-LSTM models for the high-stress level.
Dataset ( S P I ) α ( S N I ) α N ( S P I ) α N ( S N I ) α λ α Ranks
SRAD0.00000.01960.00000.59130.29563
BH0.00000.03790.00000.21130.10571
E4-L0.00000.00000.00001.00000.50004
E4-R0.00000.00000.00001.00000.50004
E4-(L+R)0.01870.01021.00000.78840.89425
BH+E4-(L+R)0.00640.04800.34500.00000.17252
Table 20. Comparison of the proposed CNN models for the two-stress class.
Table 20. Comparison of the proposed CNN models for the two-stress class.
DatasetExecution Environment

Training Time
1D CNN
Performance
Accuracy (%)Fuzzy EDAS Rank
Feature
Learning
ClassificationRelaxed StateStressed StateOverallRelaxed StateStressed State
SRADSingle CPU277 min 29 s92.7192.7192.7242
E4-L35 min, 18 s13 s86.5386.5386.5323
E4-R35 min, 26 s13 s88.2588.2588.2534
E4-(L+R)70 min, 26 s12 s92.3592.3592.3555
BH1 min, 46 s1 min, 29 s89.0989.0989.0966
BH+E4-(L+R)72 min, 30 s12 s95.64956495.6411
Table 21. Comparison of the proposed CNN-LSTM models for the two-stress class.
Table 21. Comparison of the proposed CNN-LSTM models for the two-stress class.
DatasetExecution EnvironmentTraining Time
1D CNN-LSTM
Performance
Accuracy (%)Fuzzy EDAS Rank
Feature LearningClassificationRelaxed StateStressed StateOverallRelaxed StateStressed State
SRADSingle CPU167 min 33 s23 s91.8091.8091.8154
E4-L35 min 18 s34 s88.8288.8288.8333
E4-R35 min 26 s31 s90.6590.6590.6533
E4-(L+R)70 min 26 s32 s94.6594.6594.6545
BH1 min 46 s33 s95.4595.4595.4522
BH+E4-(L+R)72 min 30 s21 s96.5996.5996.5911
Table 22. Comparison of the proposed CNN models for the three-stress class.
Table 22. Comparison of the proposed CNN models for the three-stress class.
DatasetExecution
Environment
Training Time
1D CNN
Performance
Accuracy (%)Fuzzy EDAS Rank
Feature
Learning
ClassificationLow-stress
State
Medium-Stress
State
High-Stress
State
OverallLow-Stress
State
Medium-Stress
State
High-Stress
State
SRADSingle CPU279 min, 6 s93.3883.7783.7779.47252
E4-L35 min, 20 s45 s89.1185.6778.2276.50443
E4-R35 min, 19 s13 s89.9784.8178.8076.79514
E4-(L+R)70 min, 39 s12 s93.1289.3085.4783.94326
BH1 min, 28 s2 min, 17 s86.5988.8682.2778.86665
BH+E4-(L+R)72 min, 7 s22 s95.4189.2986.6285.66131
Table 23. Comparison of the proposed CNN-LSTM models for the three-stress class.
Table 23. Comparison of the proposed CNN-LSTM models for the three-stress class.
DatasetExecution
Environment
Training Time
1D CNN-LSTM
Performance
Accuracy (%)Fuzzy EDAS Rank
Feature
Learning
ClassificationLow-Stress
State
Medium-Stress
State
High-Stress
State
OverallLow-Stress
State
Medium-Stress
State
High-Stress
State
SRADSingle CPU170 min, 44 s47 s94.5490.3286.3585.61411
E4-L35 min, 20 s3 min, 4 s85.9687.6878.8076.22333
E4-R35 min, 19 s33 s90.2790.4683.4082.06322
E4-(L+R)70 min, 39 s30 s92.3590.8285.0984.13566
BH1 min, 28 s34 s95.7492.0487.7887.78244
BH+E4-(L+R)72 min, 7 s31 s95.2291.7888.9187.95155
Table 24. Comparison of the proposed stress recognition models with existing schemes.
Table 24. Comparison of the proposed stress recognition models with existing schemes.
Article /
Year
Signal(s) /
Modalities
EnvironmentNo. of SubjectsData I/P
Mechanism
Deep
Learning
Approach
Stress LevelsACC (%)RCLPRCF1SPC
Proposed
Models
HR, BR, Posture, Activity, TEMP, IBI, GSR, ACCLReal-World
Driving
101D SignalCNN-LSTM296.696.496.896.696.4
38885.482.283.594
[50]/2021HR, GSR9Continues RPsCNN295.795.795.995.7-
[46]/2019Facial Images123ImagesPre-Trained MTCNN297.3----
[11]/2019ECG, VDD, EPSimulated Driving271D SignalCNN-LSTM392.894.195-97.4
[49]/2016EEG371D SignalCCNN286.1----
[53]/2019RESP and ECGCognitive Tasks
(Lab/Workplace)
181D SignalCNN-LSTM383.9--81.1-
[56]/2021EEG32SpectrogramPre-Trained AlexNet284.885.2--84.3
[55]/2019ECG20CNN282.7----
[54]/2018ECG131D Signal1D CNN280----
[52]/2017Thermographic Patterns of Breath8SpectrogramCNN284.6----
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Amin, M.; Ullah, K.; Asif, M.; Shah, H.; Mehmood, A.; Khan, M.A. Real-World Driver Stress Recognition and Diagnosis Based on Multimodal Deep Learning and Fuzzy EDAS Approaches. Diagnostics 2023, 13, 1897. https://doi.org/10.3390/diagnostics13111897

AMA Style

Amin M, Ullah K, Asif M, Shah H, Mehmood A, Khan MA. Real-World Driver Stress Recognition and Diagnosis Based on Multimodal Deep Learning and Fuzzy EDAS Approaches. Diagnostics. 2023; 13(11):1897. https://doi.org/10.3390/diagnostics13111897

Chicago/Turabian Style

Amin, Muhammad, Khalil Ullah, Muhammad Asif, Habib Shah, Arshad Mehmood, and Muhammad Attique Khan. 2023. "Real-World Driver Stress Recognition and Diagnosis Based on Multimodal Deep Learning and Fuzzy EDAS Approaches" Diagnostics 13, no. 11: 1897. https://doi.org/10.3390/diagnostics13111897

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop