Next Article in Journal
Statistical-QoS Guaranteed Energy Efficiency Optimization for Energy Harvesting Wireless Sensor Networks
Next Article in Special Issue
Benchmarking Foot Trajectory Estimation Methods for Mobile Gait Analysis
Previous Article in Journal
Line-Constrained Camera Location Estimation in Multi-Image Stereomatching
Previous Article in Special Issue
A Wireless ExG Interface for Patch-Type ECG Holter and EMG-Controlled Robot Hand
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Biological Signal-Based Stress Monitoring Framework for Children Using Wearable Devices

1
Department of Industrial and Management Engineering, Kyonggi University, Suwon 16227, Korea
2
Department of Industrial and Management Engineering, Incheon National University, Incheon 22012, Korea
3
Department of Library and Information science, Incheon National University, Incheon 22012, Korea
*
Authors to whom correspondence should be addressed.
Sensors 2017, 17(9), 1936; https://doi.org/10.3390/s17091936
Submission received: 5 June 2017 / Revised: 8 August 2017 / Accepted: 9 August 2017 / Published: 23 August 2017
(This article belongs to the Special Issue Wearable and Ambient Sensors for Healthcare and Wellness Applications)

Abstract

:
The safety of children has always been an important issue, and several studies have been conducted to determine the stress state of a child to ensure the safety. Audio signals and biological signals including heart rate are known to be effective for stress state detection. However, collecting those data requires specialized equipment, which is not appropriate for the constant monitoring of children, and advanced data analysis is required for accurate detection. In this regard, we propose a stress state detection framework which utilizes both audio signal and heart rate collected from wearable devices, and adopted machine learning methods for the detection. Experiments using real-world data were conducted to compare detection performances across various machine learning methods and noise levels of audio signal. Adopting the proposed framework in the real-world will contribute to the enhancement of child safety.

1. Introduction

Recently, a rapid growth in the number of instances of child abuse in nursery schools is being reported in Korea [1], and therefore, the necessity for real-time child monitoring is getting attention. Monitoring the stress state of children aged around three to five is particularly important, as their linguistic and physical abilities are immature, making it hard for their parents to be informed on their condition. In actual practice, a walkie-talkie for infants or CCTV in a nursery school are used for child monitoring, both of which have the limited coverage and provide only partial information related to the stress state of a child.
There is a limited number of stress detection studies directly conducted for infants or children. Most of them utilized audio signal collected from children for crying detection [2,3,4,5], which is relatively easy to obtain. Using only audio signal for monitoring a child is not practical, as distinguishing the voice or sound of a child in real circumstances, where many children are gathered in one place, is almost impossible. In such a condition, previously proposed stress detection methods would work poorly. For instance, the state of a child might be determined to be stressed using the previous methods even if the child is not crying when other children begin to cry. Therefore, utilizing biological signals—which has proven to be effective in previous studies on the stress detection of adults [6,7]—in addition to the audio signal will help reducing the false positives of the stress state detection of a child.
However, specialized equipment is required for the acquisition of most biological signals, and that equipment is sometimes too heavy or intrusive, making it unsuitable as a device for the constant monitoring of a child. Sensors for acquiring the brainwaves or electrodermal signal of the child need to be attached to the forehead or skin of a child, which is not possible in a real-world situation. On the other hand, heart rate can be collected using an unobtrusive device such as a wearable band with fairy accurate performances attributed to advents in the sensing technology.
To this end, we propose a stress detection framework for children using audio signal and biological signal (heart rate) acquired from a wearable device. Each child’s biological signal is collected by using a wearable device attached to their wrist. Then, the signal is transmitted to a server, and the child’s stress state is classified using a learning-based stress detection algorithm introduced in this paper. An alert is provided to the smart devices of their parents when the stress state of the child is detected.
Specifically, a three-step stress detection algorithm is introduced to provide accurate detection performances. First, raw audio signals are preprocessed to extract meaningful features, and are combined with heart rate. Second, features having high discriminative power for the stress and normal states are selected. Then, a classifier is learned from training data composed of the selected features, and stress states of children are determined using the classifiers.
The paper is organized as follows. The proposed stress monitoring framework and the stress detection algorithms are introduced in Section 2. In Section 3, the performances of the proposed framework are evaluated using real-world data, and the paper concludes in Section 4.

2. Literature Review

Table 1 shows the summary of previous research on stress state detection in terms of detection target, utilized method, and data. The previous studies are divided into two groups according to the target—child or adult.
Only a few studies have been conducted for infants or children. Audio signal collected from children has mainly been used in previous studies [2,3,4,5], attributed to the ease of data collection. Most studies adopted machine learning methods such as k-nearest neighbor [4] and hidden Markov model [3], since they can learn stress detection classifiers from data composed of multiple features without using explicitly defined rules or indices.
Unlike stress detection for children, diverse types of data such as biological signals have been used in the stress detection studies for adults. Biological signals include brainwave [10], electrodermal signal [8,9,10,11], and heart rate [6,7]. Accelerometer data [10] and respiration data [12] were also utilized. While most studies employed machine learning methods such as decision tree (DT) [7,8,10], naive Bayes (NB) [10], and support vector machine (SVM) [7,8,9,10,11,12] introduced indices for discriminating stress states.
The accuracy of stress detection performances in previous studies were around 80% to 90%. For instance, Healey [12] obtained an accuracy of around 97% by utilizing features from durations of 5 min. Setz and Sun [9,10] reported that their methods respectively yielded 82.8% and 91.0% accuracy by using SVM.
In addition, there are commercial devices for detecting stress states, such as iCalm [14] and ParentGuardian [15]. A summary comparison of the products and the proposed framework is shown in Table 2.
Both iCalm and ParentGuardian aim to detect the stress state of users like the proposed framework. However, iCalm is not only for infants or children but also for adults and utilizes diverse types of data by attaching the device to wrist and foot. ParentGuardian is generally for children, but is only tested for children with a special case. More importantly, both products do not utilize audio signal for detection.

3. Biological Signal-Based Stress Detection Framework for Children

3.1. Stress Detection Framework

In this section, we introduce a stress detection framework for children using wearable devices. Figure 1 shows the steps and elements of the proposed framework.
The proposed framework is composed of three elements: child-side, server-side, and parents-side. In the child-side, the audio signal and heart rate of a child are continuously sensed and saved in a wearable device attached to the child’s wrist. The collected data are transferred to the server at fixed intervals. On the server-side, the stress state of the child is determined by analyzing the transmitted data in a certain length. Then, on the parents-side, the detected state of the child is provided in real-time, and an alert is generated when the stress state is detected. Although the proposed framework is supposed to be real-time monitoring, there exists a latency since the collected data from a child is transmitted to a server at a fixed interval. However, the latency can be ignored by minimizing the interval, making the proposed framework similar to a real-time monitoring. Details of the stress detection algorithm are provided in Section 3.2.

3.2. Stress Detection Algorithm

3.2.1. Overview

After the audio signal and heart rate of a child are collected and transferred to the server, the stress state of the child is determined by using a learning-based stress detection algorithm. The overview of the algorithm is presented in Figure 2, which is composed of training and test phases. In Figure 2, solid and broken lines indicate the training and test phases, respectively, and shaded boxes indicate the detection steps, where the respective sections are noted in round brackets.
In the training phase, the stress detection method is developed after performing the following three steps. Firstly, meaningful features are extracted from the raw data. Audio signal is time-series data, and extracting meaningful features is one of the most important tasks for accurate classification [16]. Then, features with the highest discriminative power for stress state detection are selected, since irrelevant features can degrade the detection performance [17], and small number of features contributes to more efficient classification [18]. Lastly, a stress detection method is developed by training a machine learning method using the data composed of values of the selected features and a corresponding stress state label called training data.
In the test phase, the momentary stress state of a child is determined. Therefore, the test phase is repeatedly executed in real-time, unlike the training phase which is executed only for once. A child’s raw data collected in real-time is transformed to test data, which is composed of the values of the selected features in the training phase. Then, whether or not the child is in a stress state is determined by analyzing the test data by using the stress detection method from the training phase.

3.2.2. Feature Extraction

It is important to extract meaningful features from the signal for accurate stress state detection, since we utilize audio signal which is time-series data. Diverse features are extracted from raw data using jAudio [19], which is an implementation of feature extraction algorithms for analyzing audio signals in java. Table 3 shows the 27 feature types provided in jAudio, which can be categorized into three groups according to the preprocessed data. Most feature types are calculated using the output of a discrete Fourier transform. Others are calculated using beat histogram or frequency information from the raw signal. Details of the feature types are provided in [19].
Since some feature types generate multidimensional vectors such as MFCC (mel-frequency cepstral coefficients) and beat histograms, while others generate single values such as RMS (root mean square) and spectral centroid, the total number of extracted features from jAudio is 136. The values of a feature for a certain duration are aggregated. Note that we utilized a general aggregator function including mean and standard deviation.
Figure 3 shows (a) graphs of raw audio signal for the two states (normal (upper) and stress (lower)), and (b) a heatmap of the normalized values of the extracted features for every ten-second duration of the raw signal. Specifically, each column in the heatmap indicates one of the extracted features, and each row indicates one of the durations . For instance, a cell located in the third row and the fifth column is a value of the fifth feature calculated using the values in the third duration.
The amplitude of the raw signal in normal state is much smaller than that of the signal in stress state—about 1000 times smaller. It is noticeable that there are only a few features whose values have highly distinguishing patterns between stress and normal states, and most features show similar patterns. Therefore, selecting and utilizing the features with distinguishing patterns will generate good performance for the classification of the two stress states of children.
In addition to the extracted features of audio signal, we utilize the child’s heart rate for the detection. Particularly, we utilized the average of element heart rates for ten seconds as a heart rate, where an element heart rate is calculated using the duration of two consecutive heart beats, to obtain more accurate values for a short duration. Heart rate at the i-th duration is denoted by h i , and calculated using Equation (1) for 0 j n b .
h i = 1 n b + 1 j 60 b j + 1 b j ,
where n b indicates the total number of heart beats at the i-th duration , and b j indicates the time when the j-th heart beat occurred.

3.2.3. Feature Selection

For more accurate and efficient detection, feature selection was conducted to eliminate irrelevant features for the detection. Feature selection methods are categorized into filtering and wrapper [20] approaches. The filtering approach observes the relationship between values of a feature and their labels in terms of a certain criteria, and features are selected according to the scores of features calculated using the criteria. The wrapper approach repeatedly performs classification using different subsets of features in a predefined order, and compares their performance in order to select a subset with the best performance.
In the filtering approach, we adopted chi-square and information gain as the criteria, which were known to be the most effective for feature selection in comparison studies [21]. In the wrapper approach, we chose SVM as classifier, which is known to consistently show good performance [22] and be sensitive to whole features. We denote chi-square, information gain, and SVM wrapper as CHI, IG, and SVMW, respectively, in the following for simplicity.
  • Chi-square-based selection
    CHI utilizes the correlation between a feature and stress states by measuring the divergence of observed data from the expected distribution which assumes that the feature and labels are independent. The score for CHI is obtained as the sum of the square of the difference between observed value and expected value of a feature over the expected value. According to the score, the predefined number of features, denoted by n f , are selected.
  • Information gain-based selection
    IG evaluates a feature by measuring the information gain with respect to the stress states. The score for IG is obtained as the difference in entropy when a feature is given or not. According to the score, the top n f features are selected.
  • SVM wrapper-based selection
    SVMW utilizes SVM as a classifier for the performance evaluation of subsets. For the subset generation, a best-first search is utilized, which known to work best for SVM [20]. Accuracy is adopted as an evaluation metric. According to the accuracy obtained by classification using subsets, the features included in the best subset are selected.

3.2.4. Detection Model Training

Machine learning methods are trained for the stress state detection of children by using the selected features. Machine learning methods are widely utilized for classification and prediction problems such as energy consumption prediction [23], sentiment analysis [24], and scientific success prediction [25]. For the detection, we adopted the three most well-known machine learning methods: DT, NB, and SVM. Details of the models are provided in the following paragraphs.
We tried to detect the stress state, denoted by y i , of a child for a duration, where i is an index of the duration and y i { 0 , 1 } , 0 for normal state and 1 otherwise. Specifically, the length of a duration is predefined as 10 s. The value of the selected features for i-th duration is presented as a vector X i which is composed of x i , j where j is an index of the selected features and 1 j n f . Therefore, a classifier is learned using training data composed of instances, denoted by ( X i , y i ) for i = 0 , , n d , where n d is the total number of durations.
  • Decision tree-based detection
    DT is a tree-shaped classifier where each node is composed of a feature and a corresponding classification value.When an instance is given to the root node, each node classifies the instance according to its feature and value pair. We utilized the C4.5 algorithm [26] which is an extension of ID3 [27] to handle continuous features as we examine time-series signals for the detection. The Gini index was adopted for the feature selection in each node.
  • Naive Bayes-based detection
    NB [28] uses Bayes’ rule for the computation of the probability of a given X i to be in y i . A formal representation of the probability is shown in Equation (2). It assumes that, given a label, features are conditionally independent. The probabilities for features are estimated from data using maximum likelihood estimation.
    P ( y i | X i ) = P ( X i | y i ) P ( y i ) P ( X i ) P ( x i , 1 , x i , 2 , , x i , n f ) P ( y i ) = j n f P ( x i , j | y i ) P ( y i ) .
  • Support vector machine-based detection
    SVM [29] is one of the most well-known machine learning methods, and is widely applied to diverse domains (e.g., document classification) [30]. It finds the maximum margin among instances of normal and stress states. As a result, SVM shows relatively stable performances regardless of the number of training data and features. Equation (3) is a Lagrangian dual problem of the objective function of SVM. The optimal solution can be obtained by solving a quadratic programming.
    min 1 2 k l y k y l α k α l ( X k · X l + λ δ k , l ) l α l s . t . 0 α l C l α l y l = 0
    SVM has the advantage that it is able to classify data which are not linearly separated by using kernel function which maps a vector into a higher dimension. In this paper, we considered radial and linear kernels for comparison, and named them as SVM-R and SVM-L, respectively.

3.2.5. Stress Detection

Using the trained methods, the stress state of a child at the i -th duration is determined. From the audio signal and heart rate collected at the i -th duration, X i is constructed according to the selected features. Then, the trained method determines the stress state of the child at the i -th druation, y ^ i , which maximizes the probability using Equation (4).
y ^ i = arg max y i P ( y i | X i ) .

4. Experiment

4.1. Stress Detection Device Prototype

A prototype framework was implemented for the evaluation of the proposed framework. Figure 4 and Figure 5 show the child-side, server-side, and parents-side elements in the framework. The external and internal views of the prototype device are presented in Figure 5a,b, respectively. Note that the device is a prototype and that the wearing sensation was not considered. Inside the device, there are two sensors: a microphone for acquiring audio signal and a heart rate sensor which collects heart rate by attaching to the inner side of the wrist.
Figure 4a is a snapshot of the stress detection program which collects test data from a wearable device with a time-stamp and detects the stress state of a child at that time using the trained stress detection method. Then, the results are sent to the application on the parents-side, reporting the stress state of their child. Figure 4b shows a screenshot of the application when a child is in normal state (left) or in stress state (right).

4.2. Experiment Settings

We have conducted experiments to observe the performances and characteristics of the stress detection method in the proposed framework. For the evaluation, we have collected real-world audio signal and heart rate. The collected data were divided into instances with length of ten seconds. The total number of instances was 262, where the numbers of the instances in stress and normal states were, respectively, 138 and 124.
Note that the dataset utilized in the experiments was syntactically generated, since collecting data from children in a stress state may cause an ethical issue. The dataset was generated by combining audio signal obtained from [31] which is a collection of sounds from a number of children and heart rate obtained from six subjects to conform with the stress and normal states. Moreover, we assumed that children who are crying are in a stress state and marked the data as a stress state. For example, audio signal and heart rate of a child who was having fun while running and screaming were used as one of the instances in normal state. For evaluation, we have employed k-fold cross validation, where k was set 10 to minimize randomness.
To examine the effect of diverse settings on the detection performance, we have evaluated the performances of the method according to the utilization of heart rate, the number of the selected features, and the noise level. Specifically, the audio signal with noise was considered to investigate the robustness of the proposed method in real circumstances, where other sounds exist. We utilized Adobe Audition CC [32] to generate the audio signal with noise, where white noise is generated with the intensity of the noise on a scale of 2 to 40. As the intensity gets higher, the noise becomes more erratic, harsher, and louder. Therefore, the noise level indicates the strength of white noise added to the original audio signal, and we considered six levels: 0, 5, 10, 15, 25, and 40.
We adopted accuracy as an evaluation measure. Accuracy is one of the most widely utilized metrics for classification problems [30], and is defined as the ratio of the number of instances which are correctly classified over the number of all instances, as shown in Equation (5).
A c c u r a c y = T P + T N T P + F P + F N + T N ,
where T P (true positive), F P (false positive), F N (false negative), and T N (true negative) respectively indicate the numbers of instances when the predicted and the actual states are stress, the predicted state is stress while the actual state is normal, the predicted state is normal while the actual state is stress, and the predicted and actual states are normal (Table 4).
In addition to the accuracy, we adopted recall and precision for detailed evaluation. Recall indicates the sensibility of a model and is calculated as the ratio of the number of instances which are correctly classified over the number of instances which are actually in stress state, as per Equation (6).
R e c a l l = T P T P + F P .
Precision is calculated as the ratio of the number of instances which are correctly classified over the number of instances which are predicted as stress state, as per Equation (7).
P r e c i s i o n = T P T P + F N .

4.3. Experimental Results

4.3.1. Feature Selection

Features with high discriminative power were selected among the features extracted from the audio signal. We considered three methods—CHI, IG, and SVMW—for the feature selection. Table 5 shows the top five selected features according to the three methods.
It is noticeable that there were features which commonly appeared across the selection methods, such as MFCC overall standard deviation and heart rate. MFCC-related features were most frequently selected for all selection methods, and among them, MFCC overall standard deviation showed highest ranks. Heart rate ranked the first for both CHI and IG, and although heart rate was not included in the top five features of SVMW, it also ranked the ninth for SVMW. Moreover, the rank of heart rate for SVMW got higher as the noise level increased. Heart rate ranked ninth, sixth, fifth, and first by SVMW according to the noise level of 0, 5, 10, and 40, respectively.
While the top five selected features of CHI and IG were similar, those of SVMW differed from those of CHI and IG. This is explained by the difference in the concept of the filtering and wrapper approaches. CHI and IG basically examine the relationship between features and labels (particularly correlation), while SVMW heuristically tests the subsets of features for the detection. Therefore, SVMW incorporates the interaction among features, resulting in more diverse features than CHI and IG.
Moreover, to examine the discriminative power of the selected features, a scatter matrix is provided in Figure 6. We considered the four features which appeared most commonly across the selection methods, including heart rate (HR), MFCC overall standard deviation (MFCC), magnitude spectrum overall average (Magnitude), and power spectrum overall average (Power). In Figure 6, diagonal plots show the histogram of stress states according to the value of each feature, and non-diagonal plots show the scatter plot of feature pairs.
Scatter plot of HR and MFCC seems most discriminative for stress and normal states. When a child is in normal state, smaller values of MFCC and extreme values of HR are expected. In terms of Power and Magnitude, most instances have small values, while some of normal state have extremely large values.

4.3.2. Performance Comparison

We conducted three performance comparison experiments. First, state detection performances according to the employed data and method were evaluated. Figure 7 shows the accuracy of the four detection methods—DT, NB, SVM-R, and SVM-L— according to the utilized data, audio signal only, heart rate only, and both audio signal and heart rate in terms of accuracy, recall, and precision. Note that all extracted features were employed in this experiment.
The best accuracy was 82.18 when both data was utilized and SVM-R was employed, while the worst one was 65.27 when only heart rate was utilized and SVM-R was employed. In terms of the utilized data, for the three methods except for NB, the best accuracy was obtained when both audio signal and heart rate were employed. DT and SVM-R performed the best in terms of the employed methods, as the average accuracies of methods DT, NB, SVM-R, and SVM-L across the utilized data were 79.47, 68.37, 76.40, and 74.77, respectively. The low accuracy of NB implies that there may exist correlations among features which NB ignores. The differences between recall and precision of NB and SVM-L are large, while those of DT and SVM-R are relatively small, implying that NB tends to classify instances as stress states excessively whereas SVM-L does the opposite. Utilizing heart rate contributed to conservative classification, since precisions were higher than recalls for most cases where heart rate was utilized. Table 6 shows the results of t-test conducted on the accuracies obtained by performing 10-fold cross-validation. Except for the comparison between DT and SVM-R when only audio signal was utilized, alternative hypothesis is accepted at significance level of 0.05.
Second, the detection performances of the proposed framework using the selected features are shown in Figure 8. Performances of the four methods—DT, NB, SVM-R, and SVM-L—are presented according to n f , 10, 30, 50, and 100, and the feature selection methods, CHI (left), IG (middle), and SVMW (right). The upper plots in Figure 8 show the accuracies when only audio signal was used, and the lower plots show those when both audio and heart rate were used. For comparison purposes, detection accuracies using all features are provided on the right-side of the graphs.
The best accuracy was 93.47 when the feature selection method, detection method, and n f were SVMW, SVM-L, and 100, respectively, while the worst accuracy was 64.01 when the feature selection method, detection method, and n f were CHI, NB, and 30, respectively. Overall, detection performances were better when feature selection was conducted, since irrelevant features were removed from the training dataset. This conforms with the well-known fact that the performances of a machine learning method degrade when irrelevant features are utilized. The average accuracies of the feature selection methods CHI, IG, and SVMW across the other factors were 76.37, 77.54, and 84.65, respectively, implying that the wrapper approach outperforms the filtering approach.
Moreover, it is noticeable that as n f increases the accuracies of SVM-R and SVM-L increase, while they remain still or sometimes decrease for DT and NB. This can be explained by the fact that NB is robust to the irrelevant features, and DT internally selects good features during training.
Last, to evaluate the performance of the proposed method in a real-world situation, we conducted detection on data with noise. Figure 9 shows the accuracies of the four methods—DT (upper left), NB (upper right), SVM-R (lower left), and SVM-L (lower right)— according to the five noise levels, 5, 10, 15, 25, 40, and utilized data with audio only and with both audio and heart rate. Note that SVMW—which performed the best in the previous experiment—was adopted as a feature selection method, and the accuracies were averaged a across n f , 10, 30, 50, and 100.
Overall, it was observed that the accuracies decreased as the noise level increased, as expected. The accuracy decrements were much larger between 5 and 15 than between 15 and 25. Particularly, when heart rate was utilized together with audio signal, the performance was more robust than using only audio signal, implying that utilizing heart rate not only improves accuracy but also makes the method robust to environment.

5. Conclusions

In this paper, we proposed a stress detection framework for children using audio signal and heart rate acquired from a wearable device. The proposed framework is composed of three parts: child-side, where data is collected; server-side, where the stress detection is conducted; and parents-side, where detection results are presented. The stress detection algorithm is divided into two phases: training phase, where detection method is developed; and test phase, where the real-time stress detection is conducted.
Both audio signal and heart rate of a child are utilized for the stress detection. Three feature selection methods—CHI, IG, and SVMW—were employed to determine the most effective features from raw audio signal, and four detection methods—DT, NB, SVM-R, and SVM-L—were adopted for performance comparison. SVMW-based feature selection and SVM-L showed the best performance. Moreover, the accuracy of the proposed framework using audio signal with diverse levels of noise was evaluated to examine the performances of the proposed method in a real situation. In conclusion, eliminating irrelevant features improved the performances, and utilizing both heart rate and audio signal enhanced the performance and made the method more robust to noise.
The advantages of the proposed framework over the previous studies are as follows. First, the proposed method is more robust to noise in the audio signal by utilizing heart rate in addition to audio signal, so it is able to identify the state of a target child even if the child is in noisy circumstances. Second, the proposed method may detect the stress state of a child with special conditions such as autism by analyzing the heart rate of the child along with the audio signal, even if a child does not make any noise. However, a specially trained model which uses data collected from the children with a special condition to reflect the characteristics of those children would work the best. Third, no additional equipment except for a wrist band is required for the detection. Complex models using diverse features and highly computational methods may perform better in an experimental environment, but they are not practical.
For future work, we plan to extend our research in terms of utilized data, features, and methods. We will conduct a large-scale experiment, and utilize additional signals such as accelerometer and electrodermal data, which were effective in previous work for adults. Lastly, privacy issues should be considered for the adoption of the proposed method since it utilizes the human-generated data. Therefore, we will adopt an on-device method, where data analysis is conducted only in a device without sending private data to the outside, in order to resolve the privacy problem.

Acknowledgments

This work was supported by Incheon National University (International Cooperative) Research Grant in 2014, and by the National Research Foundation (NRF) funded by the Ministry of Science, ICT & Future Planning (2017R1C1B5017766).

Author Contributions

Y.C. and Y.-M.J. contributed to the conducting experiments and writting manuscript, and K.K. and L.W. equally corresponds to the manuscript. All authors read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ryu, J. Child Maltreatment and Improvement Direction for Child Protection System. In Health and Welfare Policy Forum; Korea Institute for Health and Social Affairs: Seoul, Korea, 2017; pp. 5–23. [Google Scholar]
  2. Abou-Abbas, L.; Alaie, H.F.; Tadj, C. Automatic detection of the expiratory and inspiratory phases in newborn cry signals. Biomed. Signal Process. Control 2015, 19, 35–43. [Google Scholar] [CrossRef]
  3. Rosales-Pérez, A.; Reyes-García, C.A.; Gonzalez, J.A.; Reyes-Galaviz, O.F.; Escalante, H.J.; Orlandi, S. Classifying infant cry patterns by the Genetic Selection of a FuzzyModel. Biomed. Signal Process. Control 2015, 17, 38–46. [Google Scholar] [CrossRef]
  4. Cohen, R.; Lavner, Y. Infant cry analysis and detection. In Proceedings of the IEEE Convention of Electrical Electronics Engineers in Israel, Eilat, Israel, 14–17 November 2012; pp. 1–5. [Google Scholar]
  5. Ruvolo, P.; Movellan, J. Automatic cry detection in early childhood education settings. In Proceedings of the IEEE International Conference on Development and Learning, Monterey, CA, USA, 9–12 August 2008; pp. 204–208. [Google Scholar]
  6. Melillo, P.; Bracale, M.; Pecchia, L. Nonlinear Heart Rate Variability features for real-life stress detection. Case study: Students under stress due to university examination. Biomed. Eng. Online 2011, 10. [Google Scholar] [CrossRef] [PubMed]
  7. Riganello, F.; Sannita, W.G. Residual brain processing in the vegetative state. J. Psychophysiol. 2009, 23, 18–26. [Google Scholar] [CrossRef]
  8. Kurniawan, H.; Maslov, A.V.; Pechenizkiy, M. Stress detection from speech and Galvanic Skin Response signals. In Proceedings of the IEEE International Symposium on Computer-Based Medical Systems, Porto, Portugal, 20–22 June 2013; pp. 209–214. [Google Scholar]
  9. Setz, C.; Arnrich, B.; Schumm, J.; La Marca, R.; Tröster, G.; Ehlert, U. Discriminating stress from cognitive load using a wearable EDA device. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 410–417. [Google Scholar] [CrossRef] [PubMed]
  10. Sun, F.T.; Kuo, C.; Cheng, H.T.; Buthpitiya, S.; Collins, P.; Griss, M. Activity-aware mental stress detection using physiological sensors. In Proceedings of the International Conference on Mobile Computing, Applications, and Services, Santa Clara, CA, USA, 25–28 October 2010; pp. 211–230. [Google Scholar]
  11. Bakker, J.; Pechenizkiy, M.; Sidorova, N. What’s Your Current Stress Level? Detection of Stress Patterns from GSR Sensor Data. In Proceedings of the IEEE International Conference on Data Mining Workshops, Vancouver, BC, Canada, 11 December 2011; pp. 573–580. [Google Scholar]
  12. Healey, J.A.; Picard, R.W. Detecting stress during real-world driving tasks using physiological sensors. IEEE Trans. Intell. Transp. Syst. 2005, 6, 156–166. [Google Scholar] [CrossRef]
  13. Tsapeli, F.; Musolesi, M. Investigating causality in human behavior from smartphone sensor data: A quasi-experimental approach. EPJ Data Sci. 2015, 4, 24. [Google Scholar] [CrossRef]
  14. Fletcher, R.R.; Dobson, K.; Goodwin, M.S.; Eydgahi, H.; Wilder-Smith, O.; Fernholz, D.; Kuboyama, Y.; Hedman, E.B.; Poh, M.Z.; Picard, R.W. iCalm: Wearable sensor and network architecture for wirelessly communicating and logging autonomic activity. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 215–223. [Google Scholar] [CrossRef] [PubMed]
  15. App Paired With Sensor Measures Stress and Delivers Advice to Cope in Real Time. Available online: http://jacobsschool.ucsd.edu/news/news_releases/release.sfe?id=1526 (accessed on 23 August 2017).
  16. Birvinskas, D.; Jusas, V.; Martisius, I.; Damasevicius, R. EEG dataset reduction and feature extraction using discrete cosine transform. In Proceedings of the European Symposium on Computer Modeling and Simulation, Valetta, Malta, 14–16 November 2012; pp. 199–204. [Google Scholar]
  17. Frigui, H.; Nasraoui, O. Unsupervised learning of prototypes and attribute weights. Pattern Recognit. 2004, 37, 567–581. [Google Scholar] [CrossRef]
  18. Geng, X.; Liu, T.Y.; Qin, T.; Li, H. Feature selection for ranking. In Proceedings of the International ACM SIGIR Conference on Research and Development in Information Retrieval, Amsterdam, The Netherlands, 23–27 July 2007; pp. 407–414. [Google Scholar]
  19. Mcennis, D.; Mckay, C.; Fujinaga, I. JAudio: A feature extraction library. In Proceedings of the International Conference on Music Information Retrieval, London, UK, 11–15 September 2005. [Google Scholar]
  20. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef]
  21. Forman, G. An extensive empirical study of feature selection metrics for text classification. J. Mach. Learn. Res. 2003, 3, 1289–1305. [Google Scholar]
  22. Joachims, T. Text categorization with support vector machines: Learning with many relevant features. In Proceedings of the European Conference on Machine Learning, Chemnitz, Germany, 21–23 April 1998; pp. 137–142. [Google Scholar]
  23. Bogomolov, A.; Lepri, B.; Larcher, R.; Antonelli, F.; Pianesi, F.; Pentland, A. Energy consumption prediction using people dynamics derived from cellular network data. EPJ Data Sci. 2016, 5, 13. [Google Scholar] [CrossRef]
  24. Ribeiro, F.N.; Araújo, M.; Gonçalves, P.; André Gonçalves, M.; Benevenuto, F. SentiBench—A benchmark comparison of state-of-the-practice sentiment analysis methods. EPJ Data Sci. 2016, 5, 1–29. [Google Scholar] [CrossRef]
  25. Sarigöl, E.; Pfitzner, R.; Scholtes, I.; Garas, A.; Schweitzer, F. Predicting scientific success based on coauthorship networks. EPJ Data Sci. 2014, 3, 9. [Google Scholar] [CrossRef]
  26. Quinlan, J.R. C4.5: Programs for Machine Learning; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar]
  27. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
  28. Anderson, J.R.; Matessa, M. Explorations of an incremental, Bayesian algorithm for categorization. Mach. Learn. 1992, 9, 275–308. [Google Scholar] [CrossRef]
  29. Vapnik, V.N.; Vapnik, V. Statistical Learning Theory; Wiley: New York, NY, USA, 1998. [Google Scholar]
  30. Kim, K.; Chung, B.S.; Choi, Y.; Lee, S.; Jung, J.Y.; Park, J. Language independent semantic kernels for short-text classification. Expert Syst. Appl. 2014, 41, 735–743. [Google Scholar] [CrossRef]
  31. Stock Music and Sound Effects for Creative Projects. Available online: http://www.audiomicro.com/ (accessed on 23 August 2017).
  32. Adobe Audition. Available online: http://www.adobe.com/kr/products/audition.html (accessed on 23 August 2017).
Figure 1. Framework of real-time stress monitoring for children using wearable devices.
Figure 1. Framework of real-time stress monitoring for children using wearable devices.
Sensors 17 01936 g001
Figure 2. Overview of the learning-based stress detection algorithm.
Figure 2. Overview of the learning-based stress detection algorithm.
Sensors 17 01936 g002
Figure 3. Example of the (a) raw audio signal and (b) extracted features in terms of the two stress states of a child: normal and stress.
Figure 3. Example of the (a) raw audio signal and (b) extracted features in terms of the two stress states of a child: normal and stress.
Sensors 17 01936 g003
Figure 4. Prototype of (a) a stress detection program in server-side and (b) a monitoring application in parents-side.
Figure 4. Prototype of (a) a stress detection program in server-side and (b) a monitoring application in parents-side.
Sensors 17 01936 g004
Figure 5. (a) External view and (b) internal view of the prototype of the wearable device, where the audio signal and heart rate of a child are collected.
Figure 5. (a) External view and (b) internal view of the prototype of the wearable device, where the audio signal and heart rate of a child are collected.
Sensors 17 01936 g005
Figure 6. Scatter matrix of four selected features which appear most frequently among the feature selection results by the three methods (CHI, IG, and SVMW). HR: heart rate; Magnitude: magnitude spectrum overall average; MFCC: MFCC overall standard deviation; Power: power spectrum overall average.
Figure 6. Scatter matrix of four selected features which appear most frequently among the feature selection results by the three methods (CHI, IG, and SVMW). HR: heart rate; Magnitude: magnitude spectrum overall average; MFCC: MFCC overall standard deviation; Power: power spectrum overall average.
Sensors 17 01936 g006
Figure 7. Performance comparison results of the proposed framework in terms of the utilized data, audio only, heart rate only, and audio and heart rate together, and the adopted methods—decision tree (DT), naive Bayes (NB), SVM with radial kernel (SVM-R), and SVM with linear kernel (SVM-L)—according to the evaluation measures (left) accuracy; (middle) recall; and (right) precision.
Figure 7. Performance comparison results of the proposed framework in terms of the utilized data, audio only, heart rate only, and audio and heart rate together, and the adopted methods—decision tree (DT), naive Bayes (NB), SVM with radial kernel (SVM-R), and SVM with linear kernel (SVM-L)—according to the evaluation measures (left) accuracy; (middle) recall; and (right) precision.
Sensors 17 01936 g007
Figure 8. Performances of the proposed framework using the selected features according to the number of selected features and detection methods in terms of data utilized: (upper) audio only and (lower) audio and heart rate, and feature selection methods: (left) CHI , (middle) IG , and (right) SVMW.
Figure 8. Performances of the proposed framework using the selected features according to the number of selected features and detection methods in terms of data utilized: (upper) audio only and (lower) audio and heart rate, and feature selection methods: (left) CHI , (middle) IG , and (right) SVMW.
Sensors 17 01936 g008
Figure 9. Performances of stress detection using the four detection methods: (upper left) DT , (upper right) NB , (lower left) SVM-R , and (lower right) SVM-L, according to the noise level (5, 10, 15, 25, 40) and utilized data (audio only and both audio and heart rate).
Figure 9. Performances of stress detection using the four detection methods: (upper left) DT , (upper right) NB , (lower left) SVM-R , and (lower right) SVM-L, according to the noise level (5, 10, 15, 25, 40) and utilized data (audio only and both audio and heart rate).
Sensors 17 01936 g009
Table 1. Summary of previous research on stress detection in terms of their target, utilized method, and data.
Table 1. Summary of previous research on stress detection in terms of their target, utilized method, and data.
TargetMethodDataReferences
ChildMachine learningAudio signal[2]
[3]
[4]
[5]
AdultMachine learningElectrodermal data[8]
[9]
[10]
Heart rate[6]
[7]
Audio signal etc.[8]
[10]
Index-basedElectrodermal data etc.[11]
[12]
[13]
Table 2. Summary of the commercial stress detection devices in terms of their target, utilized data, and wearable design.
Table 2. Summary of the commercial stress detection devices in terms of their target, utilized data, and wearable design.
Proposed FrameworkiCalmParentGuardian
TargetInfants, childrenInfants, children, adultsADHD children
DataAudio signal, Heart rateTemperature
Motion
Electrodermal data
Blood volume pulse
Electrodermal data
Wearable designWristWrist, footWrist
Table 3. List of feature types provided in jAudio [19]. FFT: fast Fourier transform; MFCC: mel-frequency cepstral coefficient; RMS: root mean square.
Table 3. List of feature types provided in jAudio [19]. FFT: fast Fourier transform; MFCC: mel-frequency cepstral coefficient; RMS: root mean square.
Power SpectrumSpectral FluxFraction of Low-Energy Frames
Magnitude SpectrumPartial-Based Spectral FluxLinear Prediction Filter Coefficients
Magnitude Spectrum PeaksMethod of MomentsBeat Histogram
Spectral VariabilityArea Method of MomentsStrongest Beat
Spectral CentroidMFCCBeat Sum
Partial-Based Spectral CentroidArea Method of Moments of MFCCsStrength of Strongest Beat
Partial-Based Spectral SmoothnessZero CrossingsStrongest Frequency via Zero Crossings
CompactnessRMSStrongest Frequency via Spectral Centroid
Spectral Roll-off PointRelative Difference FunctionStrongest Frequency via FFT Maximum
Table 4. Confusion matrix of instances in terms of predicted and actual states: stress and normal.
Table 4. Confusion matrix of instances in terms of predicted and actual states: stress and normal.
Predicted State
StressNormal
Actual stateStressTrue positive (TP)False positive (FP)
NormalFalse negative (FN)True negative (TN)
Table 5. Top five selected features according to the three feature selection methods (chi-square, CHI; information gain, IG; and support vector machine wrapper, SVMW), according to their ranks.
Table 5. Top five selected features according to the three feature selection methods (chi-square, CHI; information gain, IG; and support vector machine wrapper, SVMW), according to their ranks.
RankCHIIGSVMW
1Heart rateHeart rateMFCC overall standard deviation
2MFCC overall standard deviationMFCC overall standard deviationSpectral flux overall standard deviation
3Magnitude spectrum overall averageMagnitude spectrum overall averageStrongest beat overall average
4Power spectrum overall averageMFCC overall averageMagnitude spectrum overall standard deviation
5MFCC overall averagePower spectrum overall averageCompactness overall average
Table 6. Results of t-test for accuracies obtained by performing 10-fold cross validation according to the utilized models and data.
Table 6. Results of t-test for accuracies obtained by performing 10-fold cross validation according to the utilized models and data.
DataModeltp-ValueMean Difference
Audio onlyNB and DT−30.870.00−14.89
NB and SVM-L−24.060.00−9.46
NB and SVM-R−33.860.00−16.60
DT and SVM-L8.230.005.42
DT and SVM-R−2.100.07−1.72
SVM-L and SVM-R−13.000.00−7.14
Heart rate onlyNB and DT−25.880.00−5.15
NB and SVM-L−3.250.01−0.35
NB and SVM-R87.200.008.32
DT and SVM-L26.420.004.81
DT and SVM-R71.240.0013.47
SVM-L and SVM-R76.020.008.67
Audio and heart rateNB and DT−26.650.00−13.28
NB and SVM-L−19.110.00−9.54
NB and SVM-R−31.430.00−15.80
DT and SVM-L4.760.003.74
DT and SVM-R−3.100.01−2.52
SVM-L and SVM-R−7.840.00−6.26

Share and Cite

MDPI and ACS Style

Choi, Y.; Jeon, Y.-M.; Wang, L.; Kim, K. A Biological Signal-Based Stress Monitoring Framework for Children Using Wearable Devices. Sensors 2017, 17, 1936. https://doi.org/10.3390/s17091936

AMA Style

Choi Y, Jeon Y-M, Wang L, Kim K. A Biological Signal-Based Stress Monitoring Framework for Children Using Wearable Devices. Sensors. 2017; 17(9):1936. https://doi.org/10.3390/s17091936

Chicago/Turabian Style

Choi, Yerim, Yu-Mi Jeon, Lin Wang, and Kwanho Kim. 2017. "A Biological Signal-Based Stress Monitoring Framework for Children Using Wearable Devices" Sensors 17, no. 9: 1936. https://doi.org/10.3390/s17091936

APA Style

Choi, Y., Jeon, Y. -M., Wang, L., & Kim, K. (2017). A Biological Signal-Based Stress Monitoring Framework for Children Using Wearable Devices. Sensors, 17(9), 1936. https://doi.org/10.3390/s17091936

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop