Next Article in Journal
Enhancing Reliability and Stability of BLE Mesh Networks: A Multipath Optimized AODV Approach
Previous Article in Journal
Hybrid Optimization Path Planning Method for AGV Based on KGWO
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning Method for Human Sleeping Pose Estimation with Millimeter Wave Radar

by
Zisheng Li
1,2,
Ken Chen
1,* and
Yaoqin Xie
1,*
1
Shenzhen lnstitute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
2
University of Chinese Academy of Sciences, Beijing 100190, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(18), 5900; https://doi.org/10.3390/s24185900
Submission received: 1 August 2024 / Revised: 3 September 2024 / Accepted: 9 September 2024 / Published: 11 September 2024
(This article belongs to the Section Radar Sensors)

Abstract

:
Recognizing sleep posture is crucial for the monitoring of people with sleeping disorders. Existing contact-based systems might interfere with sleeping, while camera-based systems may raise privacy concerns. In contrast, radar-based sensors offer a promising solution with high penetration ability and the capability to detect vital bio-signals. This study propose a deep learning method for human sleep pose recognition from signals acquired from single-antenna Frequency-Modulated Continuous Wave (FMCW) radar device. To capture both frequency features and sequential features, we introduce ResTCN, an effective architecture combining Residual blocks and Temporal Convolution Network (TCN) to recognize different sleeping postures, from augmented statistical motion features of the radar time series. We rigorously evaluated our method with an experimentally acquired data set which contains sleeping radar sequences from 16 volunteers. We report a classification accuracy of 82.74% on average, which outperforms the state-of-the-art methods.

1. Introduction

Sleep is a critical component of human health. Sleep breath disorders like obstructive sleep apnea (OSA) are linked to various health issues, with a prevalence of 9% to 38%, increasing with age [1]. Sleep apnea patients may suffer from breath restarts numerous times every night when they sleep [2], which lead to a lack of work productivity, decreased cognitive function, car accidents, and daytime sleepiness [3,4]. Moreover, OSA is known as a risk factor for several complications including hypertension, diabetes, cardiovascular disease, and heart failure in untreated patients, even an increase sudden death, and the care of concomitant neurological diseases, such as epilepsy, stroke, multiple sclerosis, and headaches also becomes burdensome [5,6,7]. OSA has caused an economic burden of over 150 billion annually in the USA [8], including lost workplace productivity, increased healthcare utilization, motor vehicle collisions and workplace accidents and injuries.
Various clinical studies have demonstrated correlation relationships between OSA and sleep postures [9]. One group [10] suggests that adopting a supine posture could notably increase the risks of OSA due to the prolapse of the tongue and the soft palate against the pharyngeal wall due to gravity, while the prone position improves the apnea hypopnea index (AHI) and sleep quality. For patients with positional obstructive sleep apnea syndrome (OSAS), the supine posture will lead to significantly higher AHI [11]. Some studies [12,13] demonstrated that the OSA that occurred in lateral position is less severe than supine position.
Apart from OSA, sleep postures can also be observed as valuable markers of several disease, and significantly affect health. Keeping the supine posture for a longer period can be related to deterioration in the condition of Parkinson’s patients [14]. Keeping the same sleep posture for a long time can lead to pressure ulcers in the elderly and post-surgery patients [15,16]. Improper sleep postures can increase the risk of sudden death [17,18].
Therefore, implementing sleep posture monitoring is a crucial role in enhancing overall health and well-being. A variety of sensors have been developed to monitor sleep postures and behaviors, including body pressure sensors, cameras (including depth cameras) and wearable devices [19]. However, there are drawbacks in those existing solutions. Pressure mats, measuring the pressure intensity distribution, have been employed to characterize sleep postural behavior and estimate sleep quality [20]. Proper placement of a pressure mat is crucial for accurate measurement. The movements of patients can cause shifts in the position of body; therefore, repositioning is necessary. Different materials of the mattress can also affect the natures of the pressure profile [21]. Video recordings using red–green–blue (RGB) or RGB–depth images can capture and enable the observation of individuals’ sleep postures directly [22,23]. Camera-based methods are sensitive to the varying light conditions, the presence of blanket coverage [24,25,26,27,28] and privacy concerns may be raised [29]. Wearable devices utilizing actigraphy or accelerometry can measure physical activity and infer motion or behavioral activities [30]. Therefore, spectrogram analysis of data from wearable devices can be employed to estimate sleep postures based on body segment movements [31]. However, these systems or sensors may be costly or disrupt sleep, discouraging practical use.
In contrast, radar-based sensors do not encounter such dilemmas, and exhibit high penetration ability, which is capable of detecting human movements even through walls [32], and can be used for vital sign detection [33]. Numerous studies have explored the application of radar-based systems for the detection of human vital signs, including heart rate (HR), breathing rate, and breathing disorders [34,35,36,37]. There were also applications utilizing radar for the monitoring of body movements [38,39] and hand gestures [40,41]. The aforementioned applications suggest that the radar acquired data contains rich information to measure the macro and micro movement of human body, and it may be promising to distinguish the feature difference among different sleep postures from radar signals. A radar-based system has been proposed for sleep posture recognition [42], which exhibits an ability to capture both time and frequency features, including the movement and direction of sleep postural transition movement. Lai et al. proposed vision transformer-based deep learning methods to classify the sleep postures under multiple radar configurations [43]. The results showed that the dual-radar configuration (side + head) with the Transformer model could achieve the best sleep posture prediction. A new method for on-bed state detection was proposed by capitalizing on chest localization to monitor on-bed presence, posture, motion, and turning, providing a holistic view of sleep behavior [44]. As far as we know, the application of radar devices for sleep posture recognition has not been sufficiently studied, and is still a challenging problem.
In this study, we propose a deep learning method for sleep posture classification from radar sequences. The radar sequences are preprocessed, and we extract the statistical motion features from frequency domain to train the classification network. We propose a two-stage deep learning model ResTCN for sleep posture classification. We first use a residual block-based network structure as feature extractor, and use TCN as classifier to handle the sequential pattern.

2. Related Work

There have been several groups conducting research for sleep posture classification with radar signals. Fan et al. [45] recruited nine male and one female participant, and recorded 2 min for each posture per participant. Their primary conclusion was that the accuracy of the logic-based approach and the CNN-based method was of the same level, but the CNN-based approach only required the data in a window size below 30 s, and a sliding window size below 5 s, while the logic approach necessitated post-recording analysis of the entire sleep session; therefore, the CNN-based approach offered a distinct advantage in terms of real-time applicability. Yue et al. [46] proposed a system BodyCompass using FMCW radio equipped with an antenna array for sleep posture detection. They collected 224 nights of data from 26 subjects. They proposed a four-layer dense network to process the Filtered Multi-path Profile, and reported the accuracy outperforms the traditional machine learning methods significantly. Liu et al. [47] collected radar signal from 10 volunteers, 20 s for each posture, with basic scenario, quilt covered scenario and static reflector interference scenario. They compared the DNN-based methods with machine learning methods, and report the effectiveness of the DNN structure. They also reported that the presence of the quilt could change the extracted feature, and the results for scenarios with the presence of the static reflectors is acceptable. Kiriazi et al. [48] collected data from 20 subjects with 2.4 GHz and 5.8 GHz devices for 90 s for each posture and each device per person. They reported that the accuracy of posture classification with dual frequency data set outperforms both single frequency data sets. Lai et al. [49] recruited 18 healthy young adults (12 males and 6 females) in their study, and collected 720 samples (18 participants × 4 postures × 10 repetitions). Each posture was recorded for 20 s. They evaluated traditional machine learning methods and deep learning methods, with different radar combinations, and concluded that the accuracy of the dual radar was significantly better than single radar. Lai et al. [43] further collected data with radar devices from three different positions. They recruited 30 volunteers and collected 1200 samples, each for 15 s. They first proposed to use Vision Transformer network structure for the posture classification problem. The Swin-transformer V2 structure with the head and side radar combination showed best accuracy. Lai et al. [50] also conducted experiment by recruiting 70 adults to collect data from devices with different position and orientation, and compared the performance of different deep networks under different combination of signals. Yao et al. [51] recruited 17 subjects and collected 1400 samples of 5 s for four postures. They proposed a encoder–decoder structure.
From the literature review, we can summarize some common points of the sleep posture classification experiments. The participants usually do not achieve sleep during experiments; instead, they are instructed to lie in different postures in a given order, and are required to stay stationary during data collection. The exclusion criterion of the participants are usually in common, including physical disability, obesity, pregnancy, or any cardio respiratory problems, and difficulties in maintaining or switching specific postures in bed. In most of the experiments, the number of volunteers varies from 10 to 20, which can satisfy the generalization requirement of data set for deep network training. The scales of data sets do not follow a common standard. If the time of data collection process is too short, the scale of the data set will not be sufficient to achieve satisfactory accuracy. On the other hand, a too long recording time can affect the comfort of the volunteers, making it difficult for the subjects to stay stationary during the data collection process. Therefore, most of the state-of-the-art experiments recorded several minutes for one posture per person.
We can see that, in the current state-of-the-art studies, some groups applied existing technologies directly to the radar collected data for sleep posture classification, such as the traditional machine learning methods like regression, SVM, Random Forest, etc. [49], and the existing deep learning models, like transformer-based network structures [43,49,50]. Some groups proposed specifically designed networks, which were based on convolutions [45,47,51] or full connections [46] only, but not utilizing the temporal information. However, the radar device captures time sequences, which contain important feature patterns for different sleep postures. Therefore, in this study, we propose a novel two-stage network structure by combining the convolution-based backbone ResNet for feature extraction and the Temporal Convolution Network (TCN) to capture long-time dependencies and sequential patterns in radar time series. We will show that the combination of the ResNet and the TCN will better recognize the features from the radar data, by achieving better accuracy over the state-of-the-art comparative methods.
Also, current methods usually tried to extract the information directly from the radar acquired data, while in this study, we propose a novel form of feature, the statistical motion features derived from the range-FFT images, to capture the micro movement of human body, such as respiration and heart-beats. We will show that the proposed statistical motion features can enhance the feature representation compared to the traditional radar data features, which essentially improve the accuracy.
Therefore, we summarize the main contributions of this manuscript:
1. We propose a novel two-stage deep learning model ResTCN for sleep posture classification with radar data. We first use ResNet as feature extractor, and use TCN as classifier to handle the sequential pattern.
2. We propose a novel form of feature, the statistical motion features to enhance the feature representation of the radar data.

3. Method

3.1. System Setup and Data Acquisition

The data used in this study is an experimentally acquired radar sequence set in our lab with an Infineon BGT60TR13C FMCW radar system. The working parameter of the radar is shown in Table 1. The radar system is mounted on the wall, positioned 0.65 m above the bed, with a pitch angle tilted approximately 45 degrees downward toward the bed to cover the interested part of the torso, as shown in Figure 1.
In this data set, we recruited 16 participants to acquire the radar sequences of four different sleep postures. Their average age was 30 (SD: 5.31, range 23–38). The mean weight and height were 172 cm (SD: 5.99 cm, range 160–180 cm) and 67.3 kg (SD: 9.83 kg, range 53–90 kg), respectively. The exclusion criteria included physical disability, obesity, pregnancy, or any cardio-respiratory problems, in addition to participants with difficulties in maintaining or switching specific postures in bed.
The participants were commenced to lie on the middle of the bed, supported by a pillow. Then, they were guided to lie in various stationary postures, as shown in Figure 2, in the specified order: (1) supine, (2) right lateral, (3) left lateral, and (4) prone, each for 10 min for data acquisition. The overall acquisition time is 640 min (16 participants × 4 postures × 10 min). In this study, we mainly focus on the recognition of the feature differences of 3 gesture categories: supine, lateral and prone; therefore, we combine the left and right lateral data into one category. Compared with the literature reviewed in the Related Work section, we can see that the design of the experiments in this study does not contradict with the principles of the existing state-of-the-art works.
Informed consent was obtained from all the volunteers. The experimental procedure for data acquisition and utilization in this research has been approved by the ethics committee of Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences. The committee reference number is SIAT-IRB-240615-H0890.

3.2. Data Processing

3.2.1. Preprocessing

Frequency Modulated Continuous Wave (FMCW) is a specialized millimeter-wave technology. The transmitting (TX) antenna of the FMCW radar transmits a signal called a “chirp”, which is a sinusoid whose frequency increases lineally with time. Once the chirp arrives at the focused object, the signal will be reflected and received at the receiving (RX) antenna. The RX signal and TX signal are mixed into a resulting signal, “IF signal”. For an object at a distance D from the radar, the discrete IF signal X I F can be represented as
X I F ( m ) = A sin ( 2 π f 0 T s m + ϕ 0 ) , m = 1 , 2 , M
where X I F is the mixed signal, f 0 = s 2 D c , and ϕ 0 = 4 π D λ . D is the distance between object and radar, s is the slope of chirp, and λ is the wavelength of radar, T s is the sampling interval, and M is the number of fast time indices. We accumulate the discrete IF signal of each chirp over time to form a M × N matrix R M N , where R M N = [ x 1 , x 2 , , x N ] , and x i = [ x I F i ( 1 ) , x I F i ( 2 ) , , x I F i ( M ) ] T is the IF signal of chirp with index i. In our case, the slow time index N is 640 (chirp rate×frame rate) for 1 s, the fast time index M = 128 is the sampling number of the IF signal of each chirp.
The raw radar signal R m n contains a DC component, and also static clutter representing uninterested objects and obstacles in the environment. To mitigate the impact of the DC noise and static clutter, the mean subtraction method is applied to the raw radar signal in two consecutive steps. The DC suppression can be expressed by Equation (2), and the static clutter suppression can be expressed by Equation (3).
R ¯ m n = R m n 1 N i = 0 N 1 R m i
Y m n = R ¯ m n 1 M i = 0 M 1 R ¯ i n

3.2.2. Range Fourier Transform (Range-FFT)

We consider each chirp as a quasi-static scenario, where the relative velocity within one chirp is 0, and the relative distance R remains constant. We apply Fourier Transform to each chirp, and obtain the spectrum where each peak represents an obstacle. This process is defined as the Range Fourier Transform (range-FFT). The spectra are stored in a K × N array, denoted as F, where
F k n = m = 0 M 1 Y m n e j 2 π m k M , k = 1 , 2 , K , n = 1 , 2 , N
where K is the frequency index, and N is the slow time index. In our case, K equals to 64, and N equals to 640 for 1 s.

3.2.3. Statistical Motion Features

In order to capture the micro movement of human body, such as respiration and heartbeats, we propose to extract the statistical motion features from the range-FFT spectrogram to train the model. A sliding window with a width of W along the slow time indexes is applied to the spectrogram, with a step of S. For an arbitrary window with index i, the standard deviation of the spectrogram for an arbitrary range bin with index k is calculated as
s t d k , i = 1 W t = i × S i × S + W 1 ( F k t F k i ¯ ) 2
F k i ¯ = 1 W t = i × S i × S + W 1 F k t , k = 1 , 2 , K , i = 1 , 2 , N W
where K is the frequency index, and NW is the total number of sliding windows. In our case, K = 64, and we choose W to be 320 (10 frames × 32 chirps), which is 0.5 s, S to be 64 (2 frames × 32 chirps), which is 0.1 s. The sliding window covers a slow time index range of 6 s, which means N W = 60 in our case.
To cover only the interested regions of the human body, we select the data inside a bin window with a size of K1 × NW. The range resolution of our radar device is 4 cm; therefore, we optimally choose K 1 = 40 , which covers a range of 1.6 m. The optimal position of the bin window is decided by maximize the sum of the standard deviations within the bin window.
The standard deviations within the optimal bin window are accumulated to form a K1 × NW feature image X, where X k , j = std k , j , K 1 = 40 , and N W = 60 for our case. Typical examples of the statistical motion feature images of different sleep postures are shown in Figure 3.

3.3. Data Augmentation

In order to improve the generality of the model, data augmentation techniques, including time shift, range shift, and mix-up were applied.
For time shift process, each feature image was moved along the slow time index by a random offset ranged in [−1, −0.5] or [0.5, 1] seconds. The time shift reduces the effect of data truncation, and improves the coverage of complete respiratory cycles. For range shift process, each feature image was moved along the range bin index by a random offset ranged in [−10, −5] or [5, 10]. The range shift aid the model to reduce the effect of the lying position difference.
The mix-up technique [52] was introduced to improve the generalization of empirical risk minimization (ERM). Two feature images from different categories and the corresponding labels were linearly combined to generate a virtual feature image, based on a random weight picked according to the beta distribution, as Formula (7), where λ Beta ( α , α ) and α < 1 . Typical examples of the augmented samples are shown in Figure 4.
X ¯ = λ x i + ( 1 λ ) x j Y ¯ = λ y i + ( 1 λ ) y j

3.4. ResTCN

In this section, we will describe the proposed scheme for sleep posture recognition in detail. The overall process of the scheme is shown in Figure 5. We first used a backbone network based on residual blocks to extract meaningful features from the augmented feature set obtained in Section 3.3. The extracted features were feed into the Classification Network for sleep posture recognition.
We proposed to use the ResNet [53] network as the Feature Extraction Network. With its skip connections and residual blocks, ResNet allowed the training of deeper network structures, and showed hierarchical feature extraction capability [54]. ResNet has demonstrated superior performance in capturing complex features and patterns in various image recognition tasks [55,56,57]. The structure and parameters of the network in this work are shown in Figure 5 in detail. The feature image is cloned to the three channels of the input image. A feature vector with a length of 1024 is obtained for classification.
We proposed to use Temporal Convolution Network (TCN) [58] as the Classification Network. TCN is well-suited for tasks requiring temporal modeling due to its parallelizable dilated convolutions and global receptive field, and shows promising results in capturing long-time dependencies and sequential patterns in time series data [59,60]. The detail structure of the network applied in our case is shown in Figure 5.
The proposed deep learning method was implemented with PyTorch1.7 on a NVIDIA 4090 GPU device. We chose the AdamW optimizer to minimize the cross entropy as Equation (8). The initial learning rate is set to 0.005, with a L2 regularization coefficient of 0.001. The learning rate was scaled down 10 times every 10 training epochs. The model was trained for 100 epochs.
J p o ( θ ) = 1 M i = 0 M y p ( i ) l o g y ^ p ( i ) ( θ )

4. Results

In this section, we evaluate the performance of the proposed sleep posture classification scheme, and compare the proposed method with state-of-the-art methods.

Sleep Posture Classification Results

To test the performance of the proposed method, We randomly chose six volunteers as the test set. From the remaining 10 volunteers, samples of eight randomly chosen volunteers were used to train the network, while the remaining two volunteers were used as the validation set. We used the following metrics to evaluate the model accuracy performance:
Accuracy = T P + T N T P + T N + F P + F N
Precision = T P T P + F P
Recall = T P T P + F N
F 1 = 2 × Precision × Recall Precision + Recall
We compared our proposed method with the following state-of-the-art methods:
1. We applied the SVM-based method [61] to our augmented feature image set. The features of the image are extracted by PCA, and the classification is achieved by SVM with Radial Basis Function (RBF) kernel. Grid-search cross-validation is applied to find the best parameters of SVM. The search sets of the parameters of the SVM, C and gamma are [0.01, 0.1, 1, 10, 100] and [0.01, 0.1, 1, 10, 100], respectively.
2. ShuffleNet [62] is a convolutional neural network architecture widely employed for classification tasks. ShuffleNet features a unique design comprising shuffle units, which facilitate efficient information exchange across layers. It is characterized by its lightweight structure, making it suitable for deployment on resource-constrained devices [63].
3. DenseNet [64], a distinctive convolutional neural network architecture, introduces a novel connectivity pattern that fosters extensive information exchange among network layers. Unlike traditional architectures, where each layer is connected only to its subsequent layer, DenseNet incorporates dense connections, enabling direct communication between all layers within a block. This dense connectivity facilitates feature reuse, and enhances gradient flow during training, effectively addressing the vanishing gradient problem. By encouraging feature propagation and fostering feature reuse, DenseNet achieves remarkable parameter efficiency while maintaining high model accuracy. With its unique design, DenseNet has emerged as a powerful tool for various computer vision tasks, demonstrating superior performance in image classification, object detection, and semantic segmentation [65,66].
4. Vision transformer (VIT) [67] is a groundbreaking deep learning model designed specifically for image classification tasks. Unlike CNN, VIT employs a self-attention mechanism to capture global information within data; therefore, it has achieved promising results in various fields [68,69]. In our case, we use VIT to analyze the the pattern of the statistical feature images of the spectrogram.
5. Swin transformer V2 [70] is an advanced deep learning model that builds upon the traditional Transformer architecture by introducing a hierarchical feature representation and shifted windows, enabling it to capture both local and global context effectively. Unlike VIT, which processes images as a whole, Swin Transformer V2 divides the image into non-overlapping windows and performs self-attention within each window, followed by shifting the windows to capture cross-window connections. Swin transformer V2 and its variants have received growing interest in image processing realms. In a recent study, Lai et al. [43] first proposed to use the vision transformer network for sleep posture classification with radar image features. They reported that the Swin transformer V2 structure with the head and side radar combination showed best accuracy. For the head radar scenario, which is the case of this manuscript, the Swin transformer V2 also outperforms the comparative methods.
All of the comparative methods were trained and evaluated under the same hardware and software platform configurations of the proposed method, with the same data processing and augmentation strategies. The comparison of accuracy performance is summarized in Table 2, and the typical confusion matrices are depicted in Figure 6. We can see that the proposed method ResTCN outperforms all other comparative methods in terms of accuracy, F1-score, Precision, and Recall. The combination of residual blocks and TCN allows the proposed network to recognize both the spatial patterns and the sequential dependencies, resulting in superior classification accuracy over the convolution-based comparative methods. SVM is a powerful method for certain classification tasks, but in our case, the ability to capture the frequency domain features is limited. VIT and Swin transformer V2 may be powerful in natural image processing, but it requires a very large scale of data for training to obtain optimal feature recognition abilities.

5. Discussion

5.1. Ablation Experiment

5.1.1. Model Architecture

In this section, we show the necessity of the combination of the residual blocks and TCN structures. We conducted a comparison experiment by applying only ResNet and TCN, respectively. When applying the ResNet, the augmented statistical motion feature images are fed into the network, and a FC layer is used for classification. When applying TCN, the feature images are flatten, and the output of the last time index is connected to a FC layer for classification. We conducted three experiments, namely ShuffleNet + TCN, DenseNet + TCN, and ResNet + LSTM for performance comparison among different combinations of backbone network structures. The results are summarized in Table 3; we can see that the ResTCN model consistently outperforms the other combinations across all evaluation metrics, including accuracy, precision, recall, and F1 score. Therefore, we can conclude that the proposed ResTCN structure show better performance. The fusion of the residual blocks and TCN structure can better recognize the features from the spectrogram-based images.

5.1.2. Statistical Motion Feature Extraction

We conducted ablation experiments to evaluate the effect of the motion feature extraction in our proposed method by training and testing the proposed network with the range-FFT images and the motion feature images. The results are detailed in Table 4. Notable improvement in classification performance is observed, highlighting the crucial role of statistical motion features in capturing the difference of micro human movement patterns among different sleep postures.

5.1.3. Data Augmentation

We compare the classification results with and without data augmentation. The results are summarized in Table 5. We can see that with the data augmentation the performance is improved. This improvement demonstrates the effectiveness of data augmentation in enriching the diversity of the training data set and improving the generalization of the model.

5.2. Parameter Selection

To determine the most suitable window size along the range axis, we conducted an investigation to compare the classification accuracy metrics under different window size. The range of window size values explored in this experiment included 30, 40 (optimal), 50, and 60 range bins. The comparison details are shown in Table 6. Compared to other window size, 40 range bins ensures a more comprehensive coverage of the interested part of the human body for respiratory feature extraction, and can effectively mitigates radar multi-path effects, minimizing the environmental impact on features.
We also study the choice of range covered by the sliding window along the slow time axis. The range of sliding window included 2 s, 6 s, and 10 s. The comparison details are shown in Table 7. Compared to the 2 s sliding window, the 6 s window shows significant improvements in all accuracy metrics, while the improvement of accuracy metrics from 6 s to 10 s is minor with greater computation load and calculation time. A smaller window size along the slow time axis also reduces the time delay for detection of posture change. Therefore, we can conclude that the 6 s window size is the optimal choice, which achieves the balance of accuracy and efficiency.

5.3. Limitations

There are several limitations in this study. The scale of the data set is not sufficiently large, which keeps the vision transformer showing superior ability in image classification tasks. In our future work, we will further collect data to fine tune the vision transformer network to recognize the features among different sleep postures.
Also, in the data we acquired in this study, we currently only include the data from healthy subjects, but not the data from subjects with OSA. Identifying OSA is a labor-intensive task, and usually requires specialized facilities and experts with qualifications. In our future work, we will work with our collaborative facilities for further data acquisition and analysis from OSA patients. Also, we will include cases with the presence of blanket coverage in the future work.

6. Conclusions

This study proposed the ResTCN, an effective architecture that leverages ResNet and TCN backbone structures to classify various human sleep postures acquired from FMCW radar device. The statistical motion features are extracted from range-FFT spectrograms, and the data augmentation techniques are introduced to address the over-fitting issue. We rigorously test the proposed network with an experimentally acquired data set. The classification results for the three sleep postures show promising accuracy, and surpass the state-of-the-art methods. In conclusion, the proposed method is promising in the application of non-contact human sleep monitoring.

Author Contributions

Z.L., K.C. and Y.X. wrote and edited the paper. Y.X. and K.C. conceptualized the study. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by grants from the National Natural Science Foundation of China (U20A20373), National Key Research and Develop Program of China (2022YFC2409000), Shenzhen Science and Technology Program (KQTD20180411185028798).

Institutional Review Board Statement

The experimental procedure for data acquisition and utilization in this research has been approved by the ethics committee of Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences. The committee reference number is SIAT-IRB-240615-H0890.

Informed Consent Statement

Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

Data are available upon request.

Acknowledgments

The authors would like to thank Yi Huang, Wei Hu and Zhenhua Hu from Shenzhen HUAYI Medical Technologies Co., Ltd. for their technical assistance.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Senaratna, C.V.; Perret, J.L.; Lodge, C.J.; Lowe, A.J.; Campbell, B.E.; Matheson, M.C.; Hamilton, G.S.; Dharmage, S.C. Prevalence of obstructive sleep apnea in the general population: A systematic review. Sleep Med. Rev. 2017, 34, 70–81. [Google Scholar] [CrossRef] [PubMed]
  2. Strollo, P.J., Jr.; Rogers, R.M. Obstructive sleep apnea. N. Engl. J. Med. 1996, 334, 99–104. [Google Scholar] [CrossRef] [PubMed]
  3. Franklin, K.A.; Lindberg, E. Obstructive sleep apnea is a common disorder in the population—A review on the epidemiology of sleep apnea. J. Thorac. Dis. 2015, 7, 1311. [Google Scholar] [PubMed]
  4. Chervin, R.D. Sleepiness, fatigue, tiredness, and lack of energy in obstructive sleep apnea. Chest 2000, 118, 372–379. [Google Scholar] [CrossRef] [PubMed]
  5. Drager, L.F.; McEvoy, R.D.; Barbe, F.; Lorenzi-Filho, G.; Redline, S. Sleep apnea and cardiovascular disease: Lessons from recent trials and need for team science. Circulation 2017, 136, 1840–1850. [Google Scholar] [CrossRef]
  6. McDermott, M.; Brown, D.L.; Chervin, R.D. Sleep disorders and the risk of stroke. Expert Rev. Neurother. 2018, 18, 523–531. [Google Scholar] [CrossRef]
  7. Faria, A.; Allen, A.H.; Fox, N.; Ayas, N.; Laher, I. The public health burden of obstructive sleep apnea. Sleep Sci. 2021, 14, 257–265. [Google Scholar]
  8. Wickwire, E.M. Value-based sleep and breathing: Health economic aspects of obstructive sleep apnea. Fac. Rev. 2021, 10, 40. [Google Scholar] [CrossRef]
  9. Menon, A.; Kumar, M. Influence of body position on severity of obstructive sleep apnea: A systematic review. Int. Sch. Res. Not. 2013, 2013, 670381. [Google Scholar] [CrossRef]
  10. Matsuzawa, Y.; Hayashi, S.; Yamaguchi, S.; Yoshikawa, S.; Okada, K.; Fujimoto, K.; Sekiguchi, M. Effect of prone position on apnea severity in obstructive sleep apnea. Intern. Med. 1995, 34, 1190–1193. [Google Scholar] [CrossRef]
  11. Richard, W.; Kox, D.; den Herder, C.; Laman, M.; van Tinteren, H.; de Vries, N. The role of sleep position in obstructive sleep apnea syndrome. Eur. Arch. Oto-Rhino-Laryngol. Head Neck 2006, 263, 946–950. [Google Scholar] [CrossRef] [PubMed]
  12. Oksenberg, A.; Khamaysi, I.; Silverberg, D.S.; Tarasiuk, A. Association of body position with severity of apneic events in patients with severe nonpositional obstructive sleep apnea. Chest 2000, 118, 1018–1024. [Google Scholar] [CrossRef] [PubMed]
  13. Oksenberg, A.; Silverberg, D.S.; Arons, E.; Radwan, H. The sleep supine position has a major effect on optimal nasal continuous positive airway pressure: Relationship with rapid eye movements and non-rapid eye movements sleep, body mass index, respiratory disturbance index, and age. Chest 1999, 116, 1000–1006. [Google Scholar] [CrossRef] [PubMed]
  14. Uchino, K.; Shiraishi, M.; Tanaka, K.; Akamatsu, M.; Hasegawa, Y. Impact of inability to turn in bed assessed by a wearable three-axis accelerometer on patients with Parkinson’s disease. PLoS ONE 2017, 12, e0187616. [Google Scholar] [CrossRef]
  15. Gorecki, C.; Closs, S.J.; Nixon, J.; Briggs, M. Patient-reported pressure ulcer pain: A mixed-methods systematic review. J. Pain Symptom Manag. 2011, 42, 443–459. [Google Scholar] [CrossRef]
  16. Reddy, M.; Gill, S.S.; Rochon, P.A. Preventing pressure ulcers: A systematic review. JAMA 2006, 296, 974–984. [Google Scholar] [CrossRef]
  17. Liebenthal, J.A.; Wu, S.; Rose, S.; Ebersole, J.S.; Tao, J.X. Association of prone position with sudden unexpected death in epilepsy. Neurology 2015, 84, 703–709. [Google Scholar] [CrossRef]
  18. Goldberg, N.; Rodriguez-Prado, Y.; Tillery, R.; Chua, C. Sudden infant death syndrome: A review. Pediatr. Ann. 2018, 47, e118–e123. [Google Scholar] [CrossRef]
  19. Li, X.; Gong, Y.; Jin, X.; Shang, P. Sleep posture recognition based on machine learning: A systematic review. Pervasive Mob. Comput. 2023, 90, 101752. [Google Scholar] [CrossRef]
  20. Enayati, M.; Skubic, M.; Keller, J.M.; Popescu, M.; Farahani, N.Z. Sleep posture classification using bed sensor data and neural networks. In Proceedings of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu, HI, USA, 18–21 July 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 461–465. [Google Scholar]
  21. Low, F.Z.; Chua, M.C.H.; Lim, P.Y.; Yeow, C.H. Effects of mattress material on body pressure profiles in different sleeping postures. J. Chiropr. Med. 2017, 16, 1–9. [Google Scholar] [CrossRef]
  22. Tam, A.Y.C.; Zha, L.W.; So, B.P.H.; Lai, D.K.H.; Mao, Y.J.; Lim, H.J.; Wong, D.W.C.; Cheung, J.C.W. Depth-Camera-Based Under-Blanket Sleep Posture Classification Using Anatomical Landmark-Guided Deep Learning Model. Int. J. Environ. Res. Public Health 2022, 19, 13491. [Google Scholar] [CrossRef] [PubMed]
  23. Masek, M.; Lam, C.P.; Tranthim-Fryer, C.; Jansen, B.; Baptist, K. Sleep monitor: A tool for monitoring and categorical scoring of lying position using 3D camera data. SoftwareX 2018, 7, 341–346. [Google Scholar] [CrossRef]
  24. Akbarian, S.; Delfi, G.; Zhu, K.; Yadollahi, A.; Taati, B. Automated non-contact detection of head and body positions during sleep. IEEE Access 2019, 7, 72826–72834. [Google Scholar] [CrossRef]
  25. Liu, S.; Ostadabbas, S. A vision-based system for in-bed posture tracking. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 1373–1382. [Google Scholar]
  26. Torres, C.; Fried, J.C.; Rose, K.; Manjunath, B.S. A multiview multimodal system for monitoring patient sleep. IEEE Trans. Multimed. 2018, 20, 3057–3068. [Google Scholar] [CrossRef]
  27. Liu, S.; Ostadabbas, S. Seeing under the cover: A physics guided learning approach for in-bed pose estimation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 236–245. [Google Scholar]
  28. Vandersmissen, B.; Knudde, N.; Jalalvand, A.; Couckuyt, I.; Bourdoux, A.; De Neve, W.; Dhaene, T. Indoor person identification using a low-power FMCW radar. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3941–3952. [Google Scholar] [CrossRef]
  29. Demiris, G.; Hensel, B.K.; Skubic, M.; Rantz, M. Senior residents’ perceived need of and preferences for “smart home” sensor technologies. Int. J. Technol. Assess. Health Care 2008, 24, 120–124. [Google Scholar] [CrossRef]
  30. Cheung, J.C.W.; So, B.P.H.; Ho, K.H.M.; Wong, D.W.C.; Lam, A.H.F.; Cheung, D.S.K. Wrist accelerometry for monitoring dementia agitation behaviour in clinical settings: A scoping review. Front. Psychiatry 2022, 13, 913213. [Google Scholar] [CrossRef]
  31. Eyobu, O.S.; Kim, Y.W.; Cha, D.; Han, D.S. A real-time sleeping position recognition system using IMU sensor motion data. In Proceedings of the 2018 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 12–14 January 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–2. [Google Scholar]
  32. Li, J.; Zeng, Z.; Sun, J.; Liu, F. Through-wall detection of human being’s movement by UWB radar. IEEE Geosci. Remote Sens. Lett. 2012, 9, 1079–1083. [Google Scholar] [CrossRef]
  33. Wang, K.; Zeng, Z.; Sun, J. Through-wall detection of the moving paths and vital signs of human beings. IEEE Geosci. Remote Sens. Lett. 2018, 16, 717–721. [Google Scholar] [CrossRef]
  34. Shyu, K.K.; Chiu, L.J.; Lee, P.L.; Tung, T.H.; Yang, S.H. Detection of breathing and heart rates in UWB radar sensor data using FVPIEF-based two-layer EEMD. IEEE Sens. J. 2018, 19, 774–784. [Google Scholar] [CrossRef]
  35. Shen, H.; Xu, C.; Yang, Y.; Sun, L.; Cai, Z.; Bai, L.; Clancy, E.; Huang, X. Respiration and heartbeat rates measurement based on autocorrelation using IR-UWB radar. IEEE Trans. Circuits Syst. II Express Briefs 2018, 65, 1470–1474. [Google Scholar] [CrossRef]
  36. Van, N.T.P.; Tang, L.; Singh, A.; Minh, N.D.; Mukhopadhyay, S.C.; Hasan, S.F. Self-identification respiratory disorder based on continuous wave radar sensor system. IEEE Access 2019, 7, 40019–40026. [Google Scholar] [CrossRef]
  37. Schires, E.; Georgiou, P.; Lande, T.S. Vital sign monitoring through the back using an UWB impulse radar with body coupled antennas. IEEE Trans. Biomed. Circuits Syst. 2018, 12, 292–302. [Google Scholar] [CrossRef] [PubMed]
  38. Otero, M. Application of a continuous wave radar for human gait recognition. In Proceedings of the Signal Processing, Sensor Fusion, and Target Recognition XIV, Orlando, FL, USA, 28–30 March 2005; SPIE: San Jose, CA, USA, 2005; Volume 5809, pp. 538–548. [Google Scholar]
  39. Rahman, T.; Adams, A.T.; Ravichandran, R.V.; Zhang, M.; Patel, S.N.; Kientz, J.A.; Choudhury, T. Dopplesleep: A contactless unobtrusive sleep sensing system using short-range doppler radar. In Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Osaka, Japan, 9–11 September 2015; pp. 39–50. [Google Scholar]
  40. Kim, Y.; Toomajian, B. Application of Doppler radar for the recognition of hand gestures using optimized deep convolutional neural networks. In Proceedings of the 2017 11th European Conference on Antennas and Propagation (EUCAP), Paris, France, 19–24 March 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1258–1260. [Google Scholar]
  41. Zhang, X.; Wu, Q.; Zhao, D. Dynamic hand gesture recognition using FMCW radar sensor for driving assistance. In Proceedings of the 2018 10th International Conference on Wireless Communications and Signal Processing (WCSP), Hangzhou, China, 18–20 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–6. [Google Scholar]
  42. Piriyajitakonkij, M.; Warin, P.; Lakhan, P.; Leelaarporn, P.; Kumchaiseemak, N.; Suwajanakorn, S.; Pianpanit, T.; Niparnan, N.; Mukhopadhyay, S.C.; Wilaiprasitporn, T. SleepPoseNet: Multi-view learning for sleep postural transition recognition using UWB. IEEE J. Biomed. Health Inform. 2020, 25, 1305–1314. [Google Scholar] [CrossRef] [PubMed]
  43. Lai, D.K.H.; Yu, Z.H.; Leung, T.Y.N.; Lim, H.J.; Tam, A.Y.C.; So, B.P.H.; Mao, Y.J.; Cheung, D.S.K.; Wong, D.W.C.; Cheung, J.C.W. Vision Transformers (ViT) for blanket-penetrating sleep posture recognition using a triple ultra-wideband (UWB) radar system. Sensors 2023, 23, 2475. [Google Scholar] [CrossRef]
  44. Fan, M.; Brahim, W.; Zhang, X.; Ma, J.; Ma, M.; Qi, A.; Luo, Y.; Qi, Y. Leveraging FMCW Radar for Monitoring On-bed States of Human Presence, Posture, and Motion. In Proceedings of the 2023 IEEE Smart World Congress (SWC), Portsmouth, UK, 28–31 August 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–8. [Google Scholar]
  45. Fan, M.; Ma, J.; Ma, M.; Qi, A. FMCW Radar-based Sleep Posture Monitoring Through Logic and Deep Learning Methods. In Proceedings of the 2023 IEEE International Conferences on Internet of Things (iThings) and IEEE Green Computing & Communications (GreenCom) and IEEE Cyber, Physical & Social Computing (CPSCom) and IEEE Smart Data (SmartData) and IEEE Congress on Cybermatics (Cybermatics), Danzhou, China, 17–21 December 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 586–591. [Google Scholar]
  46. Yue, S.; Yang, Y.; Wang, H.; Rahul, H.; Katabi, D. BodyCompass: Monitoring sleep posture with wireless signals. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2020, 4, 1–25. [Google Scholar] [CrossRef]
  47. Liu, X.; Jiang, W.; Chen, S.; Xie, X.; Liu, H.; Cai, Q.; Tong, X.; Shi, T.; Qu, W. PosMonitor: Fine-grained sleep posture recognition with mmWave radar. IEEE Internet Things J. 2023, 11, 11175–11189. [Google Scholar] [CrossRef]
  48. Kiriazi, J.E.; Islam, S.M.; Borić-Lubecke, O.; Lubecke, V.M. Sleep posture recognition with a dual-frequency cardiopulmonary Doppler radar. IEEE Access 2021, 9, 36181–36194. [Google Scholar] [CrossRef]
  49. Lai, D.K.H.; Zha, L.W.; Leung, T.Y.N.; Tam, A.Y.C.; So, B.P.H.; Lim, H.J.; Cheung, D.S.K.; Wong, D.W.C.; Cheung, J.C.W. Dual ultra-wideband (UWB) radar-based sleep posture recognition system: Towards ubiquitous sleep monitoring. Eng. Regen. 2023, 4, 36–43. [Google Scholar] [CrossRef]
  50. Lai, D.K.H.; Tam, A.Y.C.; So, B.P.H.; Chan, A.C.H.; Zha, L.W.; Wong, D.W.C.; Cheung, J.C.W. Deciphering Optimal Radar Ensemble for Advancing Sleep Posture Prediction through Multiview Convolutional Neural Network (MVCNN) Approach Using Spatial Radio Echo Map (SREM). Sensors 2024, 24, 5016. [Google Scholar] [CrossRef]
  51. Yao, Y.; Xu, L.; Xia, P.; Zhang, H.; Du, L.; Chen, X.; Fang, Z. Highly Generalized Sleep Posture Recognition Using FMCW Radar. In Proceedings of the 2023 IEEE 19th International Conference on Body Sensor Networks (BSN), Boston, MA, USA, 9–11 October 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–4. [Google Scholar]
  52. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  53. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  54. Allen-Zhu, Z.; Li, Y. What can resnet learn efficiently, going beyond kernels? In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
  55. Wu, Z.; Shen, C.; Van Den Hengel, A. Wider or deeper: Revisiting the resnet model for visual recognition. Pattern Recognit. 2019, 90, 119–133. [Google Scholar] [CrossRef]
  56. Jung, H.; Choi, M.K.; Jung, J.; Lee, J.H.; Kwon, S.; Young Jung, W. ResNet-based vehicle classification and localization in traffic surveillance systems. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 61–67. [Google Scholar]
  57. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  58. Hewage, P.; Behera, A.; Trovati, M.; Pereira, E.; Ghahremani, M.; Palmieri, F.; Liu, Y. Temporal convolutional neural (TCN) network for an effective weather forecasting using time-series data from the local weather station. Soft Comput. 2020, 24, 16453–16482. [Google Scholar] [CrossRef]
  59. Fan, J.; Zhang, K.; Huang, Y.; Zhu, Y.; Chen, B. Parallel spatio-temporal attention-based TCN for multivariate time series prediction. Neural Comput. Appl. 2023, 35, 13109–13118. [Google Scholar] [CrossRef]
  60. Wang, X.; Liu, H.; Du, J.; Dong, X.; Yang, Z. A long-term multivariate time series forecasting network combining series decomposition and convolutional neural networks. Appl. Soft Comput. 2023, 139, 110214. [Google Scholar] [CrossRef]
  61. Burges, C.J. A tutorial on support vector machines for pattern recognition. Data Min. Knowl. Discov. 1998, 2, 121–167. [Google Scholar] [CrossRef]
  62. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar]
  63. Hou, W.; Guo, H.; Yan, B.; Xu, Z.; Yuan, C.; Mao, Y. Tool wear state recognition under imbalanced data based on WGAN-GP and lightweight neural network ShuffleNet. J. Mech. Sci. Technol. 2022, 36, 4993–5009. [Google Scholar] [CrossRef]
  64. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  65. Iandola, F.; Moskewicz, M.; Karayev, S.; Girshick, R.; Darrell, T.; Keutzer, K. Densenet: Implementing efficient convnet descriptor pyramids. arXiv 2014, arXiv:1404.1869. [Google Scholar]
  66. Zhang, K.; Guo, Y.; Wang, X.; Yuan, J.; Ding, Q. Multiple feature reweight densenet for image classification. IEEE Access 2019, 7, 9872–9880. [Google Scholar] [CrossRef]
  67. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  68. Yuan, L.; Chen, Y.; Wang, T.; Yu, W.; Shi, Y.; Jiang, Z.H.; Tay, F.E.; Feng, J.; Yan, S. Tokens-to-token vit: Training vision transformers from scratch on imagenet. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 558–567. [Google Scholar]
  69. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A survey on vision transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 87–110. [Google Scholar] [CrossRef] [PubMed]
  70. Liu, Z.; Hu, H.; Lin, Y.; Yao, Z.; Xie, Z.; Wei, Y.; Ning, J.; Cao, Y.; Zhang, Z.; Dong, L.; et al. Swin transformer v2: Scaling up capacity and resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 12009–12019. [Google Scholar]
Figure 1. Experimental bedroom environment with a FMCW radar devices mounted on the wall.
Figure 1. Experimental bedroom environment with a FMCW radar devices mounted on the wall.
Sensors 24 05900 g001
Figure 2. The four postures: (a) supine, (b) left-side lying, (c) right-side lying, and (d) prone.
Figure 2. The four postures: (a) supine, (b) left-side lying, (c) right-side lying, and (d) prone.
Sensors 24 05900 g002
Figure 3. Typical signal examples of different postures. (a1a4) Typical chirp time sequences of 4 postures: supine, left-side lying, right-side lying, and prone. (b1b4) Range-FFT images of the 4 postures (1 frame, 32 chirps). (c1c4) Statistical motion feature images of the 4 postures (6 s).
Figure 3. Typical signal examples of different postures. (a1a4) Typical chirp time sequences of 4 postures: supine, left-side lying, right-side lying, and prone. (b1b4) Range-FFT images of the 4 postures (1 frame, 32 chirps). (c1c4) Statistical motion feature images of the 4 postures (6 s).
Sensors 24 05900 g003
Figure 4. Typical examples of the augmented samples. (a) Original: sample from the original feature image set. (b) Range shift: range shifted image with an offset of 10 pixels. (c) Time shift: time shifted image with an offset of 1 s. (d,e) Feature image 1 of prone, feature image 2 of left side lateral: feature images of supine and left side lateral. (f) Mix-up of feature image 1 and 2 ( λ = 0.5): mix-up image of (d,e) with λ = 0.5.
Figure 4. Typical examples of the augmented samples. (a) Original: sample from the original feature image set. (b) Range shift: range shifted image with an offset of 10 pixels. (c) Time shift: time shifted image with an offset of 1 s. (d,e) Feature image 1 of prone, feature image 2 of left side lateral: feature images of supine and left side lateral. (f) Mix-up of feature image 1 and 2 ( λ = 0.5): mix-up image of (d,e) with λ = 0.5.
Sensors 24 05900 g004
Figure 5. Architecture of ResTCN.
Figure 5. Architecture of ResTCN.
Sensors 24 05900 g005
Figure 6. The confusion matrix of the proposed method and relative SOTA methods. (a) SVM. (b) ShuffleNet. (c) DenseNet. (d) Vit. (e) Swin Transformer V2. (f) ResTCN (proposed).
Figure 6. The confusion matrix of the proposed method and relative SOTA methods. (a) SVM. (b) ShuffleNet. (c) DenseNet. (d) Vit. (e) Swin Transformer V2. (f) ResTCN (proposed).
Sensors 24 05900 g006
Table 1. The parameters of the radar.
Table 1. The parameters of the radar.
ParametersValues
Bandwidth5.0 GHz
Start frequency58 GHz
Chirp duration133 µs
Chirp repetition time463 µs
No. samples per chirp128
Frame rate20 Hz
ADC sampling rate1 MHz
Range resolution3 cm
Velocity resolution1.34 m/s
Table 2. The accuracy comparison of the proposed method and comparative methods.
Table 2. The accuracy comparison of the proposed method and comparative methods.
AlgorithmAccuracyPrecisionRecallF1Score
SVM [61]0.66270.64120.67780.6552
ShuffleNet [62]0.76520.79700.72380.7400
DenseNet [64]0.75070.80630.73250.7403
Vit [67]0.76210.75870.78380.7652
Swin transformer V2 [70]0.75670.76230.75660.7517
ResTCN (proposed)0.82740.84530.84990.8404
Table 3. The accuracy comparison of ResNet, TCN, and different combinations of network structures.
Table 3. The accuracy comparison of ResNet, TCN, and different combinations of network structures.
AlgorithmAccuracyPrecisionRecallF1Score
ResTCN (proposed)0.82740.84530.84990.8404
ResNet0.79800.81030.79600.8021
TCN0.74520.78560.72300.7343
ShuffleNet+TCN0.77660.80910.80190.7950
DenseNet+TCN0.77360.81920.82250.7995
ResNet+LSTM0.75920.78340.79040.7822
Table 4. The accuracy comparison of data with/without the motion feature extraction.
Table 4. The accuracy comparison of data with/without the motion feature extraction.
MethodAccuracyPrecisionRecallF1Score
Motion feature image0.82740.84530.84990.8404
Original range-FFT image0.68520.70180.66330.6780
Table 5. The accuracy comparison of different combinations of data augmentation options.
Table 5. The accuracy comparison of different combinations of data augmentation options.
MethodAccuracyPrecisionRecallF1Score
Proposed0.82740.84530.84990.8404
Original0.77610.79460.76880.7784
Mixup (-)0.79750.83360.81630.8108
Time shift (-)0.81290.82010.82820.8241
Range shift (-)0.80440.80840.82340.8227
Table 6. The accuracy comparison of different range window size.
Table 6. The accuracy comparison of different range window size.
Window SizeAccuracyPrecisionRecallF1Score
300.79470.82220.79020.7955
40 (optimal)0.82740.84530.84990.8404
500.81940.82340.82340.8276
600.81230.83200.81890.8241
Table 7. The accuracy comparison of different slow time window sizes.
Table 7. The accuracy comparison of different slow time window sizes.
Window SizeAccuracyPrecisionRecallF1Score
2 s0.76730.80670.73980.7448
6 s (optimal)0.82740.84530.84990.8404
10 s0.83540.85200.82020.8375
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Chen, K.; Xie, Y. A Deep Learning Method for Human Sleeping Pose Estimation with Millimeter Wave Radar. Sensors 2024, 24, 5900. https://doi.org/10.3390/s24185900

AMA Style

Li Z, Chen K, Xie Y. A Deep Learning Method for Human Sleeping Pose Estimation with Millimeter Wave Radar. Sensors. 2024; 24(18):5900. https://doi.org/10.3390/s24185900

Chicago/Turabian Style

Li, Zisheng, Ken Chen, and Yaoqin Xie. 2024. "A Deep Learning Method for Human Sleeping Pose Estimation with Millimeter Wave Radar" Sensors 24, no. 18: 5900. https://doi.org/10.3390/s24185900

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop