Next Article in Journal
Compressive Wideband Spectrum Sensing Aided Intelligence Transmitter Design
Previous Article in Journal
Enhancing Bottleneck Concept Learning in Image Classification
Previous Article in Special Issue
A Transformer Encoder Approach for Localization Reconstruction During GPS Outages from an IMU and GPS-Based Sensor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Driver Head–Hand Cooperative Action Recognition Based on FMCW Millimeter-Wave Radar and Deep Learning

College of Electronic Engineering, South China Agricultural University, Guangzhou 510642, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(8), 2399; https://doi.org/10.3390/s25082399
Submission received: 15 February 2025 / Revised: 3 April 2025 / Accepted: 9 April 2025 / Published: 10 April 2025

Abstract

:
Driver status plays a critical role in ensuring driving safety. However, the current visual recognition-based methods for detecting driver actions and status are often limited to factors such as ambient light condition, occlusion, and privacy concerns. In contrast, millimeter-wave radar offers various advantages such as high accuracy, ease of integration, insensitivity to light condition, and low cost; therefore, it has been widely used for monitoring vital signals and in action recognition. Despite this, the existing studies on driver action recognition have been hindered by limited accuracy and a narrow range of detectable actions. In this study, we utilized a 77 GHz millimeter-wave frequency-modulated continuous-wave radar to construct a dataset encompassing seven types of driver head–hand cooperative actions. Furthermore, a deep learning network model based on VGG16-LSTM-CBAM using micro-Doppler spectrograms as input was developed for action classification. The experimental results demonstrated that, compared to the existing CNN-LSTM and ALEXNET-LSTM networks, the proposed network achieves a classification accuracy of 99.16%, effectively improving driver action detection.

1. Introduction

The rapid advancement of smart solutions and information technology has revolutionized modern transportation systems [1]. With the rise of artificial intelligence, intelligent driving systems have also developed rapidly [2]. Vehicle systems now incorporate sensor-based interaction designs, such as visual, tactile, auditory, and body-sensing modalities, to assist drivers by monitoring hand gestures, voice, posture, and other vital signals, ensuring safety and convenience during driving. Drowsy driving remains one of the leading causes of road accidents; therefore, the intelligent detection of driver actions and status is essential in ensuring driving safety [3].
Currently, research has been conducted on the detection of driver head or body motion [4,5,6,7,8,9], as well as driver status [10,11,12], with a heavy reliance on visible light sensors and computer vision techniques. While these sensors provide high resolution, they face significant limitations, such as sensitivity to lighting conditions, background noise, and potential privacy concerns, particularly in public vehicles [13,14]. Infrared sensors provide high spatial resolution and functionality in low-light conditions, making them suitable for detecting nighttime pedestrian targets. However, for in-vehicle applications, their use is underexplored, as they remain susceptible to occlusion, lighting variations, and ambient heat—such as from seat heating devices—often reducing accuracy for complex actions. Furthermore, near-infrared imaging inherits privacy concerns by producing identifiable images.
Millimeter-wave radar has proven to be a valuable alternative due to its high detection accuracy, insensitivity to lighting conditions, ease of integration, and low cost. It has become widely adopted in biometric signal detection [15,16,17,18,19,20,21] and action and posture recognition [22,23,24,25,26,27,28]. It has proven effective in a variety of occupational and domestic scenarios. For example, Bresnahan et al. [29] applied millimeter-wave radar to detect deep tendon reflexes, creating spectrograms for the rapid and accurate analysis of patients’ reflex status. Arab et al. [30] used a 24 GHz millimeter-wave radar to sample and analyze eight different human body movements, applying a dual-branch CNN model to classify the actions, achieving a classification accuracy of 98.85%.
In the context of vehicle driver assistance, the use of millimeter-wave radar for driver status detection and action recognition is being continuously explored, thereby promoting safe driving [31]. In [32], frequency-modulated continuous-wave (FMCW) radar and CNN models were utilized to recognize and classify four types of head movements of a driver for abnormal behavior detection. A fine-tuned CNN model was employed to classify and recognize fatigue-related actions, with breathing signals under different human body states collected via millimeter-wave radar to analyze whether the driver was fatigued [33]. In [34], FMCW radar was used to collect six types of head movement signals to measure the distance Doppler and Doppler spectrograms during head and neck movements, proving the feasibility of using a radar to monitor the driver’s head movements. In [35], four different head movement signals from the driver were collected by installing millimeter-wave radar on the steering wheel; the action signals were classified using CNN networks. In [36], millimeter-wave radar was installed on the car dashboard to collect eight common head movement signals, which were then classified and recognized using a Deep-CNN model. However, these studies have limitations, focusing mainly on head motions and overlooking the cooperative movements of the head and hands, which are critical for accurately analyzing driver behavior in real-world situations.
In this study, we used millimeter-wave FMCW radar to collect micro-Doppler features from combined head and hand motions, creating a dataset of seven head–hand actions recorded from five volunteers. We also proposed a novel deep learning network VGG16-LSTM-CBAM, which integrated the VGG16 backbone with LSTM and CBAM attention blocks to enhance classification accuracy. The experimental results demonstrated the effectiveness of our proposed network by comparing it with other common deep learning models. The paper is organized as follows: Section 2 introduces the background theory of FMCW radar, Section 3 introduces the network model used in this experiment, and Section 4 presents the dataset. Section 5 discusses the action recognition accuracy under different networks, with conclusions drawn in Section 6.

2. Related Theories

2.1. Millimeter-Wave Radar Echo Model

In the FMCW radar system, a sequence of signals is transmitted whose frequency changes linearly with a function of time. The transmitted signal of the FMCW radar can be represented as follows:
s t = A T c o s ( 2 π f c t + π B T t 2 + ϕ 0 )
where A T is the amplitude of the transmitted signal, f c is the radar carrier frequency, T is the pulse interval, B is the bandwidth, and ϕ 0 is the initial phase of the chirp.
The corresponding received echo signal is represented as follows:
s t = A R cos ( 2 π f c ( t τ ) + π B T t τ 2 + φ 0 )
where A R is the amplitude of the echo signal, τ = 2 r / c is the frequency shift caused by the distance between the radar and the target, and c is the speed of light.
Then, the transmitted and received signals pass through the frequency mixer, and high and low-frequency signals are generated, which is represented as follows:
S I F t = A I F cos ( 2 π f c τ + π B T ( 2 t τ ) τ ) = A I F cos ( 2 π f c τ + 2 π B T τ π B T τ 2 )
where A I F is the difference frequency signal amplitude.
In practical applications, when the radar-detected target exhibits radial motion, the frequency of the received signal undergoes an additional frequency shift due to the Doppler effect. This phenomenon originates from the compression or stretching of the wavefront that electromagnetic waves experience during reflection from a moving target. Specifically, if the target moves with a velocity v relative to the radar, its echo signal will carry a frequency offset f d that is proportional to the velocity. This Doppler frequency shift can be expressed as follows:
f d = 2 v λ = 2 v f c c
where λ represents the wavelength of the transmitted signal, and c is the speed of light.

2.2. Micro-Doppler Spectrogram Acquisition

The Doppler shift described in Equation (4) applies to scenarios where the target moves uniformly as a whole. However, when there are local micro-motions in the target (such as human joint movements), its echo will be superimposed with periodic time-varying modulation components. The Doppler shift caused by such micro-motions exhibits dynamic harmonic characteristics, known as the micro-Doppler effect [37], which need to be extracted through joint time–frequency analysis. The Short-Time Fourier Transform (STFT) is a commonly used joint time–frequency analysis method for analyzing non-stationary signals, and it is also the time–frequency analysis method employed in this paper to extract micro-Doppler features. It is expressed as follows:
S ( ω , τ ) = F x t · w t τ = + x ( t ) · w ( t τ ) e j ω t d t
where x ( t ) represents the echo signal containing micro-Doppler modulation, w ( t ) is the window function, and ω is the angular frequency.
As shown in Figure 1, to obtain the micro-Doppler features of the driver’s actions, the following steps were performed: First, the radar echo signals were windowed along the fast-time dimension, followed by a distance-Fast Fourier Transform (FFT). The actual distance of the target was obtained based on the peak frequency component in the spectrum and the radar range resolution. Next, an FFT was applied along the slow-time dimension to obtain the distance-Doppler spectrum features. Subsequently, the Cell Averaging–Constant False Alarm Rate (CA-CFAR) algorithm [38] and angle-FFT were used to generate a range–angle heatmap. The Density-Based Spatial Clustering of Applications with Noise (DBSCAN) clustering algorithm was applied to extract the main motion areas of the driver, and the STFT was used to capture the time-domain information of the signal sequence. Finally, the time–frequency representation containing the driver’s action micro-Doppler features was obtained.

3. VGG16-LSTM-CBAM Network

In image recognition, various deep learning network models have been proposed, each with different capabilities for handling image classification and recognition tasks. In the dataset described in previous section, micro-Doppler feature spectrograms were generated, allowing for the application of image recognition network architectures. The proposed network structure is shown in Figure 2. This network is primarily based on the VGG16 module and modified to include 13 convolutional and 2 fully connected layers, a Long Short-Term Memory (LSTM) module, and a Convolutional Block Attention Module (CBAM) attention module. Additionally, the network also includes multiple max-pooling layers and linear rectification function (ReLU) activation function layers.

3.1. VGG16-LSTM

The VGG network model was proposed by the Visual Geometry Group at the University of Oxford in 2014 [39]. Since its introduction, several versions of the model have been developed. The VGG16 model, as its name suggests, primarily consists of 16 convolutional layers. VGG16 uses many convolution kernels of size 3 × 3, which enables the network to extract finer image features with a smaller receptive field. This design helps minimize feature loss during pooling and enhances feature extraction. The proposed network model is built upon the VGG16 model, incorporating module improvements.
After passing through the VGG16 network, the 3D tensor is flattened into a 2D tensor feature sequence and input into the LSTM network, which is known for the effective capture of hidden features and information in images. A SoftMax layer is used at the end of the network to classify and recognize the results from the fused VGG16-LSTM network, ultimately providing the predicted target. Additionally, residual connections are added to prevent overfitting caused by the deeper network layers [40]. Moreover, the network remains focused on low-level features, which helps to maintain or even improve performance after the modification.

3.2. CBAM

To help VGG16 capture more accurate key information in image processing and analysis, an attention mechanism module was also introduced into the network. This allowed the model to focus more on the critical parts of the input data during decision-making thus improving training accuracy and efficiency. The attention mechanism module added here is the CBAM, which combines channel attention and spatial attention to enhance the learning ability of the network. These two modules are connected serially; the input feature map first passes through the channel attention module to obtain a corrected feature map, which is then processed by the spatial attention module, resulting in an output feature map enhanced by CBAM.
As shown in Figure 3, the CBAM attention mechanism is introduced between VGG16 and LSTM networks, allowing the model to focus on more detailed feature information and ignore irrelevant data, thereby improving classification accuracy. In the VGG16-LSTM network, the crucial process is the feature fusion and transmission between the two networks. Therefore, the CBAM attention mechanism is placed between these networks, acting as a bridge for feature fusion and information transmission. This ensures that the features extracted by VGG16 are thoroughly filtered and integrated before being passed into the LSTM network for further feature extraction and behavior prediction. This design helps the network improve prediction accuracy without increasing its depth.

4. Dataset

4.1. Experimental Setup

The main equipment for this experiment consisted of the AWR1843BOOST millimeter-wave radar sensor [41] and the DCA1000EVM [42] data acquisition board from Texas Instruments. This device was used to collect driver action signals and generate micro-Doppler spectrograms as the experimental dataset. The AWR1843BOOST radar sensor had three transmitting antennas and four receiving antennas, and it was connected to the DCA1000EVM acquisition board, as shown in Figure 4.

4.2. Safety of Millimeter-Wave Radar

According to the specifications of the AWR1843BOOST millimeter-wave radar development board, under the typical operating conditions (3 transmit channels active with 4 receive channels; 25 ms frame interval), the total power consumption remains below ~2.0 W. Complementary measurements using the DCA1000EVM data capture card (operating at 5V DC supply with typical current draw of 700 mA) demonstrate an additional 3.5 W power requirement for real-time IF signal acquisition. This results in a combined system power budget of <6 W for simultaneous radar sensing and raw data streaming. Additionally, the total RF output power is 12 dBm (approximately 15.8 milliwatts). The horizontal 3 dB beam width is ±28°, and the elevation 3 dB beam width is ±14°. The skin penetration depth of the 77 GHz electromagnetic wave is only about 0.3 mm. In an extreme case where the electromagnetic wave is entirely transmitted into human tissue without any reflection, the absorbed power density at a distance of 20 cm from the antenna is estimated to be 0.8 W/m2. This value is significantly lower than the basic limit of 20 W/m2 set by the International Commission on Non-Ionizing Radiation Protection (ICNIRP) for the general public, thereby complying with health and safety regulations.

4.3. Data Acquisition

The driver action dataset including the head and hand movements are illustrated in Figure 5, which presents the following seven different driving scenarios: stationary, head turning (left–right), hand raised to ear (phone conversation), hand raised to mouth (smoking), head drooping forward (drowsiness), head nodding, and leaning forward (hard braking). The drowsiness action primarily focuses on head movement, such as the driver’s head drooping or nodding rapidly, leading to vertical shaking or trembling [43]. Nodding, on the other hand, involves a steady up-and-down motion of the head, which distinguishes it from the drowsiness action. The smoking action is designed to capture the driver’s hand movement by simulating the motion of bringing a cigarette to the mouth. The phone call action is designed to capture the driver raising a phone to their ear. The time–frequency spectrograms for each action are shown in Figure 6. To simulate a real driving scenario, during data collection, the subject sat on a chair facing forward, with the radar placed about 0.7 m to the right front of the target. The radar parameters are shown in Table 1.
In total, 5 participants participated in the experiment, including three males and two females, aged between 22 and 26, all with driving experience. Each participant performed each action 20 times, resulting in 700 data samples. Five-fold cross-validation was used to divide the collected data into five subsets. Each subset was used as the test set once, and the other four subsets were used for training. This process was repeated five times, and the average cross-validation accuracy was used as the model evaluation metric.

5. Experimental Results and Analysis

5.1. Performance Analysis

The model training was conducted on an GeForce RTX 4090D (Nvidia, Santa Clara, CA, USA). The hyperparameters in the deep learning network are set as follows: batch size of 8, cross-entropy loss function, and 120 epochs for training. The performance analysis is based on accuracy and confusion matrices, as shown in Figure 7. As the number of training epochs increased, accuracy gradually improved. During the first 30 training iterations, the accuracy fluctuated. After 40 epochs, the model converged, and both training and validation accuracies stabilized at around 99%.
Figure 8 shows the confusion matrix diagrams for VGG16-LSTM-CBAM, in the matrix, the rows represent the true action sequences, and the columns represent the action sequences recognized by the algorithm. A1 to A7 respectively denote: head nodding; head drooping forward; head turning; leaning forward; stationary; hand raised to ear; hand raised to mouth. The results reveal that, in the VGG16-LSTM-CBAM network, the best performance was achieved for “head turning”, “stationary”, and “hand raised to ear”, all reaching 100% accuracy. However, there are slight errors in recognizing “nodding” and “drowsiness”, as these two actions are quite similar in behavior. Additionally, there were some misclassifications for “smoking” and “hard braking”.

5.2. Comparison with Other Networks

Figure 9 illustrates the training accuracy trends of six models, namely CNN, CNN-LSTM, AlexNet-LSTM, EfficientNetB0, traditional VGG16, and VGG16-LSTM-CBAM. As shown in the figure, the training accuracy of all six networks improves with increasing iteration counts. On the same dataset, the VGG16-LSTM-CBAM network achieves an accuracy of 99.16%, representing a 6.45% improvement over the traditional CNN-LSTM and 13.73% over the baseline CNN. Notably, the traditional VGG16, whose fully connected layers consist of two 4096-dimensional dense layers, contributes over 90% of the total network parameters. In this study, we adopt a dimensionality collapse strategy to replace the original dual 4096-dimensional layers with a single 64-dimensional lightweight structure. Additionally, batch normalization (BN) layers are incorporated after each convolutional block to accelerate training convergence and mitigate overfitting. The final model achieves a 95% reduction in parameter count compared to the traditional VGG16 while improving accuracy by 1.12% (detailed in Table 2), conclusively demonstrating the superior performance of the proposed network.
Comparison of the confusion matrix diagrams for different networks is shown in Figure 10. As seen, compared to the Alexnet-LSTM, CNN-LSTM, CNN, EfficientNetB0, and VGG16 networks, the proposed VGG16-LSTM-CBAM network demonstrates a smaller discrepancy between the recognition results and the actual action sequences, achieving higher recognition accuracy. This further validates the feasibility and superiority of the method proposed in this paper for driver head–hand cooperative action recognition.
To evaluate the contribution of the modules proposed in the VGG16-LSTM-CBAM model, we further conducted four ablation experiments on our dataset (see Table 3). The baseline VGG16 model (with two 4096-dimensional fully connected (FC) layers) achieved an accuracy of 98.04%. Incorporating the LSTM temporal module enables the model to capture the sequential dependencies in input data, improving accuracy to 98.12%, while increasing the mode size slightly. The original FC-4096 layer contains over 90% of the model’s total parameters. By replacing it with an FC-64 layer, this model not only reduces computational load but also effectively inhibits overfitting risks, aligning with the lightweight concept, reducing mode size by 95%, increased accuracy by 0.56%, resulting in an accuracy of 99.16% for our final model.

6. Conclusions

In this study, we utilized an FMCW radar to construct a dataset encompassing seven types of driver’s head–hand cooperative actions. A deep learning network model based on VGG16-LSTM-CBAM using micro-Doppler spectrograms as input was developed for action classification. The experimental results demonstrated that, compared to the existing CNN-LSTM, EfficientNetB0, and ALEXNET-LSTM networks, the proposed network achieves a classification accuracy of 99.16%, effectively improving the driver actions detection and thus demonstrating the practical reference value for real-world applications. However, the human actions designed for this experiment were relatively fixed and simple. In real-world scenarios, more complex actions and interference may occur, which will increase the difficulty of detection. Therefore, future research will explore the inclusion of more complex driver actions and further recognition of facial features.

Author Contributions

Conceptualization, L.Z. and Y.D.; Methodology, L.Z.; Validation, X.C. and Z.C.; Formal analysis, L.Z., X.C. and J.Z.; Writing—original draft, L.Z.; Writing—review & editing, L.Z., J.Z. and Y.D.; Visualization, L.Z.; Project administration, Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Natural Science Foundation of Beijing under Grant L233017 and Guangdong Basic and Applied Basic Research Foundation under Grant 2023A1515011413.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data used in this paper are collected through our own experiments and are not yet publicly available. However, data may be obtained from the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Creß, C.; Bing, Z.; Knoll, A.C. Intelligent Transportation Systems Using Roadside Infrastructure: A Literature Survey. IEEE Trans. Intell. Transp. Syst. 2024, 25, 6309–6327. [Google Scholar] [CrossRef]
  2. Garikapati, D.; Shetiya, S.S. Autonomous Vehicles: Evolution of Artificial Intelligence and the Current Industry Landscape. Big Data Cogn. Comput. 2024, 8, 42. [Google Scholar] [CrossRef]
  3. Wang, G.; Wang, Y.; Sun, H.; Miao, B.; Wang, Y. A Reference Matching-Based Temperature Compensation Method for Ultrasonic Guided Wave Signals. Sensors 2019, 19, 5174. [Google Scholar] [CrossRef] [PubMed]
  4. Xing, Y.; Lv, C.; Wang, H.; Cao, D.; Velenis, E.; Wang, F.-Y. Driver Activity Recognition for Intelligent Vehicles: A Deep Learning Approach. IEEE Trans. Veh. Technol. 2019, 68, 5379–5390. [Google Scholar] [CrossRef]
  5. Saranya, D.; Nalinipriya, G.; Kanagavalli, N.; Arunkumar, S.; Kavitha, G. Deep CNN Models for Driver Activity Recognition for Intelligent Vehicles. 2020, 8, 7146–7150. Int. J. 2020, 8, 7146–7150. [Google Scholar]
  6. Chai, W.; Chen, J.; Wang, J.; Velipasalar, S.; Venkatachalapathy, A.; Adu-Gyamfi, Y.; Merickel, J.; Sharma, A. Driver Head Pose Detection From Naturalistic Driving Data. IEEE Trans. Intell. Transp. Syst. 2023, 24, 9368–9377. [Google Scholar] [CrossRef]
  7. Jiao, S.-J.; Liu, L.-Y.; Liu, Q. A Hybrid Deep Learning Model for Recognizing Actions of Distracted Drivers. Sensors 2021, 21, 7424. [Google Scholar] [CrossRef]
  8. Shafique, A.; Cao, G.; Aslam, M.; Asad, M.; Ye, D. Application-Aware SDN-Based Iterative Reconfigurable Routing Protocol for Internet of Things (IoT). Sensors 2020, 20, E3521. [Google Scholar] [CrossRef]
  9. Sajid, F.; Javed, A.R.; Basharat, A.; Kryvinska, N.; Afzal, A.; Rizwan, M. An Efficient Deep Learning Framework for Distracted Driver Detection. IEEE Access 2021, 9, 169270–169280. [Google Scholar] [CrossRef]
  10. Ansari, S.; Naghdy, F.; Du, H.; Pahnwar, Y.N. Driver Mental Fatigue Detection Based on Head Posture Using New Modified reLU-BiLSTM Deep Neural Network. IEEE Trans. Intell. Transp. Syst. 2022, 23, 10957–10969. [Google Scholar] [CrossRef]
  11. Zhao, Z.; Xia, S.; Xu, X.; Zhang, L.; Yan, H.; Xu, Y.; Zhang, Z. Driver Distraction Detection Method Based on Continuous Head Pose Estimation. Comput. Intell. Neurosci. 2020, 2020, 9606908. [Google Scholar] [CrossRef]
  12. Sen, A.; Mandal, A.; Karmakar, P.; Das, A.; Chakraborty, S. Passive Monitoring of Dangerous Driving Behaviors Using mmWave Radar. Pervasive Mob. Comput. 2024, 103, 101949. [Google Scholar] [CrossRef]
  13. Ma, J.; Ma, Y.; Li, C. Infrared and Visible Image Fusion Methods and Applications: A Survey. Inf. Fusion 2019, 45, 153–178. [Google Scholar] [CrossRef]
  14. Moon, J.; Bukhari, M.; Kim, C.; Nam, Y.; Maqsood, M.; Rho, S. Object Detection under the Lens of Privacy: A Critical Survey of Methods, Challenges, and Future Directions. ICT Express 2024, 10, 1124–1144. [Google Scholar] [CrossRef]
  15. Guo, J.; Wei, J.; Xiang, Y.; Han, C. Millimeter-Wave Radar-Based Identity Recognition Algorithm Built on Multimodal Fusion. Sensors 2024, 24, 4051. [Google Scholar] [CrossRef]
  16. Mehrjouseresht, P.; Hail, R.E.; Karsmakers, P.; Schreurs, D.M.P. Respiration and Heart Rate Monitoring in Smart Homes: An Angular-Free Approach with an FMCW Radar. Sensors 2024, 24, 2448. [Google Scholar] [CrossRef]
  17. Xiang, M.; Ren, W.; Li, W.; Xue, Z.; Jiang, X. High-Precision Vital Signs Monitoring Method Using a FMCW Millimeter-Wave Sensor. Sensors 2022, 22, 7543. [Google Scholar] [CrossRef]
  18. Chen, Y.; Yuan, J.; Tang, J. A High Precision Vital Signs Detection Method Based on Millimeter Wave Radar. Sci. Rep. 2024, 14, 25535. [Google Scholar] [CrossRef]
  19. Dai, J.; Yan, J.; Qi, Y. Millimeter Wave Radar Range Bin Tracking and Locking for Vital Sign Detection with Binocular Cameras. Appl. Sci.-BASEL 2023, 13, 6270. [Google Scholar] [CrossRef]
  20. Han, Z.; Wang, J.; Xu, Z.; Yang, S.; He, L.; Xu, S.; Wang, J. 4D Millimeter-Wave Radar in Autonomous Driving: A Survey. arXiv 2023, arXiv:2306.04242. [Google Scholar]
  21. Anitori, L.; Jong, A.D.; Nennie, F. FMCW Radar for Life-Sign Detection. In Proceedings of the IEEE Radar Conference, Pasadena, CA, USA, 4–8 May 2009. [Google Scholar]
  22. Bai, W.; Chen, S.; Ma, J.; Wang, Y.; Han, C. Gesture Recognition with Residual LSTM Attention Using Millimeter-Wave Radar. Sensors 2025, 25, 469. [Google Scholar] [CrossRef] [PubMed]
  23. Arsalan, M.; Santra, A.; Issakov, V. RadarSNN: A Resource Efficient Gesture Sensing System Based on Mm-Wave Radar. IEEE Trans. Microw. Theory Tech. 2022, 70, 2451–2461. [Google Scholar] [CrossRef]
  24. Govindaraj, V. Forward Collision Warning System with Visual Distraction Detection in Bikes. Ph.D. Thesis, University of Twente, Enschede, The Netherlands, 2021. [Google Scholar]
  25. Li, W.; Guan, Y.; Chen, L.; Sun, L. Millimeter-Wave Radar and Machine Vision-Based Lane Recognition. Int. J. Pattern Recognit. Artif. Intell. 2018, 32, 1850015. [Google Scholar] [CrossRef]
  26. Dang, X.; Fan, K.; Li, F.; Tang, Y.; Gao, Y.; Wang, Y. Multi-Person Action Recognition Based on Millimeter-Wave Radar Point Cloud. Appl. Sci. Switz. 2024, 14, 7253. [Google Scholar] [CrossRef]
  27. Li, L.; Bai, W.; Han, C. Multiscale Feature Fusion for Gesture Recognition Using Commodity Millimeter-Wave Radar. Comput. Mater. Contin. 2024, 81, 1613–1640. [Google Scholar] [CrossRef]
  28. Lim, S.; Park, C.; Lee, S.; Jung, Y. Human Activity Recognition Based on Point Clouds from Millimeter-Wave Radar. Appl. Sci. Switz. 2024, 14, 10764. [Google Scholar] [CrossRef]
  29. Bresnahan, D.G.; Koziol, S.; Li, Y. Investigation of Patellar Deep Tendon Reflex Using Millimeter-Wave Radar and Motion Capture Technologies. IEEE Access 2024, 12, 9220–9228. [Google Scholar] [CrossRef]
  30. Arab, H.; Ghaffari, I.; Chioukh, L.; Tatu, S.O.; Dufour, S. A Convolutional Neural Network for Human Motion Recognition and Classification Using a Millimeter-Wave Doppler Radar. IEEE Sens. J. 2022, 22, 4494–4502. [Google Scholar] [CrossRef]
  31. Gharamohammadi, A.; Khajepour, A.; Shaker, G. In-Vehicle Monitoring by Radar: A Review. IEEE Sens. J. 2023, 23, 25650–25672. [Google Scholar] [CrossRef]
  32. Nguyen, H.N.; Lee, S.; Nguyen, T.; Kim, Y. One-shot Learning-based Driver’s Head Movement Identification Using a Millimetre-wave Radar Sensor. IET Radar Sonar Navig. 2022, 16, 825–836. [Google Scholar] [CrossRef]
  33. Chen, H.; Han, X.; Hao, Z.; Yan, H.; Yang, J. Non-Contact Monitoring of Fatigue Driving Using FMCW Millimeter Wave Radar. ACM Trans. Internet Things 2024, 5, 3. [Google Scholar] [CrossRef]
  34. Chae, R.; Wang, A.; Li, C. FMCW Radar Driver Head Motion Monitoring Based on Doppler Spectrogram and Range-Doppler Evolution. In Proceedings of the IEEE Topical Conference on Wireless Sensors and Sensor Networks, Orlando, FL, USA, 20–23 January 2019. [Google Scholar]
  35. Jung, J.; Lim, S.; Kim, B.-K.; Lee, S. CNN-Based Driver Monitoring Using Millimeter-Wave Radar Sensor. IEEE Sens. Lett. 2021, 5, 3500404. [Google Scholar] [CrossRef]
  36. Bresnahan, D.G.; Li, Y. Classification of Driver Head Motions Using a Mm-Wave FMCW Radar and Deep Convolutional Neural Network. IEEE Access 2021, 9, 100472–100479. [Google Scholar] [CrossRef]
  37. Zhang, Y.; Narayanan, R.M. Narayanan Monopulse Radar Based on Spatiotemporal Correlation of Stochastic Signals. IEEE Trans. Aerosp. Electron. Syst. 2006, 42, 160–173. [Google Scholar] [CrossRef]
  38. Finn, H.; Johnson, R. Adaptive detection mode with threshold control as a function of spatially sampled clutter-level estimates. RCA Rev. 1968, 29, 414–465. [Google Scholar]
  39. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  41. AWR1843BOOST TI.com.cn. Available online: http://www.ti.com.cn/tool/cn/AWR1843BOOST (accessed on 3 February 2025).
  42. DCA1000EVM TI.com.cn. Available online: http://www.ti.com.cn/tool/cn/DCA1000EVM (accessed on 3 February 2025).
  43. Sheng-Yang, S.; Wen-Zhong, T.; Yan-Yang, W.; Long, L.; Li, Y.; Li, X.; Dai, Y.; Yang, H. A Review on Fatigue Driving Detection. Itm Web Conf. 2017, 12, 01019. [Google Scholar]
Figure 1. Radar signal processing flowchart. In the upper left chart, the black curves represent the transmitted signals, and the red curves depict the received signals.
Figure 1. Radar signal processing flowchart. In the upper left chart, the black curves represent the transmitted signals, and the red curves depict the received signals.
Sensors 25 02399 g001
Figure 2. VGG16-LSTM-CBAM network model.
Figure 2. VGG16-LSTM-CBAM network model.
Sensors 25 02399 g002
Figure 3. Schematic diagram of VGG-LSTM with CBAM.
Figure 3. Schematic diagram of VGG-LSTM with CBAM.
Sensors 25 02399 g003
Figure 4. Experimental setup: connection between the AWR1843BOOST radar sensor and the DCA1000EVM data acquisition board.
Figure 4. Experimental setup: connection between the AWR1843BOOST radar sensor and the DCA1000EVM data acquisition board.
Sensors 25 02399 g004
Figure 5. Schematic diagram of different actions, the images are as follows: (a) stationary; (b) head turning; (c) hand raised to ear; (d) hand raised to mouth; (e) head drooping forward; (f) head nodding; (g) leaning forward. Red arrows are used to emphasize the extent of the downward head movement.
Figure 5. Schematic diagram of different actions, the images are as follows: (a) stationary; (b) head turning; (c) hand raised to ear; (d) hand raised to mouth; (e) head drooping forward; (f) head nodding; (g) leaning forward. Red arrows are used to emphasize the extent of the downward head movement.
Sensors 25 02399 g005
Figure 6. The micro-Doppler spectrograms of different actions, the images are as follows: (a) stationary; (b) head turning; (c) hand raised to ear; (d) hand raised to mouth; (e) head drooping forward; (f) head nodding; (g) leaning forward.
Figure 6. The micro-Doppler spectrograms of different actions, the images are as follows: (a) stationary; (b) head turning; (c) hand raised to ear; (d) hand raised to mouth; (e) head drooping forward; (f) head nodding; (g) leaning forward.
Sensors 25 02399 g006aSensors 25 02399 g006b
Figure 7. VGG16-LSTM-CBAM training and validation performance: (a) loss curve; (b) accuracy graph.
Figure 7. VGG16-LSTM-CBAM training and validation performance: (a) loss curve; (b) accuracy graph.
Sensors 25 02399 g007
Figure 8. Confusion matrix diagrams for VGG16-LSTM-CBAM. A1 to A7 respectively denote: head nodding; head drooping forward; head turning; leaning forward; stationary; hand raised to ear; hand raised to mouth.
Figure 8. Confusion matrix diagrams for VGG16-LSTM-CBAM. A1 to A7 respectively denote: head nodding; head drooping forward; head turning; leaning forward; stationary; hand raised to ear; hand raised to mouth.
Sensors 25 02399 g008
Figure 9. Comparison of test accuracy between VGG16_LSTM_CBAM and the different networks performance: (a) Alexnet-LSTM; (b) CNN-LSTM; (c) CNN; (d) EfficientNetB0 and (e) VGG16.
Figure 9. Comparison of test accuracy between VGG16_LSTM_CBAM and the different networks performance: (a) Alexnet-LSTM; (b) CNN-LSTM; (c) CNN; (d) EfficientNetB0 and (e) VGG16.
Sensors 25 02399 g009aSensors 25 02399 g009b
Figure 10. Confusion matrix diagrams for different networks performance: (a) Alexnet-LSTM; (b) CNN-LSTM; (c) CNN; (d) EfficientNetB0 and (e) VGG16.
Figure 10. Confusion matrix diagrams for different networks performance: (a) Alexnet-LSTM; (b) CNN-LSTM; (c) CNN; (d) EfficientNetB0 and (e) VGG16.
Sensors 25 02399 g010aSensors 25 02399 g010b
Table 1. TI AWR1843 FMCW millimeter-wave radar configuration parameter.
Table 1. TI AWR1843 FMCW millimeter-wave radar configuration parameter.
ParametersValue
Starting frequency/GHz77
Effective bandwidth/MHz2667
Chirp samples/count128
Sampling frequency/ksps4000
Slope/(MHz/μs)21
Frame chirp count/count255
Frame period/ms270
Table 2. Comparison of different networks.
Table 2. Comparison of different networks.
NetworkAccuracyRecallF1 ScoreParameter Count (M)FLOPs (G)Model Size (MB)
VGG16-LSTM-CBAM99.16%99.13%99.15%16.2591.0061.99
VGG1698.04%97.94%97.99%304.1790.371160.33
Alexnet-LSTM98.03%98.00%98.02%4.996.1019.03
EfficientNetB097.76%98.12%97.86%4.130.7815.76
CNN-LSTM92.71%92.50%92.61%0.250.080.94
CNN85.43%84.80%85.10%0.300.031.16
Table 3. Ablation experiments.
Table 3. Ablation experiments.
NetworkAccuracyF1 ScoreModel Size (MB)
VGG16 only98.04%97.99%1160.33
+LSTM98.12%98.05%1160.89
Modified FC98.60%98.63%61.46
VGG16-LSTM-CBAM99.16%99.15%61.99
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Chen, X.; Chen, Z.; Zheng, J.; Diao, Y. Driver Head–Hand Cooperative Action Recognition Based on FMCW Millimeter-Wave Radar and Deep Learning. Sensors 2025, 25, 2399. https://doi.org/10.3390/s25082399

AMA Style

Zhang L, Chen X, Chen Z, Zheng J, Diao Y. Driver Head–Hand Cooperative Action Recognition Based on FMCW Millimeter-Wave Radar and Deep Learning. Sensors. 2025; 25(8):2399. https://doi.org/10.3390/s25082399

Chicago/Turabian Style

Zhang, Lianlong, Xiaodong Chen, Zexin Chen, Jiawen Zheng, and Yinliang Diao. 2025. "Driver Head–Hand Cooperative Action Recognition Based on FMCW Millimeter-Wave Radar and Deep Learning" Sensors 25, no. 8: 2399. https://doi.org/10.3390/s25082399

APA Style

Zhang, L., Chen, X., Chen, Z., Zheng, J., & Diao, Y. (2025). Driver Head–Hand Cooperative Action Recognition Based on FMCW Millimeter-Wave Radar and Deep Learning. Sensors, 25(8), 2399. https://doi.org/10.3390/s25082399

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop