Next Article in Journal
Secure PUF-Based Authentication Systems
Previous Article in Journal
Design of Anti-Eccentric Load Sensor for Engineering Operation Early Warning Based on Particle Swarm Optimization
Previous Article in Special Issue
Fast Low-Sidelobe Pattern Synthesis Using the Symmetry of Array Geometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human Fall Detection with Ultra-Wideband Radar and Adaptive Weighted Fusion

by
Ling Huang
*,
Anfu Zhu
,
Mengjie Qian
and
Huifeng An
School of Electrical Engineering and Information Engineering, Lanzhou University of Technology, Lanzhou 730050, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(16), 5294; https://doi.org/10.3390/s24165294
Submission received: 24 June 2024 / Revised: 3 August 2024 / Accepted: 9 August 2024 / Published: 15 August 2024
(This article belongs to the Collection Radar, Sonar and Navigation)

Abstract

:
To address the challenges in recognizing various types of falls, which often exhibit high similarity and are difficult to distinguish, this paper proposes a human fall classification system based on the SE-Residual Concatenate Network (SE-RCNet) with adaptive weighted fusion. First, we designed the innovative SE-RCNet network, incorporating SE modules after dense and residual connections to automatically recalibrate feature channel weights and suppress irrelevant features. Subsequently, this network was used to train and classify three types of radar images: time–distance images, time–distance images, and distance–distance images. By adaptively fusing the classification results of these three types of radar images, we achieved higher action recognition accuracy. Experimental results indicate that SE-RCNet achieved F1-scores of 94.0%, 94.3%, and 95.4% for the three radar image types on our self-built dataset. After applying the adaptive weighted fusion method, the F1-score further improved to 98.1%.

1. Introduction

According to a 2018 survey by the World Health Organization (WHO), there are approximately 60,000 fatal fall incidents worldwide each year, the majority of which involve elderly individuals aged 65 and older [1]. Severe falls can lead to serious consequences such as fractures, dislocations, head injuries, and sprains [2]. In the member states of the European Union, injury issues affect approximately 105,000 people. Additionally, nearly 40,000 elderly individuals are declared dead due to falls. In the United States, the fall-related mortality rate among people aged 60 and older is 36.8 per 100,000, whereas in Canada, it is 9.4 per 10,000 for the same age group. In Australia, Canada, and the United Kingdom of Great Britain and Northern Ireland, the hospitalization rate for injuries due to falls among those aged 60 and older ranges from 1.6 to 3.0 per 10,000 citizens. In Western Australia and the United Kingdom, this rate is as high as 5.5 to 8.9 per 10,000 people [3]. Accurate identification of fall types can assist medical professionals in quicker diagnosis and formulation of effective treatment plans. Furthermore, integrating this technology with smart homes and health monitoring systems can significantly enhance the quality of life for the elderly and other vulnerable groups. The application of such recognition technology can facilitate more targeted treatment measures, reduce the risk of secondary injuries caused by falls, and promote the development of public health strategies and personalized medicine [4,5].
Currently, Human Motion Recognition (HMR) primarily relies on cameras. However, this method tends to invade personal privacy and requires unobstructed and well-lit conditions [6]. Although wearable devices overcome visual limitations by using inertial sensors or plantar pressure sensors, these devices may impose physical burdens, especially for the elderly and visually impaired individuals [7]. In contrast, environmental recognition methods using WiFi, infrared sensors, or radar can detect activities without adding burdens to the subjects or invading their privacy, making them the most ideal recognition methods currently available [8,9,10,11].
In previous studies, Zhao et al. addressed the issue of optimizing video images through network backpropagation by replacing time-frequency diagrams with convolutional layers and max-pooling layers. They also utilized fusion layers and an attention mechanism-based encoder-decoder structure to fully exploit the temporal information in the raw data, achieving effective recognition of continuous behaviors [12]. Li et al. proposed an enhanced contrastive learning dual transformer method (SCL-SwinT). This method incorporates overlapping patch embeddings and cosine similarity attention in the dual transformer encoder, combined with supervised contrastive learning (SupCon). By integrating two patch-embedding methods (ResNet and MLP) and three patch-partitioning strategies (sequential, grid, and strip), the model can extract contextual information from different angles and depths, achieving over 90% recognition accuracy on the IURHA2023 dataset [13]. Chen proposed a human activity classification method based on multiconvolutional neural network (CNN) information fusion. Six different data preprocessing methods were selected, and the results were input into corresponding CNN models for parameter training. The weighted voting method was used for information fusion to obtain the final classification results [14].
Fall behavior recognition is a specialized branch of human behavior recognition. Compared to general human behavior recognition, fall behavior recognition poses greater challenges in terms of recognition difficulty and data collection. Hemi et al. designed a radar map deep learning fusion network based on MobileNet-V3, significantly improving fall recognition performance. However, their consideration of spectral weight distribution was relatively limited, and they did not further classify fall types [15]. He et al. used parallel 2DCNNs to extract features, optimized by CBAM, to achieve feature-level fusion, but their study was limited to distinguishing single fall actions from various non-fall actions, with a limited range of action types in the experimental dataset [16]. Yao et al. proposed a fall detection system based on millimeter-wave radar, achieving high accuracy and robustness through neural network and information fusion technologies. This system collected target distance, speed, and angle information using FMCW radar, generating range-speed maps, range-horizontal angle maps, and range-vertical angle maps through three independent neural networks. To improve detection accuracy, this study used an ensemble learning stacking method to fuse the features extracted by the three neural networks. Although this study trained and recognized using a large dataset containing various action types, it mainly focused on accurately detecting fall events and distinguishing non-fall actions, rather than further classifying different types of falls [17].
Building on the work of Hemi et al., this paper refines the classification of fall behavior recognition and improves decision-level fusion methods. By embedding SE modules into Dense Residual Blocks composed of dense and residual connections, we achieved adaptive adjustment of feature importance, enhancing the recognition of fall behaviors and their types. Through in-depth comparisons with other advanced neural network models, the superiority of this network has been demonstrated. Additionally, this paper proposes an adaptive weighted fusion method to weight training results, improving the accuracy of behavior recognition.
The innovations of this paper are summarized as follows:
  • Novel Deep Learning Feature Extraction Network SE-RCNet: This network combines spatial attention mechanisms and residual connections, significantly enhancing the ability to extract key features from radar images, improving target detection and behavior classification performance under complex environmental conditions.
  • By employing a differentiated weight distribution mechanism, different weights are assigned to each radar map feature based on their contribution and impact in the decision-making process. By comprehensively evaluating the influence of each radar map feature, this method can more accurately determine behavior types.
The remainder of this paper is organized as follows: Section 2 describes the design process of the self-built dataset and introduces the public datasets used. Section 3 provides a brief overview of the adaptive weighted fusion network method and its structural diagram. Section 4 presents the experiments and results analysis. Section 5 briefly summarizes the content of this paper and discusses prospects for future research.

2. Dataset

2.1. Self-Built Dataset

In our experiment, we used the PulsOn 440 radar module from Time Domain. This UWB radar has a center frequency of 4.3 GHz and a frequency range of 3.1 to 4.8 GHz. The sampling frequency is 16.387 GHz. The antenna’s RF radiation intensity complies with FCC Part 15 regulations in the United States and adheres to ETSI EN 302065 standards in Europe [18]. We set the pulse repetition frequency (PRF) to 240 Hz to achieve higher velocity resolution, enabling the radar to detect finer movements. This is crucial for accurately identifying human positions and movements. The main radar parameter settings are shown in Table 1.

2.1.1. Dataset Description

In this study, we collected a dataset containing human fall behaviors using the PulsOn 440 UWB radar module (As shown in Figure 1) in an indoor environment. During the experiments, the radar was installed at a height of 1 m, and all experiments were conducted within the effective range of the radar to ensure data consistency and validity. This dataset includes fall behavior echo data from 9 subjects (2 females and 7 males). The basic information of the participants is shown in Table 2. Due to the inclusion of participants of different genders (male and female) and varying body profiles (height and weight), although they perform the same actions, the range of their movements is not entirely identical. This diversity allows us to simulate a wide range of human movements, ensuring that the collected data cover a broad spectrum of real-world scenarios.
Each subject performed data collection for 10 different types of behaviors. These behaviors include five types of daily activities: sitting down, bending to pick up an object, standing up, walking towards the radar, and walking away from the radar; as well as five types of fall behaviors: falling while sitting down or standing up, falling sideways to the radar, falling backwards to the radar, falling at a 45-degree angle to the right of the radar, and falling at a 45-degree angle to the left of the radar. Each subject provided 30 samples for each behavior type, and a total of 2700 samples were collected for classification and recognition research. Detailed action types and descriptions are shown in Table 3:

2.1.2. Data Preprocessing

As shown in Figure 2, to filter out stationary or slowly moving interference targets, the raw data collected by the radar are first processed using Moving Target Indication (MTI). We use a three-pulse canceller to eliminate the echoes of stationary targets by comparing the phase differences of consecutive pulses. The differential processing formula is as follows:
y ( n ) = x ( n + 1 ) 2 x ( n ) + x ( n 1 )
This formula effectively suppresses the echo signals of stationary targets by performing a weighted summation of the current pulse and the two adjacent pulses, thereby highlighting the signals of moving targets. This allows subsequent processing and analysis to focus more on moving targets. However, differential processing may lead to partial loss of information on stationary or slowly moving targets, as it attenuates slowly varying signal components, causing this information to be partially or completely lost. To address this, we introduced the Hilbert transform in subsequent processing. By generating an analytic signal, it provides instantaneous attributes (such as instantaneous phase and frequency) to preserve the complete information of the target as much as possible. The Hilbert transform helps to more comprehensively retain the characteristics of the signal, especially in frequency analysis and dynamic signal processing. It reduces artifacts and improves the accuracy of signal processing, thereby mitigating the distortion and information loss caused by differential processing [19]. The generated range-time diagram is shown in Figure 3a.
Subsequently, the Short-Time Fourier Transform (STFT) is applied to obtain the spectral information for each time window. In STFT, the signal is divided into multiple time windows. The signal within each window undergoes a Fourier transform, and through integration (i.e., summation), the intensity of each frequency component is calculated. This allows us to obtain the distribution of the signal across time and frequency. The generated time–doppler diagram (as shown in Figure 3b) shows the Doppler shift of the target at different time points, reflecting the speed changes in the target. The choice of time window size and overlap parameters in STFT directly affects the spectral resolution and time resolution. Inappropriate parameters may lead to spectral information distortion. To avoid this, we experimentally determined that the optimal window size is 128 points and the overlap parameter is 110 points. A window length of 128 points can capture important frequency components in radar signal preprocessing while maintaining reasonable time resolution. An overlap of 110 points improves the time resolution, making the spectrogram smoother and more continuous, reducing information loss caused by edge effects. These parameter choices are based on multiple experimental results, providing an optimal balance between frequency and time resolution in radar signal preprocessing. This ensures that the generated images have smooth time resolution and spectral continuity, thereby reducing signal distortion.
Finally, a Fast Fourier Transform (FFT) is performed on the time-range map data to generate a range–doppler diagram (as shown in Figure 3c). The range–doppler diagram can display the distribution of targets at different distances and frequency. By analyzing the color differences in these three types of radar images, we can clearly identify the interrelationships among distance, time, and Doppler frequency features, thereby providing more comprehensive target information and a reliable data source for subsequent decision fusion.

2.2. Public Dataset

To further demonstrate the generalization ability of the model, this study also utilized the K-band ultra-wideband radar fall detection dataset published by the Journal of Radars. This dataset includes fall and non-fall behavior data collected in various test environments, covering participants of different ages, heights, and genders, thus exhibiting significant diversity and representativeness. By incorporating this dataset, the model’s performance under different conditions and among diverse populations can be comprehensively evaluated.
This study primarily used ten different types of fall behaviors from this dataset to evaluate the network model’s effectiveness in recognizing similar fall behaviors. By using these ten types of falls, we can verify whether the model can maintain high recognition accuracy when dealing with highly similar fall actions. These fall behaviors are meticulously recorded in different environments and scenarios, providing rich samples for model training and validation. The detailed action types are shown in Table 4, covering various typical fall scenarios to comprehensively assess the model’s fall detection capabilities. The inclusion of this public dataset not only enhances the credibility and generalizability of the research results but also provides a solid data foundation and reference for subsequent research.

3. The Adaptive Weighted Fusion Network Recognition Method

3.1. Model Construction

This paper proposes a human fall behavior recognition model based on SE-RCNet. The model takes 80 × 80 pixel radar images as input and significantly improves the accuracy of fall behavior classification through a series of key mechanisms.
First, the model captures the basic features of radar images through an initial extraction module composed of convolutional layers and ReLU activation functions. Then, six densely connected residual blocks (labeled as block1 to block6 in the diagram) further refine and process these features. Each residual block consists of two convolutional layers, batch normalization, and ReLU activation functions, which not only enhance the ability to capture nonlinear features but also improve information flow and feature retention through the residual learning mechanism.
Each residual block integrates a Squeeze-and-Excite (SE) module. This module first compresses the spatial dimensions through global average pooling and then performs channel recalibration, adaptively adjusting the weights of the feature channels, thereby making the model focus more on important features and improving classification performance [20].
The densely connected residual blocks achieve residual learning through element-wise addition, where the output of the second convolutional layer is added to the output of the first convolutional layer, promoting information flow and feature retention in the deep network. The outputs of each residual block are merged to ensure the complete transmission of information between the deep layers while enriching feature representation.
After each densely connected residual block, a max-pooling layer further reduces feature dimensions and computational load, providing higher-level abstract feature representations for the model. Finally, a flattening layer converts the multidimensional feature maps into one-dimensional feature vectors, which are then input into the fully connected layer to achieve fall behavior classification.
The overall network structure is shown in Figure 4.

3.2. Adaptive Weighted Fusion

3.2.1. Confidence

In classification tasks, confidence measures the certainty of the model’s predictions. The model assigns a probability value to each category, with the sum of the probabilities for all categories equaling 1. The highest probability value represents the category that the model considers most likely. In short, the category with the highest probability is the model’s most confident prediction.

3.2.2. Decision-Level Fusion Network

The original decision fusion network was designed for binary classification problems of fall versus non-fall and is not suitable for multiclass classification problems. Therefore, we have modified it. The results of the three types of radar images are calculated and aggregated. If two or more radar images produce the same output, that output is used directly. If all outputs are different, the confidence of each output is compared, and the result with the highest confidence is chosen, as shown in Figure 5.
Although we have optimized the original decision fusion network to some extent, it still has a critical flaw: the network implicitly assigns equal weights to the results of the three different radar images. This approach is clearly inconsistent with our radar image recognition results. In reality, the recognition accuracy of these three radar images is different. Using traditional decision fusion methods not only diminishes the impact of the radar image with higher recognition accuracy but also amplifies the shortcomings of the radar image with lower recognition accuracy.

3.2.3. The Adaptive Weighted Fusion Network

In this paper, we adopted an adaptive weighted fusion method to enhance the performance of the decision-level network. This is an efficient strategy for addressing behavior recognition problems in multiclass tasks based on confidence levels. In this process, the confidence of each class prediction, which is the probability value output by the model, plays a central role. For the recognition tasks of each type of radar image, we no longer use the highest confidence principle alone. Instead, we employ an adaptive weighting mechanism that dynamically adjusts their influence in the final fusion decision by considering the recognition performance of different radar images on the validation set.
Specifically, the weights in the fusion network are determined by the accuracy of each individual radar image. When dealing with radar images of different accuracies, our fusion model automatically increases the weight of the more accurate results and reduces the weight of the less accurate results. This means that the model will synthesize and select the most reliable result as the final output based on the confidence and adjusted weight of each radar image.
The advantage of this adaptive weighting strategy is that it can automatically balance and optimize the integration of multisource information without sacrificing recognition accuracy. Our model applies this strategy to the SE-RCNet architecture, achieving more precise behavior recognition in the radar image dataset. Particularly in distinguishing different types of fall behaviors, this method not only improves recognition accuracy but also enhances the model’s adaptability and robustness under varying conditions.

3.2.4. The Adaptive Weighted Fusion Network-Specific Steps

(1) Dataset Partitioning and Initial Model Training
To ensure the fairness and scientific validity of model training and evaluation, this study first randomly divides the collected dataset into training, validation, and test sets in a 6:2:2 ratio. This step not only ensures comprehensive and random data coverage but also provides a rigorous benchmark for model evaluation.
After training the SE-RCNet model on the training set, we evaluate the initial performance of the model using the validation set and obtain the initial accuracy for various action categories in the fall type recognition task. This accuracy reflects the model’s ability to recognize unseen data.
(2) Radar Map Type Weight Allocation
The model trained with the training set was used to evaluate the validation set, yielding the initial radar map recognition accuracy n = [n1, n2, n3], where nk represents the accuracy of the kth radar map type. Then, each sample in the test set was input into the model. For each individual sample, the model outputs logits for ten categories. These logits are then input into the softmax function, which outputs the predicted probability for each category. The highest probability indicates the model’s confidence.
Assume the test set outputs a logits vector z = [z1, z2, …, z10], where zi is the magnitudes of logits for various categories in the sample. The softmax function converts this vector into another vector s = [s1, s2, …, s10], where si is calculated as follows:
s i = e z i j = 1 n e z j
In this expression, e z i is the natural exponent of z i . The denominator j = 1 n e z j is the sum of the natural exponents of all z j elements. n is the total number of classes, which is 10. The result of the softmax function is a probability distribution over the categories, where each element’s probability value is between 0 and 1, and the total sum of probabilities is 1.
To find the model’s confidence and its category, we use the following formula: the maximum probability s max = max ( s ) , where s max = max(s) indicates the highest value in the vector s. The maximum category index i max = argmax ( s ) , where argmax(s) is the index of the maximum value s max in the vector s. The three radar maps’ s max values for each test set sample are normalized to obtain the normalized probability matrix m = [ m 1 ,   m 2 ,   m 3 ] .
(3) Weighted Test Evaluation
We perform the Hadamard product of the probability output n from the validation set and the highest class recognition rate m from each test set spectrum to obtain the final probability-adjusted vector w = [ w 1 , w 2 , w 3 ], where w i represents the adjusted highest class recognition rate of the i-th spectrum. The specific calculation formula is expressed as follows:
w = n m
Then, we sum the probabilities of the same class in W to obtain a new probability vector w’. We use i max = argmax ( w ) as the final output class of the fusion network.
The overall recognition flowchart is shown in Figure 6.

4. Experimental Results and Analysis

4.1. Experimental Environment Setup

The computational environment for this study is configured as follows: The computer is equipped with an Intel i7-12700 processor, 32 GB of memory, and an NVIDIA GeForce RTX 3070 graphics card. The development environment is based on Anaconda 3, using Python 3.9.18. The construction, training, and performance evaluation of the deep learning models are carried out using TensorFlow 2.10.0 and Keras 2.7.0.

4.2. Comparative Experiment

In this section, we compare the proposed SE-RCNet network and adaptive weighted network with current mainstream deep learning networks and fusion methods on both the self-built dataset and the public dataset. Due to the specificity of the fall behavior recognition task, missing important positive samples (i.e., failing to detect fall behaviors) has a more severe impact than misclassifying samples (i.e., incorrectly detecting non-fall behaviors as fall behaviors). Therefore, in addition to using accuracy as a metric, more evaluation metrics are needed to comprehensively measure the overall performance of the network.
To this end, we introduce precision, recall, and F1-score as evaluation metrics in addition to accuracy. Among these metrics, TP (true positive) represents the number of times the model correctly identifies fall behaviors, TN (true negative) represents the number of times the model correctly identifies non-fall behaviors (i.e., the total number of correctly identified non-fall behaviors), FN (false negative) represents the number of times the model fails to identify fall behaviors (i.e., incorrectly identifying fall behaviors as other behaviors), and FP (false positive) represents the number of times the model incorrectly identifies fall behaviors (i.e., incorrectly identifying other behaviors as fall behaviors). The F1-score is a comprehensive metric that considers both precision and recall, providing a more complete assessment of the model’s performance.
Accurac = TR + TN Total   Cases
Precision = TP TP + FP
Recall = TP TP + FN
F1-score = 2 × Precision × Recall Precision + Recall
The system studied in this paper consists of two main subsystems: the radar image preliminary recognition network and the fusion optimization method. Since these two systems are technically progressive, we will evaluate the performance of each system separately.
First, we will assess the performance of the radar image preliminary recognition network. This subsystem is primarily responsible for extracting features from radar images and performing initial recognition. Then, we will evaluate the performance of the fusion optimization method. This subsystem optimizes and fine-tunes the initial recognition results, thereby improving overall recognition accuracy and robustness.
On this basis, we will compare the system studied in this paper with other existing networks to comprehensively evaluate its performance in fall behavior recognition. Through this phased evaluation and comparative analysis, we can gain a deep understanding of the advantages and disadvantages of each subsystem and verify the effectiveness of our system in handling highly similar fall behaviors.

4.2.1. Network Model Comparison

This paper compares the recognition performance of different deep learning models in fall detection. The selected comparison networks include CNN-4, DenseNet121 [21], ResNet50 [22], and Xception [23].
CNN-4 is a basic convolutional neural network structure with fewer layers and parameters, making it suitable as a baseline comparison model. DenseNet121, due to its densely connected layers, promotes gradient flow and feature reuse, thereby improving the overall performance of the model. ResNet50 addresses the degradation problem in deep neural networks, demonstrating excellent performance and stability, and is considered a classic network in the field of deep learning. Xception, by adopting depthwise separable convolutions, significantly reduces computational complexity while maintaining efficient model performance. These models represent different deep learning architectures and design philosophies. By comparing their performance, we can comprehensively evaluate the superiority of the SE-RCNet model in radar image processing and behavior recognition tasks. Table 5 lists the recognition rates of all networks.
Figure 7 shows the cross-entropy loss values (as in Figure 7a,c,e) and accuracy (as in Figure 7b,d,f) curves with increasing training epochs using the SE-RCNet network for five-fold cross-validation on individual spectrograms. As seen from the figures, after 100 training iterations, the accuracy and cross-entropy loss values for all three spectrograms tend to stabilize. Table 5 lists the average results of five-fold cross-validation on the three spectrograms for each network. Experimental data indicate that our proposed SE-RCNet model performs excellently in radar spectrum feature extraction and action recognition tasks, outperforming other comparison models across multiple evaluation metrics.
As a classic image recognition network, ResNet50 also performs quite well on our custom dataset, second only to SE-RCNet. However, the performances of CNN-4, DenseNet121, and Xception networks on various metrics are slightly inferior to those of SE-RCNet and ResNet50.
To evaluate the performance differences of different deep learning models in the fall detection task, this study employed one-way analysis of variance (ANOVA), which is used to test whether there are significant differences among multiple sample means. In this study, we considered the models as the factor and the F1-scores of the models as the response variable. Statistical analysis was conducted using the SciPy library. This statistical method helps determine whether there are significant differences in accuracy among the models, providing scientific basis for model selection. The results show significant differences in F1-scores among the models across the three types of radar spectrograms. Specifically, for the time–distance spectrogram (F(4, 20) = 36.618, p < 0.001), the time–distance spectrogram (F(4, 20) = 115.425, p < 0.001), and the range–distance spectrogram (F(4, 20) = 38.679, p < 0.001), all results indicate that the differences are statistically significant. Additionally, we conducted post hoc tests using Tukey’s HSD test to examine the differences between each pair of models. The post hoc test results are shown in Figure 8.
From the recognition results on the public dataset shown in Table 6, it can be seen that SE-RCNet’s advantage also lies in its ability to recognize different types of fall behaviors. By analyzing ten different fall behaviors in the K-band ultra-wideband radar fall detection spectrum dataset, SE-RCNet can more accurately distinguish these complex action types. This further demonstrates the model’s superior performance in handling diverse data and recognizing complex behavior patterns.

4.2.2. Comparison of Fusion Methods

In this section, we evaluate the fine-tuning effect of using an adaptive weighted fusion network for radar spectrum feature extraction and action recognition tasks, based on the initial classification results of the SE-RCNet network, we compared the performance of the adaptive weighted fusion network with that of the decision fusion network. We selected accuracy, precision, recall, and F1-score as the evaluation metrics. Below are the detailed experimental results and analysis.
From Table 7, it can be seen that the adaptive weighted fusion network performs excellently across multiple evaluation metrics. It achieved an accuracy of 0.975, a precision of 0.973, a recall of 0.973, and an F1-score of 0.972. Compared to the highest recognition results from individual spectrograms, these metrics improved by 2.08%, 1.87%, 2.94%, and 2.40%, respectively. This indicates that the network not only possesses high accuracy but also exhibits outstanding balance across all performance metrics. As shown in Figure 9, after adaptive weighted fine-tuning, only a small amount of confusion occurred between falling at a 45-degree angle to the left and falling at a 45-degree angle to the right. All other action types achieved 100% recognition.
To further verify the effectiveness of the adaptive weighted fusion network, we conducted comparative experiments on public datasets. The results demonstrated in the confusion matrix of the fusion method on the public dataset in Figure 10 indicate that the adaptive weighted fusion network still outperforms traditional decision fusion networks on the public dataset. As shown in Table 8, the adaptive weighted fusion network achieved an accuracy of 0.764, a precision of 0.768, a recall of 0.769, and an F1-score of 0.768 on the public dataset, while the traditional decision fusion network scored 0.689, 0.688, 0.691, and 0.689 on the respective metrics.
These results indicate that the adaptive weighted fusion network not only performs exceptionally well on self-built datasets but also demonstrates superior performance on public datasets. This can be mainly attributed to its ability to dynamically adjust the weights of sub-models. By adaptively adjusting the weights of each sub-model, the network can optimize the contribution of each sub-model in the final decision based on the current context and characteristics of the input data. This approach allows it to better adapt to different environments and data distributions, thereby improving the overall robustness and generalization capability of the model.
Overall, the adaptive weighted fusion network demonstrates high adaptability and stability in handling radar spectrum feature extraction and motion recognition tasks. Its dynamic adjustment mechanism enables it to maintain efficient recognition performance even in complex and variable environments.

5. Conclusions

This paper proposes a novel fall behavior recognition network, SE-RCNet, which significantly enhances recognition performance by integrating residual connections and Squeeze-and-Excitation (SE) modules. Comparative experiments show that SE-RCNet achieved average F1-scores of 94.0%, 94.3%, and 95.4% when processing three types of radar maps, significantly outperforming existing deep learning models.
Additionally, to further improve recognition accuracy and effectively utilize radar data, this paper introduces an adaptive weighted fusion method. This method innovatively adjusts the weight proportions dynamically based on the actual accuracy of each radar map, overcoming the limitations of traditional decision fusion strategies in radar map weight allocation. Experimental results demonstrate that this fusion method not only outperforms traditional single radar map recognition methods but also achieves an F1-score of 98.1% on our custom dataset with SE-RCNet, surpassing the decision fusion method. This indicates that the method can effectively address false positives and false negatives in fall recognition.
To verify the generalization capability of the SE-RCNet model and its adaptive weighted fusion method, we conducted evaluations on a public fall recognition dataset. The results show that the model and fusion method perform better on this dataset compared to traditional methods, confirming their high stability and reliability across different environments when compared to other network models and fusion algorithms.
However, in real life, human behavior is continuous and complex, not just simple, isolated actions. Current fall detection systems typically can only recognize single, obvious fall events, which limits their effective use in real-world environments. While our adaptive weighted fusion method can significantly improve the model’s recognition accuracy, it also presents challenges in terms of model and algorithm deployment, potentially reducing real-time recognition. However, with the continuous improvement in hardware performance and the emergence of hardware acceleration technologies (such as GPUs and FPGAs) and the deployment of edge computing. For the challenges posed by obstacles and electromagnetic interference, which can degrade radar system performance, our network and fusion method are specifically designed to improve recognition accuracy under such challenging conditions. By applying the adaptive weighted fusion algorithm to the results of the three radar maps, we enhance the system’s adaptability and robustness across various environments. Specifically, the adaptive weighted fusion algorithm can dynamically adjust the weight parameters of each model according to different environmental conditions, thereby achieving higher recognition performance. Therefore, future work will focus on recognizing continuous behavioral actions to enhance the general applicability of fall prevention systems and reduce misrecognition rates. Additionally, we will strive to improve the system’s ability to recognize human behavior in various interference environments.

Author Contributions

Conceptualization, L.H. and A.Z.; methodology, L.H. and A.Z.; software, A.Z.; validation, A.Z. and M.Q.; formal analysis, A.Z. and M.Q.; resources, M.Q.; data collection and experiments, A.Z. and H.A.; writing—A.Z.; writing—review and editing, L.H.; visualization, H.A.; supervision, L.H.; project administration, L.H.; funding acquisition, L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by National Natural Science Foundation of China Special Program under Grant 62341311, Lanzhou Municipal Talent Innovation and Entrepreneurship Project (Project No. 2022-RC-7).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

The public datasets used in this study are available from https://radars.ac.cn/web/data/getData?newsColumnId=878070cd-079e-454e-a14f-ebccc8cb5ba0 (accessed on 26 May 2024). The self-built dataset presented in this study is available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, X.; Ellul, J.; Azzopardi, G. Elderly Fall Detection Systems: A Literature Survey. Front. Robot. AI 2020, 7, 71. [Google Scholar] [CrossRef] [PubMed]
  2. Bhasin, S.; Gill, T.M.; Reuben, D.B.; Latham, N.K.; Ganz, D.A.; Greene, E.J.; Dziura, J.; Basaria, S.; Gurwitz, J.H.; Dykes, P.C.; et al. A Randomized Trial of a Multifactorial Strategy to Prevent Serious Fall Injuries. N. Engl. J. Med. 2020, 383, 129–140. [Google Scholar] [CrossRef] [PubMed]
  3. Kamińska, M.S.; Brodowski, J.; Karakiewicz, B. Fall Risk Factors in Community-Dwelling Elderly Depending on Their Physical Function, Cognitive Status and Symptoms of Depression. Int. J. Environ. Res. Public Health 2015, 12, 3406–3416. [Google Scholar] [CrossRef] [PubMed]
  4. Morris, M.E.; Webster, K.; Jones, C.; Hill, A.-M.; Haines, T.; McPhail, S.; Kiegaldie, D.; Slade, S.; Jazayeri, D.; Heng, H.; et al. Interventions to Reduce Falls in Hospitals: A Systematic Review and Meta-Analysis. Age Ageing 2022, 51, afac077. [Google Scholar] [CrossRef] [PubMed]
  5. Leland, N.E.; Elliott, S.J.; O’Malley, L.; Murphy, S.L. Occupational Therapy in Fall Prevention: Current Evidence and Future Directions. Am. J. Occup. Ther. 2012, 66, 149–160. [Google Scholar] [CrossRef] [PubMed]
  6. Wang, Y.; Chi, Z.; Liu, M.; Li, G.; Ding, S. High-Performance Lightweight Fall Detection with an Improved YOLOv5s Algorithm. Machines 2023, 11, 818. [Google Scholar] [CrossRef]
  7. Subramaniam, S.; Faisal, A.I.; Deen, M.J. Wearable Sensor Systems for Fall Risk Assessment: A Review. Front. Digit. Health 2022, 4, 921506. [Google Scholar] [CrossRef] [PubMed]
  8. Shen, M.; Tsui, K.-L.; Nussbaum, M.A.; Kim, S.; Lure, F. An Indoor Fall Monitoring System: Robust, Multistatic Radar Sensing and Explainable, Feature-Resonated Deep Neural Network. IEEE J. Biomed. Health Inform. 2023, 27, 1891–1902. [Google Scholar] [CrossRef] [PubMed]
  9. Ding, C.; Hong, H.; Zou, Y.; Chu, H.; Zhu, X.; Fioranelli, F.; Le Kernec, J.; Li, C. Continuous Human Motion Recognition With a Dynamic Range-Doppler Trajectory Method Based on FMCW Radar. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6821–6831. [Google Scholar] [CrossRef]
  10. Yang, Y.; Hou, C.; Lang, Y.; Yue, G.; He, Y.; Xiang, W. Person Identification Using Micro-Doppler Signatures of Human Motions and UWB Radar. IEEE Microw. Wirel. Compon. Lett. 2019, 29, 366–368. [Google Scholar] [CrossRef]
  11. Taylor, W.; Dashtipour, K.; Shah, S.A.; Hussain, A.; Abbasi, Q.H.; Imran, M.A. Radar Sensing for Activity Classification in Elderly People Exploiting Micro-Doppler Signatures Using Machine Learning. Sensors 2021, 21, 3881. [Google Scholar] [CrossRef] [PubMed]
  12. Zhao, R.; Ma, X.; Liu, X.; Liu, J. An End-to-End Network for Continuous Human Motion Recognition via Radar Radios. IEEE Sens. J. 2021, 21, 6487–6496. [Google Scholar] [CrossRef]
  13. Li, X.; Chen, S.; Zhang, S.; Zhu, Y.; Xiao, Z.; Wang, X. Advancing IR-UWB Radar Human Activity Recognition With Swin Transformers and Supervised Contrastive Learning. IEEE Internet Things J. 2024, 11, 11750–11766. [Google Scholar] [CrossRef]
  14. Chen, Y.; Wang, W.; Liu, Q.; Sun, Y.; Tang, Z.; Zhu, Z. Human Activity Classification with Radar Based on Multi-CNN Information Fusion. In Proceedings of the IET International Radar Conference (IET IRC 2020), Online, 4–6 November 2020; Institution of Engineering and Technology: Stevenage, UK, 2021; pp. 538–543. [Google Scholar]
  15. He, M.; Ping, Q.; Dai, R. Fall Detection Based on Deep Learning Fusing Ultrawideband Radar Spectrograms. J. Radars 2023, 12, 343–355. [Google Scholar] [CrossRef]
  16. He, J.; Ren, Z.; Zhang, W.; Jia, Y.; Guo, S.; Cui, G. Fall Detection Based on Parallel 2DCNN-CBAM With Radar Multidomain Representations. IEEE Sens. J. 2023, 23, 6085–6098. [Google Scholar] [CrossRef]
  17. Yao, Y.; Liu, C.; Zhang, H.; Yan, B.; Jian, P.; Wang, P.; Du, L.; Chen, X.; Han, B.; Fang, Z. Fall Detection System Using Millimeter-Wave Radar Based on Neural Network and Information Fusion. IEEE Internet Things J. 2022, 9, 21038–21050. [Google Scholar] [CrossRef]
  18. Data Sheet/User Manual, PulsON® 440. Available online: https://fccid.io/NUF-P440-A/User-Manual/User-Manual-2878444.pdf (accessed on 10 August 2024).
  19. Battista, B.M.; Knapp, C.; McGee, T.; Goebel, V. Application of the Empirical Mode Decomposition and Hilbert-Huang Transform to Seismic Reflection Data. GEOPHYSICS 2007, 72, H29–H37. [Google Scholar] [CrossRef]
  20. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef]
  21. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 2261–2269. [Google Scholar]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 770–778. [Google Scholar]
  23. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1800–1807. [Google Scholar]
Figure 1. PulsOn 440 radar module.
Figure 1. PulsOn 440 radar module.
Sensors 24 05294 g001
Figure 2. Diagram of Graph Generation.
Figure 2. Diagram of Graph Generation.
Sensors 24 05294 g002
Figure 3. Three Radar Spectrograms for Walking Away from the Radar. (a) Range-Time Diagram. (b) Time-Doppler Diagram. (c) Range-Doppler Diagram.
Figure 3. Three Radar Spectrograms for Walking Away from the Radar. (a) Range-Time Diagram. (b) Time-Doppler Diagram. (c) Range-Doppler Diagram.
Sensors 24 05294 g003
Figure 4. SE-RCNet Model Architecture Diagram.
Figure 4. SE-RCNet Model Architecture Diagram.
Sensors 24 05294 g004
Figure 5. Decision Fusion Diagram.
Figure 5. Decision Fusion Diagram.
Sensors 24 05294 g005
Figure 6. Adaptive Weighted Fusion Architecture Based on SE-RCNet.
Figure 6. Adaptive Weighted Fusion Architecture Based on SE-RCNet.
Sensors 24 05294 g006
Figure 7. The accuracy and Cross-Entropy Loss Curves of Single-Spectrogram Detection for 5-Fold Cross-Validation over the Number of Training Epochs. (a) Cross-entropy loss variation in the range–distance diagram. (b) Accuracy variation in the range–distance diagram. (c) Cross-entropy loss variation in the distance–time diagram. (d) Accuracy variation in the distance–time diagram. (e) Cross-entropy loss variation in the time–distance diagram. (f) Accuracy variation in the time–distance diagram.
Figure 7. The accuracy and Cross-Entropy Loss Curves of Single-Spectrogram Detection for 5-Fold Cross-Validation over the Number of Training Epochs. (a) Cross-entropy loss variation in the range–distance diagram. (b) Accuracy variation in the range–distance diagram. (c) Cross-entropy loss variation in the distance–time diagram. (d) Accuracy variation in the distance–time diagram. (e) Cross-entropy loss variation in the time–distance diagram. (f) Accuracy variation in the time–distance diagram.
Sensors 24 05294 g007
Figure 8. Boxplot of Model F1-Scores on Three Types of Spectrograms. (a) Boxplot based on time–doppler diagram. (b) Boxplot based on distance–time diagram. (c) Boxplot based on range–doppler diagram.
Figure 8. Boxplot of Model F1-Scores on Three Types of Spectrograms. (a) Boxplot based on time–doppler diagram. (b) Boxplot based on distance–time diagram. (c) Boxplot based on range–doppler diagram.
Sensors 24 05294 g008
Figure 9. Confusion Matrix of Fusion Methods in the Self-Built Dataset. (a) Decision Fusion Confusion Matrix Based on the Self-Built Dataset. (b) Adaptive Weighted Fusion Confusion Matrix Based on the Self-Built Dataset.
Figure 9. Confusion Matrix of Fusion Methods in the Self-Built Dataset. (a) Decision Fusion Confusion Matrix Based on the Self-Built Dataset. (b) Adaptive Weighted Fusion Confusion Matrix Based on the Self-Built Dataset.
Sensors 24 05294 g009
Figure 10. Confusion Matrix of Fusion Methods in the Public Datasets. (a) Decision Fusion Confusion Matrix Based on the Public Dataset. (b) Adaptive Weighted Fusion Confusion Matrix Based on the Public Dataset.
Figure 10. Confusion Matrix of Fusion Methods in the Public Datasets. (a) Decision Fusion Confusion Matrix Based on the Public Dataset. (b) Adaptive Weighted Fusion Confusion Matrix Based on the Public Dataset.
Sensors 24 05294 g010
Table 1. Radar Parameter Settings for Self-Built Dataset.
Table 1. Radar Parameter Settings for Self-Built Dataset.
Radar ModelPulsOn 440
pulse repetition frequency center frequency
frequency band range
sampling frequency
240 Hz
4.3 GHz
3.1~4.8 GHz
16.387 GHz
sampling time5 s
Table 2. Basic Information of Participants.
Table 2. Basic Information of Participants.
Subject IDGenderHeightWeightSubject IDGenderHeightWeight
1F165546M17573
2M170717F16859
3M168628M16967
4M172659M17066
5M17881
Table 3. Description of Actions in the Self-Built Dataset.
Table 3. Description of Actions in the Self-Built Dataset.
NumberBehavior TypeDescription
0Sitting downThe subject sits down from a standing position
1Falling while sitting/standingThe subject falls while at9tempting to sit or stand
2Bending to pick something upThe subject bends over to pick up an object from the ground
3Falling sideways to the radarThe subject falls sideways relative to the radar
4Standing upThe subject stands up from a seated position
5Falling backwards to the radarThe subject falls backwards relative to the radar
6Walking towards the radarThe subject walks towards the radar
7Walking away from the radarThe subject walks away from the radar
8Falling towards the radar at a 45-degree angle to the rightThe subject falls towards the radar at a 45-degree angle to the right
9Falling towards the radar at a 45-degree angle to the leftThe subject falls towards the radar at a 45-degree angle to the left
Table 4. Introduction to Public Human Fall Datasets.
Table 4. Introduction to Public Human Fall Datasets.
Action NumberAction Type
0Tripped while going upstairs
1Tripped while going downstairs
2Slipped while sitting down backward
3Fell while standing up from sitting
4Fainted with back facing the radar
5Slipped with back facing the radar
6Fell at 45 degrees to the right facing the radar
7Fell at 45 degrees to the left facing the radar
8Fainted facing the radar
9Fell facing the radar
Table 5. Cross-Validation Performance on Custom Spectrograms.
Table 5. Cross-Validation Performance on Custom Spectrograms.
ModelAccuracyPrecisionRecallF1-Score
TDTRRDTDTRRDTDTRRDTDTRRD
SE-RCNet0.9330.9410.9600.9350.9420.9550.9450.9440.9530.9400.9430.954
CNN-40.8980.9140.9310.8980.9140.9390.8940.9040.9290.8960.9060.933
DenseNet1210.8880.9300.9490.8910.9310.9490.9320.9280.9470.9080.9290.948
ResNet500.9300.9360.9520.9350.9350.9560.9360.9390.9500.9360.9370.953
Xception0.8750.8840.9140.8850.8940.9150.8940.8880.9090.8890.8910.911
Table 6. Cross-Validation Performance on Public Spectrograms.
Table 6. Cross-Validation Performance on Public Spectrograms.
ModelAccuracyPrecisionRecallF1-Score
TDTRRDTDTRRDTDTRRDTDTRRD
SE-RCNet0.5750.6980.6650.5700.7220.6740.5620.7210.6780.5660.7210.676
CNN-40.5500.6760.6200.5490.6890.6400.5300.6950.6040.5550.6920.607
DenseNet1210.5100.6690.5590.5050.6900.6320.5150.6810.6040.5070.6860.607
ResNet500.5480.6890.6630.5180.6890.6710.5360.6950.6630.5240.6920.667
Xception0.3300.5120.4950.3250.5510.5830.3260.5570.5390.3250.5540.553
Table 7. Comparison of Fusion Method Results in the Self-Built Dataset.
Table 7. Comparison of Fusion Method Results in the Self-Built Dataset.
AccuracyPrecisionRecallF1-Score
TD Diagram (Before Fusion)0.9370.9330.9380.935
TR Diagram (Before Fusion)0.9450.9410.9430.942
RD Diagram (Before Fusion)0.9630.9630.9530.958
Adaptive Weighted Fusion0.9830.9810.9810.981
Improvement (Adaptive Weighted Fusion)2.08%1.87%2.94%2.40%
Decision Fusion0.9750.9730.9730.972
Improvement (Decision Fusion)1.25%1.04%2.10%1.46%
Table 8. Comparison of Fusion Method Results in Public Datasets.
Table 8. Comparison of Fusion Method Results in Public Datasets.
Fusion MethodsAccuracyPrecisionRecallF1-Score
Adaptive Weighted Fusion0.7640.7680.7690.768
Decision Fusion0.6890.6880.6910.689
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, L.; Zhu, A.; Qian, M.; An, H. Human Fall Detection with Ultra-Wideband Radar and Adaptive Weighted Fusion. Sensors 2024, 24, 5294. https://doi.org/10.3390/s24165294

AMA Style

Huang L, Zhu A, Qian M, An H. Human Fall Detection with Ultra-Wideband Radar and Adaptive Weighted Fusion. Sensors. 2024; 24(16):5294. https://doi.org/10.3390/s24165294

Chicago/Turabian Style

Huang, Ling, Anfu Zhu, Mengjie Qian, and Huifeng An. 2024. "Human Fall Detection with Ultra-Wideband Radar and Adaptive Weighted Fusion" Sensors 24, no. 16: 5294. https://doi.org/10.3390/s24165294

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop