Next Article in Journal
Deep Learning Based Pile-Up Correction Algorithm for Spectrometric Data Under High-Count-Rate Measurements
Previous Article in Journal
Filtering and Overlapping Data for Accuracy Enhancement of Doppler-Based Location Method
Previous Article in Special Issue
A Usability Pilot Study of a Sensor-Guided Interactive System for Dexterity Training in Parkinson’s Disease
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Slip, Trip, and Fall Prevention: Real-World Near-Fall Detection with Advanced Machine Learning Technique

1
Institute for Occupational Safety and Health of the German Social Accident Insurance (IFA), 53757 Sankt Augustin, Germany
2
RheinAhrCampus, Koblenz University of Applied Sciences, 53424 Remagen, Germany
3
German Research Center for Artificial Intelligence (DFKI), 28359 Bremen, Germany
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(5), 1468; https://doi.org/10.3390/s25051468
Submission received: 7 February 2025 / Revised: 25 February 2025 / Accepted: 26 February 2025 / Published: 27 February 2025
(This article belongs to the Special Issue Sensors for Human Activity Recognition: 3rd Edition)

Abstract

:
Slips, trips, and falls (STFs) are a major occupational hazard that contributes significantly to workplace injuries and the associated financial costs. The application of traditional fall detection techniques in the real world is limited because they are usually based on simulated falls. By using kinematic data from real near-fall incidents that occurred in physically demanding work environments, this study overcomes this limitation and improves the ecological validity of fall detection algorithms. This study systematically tests several machine-learning architectures for near-fall detection using the Prev-Fall dataset, which consists of high-resolution inertial measurement unit (IMU) data from 110 workers. Convolutional neural networks (CNNs), residual networks (ResNets), convolutional long short-term memory networks (convLSTMs), and InceptionTime models were trained and evaluated over a range of temporal window lengths using a neural architecture search. High-validation F1 scores were achieved by the best-performing models, particularly CNNs and InceptionTime, indicating their effectiveness in near-fall classification. The need for more contextual variables to increase robustness was highlighted by recurrent false positives found in subsequent tests on previously unobserved occupational data, especially during biomechanically demanding activities such as bending and squatting. Nevertheless, our findings suggest the applicability of machine-learning-based STF prevention systems for workplace safety monitoring and, more generally, applications in fall mitigation. To further improve the accuracy and generalizability of the system, future research should investigate multimodal data integration and improved classification techniques.

1. Introduction

Slips, trips, and falls (STF incidents) remain a significant occupational hazard, leading to substantial rates of absence from work and adverse outcomes. In 2023, German Social Accident Insurance (DGUV) reported 171,976 STF incidents, resulting in 7 fatalities and 2291 disability benefit cases [1], underscoring the urgent need for effective prevention strategies. Such incidents are particularly prevalent in sectors such as logistics, transport, and traffic.
Recent advancements highlight the potential of machine learning and sensor technologies for mitigating STF risks. Wang et al. improved near-fall detection among older adults by using body-worn sensors, demonstrating enhanced accuracy and reliability [2]. Similarly, Nikolov et al. emphasized the importance of realistic datasets for fall detection, utilizing low-cost, human-like dummies to generate thermal surveillance data [3].
Innovative fall detection methods continue to emerge. Hsu et al. enhanced detection accuracy in complex environments by utilizing a generative adversarial network (GAN) in a sequence-to-sequence denoising approach, which effectively encodes spatiotemporal correlations for more robust feature extraction [4]. Rahemtulla et al. introduced wearable electronic textiles capable of continuous fall and near-fall detection [5]. Schumann et al. further advanced robustness in fall detection by integrating kinematic data with environmental factors, employing gait analysis and machine learning to identify fall risks in individuals with multiple sclerosis [6].
Despite these advancements, the workplace applicability of current approaches remains limited, owing to their reliance on simulated falls—intentionally induced falls performed in controlled settings, often performed by trained participants or mannequins. This research addresses these gaps by utilizing kinematic data from actual STF events involving physically demanding occupations. Near-falls, characterized by tripping, slipping, or misstepping without falling, are key indicators of future STF risks. Building on the Prev-Fall dataset, this study examines kinematic responses to unexpected gait perturbations in parcel delivery workers and steelworkers [7]. The dataset offers diverse, real-world data, addressing the limitations of earlier studies constrained by simulated falls or narrow demographics.
To enhance detection capabilities, the Prev-Fall dataset was employed in this study to train machine-learning algorithms for near-fall detection. While early-threshold-based classifiers yielded limited success, advanced machine-learning models substantially improved accuracy and reliability. These advancements underscore the necessity of high-quality, realistic data for developing robust STF prevention systems.

Dataset Differentiation

STF prevention relies on two primary types of datasets: those collected via wearable sensors and those captured with ambient sensors. Ambient sensors, such as camera systems and pressure-sensitive floors, face limitations, such as dependency on camera angles, lighting conditions, and their inability to record precise biomechanical data. In contrast, wearable sensors, particularly inertial measurement units (IMUs), offer continuous monitoring and capture accurate kinematic data, which are crucial for effective fall detection and prevention. This research leverages the Prev-Fall dataset [7] collected through wearable sensors to develop machine-learning models that are both robust and reliable, ensuring suitability for real-world occupational settings.
The primary aim of this work is to perform a neural architecture search with models proven in the field of fall and near-fall detection to analyze their potential on the non-simulated near-fall data of working-age individuals of the previously published Prev-Fall dataset [7]. In order to find appropriate models, we first analyze other works in the field of fall and near-fall detection. We begin by providing an overview of existing fall and near-fall datasets and network architectures used in fall and near-fall detection. Next, we performed linear discriminant analysis and a correlation analysis of the Prev-Fall dataset [7] to gain a deeper understanding of the data. We then perform the neural architecture search before finally testing the best networks on data recorded with the same motion-capture system but in another context and not forming part of the Prev-Fall dataset.
By utilizing kinematic data from real, non-simulated STF incidents, this study seeks more accurate and reliable near-fall detection systems than those common in the field and trained on data from simulated falls [8]. This should ultimately lead to fall detection systems that can improve workplace safety by being able to forecast falls reliably and thereby reduce harm and notify emergency services or by detecting patterns in near-fall events throughout the workday, which can then be combated with traditional fall prevention measures.
To identify the optimal strategy, the study employs a systematic evaluation of multiple neural network architectures. This approach involves fine-tuning network configurations to determine the best-performing models for the dataset, which features data from 17 IMU sensors. Furthermore, the research focuses on optimizing performance scores independently of hardware constraints, inference time, or training resources. This enables an upper bound to be established for near-fall detection accuracy, based on the dataset’s resolution. The findings will also serve as a benchmark for future studies that aim to reduce sensor requirements or computational demands.

2. State-of-the-Art STF Data and Detection

2.1. Established Fall and Near-Fall Datasets

Fall detection is a well-established research domain supported by datasets such as SisFall [9], UP-Fall [10], and KFall [11]. In addition to these established datasets, however, many studies rely on data collected independently [12,13]. This reliance introduces challenges, including limited participant numbers and the prevalence of simulated falls [7]. Simulated falls, being biomechanically distinct from real falls, reduce the ecological validity and applicability of the findings derived from such datasets [8]. Near-fall detection, while it remains a growing field, faces similar limitations. Previous studies often feature small participant groups and rely on simulated near-falls, limiting the robustness and generalizability of results.
To address these issues, Hellmers et al. utilized data from the SeFallED and CareFall studies, which featured spontaneous, unpredictable near falls induced on a perturbation treadmill. Participants were equipped with six IMU sensors capturing accelerometer and gyroscope data. These studies included a substantial sample size of 87 participants with a median age of 73 years, making them noteworthy contributions to the field [14]. However, the high median age reflects a broader trend in fall research to focus predominantly on older adults, owing to their increased risk of falling. This age bias poses challenges for generalizability. As Sucerquia, López, and Vargas-Bonilla demonstrated, biomechanical differences in the fall data between young and older participants are significant, which can impact classifier performance when this data is applied across age groups [9].
This study relies on the Prev-Fall dataset to address this gap, providing an ideal resource for studying near falls among working-age individuals [7,15]. This dataset includes data from 110 participants, equipped with 17 IMUs that capture accelerometer, gyroscope, and magnetometer data. This level of detail enables a comprehensive analysis of biomechanical processes during both the near-fall and recovery phases, making it a valuable asset for advancing near-fall detection systems.

2.2. State-of-the-Art Fall and Near-Fall Detection

Fall and near-fall detection is typically framed as a binary classification problem, where models must identify whether or not an activity constitutes a fall/near fall. In some cases, this is extended to a three-class problem, distinguishing between falls, near falls, and activities of daily living (ADLs) [5]. To be relevant in real-world applications, fall detection must operate in real-time or near real-time, enabling timely interventions such as alerting emergency services [16]. Two primary approaches exist for real-time fall detection. Threshold-based classifiers, which immediately output results when predefined thresholds are exceeded as single points of data arrive, offer simple and fast solutions but lack context. A more robust alternative involves the use of a sliding window to preprocess data, allowing models to generate intermediary predictions as chunks of data stream in [17]. When near-fall detection is combined with fall detection, real-time processing is essential, to provide situational awareness and timely alerts. However, in scenarios such as workplace safety monitoring, near-fall detection can also be performed offline to analyze temporal or spatial patterns.
Threshold-based classifiers, while historically prevalent, have become less common in recent studies, owing to their limitations in handling complex patterns [9]. Modern approaches increasingly rely on machine-learning techniques, such as k-nearest-neighbor (k-NN) classifiers [18,19,20], decision trees [18,19,21], and support vector machines (SVMs) [11,14,19]. These models often require feature engineering, demanding significant domain expertise. By contrast, neural networks excel at learning features directly from raw data, a critical advantage in near-fall detection, where no standardized ergonomic definition exists [22]. Their adaptability makes them ideal for capturing complex biomechanical patterns. Examples of such neural networks are:
  • Convolutional neural networks (CNNs): CNNs are effective at detecting local patterns in data. Variants such as the tiny-CNN [23] and DAG-CNN [24] demonstrate impressive speed and accuracy in fall detection tasks;
  • Residual neural networks (ResNets): ResNets address the vanishing gradient and degradation problems [25] and have been successfully applied to fall detection [26];
  • Hybrid architectures such as the convolutional long short-term memory (convLSTM) combine the advantages of both local pattern recognition and time dependency tracking [14];
  • InceptionTime: designed specifically for time-series classification. InceptionTime offers strong potential for fall and near-fall detection with the use of wearable sensors, though it remains underexplored in this domain [2,27].
Although transformer and diffusion-based models have shown encouraging results in natural language processing and computer vision, their applicability to wearable sensor-based fall detection is still largely unknown. Furthermore, their high processing complexity makes them less useful for real-time occupational safety applications. Leveraging the characteristics of the Prev-Fall dataset, which features high-resolution data from 17 IMUs, this study will perform a neural architecture search to evaluate the performance of CNN, ResNet, convLSTM, and InceptionTime architectures. This approach aims to identify optimal models tailored to the dataset’s unique properties, contributing to the advancement of near-fall detection systems.

3. Research Question and Hypothesis

The primary research question addressed in this study is: “How effectively can machine learning approaches classify near-fall events”? The research seeks to evaluate the potential of various machine-learning architectures to enhance the detection and prevention of STFs in workplace environments. To answer this question, the study formulates the following key hypotheses.
First, it is hypothesized that machine-learning algorithms can accurately classify near-fall events. Previous research has shown that machine-learning models outperform threshold-based methods in fall detection tasks. Given the complexity and variability of near-fall events, it is expected that machine-learning models, particularly those employing deep-learning techniques, will demonstrate high performance in distinguishing near-falls from baseline activities.
Second, it is hypothesized that real-world data will enhance the generalizability of near-fall detection models. Many existing studies rely on simulated data, which often fails to fully represent the complexity of real-world STF incidents. By using the Prev-Fall dataset, based on data collected in real workplace environments, this study expects the models to exhibit greater robustness and adaptability across different scenarios and populations. Models trained on real-world data are anticipated to outperform those based on simulated data in terms of accuracy and applicability.
In pursuit of these two hypotheses, this study will experiment with different machine-learning architectures, including CNNs, InceptionTime, convLSTM, and ResNets. The aim is to determine their effectiveness in classifying near-fall events and test the best architectures on data foreign to the Prev-Fall dataset, thereby gauging their generalization power. By rigorously testing and comparing these architectures, the study aims to identify the most effective approach to near-fall detection and advance the implementation of machine-learning-based systems for the purpose of improving workplace safety.

4. Methods

4.1. Selection and Recruitment of Participants

Participants in the Prev-Fall dataset used in this study were selected and recruited as part of the ENTRAPon research project, which is concerned with high-risk sites for STFs. The cohort consisted of 110 workers aged between 19 and 63, 55 of whom were steelworkers and 55 parcel delivery drivers. Participants had to be in good health and able to perform the exercises without assistance; persons with physical or neurological impairments were not allowed to participate. The study design was reviewed and approved by a committee to ensure compliance with ethical research standards. Furthermore, all participants provided informed consent prior to their inclusion in the study. The study on kinematic responses to gait perturbations (already uploaded) provides details on the recruitment procedure and participant requirements [7,15].

4.2. Construction of STF Course and Sensors

A specially constructed STF course was used in the experiment for the creation of the Prev-Fall dataset to replicate real-life STFs in a controlled environment. The course consisted of a 15-m-long boardwalk with eighteen integrated STF traps that caused unexpected STFs. These were divided into three categories: slips, trips, and missteps [15]. They were arranged in such a way that participants could not predict their location or type. All participants wore a full-body measurement suit equipped with 17 Xsens Link IMUs (Enschede, The Netherlands), positioned according to the standard Xsens placement protocol to ensure consistent coordinate system alignment across individuals [7,15]. This standardization minimizes inter-subject variability in sensor orientation and ensures biomechanically consistent data acquisition. These sensors collected magnetometer, gyroscope, and accelerometer data at a sampling rate of 120 Hz [7,15]. To the best of our knowledge, there is no comparable real-world dataset that captures near-fall incidents in occupational settings. This dataset was specifically designed to address the limitations of previous studies that rely on simulated falls, thus ensuring greater ecological validity [7,15].

4.3. Data Exploration

In order to further deepen the understanding of the Prev-Fall dataset [7], we first used a non-class-specific (S, T, F, and walking) linear discriminant analysis [28,29] to reduce the dimensionality of our data and visualize it. For this approach, we considered the sensor data and the ergonomic metrics separately, creating plots for each.
Furthermore, we performed a detailed correlation analysis, employing cross-correlation to gauge the importance of features for classification: a relevant feature has a high correlation with the target variable but a low correlation with other features [30]. The analysis also shows the potential impact of multicollinearity on the classification of this dataset [31] by calculating the level of the features’ mutual correlation. Lastly, cross-correlation has been used to detect data overlap by showing the level of similarity between the time series from different classes. Similar time-series data belonging to different classes have the potential to complicate the finding of a clean decision boundary [32]. The cross-correlation analysis was conducted to evaluate three primary aspects:
  • Between-feature correlation: first, we grouped the sensor data based on biomechanical characteristics: upper body, lower body, head, and others (e.g., CoM). Second, we calculated a pairwise cross-correlation between all features of one STF trial (e.g., Trial 1: left toe acceleration vs. right wrist velocity), then summed and averaged the cross-correlations across all STF trials. This resulted in a very large correlation matrix. We, therefore, further summed and averaged them across the aforementioned biomechanical characteristics and created a correlation dendrogram;
  • Within-class correlation: we calculated the average multivariate cross-correlation between each pair of STF trials from the same STF class;
  • Between-class correlation: we calculated the average multivariate cross-correlation for each pair of STF trials from different classes.
Specifically, we calculated the zero-normalized cross-correlation in order to detect structural patterns, taking the maximum correlation value for all valid time shifts. The calculation was performed using ObsPy [33].

4.4. Neural Architecture Search

In order to find a high-performing network architecture for the Prev-Fall dataset, we systematically tested different configurations of suitable network types for different window sizes. The windows contain 493 channels made up of Xsens data, including acceleration, velocity, angular acceleration and angular velocity, body angles, sensor and CoM positions, and foot contact information. We ran experiments with 500 ms, 1000 ms, and 1500 ms windows with a step size of 50% and tagged a window as belonging to an STF event if it contained 30 data points of the STF event. For the sampling rate of 120 Hz, this is equivalent to 250 ms of the STF event. To balance the significantly larger number of windows containing baseline walking on the STF course with the windows containing STF events, we rebalanced the dataset before training, and windows were randomly drawn until all classes were the size of the smallest class. We subsequently split the data randomly into 70% training and 30% validation datasets.
For the neural architecture search, we implemented an updated version of the mcfly automated deep learning on the time series framework [34]. We trained 50 different CNNs, ResNets, convLSTM, and InceptionTime neural networks for up to 20 epochs. We used early stopping when no improvement in the validation loss was observed for 5 successive epochs. We adjusted the different parameters equivalent to [34] according to the network type.
For all four network types, we utilized an adam optimizer [35] and used random search to find suitable values for weight decay (between 0.1 and 0.0001) and learning rate (between 0.1 and 0.0001). For CNNs, we varied the number of convolutional layers, the number of filters in the convolutional layers, and the number of nodes in the following fully connected layer. For the ResNets, we varied the number of convolutional blocks, the number of filters, and their size. For InceptionTime, we varied the number of inception blocks, the number of filters, and their size. For deepConvLSTMs, we varied the number of initial convolutional layers, the convolutional filter size, the number of lstm layers, and the number of features in the hidden state. The ranges of all network-specific parameters can be found in Appendix A, Table A1.
To measure network performance, we calculated the accuracy and the weighted F1 score [36] for the training and validation data.

4.5. Testing Networks on Foreign Data

Following the neural architecture search, we tested the top three networks of each network type and from all three window sizes on 25 h of as-yet-unseen data from a different dataset recorded with the same Xsens Link system and with 17 IMUs. These data were recorded in the IFA’s ergonomics laboratory and depict a range of ADLs, from walking and filling shelves at different heights with packages of different sizes and weights to using a cordless screwdriver at different heights [37]. The data are not labeled with ADL events, nor do they contain labeled STF events. The activities performed in these recordings are distinctly different from the activities in our STF training and validation set, in addition to the normal activity of walking around. We ran this test to gauge the generalization power of the trained networks with respect to the data from as-yet-unseen environments and to gauge the false-positive rate for as-yet-unseen activities. We, therefore, calculated the false-positive rate as a quantitative performance metric and checked the timestamps of the windows in which the false positives occurred to gain a qualitative impression of which activities lead to false positives.

5. Results

5.1. Data Exploration

To evaluate the quality and redundancy of the sensor data, we analyzed the between-feature correlation. The analysis particularly revealed high correlations between features within the same anatomical region (e.g., left foot and left toe; between the different vertebrae) and for related metrics in the same anatomical region (e.g., left foot acceleration and left foot velocity, joint angles of vertebrae). The correlation between features from different anatomical regions ranged from −0.65 to 0.99 but was on average low, at −0.06 for negative and 0.16 for positive correlations. A Dendrogram of averaged correlations between the different feature groups can be seen in Figure 1.
For the purpose of the neural architecture search, we decided to retain the few features with high intra-group correlations, as they provide complementary information on anatomy and are, thus, critical for model performance and interpretability. Redundancy was intentionally preserved to ensure a comprehensive representation of biomechanical processes, leveraging the robustness of machine-learning models such as InceptionTime and ResNet against correlated inputs.
The results of the intra-class and inter-class correlateon analysis can be seen in Table 1 and Table 2, respectively. They show that the multivariate zero-normalized cross-correlations within a class are only marginally higher than between classes and could suggest that linear relationships are not sufficient for a proper understanding of the complexity of the data.
The two graphs in Figure 2 present the results of a linear Discriminant Analysis (LDA) applied to the Prev-Fall dataset, reducing dimensions and thus yielding distinct clusters for the different event types-namely baseline, misstep, trip, and slip. The primary objective of these visualizations is to assess the extent to which LDA effectively separates these classes in a reduced three-dimensional space, where each axis (LD1, LD2, and LD3) represents one of the discriminant components. The LDA components are not used as the input in further data analysis steps or the neural architecture search. Rather, LDA was employed to reduce the high-dimensional feature space to a lower-dimensional representation, enabling intuitive visualization and facilitating the interpretability of class separability, with the objective of identifying the underlying patterns within the data. The two graphs in Figure 2 show the same LDA from different perspectives. A relatively high separation between the classes is observed, particularly for the trip class (green), which is clearly isolated from the other classes. The Slip (purple) and Misstep (red) classes appear to be more densely packed than the Baseline (blue) class, which scatters the most strongly. Compared to the clear separation seen from the trip cluster, the other three clusters are, at any rate, close together and less differentiated.

5.2. Neural Architecture Search

For this study, we implemented a neural architecture search framework, which facilitated automated training and hyperparameter optimization to solve our time-series classification problem. This framework enabled the efficient exploration of multiple deep-learning architectures, including CNN, ResNet, convLSTM, and InceptionTime, with the goal of identifying the most accurate models for detecting near-fall events. Our focus lay on using Xsens sensor data, and we excluded the advanced ergonomic metrics present in the Prev-Fall dataset.
The three best-performing models independent of window size can be seen in Table 3.
Table A2, Table A3 and Table A4 in Appendix B summarize the top-performing models across different model types, ranked by their highest validation F1 score. For each model type, the three models with the best validation performances are reported. Additionally, the tables provide the initial learning rate, which influences the convergence speed, and the regularization rate, which helps to reduce overfitting and improve generalizability.
Each model’s architecture type is specified, distinguishing models according to structural characteristics. Key performance metrics, such as the best F1 score and accuracy achieved in both training and validation, reveal the models’ effectiveness and generalization capability.
CNNs and InceptionTime demonstrated exceptional consistency over all of the window lengths, with the models achieving higher F1 scores correlating with better generalization capabilities and lower validation losses. The model achieving the highest performance in terms of weighted F1 score on the validation dataset is the InceptionTime model, which attained a validation F1 score of 0.9632 at a window size of 500 ms, potentially attributed to its architectural design and hyperparameter tuning, including a learning rate of 0.000197 and a regularization rate of 0.013995.
Some models exhibited slight discrepancies between training and validation metrics. Overall, however, these architectures showcased robustness and reliability in feature extraction and classification tasks.
Figure 3 shows that the different window sizes impacted the weighted f1 score on the validation dataset only slightly. Figure A1, Figure A2, Figure A3 and Figure A4 in Appendix C further support this observation. They illustrate the progression of training and validation accuracy over epochs for various neural network architectures. The figures in Appendix C are organized by model types, such as CNN and InceptionTime, and further differentiated by time intervals of 500 ms, 1000 ms, and 1500 ms.
Figure 4 shows the correlation of the learning and regularization rates with the evaluation metrics of the neural architecture search. All evaluation metrics are highly correlated with each other. More importantly, the learning rate correlates negatively with the evaluation metrics, and the regularization rate correlates slightly negatively with the evaluation metrics. The correlation between the learning and regularization rates is close to zero and spurious, since both are randomized independently.

5.3. Test on Foreign Data

The results of testing the networks from the neural architecture search on the Prev-Fall dataset on foreign data can be seen in Table 4, Table 5 and Table 6. The false-positive rate of the networks ranged from 10.1 to 27.8 percent, with performances being considerably better for 500 ms than 1000 ms and 1500 ms windows. Manual investigation revealed that false positives occurred particularly during movements such as bending forward (e.g., lifting boxes), squatting or kneeling (e.g., setting objects down), and raising the hands—especially when the arms were extended outward.

6. Discussion

6.1. Insights from the Data Analysis

The correlation analysis provided important insights into the interdependence of features across different sensor groups, revealing both redundancy and complementary relationships. As expected, we found especially high correlations for the same metrics in the same region (i.e., velocity of left foot and left toe) and for related metrics such as velocity and acceleration. However, when different metrics from different regions were compared, the averages were generally low (−0.06 for negative and 0.16 for positive correlations).
The low correlation values across most feature groups highlight the independence of these features, underscoring their potential complementary value in classification. For instance, the moderate correlations with only the pelvic region of external features (“Others”) suggest that, while these features may not independently predict STF events, their inclusion probably enhances the robustness of the models by capturing additional variance in the data. The observed low correlations indicate that models such as InceptionTime and CNNs, which excel at feature extraction from raw and heterogeneous inputs, are well-suited for this dataset. Conversely, models requiring manual feature engineering (e.g., SVMs or decision trees) might struggle to leverage this complexity, as they rely on pre-selected, highly correlated features. Future work could investigate whether dimensionality reduction techniques such as PCA or feature selection based on correlation thresholds improve computational efficiency without sacrificing model performance.
The linear discriminant analysis (LDA) demonstrated varying degrees of class separability, with the trip class showing consistent isolation across projections (Figure 2). This indicates that the kinematic signatures of trips are distinct, probably owing to unique biomechanical features such as abrupt forward shifts in the CoM. In contrast, misstep and slip events exhibited significant overlap, reflecting shared instability characteristics such as lateral CoM shifts or irregular gait recovery patterns. The baseline class, which consistently overlaps with both missteps and slips, further complicates classification. This overlap can be attributed to dynamic movements during normal walking, such as rapid directional changes or decelerations, which mimic early indicators of instability. While LDA captured meaningful separability, the overlaps between missteps, slips, and baseline activities highlight the need for additional features or nonlinear techniques. The observed overlaps between missteps, slips, and baseline activities indicate that a simple linear separation of classes is not feasible. Nonlinear approaches, such as neural networks, are therefore required to capture the complex, nonlinear relationships present in the data. The correlation analysis suggests that including biomechanical variables such as joint torques, stride asymmetry, or temporal patterns in CoM trajectory could enhance class separability. Additionally, nonlinear dimensionality reduction methods such as t-SNE [38] could complement LDA by better capturing the nonlinear relationships between features.

6.2. Neural Network Performances

Overall, we found that the network types we identified as popular for fall and near-fall detection work well on the Prev-Fall dataset, with no architecture being completely unable to tackle the problem. While the performance of the neural networks on our dataset is lower than on datasets discussed in Section 2, this can be explained by the higher difficulty of detecting near-falls—a more subtle event than falls—and by the differentiation between multiple types of near falls rather than the issue being treated as a binary detection problem.
The neural architecture search highlighted distinct strengths and weaknesses across model types and time windows (Figure 3, Table A2, Table A3 and Table A4). CNNs performed exceptionally well for short time intervals (500 ms), achieving validation F1 scores of up to 0.954. Their ability to capture localized patterns, such as sudden spikes in acceleration or angular velocity, makes them ideal for real-time applications. Surprisingly, their performance barely declined for longer windows (1000 ms, 1500 ms), suggesting that the CNNs’ difficulties in integrating temporal dependencies over extended periods did not meaningfully come into effect here. InceptionTime consistently outperformed other models for longer intervals (1000 ms and 1500 ms), with F1 scores approaching 0.96, but it also performed the best overall for a window size of only 500 ms. Its multiscale temporal feature extraction capabilities probably enabled it to capture both short-term and long-term patterns that are critical for distinguishing between STF events. The architecture’s scalability and robust generalization make it a strong candidate for offline analysis or retrospective evaluations of workplace safety. ResNets demonstrated stable performance across all time intervals, but did not match the F1 scores of InceptionTime. While ResNets effectively addressed issues such as vanishing gradients, their limited ability to capture multiscale temporal patterns may explain their slightly lower performance for STF classification. An analysis of false positives revealed that biomechanical movements such as bending and squatting are often misclassified because their sensor patterns resemble actual falls. This suggests that existing models lack kinematic context integration, which is essential for distinguishing controlled movements from near-fall events. Future improvements could incorporate context-aware kinematic features, such as motion dynamics and postural stability, along with task-specific biomechanical constraints to improve classification accuracy and reduce false positives.
The hyperparameter correlation analysis (Figure 4) revealed that medium learning rates (0.001–0.005) and regularization rates (0.001–0.01) were associated with the best performance. Excessively high learning rates led to unstable models, while low regularization occasionally resulted in overfitting. These findings emphasize the importance of careful hyperparameter tuning to balance model stability and generalization.

6.3. False Positives and Generalization

Testing the models resulting from the neural architecture search on as-of-yet-unseen data gives us insights into their generalization abilities. Since this test data was not labeled and did not contain near-falls either way, we measured the false-positive rate to gauge how many false alarms the best networks would send. While the models exhibited promising generalization capabilities on unseen foreign data, capable of achieving false-positive rates below 20%, this would still be too high for real-world applications without further techniques to eliminate erroneous alarms.
False positives were observed, particularly during biomechanically complex movements such as bending, squatting, or raising the hands. These activities involve significant shifts in the CoM or angular momentum, which resemble instability patterns seen in true near falls. For example, bending forward to lift an object causes the CoM to move anteriorly relative to the base of support (BoS), mimicking the destabilization observed during actual falls. Similarly, raising the hands while extending the arms redistributes mass, creating kinematic patterns that can be mistaken for balance recovery attempts. These findings highlight a critical challenge for real-world deployment: distinguishing between stability-critical movements and true near falls. Incorporating additional contextual data, such as task-specific information or environmental cues, as well as additional ADL data, could help reduce false positives. Alternatively, hybrid approaches that combine machine learning with rule-based systems as common in the fields of computer vision and natural language processing may improve classification reliability by integrating domain knowledge into the decision-making process.

7. Conclusions

The findings of this study demonstrate the significant potential of advanced machine-learning models for near-fall detection, particularly in real-world occupational settings. By leveraging high-quality kinematic data from actual STF events, this research bridges critical gaps left by prior studies that relied heavily on simulated datasets or were narrowly focused on older populations. The inclusion of diverse participants, representing physically demanding professions, such as parcel delivery workers and steelworkers, underscores the broader applicability of the proposed methods to working-age individuals—a demographic often under-represented in STF research.
One of the primary practical contributions of this study lies in its evaluation of neural network architectures tailored to time-series data. The consistent performance of CNNs for shorter time intervals highlights their suitability for real-time near-fall detection systems, which are critical in workplace environments where timely intervention can prevent injuries. CNNs’ ability to operate with lightweight computational requirements further strengthens their feasibility for integration into wearable devices, such as smart clothing or footwear with embedded sensors, enabling continuous monitoring without imposing a significant hardware burden. In contrast, the superior performance of InceptionTime models across longer time intervals positions them as ideal candidates for offline analysis or retrospective safety evaluations. This capability is especially valuable for post hoc investigations, allowing safety managers to identify temporal or spatial patterns in STF events and design targeted interventions. For instance, near-fall hot spots within a facility or recurrent hazardous behaviors can be identified and mitigated, improving workplace safety protocols over time.
Despite these advances, challenges remain. The overlap between baseline activities and near falls, as revealed by the LDA results, emphasizes the need for more sophisticated feature engineering or multimodal integration. Future work should explore the inclusion of external contextual data, such as environmental information (e.g., surface type or incline) or additional sensor modalities (e.g., visual data from cameras and force sensors embedded in flooring), as well as the integration of additional contextual features, such as task-specific metadata or ergonomic risk factors, to improve model robustness. In addition, a hybrid approach combining feature engineering and deep learning may help mitigate false positives in biomechanically complex activities. These enhancements could provide a richer context for distinguishing between complex stability-critical movements and true near-fall events, thereby reducing false positives.
Scalability also represents a key area of focus. While the Prev-Fall dataset provides a robust foundation, extending the dataset to include a broader range of occupations, environmental conditions, and demographic profiles would enhance the generalizability of the models. Such expansions would not only improve the reliability of near-fall detection systems across industries but also pave the way for developing population-specific models tailored to different worker groups or workplace risks.
The practical implications of this work extend beyond workplace safety. The methodologies and insights presented here have the potential to inform near-fall detection systems in other domains, such as elder care or rehabilitation. For instance, applying these models in home settings could provide continuous monitoring for at-risk individuals, offering early interventions and personalized fall prevention strategies.
In summary, this study represents a significant step forward in the development of near-fall detection systems, addressing critical gaps in prior research and setting new benchmarks for model performance in real-world settings, and represents one of the first systematic evaluations of deep-learning models for near-fall detection using a real-world dataset, demonstrating the limitations of existing approaches in handling biomechanically complex movements. Our findings provide critical insights for developing robust and context-aware fall prevention systems. The integration of cutting-edge machine-learning architectures with high-quality kinematic data ensures both accuracy and practical applicability. As these systems continue to evolve, their deployment in occupational settings has the potential to drastically reduce STF-related injuries, improving worker safety and productivity on a global scale. By laying the foundation for more robust, generalizable, and context-aware detection systems, this research contributes a transformative framework for advancing occupational safety and injury prevention.

Author Contributions

Conceptualization, M.S.; Methodology, M.S.; Software, M.S. and K.S.-R.; Investigation, K.S.-R.; Writing—original draft, M.S. and K.S.-R.; Writing—review & editing, U.F.; Supervision, A.F. and U.F.; Project administration, M.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this study are available on request from the corresponding author. The data are currently not publicly available, as they still need to be technically prepared for dissemination and, for legal reasons, all participating institutions must be involved in any dissemination.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Table containing the hyperparameters randomized for the different model architectures during the neural architecture search.
Table A1. Table containing the hyperparameters randomized for the different model architectures during the neural architecture search.
ModelHyperparameterRange
CNNNumber of Layers1–10
CNNNumber of Filters10–100
CNNNumber of Fully Connected Nodes10–2000
ResNetNumber of Residual Blocks2–5
ResNetNumber of Filters32–128
ResNetKernel Size8–32
deepConvLSTMNumber of Conv Layers1–10
deepConvLSTMNumber of Conv Filters10–100
deepConvLSTMNumber of LSTM Layers1–5
deepConvLSTMSize of LSTM Hidden State10–100
InceptionTimeNumber of Inception Modules3–6
InceptionTimeNumber of Filters32–96
InceptionTimeKernel Size10–100

Appendix B

Table A2. Results of neural architecture search with window length of 500 ms.
Table A2. Results of neural architecture search with window length of 500 ms.
A.ModelLearning RateRegulation RateBest Train F1Best Val F1Best Train AccuracyBest Val. AccuracyBest Train LossBest Val. Loss
500 msCNN0.00500.00150.94960.95420.94340.92970.16210.1660
CNN0.00050.00100.98110.95360.97480.92910.09680.1871
CNN0.00060.00400.97740.95190.97110.92740.10610.1919
inception time0.00010.01390.96400.96320.95780.93850.79600.7998
inception time0.00090.00010.96470.95730.95850.93260.77910.7904
inception time0.00110.00020.96060.95500.95440.93030.78530.7925
ResNet0.00070.00010.96440.95010.9580.92560.78150.7954
ResNet0.00080.00080.94000.94150.93390.91740.80570.8048
ResNet0.00150.00320.94220.94150.93610.91740.81570.8124
deepConvLSTM0.00010.00220.91950.91660.91340.89280.27180.3176
deepConvLSTM0.00020.00130.91290.91190.90720.88810.30590.2910
deepConvLSTM0.00040.00020.90280.91140.89680.88750.29400.3012
Table A3. Results of neural architecture search with window length of 1000 ms.
Table A3. Results of neural architecture search with window length of 1000 ms.
ModelLearning RateRegulation RateBest Train F1Best Val F1Best Train AccuracyBest Val. AccuracyBest Train LossBest Val. Loss
1000 msCNN0.00300.00230.95890.95880.94870.91230.14920.1588
CNN0.00070.00100.98140.95810.97100.91150.0990.1720
CNN0.00040.01100.96530.95230.95510.90610.18410.2152
inception time0.00250.00030.96540.95900.95510.91230.78100.7904
inception time0.00160.00010.96610.95390.95590.90760.77920.789
inception time0.00020.01340.95490.95390.94480.90760.80830.8189
ResNet0.00060.00220.95650.95310.94630.90680.79820.8038
ResNet0.00090.00020.96480.95230.95450.90610.78210.7950
ResNet0.00010.00100.97490.95140.96450.90530.77470.8011
deepConvLSTM0.00780.00010.92030.90230.91010.85830.28840.3421
deepConvLSTM0.00020.00020.83560.82460.82720.78320.48650.5048
deepConvLSTM0.00030.00090.79230.77060.78450.73390.62430.6619
Table A4. Results of neural architecture search with window length of 1500 ms.
Table A4. Results of neural architecture search with window length of 1500 ms.
ModelLearning RateRegulation RateBest Train F1Best Val F1Best Train AccuracyBest Val. AccuracyBest Train LossBest Val. Loss
1500 msCNN0.00070.00020.99660.96220.99170.92980.06720.1875
CNN0.00100.00060.9860.95590.98150.92360.07210.183
CNN0.00050.00130.98740.95490.98260.92270.11400.1872
inception time0.00140.00170.97100.95770.96620.92540.78660.7940
inception time0.00140.00160.9667670.9549280.9620170.9227350.7909440.802259
inception time0.00120.00010.97430.95400.96950.92180.77250.7919
ResNet0.00010.00560.97450.95300.96970.92090.77720.8029
ResNet0.00400.00010.97120.95210.96640.92000.77560.7958
ResNet0.00010.00060.97360.95030.96890.91820.77280.7973
deepConvLSTM0.00450.00020.88350.89010.87930.86050.35480.3344
deepConvLSTM0.00020.00020.83570.75030.83090.72020.45450.7359
deepConvLSTM0.00260.00010.70980.73220.71100.71310.60110.5363

Appendix C

Figure A1. Validation and training accuracy progression during the training of the best-performing InceptionTime model for all three window lengths.
Figure A1. Validation and training accuracy progression during the training of the best-performing InceptionTime model for all three window lengths.
Sensors 25 01468 g0a1
Figure A2. Validation and training accuracy progression during the training of the best-performing CNN model for all three window lengths.
Figure A2. Validation and training accuracy progression during the training of the best-performing CNN model for all three window lengths.
Sensors 25 01468 g0a2
Figure A3. Validation and training accuracy progression during the training of the best-performing ResNet model for all three window lengths.
Figure A3. Validation and training accuracy progression during the training of the best-performing ResNet model for all three window lengths.
Sensors 25 01468 g0a3
Figure A4. Validation and training accuracy progression during the training of the best-performing convLSTM model for all three window lengths.
Figure A4. Validation and training accuracy progression during the training of the best-performing convLSTM model for all three window lengths.
Sensors 25 01468 g0a4

References

  1. Deutsche Gesetzliche Unfallversicherung e.V. Statistik Arbeitsunfallgeschehen 2023; Deutsche Gesetzliche Unfallversicherung e.V.: Berlin, Germany, 2024. [Google Scholar]
  2. Wang, S.; Miranda, F.; Wang, Y.; Rasheed, R.; Bhatt, T. Near-Fall Detection in Unexpected Slips during Over-Ground Locomotion with Body-Worn Sensors among Older Adults. Sensors 2022, 22, 3334. [Google Scholar] [CrossRef] [PubMed]
  3. Nikolov, I.; Liu, J.; Moeslund, T. Imitating Emergencies: Generating Thermal Surveillance Fall Data Using Low-Cost Human-like Dolls. Sensors 2022, 22, 825. [Google Scholar] [CrossRef]
  4. Hsu, W.-W.; Guo, J.-M.; Chen, C.-Y.; Chang, Y.-C. Fall Detection with the Spatial-Temporal Correlation Encoded by a Sequence-to-Sequence Denoised GAN. Sensors 2022, 22, 4194. [Google Scholar] [CrossRef] [PubMed]
  5. Rahemtulla, Z.; Turner, A.; Oliveira, C.; Kaner, J.; Dias, T.; Hughes-Riley, T. The Design and Engineering of a Fall and Near-Fall Detection Electronic Textile. Materials 2023, 16, 1920. [Google Scholar] [CrossRef]
  6. Schumann, P.; Scholz, M.; Trentzsch, K.; Jochim, T.; Śliwiński, G.; Malberg, H.; Ziemssen, T. Detection of Fall Risk in Multiple Sclerosis by Gait Analysis—An Innovative Approach Using Feature Selection Ensemble and Machine Learning Algorithms. Brain Sci. 2022, 12, 1477. [Google Scholar] [CrossRef] [PubMed]
  7. Schneider, M.; Reich, K.; Hartmann, U.; Hermanns, I.; Kaufmann, M.; Kluge, A.; Fiedler, A.; Frese, U.; Ellegast, R. Acquisition of Data on Kinematic Responses to Unpredictable Gait Perturbations: Collection and Quality Assurance of Data for Use in Machine Learning Algorithms for (Near-)Fall Detection. Sensors 2024, 24, 5381. [Google Scholar] [CrossRef]
  8. Stack, E. Falls are unintentional: Studying simulations is a waste of faking time. J. Rehabil. Assist. Technol. Eng. 2017, 4, 2055668317732945. [Google Scholar] [CrossRef]
  9. Sucerquia, A.; López, J.D.; Vargas-Bonilla, J.F. SisFall: A Fall and Movement Dataset. Sensors 2017, 17, 198. [Google Scholar] [CrossRef]
  10. Martínez-Villaseñor, L.; Ponce, H.; Brieva, J.; Moya-Albor, E.; Núñez-Martínez, J.; Peñafort-Asturiano, C. UP-Fall Detection Dataset: A Multimodal Approach. Sensors 2019, 19, 1988. [Google Scholar] [CrossRef]
  11. Yu, X.; Jang, J.; Xiong, S. A Large-Scale Open Motion Dataset (KFall) and Benchmark Algorithms for Detecting Pre-impact Fall of the Elderly Using Wearable Inertial Sensors. Front. Aging Neurosci. 2021, 13, 692865. [Google Scholar] [CrossRef]
  12. Ramachandran, A.; Karuppiah, A. A Survey on Recent Advances in Wearable Fall Detection Systems. BioMed Res. Int. 2020, 2020, 2167160. [Google Scholar] [CrossRef]
  13. Wang, X.; Ellul, J.; Azzopardi, G. Elderly Fall Detection Systems: A Literature Survey. Front. Robot. AI 2020, 7, 71. [Google Scholar] [CrossRef] [PubMed]
  14. Hellmers, S.; Krey, E.; Gashi, A.; Koschate, J.; Schmidt, L.; Stuckenschneider, T.; Hein, A.; Zieschang, T. Comparison of machine learning approaches for near-fall-detection with motion sensors. Front. Digit. Health 2023, 5, 1223845. [Google Scholar] [CrossRef] [PubMed]
  15. Schneider, M.; Weber, A.; Kaufmann, M.; Kluge, A.; Hartmann, U.; Karamanidis, K.; Ellegast, R. Generation of Consistent Slip, Trip and Fall Kinematic Data via Instability Detection and Recovery Performance Analysis for Use in Machine Learning Algorithms for (Near) Fall Detection. In Digital Human Modeling and Applications in Health, Safety, Ergonomics and Risk Management-14th International Conference; Volume 14029 of Lecture Notes in Computer Science; Springer Nature: Cham, Switzerland, 2023; pp. 298–305. [Google Scholar]
  16. Núñez-Marcos, A.; Arganda-Carreras, I. Transformer-based fall detection in videos. Eng. Appl. Artif. Intell. 2024, 132, 107937. [Google Scholar] [CrossRef]
  17. Rodriguez, J.; Mercuri, M.; Karsmakers, P.; Soh, P.J.; Leroux, P.; Schreurs, D. Automatic fall detector based on sliding window principle. In Proceedings of the 34th WIC Symposium on Information Theory in the Benelux and the Third joint WIC/IEEE SP Symposium on Information Theory and Signal Processing in the Benelux, Louvain-la-Neuve, Belgium, 19–20 May 2016; Werkgemeenschap voor Informatie-en Communicatietheorie (WIC): Heverlee, Belgium, 2013; pp. 215–219. [Google Scholar]
  18. Ntanasis, P.; Pippa, E.; Özdemir, A.T.; Barshan, B.; Megalooikonomou, V. Investigation of Sensor Placement for Accurate Fall Detection. In Wireless Mobile Communication and Healthcare, 6th International Conference, MobiHealth 2016, Milan, Italy, 14–16 November 2016, Proceedings 6; Springer International Publishing: New York, NY, USA, 2017; pp. 225–232. [Google Scholar]
  19. Ribeiro, N.F.; Pereira, A.; Figueiredo, J.; Afonso, J.A.; Santos, C.P. A Machine Learning Approach for Near-Fall Detection Based on Inertial and Force Data While Using a Conventional Rollator. In International Conference on NeuroRehabilitation; Springer International Publishing: Cham, Switzerland, 2022; pp. 345–349. [Google Scholar]
  20. Safira, S.M.; Gunawan, P.H.; Kurniawan, W.Y.; Putra, I.G.K.K.; Reganata, G.P.; Patrianingsih, N.K.W.; Dharma, I.G.W.S.; Sugianta, I.K.A.; Udhayana, K.F.R. Comparative Analysis of the K-Nearest Neighbor Method and Support Vector Machine in Human Fall Detection. In Proceedings of the 2023 3rd International Conference on Intelligent Cybernetics Technology & Applications (ICICyTA), Denpasar, Bali, Indonesia, 13–15 December 2023; pp. 148–153. [Google Scholar]
  21. Reddy, I.V.S.; Lavanya, P.; Selvakumar, V. Machine learning classifiers for fall detection leveraging LoRa communication network. Int. J. Reconfigurable Embed. Syst. 2024, 13, 76–84. [Google Scholar] [CrossRef]
  22. Maidan, I.; Freedman, T.; Tzemah, R.; Giladi, N.; Mirelman, A.; Hausdorff, J. Introducing a new definition of a near fall: Intra-rater and inter-rater reliability. Gait Posture 2013, 39, 645–647. [Google Scholar] [CrossRef]
  23. Yu, X.; Park, S.; Kim, D.; Kim, E.; Kim, J.; Kim, W.; An, Y.; Xiong, S. A practical wearable fall detection system based on tiny convolutional neural networks. Biomed. Signal Process. Control 2023, 86, 105325. [Google Scholar] [CrossRef]
  24. Choi, A.; Kim, T.H.; Yuhai, O.; Jeong, S.; Kim, K.; Kim, H.; Mun, J.H. Deep Learning-Based Near-Fall Detection Algorithm for Fall Risk Monitoring System Using a Single Inertial Measurement Unit. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 30, 2385–2394. [Google Scholar] [CrossRef]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  26. Ramanathan, A.; McDermott, J. Fall detection with accelerometer data using Residual Networks adapted to multi-variate time series classification. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar]
  27. Fawaz, H.I.; Lucas, B.; Forestier, G.; Pelletier, C.; Schmidt, D.F.; Weber, J.; Webb, G.I.; Idoumghar, L.; Muller, P.-A.; Petitjean, F. InceptionTime: Finding AlexNet for time series classification. Data Min. Knowl. Discov. 2020, 34, 1936–1962. [Google Scholar] [CrossRef]
  28. Tharwat, A.; Gaber, T.; Ibrahim, A.; Hassanien, A.E. Linear discriminant analysis: A detailed tutorial. AI Commun. 2017, 30, 169–190. [Google Scholar] [CrossRef]
  29. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  30. Gupta, N.; Patel, H.; Afzal, S.; Panwar, N.; Mittal, R.S.; Guttula, S.; Jain, A.; Nagalapatti, L.; Mehta, S.; Hans, S. Data Quality Toolkit: Automatic assessment of data quality and remediation for machine learning datasets. arXiv 2021, arXiv:2108.05935. [Google Scholar]
  31. Lieberman, M.G.; Morris, J.D. The precise effect of multicollinearity on classification prediction. Mult. Linear Regres. Viewp. 2014, 40, 5–10. [Google Scholar]
  32. Santos, M.S.; Abreu, P.H.; Japkowicz, N.; Fernández, A.; Santos, J. A unifying view of class overlap and imbalance: Key concepts, multi-view panorama, and open avenues for research. Inf. Fusion 2023, 89, 228–253. [Google Scholar] [CrossRef]
  33. Beyreuther, M.; Barsch, R.; Krischer, L.; Megies, T.; Behr, Y.; Wassermann, J. ObsPy: A Python toolbox for seismology. Seism. Res. Lett. 2010, 81, 530–533. [Google Scholar] [CrossRef]
  34. van Kuppevelt, D.; Meijer, C.; Huber, F.; van der Ploeg, A.; Georgievska, S.; van Hees, V. Mcfly: Automated deep learning on time series. SoftwareX 2020, 12, 100548. [Google Scholar] [CrossRef]
  35. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  36. Lee, M.C.H.; Braet, J.; Springael, J. Performance Metrics for Multilabel Emotion Classification: Comparing Micro, Macro, and Weighted F1-Scores. Appl. Sci. 2024, 14, 9863. [Google Scholar] [CrossRef]
  37. Deutsche Gesetzliche Unfallversicherung e. V. (DGUV). Evaluation of Physical Loads in Virtual Reality for Conceptual Work Design. 22 August 2023. Available online: https://www.dguv.de/ifa/forschung/projektverzeichnis/ifa0507-2.jsp (accessed on 16 November 2024).
  38. van der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Dendrogram showing the correlation distance between features in different anatomical regions averaged across all trials. The “Others” category contains features that cannot be clearly assigned to specific anatomical regions or sensor modalities. These include whole-body metrics (e.g., center of mass (CoM) measures). The inclusion of “Others” ensures that the analysis captures all potentially relevant feature interrelationships.
Figure 1. Dendrogram showing the correlation distance between features in different anatomical regions averaged across all trials. The “Others” category contains features that cannot be clearly assigned to specific anatomical regions or sensor modalities. These include whole-body metrics (e.g., center of mass (CoM) measures). The inclusion of “Others” ensures that the analysis captures all potentially relevant feature interrelationships.
Sensors 25 01468 g001
Figure 2. LDA of sensor data of the STF dataset. The two graphs are different perspectives of the same plot.
Figure 2. LDA of sensor data of the STF dataset. The two graphs are different perspectives of the same plot.
Sensors 25 01468 g002
Figure 3. Weighted F1 score on the validation dataset of the best-performing model of each model type for the three different window lengths.
Figure 3. Weighted F1 score on the validation dataset of the best-performing model of each model type for the three different window lengths.
Sensors 25 01468 g003
Figure 4. Correlation between hyperparameters and best metrics for ML models.
Figure 4. Correlation between hyperparameters and best metrics for ML models.
Sensors 25 01468 g004
Table 1. Within-class correlation.
Table 1. Within-class correlation.
ClassCorrelation Coefficient
Misstep0.055
Slip0.058
Trip0.073
Table 2. Between-class correlation.
Table 2. Between-class correlation.
Class PairCorrelation Coefficient
Misstep–Slip0.030
Misstep–Trip0.015
Slip–Trip0.040
Table 3. Showing the top 3 networks according to the validation-weighted F1 score from the neural architecture search on the rebalanced Prev-Fall dataset.
Table 3. Showing the top 3 networks according to the validation-weighted F1 score from the neural architecture search on the rebalanced Prev-Fall dataset.
ModelWindow SizeValidation AccuracyValidation F1 Score
InceptionTime500 ms0.93850.9632
CNN1500 ms0.92980.9622
InceptionTime1000 ms0.91230.9590
Table 4. This table shows the false positive rate on as-yet-unseen data spliced into windows with a size of 1500 ms for the best three networks of each network type from the neural architecture search performed on the Prev-Fall-Dataset.
Table 4. This table shows the false positive rate on as-yet-unseen data spliced into windows with a size of 1500 ms for the best three networks of each network type from the neural architecture search performed on the Prev-Fall-Dataset.
ExperimentModelRankFP Rate
1500 msCNN10.2284
CNN20.2785
CNN30.1579
InceptionTime10.1595
InceptionTime20.1596
InceptionTime30.2028
ResNet10.1988
ResNet20.2211
ResNet30.2232
deepConvLSTM10.2151
deepConvLSTM20.2482
deepConvLSTM30.2631
Table 5. This table shows the false positive rate on as-yet-unseen data spliced into windows with a size of 1000 ms for the best three networks of each network type from the neural architecture search performed on the Prev-Fall-Dataset.
Table 5. This table shows the false positive rate on as-yet-unseen data spliced into windows with a size of 1000 ms for the best three networks of each network type from the neural architecture search performed on the Prev-Fall-Dataset.
ExperimentModelRankFP Rate
1000 msCNN10.2230
CNN20.1948
CNN30.2499
InceptionTime10.2069
InceptionTime20.2101
InceptionTime30.1565
ResNet10.2326
ResNet20.1982
ResNet30.1804
deepConvLSTM10.2418
deepConvLSTM20.2310
deepConvLSTM30.2595
Table 6. This table shows the false positive rate on as-yet-unseen data spliced into windows with a size of 500 ms for the best three networks of each network type from the neural architecture search performed on the Prev-Fall-Dataset.
Table 6. This table shows the false positive rate on as-yet-unseen data spliced into windows with a size of 500 ms for the best three networks of each network type from the neural architecture search performed on the Prev-Fall-Dataset.
ExperimentModelRankFP Rate
500 msCNN10.2019
CNN20.1588
CNN30.2028
InceptionTime10.1013
InceptionTime20.1675
InceptionTime30.2000
ResNet10.1813
ResNet20.1862
ResNet30.2121
deepConvLSTM10.2048
deepConvLSTM20.2257
deepConvLSTM30.2549
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Schneider, M.; Seeser-Reich, K.; Fiedler, A.; Frese, U. Enhancing Slip, Trip, and Fall Prevention: Real-World Near-Fall Detection with Advanced Machine Learning Technique. Sensors 2025, 25, 1468. https://doi.org/10.3390/s25051468

AMA Style

Schneider M, Seeser-Reich K, Fiedler A, Frese U. Enhancing Slip, Trip, and Fall Prevention: Real-World Near-Fall Detection with Advanced Machine Learning Technique. Sensors. 2025; 25(5):1468. https://doi.org/10.3390/s25051468

Chicago/Turabian Style

Schneider, Moritz, Kevin Seeser-Reich, Armin Fiedler, and Udo Frese. 2025. "Enhancing Slip, Trip, and Fall Prevention: Real-World Near-Fall Detection with Advanced Machine Learning Technique" Sensors 25, no. 5: 1468. https://doi.org/10.3390/s25051468

APA Style

Schneider, M., Seeser-Reich, K., Fiedler, A., & Frese, U. (2025). Enhancing Slip, Trip, and Fall Prevention: Real-World Near-Fall Detection with Advanced Machine Learning Technique. Sensors, 25(5), 1468. https://doi.org/10.3390/s25051468

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop