Next Article in Journal
Indirect Thermographic Measurement of the Temperature of a Transistor Die during Pulse Operation
Previous Article in Journal
Investigating the Internal Deterioration of the Auriga Statue of Mozia Island, Sicily, through Ultrasonic and Ground-Penetrating Radar Studies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Falling Detection of Toddlers Based on Improved YOLOv8 Models

by
Ziqian Yang
1,2,*,
Baiyu Tsui
1,2,
Jiachuan Ning
3 and
Zhihui Wu
1,2
1
College of Furnishings and Industrial Design, Nanjing Forestry University, Nanjing 210037, China
2
Jiangsu Co-Innovation Center of Efficient Processing and Utilization of Forest Resources, Nanjing 210037, China
3
Qingdao Grace Chain Software Ltd., Qingdao 266071, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(19), 6451; https://doi.org/10.3390/s24196451 (registering DOI)
Submission received: 22 August 2024 / Revised: 21 September 2024 / Accepted: 4 October 2024 / Published: 6 October 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
If toddlers are not promptly checked and rescued after falling from relatively high locations at homes, they are at risk of severe health complications. We present a toddler target extraction method and real-time falling alarm. The procedure is executed in two stages: In stage I, a GELAN-integrated YOLOv8 model is used to extract the body features. Based on this, a head capture technique is developed to obtain the head features. In stage II, the “safe zone” is calculated through Generalized Hough Transform (GHT). The spatial location is compared to the preceding stage’s two centers of mass points, K for the toddler’s body and H for the head. Position status detection is performed on the extracted data. We gathered 230 RGB-captured daily videos of toddlers aged 13 to 30 months playing and experiencing upside-down falls. We split 500 video clips (×30 FPS) from 200 videos into 8:2 training and validation sets. A test set of 100 clips (×30 FPS) was cut from another 30 videos. The experimental results suggested that the framework has higher precision and recall in detection, as well as improved mean average precision and F1 scores compared to YOLOv3, v5, v6, and v8. It meets the standard FPS requirement for surveillance cameras and has an accuracy of 96.33 percent.

1. Introduction

Falling is a natural aspect of a child’s development as they learn motor and cognitive skills. It is especially prevalent when learning to walk, climb, jump, and explore their surroundings [1]. As children grow older, they are more likely to engage in challenging activities, sometimes called “risk-taking behaviors.” The Injury Review of Chinese Children and Adolescents [2] reveals that the mortality rate from falls among children aged 1–4 is disturbingly high, with falls being the leading cause of death in this population. The danger of falling is increased by up to eight times due to immature skeletal systems and a lack of self-protection habits in this age range. The Report on the Epidemiological Review of Injuries to Children and Youth in China [3] states that 46.34% of all cases of injuries occur at home. Falls account for the majority of injuries at home (40.18%). The most common falls among children, mainly those ages 1–3, are from stairs, baby walkers, furniture, or play equipment.
This paper focus on vision-based techniques for fall detection because of their benefits over wearable sensors. Vision-based techniques are less intrusive and eliminate the motion disturbance of toddlers wearing specific test clothes. With the proliferation of cameras in modern cities, we can expand the applicability of fall detection models from smart homes to public spaces and beyond.
The visual equipment mainly consists of several types of cameras, such as depth, thermal, and monocular. Mecocci et al. [4] employed a depth camera to detect falls in hospital rooms, with a sensitivity of 82.1% and a specificity of 97.8%. Booranrom et al. [5] utilized a Kinect camera to analyze ten volunteers’ falls and achieved 97% accuracy. Song et al. [6] suggested a thermal camera-based monitoring system, tested it on five volunteers, and evaluated the thermal images using SVN with a 99.7% accuracy. Kittipanya-Ngam et al. [7] monitored falls by recognizing the bed’s position relative to the person’s head. However, the method only recognizes the edges of the bed. It identifies the head features by skin color, which may cause serious errors. Banerjee et al. [8] estimated the location of the bed-ridden person by determining the person’s area, which reduces misjudgment of fall detection. Still, the study could not provide early warning of the falling behavior. Ni et al. [9] similarly proposed a monitoring method that utilizes visual information from RGB-D for fall warnings. Brulin et al. [10] perform fall detection using a gesture recognition algorithm based on fuzzy logic analysis. Still, due to the specificity of the algorithm and the camera, it cannot work during the night, which results in it being unable to address real-life applications. Two-dimensional cameras are a more cost-effective alternative to three-dimensional range sensors.
Takeda et al. [11] used NeNet to determine the relationship between humans and beds. Still, this method required complex parameter settings during picture segmentation processing, resulting in low computing efficiency and preventing real-time monitoring. Shen et al. [12] recognized the human head using template matching before extracting the entire body using region growth. However, this technology necessitates a more static stance for the human body and cannot accurately detect changing movements. Zhao et al. [13] detected the human head first and used its position to collect information about the upper half of the human body. However, if the subject’s body is obscured, information about the upper body cannot be accurately extracted, and the system’s operation time is considerable, making real-time monitoring impossible. Other studies [14,15,16,17] primarily identify the human body following a fall. They detect the person lying on the floor and do not consider dynamic information, making it difficult to discern between typical lying and accidental falls.
Researchers modified the feature extraction module’s design to enhance detection accuracy. Zhao et al. [18] adds the SDI (Shuffle Dimensionally Integrated) attention module to the Efficient Layer Aggregation Networks(ELAN) in the Backbone and Spatial Pyramid Pooling, Cross Stage Partial Channel (SPPCSPC) in the head, enabling the YOLOv7 model to focus on key areas or typical features. Lyu et al. [19] improved the SPP network by incorporating an extra 1 × 1 max-pooling layer, resulting in notable improvements in detection performance. Chen et al. [20] improved the YOLOv5 model’s extraction ability by incorporating spatial attention and average pooling layers into the Efficient Channel Attention(ECA) and SPP networks, respectively. Pereira GA [21] analyzed YOLOv8 variations and found that YOLOv8m achieved a good mix of computational efficiency and detection performance. Sun et al. [22] used Single Shot Detector (SSD) MobileNet to remove the non-human key points detected by OpenPose, and the SVDD classification algorithm classifies the values. In summary, these studies have shown high accuracy, sensitivity, and specificity, but the datasets used are laboratory environments and mostly simulated activities of daily living (ADLs) and falls, which have significant limitations.
Creating datasets from publicly available data or real-world surveillance footage makes a lot of sense for training detection models. The approach should adapt to a wide range of surveillance images, including those with differing small distractions and occlusions in the video and varying ambient settings and lighting conditions. He et al. [23] use an attention-integrated YOLOv5-alike model to extract infant portraits. Then, a single-shot multi-box detector filters out spurious targets and accurately recognizes the position. The dataset was derived from actual surveillance captured in baby pools. Furthermore, additional actual surveillance was used as the test set. Marcos et al. [24] present a transformer-based fall detection method. They use transformers to extract characteristics from raw RGB frames, eliminating the need for further skeleton and position computations. They evaluate the fall detection model on the UP-Fall dataset [25] and on the UR fall dataset [26]. Qin et al. [27] replaced YOLOv8′s C2f module in the backbone network with the C2Dv3 module to capture more intricate features. The DyHead block was utilized in the Neck section to unify multiple attentions, increasing the detection accuracy of targets at different scales. Maudsley-Barton [28] employs an LSTM autoencoder to calculate a distance measure on the KINECAL dataset. Using this measure, he develops a new scoring system that places individuals with varying fall risks on a continuous scale.
In everyday scenarios, toddlers are prone to sudden and swift falls from furniture. Surveillance cameras integrated with the YOLO recognition algorithm offer an effective solution to these challenges. They are capable of capturing data on multiple individuals concurrently, eliminating the need for additional sensors. These cameras boast swift processing capabilities, facilitating real-time data analysis. Furthermore, they are economically viable and enhance the overall user experience.
Our study focused on toddlers, characterized by their disproportionately large heads and tendency to crawl and roll on furniture like beds and sofas. To accurately determine falls, we consider the combined center of mass for both the body and head in relation to the furniture edges.
We propose a toddler target information extraction approach based on improved YOLOv8.This method can accurately recognize the target subject, even in cases where the person is overlapping with furniture. Simultaneously, given that the positions of furniture and the ground are typically unchanging, We use GHT to identify forms and extract the shape of the “safe zone” surface using a predefined template. We then calculate the positional relationship between the human body target, head target, and the “safe zone” to generate a real-time alarm for falling behavior, as depicted in Figure 1.
Due to the specificity of the toddler population, we first built our dataset by collecting 230 daily surveillance videos of toddler falls captured by RGB cameras. These videos primarily recorded the daily activities of 13- to 30-month-old toddlers playing near the edge of a piece of furniture or a guardrail before falling. The length of these videos varied from 10 min to 15 s. One hundred and eight parents agreed to use video surveillance videos for research purposes.
The main contributions are as follows:
  • This study use GHT to identify forms and extract the shape of the “safe zone” surface using a predefined template.
  • A new dataset of 500 video clips (×30 FPS) is created using 200 real-time daily videos collected from 100 parents. Another test set of 100 falling video clips (×30 FPS) is created using 30 real-time daily videos collected from 8 parents. The datasets collected from daily surveillance video feature varying lighting conditions, camera settings, furniture placement, and furnishing arrangement, increasing the diversity of the generated dataset and enhancing the framework’s generalization capacity.
  • Given that the standard YOLOv8 method is insufficient for collecting body and head information at higher precision, we have improved the original YOLOv8 program. We recommend replacing the C2f module in YOLO with the GELAN module. One advantage of GELAN is that it uses the convolutional network layer and can contain any computational module. It increases the overall flexibility of the network topology and allows it to fulfill the needs of our information extraction techniques successfully.
  • A real-time system is proposed to detect instances of toddlers falling in their home surroundings. This technology efficiently and precisely captures data of toddler targets in challenging environments, converting the human body into a rectangular shape, simplifying the data without sacrificing accuracy, and increasing the usefulness of real-time artificial intelligence for fall detection. It is an early warning preparation for a later head injury assessment. After thorough enhancement and streamlining, it may be deployed on the hardware infrastructure of smart home devices.
The remaining sections of this work are structured as follows. In Section 2, the detection approach’s details of toddlers falling are designed. Section 3 presents the conducted experiments and discusses their outcomes. The framework’s benefits and constraints are examined in Section 4. Conclusions are derived in Section 5.

2. The Proposed Methods

Toddlers are frequently put in flat areas of the home with guardrails and on beds and sofas. The primary challenge in monitoring toddlers’ falls from furniture is obtaining information about their bodies efficiently. Our studies have determined that OpenPose is unsuitable with our framework, even though we initially considered it for a comparative study together with YOLO. The main reasons are as follows:
  • Posture changes: Toddlers show a variety of postural modifications. Their limbs are often more flexible than adults, who often maintain an upright posture and move regularly. This variation can make it difficult to identify posture via skeletal analysis.
  • Differences in physical characteristics: Toddlers have shorter limbs, larger heads, and notably distinct proportions compared to adults. OpenPose may not capture these traits well, leading to imprecise detection outcomes.
  • Differences in the coherence of movement: Toddlers have less predictable and coherent motions and tend to move quicker and erratically, especially when lying down or crawling, making it challenging for the OpenPose model to track them effectively.
  • Problems with loose clothing covering: Toddlers typically wear clothing that fits loosely, which can conceal joints and make it more challenging to identify them.
  • The problem of real-time availability: OpenPose primarily captures human skeletal point data, and excess recognition data may cause interference. In real applications, the approach requires significant computational resources and processes at a slower speed.
Consequently, YOLO was ultimately selected to extract the body information of toddlers.
We carefully examined the toddler’s fall video without interruption to ensure that our analysis was accurate. YOLO-based body information extraction models received keyframes. This model accurately identified the child’s head and torso from the input image.
Following that, manual and automatic parsing techniques were used to thoroughly evaluate the furniture and floor, determining the optimal ‘safe zone’. Given the architecture of our technique, this procedure may be carried out in real-time on a conventional computer.
Finally, by examining the relative placements of these two zones and using a state delineation method, we may precisely define the child’s current state.
Our classification approach identifies whether a toddler is safe or at risk of falling. Using the previously mentioned basic methodology, we created an accurate target identification algorithm and a unique classifier for detecting potential hazards. Figure 2 depicts the workflow for our framework.

2.1. Extracting Body Information of Toddler

To effectively solve the issue, we propose a technique for extracting body information based on the YOLO algorithm. Subsequently, a Generalized Efficient Layer Aggregation Network (GELAN) is developed, utilizing an ELAN, to substitute the C2f module in the backbone network. This replacement is intended to improve the network’s ability to extract features and give a more lightweight model.
GELAN combines the Cross Stage Partial Network (CSPNet) and ELAN procedures, integrating RepConv to achieve more convincing results. The reasoning speed is unaffected when the network environment is minimally influenced due to the lightweight processing and tiny amount of data.
The human head is more recognized and has consistent visual traits than other body parts. As a result, our approach prioritizes detecting body information before retrieving a toddler’s head details. First, the child’s physical parts are properly identified. The head data are then retrieved and verified using positioning body information and a GELAN module to focus on the child’s head, an improved alternative to the C2f module in YOLO. Potential body areas are identified based on the spatiotemporal consistency seen between consecutive frames to maximize efficiency. Figure 3 displays the major technical route for retrieving information from child bodies.
The GELAN module, integrated into our framework, merges two neural networks: CSPNet and ELAN. Both are designed via gradient pathways, allowing us to combine their strengths into a GELAN. This network excels at lightweight construction, fast inference, and high accuracy. This makes it ideal for devices with little resources. Our framework extends the functionality of the ELAN architecture. The internal framework structure is shown in Figure 4.
GELAN integrates arbitrary computing blocks and a convolutional network layer. This flexibility improves the network structure, making it suitable for information extraction. Figure 4 shows how modularity and partitioning increase the network’s adaptability and customizability. GELAN enables several computing blocks, solving current architecture restrictions and providing a scalable deep-learning solution.
In this study, leveraging GELAN’s flexibility, we can efficiently and swiftly obtain the toddler’s body information, laying a solid foundation for subsequent head information extraction.

2.2. Focusing on Toddler’s Head Information

We extract specific features from the toddler’s head after obtaining human body information using the improved YOLOv8 model, as shown in Figure 5. Head features are essential for making the final assessment.
Based on toddler body information extracted in the previous phase, we created a specialized model for recognizing heads. The model is employed for the following reasons:
  • The method is robust to image noise or pixel module processing.
  • The method boasts robust feature selection and verification capabilities, effectively preventing the extraction of non-target features. Additionally, it significantly reduces the computation time.
  • The approach is trained using the minimal entropy concept and may be readily parallelized. However, false alarms may still occur. As a result, the precise head location requires further localization.
This head-recognized model separates human body information into the head and the central torso section [29]. The fundamental concept of this algorithm involves partitioning the human body into two interconnected cubes. Both cubes have the same axis orientation along the line segment, linking the center of the head and the center of the body. This categorization reduces the intricacy of the structure of the human body but also enables the algorithm to analyze visual input that contains sequential patterns more effectively.
Given that we deal with sequential visual data, the algorithm uses spatiotemporal contextual information to reduce the search area for body detection (particularly the head). This implies that when the algorithm examines a new video frame, it anticipates where the human body will most likely be in the current frame based on information from the previous frame, limiting the search region and increasing detection efficiency.
If the initial frame is being analyzed or no toddler body is detected within the first two seconds, the search area will include the entire frame. This is due to the algorithm’s insufficient historical data to establish a search reference in such circumstances. Once the algorithm accurately identifies the location of the torso and head of the human body in the frame, it will provide the current center point position information as point i . This information is crucial for tracking and recognizing the body in subsequent tasks.
We recognize the algorithm’s potential for misdetection or omission by categorizing the picture characteristics using a Random Forest (RF) classifier [30]. This classification allows us to determine whether a detected feature corresponds to the head.
Furthermore, the algorithm can deal with situations where the human body cannot be precisely identified. During this procedure, the algorithm gradually increases the region of head detection. The algorithm’s capacity to dynamically update the search zone allows it to adapt to changing conditions and fluctuations in human motion successfully.
Algorithm 1 is for body and head information extraction.
Algorithm 1. Body Information Extraction
Inputs: Images to be processed
Output: Center of mass point K of the child’s body, center of mass point H of the head
 1. Segment the information of the toddler’s surveillance frame;
 2. Capture the toddler’s body information and extract the centroid K;
 3. Decompose the body information into two interconnected cubes: the head and the main torso part;
 4. Based on the information in the in-1 frame or the in-frame, predict the possible locations of the human body in the current frame and narrow down the search range;
  IF:
  Human body has been detected, output center of mass point H of the head;
  Otherwise:
  Expanding the search area to the entire frame to look for body information that output center of mass point H of the head.

2.3. “Safe Zone” Demarcation

It is critical to recognize the current state of young children to provide appropriate care. However, status assessments require more than retrieving information about a child’s body and head. This is because a child’s status is influenced not just by their physical and mental health, but also by their surroundings. As a result, we suggest a solution to this problem by establishing “safe zones”.
In this study, we initially restricted the “safe zone” to the bed and sofa at home. Most toddlers in these two regions climb over guardrails, tumbling in the dataset we collected. We can improve child safety by properly labelling mattresses, couches, and guardrails.
In future studies, we should examine dining chairs, tables, and strollers to ensure toddler safety. Labeling and tracking these household objects will assist us in understanding a child’s health.
We partition the “safe region” using Hough Transform to ensure precision [31]. This method is resilient to interference and insensitive to residual straight lines, noise, and other non-linear features in the image. It is tolerant of feature border description gaps and less impacted by image noise. It may be modified to meet specific needs and applies to beds, sofas, and other home furnishings. The Hough Transform method, known for its universality and simplicity of implementation, has been widely utilized in object identification, such as hand gestures [32] and airplanes [33].
The Hough Transform converts a curve or straight line of the same shape in one coordinate system to a point in another, creating a statistical peak problem rather than a problem of identifying arbitrary forms. The bedroom environment photos contain blankets and pillows, making it hard to locate the bed surface. We apply a Generalized Hough Transform (GHT) [34]. It can identify forms and extract the shape of the bed surface using a predefined template. Figure 6 displays the stage of GHT-based bed surface identification. It filters by the detected contour’s area to find contours representing bed surface components while avoiding contours depicting other items. The specific Algorithm 2 is displayed below.
Algorithm 2. GHT based bed surface detection
Inputs: Colorful image to be detected
Output: Detected bed boundary line
  1. Convert the color image to be detected into gray scale;
  2. Remove the high frequency signal and smooth the image by Gaussian kernel denoising;
  3. Use gradient operator, Laplace operator, canny, and sobel for edge extraction;
  4. Perform the edge point judgment by the principle of binarization, i.e., gray scale value = 0/255;
  5. Prepare two containers, one for displaying the Hough-space profile, an array Hough-space used to store the values of voting;
  6. Take the local maxima, set the threshold, filter the interfering straight line, draw the straight line, and calibrate the corner points.
  Output the bed boundary

2.4. State Label Delineation

Based on the previously extracted body feature information, we obtain the center of mass point K of the toddler’s body and the center of mass point H of the head, as shown in Figure 7.
We added the safety icon S p . We identify the toddler as S 1 while they are safe, when they may fall, and S 3 when they have already fallen. Security symbol division is shown clearly in Algorithm 3.
Algorithm 3. Security Symbol Division
Inputs: center of mass point K of the child’s body, center of mass point H of the head, safety zone S p
Output: safety markers S1, S2, S3
If:
 H is within S p , then it is currently in S1;
 Point H leaves the region, and K is within S p , the current output is S2;
Otherwise:
 The output is S3.
According to the algorithm above, we established a particular determination technique for the division of the child’s state and the safety mark, representing the three states in which the kid is. The precise delineation of the circumstances and the basis are shown in Table 1.

3. Experiments and Results

We performed experiments on our designated datasets to fully illustrate the efficacy of our suggested detector and approach. The experimental results validate the efficacy of our approach.

3.1. Dataset

To assess the validity of the proposed methodology, we used fall events as test cases. We collected 230 videos of toddlers falling in the home environment from parents. These videos focused on the daily activities of toddlers between 13 and 30 months of age, particularly when they fell from their beds or sofas. We included 108 toddlers in our sample pool and covered diverse lighting conditions, camera settings, and indoor furniture layouts. These measures were intended to enhance the breadth and representativeness of the dataset, and some of the representative data are shown in Figure 8.
Each video lasted 10 s to 10 min and vividly captured the toddler’s daily activities, including falls on the bed or sofa. These show toddlers’ daily activities, such as falling forward or backwards on the bed or being held by parents after falling without hitting the ground. Based on this rapid fall movement of the head, we carefully selected two to three clips for each video.
Subsequently, we take 500 video clips (×30 FPS) extracted from the 200 videos into a training set and a validation set in the ratio of 8:2. In addition, a separate test set of 100 video clips (×30 FPS) cut from another 30 videos concentrating on children’s falling behaviors at home without direct adult supervision was created to completely evaluate our framework’s effectiveness.
The experimental model operates on Windows 11 Professional Edition (Microsoft Corporation, Redmond, WA, USA). It was configured with a CPU: Intel i5 13600KF (Intel Corporation, Santa Clara, CA, USA) and GPUs: Nvidia GTX 4070 Ti 12 GB (Nvidia Corporation, Santa Clara, CA, USA), employing Python 3.8, PyTorch 2.0.1, and cuDNN 8.9.1.

3.2. Experimental Setting

Considering the dataset size and training efficiency, we set the batch size to 16 and planned 300 Epochs for training. All photos within the dataset were normalized to 640 × 640 pixels and transmitted over the network to train the extraction model. The initial learning rate was set to 0.01, while the final learning rate was adjusted to 0.10 to optimize the training process. For the toddler information model, we specifically used images scaled to 300 × 300 pixels from the dataset for training. In addition, a weight decay coefficient of 0.0005 was carefully set, and the learning rate was defined to vary from 0.0002 to 0.02, aiming to refine the model’s training effect further. To comprehensively and accurately evaluate the performance of our model on the task of classifying the state of a toddler, we set up four clear and mutually exclusive result classifications, which cover all possible scenarios as follows:
  • The first is that the system correctly identifies the toddler’s current state.
  • The second is that an event that did not result in a fall was wrongly classified as an alarm state.
  • The third one is that no fall occurs; hence, the algorithm does not segment it.
  • The fourth is a fall event that the system does not recognize.
Based on the four different scenarios mentioned above, we segmented them into four categories: TP, FP, TN, and FN, using a specialized measurement method.
  • True Positive (TP): A fall occurred, and the system accurately classified the fall.
  • False Positive (FP): The fall did not occur, but the system misclassified it.
  • True negative (TN): No fall occurred, and the system accurately identified it.
  • False negative (FN): A fall occurred, but the system incorrectly labeled it.
The evaluation of classification models is fundamental to machine learning. Model selection and performance evaluation frequently use numerous indicators for a holistic approach. These numbers show the model’s classification accuracy and performance in different settings. In practice, we must choose evaluation metrics based on task needs and background information.

3.3. GELAN Improved Results

To verify the enhanced performance of YOLOv8 on our custom dataset, we conducted a comparative analysis with the original YOLOv8 model, YOLOv3, YOLOv5, and YOLOv6. The results in Table 2 show that YOLOv3 and YOLOv5 have poor detection performance in the toddler head and body information extraction scenario, barely reaching about 85%. In terms of metrics, the improved YOLOv8 network uses GELAN and has a mAP@50 score of 97.26%; the average accuracy is 11.65, 17.32, 7.26, and 3.15 percentage points higher than that of YOLOv3, YOLOv5, YOLOv6, and YOLOv8, respectively. All data in the table are mean values. Regarding accuracy and recall, the improved YOLOv8 model achieves 96.33% and 97.87%, respectively, higher than the original YOLOv8 model. In addition, the FPS of the enhanced algorithm is 69.97, and the improved model is more suitable for small target detection.
The improved YOLOv8 model proposed in this paper achieves significant performance gains in head and body information extraction from toddlers by replacing the original c2f module with the GELAN module. This performance improvement is mainly due to the following reasons:
  • More efficient feature extraction: The GELAN module combines the advantages of multiple network structures to enhance the model’s ability to recognize toddlers’ body features through effective hierarchical aggregation and the combination of CSPNet and ELAN. This combination not only enhances the richness of the features, but also speeds up the feature extraction by keeping its time difference within 3 s.
  • Adaptable: The improved YOLOv8 model is optimized for complex background and small target detection. The flexibility of the GELAN module allows the network to better adapt to the characteristics of young children’s large changes in body size and dynamics, which improves the robustness of the detection and increases the accuracy by nearly 20% compared to the original YOLO model.
  • Optimized network structure: GELAN reduces the computational burden of the model through more efficient data processing and simplification of the network structure, while maintaining high accuracy with an output frame rate of 69.97 FPS, which makes the model more suitable for running on resource-limited devices and improves the model’s practicality and scalability.

3.4. Human Information Extraction Results

To evaluate the effectiveness of each daily activity sample in the self-built dataset, we randomly selected 15 discontinuous frames from each video clip. There are a total of 7500 frames (15 × 500) for performance evaluation. We identified and labeled the toddler body information for each frame in boxes. The selected frames will be divided into an 8:2 ratio for training and testing. Among them, the training set randomly selected proportionally is used for Random Forest (RF) training and the rest are tested. The following criteria are used to evaluate the accuracy of human body and head information extraction:
T i = r i < λ , 0 r i λ , 1
where T i denotes whether the toddler body and head information has been successfully extracted in the frame; r i denotes the intersection of the extracted human body and head information rectangles, and the actual information rectangles; and λ is a predefined threshold. Our rule states that the higher λ is, the more stringent the criteria for effectively extracting the toddler`s upper body. If r i is less than λ , it indicates that the toddler body information extracted in the frame is inaccurate and outputs it as 0. On the contrary, it suggests that it is accurate and outputs it as 1. Given that it is impossible to achieve a 100% accurate rate in the actual application of the algorithm, we can disregard the effect of the error on the results when the intersection between the measured and actual information reaches a specific value. We first converted the image to be measured into a histogram, essentially the probability density distribution of points with varying pixel values. Each bar in the histogram is divided by the total number of points to obtain the a priori probability of its occurrence, which equals one. Then, the image is traversed, separating the pixel points into two portions, A and B, and calculating the sum of the probability of occurrence of A and B pixels, the mean UA, UB, and their respective variances, sigmaA and sigmaB. All acquired parameters are combined to form an objective function, and the minimum objective function is used to determine the suitable threshold. The calculation above yielded the algorithm’s threshold value λ to be 0.8.
A = i = 1 n r i n
In the above Equation (2), A is the accuracy of the whole extraction result and n is the total test sample size. The final extracted result is 95.3%, and we can conclude that the method has a strong advantage for extracting human body and head information. At the same time, the method has a strong learning ability. If the human body information in a certain frame is missing, it can still extract the human body information in the next frame efficiently through learning. Figure 9 shows the results of the method in the inspection of the image, as shown in the figure in the 70th frame, the detection of the omission. However, the subsequent detection is not affected by the still accurate extraction.
To highlight the superiority of our toddler upper body identification approach, we compare it with the popular methods. Zhao et al. [35] used effective human upper body detection method to extract the center of the human head and upper body by using a RF approach. Shen et al. [12] extracted LTDP visual descriptors for human body detection and combined with SVM classifier. The final performance comparison results are listed in Table 3. We can clearly observe that the human head and body information extraction method by improving YOLO is greatly superior to the methods of Zhao and Shen. We can observe that these two comparison methods cannot handle the human–bed position interactions and substantial postural fluctuations that occur during the fall-from-bed phase, resulting in unsatisfactory human body extraction results.

3.5. State Classification Results

We tested the model’s reliability using 300 additional video clips of toddlers falling at home. We renamed the safe, warning, and hazardous as states 1, 2, and 3, respectively, resulting in two out of sixty safe state samples being misdiagnosed as warning states and one out of sixty hazardous state samples. In contrast, three of the ninety-six warning state samples and five of the one hundred forty-four danger state samples were misdiagnosed as safe. Table 4 displays the number of young children identified in various states.
The accuracy and recall curves of the alarm algorithm studied in this paper are shown in Figure 10.
We can see that both values remain over 0.84 during the training process, and practically all of them are close to one. Based on YOLO enhancement, it demonstrates that our suggested fall alarm algorithm performs well and can play an important role in state recognition and alert, with the specific formula presented in Equation (3).
Re c a l l = T P T P + F N
This study evaluates the algorithm’s overall performance, which is displayed in Table 5. The precision of state classification for toddlers is 98.72%, with an accuracy rate of 96.33%. This implies that the method presented in this study may efficiently identify the current condition of the toddler body with a low leakage detection rate. Meanwhile, its specificity is 95%, and its f-score is 97.68%, indicating that the approach can distinguish between falling and daily activity behavior with fewer false positives.

4. Discussion

Based on its structure and performance, the detection software proposed in this work has the potential to be extended to public settings, as proved by the task of detecting toddlers falling. The detection technique is separated into two steps, aiming to accurately and conveniently detect toddler falling in real-time. In stage I, the features of the toddler’s body and head are separately extracted. In stage II, position status detection is conducted on the retrieved data using the “safe zone”. The experiment results suggest that the framework can reduce false and missed detections in complex household settings compared to other initial YOLO models.
The suggested detection system can be expanded functionally and architecturally to encompass outdoor situations. For instance, at community playgrounds, surveillance cameras near slides and climbing frames can effectively monitor cases of falls and promptly send real-time alert warnings. It can also be extended to help with recreational facilities for the elderly [36].
While our technique yields good detection outcomes, it also exhibits limitations.
  • The computer’s performance influences the computing time of the recognition algorithm in this article. In practice, computing platforms may behave differently from our specific setting in research. The algorithm’s GPU usage, memory requirements, and energy consumption will be assessed in our future work to discover resource bottlenecks. Based on the evaluation results, we will incorporate lightweight optimization strategies like model trimming and quantization to make the algorithm perform efficiently on traditional CPUs and edge devices utilized in real-time applications. These changes will improve our research’s practicality and sustainability.
  • We did not examine the safety zone of other pieces of furniture, such as dining chairs, dining tables, strollers, and other surfaces that toddlers frequently interact with and are at a high risk of falling from. Our future research should focus on expanding this further.
  • Based on the error results of our detection, we should develop corresponding correction mechanisms. Firstly, certain toys and patterns may still be mistaken for toddlers, leading to error detection. These confused objects are so close to toddlers in texture, color, and form that even video viewers find it difficult to distinguish between them. To effectively filter out small noises, the algorithm requires more optimization. Secondly, our algorithm is sensitive to noise caused by unusual photos and changing scene conditions. When confronted with a situation in which a toddler is in bed with a quilt covering the body, the system fails to capture the physical features. Thirdly, The RGB camera-based method remains unrecognizable due to excessively dim illumination. We should enhance the technique’s capacity to identify tiny, illegible targets, adjust model parameters, and enhance the complexity of model training.

5. Conclusions

This study suggests a toddler fall alarm approach based on the improved YOLOv8 model. The detection technique is divided into two steps. The first stage begins with extracting video frames, and the toddler body feature information is gathered by processing the surveillance video captured by improved YOLOv8. The toddler’s body feature data from the previous phase is then processed and incorporated using a head capture technique. In the second stage, the “safe zone” is computed using GHT, and the spatial location is compared to the two centers of mass points, K for the toddler’s body and H for the head from the previous stage. Position status detection is performed on the extracted data. The experimental findings demonstrate that the proposed framework outperforms YOLOv3, v5, v6, and v8 regarding detection precision, recall, mean average precision, and F1 scores. Additionally, it meets the typical FPS requirement for general surveillance cameras. It has an accuracy of 96.33 percent, a sensitivity of 96.67 percent, and a specificity of 95 percent.

Author Contributions

Conceptualization, methodology, writing—review and editing, supervision, Z.Y.; data curation, formal analysis, investigation, writing—original draft preparation, B.T.; consulting, J.N.; supervision, Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by 2022 Jiangsu Province Ecological Health Home Furnishing Industry-University-Research International Cooperation Joint Support for laboratory projects in the form of a grant to CZ, WG, and XL [20220602], The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

Guardians of the toddlers in the collected dataset have requested that the dataset not be made publicly available.

Conflicts of Interest

J.N. is employed by the company Zhejiang New Rise Digital Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SDIShuffle Dimensionally Integrated
ELANLong Short-Term Memory Efficient Layer Aggregation Networks
SPPCSPCSpatial Pyramid Pooling, Cross Stage Partial
STGCNSpatio-Temporal Graph Convolutional Network
ECAEfficient Channel Attention
SSDSingle Shot Detector
ADLsActivities of Daily Living
GELANGeneralized Efficient Layer Aggregation Network
CSPNetCross Stage Partial Network
GHTGeneralized Hough Transform
RFRandom Forest
TPTrue Positive
FPFalse Positive
TNTrue Negative
FNFalse Negative

References

  1. Bartlett, S.N. The problem of children’s injuries in low-income countries: A review. Health Policy Plan. 2002, 17, 1–13. [Google Scholar] [CrossRef]
  2. Li, L.; Mao, X.; Shi, Q.; Ma, A.; Wang, L. Meta-analysis of the incidence of fall injuries among Chinese children and adolescents, 2002–2012. China School Health 2014, 10, 1534. [Google Scholar]
  3. The National Center for Chronic and Noncommunicable Disease Control and Prevention China (NCNCD). Review of Chinese Children and Adolescents; NCNCD: Beijing, China, 2018. [Google Scholar]
  4. Mecocci, A.; Micheli, F.; Zoppetti, C.; Baghini, A. Automatic Falls Detection in Hospital-Room Context. In Proceedings of the 7th IEEE International Conference on Cognitive Infocommunications (CogInfoCom), Wroclaw, Poland, 16–18 October 2016; pp. 127–132. [Google Scholar]
  5. Mongkolnam, P.; Booranrom, Y.; Watanapa, B.; Visutarrom, T.; Chan, J.H.; Nukoolkit, C. Smart bedroom for the elderly with gesture and posture analyses using Kinect. Maejo Int. J. Sci. Technol. 2017, 11, 1–16. [Google Scholar]
  6. Song, K.S.; Nho, Y.H.; Kwon, D.S. Histogram based Fall Prediction of Patients using a Thermal Imagery Camera. In Proceedings of the 14th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Jeju, Republic of Korea, 28 June–1 July 2017; pp. 161–164. [Google Scholar]
  7. Kittipanya-Ngam, P.; Guat, O.S.; Lung, E.H. Computer vision applications for patients monitoring system. In Proceedings of the International Conference on Information Fusion, Singapore, 9–12 July 2012. [Google Scholar]
  8. Banerjee, T.; Enayati, M.; Keller, J.M.; Skubic, M.; Popescu, M.; Rantz, M. Monitoring Patients in Hospital Beds Using Unobtrusive Depth Sensors. In Proceedings of the 36th Annual International Conference of the IEEE-Engineering-in-Medicine-and-Biology-Society (EMBC), Chicago, IL, USA, 26–30 August 2014; pp. 5904–5907. [Google Scholar]
  9. Ni, B.B.; Dat, N.C.; Moulin, P. RGB-D camera based get-up event detection for hospital fall prevention. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Kyoto, Japan, 25–30 March 2012; pp. 1405–1408. [Google Scholar]
  10. Brulin, D.; Benezeth, Y.; Courtial, E. Posture Recognition Based on Fuzzy Logic for Home Monitoring of the Elderly. IEEE Trans. Inf. Technol. Biomed. 2012, 16, 974–982. [Google Scholar] [CrossRef]
  11. Takeda, F. Proposal of an Awakening Behavior Detection System for Medical Use and Adaptation for Fluctuation of the Brightness Quantity with Infrared Camera Device Kinect. In Proceedings of the 9th International Conference on Signal-Image Technology and Internet-Based Systems (SITIS), Kyoto, Japan, 2–5 December 2013; pp. 714–719. [Google Scholar]
  12. Shen, Y.; Hao, Z.; Wang, P.; Ma, S.; Liu, W. A Novel Human Detection Approach Based on Depth Map via Kinect. In Proceedings of the 26th IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Portland, OR, USA, 23–28 June 2013; pp. 535–541. [Google Scholar]
  13. Zhao, F.; Cao, Z.G.; Xiao, Y.; Mao, J.; Yuan, J.S. Real-Time Detection of Fall from Bed Using a Single Depth Camera. IEEE Trans. Autom. Sci. Eng. 2019, 16, 1018–1032. [Google Scholar] [CrossRef]
  14. Volkhardt, M.; Schneemann, F.; Gross, H.M. Fallen Person Detection for Mobile Robots using 3D Depth Data. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics (SMC), Manchester, UK, 13–16 October 2013; pp. 3573–3578. [Google Scholar]
  15. Antonello, M.; Carraro, M.; Pierobon, M.; Menegatti, E. Fast and Robust Detection of Fallen People from a Mobile Robot. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 4159–4166. [Google Scholar]
  16. Nishi, K.; Miura, J. A head position estimation method for a variety of recumbent positions for a care robot. Mechatronics 2015, 2015, 61–64. [Google Scholar] [CrossRef]
  17. Wang, S.; Zabir, S.; Leibe, B. Lying Pose Recognition for Elderly Fall Detection. In Proceedings of the Robotics: Science and Systems VII, Los Angeles, CA, USA, 27–30 June 2011; Durrant-Whyte, H., Roy, N., Abbeel, P., Eds.; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
  18. Zhao, D.; Song, T.; Gao, J.; Li, D.; Niu, Y. YOLO-Fall: A Novel Convolutional Neural Network Model for Fall Detection in Open Spaces. In Proceedings of the IEEE Access, Manchester, UK, 6 February 2024; pp. 26137–26149. [Google Scholar]
  19. Lyu, L.; Liu, Y.; Xu, X.; Yan, P.; Zhang, J. EFP-YOLO: A quantitative detection algorithm for marine benthic organisms. Ocean Coast. Manag. 2023, 243, 106770. [Google Scholar] [CrossRef]
  20. He, C. Image Compressive Sensing via Multi-scale Feature Extraction and Attention Mechanism. In Proceedings of the 2020 International Conference on Intelligent Computing, Automation and Systems (ICICAS), Chongqing, China, 11–13 December 2020; pp. 266–270. [Google Scholar]
  21. Pereira, G.A. Fall Detection for Industrial Setups Using YOLOv8 Variants. arXiv 2024, arXiv:2408.04605. [Google Scholar]
  22. Sun, G.; Wang, Z. Fall detection algorithm for the elderly based on human posture estimation. In Proceedings of the 2020 Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC), Dalian, China, 14–16 April 2020; pp. 172–176. [Google Scholar]
  23. He, Q.; Zhang, H.; Mei, Z.; Xu, X. High accuracy intelligent real-time framework for detecting infant drowning based on deep learning. Expert Syst. Appl. 2023, 228, 120204. [Google Scholar] [CrossRef]
  24. Núñez-Marcos, A.; Arganda-Carreras, I. Transformer-based fall detection in videos. Eng. Appl. Artif. Intell. 2024, 132, 107937. [Google Scholar] [CrossRef]
  25. Martínez-Villaseor, L.; Ponce, H.; Brieva, J.; Moya-Albor, E.; Peafort-Asturiano, C.J.S. UP-Fall Detection Dataset: A Multimodal Approach. Sensors 2019, 19, 1988. [Google Scholar] [CrossRef] [PubMed]
  26. Kwolek, B.; Kepski, M.J.C.M.; Biomedicine, P.i. Human fall detection on embedded platform using depth maps and wireless accelerometer. Comp. Methods Programs Biomed. 2014, 117, 489–501. [Google Scholar] [CrossRef] [PubMed]
  27. Qin, Y.; Miao, W.; Qian, C. A High-Precision Fall Detection Model Based on Dynamic Convolution in Complex Scenes. Electronics 2024, 13, 1141. [Google Scholar] [CrossRef]
  28. Maudsley-Barton, S.; Yap, M.H. Objective falls risk assessment using markerless motion capture and representational machine learning. Sensors 2024, 24, 4593. [Google Scholar] [CrossRef]
  29. Wagh, K.V.; Kulkarni, R.K. Human Tracking System. In Proceedings of the International Conference on Electronics and Communication Systems (ICECS), Coimbatore, India, 13–14 February 2014. [Google Scholar]
  30. Rhodes, J.S.; Cutler, A.; Moon, K.R. Geometry and Accuracy Preserving Random Forest Proximities. IEEE Trans. Pattern Anal. Mach. 2023, 45, 10947–10959. [Google Scholar] [CrossRef]
  31. Ballard, D.H. Generalizing the Hough transform to detect arbitrary shapes. Pattern Recognit. 1981, 13, 111–122. [Google Scholar] [CrossRef]
  32. Liao, B.; Li, J.; Ju, Z.; Ouyang, G. Hand gesture recognition with generalized hough transform and DC-CNN using realsense. In Proceedings of the 2018 Eighth International Conference on Information Science and Technology (ICIST), Cordoba, Spain, 30 June–6 July 2018; pp. 84–90. [Google Scholar]
  33. Chen, H.; Gao, T.; Qian, G.; Chen, W.; Zhang, Y. Tensored generalized hough transform for object detection in remote sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3503–3520. [Google Scholar] [CrossRef]
  34. Zhang, C.; Wang, F.; Zou, Y.; Dimyadi, J.; Guo, B.H.; Hou, L. Automated UAV image-to-BIM registration for building façade inspection using improved generalised Hough transform. Autom. Constr. 2023, 153, 104957. [Google Scholar] [CrossRef]
  35. Xia, L.; Chen, C.C.; Aggarwal, J.K. Human detection using depth information by Kinect. In Proceedings of the IEEE Conference on Computer Vision & Pattern Recognition Workshops, Washington, DC, USA, 20–25 June 2011. [Google Scholar]
  36. Zhou, C.; Xu, X.; Huang, T.; Kaner, J. Effect of different postures and loads on joint motion and muscle activity in older adults during overhead retrieval. Front. Physiol. 2024, 14, 1303577. [Google Scholar] [CrossRef] [PubMed]
Figure 1. General structure of child falling detection system.
Figure 1. General structure of child falling detection system.
Sensors 24 06451 g001
Figure 2. General structure of toddler falling detection system.
Figure 2. General structure of toddler falling detection system.
Sensors 24 06451 g002
Figure 3. Overall structure of toddler’s body information extraction.
Figure 3. Overall structure of toddler’s body information extraction.
Sensors 24 06451 g003
Figure 4. The GELAN module architecture.
Figure 4. The GELAN module architecture.
Sensors 24 06451 g004
Figure 5. Focus on the head information.
Figure 5. Focus on the head information.
Sensors 24 06451 g005
Figure 6. GHT-based bed surface identification.
Figure 6. GHT-based bed surface identification.
Sensors 24 06451 g006
Figure 7. Head and body center of gravity points.
Figure 7. Head and body center of gravity points.
Sensors 24 06451 g007
Figure 8. Images in self-built fall dataset of toddlers.
Figure 8. Images in self-built fall dataset of toddlers.
Sensors 24 06451 g008
Figure 9. Live examples of toddler body extraction.
Figure 9. Live examples of toddler body extraction.
Sensors 24 06451 g009
Figure 10. Recall and accuracy curves for 300 test samples.
Figure 10. Recall and accuracy curves for 300 test samples.
Sensors 24 06451 g010
Table 1. Child status division.
Table 1. Child status division.
Position RelationSafety SymbolState Type
H ∈ SpS1security status
H ⊄ Sp, K ∈ SpS2alarm status
H ⊄ Sp, K ∈ SpS3dangerous status
Table 2. Comparison of detection performance of different YOLO models.
Table 2. Comparison of detection performance of different YOLO models.
ModelPRF1[email protected]
YOLOv30.83940.82220.79030.8561
YOLOv50.76990.78100.78010.7994
YOLOv60.85880.88170.85910.9
YOLOv80.9050.92710.92110.9411
Improved YOLOv80.94910.95530.95720.9726
Table 3. Human body extraction results.
Table 3. Human body extraction results.
ResearchZhao et al. [35]Shen et al. [12] Ours
Accuracy0.780.730.953
Table 4. Child status identification results.
Table 4. Child status identification results.
Safety StatusWarning StatusDangerous Status
6096144
Table 5. Results of the infant status classification.
Table 5. Results of the infant status classification.
SensitivitySpecificityAccuracyPrecisionF-Value
96.67%95%96.33%98.72%97.68%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Z.; Tsui, B.; Ning, J.; Wu, Z. Falling Detection of Toddlers Based on Improved YOLOv8 Models. Sensors 2024, 24, 6451. https://doi.org/10.3390/s24196451

AMA Style

Yang Z, Tsui B, Ning J, Wu Z. Falling Detection of Toddlers Based on Improved YOLOv8 Models. Sensors. 2024; 24(19):6451. https://doi.org/10.3390/s24196451

Chicago/Turabian Style

Yang, Ziqian, Baiyu Tsui, Jiachuan Ning, and Zhihui Wu. 2024. "Falling Detection of Toddlers Based on Improved YOLOv8 Models" Sensors 24, no. 19: 6451. https://doi.org/10.3390/s24196451

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop