Next Article in Journal
Correlation-Assisted Pixel Array for Direct Time of Flight
Previous Article in Journal
Targeted FT-NIR and SERS Detection of Breast Cancer HER-II Biomarkers in Blood Serum Using PCB-Based Plasmonic Active Nanostructured Thin Film Label-Free Immunosensor Immobilized with Directional GNU-Conjugated Antibody
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

NABNet: Deep Learning-Based IoT Alert System for Detection of Abnormal Neck Behavior

School of Computer Science, Hangzhou Dianzi University, Hangzhou 310018, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(16), 5379; https://doi.org/10.3390/s24165379
Submission received: 16 July 2024 / Revised: 10 August 2024 / Accepted: 15 August 2024 / Published: 20 August 2024
(This article belongs to the Section Internet of Things)

Abstract

:
The excessive use of electronic devices for prolonged periods has led to problems such as neck pain and pressure injury in sedentary people. If not detected and corrected early, these issues can cause serious risks to physical health. Detectors for generic objects cannot adequately capture such subtle neck behaviors, resulting in missed detections. In this paper, we explore a deep learning-based solution for detecting abnormal behavior of the neck and propose a model called NABNet that combines object detection based on YOLOv5s with pose estimation based on Lightweight OpenPose. NABNet extracts the detailed behavior characteristics of the neck from global to local and detects abnormal behavior by analyzing the angle of the data. We deployed NABNet on the cloud and edge devices to achieve remote monitoring and abnormal behavior alarms. Finally, we applied the resulting NABNet-based IoT system for abnormal behavior detection in order to evaluate its effectiveness. The experimental results show that our system can effectively detect abnormal neck behavior and raise alarms on the cloud platform, with the highest accuracy reaching 94.13%.

1. Introduction

Sedentary lifestyles are common among young people today due to high dependence on electronic products such as mobile phones and computers. The incorrect sitting position is one of the manifestations of neck diseases [1,2]. Early detection and monitoring of abnormal behaviors can improve the life quality of sedentary people while reducing pressure on medical resources and providing significant commercial value.
Abnormal behavior can be defined as actions in which people may be in danger [3]. In this paper, we identify long-term head tilts and dropped heads as abnormal neck behaviors. Traditional methods rely on ambient devices to obtain environmental data and human data for abnormal behavior monitoring [4,5,6]. This approach is sensitive to environmental noise and lacks flexibility.
In recent years, studies about abnormal behavior detection have indicated that convolutional neural network (CNN)-based computer vision possesses high robustness and accuracy for detection [3,7,8]. While object detection-based methods are suitable for full-body abnormal behaviors such as falls and repeated activities [9,10], they are ineffective for detecting abnormal neck behavior, which involves smaller localized movements.
In this paper, we propose an intelligent detection and monitoring solution for abnormal neck behavior in indoor scenarios, called Neck Abnormal Behavior Network (NABNet). NABNet is designed to detect head tilt and dropped head events. First, NABNet uses YOLOv5s to extract features from the input in order to detect and track objects [11]. To improve the accuracy of small target detection, the Coordinate Attention (CA) [12] mechanism is added to YOLOv5s. Second, we obtain the key points and joint information of the body using Lightweight OpenPose [13]. Abnormal behavior events are detected by body orientation judgment, angle calculation, and correction. The high computational overhead of existing CNN-based models makes them challenging to execute on edge-end IoT devices while maintaining good real-time performance. Thus, we built our proposed NABNet IoT-based alert system to quickly and cheaply deployed on edge devices. Abnormal neck behaviors are detected in real-time, allowing alerts to be sent to healthcare professionals or caregivers.
Our main contributions can be summarized as follows:
  • This paper proposes a deep learning-based solution for monitoring and detecting abnormal neck behavior in sedentary people; specifically, the proposed NABNet detects head tilt and dropped head events.
  • The features inherent in abnormal neck behaviors are fully considered to alleviate false alarms. NABNet combines YOLOv5s with a CA mechanism to enhance the robustness of object detection, then uses OpenPose-guided skeleton and angle relationship information to judge the neck position.
  • A NABNet-based detection system was deployed on edge-end IoT devices, and its performance was tested in practical scenarios. Our experimental results demonstrate the effectiveness of NABNet for detecting abnormal neck behavior.

2. Related Works

2.1. Object Detection

In recent years, object detection methods based on neural network algorithms have become widely used in industrial fields, in particular for intelligent monitoring systems [14,15]. Current mainstream object detection algorithms are mainly categorized into two-stage and single-stage approaches [3]. The former approach generates a series of samples of the candidate box and then uses a CNN for sample classification. Two-stage approaches include networks such as the R-CNN [16], faster R-CNN [17], R-FCN [18], and Libra R-CNN [19]. The latter approach does not have to generate a candidate box, instead directly transforming the problem of target box localization into a regression problem [20]. Single-stage approaches include networks such as YOLO [21], SSD [22], and RetinaNet [23]. Single-stage algorithms are superior in terms of speed and have increasingly been applied in practical detection applications, including medical cancer cell detection [24], face detection [25], product detection [26], and more. However, their detection accuracy is slightly lower than two-stage algorithms, and they are not good at detecting small objects [27].
Thus, these state-of-the-art object detectors are ineffective for detecting abnormal neck behavior, as they cannot capture subtle feature differences such as neck rotation, which are critical for the classification and localization of abnormal neck behaviors. In this paper, the neck behavior features are learned by combining object detection and pose estimation to obtain better abnormal behavior detection results.

2.2. Abnormal Behavior Detection

Current methods for abnormal behavior detection can be coarsely classified into three types: ambient device-based, wearable-based, and computer vision-based [28,29]. The ambient device-based [30,31] and wearable sensor-based [28,32] methods require dedicated devices to collect ambient data (e.g., floor pressure, sound, vibration) or object movement data (e.g., speed, acceleration). However, ambient device-based methods are sensitive to environmental noise, leading to high false positive rates [29]. Ma, C. et al. [4] used pressure sensors to detect abnormal behaviors of people in wheelchairs, using a fuzzy inference system to evaluate movement and posture conversion intensity. Based on pressure sensors, PIR sensors, etc., Arifoglu, D. et al. [5] identified abnormal behaviors such as forgetting or repetition in the elderly. Tokas, P. et al. [6] used a Microsoft Kinect sensor to detect normal and abnormal sitting postures, achieving the highest accuracy of 92.85% with Random Forest and Support Vector Machine classifiers. Wearable-based approaches require individuals to wear sensors, which can cause discomfort and inconvenience. On the other hand, computer vision-based approaches usually use cameras to capture images, overcoming the limitations of fixed device placement inherent in wearable-based methods and allowing for simultaneous and visible capture of multiple events.
Many image-based abnormal behavior recognition works have emerged thanks to the popularity of camera monitoring and the powerful feature extraction ability of deep learning. The core technology of abnormal behavior detection is to recognize the category and location of abnormal behaviors in images captured by camera monitoring [3]. An increasing number of studies have been emerging into abnormal behavior detection as a subtask of object detection. Alruwaili, M. et al. [33] proposed a Yolov5-based real-time detection and tracking model for people with paralysis, limb defects, and other diseases. Wang L. et al. [2] proposed an attention-based spatiotemporal network to identify the behavior and location of abnormal activities in the elderly, such as multiple toilet visits, forgetting to wash dishes, etc. Fang, M.-T. et al. [34] proposed a real-time abnormal behavior detection method using improved YOLOv3. These methods attempt to handle abnormal behavior detection through improved generic object detection methods based on deep learning and temporal surveillance information.
In addition, many methods try to define different kinds of abnormal behavior in advance depending on the specific scenario. Mehmood, A. et al. [35] defined human falls, certain types of suspicious behavior, and violent acts as abnormal activities. They provided a lightweight framework to represent and differentiate between normal and abnormal events effectively. Fu, Y. et al. [8] proposed a lightweight GD-YOLO network based on YOLOv7 for detecting abnormal behaviors, including smoking and using mobile phones.
In this paper, based on behavior in office scenarios, we define head tilt and dropped head postures as abnormal behaviors of the neck. We focus on detecting abnormal behavior of the neck joints and combine object detection with pose estimation to better detect abnormal behavior.

3. Methods

3.1. Overview

This paper proposes an abnormal behavior detection network called NABNet based on YOLOv5s and Lightweight OpenPose for detecting head tilt postures of sedentary people in indoor scenarios. Furthermore, we present a NABNet-based IoT system that leverages the trained NABNet and hardware equipment, as shown in Figure 1. Video is collected by a camera and processed on edge devices to detect abnormal behaviors, with the results uploaded to the cloud for issuing alarms. The system’s main components are video gathering, image transmission and processing, and NABNet-based detection parts. Based on the YOLOv5s object detection network and the Lightweight OpenPose pose estimation network, NABNet extracts the detailed behavior characteristics of the head and neck from global to local to detect abnormal head tilt and dropped head behaviors.

3.2. Object Detection and Tracking

YOLOv5s is a CNN-based object detection model composed of backbone, neck, and head networks. The input is processed by the backbone network (CSPDarknet53 [36]) for feature extraction, the neck network for feature fusion, and the head network for object prediction.The backbone network uses Focus, CBL, CSP, and SPP structures to reduce the calculation without reducing the accuracy [37]. The neck network uses top-down FPN and bottom-up PAN modules for feature fusion. Finally, we adopt the GIOU loss to supervise the network [38]. The CA mechanism is embedded in the backbone structure of YOLOv5s to enhance the feature extraction and object attention capability, as shown in Figure 2. Unlike other mechanisms that convert the features extracted by the network into a single feature vector through 2D global pooling, the CA mechanism decomposes channel attention into two 1D feature encoding processes and then separately aggregates features along two spatial directions; one spatial direction captures long-range dependencies, while the other retains precise positional information. Two feature maps with specific spatial dependencies are obtained by applying an activation function, enhancing attention towards the target of interest.
We track the object to analyze the behavior category according to the object’s motion mode. For this, we design an extended Kalman filter-based tracker, as shown in Figure 3. The tracking process is divided into two parts: initialization and tracking; in this paper, the frame of YOLOv5s detection failure is taken as the third frame of the tracker, while the first two frames are taken as the tracker’s initial state. After initialization, the prior state and error covariance matrix are calculated, and multiple candidate regions are created near the preceding state. The candidate region is matched with the tracking result of the previous frame using the difference hash to select the best-matching candidate region. The matching value is compared with a threshold. If this threshold is exceeded, the best-matching candidate region is used as the observation value to update the system; otherwise, the matching is considered to have failed, and waiting for the detector to detect the object again is necessary.
The traditional Kalman filter assumes a Gaussian distribution and linearity; however, object tracking of the body is nonlinear due to object deformation and environmental changes. The extended Kalman filter can solve the problem of nonlinearity via Taylor expansion [39]. The state equation and observation equation of the extended Kalman filter are as follows:
x k = f ( x k 1 ) + w k 1 ,
z k = h ( x k ) + v k ,
where x k and x k 1 denote the state vector with k and k 1 , respectively, f ( ) is the state transition function, and w k 1 is the process noise, which is assumed to conform to the multivariate normal distribution with a mean value of 0 and covariance matrix of Q. The process noise is due to uncertain factors in practical scenarios, such as sudden acceleration, deceleration, and turning. In addition, z k is the observation vector, h ( ) is the state observation function which converts the state space into the observation space, and v k is the observation noise, which is assumed to conform to a multivariate normal distribution with a mean value of 0 and covariance matrix of R.

3.3. Detection of Abnormal Neck Behavior

Because the neck has subtle movements, using the object detection method directly is impossible, as it can only make judgments based on violent movements (see Section 3.2). Inspired by Maji D. et al. [40], this paper uses the YOLOv5s object detection model combined with the Lightweight OpenPose pose estimation method [13,41] to determine the neck’s state. The advantages of this approach are as follows: (1) combining YOLOv5s with Lightweight OpenPose increases the robustness of the model against occlusion compared to using Lightweight OpenPose alone, which can cause detection errors if key points and joint information are lost; (2) having Lightweight OpenPose only process the objects extracted by YOLOv5s helps it to ignore irrelevant objects, which reduces the amount of calculation and improves real-time performance.
We determine abnormal behavior according to the neck angle. The evaluation criteria of the dropped head state consist of the ratio of the neck-to-nose vector to the vector from the neck to the shoulder on one side. The evaluation criteria of the head tilt state are the angle between the neck and the shoulder. When the camera is not in the front position of the object, there is a difference between the detected angle and the actual angle of the neck, as shown in Figure 4. To solve this problem, we first determine the body’s orientation, then use the camera calibration to obtain the mapping relationship between the two-dimensional and three-dimensional object, and finally obtain the angle between the object and the camera by calculating the Euler angle. Then, the angle correction is carried out by affine transformation [42] to reduce the timidity of the angles, as shown in Figure 5. The affine transformation is as follows:
u c v c 1 = cos ϕ 0 0 sin ϕ 1 0 0 0 1 u v 1
where ϕ is the yaw angle and w is the neck angle, determined as follows:
w = a r c cos ( a c · b c a c b c ) .

4. NABNet-Based IoT Alert System

We constructed a NABNet-based IoT alert system for abnormal behavior detection to identify head tilt and dropped head postures. The structure of the system is illustrated in Figure 6. The edge device must process the video the camera collects to track the object and determine the abnormal behavior, which requires matching computational power. The selected edge devices must also be extensible to enhance the system’s availability. We chose the Raspberry Pi 3B as the control terminal. The Raspberry Pi 3B is known for its compact size and powerful functionality and is widely used in fields such as smart homes, media devices, and industrial control. Its primary parameters are detailed in Table 1. We selected the DF200 camera, which has a resolution of 1280 × 1080P and a speed of 30 frames per second.
The system captures indoor video through camera monitoring and transmits it to edge devices for analysis using NABNet. The edge devices continue to track objects even when no abnormal behavior events are detected. In the event of a detection, the system interacts with the cloud and issues an alert.

5. Experiments

In this section, we introduce the experiment methodology and evaluate the performance of NABNet. We also carry out a series of ablation studies.

5.1. Setup

We conducted field experiments to evaluate the effectiveness of our approach, consisting of a camera and 1.7 m bracket, embedded devices, and a computer, as shown in Figure 7. We selected 80 participants, including 60 males and 20 females, aged between 24 and 30 years old and varying in height between 1.55 m and 1.85 m. The participants performed positive pose (head tilt or dropped head) and negative pose (front position) at distances of 0.5 m and 0.75 m from the camera.
The performance of the proposed model was evaluated using the Precision, Recall, and Accuracy:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
A c c u r a c y = T P + T N T P + F P + T N + F N
where T P is the number of positive samples correctly detected, F N is the number of negative samples incorrectly detected, F P is the number of positive samples falsely classified as negative ones, and T N is the number of negative samples correctly detected.

5.2. Evaluation of NABNet-Based IoT System

The participants performed positive and negative poses facing the camera in the front position, 30° right rotation, 30° left rotation, 60° right rotation, and 60° left rotation, as shown in Figure 8. The sample number was 160 (80 persons × two poses). The experimental results in Table 2 show that the detection accuracy in the front position and at 30° rotation is higher than that at 60° rotation. In general, as the rotation angle increases, the detection accuracy gradually decreases.
Figure 9 shows the abnormal behavior neck detection results obtained with our multi-node system; (a) shows results for the control group when the participant is standing and is detected correctly by the system, while (b) shows the results when the participants are partially occluded. In the latter case, the system is still able to detect abnormal behavior of the neck. The multi-node detection and final results are shown above the monitoring video, with the final result comprising a comprehensive judgment of the multi-node detection results. The remote monitoring data on the cloud platform are shown below the monitoring video, indicating whether an abnormal behavior event occurs and an alarm was realized.

5.3. Ablation Studies

Ablation experiments were conducted on the different improvement stages of NABNet, including the CA mechanism, tracker, and angle correction. We selected five groups of experimental fragments from the Multicam Fall Database [43], Le2i [44], SIMPLE Fall Detection Dataset [45], and one group photographed in a low illumination scene.
Table 3 shows the influence of the different components on our model. It can be observed that incorporating the CA mechanism leads to a 3.58% improvement in accuracy, indicating that adding the CA mechanism enhances the model’s accuracy in detecting small targets.
To verify the effectiveness of the proposed tracker, we compared it with Kernelized Correlation Filters (KCF) [46] and Tracking Learning Detection (TLD) [47]. Table 3 shows that the accuracy of our tracker is on par with KCF, while the accuracy of TLD is significantly lower, indicating that TLD incurs errors in object tracking. In order to better meet the system’s real-time performance requirements, we tested the frame rates of three different trackers, with the results presented in Table 4. Our tracker achieved the highest frame rate, with KCF being 25.57 frames per second (fps) slower. It demonstrates that our algorithm offers superior detection speed and improved real-time capabilities while maintaining comparable accuracy.
By utilizing angle correction via affine transformation, our method can correct the object to the forward position when facing the camera at an angle to obtain the actual neck angle information. As shown in the last two lines of Table 3, the accuracy of the model is improved by 8.72% after affine transformation.

6. Conclusions

Focusing on the local behavior of abnormal neck postures, this paper proposes an abnormal behavior detection network called NABNet based on object detection and pose estimation. To accommodate devices with limited computing power, we have also developed an IoT detection and alert system for edge devices based on NABNet, which can realize monitoring and early alerts on a cloud platform. Our experimental results show the effectiveness of the proposed approach.
As with the majority of studies, the design of the current study is subject to limitations. We have not yet validated the effectiveness of our approach on a large-scale public dataset specifically for neck abnormal behavior due to the absence of such a dataset, which highlights the need for further research in this area. In the future, we will consider increasing the number of key points and combining NABNet with other pose estimation models to improve its ability to detect multiple abnormal behaviors.

Author Contributions

H.Q. (Hongshuai Qin): Methodology, writing—original draft, writing—review and editing, project administration. M.C.: Validation, formal analysis. H.Q. (Huibin Qin): Conceptualization, writing—review and editing, funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Zhejiang Key Research and Development Program of China under Grant 2017C01027.

Data Availability Statement

The datasets generated and analyzed during the current study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Doewes, R.I.; Gharibian, G.; Zaman, B.A.; Akhavan-Sigari, R. An updated systematic review on the effects of aerobic exercise on human blood lipid profile. Curr. Probl. Cardiol. 2023, 48, 101–108. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, L.; Zhou, Y.; Li, R.; Ding, L. A fusion of a deep neural network and a hidden Markov model to recognize the multiclass abnormal behavior of elderly people. Knowledge-Based Syst. 2022, 252, 109351. [Google Scholar] [CrossRef]
  3. Liu, C.; Zhang, Y.; Xue, Y.; Qian, X. AJENet: Adaptive joints enhancement network for abnormal behavior detection in office scenario. IEEE Trans. Circuits Syst. Video Technol. 2023, 252, 1427–1440. [Google Scholar] [CrossRef]
  4. Ma, C.; Du, J.; Gravina, R. Abnormal behavior detection based on activity level using fuzzy inference system for wheelchair users. Human-Centric Comput. Inf. Sci. 2022, 12, 10.22967. [Google Scholar]
  5. Arifoglu, D.; Wang, Y.; Bouchachia, A. Detection of dementia-related abnormal behaviour using recursive auto-encoders. Sensors 2021, 21, 260. [Google Scholar] [CrossRef] [PubMed]
  6. Tokas, P. Machine learning based text neck syndrome detection using Microsoft Kinect sensor. Mater. Today Proc. 2023, 80, 3751–3756. [Google Scholar] [CrossRef]
  7. Alruwaili, M.; Siddiqi, M.H.; Atta, M.N.; Arif, M. Deep learning and ubiquitous systems for disabled people detection using YOLO models. Comput. Hum. Behav. 2024, 154, 108150. [Google Scholar] [CrossRef]
  8. Fu, Y.; Ran, T.; Xiao, W.; Yuan, L.; Zhao, J.; He, L.; Mei, J. GD-YOLO: An improved convolutional neural network architecture for real-time detection of smoking and phone use behaviors. Digit. Signal Process. 2024, 151, 104554. [Google Scholar] [CrossRef]
  9. Cao, C.; Lan, C.; Zhang, Y.; Zeng, W.; Lu, H.; Zhang, Y. Skeleton-based action recognition with gated convolutional neural networks. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 3247–3257. [Google Scholar] [CrossRef]
  10. Lentzas, A.; Vrakas, D. Non-intrusive human activity recognition and abnormal behavior detection on elderly people: A review. Artif. Intell. Rev. 2020, 53, 1975–2021. [Google Scholar] [CrossRef]
  11. YOLOv5. Available online: https://github.com/ultralytics/yolov5 (accessed on 22 November 2022).
  12. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 25 June 2021; pp. 13713–13722. [Google Scholar]
  13. Osokin, D. Real-time 2d multi-person pose estimation on cpu: Lightweight openpose. arXiv 2018, arXiv:1811.12004. [Google Scholar]
  14. Zhang, D.; Han, J.; Cheng, G.; Yang, M.H. Weakly supervised object localization and detection: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5866–5885. [Google Scholar] [CrossRef] [PubMed]
  15. Chaoxia, C.; Shang, W.; Zhang, F.; Cong, S. Weakly aligned multimodal flame detection for fire-fighting robots. IEEE IEEE Trans. Ind. Inform. 2022, 19, 2866–2875. [Google Scholar] [CrossRef]
  16. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 80–587. [Google Scholar]
  17. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 2866–2875. [Google Scholar] [CrossRef] [PubMed]
  18. Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. Neural Inf. Process. Syst. 2016, 29, 379–387. [Google Scholar]
  19. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra r-cnn: Towards balanced learning for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 821–830. [Google Scholar]
  20. Chen, Z.; Chen, D.; Zhang, Y.; Cheng, X.; Zhang, M.; Wu, C. Deep learning for autonomous ship-oriented small ship detection. Saf. Sci. 2020, 130, 104812. [Google Scholar] [CrossRef]
  21. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  22. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. pp. 21–37. [Google Scholar]
  23. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  24. Li, Y.; Xue, Y.; Li, L.; Zhang, X.; Qian, X. Domain adaptive box-supervised instance segmentation network for mitosis detection. IEEE Trans. Med. Imag. 2022, 41, 2469–2485. [Google Scholar] [CrossRef] [PubMed]
  25. Li, X.; Lai, S.; Qian, X. Dbcface: Towards pure convolutional neural network face detection. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1792–1804. [Google Scholar] [CrossRef]
  26. Liu, C.; Da, Z.; Liang, Y.; Xue, Y.; Zhao, G.; Qian, X. Product recognition for unmanned vending machines. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 1584–1597. [Google Scholar] [CrossRef] [PubMed]
  27. Li, H.; Wu, D.; Zhang, W.; Xiao, C. YOLO-PL: Helmet wearing detection algorithm based on improved YOLOv4. Digit. Signal Process. 2023, 144, 104283. [Google Scholar] [CrossRef]
  28. Qiu, J.; Yan, X.; Wang, W.; Wei, W.; Fang, K. Skeleton-based abnormal behavior detection using secure partitioned convolutional neural network model. IEEE J. Biomed. Health Inform. 2021, 26, 5829–5840. [Google Scholar] [CrossRef] [PubMed]
  29. Naser, A.; Lotfi, A.; Mwanje, M.D.; Zhong, J. Privacy-preserving, thermal vision with human in the loop fall detection alert system. IEEE T. Hum.-Mach. Syst. 2022, 53, 164–175. [Google Scholar] [CrossRef]
  30. Jin, F.; Zhang, R.; Sengupta, A.; Cao, S.; Hariri, S.; Agarwal, N.K.; Agarwal, S.K. Multiple patients behavior detection in real-time using mmWave radar and deep CNNs. In Proceedings of the 2019 IEEE Radar Conference, Boston, MA, USA, 22–26 April 2019; pp. 1–6. [Google Scholar]
  31. Okumura, N.; Yamanoi, Y.; Kato, R.; Yamamura, O. Fall detection and walking estimation using floor vibration for solitary elderly people. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019; pp. 1437–1442. [Google Scholar]
  32. Santos, G.L.; Endo, P.T.; Monteiro, K.H.d.C.; Rocha, E.d.S.; Silva, I.; Lynn, T. Accelerometer-based human fall detection using convolutional neural networks. Sensors 2019, 19, 1644. [Google Scholar] [CrossRef] [PubMed]
  33. Alruwaili, M.; Atta, M.N.; Siddiqi, M.H.; Khan, A.; Khan, A.; Alhwaiti, Y.; Alanazi, S. Deep Learning-Based YOLO Models for the Detection of People With Disabilities. IEEE Access 2023, 12, 2543–2566. [Google Scholar] [CrossRef]
  34. Fang, M.-T.; Chen, Z.-J.; Przystupa, K.; Li, T.; Majka, M.; Kochan, O. Examination of abnormal behavior detection based on improved YOLOv3. Electronics 2021, 10, 197. [Google Scholar] [CrossRef]
  35. Mehmood, A. Lightanomalynet: A lightweight framework for efficient abnormal behavior detection. Sensors 2021, 21, 8501. [Google Scholar] [CrossRef] [PubMed]
  36. Bochkovskiy, A.; Wang, C.; Liao, H. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  37. Gündüz, M.; Işık, G. A new YOLO-based method for real-time crowd detection from video and performance analysis of YOLO models. J. Real-Time Image Process. 2023, 20, 5–17. [Google Scholar] [CrossRef]
  38. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognitio, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  39. Dai, Y.; Yu, S.; Yan, Y. An adaptive EKF-FMPC for the trajectory tracking of UVMS. IEEE J. Ocean. Eng. 2019, 45, 699–713. [Google Scholar] [CrossRef]
  40. Maji, D.; Nagori, S.; Mathew, M.; Poddar, D. Yolo-pose: Enhancing yolo for multi person pose estimation using object keypoint similarity loss. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2637–2646. [Google Scholar]
  41. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  42. Stearns, C.; Kannappan, K. Method for 2-D Affine Transformation of Images. US Patent US5475803, 12 December 1995. Application No. 07/911832, 10 July 1992. [Google Scholar]
  43. Auvinet, E.; Rougier, C.; Meunier, J.; St-Arnaud, A.; Rousseau, J. Multiple cameras fall dataset. DIRO-Université Montréal Tech. Rep. 2010, 1350, 24. [Google Scholar]
  44. Charfi, I.; Miteran, J.; Dubois, J.; Atri, M.; Tourki, R. Optimized spatio-temporal descriptors for real-time fall detection: Comparison of support vector machine and Adaboost-based classification. J. Electron. Imag. 2013, 22, 041106. [Google Scholar] [CrossRef]
  45. Chua, J.-L.; Chang, Y.C.; Lim, W.K. A simple vision-based fall detection technique for indoor video surveillance. Signal Image Video Process. 2015, 9, 623–633. [Google Scholar] [CrossRef]
  46. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 37, 583–596. [Google Scholar] [CrossRef] [PubMed]
  47. Kalal, Z.; Mikolajczyk, K.; Matas, J. Tracking-learning-detection. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 1409–1422. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Schematic of the proposed NABNet-based IoT system for abnormal behavior detection.
Figure 1. Schematic of the proposed NABNet-based IoT system for abnormal behavior detection.
Sensors 24 05379 g001
Figure 2. CA mechanism structure.
Figure 2. CA mechanism structure.
Sensors 24 05379 g002
Figure 3. Flowchart of object tracking.
Figure 3. Flowchart of object tracking.
Sensors 24 05379 g003
Figure 4. (a) The neck state when the object side is towards the camera and (b) the actual neck state.
Figure 4. (a) The neck state when the object side is towards the camera and (b) the actual neck state.
Sensors 24 05379 g004
Figure 5. Angle correction via affine transformation. a denotes the vector between the neck and the head, and b represents the vector between the neck and the shoulder.
Figure 5. Angle correction via affine transformation. a denotes the vector between the neck and the head, and b represents the vector between the neck and the shoulder.
Sensors 24 05379 g005
Figure 6. NABNet-based IoT alert system structure.
Figure 6. NABNet-based IoT alert system structure.
Sensors 24 05379 g006
Figure 7. Experimental setup.
Figure 7. Experimental setup.
Sensors 24 05379 g007
Figure 8. Representative images of positive and negative poses.
Figure 8. Representative images of positive and negative poses.
Sensors 24 05379 g008
Figure 9. Illustration of abnormal neck behavior detection displayed on the server screens and cloud.
Figure 9. Illustration of abnormal neck behavior detection displayed on the server screens and cloud.
Sensors 24 05379 g009
Table 1. Primary parameters of the Raspberry Pi 3B.
Table 1. Primary parameters of the Raspberry Pi 3B.
ParametersConfiguration
SOCCM2711
CPUARM Cortex-A72 1.5 GHz
GPUBroadcom VideoCore IV
Memory4 GB LPDDR4
Power5 V Micro USB
Supported SystemsRaspbian/Ubuntu/Windows10/Linux
Table 2. Evaluation of our system based on the obtained samples.
Table 2. Evaluation of our system based on the obtained samples.
0.5 m0.75 m
PositionRecall (%)Precision (%)Accuracy (%)Recall (%)Precision (%)Accuracy (%)
Front position97.5096.3096.8896.2596.2596.25
30° right rotation97.5098.7398.1395.0097.4496.25
30° left rotation95.0097.4496.2597.5093.9895.63
60° right rotation86.2592.0089.3886.2587.3486.88
60° left rotation88.7591.0390.0087.5088.6188.13
Mean9395.194.1392.592.7292.63
Table 3. Influence of components on our model.
Table 3. Influence of components on our model.
ModelCA MechanismTrackerAngle CorrectionAccuracy (%)
YOLOv5s 67.44
YOLOv5s 71.02
YOLOv5sKCF 87.45
YOLOv5sTLD 78.63
YOLOv5sOurs 85.41
YOLOv5sOurs94.13
Table 4. Comparison experiments of the detection frame rate with different trackers.
Table 4. Comparison experiments of the detection frame rate with different trackers.
TrackerDetection Frame Rate (fps)
KCF17.61
TLD9.95
Ours43.18
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qin, H.; Cai, M.; Qin, H. NABNet: Deep Learning-Based IoT Alert System for Detection of Abnormal Neck Behavior. Sensors 2024, 24, 5379. https://doi.org/10.3390/s24165379

AMA Style

Qin H, Cai M, Qin H. NABNet: Deep Learning-Based IoT Alert System for Detection of Abnormal Neck Behavior. Sensors. 2024; 24(16):5379. https://doi.org/10.3390/s24165379

Chicago/Turabian Style

Qin, Hongshuai, Minya Cai, and Huibin Qin. 2024. "NABNet: Deep Learning-Based IoT Alert System for Detection of Abnormal Neck Behavior" Sensors 24, no. 16: 5379. https://doi.org/10.3390/s24165379

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop