Next Article in Journal
A Wearable Personalised Sonification and Biofeedback Device to Enhance Movement Awareness
Previous Article in Journal
Data Analytics for Predicting Situational Developments in Smart Cities: Assessing User Perceptions
Previous Article in Special Issue
Non-Contact Vision-Based Techniques of Vital Sign Monitoring: Systematic Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep-Learning-Based CPR Action Standardization Method

1
Jiangsu Tuoyou Information Intelligent Technology Research Institute Co., Ltd., Nanjing 210012, China
2
School of Computer Science, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
3
School of Internet of Things, Nanjing University of Posts and Telecommunications, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(15), 4813; https://doi.org/10.3390/s24154813
Submission received: 5 June 2024 / Revised: 14 July 2024 / Accepted: 23 July 2024 / Published: 24 July 2024
(This article belongs to the Special Issue AI-Based Automated Recognition and Detection in Healthcare)

Abstract

:
In emergency situations, ensuring standardized cardiopulmonary resuscitation (CPR) actions is crucial. However, current automated external defibrillators (AEDs) lack methods to determine whether CPR actions are performed correctly, leading to inconsistent CPR quality. To address this issue, we introduce a novel method called deep-learning-based CPR action standardization (DLCAS). This method involves three parts. First, it detects correct posture using OpenPose to recognize skeletal points. Second, it identifies a marker wristband with our CPR-Detection algorithm and measures compression depth, count, and frequency using a depth algorithm. Finally, we optimize the algorithm for edge devices to enhance real-time processing speed. Extensive experiments on our custom dataset have shown that the CPR-Detection algorithm achieves a mAP0.5 of 97.04%, while reducing parameters to 0.20 M and FLOPs to 132.15 K. In a complete CPR operation procedure, the depth measurement solution achieves an accuracy of 90% with a margin of error less than 1 cm, while the count and frequency measurements achieve 98% accuracy with a margin of error less than two counts. Our method meets the real-time requirements in medical scenarios, and the processing speed on edge devices has increased from 8 fps to 25 fps.

1. Introduction

Out-of-hospital cardiac arrest (OHCA) is a critical medical emergency with a substantial impact on public health, exhibiting annual incidence rates of approximately 55 per 100,000 people in North America and 59 per 100,000 in Asia. Without timely intervention, OHCA can lead to irreversible death within 10 min [1]. Studies have demonstrated that CPR and AED defibrillation performed by nearby volunteers or citizens significantly improve survival rates [1,2,3]. Standard CPR procedures are known to enhance survival outcomes in cardiac arrest patients [3]. However, the dissemination of CPR skills remains limited in many countries, primarily relying on mannequins and instructors, leading to high costs and inefficiencies. Traditional AED devices also lack the capability to prevent harm caused by improper operation [4,5].
Current CPR methods have several limitations, particularly in their effectiveness during real emergency situations. Traditional CPR training relies heavily on classroom simulations, which cannot replicate the pressure and urgency of actual cardiac arrest scenarios. This can lead to improper performance during real emergencies [6]. Although virtual reality (VR) and augmented reality (AR) technologies are being used to enhance CPR training, they remain primarily educational tools and are not widely integrated into real-time emergency applications [7,8]. Moreover, mainstream CPR techniques have not fully incorporated artificial intelligence (AI) assistance; advancements have focused more on mechanical devices and VR/AR training rather than real-time AI intervention [8,9]. Recent advancements in CPR algorithms have started to address these issues by integrating computational models and machine-learning techniques. For instance, the use of integrated computational models of the cardiopulmonary system to evaluate current CPR guidelines has shown potential in improving CPR effectiveness [10]. Additionally, machine learning has been used to identify higher survival rates during extracorporeal cardiopulmonary resuscitation, significantly enhancing survival outcomes [11]. The future trends in CPR technology indicate that the combination of AI and machine learning will continue to evolve, potentially predicting and shaping technological innovations in this field [12]. To bridge the gap between training and real-time application, this paper proposes the first application of posture-estimation and object-detection algorithms on AEDs to assist in real-time CPR action standardization, extending their use to actual emergency rescues. This innovative approach addresses the lack of real-time AI-assisted intervention in current CPR methods, thereby improving the accuracy and effectiveness of lifesaving measures during OHCA incidents. By integrating AI technology into AED devices, we aim to provide immediate feedback and corrective actions during CPR, potentially increasing survival rates and reducing risks associated with improper CPR techniques. This approach represents a significant advancement over traditional methods, which lack the ability to dynamically adjust in real-time and guide rescuers [13,14].
To enhance real-time medical interventions, advanced pose estimation techniques like OpenPose are highly beneficial. Developed by the Perceptual Computing Lab at Carnegie Mellon University, OpenPose is a pioneering open-source library for real-time multi-person pose estimation. It detects human body, hand, facial, and foot keypoints simultaneously [15]. Initially, OpenPose used a dual-branch CNN architecture to produce confidence maps and part affinity fields (PAFs) for associating body parts into a coherent skeletal structure. Subsequent improvements focused on refining PAFs, integrating foot keypoint detection, and introducing multi-stage CNNs for iterative prediction refinement [16,17]. Supported by continuous research and updates, OpenPose remains robust and efficient for edge computing and real-time applications [18], solidifying its status as a leading tool in diverse and complex scenarios.
In addition, deploying neural-network models on AED edge devices to recognize and standardize rescuers’ CPR actions can effectively improve the survival rate of cardiac arrest patients. However, deploying neural-network models on embedded systems faces challenges, such as high weight, insufficient computational power, and low running speed [19]. Most early lightweight object detection models were based on MobileNet-SSD (single shot multibox detector) [20]. Installing these models on some high-end smartphones can achieve sufficiently high running speeds [21]. However, due to insufficient ARM cores for running neural networks, model execution speed is slow on low-cost advanced RISC machine (ARM) devices [22].
In recent years, various lightweight object-detection networks have been proposed and widely applied in traffic management [23,24,25,26], fire warning systems [27], anomaly detection [28,29,30], and facial recognition [31,32,33]. Redmon et al. [34] introduced an end-to-end object-detection model using Darknet53, incorporating k-means clustering for anchor boxes, multi-label classification for class probabilities, and a feature pyramid network for multi-scale bounding box prediction. Wong et al. [35] developed Yolo Nano, a compact network for embedded object detection with a model size of approximately 4.0 MB. Hu et al. [36] improved the Yolov3-tiny network by using depthwise distributed convolutions and squeeze-and-excitation blocks, creating Micro-Yolo to reduce parameters and optimize performance. Lyu [37] proposed NanoDet, an anchor-free model using generalized focal loss and GhostPAN for enhanced feature fusion, increasing accuracy on the COCO dataset by 7% mAP. Ge et al. [38] modified Yolo to an anchor-free mode with a decoupled head and SimOTA strategy, significantly enhancing performance. For example, Yolo Nano achieved 25.3% AP on the COCO dataset with only 0.91 M parameters and 1.08 G FLOPs, surpassing NanoDet by 1.8%, while the improved Yolov3 AP increased to 47.3%, exceeding the current best practice by 3.0%. Yolov5 Lite [39] optimized inference speed by adding shuffle channels and pruning head channels while maintaining high accuracy. Dogqiuqiu [40] developed the Yolo-Fastest series for single-core real-time inference, reducing CPU usage. Yolo-FastestV2 used the ShufflenetV2 backbone, decoupled the detection head, reduced parameters, and improved the anchor-matching mechanism. Dogqiuqiu [41] further proposed FastestDet, simplifying to a single detection head, transitioning to anchor-free, and increasing candidate objects across grids for ARM platforms. However, for our dataset, FastestDet underperformed, mainly due to its single detection head design, limiting the utilization of features with different receptive fields and lacking sufficient feature fusion, resulting in insufficient accuracy in locating small objects.
This paper proposes a standardized CPR action-detection method based on AED, utilizing skeletal points to assist in posture estimation. We develop the CPR-Detection algorithm based on Yolo-FastestV2, which includes a novel compression depth-calculation method that maps actual depth by analyzing the wristband’s displacement. Additionally, we optimize the computation for edge devices to enhance their speed and accuracy. The main contributions of this paper include:
(1)
Introducing a novel method called deep-learning-based CPR action standardization (DLCAS) and developing a custom CPR action dataset. Additionally, we incorporated OpenPose for pose estimation of rescuers.
(2)
Proposing an object-detection model called CPR-Detection and introducing various methods to optimize its structure. Based on this, we developed a new method for measuring compression depth by analyzing wristband displacement data.
(3)
Proposing an optimized deployment method for automated external defibrillator (AED) edge devices. This method addresses the issues of long model inference time and low accuracy that exist in current edge device deployments of deep-learning algorithms.
(4)
Conducting extensive experimental validation to confirm the effectiveness of the improved algorithm and the feasibility of the compression depth-measurement scheme.

2. Methods

As shown in Figure 1, the overall workflow of this study is divided into three parts. The first part involves the experimental preparation phase, which includes dataset collection, image pre-processing and augmentation, dataset splitting, training, and then testing the trained model to obtain performance metrics. The second part presents the flowchart of the DLCAS, covering pose estimation, object-detection network, and depth measurement, ultimately yielding depth, compression count, and frequency. The third part describes the model’s inference and application. The captured images, processed through the optimized AED edge devices, eventually become CPR images with easily assessable metrics.
In this section, we first introduce the principles of OpenPose, followed by the design details of CPR-Detection. Next, we explain the depth measurement scheme based on object-detection algorithms. Finally, we discuss the optimization of computational methods for edge devices.

2.1. OpenPose

In edge computing devices for medical posture assessment, processing speed and real-time performance are crucial. Therefore, we chose OpenPose for skeletal-point detection due to its efficiency and accuracy. OpenPose employs a dual-branch architecture that generates confidence maps for body-part detection and part affinity fields (PAFs) to assemble these parts into a coherent skeletal structure. This method enables precise and real-time posture analysis, which is essential for medical applications. Traditional pose-estimation algorithms often involve complex computations that delay processing. OpenPose optimizes this process by focusing on key points and their connections, significantly reducing computational load and improving speed. It detects body parts independently before associating them, enhancing accuracy and efficiency by minimizing redundant computations. Overall, OpenPose allows for accurate and swift identification and assessment of human postures, making it ideal for real-time medical applications. Its efficient processing and reduced computational overhead make it suitable for deployment in edge computing devices used in emergency medical care, ensuring both reliability and speed in critical situations.
As shown in Figure 2, the workflow of OpenPose starts with feature extraction through a backbone network. These features pass through Stage 0, producing keypoint heatmaps and PAFs. Keypoint heatmaps indicate confidence scores for the presence of keypoints at each location, while PAFs encode the associations between pairs of keypoints, capturing spatial relationships between different body parts. These outputs are refined in subsequent stages, iteratively improving accuracy. Finally, the keypoint heatmaps and PAFs are processed to generate the final skeletal structure, combining keypoints according to the PAFs to form a coherent and accurate representation of the human pose. This method ensures precise and real-time posture analysis, making it highly effective for applications in medical posture assessment, particularly in edge computing devices used in emergency medical care, ensuring both reliability and speed in critical situations [16].

2.2. CPR-Detection

In this study, we provide a detailed explanation of CPR-Detection. As illustrated in Figure 3, the model consists of three components: the backbone network ShuffleNetV2, the STD-FPN feature-fusion module, and the detection head. The STD-FPN feature-fusion module incorporates the MLCA attention mechanism, and the detection head integrates PConv position-enhanced convolution.

2.2.1. PConv

In edge computing devices for medical emergency care, we need to prioritize processing speed due to performance and real-time processing requirements. Therefore, we chose partial convolution (PConv) to replace depthwise separable convolution (DWSConv) in Yolo-FastestV2. PConv offers higher efficiency while maintaining performance, meeting the needs for real-time processing [42].
As shown in Figure 4a, DWSConv works by first performing depthwise convolution on the input feature map, grouping by channels, and then using 1 × 1 convolution to integrate all channel information. However, this depthwise convolution can lead to computational redundancy in practical applications. The principle of PConv, illustrated in Figure 4b, involves performing regular convolution operations on a portion of the input channels while leaving the other channels unchanged. This design significantly reduces computational load and memory access requirements because it processes only a subset of feature channels. PConv only performs convolution on a specific proportion of the input features, resulting in lower FLOPs compared to DWSConv, thereby reducing computational overhead and improving model efficiency. In summary, PConv enhances the network’s feature representation capability by focusing on crucial spatial information without sacrificing detection performance.
This strategy not only improves the network’s processing speed but also enhances the extraction and focus on key feature channels, making it essential for real-time object-detection systems. Additionally, by reducing redundant computations, the application of PConv lowers model complexity and increases model generalization, ensuring robustness and efficiency in complex medical emergency scenarios. Therefore, PConv is an ideal convolution method for medical emergency devices, enabling real-time object detection while ensuring reliability and efficiency on edge computing devices.

2.2.2. MLCA

In emergency medical scenarios, complex backgrounds can interfere with the effective detection of wristbands. To address this, we introduce the mixed local channel attention (MLCA) module to enhance the model’s performance in processing channel-level and spatial-level information. As illustrated in Figure 5, MLCA combines local and global context information to improve the network’s feature representation capabilities. This focus on critical features enhances both the accuracy and efficiency of target detection [43].
The core of MLCA lies in its ability to process and integrate both local and global feature information simultaneously. Specifically, MLCA first performs two types of pooling operations on the input feature vector: local pooling, which captures fine-grained spatial details, and global pooling, which extracts broader contextual information. These pooled features are then sent to separate branches for detailed analysis. Each branch output is further processed by convolutional layers to extract cross-channel interaction information. Finally, the pooled features are restored to their original resolution through an unpooling operation and fused using an addition operation, achieving comprehensive attention modulation. Compared to traditional attention mechanisms, such as SENet [44] or CBAM [45], MLCA offers the advantage of considering both global dependencies and local feature sensitivity. This is particularly important for accurately locating small-sized targets. Moreover, the design of MLCA emphasizes computational efficiency. Despite introducing a complex context fusion strategy, its implementation ensures that it does not significantly increase the network’s computational burden, making it well-suited for integration into resource-constrained edge devices. In performance evaluations, MLCA demonstrates significant advantages. Experimental results show that models incorporating MLCA achieve a notable percentage increase in mAP0.5 compared to the original models while maintaining low computational complexity.
Overall, MLCA is an efficient and practical attention module ideal for target detection tasks in emergency medical scenarios requiring high accuracy and real-time processing.

2.2.3. STD-FPN

In recent years, ShuffleNetV2 [46] has emerged as a leading network for lightweight feature extraction, incorporating innovative channel split and channel shuffle designs that significantly reduce computational load and the number of parameters while maintaining high accuracy. Compared to its predecessor, ShuffleNetV1 [47], ShuffleNetV2 demonstrates greater efficiency and scalability, with substantial innovations and improvements in its structural design and complexity management. The network is divided into three main stages, each containing multiple ShuffleV2Blocks. Data first passes through an initial convolution layer and a max pooling layer, progressively moving through the stages, and ultimately outputs feature maps of three different dimensions. The entire network optimizes feature extraction performance by minimizing memory access.
As shown in Figure 6a, the FPN structure of Yolo-FastestV2 utilizes the feature map from the third ShuffleV2Block in ShuffleNetV2, combined with 1 × 1 convolution to predict large objects. These feature maps are then upsampled and fused with the feature maps from the second ShuffleV2Block to predict smaller objects. However, Yolo-FastestV2’s FPN only uses two layers of shallow feature maps, limiting the acquisition of rich positional information and affecting the semantic information extraction and precise localization of small objects. Considering that AED devices are typically placed within 50 cm to 75 cm from the patient, and the wristband is a small-scale target, we propose an improved FPN structure named STD-FPN (see Figure 6b), which effectively merges shallow and deep feature maps from ShuffleV2Block, focusing on small-object detection. Each output from the ShuffleV2Block is defined as S i , i [ 1 , 3 ] . After processing through the MLCA module, it becomes C i . First, C 1 is globally pooled to reduce its size by a factor of four to get C 1 , which is then concatenated with C 3 . This concatenated feature undergoes Convolution-BatchNormalization-ReLU(CBR), forming the input for the first detection head. The second detection head, designed for small objects, processes C 2 through CBR operations to match the channel count of C 1 and then upsamples C 2 along all dimensions using a specified scaling factor. C 2 is element-wise added to C 1 , followed by the CBR operation.
After each feature-fusion step, a 1 × 1 convolution is applied. During the entire model training process, convolution helps extract effective features from previous feature maps and reduces the impact of noise. By using additive feature fusion, shallow and deep features are fully integrated, producing fused feature maps rich in object positional information, thus enhancing the original model’s localization capability.

2.3. Depth Measurement Method

Image processing often involves four coordinate systems: the world coordinate system, the camera coordinate system, the image coordinate system, and the pixel coordinate system. Typically, the transformation process starts from the world coordinate system, passes through the camera coordinate system and the image coordinate system, and finally reaches the pixel coordinate system [48]. Assume world coordinate point P w = x w , y w , z w T , camera coordinate point P c = x c , y c , z c T , image coordinate point m = ( x p , y p , 1 ) T , and pixel coordinate point P i x = ( u , v , 1 ) T . The transformation from the world coordinate point P w = x w , y w , z w T to the camera coordinate point P c = x c , y c , z c T is given by Formula (1).
x c y c z c 1 = R T 0 1 x w y w z w 1
In this formula, the orthogonal rotation matrix R = r 11 r 12 r 13 r 21 r 22 r 23 r 31 r 32 r 33 and the translation matrix T = t x t y t z T . Assume the center O of the projective transformation as the origin of the camera coordinate system, and the distance from this point to the imaging plane is the focal length f. According to the principle of similar triangles, Formula (2) can be obtained to transform from the camera coordinate point P c = x c , y c , z c T to the image coordinate point m = ( x p , y p , 1 ) T :
z c m = f 0 0 0 0 f 0 0 0 0 1 0 P c
Assume that the length and width of a pixel are d x , d y , respectively. Pixel coordinate point P i x = ( u , v , 1 ) T , then
u v 1 = 1 / d x 0 0 0 1 / d y 0 0 0 1 x p y p 1
In summary, combining Formulas (1)–(3), the transformation matrix K from the camera coordinate point P c = x c , y c , z c T to the pixel coordinate point P i x = ( u , v , 1 ) T can be obtained:
K = 1 / d x 0 0 0 1 / d y 0 0 0 1 f 0 x 0 0 f y 0 0 0 1 = f x 0 u 0 0 f y v 0 0 0 1
Among them, f x = f / d x and f y = f / d y are called the scale factors of the camera in the u-axis and v-axis directions:
z c u v 1 = K · R T 0 1 x w y w z w 1
Equation (5) represents the transformation from world coordinates to pixel coordinates. The above explanation covers the principles of camera imaging. Building on this foundation, we propose a new depth measurement method.
In conventional monocular camera distance measurement, directly measuring depth is challenging because it lacks stereoscopic information. To address this issue, this study employs an innovative approach, as shown in Figure 7, using a fixed-length marker wristband as a depth-calibration tool. By applying the principles of camera imaging, we can accurately calculate the distance between the camera and the marker wristband. Ultimately, by comparing the known length of the marker with the image captured by the camera, we achieve precise mapping calculations of real-world compression depth.
During the execution of the program, it is necessary to read the detection frame displacement, denoted by B 0 , at the current window resolution. The resolution conversion function f converts the detection frame displacement B 0 at the current window resolution to the pixel height B p at the ideal camera resolution, i.e.:
B p = f B 0
From Figure 6, B p is the vertical displacement of the marker captured by the camera, L is the focal length of the camera, L is the horizontal distance between the marker and the camera, R is half of the vertical displacement of the marker, H is the vertical displacement of the marker, and the following equation is obtained:
tan ( a ) = A p 2 L tan ( b ) = B p 2 L tan ( b ) = R L
The following is obtained from Equation (7):
tan ( a ) tan ( b ) = A p B p
Substituting tan ( b ) = R L from Equation (8) yields:
L = R × A p B p × tan ( a )
In summary:
H = 2 tan ( b ) × L = B p × L L
H is the realistic depth of compression that we seek.

2.4. Edge Device Algorithm Optimization

Given the limited computational power of existing edge devices, a special optimization method is needed to enhance the timeliness of CPR action recognition, which requires high accuracy and real-time processing. As illustrated in Figure 8, the deep-learning algorithm model is first converted into weights compatible with the corresponding NPU. During this conversion process, MMSE algorithms and lossless pruning are employed to obtain more lightweight weights. Next, a multithreading scheme is designed. Two threads on the CPU handle the algorithm’s pre-processing and post-processing, while one thread on the NPU handles the inference phase. The RGA method is applied to image processing during both the pre- and post-processing stages. Finally, NEON instructions are used during the algorithm’s compilation phase.
By using the MMSE algorithm for weight quantization and applying RGA and NEON acceleration, the algorithm’s size is reduced, computational overhead is minimized, and inference speed is increased. Lossless pruning during model quantization effectively prevents accuracy degradation. The multithreading design enables asynchronous processing between the CPU and NPU, significantly improving the model’s performance on edge devices.

3. Experiments and Results

3.1. Datasets

The dataset used in this study consists of video frames of CPR actions captured by student volunteers from Nanjing University of Posts and Telecommunications in various scenarios. These videos encompass different indoor and outdoor environments and lighting conditions. The environments include objects with colors similar to the marker wristbands. The volunteer group comprised students with and without first aid knowledge to ensure data diversity and broad applicability. Videos are selected based on clarity, shooting angle, and visibility of the marker wristbands. Videos with low image quality due to blurriness, overexposure, or unclear markers are excluded to maintain high quality and consistency in the dataset. The original dataset contains 1479 images, which are augmented to 8874 images. To ensure the model’s robustness and generalization ability, the dataset is divided into training, testing, and validation sets in an 8:1:1 ratio, comprising 7081, 897, and 896 images, respectively. The experiments focus on a single object type, the marker wristband, ensuring the model specifically targeted this object.

3.2. Experimental Setting and Evaluation Index

The marker wristband used in the experiments is 33.40 cm long, 3.80 cm wide, and fluorescent green. The experiments are conducted on an NVIDIA GEFORCE RTX 6000 GPU with 24 GB of memory to ensure efficient training. The model is trained without using pre-trained weights. Image processing and data-augmentation techniques are employed to reduce overfitting and improve recognition accuracy. The training parameters are set as follows: image resolution of 352 × 352 , 300 epochs, a learning rate of 0.001, and a batch size of 512. To ensure annotation accuracy and consistency, professionally trained volunteers use the LabelMe tool to annotate images, accurately marking each wristband within the boundary boxes to avoid unnecessary noise. During the training phase, we implement basic image quality control measures, including checking image clarity, brightness, and contrast. All images are cropped and scaled to a uniform 352 × 352 pixels to standardize the input data format. To enhance the model’s generalization ability and reduce overfitting, various data-augmentation techniques are applied. These included random rotation, horizontal and vertical flipping, random scaling, and slight color transformations (such as hue and saturation adjustments) to simulate different lighting conditions. These steps ensure the dataset’s quality, making the model more robust and reliable. The training process of the dataset is illustrated in Figure 9a, showing batch 0, while Figure 9b shows the testing of batch 0 using the dataset labels.
True positives (TP) refer to the number of instances where the actual condition is “yes” and the model also predicts “yes”. True negatives (TN) refer to the number of instances where the actual condition is “no” and the model correctly predicts “no”. False positives (FP) occur when the model incorrectly predicts “yes” for an actual “no” scenario, leading to false alarms. Conversely, False negatives (FN) occur when the model incorrectly predicts “no” for an actual “yes” scenario [49]. Precision and recall are calculated using Equations (11) and (12), respectively [50,51].
Precision P = TP TP + FP
Recall ( R ) = TP TP + FN

3.3. OpenPose for CPR Recognition

During the process of performing CPR with an AED device, some errors may be difficult to detect through direct observation by a physician. Therefore, it is necessary to use OpenPose to draw skeletal points. As shown in Figure 10, three common incorrect CPR scenarios are identified: obscured arm movements due to dark clothing, kneeling on one knee, and non-vertical compressions. In the first scenario, dark clothing reduces the contrast with the background, making it difficult to clearly distinguish the edges of the arms. This issue is exacerbated in low-light conditions, making arm movements even more blurred and harder to identify. In the second scenario, kneeling on one knee causes the rescuer’s body to be unstable, affecting the stability and effectiveness of the compressions. In the third scenario, non-vertical compressions cause the force to be dispersed, preventing it from being effectively concentrated on the patient’s chest, thereby affecting the depth and effectiveness of the compressions. These issues can all be addressed using OpenPose. After posture recognition, physicians can remotely provide voice reminders, allowing for the immediate correction of these otherwise difficult-to-detect incorrect postures.

3.4. Ablation Experiment

CPR-Detection is an improved object-detection model designed to optimize recognition accuracy and speed. In medical CPR scenarios, due to the limited computational power of edge devices, smaller image inputs (352 × 352 pixels) are typically used to achieve the highest possible mAP0.5. To assess the specific impact of the new method on mAP0.5, ablation experiments are conducted on Yolo-FastestV2. The study independently and jointly tests the effects of the PConv, MLCA, and STD-FPN modules on model performance. The results, as shown in Table 1, clearly demonstrate that these modules, whether applied alone or in combination, enhance the model’s mAP0.5: introducing PConv improves mAP0.5 by 0.44%, optimizing the extraction and representation of positional features [42]. Using MLCA increases mAP0.5 by 0.44%, effectively enhancing the model’s ability to process channel-level and spatial-level information [43]. Applying the STD-FPN structure results in a 0.11% mAP0.5 improvement, optimizing feature fusion and positional enhancement. Combining PConv and MLCA boosts mAP0.5 to 96.87%, achieving a 0.83% increase. The combination of PConv and STD-FPN raises mAP0.5 by 0.95%, better integrating local and global features. The combined use of all three modules increases mAP0.5 by 1.00%, slightly increasing FLOPs but reducing the number of parameters.
These improvements significantly enhance the model’s ability to recognize small targets in CPR scenarios, ensuring higher accuracy while maintaining real-time detection, and demonstrating the superiority of the CPR-Detection model. The combined use of the three modules fully leverages their unique advantages, enabling the model to adapt flexibly to different input sizes and application scenarios, providing an ideal object-detection solution for medical emergency scenarios that demand high accuracy and speed.

3.5. Compared with State-of-the-Art Models

To evaluate the impact of the proposed method on the model’s feature-extraction capabilities, the CPR-Detection model is compared with six state-of-the-art lightweight object-detection models, including FastestDet and Yolo-FastestV2 based on the YoloV5 architecture, as well as other official lightweight models. This comparison aims to demonstrate the effectiveness of the new method in improving model performance. Compared to Yolo-FastestV2, the improved CPR-Detection model significantly enhance feature-extraction capabilities. Table 2 presents a quantitative comparison of these models in terms of FLOPs, parameter count, mAP0.5, and mAP0.5:0.95.
As shown in Table 2, the comparison results of CPR-Detection with other models in terms of mAP0.5 are as follows: CPR-Detection’s mAP0.5 improved by 1.02% compared to YoloV7-Tiny; by 6.84% compared to NanoDet-m; by 11.46% compared to FastestDet; and by 1.00% compared to Yolo-FastestV2. Although CPR-Detection’s mAP0.5 is slightly lower than YoloV3-Tiny and YoloV5-Lite (1.45% and 1.16% lower, respectively), it has fewer parameters and lower computational costs compared to these models. This balance strikes an optimal point between speed and accuracy, making it an ideal choice for medical emergency scenarios with limited computational resources.

3.6. Measurement Results

One of the key parameters in CPR is the number and frequency of compressions. In this study, we identify each effective compression by analyzing the peaks and troughs of hand movements in the video, with each complete peak–trough cycle representing one compression. The frequency is calculated based on the number of effective compressions occurring per unit of time. Extensive testing shows that the accuracy of compression count and frequency exceeds 98%, with depth accuracy over 90% and errors generally within 1 cm. The errors in count and frequency are mainly due to initial fluctuations of the marker, while depth errors were often caused by inconsistencies in marker performance under different experimental conditions, such as camera angle and lighting changes. The video analysis-based method for measuring CPR compression count, frequency, and depth proposed in this study is highly accurate and practical. It is crucial for guiding first responders in performing standardized CPR, significantly enhancing the effectiveness of emergency care. Although there are some errors, further optimization of the algorithm and improvements in data-collection methods are expected to enhance measurement accuracy.
Figure 11a shows the depth variance distribution for 100 compressions. Most data points have depth errors within ±1 cm, meeting CPR operational standards and demonstrating the high accuracy of the measurement system. However, a few data points exceed a 1 cm depth error, likely due to changes in experimental conditions, such as slight adjustments in camera angle or lighting intensity, which can affect the visual recognition accuracy of the wristband. Figure 11b illustrates the accuracy for each of the 100 measurement tests conducted. A 90% accuracy threshold is set to evaluate the system’s performance. Results indicate that the vast majority of measurements exceed this threshold, confirming the system’s high reliability in most cases. However, there are a few instances where accuracy falls below 90%, highlighting potential weaknesses in the system, such as improper actions, insufficient device calibration, or environmental interference. Future work will focus on diagnosing and addressing these issues to improve the overall performance and reliability of the system.

3.7. AED Application for CPR

When using the AED edge device, the user should wear the wristband on their arm and prepare for CPR. The usage process is as follows. After activating the AED edge device, the data-collection unit starts automatically. Once the intelligent emergency function is initiated, the device automatically activates the AI recognition module, capturing real-time images of the emergency scene and collecting data for AI image recognition. During CPR, the AI recognition module uses multiple algorithms to assess whether the procedure meets standards. The voice playback and video display modules provide corrective prompts based on AI processing feedback. The storage module continuously records device operation, emergency events, detection, and AI recognition feedback. Medical emergency personnel can view real-time audio-visual information, location data, AED data, and AI recognition feedback sent by the intelligent module via the emergency platform server. The server also transmits this data back to the device. The intelligent module connects to the emergency platform server through the communication module, retrieves the server’s audio-visual data, and plays it through the voice playback and video display modules. As illustrated in Figure 12, our algorithm’s effectiveness in practical applications is demonstrated. We capture two frames from the AED edge device video after activation, showing the displayed activation time, compression count, frequency, and depth. Additionally, we used OpenPose to visualize skeletal points, capturing the arm’s local motion trajectory during compressions [16]. This helps doctors assess the correctness of the posture via the emergency platform server.
As shown in Figure 13, after optimizing the algorithm on the edge device, the initial frame rate of 8 FPS was significantly improved. By applying quantization methods, the frame rate increased by 5 FPS. Pruning techniques add another 2 FPS, and the asynchronous method contributed an additional 7 FPS. Further enhancements are achieved with RGA and NEON, which improve the frame rate by 1 FPS and 2 FPS, respectively. Overall, the frame rate increases from 8 FPS to 25 FPS, validating the feasibility of these optimization methods.

4. Discussion

The application of artificial intelligence in CPR action standardization addresses the limitations of traditional methods. Traditional CPR training relies on classroom simulations, which fail to replicate the stress of actual cardiac arrest events, while VR and AR technologies, though educational, lack real-time application [6,7,8]. Unlike mainstream techniques that have not fully embraced AI, DLCAS pioneers real-time AI interventions on AEDs, offering immediate feedback and corrective actions to improve CPR accuracy and survival rates. By utilizing advanced deep-learning methods like OpenPose, the CPR-Detection algorithm, and edge device optimization, DLCAS achieves high precision in posture detection and compression metrics. Specifically, it boasts a mean average precision (mAP) of 97.04% and impressive accuracy in depth and count measurements. Furthermore, DLCAS is optimized for edge devices, enhancing processing speed from 8 to 25 fps to meet emergency demands.
In the third part of this study, we evaluate the effectiveness of the DLCAS method through a series of experiments. The figures and quantitative performance metrics of the experimental results highlight the superiority of our approach. Qualitatively, Figure 10, Figure 11 and Figure 12 demonstrate significant improvements in our method’s ability to capture arm movements and compression depth accuracy. Additionally, Table 1 and Table 2 present comprehensive quantitative results across these datasets, consistently indicating that our proposed CPR-Detection algorithm outperforms existing models in terms of accuracy and efficiency. Section 3.7 provides a detailed account of how we optimize the algorithm for edge devices to ensure high performance in practical applications.
Our method demonstrates exceptional performance in both quantitative and qualitative experiments, owing to several key innovations. We employ OpenPose for accurate and rapid recognition of human body poses, facilitating physicians’ assessment of posture accuracy via emergency platform servers. In our CPR-Detection approach, we choose PConv over DWSConv to ensure higher efficiency without compromising performance, effectively meeting real-time processing demands. The incorporation of MLCA modules enhances our model’s ability to manage channel-level and spatial-level information. STD-FPN comprehensively integrates shallow and deep features, generating fusion-feature maps rich in positional details that enhance the model’s localization capabilities. Additionally, our depth measurement method guarantees precise mapping of real-world compression depths, while edge-device algorithm optimization ensures efficient performance on edge devices.
The proposed method, while achieving promising results, still has some issues that need to be addressed. Given the strict requirements for data accuracy in medical applications, it is crucial to enhance the accuracy of our model and the stability of the detection boxes in our target-detection algorithm [53]. Additionally, our method relies on the use of marked wristbands, which can consume valuable time in emergency scenarios. In subsequent work, components such as infrared rangefinders will be added to enable distance measurement without the use of a wristband [54]. Reducing the time required for this step would significantly improve the safety of the person being rescued [55].
To address these challenges, future research will focus on several key areas: (1) adopting advanced techniques like dynamic parameter regularization to improve the accuracy and stability of detection boxes by dynamically adjusting regularization parameters throughout the training process [56]; (2) developing markerless motion capture and advanced image-processing algorithms such as infrared rangefinders to eliminate the need for marked wristbands, thereby reducing setup time and increasing the efficiency of emergency interventions [57]; (3) enhancing neural-network interpretability by utilizing techniques such as heat mapping, which will help clinicians better understand and trust AI-assisted decisions [58].

5. Conclusions

In this paper, we aim to address the issue related to the lack of standardized cardiopulmonary resuscitation (CPR) actions in automated external defibrillators (AEDs). We propose the deep-learning-based CPR action standardization (DLCAS) method. The first part of DLCAS utilizes OpenPose to identify skeletal points, enabling remote doctors to correct rescuers’ posture through networked AED devices. In the second part of DLCAS, we design the CPR-Detection network. This network uses partial convolution (PConv) to enhance feature representation by focusing on critical spatial information. Additionally, we employ mixed local channel attention (MLCA) on our custom small-target detection-feature pyramid network (STD-FPN). MLCA combines local and global contextual information, improving detection accuracy and efficiency. STD-FPN effectively merges shallow and deep-image features, enhancing the model’s localization capability. Based on CPR-Detection, we introduce a new depth algorithm to measure the rescuers’ compression depth, count and frequency. In the third part of DLCAS, we apply computational optimization methods, including multi-threaded CPU and NPU asynchronous design, RGA, and NEON acceleration, significantly boosting real-time processing efficiency. Extensive experiments on our custom dataset have shown that our method effectively addresses the issue of AED devices’ inability to standardize CPR actions. Furthermore, our method improves the stability and speed of edge devices, validating the applicability of the DLCAS method in current medical scenarios through performance testing.

Author Contributions

Conceptualization, Y.L. and M.Y.; methodology, Y.L. and M.Y.; software, Y.L. and M.Y.; validation, Y.L., M.Y. and W.W.; formal analysis, Y.L., M.Y., W.W. and J.L.; investigation, Y.L., M.Y. and J.L.; resources, Y.L., M.Y. and W.W.; data curation, Y.L., M.Y. and W.W.; writing—original draft preparation, Y.L.; writing—review and editing, M.Y.; visualization, M.Y.; supervision, Y.L.; project administration, S.L. and Y.J.; funding acquisition, S.L. and Y.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key R&D Program of China (2023YFB2904000, 2023YFB2904004), the Jiangsu Key Development Planning Project (BE2023004-2), the Natural Science Foundation of Jiangsu Province (Higher Education Institutions) (20KJA520001), the 14th Five-Year Plan project of Equipment Development Department (315107402), the Jiangsu Hongxin Information Technology Co., Ltd. Project (JSSGS2301022EGN00), the Postgraduate Research & Practice Innovation Program of the Jiangsu Province (KYCX24_1204) and the Future Network Scientific Research Fund Project (No. FNSRFP-2021-YB-15).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Due to privacy protection for student volunteers, the data supporting the reported results will be made available upon request after acceptance and following privacy protection inquiries.

Acknowledgments

We would like to express our gratitude to Yuwell for their financial support and the provision of equipment used in our experiments. Their generous contributions were instrumental in the successful completion of this research.

Conflicts of Interest

Yongyuan Li is temporarily affiliated with Jiangsu Tuoyou Information Intelligent Technology Research Institute Co., Ltd. to perform research activities under Yimu Ji’s supervision, the affiliation does not provide any financial support or funding.

References

  1. Berdowski, J.; Berg, R.A.; Tijssen, J.G.; Koster, R.W. Global incidences of out-of-hospital cardiac arrest and survival rates: Systematic review of 67 prospective studies. Resuscitation 2010, 81, 1479–1487. [Google Scholar] [CrossRef] [PubMed]
  2. Yan, S.; Gan, Y.; Jiang, N.; Wang, R.; Chen, Y.; Luo, Z.; Zong, Q.; Chen, S.; Lv, C. The global survival rate among adult out-of-hospital cardiac arrest patients who received cardiopulmonary resuscitation: A systematic review and meta-analysis. Crit. Care 2020, 24, 61. [Google Scholar] [CrossRef] [PubMed]
  3. Song, J.; Guo, W.; Lu, X.; Kang, X.; Song, Y.; Gong, D. The effect of bystander cardiopulmonary resuscitation on the survival of out-of-hospital cardiac arrests: A systematic review and meta-analysis. Scand. J. Trauma Resusc. Emerg. Med. 2018, 26, 86. [Google Scholar] [CrossRef] [PubMed]
  4. Gräsner, J.T.; Wnent, J.; Herlitz, J.; Perkins, G.D.; Lefering, R.; Tjelmeland, I.; Koster, R.W.; Masterson, S.; Rossell-Ortiz, F.; Maurer, H.; et al. Survival after out-of-hospital cardiac arrest in Europe—Results of the EuReCa TWO study. Resuscitation 2020, 148, 218–226. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, S.; Yu, Q.; Wang, S.; Yang, D.; Su, L.; Zhao, X.; Kuang, H.; Zhang, P.; Zhai, P.; Zhang, L. CPR-Coach: Recognizing Composite Error Actions based on Single-class Training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  6. Rodríguez-Matesanz, M.; Guzmán-García, C.; Oropesa, I.; Rubio-Bolivar, J.; Quintana-Díaz, M.; Sánchez-González, P. A New Immersive virtual reality station for cardiopulmonary resuscitation objective structured clinical exam evaluation. Sensors 2022, 22, 4913. [Google Scholar] [CrossRef] [PubMed]
  7. Krasteva, V.; Didon, J.P.; Ménétré, S.; Jekova, I. Deep Learning Strategy for Sliding ECG Analysis during Cardiopulmonary Resuscitation: Influence of the Hands-Off Time on Accuracy. Sensors 2023, 23, 4500. [Google Scholar] [CrossRef]
  8. Xie, J.; Wu, Q. Design and Evaluation of CPR Emergency Equipment for Non-Professionals. Sensors 2023, 23, 5948. [Google Scholar] [CrossRef]
  9. Tang, X.; Wang, Y.; Ma, H.; Wang, A.; Zhou, Y.; Li, S.; Pei, R.; Cui, H.; Peng, Y.; Piao, M. Detection and Evaluation for High-Quality Cardiopulmonary Resuscitation Based on a Three-Dimensional Motion Capture System: A Feasibility Study. Sensors 2024, 24, 2154. [Google Scholar] [CrossRef]
  10. Daudre-Vignier, C.; Bates, D.G.; Scott, T.E.; Hardman, J.G.; Laviola, M. Evaluating current guidelines for cardiopulmonary resuscitation using an integrated computational model of the cardiopulmonary system. Resuscitation 2023, 186, 109758. [Google Scholar] [CrossRef]
  11. Crespo-Diaz, R.; Wolfson, J.; Yannopoulos, D.; Bartos, J.A. Machine learning identifies higher survival profile in extracorporeal cardiopulmonary resuscitation. Crit. Care Med. 2024, 52, 1065–1076. [Google Scholar] [CrossRef]
  12. Semeraro, F.; Schnaubelt, S.; Hansen, C.M.; Bignami, E.G.; Piazza, O.; Monsieurs, K.G. Cardiac arrest and cardiopulmonary resuscitation in the next decade: Predicting and shaping the impact of technological innovations. Resuscitation 2024, 200, 110250. [Google Scholar] [CrossRef] [PubMed]
  13. Shrimpton, A.J.; Brown, V.; Vassallo, J.; Nolan, J.; Soar, J.; Hamilton, F.; Cook, T.; Bzdek, B.R.; Reid, J.P.; Makepeace, C.; et al. A quantitative evaluation of aerosol generation during cardiopulmonary resuscitation. Anaesthesia 2024, 79, 156–167. [Google Scholar] [CrossRef] [PubMed]
  14. Kao, C.L.; Tsou, J.Y.; Hong, M.Y.; Chang, C.J.; Tu, Y.F.; Huang, S.P.; Su, F.C.; Chi, C.H. A novel CPR-assist device vs. established chest compression techniques in infant CPR: A manikin study. Am. J. Emerg. Med. 2024, 77, 81–86. [Google Scholar] [CrossRef] [PubMed]
  15. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
  16. Cao, Z.; Martinez, G.H.; Simon, T.; Wei, S.; Sheikh, Y.A. OpenPose: Realtime Multi-Person 2D Pose Estimation using Part Affinity Fields. IEEE Trans. Pattern Anal. Mach. Intell. 2019. [Google Scholar] [CrossRef] [PubMed]
  17. Simon, T.; Joo, H.; Matthews, I.; Sheikh, Y. Hand keypoint detection in single images using multiview bootstrapping. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1145–1153. [Google Scholar]
  18. Joo, H.; Liu, H.; Tan, L.; Gui, L.; Nabbe, B.; Matthews, I.; Kanade, T.; Nobuhara, S.; Sheikh, Y. Panoptic studio: A massively multiview system for social motion capture. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 3334–3342. [Google Scholar]
  19. Gholami, A.; Kwon, K.; Wu, B.; Tai, Z.; Yue, X.; Jin, P.; Zhao, S.; Keutzer, K. SqueezeNext: Hardware-Aware Neural Network Design. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  20. Howard, A.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  21. Zeng, T.; Li, S.; Song, Q.; Zhong, F.; Wei, X. Lightweight tomato real-time detection method based on improved YOLO and mobile deployment. Comput. Electron. Agric. 2023, 205, 107625. [Google Scholar] [CrossRef]
  22. Cong, S.; Zhou, Y. A review of convolutional neural network architectures and their optimizations. Artif. Intell. Rev. 2023, 56, 1905–1969. [Google Scholar] [CrossRef]
  23. Wei, Y.; Zhao, L.; Zheng, W.; Zhu, Z.; Zhou, J.; Lu, J. SurroundOcc: Multi-Camera 3D Occupancy Prediction for Autonomous Driving. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023. [Google Scholar]
  24. Wu, D.; Liao, M.W.; Zhang, W.T.; Wang, X.G.; Bai, X.; Cheng, W.Q.; Liu, W.Y. Correction to: YOLOP: You Only Look Once for Panoptic Driving Perception. Mach. Intell. Res. 2023, 20, 952. [Google Scholar] [CrossRef]
  25. Xu, M.; Wang, X.; Zhang, S.; Wan, R.; Zhao, F. Detection algorithm of aerial vehicle target based on improved YOLOv3. J. Phys. Conf. Ser. 2022, 2284, 012022. [Google Scholar] [CrossRef]
  26. Jamiya, S.S.; Rani, P.E. An Efficient Method for Moving Vehicle Detection in Real-Time Video Surveillance. In Proceedings of the Advances in Smart System Technologies, Osijek, Croatia, 14–16 October 2020. [Google Scholar]
  27. Wu, S.; Zhang, L. Using Popular Object Detection Methods for Real Time Forest Fire Detection. In Proceedings of the 2018 11th International Symposium on Computational Intelligence and Design (ISCID), Hangzhou, China, 8–9 December 2018. [Google Scholar]
  28. Mishra, S.; Jabin, S. Anomaly detection in surveillance videos using deep autoencoder. Int. J. Inf. Technol. 2024, 16, 1111–1122. [Google Scholar] [CrossRef]
  29. Ali, M.M. Real-time video anomaly detection for smart surveillance. IET Image Process. 2023, 17, 1375–1388. [Google Scholar] [CrossRef]
  30. Sun, S.; Xu, Z. Large kernel convolution YOLO for ship detection in surveillance video. Math. Biosci. Eng. 2023, 20, 15018–15043. [Google Scholar] [CrossRef] [PubMed]
  31. Zhang, X.; Xuan, C.; Xue, J.; Chen, B.; Ma, Y. LSR-YOLO: A High-Precision, Lightweight Model for Sheep Face Recognition on the Mobile End. Animals 2023, 13, 1824. [Google Scholar] [CrossRef] [PubMed]
  32. Yu, F.; Zhang, G.; Zhao, F.; Wang, X.; Liu, H.; Lin, P.; Chen, Y. Improved YOLO-v5 model for boosting face mask recognition accuracy on heterogeneous IoT computing platforms. Internet Things 2023, 23, 100881. [Google Scholar] [CrossRef]
  33. Sun, F. Face Recognition Analysis Based on the YOLO Algorithm. In Proceedings of the 4th International Conference on Computing and Data Science (CONF-CDS 2022), Macau, China, 16 July 2022. [Google Scholar]
  34. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  35. Wong, A.; Famuori, M.; Shafiee, M.J.; Li, F.; Chwyl, B.; Chung, J. YOLO Nano: A Highly Compact You Only Look Once Convolutional Neural Network for Object Detection. In Proceedings of the 2019 Fifth Workshop on Energy Efficient Machine Learning and Cognitive Computing-NeurIPS Edition (EMC2-NIPS), Vancouver, BC, Canada, 13 December 2019. [Google Scholar]
  36. Hu, L.; Li, Y. Micro-YOLO: Exploring Efficient Methods to Compress CNN based Object Detection Model. In Proceedings of the International Conference on Agents and Artificial Intelligence, Online, 4–6 February 2021. [Google Scholar]
  37. Lyu, R. Nanodet-Plus: Super Fast and High Accuracy Lightweight Anchor-Free Object Detection Model. 2021. Available online: https://github.com/RangiLyu/nanodet (accessed on 1 April 2024).
  38. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  39. Jocher, G.; Nishimura, K.; Mineeva, T.; Vilarino, R. yolov5. Code Repos. 2020, 9. [Google Scholar]
  40. Dog-Qiuqiu, A. Dog-Qiuqiu/Yolo-Fastest: Yolo-Fastest-v1. 1.0 2021. Available online: https://github.com/dog-qiuqiu/Yolo-FastestV2 (accessed on 30 December 2023).
  41. Ma, X. Fastestdet: Ultra Lightweight Anchor-Free Realtime Object Detection Algorithm. 2022. Available online: https://github.com/dog-qiuqiu/FastestDet (accessed on 12 January 2024).
  42. Chen, J.; Kao, S.H.; He, H.; Zhuo, W.; Wen, S.; Lee, C.H.; Chan, S.H.G. Run, Don’t walk: Chasing higher FLOPS for faster neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 12021–12031. [Google Scholar]
  43. Wan, D.; Lu, R.; Shen, S.; Xu, T.; Lang, X.; Ren, Z. Mixed local channel attention for object detection. Eng. Appl. Artif. Intell. 2023, 123, 106442. [Google Scholar] [CrossRef]
  44. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  45. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  46. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
  47. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar]
  48. You, Z.; Luan, Z.; Wei, X. General lens distortion model expressed by image pixel coordinate. Opt. Tech. 2015, 41, 265–269. [Google Scholar]
  49. Dewi, C.; Chen, R.C. Random forest and support vector machine on features selection for regression analysis. Int. J. Innov. Comput. Inf. Control 2019, 15, 2027–2037. [Google Scholar]
  50. Yuan, Y.; Xiong, Z.; Wang, Q. An incremental framework for video-based traffic sign detection, tracking, and recognition. IEEE Trans. Intell. Transp. Syst. 2016, 18, 1918–1929. [Google Scholar] [CrossRef]
  51. Dewi, C.; Chen, R.C.; Tai, S.K. Evaluation of robust spatial pyramid pooling based on convolutional neural network for traffic sign recognition system. Electronics 2020, 9, 889. [Google Scholar] [CrossRef]
  52. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7464–7475. [Google Scholar]
  53. Ahmed, S.F.; Alam, M.S.B.; Afrin, S.; Rafa, S.J.; Rafa, N.; Gandomi, A.H. Insights into Internet of Medical Things (IoMT): Data fusion, security issues and potential solutions. Inf. Fusion 2024, 102, 102060. [Google Scholar] [CrossRef]
  54. Kim, D.; Kang, J.; Na, K.S. Development of smart glasses monitoring viewing distance using an infrared distance measurement sensor. Investig. Ophthalmol. Vis. Sci. 2024, 65, 2754. [Google Scholar]
  55. Choi, Y.; Park, J.H.; Jeong, J.; Kim, Y.J.; Song, K.J.; Shin, S.D. Extracorporeal cardiopulmonary resuscitation for adult out-of-hospital cardiac arrest patients: Time-dependent propensity score-sequential matching analysis from a nationwide population-based registry. Crit. Care 2023, 27, 87. [Google Scholar] [CrossRef] [PubMed]
  56. Pu, J.C.; Chen, Y. Data-driven forward-inverse problems for Yajima–Oikawa system using deep learning with parameter regularization. Commun. Nonlinear Sci. Numer. Simul. 2023, 118, 107051. [Google Scholar] [CrossRef]
  57. Tian, Z.; Weng, D.; Fang, H.; Shen, T.; Zhang, W. Robust facial marker tracking based on a synthetic analysis of optical flows and the YOLO network. Vis. Comput. 2024, 40, 2471–2489. [Google Scholar] [CrossRef]
  58. Wang, Z.; Zhou, Y.; Han, M.; Guo, Y. Interpreting convolutional neural network by joint evaluation of multiple feature maps and an improved NSGA-II algorithm. Expert Syst. Appl. 2024, 255, 124489. [Google Scholar] [CrossRef]
Figure 1. Overall working flowchart.
Figure 1. Overall working flowchart.
Sensors 24 04813 g001
Figure 2. Overall framework of OpenPose.
Figure 2. Overall framework of OpenPose.
Sensors 24 04813 g002
Figure 3. Overall framework of CPR-Detection.
Figure 3. Overall framework of CPR-Detection.
Sensors 24 04813 g003
Figure 4. (a) DWSConv. (b) PConv (The * in the figure indicates convolution calculation).
Figure 4. (a) DWSConv. (b) PConv (The * in the figure indicates convolution calculation).
Sensors 24 04813 g004
Figure 5. Mixed local channel attention (MLCA).
Figure 5. Mixed local channel attention (MLCA).
Sensors 24 04813 g005
Figure 6. (a) The FPN of Yolo-FastestV2. (b) Small-target detection-feature pyramid network.
Figure 6. (a) The FPN of Yolo-FastestV2. (b) Small-target detection-feature pyramid network.
Sensors 24 04813 g006
Figure 7. Depth ranging schematic.
Figure 7. Depth ranging schematic.
Sensors 24 04813 g007
Figure 8. Edge device computing optimization flow chart.
Figure 8. Edge device computing optimization flow chart.
Sensors 24 04813 g008
Figure 9. (a) Train batch 0 with datasets. (b) Test batch 0 labels with datasets.
Figure 9. (a) Train batch 0 with datasets. (b) Test batch 0 labels with datasets.
Sensors 24 04813 g009
Figure 10. Common incorrect posture images (including RGB, 2D Pose, Combined).
Figure 10. Common incorrect posture images (including RGB, 2D Pose, Combined).
Sensors 24 04813 g010
Figure 11. (a) Difference between actual depth and measured depth. (b) Measurement accuracy over time.
Figure 11. (a) Difference between actual depth and measured depth. (b) Measurement accuracy over time.
Sensors 24 04813 g011
Figure 12. Application scenario flowchart.
Figure 12. Application scenario flowchart.
Sensors 24 04813 g012
Figure 13. FPS improvement through various optimization steps.
Figure 13. FPS improvement through various optimization steps.
Sensors 24 04813 g013
Table 1. Validation of the Proposed Method on Yolo-FastestV2.
Table 1. Validation of the Proposed Method on Yolo-FastestV2.
IndexBASEPConvMLCASTD-FPNFLOPsParametersmAP0.5mAP0.5:0.95
1114.12 K238.50 K96.0472.55
2105.98 K213.30 K96.4873.89
3114.36 K238.52 K96.4875.09
4159.53 K229.38 K96.1571.12
5131.83 K204.18 K96.9975.16
6106.22 K213.32 K96.8776.57
7132.15 K204.20 K97.0475.13
Table 2. Model Comparison.
Table 2. Model Comparison.
MethodSizeFLOPsParametersmAP0.5mAP0.5:0.95
YoloV3-Tiny [34]352 × 3521.97 G8.66 M98.4980.42
YoloV7-Tiny [52]352 × 35213.2 G6.01 M96.0266.05
NanoDet-m [37]352 × 3520.87 G0.96 M90.2065.70
Yolo-FastestV2 [40]352 × 3520.11 G0.23 M96.0472.55
FastestDet [41]352 × 3520.13 G0.23 M85.5852.90
YoloV5-Lite [39]352 × 3523.70 G1.54 M98.2077.20
CPR-Detection352 × 3520.13 G0.20 M97.0475.13
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Yin, M.; Wu, W.; Lu, J.; Liu, S.; Ji, Y. A Deep-Learning-Based CPR Action Standardization Method. Sensors 2024, 24, 4813. https://doi.org/10.3390/s24154813

AMA Style

Li Y, Yin M, Wu W, Lu J, Liu S, Ji Y. A Deep-Learning-Based CPR Action Standardization Method. Sensors. 2024; 24(15):4813. https://doi.org/10.3390/s24154813

Chicago/Turabian Style

Li, Yongyuan, Mingjie Yin, Wenxiang Wu, Jiahuan Lu, Shangdong Liu, and Yimu Ji. 2024. "A Deep-Learning-Based CPR Action Standardization Method" Sensors 24, no. 15: 4813. https://doi.org/10.3390/s24154813

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop