Next Article in Journal
Accurate Computations with Generalized Green Matrices
Previous Article in Journal
A New Generalization of the Inverse Generalized Weibull Distribution with Different Methods of Estimation and Applications in Medicine and Engineering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLO-ABD: A Multi-Scale Detection Model for Pedestrian Anomaly Behavior Detection

1
School of Computer Science and Engineering, Sichuan University of Science and Engineering, Yibin 644000, China
2
Sichuan Big Data Visualization Analysis Technology Engineering Laboratory, Yibin 644000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Symmetry 2024, 16(8), 1003; https://doi.org/10.3390/sym16081003
Submission received: 4 July 2024 / Revised: 31 July 2024 / Accepted: 5 August 2024 / Published: 7 August 2024
(This article belongs to the Section Computer)

Abstract

:
Public safety and intelligent surveillance systems rely on anomaly detection for effective monitoring. In real-world pedestrian detection scenarios, Pedestrians often exhibit various symmetrical features such as body contours, facial features, posture, and clothing. However, the accuracy of pedestrian anomaly detection is affected by factors such as complex backgrounds, pedestrian obstruction, and small target sizes. To address these issues, this study introduced YOLO-ABD, a lightweight method for anomaly behavior detection that integrated small object detection and channel shuffling. This approach enhanced the YOLOv8n baseline model by integrating a small-object detection mechanism at the head and employing the symmetric GSConv convolutional module in the backbone network to improve perceptual capabilities. Furthermore, it incorporated the SimAM attention mechanism to mitigate complex background interference and thus enhance target detection performance. Evaluation on the IITB-Corridor dataset showed mAP50 and mAP50-95 scores of 89.3% and 60.6%, respectively. Generalization testing on the street-view-gdogo dataset further underscored the superiority of YOLO-ABD over advanced detection algorithms, demonstrating its effectiveness and generalization capabilities. With relatively fewer parameters, YOLO-ABD provided an excellent lightweight solution for pedestrian anomaly detection.

1. Introduction

Human anomaly detection involves real-time analysis and recognition of human behaviors in surveillance videos or images, aiming to identify deviations from typical behavior patterns [1,2]. Such anomalies may include activities like falling, climbing, running, and carrying hazardous items. This technology has widespread applications in public safety and intelligent monitoring systems, facilitating prompt identification of security threats and emergencies to effectively mitigate accidents [3,4,5,6].
With advancements in modeling, mainstream methods for anomaly behavior detection can be categorized into three types: frame reconstruction-based methods, frame prediction-based methods, and end-to-end anomaly score calculation methods [7,8,9,10,11,12,13,14]. Frame reconstruction methods leveraging deep learning often rely on Auto-Encoders [15]. Gong et al. [16] introduced the Memory-Augmented Autoencoder (MemAE), enhancing reconstruction by incorporating a memory module. MemAE selects the most relevant memory items for reconstruction using the encoded query, ensuring that testing results closely resemble normal samples. This approach aims to minimize the reconstruction error for normal behavior while maximizing it for anomalous behavior. Frame prediction-based methods involve feeding a video segment into a predictor to predict a frame; significant deviations between the predicted and current input frames indicate abnormal behavior [17,18]. In [19], a frame prediction network based on multipath ConvGRU was employed, capable of handling semantically informative objects and areas of various scales, capturing spatial-temporal dependencies in normal videos. To mitigate interference from background noise, a noise tolerance loss was introduced during training. However, samples in a static state may be misclassified as background due to the lack of differences between adjacent action frames, potentially leading to missed detections [20,21].
The YOLOv8 model adopts an end-to-end detection approach, enabling real-time object detection with high accuracy across diverse scenes and object types [22]. Its concise and lightweight architecture makes it suitable for deployment in resource-constrained environments. However, anomaly behavior detection presents challenges such as variations in spatial orientation, size, shape, overlapping backgrounds, and limited detection precision. To address these challenges, this paper proposes an enhanced YOLOv8-based method for anomaly behavior detection, termed YOLO-ABD, integrating small object detection and Group Shuffle Convolution (GSConv) [23]. The key contributions of this study are summarized as follows:
  • Introducing an end-to-end pedestrian anomaly detection method that utilizes the SimAM attention mechanism [24] to reduce background interference. Additionally, it incorporates a custom-designed small-object detection head to identify pedestrian anomalies at various scales.
  • Integrating the GSConv (Group Shuffle Convolution) module with a symmetrical structure enhances the model’s accuracy. Moreover, adopting shuffling strategies decreases the computational complexity, thereby achieving the goal of lightweighting the model.
  • The proposed method is trained and validated on public datasets for anomaly behavior detection. Generalization testing in traffic scene detection demonstrates significant performance improvements over existing methods.

2. Related Works

Classical methods for anomaly detection typically involve extracting features from videos or images to identify abnormal behaviors. Motion features are derived from various dimensions such as optical flow maps, heat maps, and depth maps. These motion features can be integrated with trajectory information or motion magnitude to fuse high-level semantic features with low-level details, achieving precise detection of abnormal behaviors [25,26,27,28]. For instance, Xie et al. [29] applied spatiotemporal representation learning to identify behaviors such as sleeping and using mobile phones by analyzing college students’ movement trajectories. Banerjee et al. [30] introduced a deep convolutional network architecture for detecting behavioral patterns of students and teachers in laboratory settings. Guan et al. [31] employed a 3D-CNN to extract features from optical flow and motion history images, using LSTM networks to capture spatiotemporal features for anomaly detection. These studies highlight the effectiveness of classical approaches in recognizing abnormal behaviors. However, these methods may struggle with detecting diverse abnormal behaviors across different scenarios and can be susceptible to interference from complex backgrounds, which poses challenges for multi-person anomaly detection in complex environments.
Modern anomaly detection methods also incorporate multi-object detection techniques to identify abnormal behaviors [32,33,34,35,36]. Researchers apply these approaches to detect abnormal behaviors in individuals or crowds within specific contexts. Object detection frameworks broadly fall into two categories: two-stage and single-stage approaches [37]. Two-stage algorithms, exemplified by Faster R-CNN, have been used for anomaly detection and classification, enhancing accuracy but requiring substantial computational resources and being less suitable for real-time applications, as demonstrated by Mansour et al. [38] and Hongchao et al. [39], who proposed enhanced models for identifying hazardous behaviors in factory workers.
In contrast, single-stage algorithms, exemplified by the YOLO series, are also employed for anomaly detection [40,41]. For example, in [42], the YOLOv3 model was used to detect unsafe behaviors at gas stations like smoking and mobile phone usage. Benjumea et al. [43] modified YOLOv5 to improve its ability to detect small objects in autonomous driving scenarios, while Xiao et al. [44] applied YOLOv5 to monitor safety in substations. The YOLO series simplifies the detection process, offering efficient computation and faster speeds, making it suitable for practical scenarios.

3. Methodology

3.1. The General Structure of YOLO-ABD

In the network architecture of YOLO-ABD, we adopt YOLOv8n as the base model, enhancing it with a small-object detection head and integrating GSConv convolutional blocks along with the SimAm attention mechanism for improvement. The network structure, illustrated in Figure 1, consists of three main components: the backbone, the neck, and the detection head. In the backbone, GSConv modules replace the original Conv modules of YOLOv8n, while the SimAm attention is integrated before the SPPF module. GSConv maintains implicit channel connections to reduce model complexity while retaining the learning capacity of standard convolutions, thereby enhancing feature extraction for abnormal behavior characterization. The SimAm attention mechanism mitigates background interference while maintaining parameter efficiency, thereby improving accuracy. To further enhance the network’s capability to perceive abnormal behaviors across multiple scales, new pathways and detection heads for small objects are designed within the neck and head sections to prevent overlooking or misclassifying distant pedestrians.

3.2. Baseline Model

YOLOv8 models are categorized into five versions based on network depth, width, and maximum number of channels: n, s, m, l, and x. Among these, YOLOv8n stands out with the least parameters and floating-point operations, making it highly efficient and suitable for real-time deployment while maintaining high accuracy. This paper adopts YOLOv8n as the baseline model, depicted in Figure 2. YOLOv8 consists of three main components: backbone, neck, and head. The backbone includes Conv modules, C2f modules, and an SPPF module focused on extracting features from input images. Conv modules process images or feature maps, generating new maps with reduced resolution and increased channels. C2f modules excel in deep and multi-scale feature extraction. The neck integrates multi-scale features through a combination of Feature Pyramid Network (FPN) [45] and Path Aggregation Network (PAN) [46], employing techniques like upsampling, channel concatenation, and deep feature extraction. These features are then passed to the head, which utilizes a decoupled head structure [47] to separate classification and prediction tasks. This design enables precise object detection regarding size and position using feature maps of varying scales.

3.3. Small Object Detection Head

In the IITB-Corridor dataset [48], distant pedestrians occupy a relatively small portion of the image. When resizing the dataset images to 640 × 640 pixels, the input image includes numerous small targets. However, after multiple upsampling and pooling operations in the neck network, many detection features associated with small targets are lost, resulting in missed detections. The baseline model includes detection heads sized at 80 × 80, 40 × 40, and 20 × 20. The 80 × 80 detection head’s receptive field is overly extensive for small targets, causing the baseline model to struggle in accurately detecting abnormal behaviors of distant pedestrians.
To address this issue, we enhanced the baseline model by introducing a 160 × 160 small target detection head. The structural diagram for this new detection head is depicted in Figure 1. Minimal modifications were made to the original model structure, as shown in Figure 1 for clarity. The initially upsampled feature map in the neck layer, originally sized at 40 × 40, undergoes two additional upsampling steps to produce a detailed 160 × 160 feature map rich in small target information. This feature map is then concatenated with the 160 × 160 feature map from the second layer of the backbone network, enhancing the model’s capability to detect small target behaviors at the 160 × 160 scale. Finally, this refined feature map is fed into the detection head layer, creating a new small target detection head that effectively reduces missed and false detections of abnormal behaviors across various scales.

3.4. GSConv Module

Currently, lightweight network design heavily relies on Depth-wise Separable Convolution (DSConv) to reduce model parameters and floating-point operations [49,50,51]. DSConv conducts convolutions independently across three channels, which minimizes redundant feature information but also fully separates the channel information of input data, thereby reducing the model’s feature extraction capability compared to dense channel convolutions, such as standard convolutions. To address this limitation, Zhang et al. [52] introduced “ShuffleNet”, which employs pointwise group convolution and channel shuffle operations to significantly decrease computational cost while maintaining accuracy. GhostNet [53] generates multiple feature maps through a series of low-cost linear transformations, effectively revealing intrinsic feature information.
Li et al. [23] introduced GSConv as an alternative to standard convolutions, offering a computational cost approximately 60–70% lower while retaining comparable learning capability. Illustrated in Figure 3, GSConv initially applies a standard convolution to the input feature map with C 1 channels, resulting in an intermediate feature map with C 2 /2 channels. Subsequently, this intermediate feature map undergoes Depth-wise Separable Convolution (DSConv) to produce another intermediate feature map, also with C 2 /2 channels. Finally, the two intermediate feature maps are concatenated and shuffled to yield an output feature map with C 2 channels. This approach mitigates information loss that can occur with DSConv’s channel separation, while maintaining an output similar to that of standard convolution. The time complexity formulas for standard convolution (SC), Depth-wise Separable Convolution (DSC), and GSConv are shown in Equations (1), (2) and (3), respectively:
O W · H · K 1 · K 2 · C 1 · C 2
O W · H · K 1 · K 2 · C 1
O W · H · K 1 · K 2 · C 2 / 2 · C 1 + 1
In the above formulas, W and H represent the width and height of the output feature map, K 1 and K 2 are the sizes of the convolution kernels, C 1 is the number of input feature channels, and C 2 is the number of output feature channels.

3.5. SimAm Attention Module

To enhance the model’s capability in detecting objects and mitigate interference from complex backgrounds, Liu et al. [54] introduced the SimAM attention mechanism before the detection head, effectively reducing background noise interference, which allowed the model to focus more precisely on the small target of corn silk, significantly enhancing the model’s ability to extract distinguishable features. In the study in [55], to address the challenges of small target detection and significant sea surface interference in maritime environments, the authors enhanced the YOLOv7 model by incorporating a small-target detection head to better detect tiny targets. Additionally, they integrated the SimAM module to identify attention regions within the scene, thereby reducing sea surface interference and markedly improving the model’s performance. Based on the case study, the SimAM module demonstrates a significant impact on reducing background interference and enhancing model performance in object detection tasks.
This paper introduces SimAm, a parameter-free attention mechanism. SimAm, rooted in neural network theory, distinguishes itself from existing attention mechanisms [56,57], which typically focus on either the channel or spatial domain. Unlike CBAM [58] and GAM [59], which extract features from both channel and spatial dimensions and merge them into a hybrid attention mechanism, SimAm integrates spatial, channel, and feature dimensions to generate 3D weights. Figure 4 illustrates the generation of these 3D weights.
In visual neuroscience, the activity of an active neuron can induce spatial suppression in the surrounding neurons. SimAm utilizes this concept by prioritizing neurons that induce spatial suppression. The priority of each neuron is determined by an energy function, represented by Equation (4).
e t ( w t , b t , x i ) = 1 M 1 i = 1 M 1 ( 1 ( w t x i + b t ) ) 2 + ( 1 ( w t + b t ) ) 2 + λ w t 2
In the formula, t and x i represent the target neuron and other neurons of a single channel in the input features, respectively. Here, i is the index in the spatial dimension, and M = H × W denotes the number of neurons in that single channel. The weight and bias of the target neuron, denoted as w t and b t , respectively, are computed using the formulas shown in (5) and (6):
w t = 2 ( t μ t ) ( t μ t ) 2 + 2 σ t 2 + 2 λ
b t = 1 2 ( t + μ t ) w t
In the formula, μ t = 1 M 1 i = 1 M 1 x i and σ t 2 = 1 M 1 i = 1 M 1 ( x i μ t ) 2 . When ensuring that all pixels within a single channel follow the same distribution, the minimum energy function can be obtained:
e t * = 4 ( σ ^ + λ ) ( t μ ^ ) 2 + 2 σ ^ 2 + 2 λ
In Formula (7), μ ^ and σ ^ represent the mean and variance of all neurons except t. A smaller e t * indicates higher importance, and the importance calculation is 1 / e t * . Scaling the energy of neurons for feature refinement within each neuron yields the final feature map, as shown in Formula (8):
X ˜ = s i g m o i d ( 1 E ) X
To assess the effectiveness of the SimAm attention module, two images showcasing abnormal behaviors were selected from the IITB-Corridor dataset. Feature extraction was conducted using both the baseline model and the SimAm-enhanced model, followed by visualization of the resulting feature heatmaps, as shown in Figure 5. It is evident that following the integration of the SimAM attention module, the neural network’s focus on human actions exhibits a higher degree of congruence with the human brain’s attention to action-related features. The distribution of heatmap values (represented by warm-colored regions) is more prominently concentrated on the human body or the actions being performed by it, thereby mitigating the interference caused by extraneous background elements. Thus, the introduction of the SimAM attention module has significantly enhanced the baseline model’s capability to detect abnormal human behavior.

4. Experiments

4.1. Dataset

The IITB-Corridor dataset [48] was curated to investigate abnormal human activities within the corridors of the Indian Institute of Technology Bombay. These videos were captured by fixed cameras overlooking pedestrian corridors on the Mumbai campus. The dataset encompasses a variety of both normal and abnormal behaviors. Normal activities typically include common behaviors such as walking and standing, whereas abnormal behaviors encompass relatively rare and potentially hazardous activities like running, fighting, playing soccer, and riding a bicycle. Notably, some abnormal behaviors are naturally occurring incidents captured by surveillance videos, reflecting unexpected situations in daily life, while others are intentionally arranged by the dataset creators to simulate behavior patterns in specific contexts. These abnormal behaviors vary at both individual and group levels, providing a rich analytical dimension for research. Furthermore, in public settings such as campus hallways, video data are usually obtained through surveillance cameras, exhibiting several distinct characteristics. Firstly, the high density of people in the data results in numerous pedestrians within the frame. Secondly, due to the limitations of camera position and angle, distant human targets are small and difficult to identify, while close human targets may be partially out of the camera’s view. Additionally, overlapping individuals and shadow occlusion significantly increase the difficulty of behavior recognition. These features not only complicate data analysis but also pose higher demands on the accuracy and robustness of algorithms.
To manage the extensive collection of abnormal and non-abnormal video frames in the IITB-Corridor dataset, abnormal frames were sampled at a rate of three frames per second from each video depicting abnormal behavior. Subsequently, all abnormal behaviors were annotated using the labelImg annotation tool, as illustrated in Table 1.The adjusted dataset consists of 18,674 video frame images, covering eight categories of abnormal behaviors: Bag Exchange, Cycling, Suspicious Object, Running, Fighting, Hiding, Playing With Ball, and Protest, with representative abnormal frame images shown in Figure 6. In this paper, the first three letters of each category denote a specific class (e.g., “Bag” corresponds to “Bag Exchange”). To facilitate model training and validation, the dataset was divided into training and validation sets in an 8:2 ratio.
The street-view-gdogo dataset [60] contains images captured by fixed traffic cameras positioned along urban roads in Turkey. Through manual annotation, common objects encountered in urban street traffic, such as Person, Car, Bus, Motor, and Bicycle, were categorized into five classes. The dataset comprises a total of 6685 images, which were divided into training, validation, and test sets at a ratio of 8:1:1.

4.2. Training Setting

The experiment on abnormal behavior detection was conducted on the Ubuntu 20.04 operating system, utilizing Python 3.8 as the programming language and CUDA 12.2 and PyTorch 1.9.0 for developing the deep learning framework. The hardware configuration included an NVIDIA GeForce RTX 3080 GPU with 10GB of memory and an Intel Xeon Silver 4210R CPU with 128GB of memory. Experimental parameters were set as follows: a batch size of 16 and training for 200 epochs and employing the Adam optimizer with an initial learning rate of 0.01, weight decay set to 0.0005, and a confidence threshold for object detection of 0.25. Image preprocessing involved resizing images to dimensions of 640 × 640 pixels.

4.2.1. Evaluating Indicators

To evaluate the practical performance of YOLO-ABD, the experiments employ established model evaluation metrics from the object detection domain. These metrics include precision, recall, average precision (AP), mean average precision (mAP), giga floating-point operations per second (GFLOPs), parameters, and frames per second (FPS). Precision measures the ratio of correctly identified positive samples to all samples classified as positive by the model, while recall evaluates the ratio of correctly identified positive samples to all actual positive samples. The definitions of precision and recall are as follows:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
TP represents true positives, indicating the number of positive samples correctly predicted as positive by the model; FP represents false positives, indicating the number of negative samples incorrectly predicted as positive by the model; FN denotes false negatives, indicating the number of positive samples incorrectly predicted as negative by the model. The precision–recall (PR) curve illustrates the relationship between recall (X-axis) and precision (Y-axis) across various thresholds. AP represents the area under the PR curve for each class, while mAP denotes the average AP value across all classes. The calculation formulas are as follows:
A P = 0 1 P ( R ) d R
m A P = 1 N i = 1 N A P i
mAP@50 denotes the mean average precision at an IoU threshold of 0.5, while mAP@50:95 signifies the mean average precision across IoU thresholds ranging from 0.5 to 0.95. Parameters and GFLOPs are utilized to assess model size and computational efficiency, as well as to evaluate the model’s hardware requirements. Frames per second (FPS) represents the number of images the model can process per second, providing insights into the model’s real-time performance and detection speed. These metrics collectively offer comprehensive insights into the effectiveness and efficiency of the model in practical scenarios.

4.2.2. Result Analysis

To validate the performance improvement of YOLO-ABD, comparative experiments were conducted on the IITB-Corridor dataset to evaluate various object detection models. Key evaluation metrics such as AP, mAP@50, GFLOPs, and parameters were used to assess model performance and compare it with several state-of-the-art methods. Each category of abnormal behavior is represented by the first three letters: “Bag” for “Bag Exchange”, “Cyc” for “Cycling”, “Sus” for “Suspicious Object”, “Run” for “Running”, “Fig” for “Fighting”, “Hid” for “Hiding”, “Pla” for “Playing With Ball”, and “Pro” for “Protest”. The experimental results are summarized in Table 2, offering a comprehensive comparison of model performance.
All models were trained from scratch to ensure fairness, without employing pre-training methods. Analysis of the experimental results indicates that YOLO-ABD achieves superior detection performance on the IITB-Corridor dataset, with an mAP@50 score of 89.3%. YOLO-ABD shows a performance improvement of 6.8 percentage points compared to the early-stage two-stage object detection method Faster R-CNN [61]. Significant advancements are observed compared to the YOLO series, particularly in the “Cyc” and “Run” categories. Compared to YOLOv3, YOLO-ABD demonstrates superior performance across most categories, notably in “Run”, “Pla”, and “Pro”, with a 2.6 percentage point increase in mAP@50. Compared to the YOLOv5s and YOLOv5n models, YOLO-ABD achieves higher accuracy across all categories, with mAP@50 scores improving by 8.1 and 12.9 percentage points, respectively. When compared to the YOLOv6s and YOLOv6n models, YOLO-ABD notably enhances accuracy in the “Pro” category, with mAP@50 scores increasing by 4.8 and 8.1 percentage points, respectively. Against the YOLOv8s and YOLOv8n models, YOLO-ABD surpasses them in nearly all categories, particularly in “Run”, “Pla”, and “Pro”, with significant improvements in accuracy and mAP@50 scores. Compared to the latest YOLOv9t and YOLOv10n models, the YOLO-ABD model achieves higher accuracy across all categories, with mAP@50 scores improving by 8.3 and 6.7 percentage points, respectively. Overall, YOLO-ABD exhibits significant performance advantages by achieving higher mAP@50 values for pedestrian anomaly detection while reducing the number of parameters. It retains the lightweight characteristics of the YOLO series, demonstrating that the proposed model offers considerable advancements.
The PR curves shown in Figure 7 provide a visual comparison between our proposed method and the YOLO variants with similar computational complexity on the IITB-Corridor dataset. Our method exhibits a smoother curve, positioned closer to the top-right corner, indicating superior precision and recall performance. This demonstrates a more balanced detection capability, with notable improvements across various classes, particularly evident in the “Bag Exchange” category. These findings underscore the enhanced effectiveness of our proposed method in detecting abnormal behaviors, validating its superiority over existing YOLO variants.
Comparison of the actual detection results between the baseline model and our proposed model, shown in Figure 8 and Figure 9, respectively, reveals significant differences. The baseline model YOLOv8n demonstrates noticeable missed detections on the IITB-Corridor dataset, particularly in scenarios involving crowded and overlapping individuals. In contrast, our proposed model exhibits enhanced accuracy, effectively detecting abnormal targets even under challenging conditions. This improvement is reflected in the higher accuracy and precision of the detection boxes for individuals. Upon analysis of the results depicted in Figure 8 and Figure 9, our proposed model demonstrates superior performance in detecting abnormal behaviors such as “Cycling”, “Suspicious Object”, “Protest”, and “Fighting” compared to the baseline model. Notably, it accurately identifies the “Protest” abnormal behavior, which the baseline model fails to detect, without any missed detections, highlighting the efficacy of our approach. In summary, these findings validate the advanced and effective performance of our proposed model.

4.2.3. Generalization Study

To evaluate the generalization performance of YOLO-ABD in object detection tasks, this study utilized the street-view-gdogo traffic object detection dataset for testing and compared it with several contemporary advanced object-detection models. The experimental results, summarized in Table 3, demonstrate the performance of our model across five classes: Bicycle, Bus, Car, Motor, and Person. Our model achieved notable accuracy rates of 93.5%, 95.4%, 97%, 89.9%, and 88.1% for these respective classes. Comparative analysis with the baseline model YOLOv8n reveals that our model consistently outperforms in most categories, with a significant 2.6% increase observed in the Person class. These findings highlight the effectiveness and robust generalization capabilities of our model across diverse scenarios.

4.2.4. Ablation Study

Table 4 compares the accuracy of the YOLOv8n model integrated with three different convolution modules (SC, DSC, and GSConv) on the IITB-Corridor dataset. To ensure the reliability of the evaluation, each convolution module was applied to the backbone network under identical conditions. The results indicate that the GSConv module achieved superior performance with a lower time complexity. The advantage of the GSConv module lies in the efficient collaboration between its main and auxiliary branches: the main branch captures channel features using the SC module, while the auxiliary branch focuses on spatial information using the DSC module, with a final fusion of these features through a mixing operation. Consequently, the GSConv module captures more spatial and channel features while maintaining lower computational costs, a feat that other convolution methods struggle to achieve simultaneously in optimizing computational efficiency, spatial information, and channel features. Figure 10 displays the visualized feature maps of the SC, DSConv, and GSConv modules. The texture features of the GSConv module are noticeably more similar to the SC module than those of the DSConv module compared to the SC module.
To validate the effectiveness of YOLO-ABD and assess the contributions of SimAm, GSConv, and the small-object detection heads to the model’s accuracy and computational efficiency, eight sets of comparative experiments were conducted. Using YOLOv8n as the baseline model, ablation experiments were conducted on the IITB-Corridor dataset with an input image size of 640 × 640. In the experiments, the positions and quantities of module replacements correspond precisely to the improvements in the YOLO-ABD model. The experimental results are shown in Table 5.
From the experimental findings, it is evident that integrating the SimAm module alone improved the model’s mAP50 metric by 0.3 percentage points. The SimAm module enhances performance without adding parameters, thus avoiding increased model complexity or computational cost, and has negligible impact on floating-point operations.Introducing the GSConv module alone improved the model’s mAP50 and mAP50-90 metrics on the dataset by 1.1 and 0.7 percentage points, respectively, while reducing floating-point operations by 0.4 GFLOPs. Adding the small-object detection head alone enhanced the model’s mAP50 and mAP50-90 metrics by 0.4 and 0.3 percentage points, respectively. However, integrating the small-object detection path increased floating-point operations from 8.1 to 11.8 GFLOPs, slightly reducing detection speed, though still suitable for real-time applications.
Combining pairs of the three modules significantly improved both mAP50 and mAP50-90 metrics. By incorporating a small-object detection head and the GSConv module, the mAP50 metric was increased to 89%, and the mAP50-90 metric was improved to 59.2%. This enhancement was achieved while keeping additional floating-point operations to a minimum and significantly reducing the parameter count of the baseline model. Integrating all three modules seamlessly merged with the baseline model, maximizing efficiency and resulting in improvements of 2.9 and 3.2 percentage points in mAP50 and mAP50-90 metrics, respectively. Although integrating the small-object detection head increased floating-point operations, the GSConv module effectively reduced the baseline model’s overall floating-point operations.
Ablation experiments were conducted by individually integrating different modules based on the baseline model, followed by anomaly behavior testing and comparative diagram generation, as depicted in Figure 11. These diagrams illustrate that while individual integration of SimAm, GSConv, and the small-object detection head each improves performance over the original YOLOv8n model, their simultaneous integration yields substantial enhancements in anomaly behavior detection performance.

5. Conclusions

This study introduces the YOLO-ABD model, a lightweight pedestrian anomaly detection method based on the YOLOv8n baseline network. YOLO-ABD addresses the challenge of small-object detection performance by incorporating a dedicated small target detection head. The model integrates the GSConv module to optimize channel connections while minimizing computational cost. Additionally, the inclusion of the SimAM attention module further reduces background interference and enhances model accuracy. Experimental results on the IITB-Corridor dataset demonstrate that YOLO-ABD achieves mAP50 and mAP50-95 scores of 89.3% and 60.6%, respectively, without requiring additional pre-training. Compared to current state-of-the-art object detection algorithms, YOLO-ABD strikes an optimal balance between speed and accuracy. It effectively reduces false negatives and false positives caused by background interference and occlusion, achieving an excellent combination of accuracy and lightweight design. Experiments on the street-view-gdogo dataset highlight the generalization capability of YOLO-ABD.
This study proposes a supervised learning-based method for pedestrian anomaly detection. Given the scarcity of pedestrian anomaly behavior data, initial experiments require extensive manual annotation of pedestrian anomaly behaviors, resulting in substantial data processing and manual labor. Future research could explore the use of unsupervised or semi-supervised learning methods to reduce the dependency of anomaly detection models on labeled data, achieving high performance and accuracy with a small amount of data.

Author Contributions

Conceptualization, C.H. and K.L.; methodology, C.H. and K.L.; software, K.L.; validation, K.L.; formal analysis, C.H.; investigation, K.L. and Y.W.; resources, K.L. and R.S.; data curation, K.L.; writing—original draft preparation, K.L.; writing—review and editing, C.H.; visualization, K.L. and Y.W.; supervision, C.H. and Y.W.; project administration, C.H. and R.S.; funding acquisition, Y.W. and R.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Scientific Research and Innovation Team Program of Sichuan University of Science and Technology under Grant SUSE652A006.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets can be provided by the corresponding author upon reasonable request.

Acknowledgments

The authors would like to express their heartfelt gratitude to those people who have helped with this manuscript and to the reviewers for their comments on the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Pang, G.; Shen, C.; Cao, L.; Hengel, A.V.D. Deep learning for anomaly detection: A review. ACM Comput. Surv. 2021, 54, 1–38. [Google Scholar] [CrossRef]
  2. Nassif, A.B.; Talib, M.A.; Nasir, Q.; Dakalbab, F.M. Machine learning for anomaly detection: A systematic review. IEEE Access 2021, 9, 78658–78700. [Google Scholar] [CrossRef]
  3. Ristea, N.C.; Madan, N.; Ionescu, R.T.; Nasrollahi, K.; Khan, F.S.; Moeslund, T.B. Self-supervised predictive convolutional attentive block for anomaly detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13576–13586. [Google Scholar]
  4. Kwon, H. Adversarial image perturbations with distortions weighted by color on deep neural networks. Multimed. Tools Appl. 2023, 82, 13779–13795. [Google Scholar] [CrossRef]
  5. Chen, B.; Wang, X.; Bao, Q.; Jia, B.; Li, X.; Wang, Y. An unsafe behavior detection method based on improved YOLO framework. Electronics 2022, 11, 1912. [Google Scholar] [CrossRef]
  6. Liu, B.; Yu, C.; Chen, B.; Zhao, Y. YOLO-GP: A Multi-Scale Dangerous Behavior Detection Model Based on YOLOv8. Symmetry 2024, 16, 730. [Google Scholar] [CrossRef]
  7. Ravanbakhsh, M.; Nabi, M.; Sangineto, E.; Marcenaro, L.; Regazzoni, C.; Sebe, N. Abnormal event detection in videos using generative adversarial nets. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 1577–1581. [Google Scholar]
  8. Lv, H.; Chen, C.; Cui, Z.; Xu, C.; Li, Y.; Yang, J. Learning normal dynamics in videos with meta prototype network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15425–15434. [Google Scholar]
  9. Yajing, L.; Zhongjian, D. Abnormal behavior detection in crowd scene using YOLO and Conv-AE. In Proceedings of the 2021 33rd Chinese Control and Decision Conference (CCDC), Kunming, China, 22–24 May 2021; pp. 1720–1725. [Google Scholar]
  10. Dong, F.; Zhang, Y.; Nie, X. Dual discriminator generative adversarial network for video anomaly detection. IEEE Access 2020, 8, 88170–88176. [Google Scholar] [CrossRef]
  11. Lee, S.; Kim, H.G.; Ro, Y.M. BMAN: Bidirectional multi-scale aggregation networks for abnormal event detection. IEEE Trans. Image Process. 2019, 29, 2395–2408. [Google Scholar] [CrossRef]
  12. Ullah, W.; Hussain, T.; Ullah, F.U.M.; Lee, M.Y.; Baik, S.W. TransCNN: Hybrid CNN and transformer mechanism for surveillance anomaly detection. Eng. Appl. Artif. Intell. 2023, 123, 106173. [Google Scholar] [CrossRef]
  13. Pang, G.; Yan, C.; Shen, C.; Hengel, A.V.D.; Bai, X. Self-trained deep ordinal regression for end-to-end video anomaly detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12173–12182. [Google Scholar]
  14. Hao, Y.; Tang, Z.; Alzahrani, B.; Alotaibi, R.; Alharthi, R.; Zhao, M.; Mahmood, A. An end-to-end human abnormal behavior recognition framework for crowds with mentally disordered individuals. IEEE J. Biomed. Health Inf. 2021, 26, 3618–3625. [Google Scholar] [CrossRef]
  15. Chen, S.; Guo, W. Auto-encoders in deep learning—A review with new perspectives. Mathematics 2023, 11, 1777. [Google Scholar] [CrossRef]
  16. Gong, D.; Liu, L.; Le, V.; Saha, B.; Mansour, M.R.; Venkatesh, S.; Hengel, A.V.D. Memorizing normality to detect anomaly: Memory-augmented deep autoencoder for unsupervised anomaly detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1705–1714. [Google Scholar]
  17. Luo, W.; Liu, W.; Lian, D.; Gao, S. Future frame prediction network for video anomaly detection. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 7505–7520. [Google Scholar] [CrossRef]
  18. Li, S.; Fang, J.; Xu, H.; Xue, J. Video frame prediction by deep multi-branch mask network. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 1283–1295. [Google Scholar] [CrossRef]
  19. Wang, X.; Che, Z.; Jiang, B.; Xiao, N.; Yang, K.; Tang, J.; Qi, Q. Robust unsupervised video anomaly detection by multipath frame prediction. IEEE Trans. Neural Networks Learn. Syst. 2022, 33, 2301–2312. [Google Scholar] [CrossRef] [PubMed]
  20. Li, C.; Li, H.; Zhang, G. Future frame prediction based on generative assistant discriminative network for anomaly detection. Appl. Intell. 2023, 53, 542–559. [Google Scholar] [CrossRef]
  21. Straka, Z.; Svoboda, T.; Hoffmann, M. PreCNet: Next-frame video prediction based on predictive coding. IEEE Trans. Neural Netw. Learn. Syst. 2023, 1–15. [Google Scholar] [CrossRef] [PubMed]
  22. Hussain, M. YOLOv1 to v8: Unveiling Each Variant—A Comprehensive Review of YOLO. IEEE Access 2024, 12, 42816–42833. [Google Scholar] [CrossRef]
  23. Li, H.; Li, J.; Wei, H.; Liu, Z.; Zhan, Z.; Ren, Q. Slim-neck by GSConv: A better design paradigm of detector architectures for autonomous vehicles. arXiv 2022, arXiv:2206.02424. [Google Scholar]
  24. Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. Simam: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 11863–11874. [Google Scholar]
  25. Cheoi, K.J. Temporal saliency-based suspicious behavior pattern detection. Appl. Sci. 2020, 10, 1020. [Google Scholar] [CrossRef]
  26. Smoliński, A.; Forczmański, P.; Nowosielski, A. Processing and Integration of Multimodal Image Data Supporting the Detection of Behaviors Related to Reduced Concentration Level of Motor Vehicle Users. Electronics 2024, 13, 2457. [Google Scholar] [CrossRef]
  27. Xie, B.; Guo, H.; Zheng, G. Mining Abnormal Patterns in Moving Target Trajectories Based on Multi-Attribute Classification. Mathematics 2024, 12, 1924. [Google Scholar] [CrossRef]
  28. Lei, J.; Sun, W.; Fang, Y.; Ye, N.; Yang, S.; Wu, J. A Model for Detecting Abnormal Elevator Passenger Behavior Based on Video Classification. Electronics 2024, 13, 2472. [Google Scholar] [CrossRef]
  29. Xie, Y.; Zhang, S.; Liu, Y. Abnormal Behavior Recognition in Classroom Pose Estimation of College Students Based on Spatiotemporal Representation Learning. Trait. Signal 2021, 38, 89–95. [Google Scholar] [CrossRef]
  30. Banerjee, S.; Ashwin, T.S.; Guddeti, R.M.R. Multimodal behavior analysis in computer-enabled laboratories using nonverbal cues. Signal Image Video Proces. 2020, 14, 1617–1624. [Google Scholar] [CrossRef]
  31. Guan, Y.; Hu, W.; Hu, X. Abnormal behavior recognition using 3D-CNN combined with LSTM. Multimed. Tools Appl. 2021, 80, 18787–18801. [Google Scholar] [CrossRef]
  32. Rashmi, M.; Ashwin, T.S.; Guddeti, R.M.R. Surveillance video analysis for student action recognition and localization inside computer laboratories of a smart campus. Multimed. Tools Appl. 2021, 80, 2907–2929. [Google Scholar] [CrossRef]
  33. Lentzas, A.; Vrakas, D. Non-intrusive human activity recognition and abnormal behavior detection on elderly people: A review. Artif. Intell. Rev. 2020, 53, 1975–2021. [Google Scholar] [CrossRef]
  34. Lina, W.; Ding, J. Behavior detection method of OpenPose combined with Yolo network. In Proceedings of the 2020 International Conference on Communications, Kuala Lumpur, Malaysia, 3–5 July 2020; pp. 326–330. [Google Scholar]
  35. Ganagavalli, K.; Santhi, V. YOLO-based anomaly activity detection system for human behavior analysis and crime mitigation. Signal Image Video Process. 2024, 18, 417–427. [Google Scholar] [CrossRef]
  36. Zhou, T.; Zheng, L.; Peng, Y.; Jiang, R. A survey of research on crowd abnormal behavior detection algorithm based on YOLO network. In Proceedings of the 2022 2nd International Conference on Consumer Electronics and Computer Engineering (ICCECE), Guangzhou, China, 14–16 January 2022; pp. 783–786. [Google Scholar]
  37. Maity, M.; Banerjee, S.; Chaudhuri, S.S. Faster r-cnn and yolo based vehicle detection: A survey. In Proceedings of the 2021 5th International Conference on Computing Methodologies and Communication (ICCMC), Erode, India, 8–10 April 2021; pp. 1442–1447. [Google Scholar]
  38. Mansour, R.F.; Escorcia-Gutierrez, J.; Gamarra, M.; Villanueva, J.A.; Leal, N. Intelligent video anomaly detection and classification using faster RCNN with deep reinforcement learning model. Image Vis. Comput. 2021, 112, 104229. [Google Scholar] [CrossRef]
  39. Su, H.; Ying, H.; Zhu, G.; Zhang, C. Behavior Identification based on Improved Two-Stream Convolutional Networks and Faster RCNN. In Proceedings of the 2021 33rd Chinese Control and Decision Conference (CCDC), Kunming, China, 22–24 May 2021; pp. 1771–1776. [Google Scholar]
  40. Chen, N.; Man, Y.; Sun, Y. Abnormal cockpit pilot driving behavior detection using YOLOv4 fused attention mechanism. Electronics 2022, 11, 2538. [Google Scholar] [CrossRef]
  41. Chen, H.; Zhou, G.; Jiang, H. Student behavior detection in the classroom based on improved YOLOv8. Sensors 2023, 23, 8385. [Google Scholar] [CrossRef]
  42. Chang, J.; Zhang, G.; Chen, W.; Yuan, D.; Wang, Y. Gas station unsafe behavior detection based on YOLO-V3 algorithm. China Saf. Sci. J. 2023, 33, 31–37. [Google Scholar]
  43. Benjumea, A.; Teeti, I.; Cuzzolin, F.; Bradley, A. YOLO-Z: Improving small object detection in YOLOv5 for autonomous vehicles. arXiv 2021, arXiv:2112.11798. [Google Scholar]
  44. Xiao, Y.; Wang, Y.; Li, W.; Sun, M.; Shen, X.; Luo, Z. Monitoring the Abnormal Human Behaviors in Substations based on Probabilistic Behaviours Prediction and YOLO-V5. In Proceedings of the 2022 7th Asia Conference on Power and Electrical Engineering (ACPEE), Hangzhou, China, 15–17 April 2022; pp. 943–948. [Google Scholar]
  45. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  46. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  47. Wang, H.; Jin, Y.; Ke, H.; Zhang, X. DDH-YOLOv5: Improved YOLOv5 based on Double IoU-aware Decoupled Head for object detection. J. Real-Time Image Process. 2022, 19, 1023–1033. [Google Scholar] [CrossRef]
  48. Rodrigues, R.; Bhargava, N.; Velmurugan, R.; Chaudhuri, S. Multi-timescale trajectory prediction for abnormal human activity detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 2626–2634. [Google Scholar]
  49. Gennari, M.; Fawcett, R.; Prisacariu, V.A. DSConv: Efficient Convolution Operator. arXiv 2019, arXiv:1901.01928. [Google Scholar]
  50. Guo, J.; Teodorescu, R.; Agrawal, G. Fused DSConv: Optimizing sparse CNN inference for execution on edge devices. In Proceedings of the 2021 IEEE/ACM 21st International Symposium on Cluster, Cloud and Internet Computing (CCGrid), Melbourne, Australia, 10–13 May 2021; pp. 545–554. [Google Scholar]
  51. Alalwan, N.; Abozeid, A.; ElHabshy, A.A.; Alzahrani, A. Efficient 3D deep learning model for medical image semantic segmentation. Alex. Eng. J. 2021, 60, 1231–1239. [Google Scholar] [CrossRef]
  52. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar]
  53. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. Ghostnet: More features from cheap operations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1580–1589. [Google Scholar]
  54. Liu, W.; Quijano, K.; Crawford, M.M. YOLOv5-Tassel: Detecting tassels in RGB UAV imagery with improved YOLOv5 based on transfer learning. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2022, 15, 8085–8094. [Google Scholar] [CrossRef]
  55. Zhao, H.; Zhang, H.; Zhao, Y. Yolov7-sea: Object detection of maritime uav images based on improved yolov7. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 233–238. [Google Scholar]
  56. Jin, X.; Xie, Y.; Wei, X.S.; Zhao, B.R.; Chen, Z.M.; Tan, X. Delving deep into spatial pooling for squeeze-and-excitation networks. Pattern Recognit. 2022, 121, 108159. [Google Scholar] [CrossRef]
  57. Jaderberg, M.; Simonyan, K.; Zisserman, A. Spatial transformer networks. Adv. Neural Inf. Process. Syst. 2015, 28, 2017–2025. [Google Scholar]
  58. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  59. Liu, Y.; Shao, Z.; Hoffmann, N. Global attention mechanism: Retain information to enhance channel-spatial interactions. arXiv 2021, arXiv:2112.05561. [Google Scholar]
  60. FSMVU. Street View Dataset. 2023. Available online: https://universe.roboflow.com/fsmvu/street-view-gdogo (accessed on 5 September 2023).
  61. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef]
Figure 1. Network structure of YOLO-ABD.
Figure 1. Network structure of YOLO-ABD.
Symmetry 16 01003 g001
Figure 2. Network structure of YOLOv8 model.
Figure 2. Network structure of YOLOv8 model.
Symmetry 16 01003 g002
Figure 3. Symmetric structure of the GSConv module.
Figure 3. Symmetric structure of the GSConv module.
Symmetry 16 01003 g003
Figure 4. Schematic diagram of SimAM attention module.
Figure 4. Schematic diagram of SimAM attention module.
Symmetry 16 01003 g004
Figure 5. Comparison of Extracted Features Across Different Models. (a) Original. (b) Baseline model. (c) Baseline model with SimAm.
Figure 5. Comparison of Extracted Features Across Different Models. (a) Original. (b) Baseline model. (c) Baseline model with SimAm.
Symmetry 16 01003 g005
Figure 6. Partial images of IITB-Corridor.
Figure 6. Partial images of IITB-Corridor.
Symmetry 16 01003 g006
Figure 7. Comparison of PR curves for YOLO series models with similar floating-point operation counts.
Figure 7. Comparison of PR curves for YOLO series models with similar floating-point operation counts.
Symmetry 16 01003 g007
Figure 8. Actual detection results of the baseline model on the IITB-Corridor dataset.
Figure 8. Actual detection results of the baseline model on the IITB-Corridor dataset.
Symmetry 16 01003 g008
Figure 9. Actual detection results of our model on the IITB-Corridor dataset.
Figure 9. Actual detection results of our model on the IITB-Corridor dataset.
Symmetry 16 01003 g009
Figure 10. Three type features of the different basic convolutional of the scale P2 of the Yolov8n. (a) features from SC; (b) features from DSConv; (c) features from GSConv.
Figure 10. Three type features of the different basic convolutional of the scale P2 of the Yolov8n. (a) features from SC; (b) features from DSConv; (c) features from GSConv.
Symmetry 16 01003 g010
Figure 11. Comparison of the ablation study result.
Figure 11. Comparison of the ablation study result.
Symmetry 16 01003 g011
Table 1. Categories and number of labels in the IITB-Corridor dataset.
Table 1. Categories and number of labels in the IITB-Corridor dataset.
DatasetAbnormal BehaviorNumber of Boxes
IITB-CorridorBag Exchange209
Cycling577
Suspicious Object2255
Running2301
Fighting2072
Hiding396
Playing With Ball2058
Protest5575
Table 2. Comparison results on the IITB-Corridor dataset.
Table 2. Comparison results on the IITB-Corridor dataset.
MethodsBagCycSusRunFigHidPlaPromAP50GFLOPsParameterFPS
Faster-RCNN76.5%87.3%96.0%70.2%90.4%85.2%66.9%87.5%82.5%---
YOLOv381.0%78.6%97.7%71.3%94.6%98.1%81.7%90.5%86.7%18.912.1M467
YOLOv5s70.4%82.0%94.8%55.4%89.0%95.7%66.8%89.2%80.4%23.88.6M381
YOLOv5n48.6%76.0%94.1%55.8%88.8%92.4%66.8%88.5%76.4%7.12.3M503
YOLOv6s73.7%85.3%96.9%67.2%93.6%95.0%72.8%91.4%84.5%44.015.5M352
YOLOv6n68.8%81.6%96.4%59.7%91.7%94.2%67.1%89.6%81.2%11.84.2M509
YOLOv8s76.7%88.1%96.6%62.0%91.1%96.5%72.4%90.9%84.3%28.511.1M380
YOLOv8n75.2%84.4%96.9%79.6%94.8%94.8%73.7%91.5%86.4%10.72.49M347
YOLOv9t64.9%75.8%94.5%70.1%93.3%93.7%65.7%89.6%81.0%8.72.86M504
YOLOv10n69.3%77.0%95.1%77.8%89.1%92.9%69.2%90.5%82.6%8.22.57M689
Ours76.7%92.5%96.7%84.6%95.6%95.7%80.5%92.2%89.3%11.42.56M436
Table 3. Comparison results on the street-view-gdogo dataset.
Table 3. Comparison results on the street-view-gdogo dataset.
MethodsBicycleBusCarMotorPersonmAP50GFLOPsParameterFPS
Faster-RCNN86.7%89.8%94.7%80.9%79.1%86.3%---
YOLOv392.1%94.2%94.3%80.9%68.2%85.9%18.912.1M95
YOLOv5s93.4%95.7%96.6%89.8%86.6%92.3%23.88.6M806
YOLOv5n91.9%96.1%96.6%89.3%84.9%91.8%7.12.3M111
YOLOv6s93.4%95.7%96.3%90.3%85.2%92.2%44.015.5M63
YOLOv6n88.2%91.3%95.7%86.8%76.6%87.7%11.84.2M105
YOLOv8s94.7%96.8%96.7%91.1%87.3%93.3%28.511.1M82
YOLOv8n92.5%95.7%96.5%89.1%85.5%91.9%8.12.49M65
YOLOv9t89.3%94.2%95.9%87.3%76.1%88.5%8.72.86M92
YOLOv10n91.2%94.9%96.3%87.5%83.1%90.6%8.22.57M139
Ours93.5%95.4%97%89.9%88.1%92.8%11.42.56M122
Table 4. Comparisons of SC, DSC, and GSConv modules on the IITB-Corridor dataset.
Table 4. Comparisons of SC, DSC, and GSConv modules on the IITB-Corridor dataset.
Basic Convolutional MethodmAP50mAP50-95GFLOPsParameterFPS
SC86.4%57.4%8.12.86M504
DSC85.6%55.8%7.32.5M505
GSConv87.5%58.1%7.72.68M552
Table 5. Experimental study of ablation on the IITB-Corridor dataset.
Table 5. Experimental study of ablation on the IITB-Corridor dataset.
SimAmGSConvSmallPrecisionRecallmAP50mAP50-95GFLOPsParameterFPS
---83.8%79.3%86.4%57.4%8.12.86M504
--83.8%79.5%86.7%57.4%8.12.86M507
--84.4%78.1%87.5%58.1%7.72.68M552
-85.0%79.0%88.0%58.2%7.72.68M504
--85.1%79.4%86.8%57.7%11.82.75M420
-85.0%81.1%87.4%58.2%11.82.75M405
-84.1%83.0%88.9%59.2%11.42.56M422
83.9%81.5%89.3%60.6%11.42.56M436
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hua, C.; Luo, K.; Wu, Y.; Shi, R. YOLO-ABD: A Multi-Scale Detection Model for Pedestrian Anomaly Behavior Detection. Symmetry 2024, 16, 1003. https://doi.org/10.3390/sym16081003

AMA Style

Hua C, Luo K, Wu Y, Shi R. YOLO-ABD: A Multi-Scale Detection Model for Pedestrian Anomaly Behavior Detection. Symmetry. 2024; 16(8):1003. https://doi.org/10.3390/sym16081003

Chicago/Turabian Style

Hua, Caijian, Kun Luo, Yadong Wu, and Rui Shi. 2024. "YOLO-ABD: A Multi-Scale Detection Model for Pedestrian Anomaly Behavior Detection" Symmetry 16, no. 8: 1003. https://doi.org/10.3390/sym16081003

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop