Next Article in Journal
Surface Defect Detection of Bearing Rings Based on an Improved YOLOv5 Network
Previous Article in Journal
On the Distribution of Muscle Signals: A Method for Distance-Based Classification of Human Gestures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Online Video Anomaly Detection

1
School of Information Science and Technology, Nantong University, Nantong 226019, China
2
State Key Lab. for Novel Software Technology, Nanjing University, Nanjing 210023, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(17), 7442; https://doi.org/10.3390/s23177442
Submission received: 9 May 2023 / Revised: 11 August 2023 / Accepted: 24 August 2023 / Published: 26 August 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
With the popularity of video surveillance technology, people are paying more and more attention to how to detect abnormal states or events in videos in time. Therefore, real-time, automatic and accurate detection of abnormal events has become the main goal of video-based surveillance systems. To achieve this goal, many researchers have conducted in-depth research on online video anomaly detection. This paper presents the background of the research in this field and briefly explains the research methods of offline video anomaly detection. Then, we sort out and classify the research methods of online video anomaly detection and expound on the basic ideas and characteristics of each method. In addition, we summarize the datasets commonly used in online video anomaly detection and compare and analyze the performance of the current mainstream algorithms according to the evaluation criteria of each dataset. Finally, we summarize the future trends in the field of online video anomaly detection.

1. Introduction

Video surveillance systems are widely used in shopping malls, hospitals, banks, streets, smart cities and other places to enhance the safety of public life and property [1]. However, the massive video data generated by the rapid development of video surveillance systems have brought great challenges to abnormal event detection based on manual interpretation. The traditional way of finding anomalies by manually viewing the monitored image records not only consumes a lot of human resources, but also may cause errors or omissions that cannot be remedied in time [2]. Therefore, it is very important to develop a technology that can automatically analyze and find abnormal conditions in surveillance videos without relying on a lot of manpower, which is video anomaly detection technology.
Video anomaly detection is the task of identifying abnormal events from video sequences that reflect frames that are significantly different from normal events. A video anomaly is usually an abnormal appearance or motion attribute in a video [3]. Due to the scarcity and diversity of abnormal samples, video anomaly detection methods usually only model the normal sample distribution. In the detection process, video frames or video clips that deviate from the normal sample distribution are regarded as anomalies [4]. In video surveillance scenarios, common examples of anomalies are pedestrians driving on sidewalks, pedestrians walking on lawns and vehicles incorrectly parking, as shown in Figure 1 [5].
On the one hand, online video anomaly detection usually refers to actively learning frames of normal events in a video sequence in the system or model, constantly updating the definition of a normal event. On the other hand, it refers to reducing the processing time of each incoming video frame online for fast spatio-temporal localization of abnormal events and real-time video anomaly detection. Ideal video anomaly detection is defined as the process of detecting and tracking abnormal events in real time [6]. However, most video anomaly detection methods are still offline. Considering that video anomaly detection can play an important role in ensuring security and confidentiality and preventing potential disasters, video anomaly detection systems should have real-time decision-making capabilities. For example, traffic accidents, robberies, fires in remote areas and other events need to be detected and solved in time through real-time detection of abnormal events [7]. However, the research focusing on online and real-time detection methods is very limited. Therefore, online (real-time) video anomaly detection has important research significance.
In summary, this paper briefly introduces the offline video anomaly detection method, mainly analyzes and studies the online video anomaly detection based on sample annotation and the technical methods used, and further classifies the methods on this basis, as shown in Figure 2. At the same time, the mainstream datasets of online video anomaly detection are sorted out, and the development direction of online video anomaly detection is prospected.

2. Common Datasets and Assessment Criteria

2.1. Common Single-Scene Anomaly Detection Datasets

Table 1 lists the specific information of current scene datasets commonly used for online video anomaly detection.
(1)
UMN dataset. The UMN dataset [8] has 11 videos containing three different scenes: campus lawn, indoor and square. This dataset belongs to the global abnormal behavior dataset.
(2)
UCSD Pedestrian dataset. The UCSD Pedestrian dataset [9] was captured on a university campus pavement. The current frame-level anomaly detection accuracies achieved on the UCSD Ped1 and UCSD Ped2 datasets are 97.4% and 97.8%, respectively. Although the accuracy rates are relatively high, this dataset is still one of the few more popular benchmark datasets due to its relatively adequate number of anomalies and anomaly types.
(3)
CUHK Avenue dataset. The CUHK Avenue dataset [10] contains 16 training videos and 21 test videos with a total of 47 abnormal events, including throwing objects, loitering and running. The size of the people changes depending on the position and angle of the camera.
(4)
ShanghaiTech (Shanghai, China) dataset. The ShanghaiTech dataset [11] contains 13 scenes with complex lighting conditions and camera angles, containing 130 abnormal events and over 270,000 training frames. In addition, pixel-level annotations of the abnormal events are given.
(5)
UCF-Crime dataset. The UCF-Crime dataset [12] is a large dataset of 128 h of video. It consists of 1900 continuous unedited surveillance videos containing a total of 13 real-life aberrant behaviors, namely abuse, arrest, arson, assault, road accident, burglary, explosion, combat, robbery, shooting, theft, shoplifting and vandalism.
(6)
The DOTA dataset. The DOTA dataset [13] is the first traffic anomaly dataset to provide detailed spatio-temporal annotation of abnormal objects. It contains 4677 videos, and each video contains exactly one abnormal event. The anomalies consider seven common traffic participant categories, namely pedestrians, cars, trucks, buses, motorbikes, bicycles and riders.

2.2. Evaluation Criteria

There are two main criteria for evaluating the effectiveness of anomaly detection: (1) the receiver operating characteristic (ROC) and its corresponding area under the curve (AUC) and (2) the equal error rate (EER). There are two levels of AUC evaluation criteria: the pixel level and the frame level. The frame-level criterion determines the correctness of anomaly detection on a frame-by-frame basis and only determines whether a frame contains an anomaly, without the need to accurately locate the spatial location of the anomaly. The pixel-level criterion requires the spatial location of the anomaly to be accurately located, and the detection is deemed successful when the detected anomaly area overlaps with the real anomaly area by 40%. For the evaluation criteria of EER, some literature argues that abnormal events are rare in real life and there is a serious sample imbalance between normal and abnormal events, so using EER as a metric can be misleading in practical applications [14,15].

3. Offline Video Anomaly Detection Methods

Depending on the data and the method, learning methods for anomaly detection can be broadly classified into supervised, unsupervised, weakly supervised and self-supervised learning. They can also be divided into reconstruction-based and prediction-based video anomaly detection methods according to deep learning methods.
In offline video anomaly detection methods, the learning methods of anomaly detection can be roughly divided into supervised learning, unsupervised learning, weakly supervised learning and self-supervised learning. Among them, the generative cooperative learning method is a truly unsupervised anomaly detection algorithm based on unlabeled training videos [16]. Implicit dual-path automatic encoding and a normal feature distribution model based on a normalized flow generation model can be used for anomaly detection [17]. Three-dimensional convolutional auto-encoders can be trained for anomaly detection by denoising reconstruction errors and adversarial learning strategies, which can better distinguish normal and abnormal events even without any supervised information [18]. Weakly supervised video anomaly detection methods can capture abnormal events at different scales from local to global spatio-temporal perception, and they can identify abnormal segments from the entire video segment marked as abnormal [19]. A multi-scale convolutional neural network (CNN) can be captured from video using relation-aware feature extractors, which is a new weakly supervised method [20]. Video anomaly detection methods based on self-supervised learning generally model and generate feature-level anomalies by using offline learning dictionaries and self-supervised learning [21]. A self-supervised learning method can solve a multi-label fine-grained classification problem [22], and CutPaste self-supervised learning can be used to improve and train the network to better adapt to the data [23].
State-of-the-art video anomaly detection methods use deep learning techniques and are divided into reconstruction-based and prediction-based video anomaly detection methods. Among them, the cascade reconstruction model combined with the frame reconstruction network and the optical flow prediction network can be used to increase the reconstruction error of video anomalies [24], so as to better distinguish normal and abnormal events. A generative assistant discriminative network is used to distinguish normal future frames from noisy future frames by introducing a helper, and this method can better learn discriminative features to distinguish small differences between normal and abnormal events [25]. Other methods use convolutional auto-encoder architectures to address large reconstruction errors caused by abnormal events with irregular appearance or motion behavior [26], and noise modulates the end-to-end training of generative adversarial networks to regulate the generalization ability of the reconstruction network [27]. The encoder and decoder are used to generate reconstruction errors to separate abnormal events from normal events to detect abnormal events [28]; in order to alleviate the loss caused by feature compression, the authors of [29] proposed the skip attention gate mechanism to complete the spatio-temporal information fusion between the encoder and the decoder. Patch-wise prediction method bidirectional architecture [30] combines the prediction error of appearance and motion and the commit error between original features and memory items to calculate the anomaly score and determine whether a frame is abnormal [31]. An encoder based on a 3D CNN and a decoder based on a 2D CNN constitute a network with spatio-temporal consistency enhancement. This network uses the difference in prediction quality between normal and abnormal events to infer whether an anomaly occurs [32]. Methods such as the adaptive Markov jumping particle filter [33] and active-Langevin model for predicting abnormal behavior have also been used for video anomaly detection [34].
Although most video-based abnormal event detection methods work well in simple scenes, they often suffer from low detection rates and high false alarm rates in complex motion scenes. Therefore, video anomaly detection methods should shift from offline detection to online (real-time) detection, so as to find abnormal events in time and make appropriate decisions. So, real-time abnormal event detection in video streams is a worthy future research direction.

4. Online Video Anomaly Detection Methods

In order to solve the problem of detecting and identifying anomalies in video accurately and in real time, this section mainly summarizes an overview of online video anomaly detection methods based on sample annotation, technical approaches and other methods and further subclassifies them to obtain a clearer understanding.

4.1. Online Anomaly Detection Based on Sample Annotation

4.1.1. Online Anomaly Detection Based on Supervised Learning

Supervised learning is the process of mapping all data samples to different class labels through training a model with known data samples and their one-to-one corresponding labels. However, due to the scarcity of abnormal videos, online video anomaly detection methods based on supervised learning are relatively rare. As shown in Table 2, to address the challenges and limitations of anomaly detection localization in real-time surveillance, a multi-objective neural anomaly detection framework has been proposed [35]. This method combined a future frame detector based on generative adversarial networks and a lightweight object detector (Yolov3). The extracted features were used for online anomaly detection by nonparametric statistical algorithms. Intelligent traffic monitoring systems need to be able to detect anomalies such as traffic accidents in real time, and [36] adopted the Yolo object detection framework and used a more heuristic approach to detect anomalies. Methods included background estimation, road mask extraction and adaptive thresholding to remove some erroneous anomalies. Finally, the final anomaly was detected and analyzed using a decision tree. In order to ensure the online working mode and fast response of autonomous driving anomaly detection, in [37], researchers used the video swin transformer model to model short-term memory through the current frame and a few previous frames, used temporal correlation for behavior detection, introduced the LSTM module to model the past long-term memory, and obtained the overall context information from the video in real-time and online scenarios. Luo et al. [38] jointly used features extracted by spatially optimized temporal coherent sparse coding and features extracted by time-stacked recurrent neural network auto-encoders in real-time anomaly detection. Meanwhile, the distance between sparse codes was weighted by the similarity between adjacent frames.
Although these online anomaly detection methods based on supervised learning can detect the labeled types of anomalies, their performance deteriorates sharply in complex scenes. Since there are many types of anomalies and they cannot all be manually labeled, anomaly detection approaches turn to unsupervised learning methods.

4.1.2. Online Anomaly Detection Based on Unsupervised Learning

Unsupervised learning does not need labeling of the data, which not only saves costs but makes the learned features more adaptive and richer. Therefore, unsupervised learning methods are widely used in online anomaly detection. As shown in Table 2, Chaker et al. [39] used a window-based method to construct a spatio-temporal cuboid; modeled crowd behavior through local social networks; and realized online abnormal behavior detection, localization, and global updating. Dang et al. [40] proposed a method for unsupervised anomaly detection, map coding and anomaly-aware path planning using aerial robots. This method uses deep learning features and support vector machines combined with Bayesian fusion on a map for multi-view anomaly consensus. Real-time detection of abnormal events in video streams is a difficult challenge due to the volatility of anomaly definitions, time constraints and the adaptability of parametric models. To address the problem, ref. [41] used deep learning features for image classification and used Bayesian technology to fuse, encode and update the abnormal information on the occupancy map reconstructed in real time, so as to realize the detection and location of environmental anomalies for robots. In addition, a path-planning method was proposed to maximize the sensor observation entropy of the automatic exploration location area and abnormal area. In [13], authors introduced a new large-scale benchmark dataset detection of traffic anomaly (DOTA) and proposed a new video anomaly detection evaluation metric called the spatio-temporal area under the curve (STAUC). Researchers predicted the future position of an object in a short period of time by using a multi-stream RNN framework. Most existing methods used manually labeled data and defined parameters to train offline detectors and have difficulty modeling in changing scenes. An online self-organizing map (SOM) was proposed and achieved more satisfactory results for local anomaly detection in outdoor crowded scenes [12]. Compared with the original SOM [42], the online SOM model has the ability to adjust the learning rate and domain size according to scene changes. In order to avoid the poor performance of the data space caused by the zero neural structure far away from the actual sample in SOM, Ref. [43] adopted a more flexible neural network—an online growth neural network. The model achieved anomaly detection in the changing data space by self-using parameter learning operations online, estimating the learning efficiency and adaptively adjusting the network size. To some extent, it effectively reduced false alarms and missed detection caused by model aging and threshold aging in unstable environments. The incremental spatio-temporal learner in [44] updated and distinguished new abnormal and normal behaviors in real time through active learning with fuzzy aggregation for anomaly detection and localization in real-time surveillance video streams. In addition, Monakhov et al. [45] proposed an algorithm for anomaly detection in complex videos such as surveillance video. The algorithm is based on the HTM architecture of a grid, which simplifies data by segmentation technology and realizes real-time unsupervised anomaly detection. It has good anti-noise ability and online learning ability and can combat concept drift when performing real-time anomaly detection in video surveillance. The authors of [46] studied an anomaly detection problem for a video dataset taken by drones in parking lots. These researchers proposed a new anomaly detection technique using multiple feature extraction methods, such as GoogleNet, HOG, PCA-HOG and HOG3D, to classify the extracted features into two categories, namely abnormal and normal, and constructed four one-class support vector machine (OCSVM)-based models for anomaly event detection. The goal of this research was to enable automated abnormal event detection to assist UAV surveillance tasks. Different from the above methods, the GridNet method proposed in [47] utilizes deep auto-encoders for anomaly detection under image-agnostic conditions. By using the meta-representation of the scene image and introducing a multi-part loss function consisting of an adjustable variational loss term and a new loss term, namely the Soft-F1 loss term, the proposed method can accurately detect object anomalies in a scene image.
These online anomaly detection methods based on unsupervised learning can reduce the need for manual annotation resources, but the performance is poor in complex scenarios. One of the main difficulties faced by video analysis is the huge amount of data, and only a small number of videos contain important information. Therefore, developing algorithms to automatically detect abnormal events in streaming or archived videos can significantly improve the efficiency of video detection. This is also the focus of future research.

4.1.3. Online Anomaly Detection Based on Weakly Supervised Learning

Semi-supervised learning is an important type of weakly supervised learning that can improve the learning effect by using labeled and unlabeled data without fully supervised tasks. As shown in Table 2, Ref. [48] introduced an auto-encoder framework based on multi-modal semi-supervised deep learning convolution–bidirectional long short-term memory model (CNN-BiLSTM). The proposed method generates video clips by segmenting the online video stream from RGB + D sensors into fixed-size video frames, and then it uses a CNN-BiLSTM auto-encoder semi-supervised framework that only utilizes weakly labeled video samples for training and is able to perform anomaly detection without abnormal video samples. In addition, a framework for multivariate, data-driven and sequential detection of anomalies in high-dimensional systems based on the availability of labeled data was proposed in [49], which is suitable for both semi-supervised and supervised environments. These methods provide an effective framework for semi-supervised learning, which can be applied to various fields to improve the efficiency and accuracy of machine learning.
By using a large number of coarsely labeled samples, weakly supervised learning can use data more effectively than semi-supervised learning to improve the performance of the model. However, video anomaly detection based on weak supervision has difficulty in distinguishing abnormal and normal events in complex situations during training, which leads to non-optimal separation boundaries. To solve this problem, some authors extracted features from a pre-trained 3D CNN [50] and considered the local spatio-temporal differences by introducing a dissimilarity attention module. A ranking loss function considering the detection score of the dissimilarity attention module (DAM) and the temporal attention weight of the DAM for gradient calculation was used for optimization. To enhance the detection of abnormal segments in video streams, a transformer-based multi-scale long–short-term context model for online anomaly localization was proposed [51]. The locator is able to recover useful information from historical observations and compensate for future semantics by anticipation, providing critical context for detecting what is happening. However, this method required considerable computational cost. To address issues such as the lack of ability to model the temporal relationship between video clips and the inability to extract enough discriminative features to distinguish between normal and abnormal clips, a weakly supervised temporal discrimination algorithm was proposed by Huang et al. A transformer-styled temporal feature aggregator (TTFA) was used to aggregate temporal features in [52], which addressed the issue of temporal relationship modeling and feature discrimination. Additionally, a self-guided discriminative feature encoder (SDFE) was introduced to enhance the separability of features. The performance of temporal modeling and anomaly detection in video clips can be improved with the application of these methods. Online video anomaly detection based on weak supervision is an important research topic in the field of intelligent surveillance. Previous research has tended to use a unified single-stage framework that struggles to address both online constraints and weakly supervised settings. Online video anomaly detection using a two-stage- or multi-stage-based framework is a worthwhile research direction for the future.
Table 2. Summary of online anomaly detection methods based on sample labeling.
Table 2. Summary of online anomaly detection methods based on sample labeling.
ReferenceMethodArchitectureRemarks
Doshi et al.
[35]
Supervised learningGAN and YOLOv3A multi-objective neural anomaly detection framework was proposed.
Aboah et al.
[36]
YOLO and
decision-making tree
The Yolo object detection framework was adopted, and the final anomaly was detected and analyzed using a decision tree.
Rossi et al.
[37]
Transformer and LSTMThe video swin transformer model was used, and the LSTM module was introduced to model long short-term memory.
Luo et al.
[38]
Sparse-coding-inspired DNNUsing spatially optimized temporally coherent sparse coding and temporally stacked recurrent neural network auto-encoders.
Chaker et al.
[39]
Unsupervised learningSocial network modelThe detection, location and global update of online abnormal behavior were realized.
Dang et al.
[40]
AlexNet and one-class SVM + Bayesian encodingDeep learning features and Bayesian technology were used to detect and locate environmental anomalies, and a path-planning method was proposed.
Zhao et al.
[41]
Dynamic sparse codingOnline sparse reconstruction methods and online reconfigurability were proposed to detect abnormal events in videos.
Yao et al.
[13]
Two-stream RNNSA new dataset and evaluation metric STAUC were introduced to perform online anomaly detection.
Sultani et al.
[12]
Deep multiple-instance learningA MIL algorithm with sparsity and smoothness constraints was proposed.
Harada et al.
[43]
Online growing neural gas The anomaly detection of the changing data space was realized.
Nawaratne et al. [44]Incremental
spatio-temporal learner
It updated and distinguished new abnormal and normal behaviors in real time and was used for anomaly detection and location in monitoring.
Monakhov et al. [45]GridHTMThe visual analysis method based on the extended HTM algorithm was studied, and the data were simplified by the segmentation technique.
Chriki et al.
[46]
One-class support
fvector machine
Four different features were extracted with GoogleNet, HOG, PCA-HOG and HOG3D for automated abnormal event detection to form four OCSVM models.
Bozcan et al.
[47]
GridNetAchieved anomaly detection that was insensitive to vision-related issues and proposed a novel loss function based on grid representation.
Khaire et al.
[48]
Weakly
supervised learning
CNN-BiLSTMThe proposed CNN-BiLSTM auto-encoder framework implements semi-supervised anomaly detection.
Mozaffari et al.
[49]
KNNPresenting a framework for multivariate, data-driven and sequential detection of anomalies in high-dimensional systems based on the availability of labeled data.
Majhi et al.
[50]
Dissimilarity
attention module
A dissimilarity attention module with local context information was introduced into anomaly detection for real-time applications.
Liu et al.
[51]
TPP and OALAccurate segment-level pseudo-label generation and future semantic prediction were achieved through a two-stage framework of “decoupling and parsing”.
Huang et al.
[52]
TTFA and SDFEProposing a transformer-style TTFA and a boot SDFE.

4.2. Online Anomaly Detection Based on Technical Methods

From the perspective of technical methods, online anomaly detection methods can be divided into two aspects: those based on deep learning and those based on transfer learning.

4.2.1. Online Anomaly Detection Based on Deep Learning

Due to the excellent performance of deep learning, many deep-learning-based frameworks are available, and these are compatible with many platforms to enable complex problems to be solved. The current mainstream method is to identify patterns of abnormal behavior through deep learning. As shown in Table 3, traditional online anomaly detection solutions are limited to operating on videos with only a few abnormal frames. In order to break this limitation, the perceptron proposed in [53] was optimized online to reconstruct video frames pixel by pixel from their frequency information. Based on the movement of information between adjacent frames, the incremental learner updated the parameters of the multi-layer perceptron after observing each frame, thus allowing the detection of abnormal events along the video stream. An incremental learner was also utilized to online optimize the multi-layer perceptron to produce results that detect abnormal events along the video stream. Ata-Ur-Rehman et al. [54] proposed a new particle prediction model and a weighted likelihood model to accurately detect abnormal video frames and abnormal regions in video frames. At the same time, the above two models can effectively extract and utilize information in the form of size, motion and unknown features from video frames. Currently, no fully automatic monitoring techniques can simultaneously detect and interpret abnormal events, and most require human intervention to identify the nature of abnormal events and select appropriate countermeasures. To solve this problem, in [55], a classification model was used for anomaly detection and compressed to efficiently accommodate resource-constrained devices. Additionally, abnormal activities were effectively classified using a visual transformer with a sequential learning model and attention for anomaly recognition. Due to the data-intensive end-to-end training of neural networks, the feature representation extracted from the spatio-temporal features of video was not interpretable. In [56], the authors proposed an interpretable framework with zero-shot cross-domain adaptability. The asymptotic optimality of the used sequence detector was derived in terms of minimizing the average detection delay in the maximal sense while controlling the false alarm rate. A new architecture based on deep learning was provided for the detection of violent acts in videos in [57]. The architecture consisted mainly of 3D DenseNet, a multi-head self-attention layer, and a bidirectional convolutional LSTM module. The detection of abnormal behavior was achieved by encoding the relevant spatio-temporal features. With the design of this architecture, the spatio-temporal information in the video was effectively captured, and the accuracy and reliability of abnormal behavior detection were improved. In [58], researchers proposed the first important surveillance scene dataset considering indoor and outdoor environments and adopted CNN-based optical flow estimation for the first time. It represented motion in industrial surveillance video by adding residual layer blocks encoded by a temporal optical flow feature (TOFF) and used Mask R-CNN to segment instances of important objects in the scene. To address the challenge of complex abnormal behavior analysis, a system that starts with anomaly detection was introduced in [59]. The image of the pedestrian in the video is transmitted to the abnormal behavior analysis module through TCP/IP communication. Subsequently, real-time recognition of behaviors is performed using 3DResNet, enabling the analysis and recognition of complex abnormal behaviors. In [60], researchers proposed a lightweight framework containing convolutional neural networks, trained on input frames obtained in a computationally cost-effective method, which effectively described and distinguished between normal and abnormal events. Due to the insufficient availability of real-world datasets, the first comprehensive abnormal event generation system capable of simulating typical specific abnormal events was developed in [61]. By combining 3D CNNs and non-local mechanisms, the performance of anomaly video detection on synthetic datasets was improved. A recurrent 3D GAN was proposed for adaptation to reduce the domain gap, enabling the conversion of synthetic data into realistic video sequences. In order to realize the automatic detection of abnormal behavior in an examination room, the authors [62] studied an improved algorithm based on yolov3, which used the K-means algorithm, Giouloss algorithm, Focal loss algorithm and Darknet32 algorithm. In addition, The frame-alternate dual-thread method was used to optimize the detection process. Direkoglu et al. [63] first proposed a new concept of multiplying the angle difference by the optical flow amplitude calculated in the current frame to form the motion information image (MII). The proposed motion information image system (MIIS) can well distinguish normal and abnormal events. The proposal of the MII model and its combination with a CNN was a new method for global abnormal crowd behavior detection.
The false alarm rate is a key metric for evaluating the performance of online video anomaly detection. However, there are relatively few online anomaly detection methods based on deep learning in the current methods, so it is worthwhile to establish an online anomaly detection framework from the perspective of deep learning. At the same time, the timely location of abnormal events is also the development direction of future research. This involves not only the detection and location of the abnormal event itself, but also the accurate detection of the occurrence time of the abnormal event. Therefore, one of the future research directions is to comprehensively consider the time factor to achieve the precise location of abnormal events in anomaly detection. In summary, establishing an online anomaly detection framework from the perspective of deep learning and taking the time factor into account to achieve accurate detection and localization of abnormal events is a worthy direction for future research and development.

4.2.2. Online Anomaly Detection Based on Transfer Learning

The currently proposed methods and models only perform well on some publicly available datasets but often do not perform well when used in real-world complex video surveillance. As shown in Table 3, in order to produce models with high generalization, some scholars have used transfer learning methods to improve them. In [64], the k-nearest neighbor method was employed to remove misclassified fixed objects, and then K-means clustering was used to identify potential abnormal regions. Finally, the occurrence time of the anomaly was determined by the backtracking anomaly detection algorithm, which calculated the similarity statistics. To address the reason why online decision making was neglected in the field of surveillance video, some researchers have proposed a hybrid use of transfer learning via neural networks and statistical k-nearest neighbor (KNN) decision-making methods to detect video anomalies forever and in an online manner with limited training. For video anomaly detection, the authors of several papers significantly reduced the training complexity by leveraging transfer learning, while presenting a new framework for statistical arbitrary shot sequence anomaly detection, which was able to continuously learn from very few samples. A continuous learning algorithm was also proposed, which consisted of a feature extraction module based on transfer learning and a statistical anomaly detection module that made online decisions on decision rules and continuously updated them [65,66]. To overcome the challenges of lack of modularity, cross-domain adaptability, interpretability and real-time anomaly event detection in existing video anomaly detection, the authors of [67] proposed the first multi-task learning framework capable of cross-adaptation, fewshot learning and continuous learning for limited-data video anomaly detection, while presenting the first semantic embedded-based approach for video anomaly detection using deep metric learning. In addition to a single modular framework, a plug-and-play architecture that includes transfer learning was introduced in [68]. It included an interpretable transfer learning feature extractor and a novel kNN-RNN-based sequential anomaly detector. The false alarm rate and threshold selection were analyzed mathematically. To address the common underuse of motion patterns and inconsistent datasets in current anomaly detection methods, ref. [69] proposed a system consisting of two parts: target tracking and suspicious activity. The system used ResNet tracking based on transfer learning to detect and track abnormal events in video surveillance in real time. This method avoided saturation performance when training deeper layers.
Table 3. Summary of online anomaly detection methods based on technical approaches.
Table 3. Summary of online anomaly detection methods based on technical approaches.
ReferenceMethodArchitectureRemarks
Ouyang et al.
[53]
Deep
learning
Randomly initialized multi-layer The perceptron online optimization method and incremental learner were proposed.
Ullah et al.
[55]
Vision transformerA visual transformer with a sequential learning model and an attention mechanism was used.
Doshi et al.
[56]
Transfer learningThe proposed semantic-embedding-based video anomaly detection method and transfer learning were used in different surveillance scenarios.
Rendón-Segador et al. [57]3D DenseNet, multi-head
self-attention mechanism and bidirectional convolutional LSTM
Violence in videos was detected by encoding relevant spatio-temporal features.
Ullah et al.
[58]
CNN and LSTMA surveillance scene dataset for indoor and outdoor environments was presented.
Kim et al.
[59]
KCF and 3DResNetAn architecture that can extract pedestrian information in real time and detect abnormal behavior was used.
Mehmood et al. [60]CNNProposing a lightweight framework that utilizes convolutional neural networks for training.
Lin et al.
[61]
3DCNN and C3DGANCombined with 3D CNN and non-local mechanism, the synthetic data were converted into realistic video sequences through an adaptive recurrent 3D GAN.
Fang et al.
[62]
CNNBased on the improved yolov3 algorithm, the detection optimization was carried out using the frame-alternating two-thread method.
Direkoglu et al. [63]MII and CNNA new concept of MII was formed and combined with CNN for global abnormal crowd behavior detection.
Doshi et al.
[64]
Transfer learningKNN and KMeans clusteringA framework was proposed to determine the occurrence time of exceptions.
Doshi et al.
[65]
Any-shot learningA new framework with continuous learning capability was proposed for anomaly detection of statistical arbitrary shot sequences.
Doshi et al.
[66]
Continual and few-shot learningA continual learning algorithm was proposed.
Doshi et al.
[67]
Multi-task learningProposing a multi-task learning framework and introducing a deep metric learning method based on semantic embedding.
Doshi et al.
[68]
MOVADCombining an interpretable transfer learning feature extractor with a novel kNN-RNN sequential anomaly detector.
Kale et al.
[69]
ResNet and DMLTransfer-learning-based ResNet object tracking and DML methods were proposed.
In the above research works, transfer learning was applied to online anomaly detection, but it was only limited to combining with neural networks, feature extractors and object trackers. So, combining transfer learning with more models and methods to build sustainable learning anomaly detection architectures is a focus of future research.

4.3. Other Methods

Instead of the above methods, as shown in Table 4, the use of Gaussian process regression models for anomaly detection in video surveillance was proposed by Li et al. [70]. Adaptive modeling of behavior patterns and anomaly detection in the current frame were achieved through online clustering algorithms and the utilization of supplementary information from previous frames. A cascade classifier for anomaly detection and localization was researched in [71]. It combined optical flow based on motion features and gradient-based appearance anomaly detection, thus simultaneously considering appearance and motion cues of video sequences and addressing the limitation of existing methods that only perform anomaly detection in the spatial dimension. Reference [72] introduced a method for automatic detection of suspicious behavior using CCTV video streams. The proposed method combined attention to motion amplitude and a gradient and used motion reactivity features extracted from optical flow to construct a suspicious behavior detection method based on a temporal saliency map. The effectiveness of the proposed method was verified by testing it on video clips containing suspicious behaviors. The framework of extracting a set of compact features based on foreground occupancy and optical flow information is a traditional approach for online anomaly detection [73]. This framework used an inference mechanism that evaluates a compact feature set to detect abnormal events through Gaussian mixture models, Markov chains and bags of words. However, the optical flow model thresholds and foreground occupancy model thresholds of this method were artificially set and did not automatically adapt to environmental transformations. Some researchers have proposed encoding spatio-temporal support regions with binary features instead of commonly using double precision features, combined foreground occupancy features and accurately classified events by relying on multiple low-complexity probability models to achieve online anomaly detection [74,75]. In [76], the authors combined the spatio-temporal gradient model with foreground detection to extract the significant spatio-temporal features on the foreground objects of video sequences, and then they used a sparse combinatorial learning algorithm to build an abnormal event detection model for anomaly detection. In order to be able to gradually learn new event samples from a video and detect anomalies in the video online based on the distance to the model and the relative frequency of each event class, a nonparametric hierarchical event model was proposed [77]. This method could lead to false alarms if the mesh was not properly delineated, localizing tiny human limbs or inaccurate optical flow estimation in crowded crowds. Feng et al. described a real-time online crowd behavior detection algorithm for video sequences [78]. The proposed algorithm was mainly based on a combination of visual feature extraction and image segmentation, which did not require an artificial neural network training phase to detect anomalies. The use of background subtraction in this method could make the algorithm less sensitive to rapid changes. The video abnormal event detection method based on multi-scale pyramid grid templates (MPGTs) proposed in [79] was innovative in multi-scale features, convolution operation and anomaly detection result fusion. Through the application of these methods, the accuracy and efficiency of video abnormal event detection can be improved.

4.4. Comparison

4.4.1. Comparison of Advantages and Disadvantages of Different Methods

This paper focuses on the basic principles of distinguishing normal video from abnormal video and classifies online video abnormality detection methods into supervised, unsupervised, weakly supervised, deep learning, transfer learning and traditional methods based on comparison and summary as shown in Table 5.
The advantage of supervised learning in online video anomaly detection is that it can use labeled data to improve model performance, but it does not perform well in complex scenes. Unsupervised learning methods can effectively reduce false alarms and missed detection caused by model aging and threshold aging and avoid the problem of concept drift, but their performance will degrade in complex scenarios. Semi-supervised learning and weakly supervised learning methods can allow anomaly detection in real-time scenarios, but it is difficult to solve the problem of online constraints and weakly supervised settings simultaneously, and they are computationally expensive. Deep learning methods have a powerful learning ability and can extract more representative feature information from data, but they are highly dependent on data and have high computational complexity. The detection effect is good for video data with fewer anomalies, while the performance may be reduced in the case of a small number of datasets. Transfer learning methods can significantly reduce the computational complexity of the training and detection phases, but they have certain limitations when combined with the model. The method based on optical flow features is sensitive to moving objects, can extract speed and direction information, and has a good detection effect on data videos with few people or optical flows. Although the method based on binary features can reduce the frame processing time and improve the operation speed, the feature extraction will lose some important information, resulting in poor performance. In summary, different methods have advantages and disadvantages in different scenarios, and it is necessary to choose a suitable method according to the specific situation.

4.4.2. Performance Comparison of Methods

Table 6 shows the performance metrics of these methods on representative scene datasets such as UCSD, ShanghaiTech, CUHK Avenue, UCF-crime, UMN and DOTA.
In general, the anomaly detection effect on the ShanghaiTech dataset is the best, and the anomaly detection effect on the UCSD Ped2 dataset is the second best. At the same time, the continuous learning regression method is used to obtain high detection accuracy on this dataset. This means that continuous learning techniques can help a model to continuously adapt to new data and improve the accuracy of the model when running for a long time. In addition, the better performance on the UCSD Ped1 dataset may be due to the fact that the scenes in this dataset are relatively simple and can be easily captured by the model for anomalies. In contrast, the AUC of the existing methods is generally low on the UCF-crime and DOTA datasets, probably because the scenes of these datasets are relatively complex, with more variations and noise. However, the UMN dataset is relatively less used. It is suggested that we need to consider the complexity and diversity of scenes when applying video anomaly detection methods and choose appropriate methods and datasets for research.

5. Summary and Outlook

In this paper, the research progress of online video anomaly detection technology is discussed in detail according to the method of example-based learning and technology. In addition, different datasets are presented, along with important anomaly details. At the same time, this paper also compares and analyzes the performance of the current mainstream algorithms according to various dataset evaluation criteria.
Current research on online methods in video anomaly detection faces the following main challenges and issues:
(1)
The models have a strong dependency on the task scenario and are difficult to directly transfer. The models only perform well on the datasets described in the paper, and the feature structure of the data is difficult to change due to strict requirements on some data formats (input size, color channels, text format, etc.); the trained feature extractors cannot be easily transferred to other tasks. Although these problems are mitigated by the “pre-training + fine-tuning” method, the transferability of the model needs to be further enhanced. Therefore, improving the transfer capability of models in online video anomaly detection is a problem worth investigating.
(2)
A real-time anomaly detection framework was built. Although most of the current methods can achieve high accuracy in anomaly detection, they cannot be deployed in real time. One of the key reasons is that the time cost of extracting effective features in the video is too high. From the perspective of practical applications, detecting anomalies in time and accurately can effectively reduce the loss caused by abnormal events. Therefore, it is necessary to design new methods for efficient video data preprocessing and feature extraction in the future, so as to break through the limitation of processing speed, so that more systems can be used for real-time detection scenarios.
(3)
A sustainable learning architecture for anomaly detection should be built. With prolonged monitoring, abnormal events are diverse and variable, and learned abnormal events may become normal events, making it difficult to identify abnormal behavior without expert interaction. Existing feature learning models trained offline are unable to adapt to such changes. Therefore, more online learning models with automatic update capability are needed to continuously update the anomaly detection knowledge base and improve the accuracy of detection in a continuous learning process.
(4)
The ability to locate the occurrence time of abnormal events should be enhanced. In video anomaly detection, it is necessary not only to detect and locate the abnormal event itself, but also to know the time when the anomaly occurs. The methods of temporal modeling and event analysis can be further explored to better understand the relationship and dynamic development of events in videos. This may include the use of methods such as spatio-temporal attention mechanisms, time-series models and event graphs.
(5)
Online and offline learning should be bridged with transformers. In practice, some decision tasks are not achievable outside of an online framework, and some environments are open-ended, meaning that strategies must be constantly adapted to handle tasks not seen during online interactions. Therefore, we believe that bridging online and offline learning is necessary. However, most of the current research advances based on decision transformers have focused on offline learning frameworks. Similar to offline reinforcement learning algorithms, distributional variation in online fine-tuning still exists; hence, there is a need for a special design of a decision transformer to address related issues. How to train an online decision transformer from scratch is a problem worth investigating.
(6)
A more appropriate online anomaly evaluation system should be found to make the evaluation of abnormal behavior more objective and comprehensive. The current evaluation metrics for video anomaly detection are mainly the area under the ROC curve (AUC) and the equal error rate (EER), but due to the unbalanced and heterogeneous nature of the data in the video anomaly detection datasets, using the equal error rate to evaluate the anomaly detection results will lead to large errors in practical application scenarios. The use of AUC to evaluate a model from frame-level and pixel-level criteria is also not very representative of the overall performance of the model. Therefore, it is important to establish a comprehensive and robust evaluation system for online video anomaly detection.

Author Contributions

All authors contributed to the content in this paper. Material preparation, data collection and analysis were performed by Y.Z. The first draft was written by Y.Z., and all authors commented on the previous version. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by National Natural Science Foundation of China under Grant 61976120, in part by Nanjing University State Key Laboratory for Novel Software Technology under Grant KFKT2019B15, in part by Nantong Science and Technology Program JC2021131 and in part by Postgraduate Research and Practice Innovation Program of Jiangsu Province KYCX22_3340 and College Students’ Innovation and Entrepreneurship Training Project Project No. 202310304160Y.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

There are no available data to be stated.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nayak, R.; Pati, U.C.; Das, S.K. A comprehensive review on deep learning-based methods for video anomaly detection. Image Vis. Comput. 2021, 106, 104078. [Google Scholar] [CrossRef]
  2. Si-Qi, W.A.N.G.; Jing-tao, H.U.; Guang, Y.U.; En, Z.H.U.; Zhi-Ping, C.A.I. A survey of video abnormal event detection. Comput. Eng. Sci. 2020, 42, 1393–1405. [Google Scholar]
  3. Saligrama, V.; Konrad, J.; Jodoin, P. Video Anomaly Identification. IEEE Signal Process. Mag. 2010, 27, 18–33. [Google Scholar] [CrossRef]
  4. Peipei, Z.H.O.U.; Qinghai, D.I.N.G.; Haibo, L.U.O.; Xinglin, H. Anomaly Detection and Location in Crowded Surveillance Videos. Acta Opt. Sin. 2018, 38, 0815007. [Google Scholar] [CrossRef]
  5. Patrikar, D.R.; Parate, M.R. Anomaly detection using edge computing in video surveillance system. Int. J. Multimed. Inf. Retr. 2022, 11, 85–110. [Google Scholar] [CrossRef] [PubMed]
  6. Lin, H.; Deng, J.D.; Woodford, B.J.; Shahi, A. Online weighted clustering for real-time abnormal event detection in video surveillance. In Proceedings of the 24th ACM International Conference on Multimedia, Amsterdam, The Netherlands, 1 October 2016; pp. 536–540. [Google Scholar]
  7. Mahmood, S.A.; Abid, A.M.; Lafta, S.H. Anomaly event detection and localization of video clips using global and local outliers. Indones. J. Electr. Eng. Comput. Sci. 2021, 24, 1063. [Google Scholar] [CrossRef]
  8. Bird, N.; Atev, S.; Caramelli, N.; Martin, R.; Masoud, O.; Papanikolopoulos, N. Real time, Online Detection of Abandoned Objects In Public Areas. In Proceedings of the 2006 IEEE International Conference on Robotics and Automation, Orlando, FL, USA, 15–19 May 2006; pp. 3775–3780. [Google Scholar]
  9. Ramachandra, B.; Jones, M. Street Scene: A new dataset and evaluation protocol for video anomaly detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 14–19 June 2020; pp. 2569–2578. [Google Scholar]
  10. Mahadevan, V.; Li, W.; Bhalodia, V. Anomaly Detection in Crowded Scenes. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1975–1981. [Google Scholar]
  11. Zhang, Y.; Zhou, D.; Chen, S.; Gao, S.; Ma, Y. Single-Image Crowd Counting Via Multi-Column Convolutional Neural Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June 2016; pp. 589–597. [Google Scholar]
  12. Sultani, W.; Chen, C.; Shah, M. Real-World Anomaly Detection in Surveillance Videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6479–6488. [Google Scholar]
  13. Yao, Y.; Wang, X.; Xu, M.; Pu, Z.; Wang, Y.; Atkins, E.; Crandall, D.J. DoTA: Unsupervised Detection of Traffic Anomaly in Driving Videos. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 444–459. [Google Scholar] [CrossRef]
  14. Del Giorno, A.; Bagnell, J.A.; Hebert, M. A Discriminative Framework for Anomaly Detection in Large Videos. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 334–349. [Google Scholar]
  15. Tudor Ionescu, R.; Smeureanu, S.; Alexe, B.; Popescu, M. Unmasking the Abnormal Events in Video. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 14–17 October 2017; pp. 2895–2903. [Google Scholar]
  16. Zaheer, M.Z.; Mahmood, A.; Khan, M.H.; Segu, M.; Yu, F.; Lee, S.I. Generative Cooperative Learning for Unsupervised Video Anomaly Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 14744–14754. [Google Scholar]
  17. Cho, M.; Kim, T.; Kim, W.J.; Cho, S.; Lee, S. Unsupervised video anomaly detection via normalizing flows with implicit latent features. Pattern Recognit. 2022, 129, 108703. [Google Scholar] [CrossRef]
  18. Sun, C.; Jia, Y.; Song, H.; Wu, Y. Adversarial 3d convolutional auto-encoder for abnormal event detection in videos. IEEE Trans. Multimed. 2020, 23, 3292–3305. [Google Scholar] [CrossRef]
  19. Li, G.; Cai, G.; Zeng, X.; Zhao, R. Scale-Aware Spatio-Temporal Relation Learning for Video Anomaly Detection. In Proceedings of the European Conference on Computer Vision; Springer: Cham, Switzerland, 2022; pp. 333–350. [Google Scholar]
  20. Purwanto, D.; Chen, Y.T.; Fang, W.H. Dance with Self-Attention: A New Look of Conditional Random Fields on Anomaly Detection In Videos. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 173–183. [Google Scholar]
  21. Wu, J.C.; Hsieh, H.Y.; Chen, D.J.; Fuh, F.S.; Liu, T.L. Self-supervised Sparse Representation for Video Anomaly Detection. In European Conference on Computer Vision 23 October 2022; Springer Nature: Cham, Switzerland, 2022. [Google Scholar]
  22. Wang, G.; Wang, Y.; Qin, J.; Zhang, D.; Bao, X.; Huang, D. Video Anomaly Detection by Solving Decoupled Spatio-Temporal Jigsaw Puzzles. In European Conference on Computer Vision, 23 October 2022; Springer: Cham, Switzerland, 2022; pp. 494–511. [Google Scholar]
  23. Li, C.L.; Sohn, K.; Yoon, J. Cutpaste: Self-Supervised Learning for Anomaly Detection And Localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 13–19 June 2021; pp. 9664–9674. [Google Scholar]
  24. Zhong, Y.; Chen, X.; Jiang, J.; Ren, F. A cascade reconstruction model with generalization ability evaluation for anomaly detection in videos. Pattern Recognit. 2022, 122, 108336. [Google Scholar] [CrossRef]
  25. Li, C.; Li, H.; Zhang, G. Future frame prediction based on generative assistant discriminative network for anomaly detection. Appl. Intell. 2023, 53, 542–559. [Google Scholar] [CrossRef]
  26. Chang, Y.; Tu, Z.; Xie, W.; Luo, B.; Zhang, S.; Sui, H.; Yuan, J. Video anomaly detection with spatio-temporal dissociation. Pattern Recognit. 2022, 122, 108213. [Google Scholar] [CrossRef]
  27. Chen, D.; Yue, L.; Chang, X. NM-GAN: Noise-modulated generative adversarial network for video anomaly detection. Pattern Recognit. 2021, 116, 107969. [Google Scholar] [CrossRef]
  28. Li, H.; Sun, X.; Li, C.; Shen, X.; Chen, J.; Chen, J.; Xie, Z. MPAT: Multi-path attention temporal method for video anomaly detection. Multimed. Tools Appl. 2023, 82, 12557–12575. [Google Scholar] [CrossRef]
  29. Georgescu, M.I.; Ionescu, R.T.; Khan, F.S. A background-agnostic framework with adversarial training for abnormal event detection in video. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 4505–4523. [Google Scholar] [CrossRef]
  30. Chen, C.; Xie, Y.; Lin, S.; Yao, A.; Jiang, G.; Zhang, W.; Qu, Y.; Qiao, R.; Ren, B.; Ma, L. Comprehensive Regularization in a Bi-directional Predictive Network for Video Anomaly Detection. In Proceedings of the American Association for Artificial Intelligence, Palo Alto, CA, USA, 22 February–1 March 2022; pp. 1–9. [Google Scholar]
  31. Cai, R.; Zhang, H.; Liu, W.; Gao, S.; Hao, Z. Appearance-Motion Memory Consistency Network for Video Anomaly Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, Palo Alto, CA, USA, 2–9 February 2021; Volume 35, pp. 938–946. [Google Scholar]
  32. Hao, Y.; Li, J.; Wang, N.; Wang, X.; Gao, G. Spatiotemporal consistency-enhanced network for video anomaly detection. Pattern Recognit. 2022, 121, 108232. [Google Scholar] [CrossRef]
  33. Slavic, G.; Baydoun, M.; Campo, D.; Marcenaro, L.; Regazzoni, C. Multilevel anomaly detection through variational auto-encoders and bayesian models for self-aware embodied agents. IEEE Trans. Multimed. 2021, 24, 1399–1414. [Google Scholar] [CrossRef]
  34. Behera, S.; Dogra, D.P.; Bandyopadhyay, M.K.; Roy, P.P. Understanding crowd flow patterns using active-Langevin model. Pattern Recognit. 2021, 119, 108037. [Google Scholar] [CrossRef]
  35. Doshi, K.; Yilmaz, Y. Online anomaly detection in surveillance videos with asymptotic bound on false alarm rate. Pattern Recognit. 2021, 114, 107865. [Google Scholar] [CrossRef]
  36. Aboah, A. A vision-based system for traffic anomaly detection using deep learning and decision trees. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 4207–4212. [Google Scholar]
  37. Rossi, L.; Bernuzzi, V.; Fontanini, T.; Bertozzi, M.; Prati, A. Memory-augmented Online Video Anomaly Detection. arXiv 2023, arXiv:2302.10719. [Google Scholar]
  38. Luo, W.; Liu, W.; Lian, D.; Tang, J.; Duan, L.; Peng, X.; Gao, S. Video anomaly detection with sparse coding inspired deep neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1070–1084. [Google Scholar] [CrossRef] [PubMed]
  39. Chaker, R.; Aghbari, Z.A.; Junejo, I.N. Social network model for crowd anomaly detection and localization. Pattern Recognit. 2017, 61, 266–281. [Google Scholar] [CrossRef]
  40. Dang, T.; Khattak, S.; Papachristos, C.; Alexis, K. Anomaly Detection and Cognizant Path Planning for Surveillance Operations using Aerial Robots. In Proceedings of the International Conference on Unmanned Aircraft Systems, Atlanta, GA, USA, 11–14 June 2019. [Google Scholar]
  41. Zhao, B.; Li, F.F.; Xing, E.P. Online Detection of Unusual Events In Videos Via Dynamic Sparse Coding. In Proceedings of the Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011. [Google Scholar]
  42. Bao, W.; Yu, Q.; Kong, Y. Uncertainty-Based Traffic Accident Anticipation with Spatio-Temporal Relational Learning. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12 October 2020; pp. 2682–2690. [Google Scholar]
  43. Sun, Q.; Liu, H.; Harada, T. Online growing neural gas for anomaly detection in changing surveillance scenes. Pattern Recognit. J. Pattern Recognit. Soc. 2017, 64, 187–201. [Google Scholar] [CrossRef]
  44. Nawaratne, R.; Alahakoon, D.; De Silva, D.; Yu, X. Spatiotemporal anomaly detection using deep learning for real-time video surveillance. IEEE Trans. Ind. Inform. 2019, 16, 393–402. [Google Scholar] [CrossRef]
  45. Monakhov, V.; Thambawita, V.; Halvorsen, P.; Riegler, M.A. GridHTM: Grid-Based Hierarchical Temporal Memory for Anomaly Detection in Videos. Sensors 2023, 23, 2087. [Google Scholar] [CrossRef] [PubMed]
  46. Chriki, A.; Touati, H.; Snoussi, H.; Kamoun, F. Deep learning and handcrafted features for one-class anomaly detection in UAV video. Multimed. Tools Appl. 2021, 80, 2599–2620. [Google Scholar] [CrossRef]
  47. Bozcan, I.; Le Fevre, J.; Pham, H.X.; Kayacan, E. Gridnet: Image-agnostic conditional anomaly detection for indoor surveillance. IEEE Robot. Autom. Lett. 2021, 6, 1638–1645. [Google Scholar] [CrossRef]
  48. Khaire, P.; Kumar, P. A semi-supervised deep learning based video anomaly detection framework using RGB-D for surveillance of real-world critical environments. Forensic Sci. Int. Digit. Investig. 2022, 40, 301346. [Google Scholar] [CrossRef]
  49. Mozaffari, M.; Doshi, K.; Yilmaz, Y. Online multivariate anomaly detection and localization for high-dimensional settings. Sensors 2022, 22, 8264. [Google Scholar] [CrossRef]
  50. Majhi, S.; Das, S.; Brémond, F. DAM: Dissimilarity Attention Module for Weakly-supervised Video Anomaly Detection. In Proceedings of the 17th IEEE International Conference on Advanced Video and Signal Based Surveillance, Washington, DC, USA, 16–19 November 2021; pp. 1–8. [Google Scholar]
  51. Liu, T.; Zhang, C.; Lam, K.M.; Kong, J. Decouple and Resolve: Transformer-Based Models for Online Anomaly Detection from Weakly Labeled Videos. IEEE Trans. Inf. Forensics Secur. 2022, 18, 15–28. [Google Scholar] [CrossRef]
  52. Huang, C.; Liu, C.; Wen, J.; Wu, L.; Xu, Y.; Jiang, Q.; Wang, Y. Weakly Supervised Video Anomaly Detection via Self-Guided Temporal Discriminative Transformer. IEEE Trans. Cybern. 2022, 29, 1–14. [Google Scholar] [CrossRef]
  53. Ouyang, Y.; Shen, G.; Sanchez, V. Look at adjacent frames: Video anomaly detection without offline training. In Proceedings of the Computer Vision-ECCV 2022 Workshops, Tel Aviv, Israel, 23–27 October 2022; Part V. Springer Nature: Cham, Switzerland, 2023; pp. 642–658. [Google Scholar]
  54. Tariq, S.; Farooq, H.; Jaleel, A.; Wasif, S.M. Anomaly detection with particle filtering for online video surveillance. IEEE Access 2021, 9, 19457–19468. [Google Scholar]
  55. Ullah, W.; Hussain, T.; Baik, S.W. Vision transformer attention with multi-reservoir echo state network for anomaly recognition. Inf. Process. Manag. 2023, 60, 103289. [Google Scholar] [CrossRef]
  56. Doshi, K.; Yilmaz, Y. Towards Interpretable Video Anomaly Detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 2655–2664. [Google Scholar]
  57. Rendón-Segador, F.J.; Álvarez-García, J.A.; Enríquez, F.; Deniz, O. Violencenet: Dense multi-head self-attention with bidirectional convolutional lstm for detecting violence. Electronics 2021, 10, 1601. [Google Scholar] [CrossRef]
  58. Ullah, F.U.M.; Obaidat, M.S.; Muhammad, K.; Ullah, A.; Baik, S.W.; Cuzzolin, F.; Rodrigues, J.J.P.C. An intelligent system for complex violence pattern analysis and detection. Int. J. Intell. Syst. 2022, 37, 10400–10422. [Google Scholar] [CrossRef]
  59. Kim, D.; Kim, H.; Mok, Y.; Paik, J. Real-time surveillance system for analyzing abnormal behavior of pedestrians. Appl. Sci. 2021, 11, 6153. [Google Scholar] [CrossRef]
  60. Mehmood, A. LightAnomalyNet: A Lightweight Framework for Efficient Abnormal Behavior Detection. Sensors 2021, 21, 8501. [Google Scholar] [CrossRef]
  61. Lin, W.; Gao, J.; Wang, Q.; Liu, X. Learning to detect anomaly events in crowd scenes from synthetic data. Neurocomputing 2021, 436, 248–259. [Google Scholar] [CrossRef]
  62. Fang, M.; Chen, Z.; Przystupa, K.; Li, T.; Majka, M.; Kochan, O. Examination of abnormal behavior detection based on improved YOLOv3. Electronics 2021, 10, 197. [Google Scholar] [CrossRef]
  63. Direkoglu, C. Abnormal crowd behavior detection using motion information images and convolutional neural networks. IEEE Access 2020, 8, 80408–80416. [Google Scholar] [CrossRef]
  64. Doshi, K.; Yilmaz, Y. Fast Unsupervised Anomaly Detection in Traffic Videos. In Proceedings of the Conference on Computer Vision and Pattern Recognition Workshops IEEE, Virtual, 14–19 June 2020. [Google Scholar]
  65. Doshi, K.; Yilmaz, Y. Any-shot Sequential Anomaly Detection in Surveillance Videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Virtual, 14–19 June 2020; pp. 934–935. [Google Scholar]
  66. Doshi, K.; Yilmaz, Y. Continual Learning for Anomaly Detection In Surveillance Videos. In Proceedings of the IEEE/CVF Conference on Computer Vision And Pattern Recognition Workshops, Virtual, 14–19 June 2020; pp. 254–255. [Google Scholar]
  67. Doshi, K.; Yilmaz, Y. Multi-Task Learning for Video Surveillance with Limited Data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3889–3899. [Google Scholar]
  68. Doshi, K.; Yilmaz, Y. A Modular and Unified Framework for Detecting and Localizing Video Anomalies. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 3982–3991. [Google Scholar]
  69. Kale, S.; Shriram, R. Suspicious activity detection using transfer learning based resnet tracking from surveillance videos. In Proceedings of the 12th International Conference on Soft Computing and Pattern Recognition, online, India, 15–18 December 2020; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 208–220. [Google Scholar]
  70. Li, N.; Wu, X.; Guo, H.; Xu, D.; Ou, Y.; Chen, Y. Anomaly detection in video surveillance via gaussian process. Int. J. Pattern Recognit. Artif. Intell. 2015, 29, 1555011. [Google Scholar] [CrossRef]
  71. Li, N.; Chang, F.; Liu, C. Spatial-temporal cascade autoencoder for video anomaly detection in crowded scenes. IEEE Trans. Multimed. 2020, 23, 203–215. [Google Scholar] [CrossRef]
  72. Cheoi, K.J. Temporal saliency-based suspicious behavior pattern detection. Appl. Sci. 2020, 10, 1020. [Google Scholar] [CrossRef]
  73. Leyva, R.; Sanchez, V.; Li, C.T. Video anomaly detection with compact feature sets for online performance. IEEE Trans. Image Process. 2017, 26, 3463–3478. [Google Scholar] [CrossRef] [PubMed]
  74. Leyva, R.; Sanchez, V.; Li, C.T. Abnormal Event Detection in Videos Using Binary Features. In Proceedings of the 40th International Conference on Telecommunications and Signal Processing, Barcelona, Spain, 5–7 July 2017; pp. 621–625. [Google Scholar]
  75. Leyva, R.; Sanchez, V.; Li, C.T. Fast Detection of Abnormal Events In Videos With Binary Features. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Calgary, AB, Canada, 15–20 April 2018; pp. 1318–1322. [Google Scholar]
  76. Geng, Y.; Du, J.; Liang, M. Abnormal event detection in tourism video based on salient spatio-temporal features and sparse combination learning. World Wide Web 2019, 22, 689–715. [Google Scholar] [CrossRef]
  77. Feng, J.; Zhang, C.; Hao, P. Online Anomaly Detection in Videos by Clustering Dynamic Exemplars. In Proceedings of the 19th IEEE International Conference on Image Processing, Orlando, FL, USA, 30 September–3 October 2012; pp. 3097–3100. [Google Scholar]
  78. Pennisi, A.; Bloisi, D.D.; Iocchi, L. Online real-time crowd behavior detection in video sequences. Comput. Vis. Image Underst. 2016, 144, 166–176. [Google Scholar] [CrossRef]
  79. Li, S.; Cheng, Y.; Zhao, L.; Wang, Y. Anomaly detection with multi-scale pyramid grid templates. Multimed. Tools Appl. 2023, 30, 1–19. [Google Scholar] [CrossRef]
Figure 1. Anomaly detection in video surveillance scenes. (a) A truck moving on the footpath (UCSD Dataset). (b) Pedestrian walking on a lawn (UCSD Dataset). (c) A person throwing an object (Avenue). (d) A person carrying a suspicious bag (Avenue). (e) Incorrect parking of vehicle (MDVD). (f) People fighting (MDVD). (g) A person catching a bag (ShanghaiTech). (h) A vehicle moving on the footpath (ShanghaiTech) [5].
Figure 1. Anomaly detection in video surveillance scenes. (a) A truck moving on the footpath (UCSD Dataset). (b) Pedestrian walking on a lawn (UCSD Dataset). (c) A person throwing an object (Avenue). (d) A person carrying a suspicious bag (Avenue). (e) Incorrect parking of vehicle (MDVD). (f) People fighting (MDVD). (g) A person catching a bag (ShanghaiTech). (h) A vehicle moving on the footpath (ShanghaiTech) [5].
Sensors 23 07442 g001
Figure 2. Video anomaly detection methods.
Figure 2. Video anomaly detection methods.
Sensors 23 07442 g002
Table 1. Anomaly detection datasets for common scenarios.
Table 1. Anomaly detection datasets for common scenarios.
DatasetsLabelingResolutionType of ExceptionDuration
UMNFrame-Level320 × 240Escape, panic27 min
UCSD Ped1Pixel-Level and
Frame-Level
238 × 158Cycling, small vehicles10 min
UCSD Ped2Pixel-Level and
Frame-Level
240 × 360Cycling, small vehicles10 min
CUHK AvenueFrame-Level640 × 360Running, throwing objects, or walking in the wrong direction30 min
ShanghaiTechFrame-Level856 × 480Cycling, pushing prams, climbing over railings, running3.01 h
UCF-CrimeFrame-Level320 × 240Abuse, arrest, arson, assault, road accidents, burglary, explosions, fighting, robbery, shooting, theft, shoplifting and vandalism128 h
DOTAFrame-Level1280 × 720Pedestrians, cars, trucks, buses, motorbikes, bicycles and riders35 min
Table 4. Summary of other methods.
Table 4. Summary of other methods.
ReferenceMethodArchitectureRemarks
Li et al.
[70]
Other methodsOnline clustering algorithmAn online clustering algorithm was used to construct the basic behavior model.
Li et al.
[71]
ST-CAAEA cuboid-patch-based spatio-temporal cascading auto-encoder (ST-CAAE) was proposed to improve the speed of anomaly detection in the next stage.
Cheoi et al.
[72]
Optical flow and gradient featuresA method for automatic detection of suspicious behavior using CCTV video streams was investigated.
Leyva et al.
[73]
Foreground occupancy and optical flow functionsThe joint response of the local spatio-temporal neighborhood model was considered to improve the detection accuracy.
Leyva et al.
[74]
Binary featuresBinary features were introduced to detect abnormal events in videos.
Leyva et al.
[75]
Binary featuresIt used binary features to encode motion information and for low-complexity probabilistic model detection.
Geng et al.
[76]
Sparse ensemble learning algorithmsReal-time abnormal event detection was realized by a sparse combination learning algorithm.
Feng et al.
[77]
Incremental learning machineA hierarchical event model for online anomaly detection in surveillance video was proposed.
Pennisi et al.
[78]
Visual feature extraction and Image segmentationA statistical analysis method combining feature detection and image segmentation was proposed.
Li et al.
[79]
Particle filteringA new particle prediction model and a weighted likelihood model were proposed.
Table 5. Advantages and disadvantages of different types of abnormal event detection methods.
Table 5. Advantages and disadvantages of different types of abnormal event detection methods.
Method
Categories
Basis of JudgementAdvantagesDisadvantages
Methods based
on
sample
annotation
Supervised
learning
The optimal model is trained by training samples and applied to new data and output results so that the model has predictive ability.Effectively use the information of data annotation to improve the prediction performance of the model.Exception types are more complex, with poor performance and longer latency in complex scenarios.
Unsupervised
learning
The unlabeled training dataset needs to be analyzed to understand the statistical regularities between samples.It reduces false alarms and missed detections caused by model threshold aging and avoids concept drift in unstable environments.Lack of ability to automatically update and distinguish anomalies online based on scene changes.
Weakly
supervised
learning
Incomplete, inexact or imprecise labeling information is called weak supervision and is widely used in label learning.Real-time online detection of anomalies without additional buffer time.The single-stage framework is not suitable for solving multi-stage problems and is computationally expensive.
Methods
based
on
technical
approach
Deep
learning
A deep neural network is constructed, and a large number of sample data are used as input to obtain the model and learn its internal laws and representation levels.Good portability is directly proportional to the amount of data, and deep learning is highly data-dependent.There are challenges and high computational complexity in dealing with online learning, noise and concept drift.
Transfer
learning
Using the similarity between data, tasks or models, a model that has learned in an old domain is applied to a new domain to solve a new problem.The computational complexity of the training and detection phases can be significantly reduced.The combination with the model has certain limitations.
Other
methods
Optical flow
features
When the object is moving, the brightness pattern of its corresponding point on the image is also moving.Optical flow expresses the information of image changes and object motion, which is suitable for detecting abnormal behavior activities in videos.When changes in optical flow are not obvious, abnormal behavior cannot be accurately detected.
Binary
features
A number of point pairs are randomly selected around a feature point, and their gray values are combined into a binary string, which is used as the feature descriptor of the feature point.Frame processing time can be reduced, greatly increasing the speed of computing.Some important feature information can be lost, resulting in a loss of accuracy.
Table 6. Performance comparison of different online video anomaly detection methods.
Table 6. Performance comparison of different online video anomaly detection methods.
Author/YearModel StructuresFrame Levels AUC/EER/%
UCSD
PED1
UCSD
PED2
ShanghaiTechCUHK AvenueUCF-CrimeUMNDOTA
Pennisi [78]/2016Statistic Analysis-/--/--/--/--/-95.0/--/-
Chaker [39]/2017SNN-/-87.9/--/--/--/--/--/-
Sun [43]/2017Online GNG93.75/-94.09/--/--/--/--/--/-
Leyva [75]/2017Binary Features-/--/--/--/--/-88.3/--/-
Sultani [12]/2018MIL-/--/--/-75.41/--/--/--/-
Leyva [76]/2018Binary Features-/48.1-/38.4-/--/--/--/--/-
Doshi [65]/2020Continual Learning-/-97.8/-71.62/-86.4/--/--/--/-
Li [71]/2020ST-CAAE-/-87.1/--/--/--/--/--/-
Doshi [35]/2021GAN-/-97.2/-70.9/--/--/--/--/-
Majhi [50]/2021Cov-DAM-/--/-88.22/--/-82.67/--/--/-
Dist-DAM-/--/-88.86/--/-82.57/--/--/-
Huang [52]/2022TTFA and SDFE-/--/-98.06/--/-84.04/--/--/-
Nawaratne [44]/2022ISTL75.2/29.891.1/8.9-/-79.8/29.2-/--/--/-
Liu [51]/2022Hybrid
Conv-Transformer
-/--/--/--/-85.18/--/--/-
Doshi [67]/2022DNN and
Continual Learning
-/-95.6/-70.12/-86.4/--/--/--/-
Doshi [68]/2022Continual Learning-/-97.2/-73.62/-88.7/--/--/--/-
Rossi [37]/2023VST-/--/--/-82.11/--/--/-82.1/-
Yao [13]/2023FOL-Ensemble-/--/--/--/--/--/-73.0/-
Ouyang [53]/2023MLP-/-96.5/-83.1/-90.2/--/--/--/-
Ullah [55]/20233D-CNN-/--/--/--/-51.0/--/--/-
Doshi [56]/2023Continual Learning-/--/-68.9/-79.0/--/--/--/-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, Y.; Song, J.; Jiang, Y.; Li, H. Online Video Anomaly Detection. Sensors 2023, 23, 7442. https://doi.org/10.3390/s23177442

AMA Style

Zhang Y, Song J, Jiang Y, Li H. Online Video Anomaly Detection. Sensors. 2023; 23(17):7442. https://doi.org/10.3390/s23177442

Chicago/Turabian Style

Zhang, Yuxing, Jinchen Song, Yuehan Jiang, and Hongjun Li. 2023. "Online Video Anomaly Detection" Sensors 23, no. 17: 7442. https://doi.org/10.3390/s23177442

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop