Next Article in Journal
AI-Based Positioning with Input Parameter Optimization in Indoor VLC Environments
Previous Article in Journal
Experimental Research on a Hybrid Algorithm for Localisation and Reconstruction of the Impact Force Applied to a Rectangular Steel Plate Structure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Monocular Camera Viewpoint-Invariant Vehicular Traffic Segmentation and Classification Utilizing Small Datasets

1
Engineering Mathematics Department, Alexandria University, Lotfy El-Sied st. off Gamal Abd El-Naser, Alexandria 11432, Egypt
2
Electrical Engineering Department, University of Business and Technology, Ar Rawdah, Jeddah 23435, Saudi Arabia
3
Electrical Engineering Department, Old Dominion University, 5115 Terminal Blvd, Norfolk, VA 23529, USA
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(21), 8121; https://doi.org/10.3390/s22218121
Submission received: 31 August 2022 / Revised: 20 October 2022 / Accepted: 21 October 2022 / Published: 24 October 2022
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)

Abstract

:
The work presented here develops a computer vision framework that is view angle independent for vehicle segmentation and classification from roadway traffic systems installed by the Virginia Department of Transportation (VDOT). An automated technique for extracting a region of interest is discussed to speed up the processing. The VDOT traffic videos are analyzed for vehicle segmentation using an improved robust low-rank matrix decomposition technique. It presents a new and effective thresholding method that improves segmentation accuracy and simultaneously speeds up the segmentation processing. Size and shape physical descriptors from morphological properties and textural features from the Histogram of Oriented Gradients (HOG) are extracted from the segmented traffic. Furthermore, a multi-class support vector machine classifier is employed to categorize different traffic vehicle types, including passenger cars, passenger trucks, motorcycles, buses, and small and large utility trucks. It handles multiple vehicle detections through an iterative k-means clustering over-segmentation process. The proposed algorithm reduced the processed data by an average of 40%. Compared to recent techniques, it showed an average improvement of 15% in segmentation accuracy, and it is 55% faster than the compared segmentation techniques on average. Moreover, a comparative analysis of 23 different deep learning architectures is presented. The resulting algorithm outperformed the compared deep learning algorithms for the quality of vehicle classification accuracy. Furthermore, the timing analysis showed that it could operate in real-time scenarios.

1. Introduction

Computer vision methods may offer more practical perceptual, contextual, and situational awareness to a highway intelligent transportation system (ITS) than any other types of sensors [1]. Utilizing vision sensors in the development of the ITS focuses on challenges that are generally difficult to solve using other sensor approaches [2]. For example, vehicles come in various sizes, shapes, and colors within the same class. Additionally, the projected shape of the vehicle differs due to pose variations concerning the camera sensor. All the previously mentioned points make the problem of vehicle detection and classification very challenging. Another complexity in ITS arises due to highway traffic outdoor scenes with varying visibility and lighting conditions. Furthermore, limited computing power may reduce the efficiency of an integrated system to perform in real time. Parallel computing techniques have decreased this limitation drastically. Finally, accounting for the unpredictable movement of vehicles through a highway traffic scene may be difficult. Off-road cameras are currently used by most state departments of transportation to monitor traffic conditions and to identify situations that disrupt traffic flow, such as road debris, stopped vehicles, or accidents. Computer vision algorithms allow these tasks to be performed automatically and efficiently [3].
Tremendous efforts to tackle different challenging aspects have been investigated deeply in the literature. While there are several other non-vision sensor-based approaches, such as the one presented in [4], the authors in this work focused on computer vision and machine-learning-related techniques utilizing monocular optical sensors.
The authors in [5] proposed a pipeline for detecting vehicles from monocular videos based on transfer learning. They weakly calibrated the camera using a 3 × 3 transformation from the image domain to the real-world domain to estimate the vehicle length and speed. Additionally, the authors adopted three popular object detection approaches, namely: SSD [6], YOLOV2 [7], and faster R-CNN [8] to classify traffic vehicles. The authors tested their approach on videos captured from different sites and demonstrated that they had achieved high accuracy and fast processing speed. The research in [9] proposed a semi-supervised vehicle type classification approach based on a broad learning scheme (BLS) [10] to reduce the cost of the training phase, and a dynamic ensemble structure is used to estimate the class type probabilities. Naghmeh et al. [11] proposed a semi-supervised Fuzzy C-Means clustering to predict different vehicle types and labels. They utilized unlabeled and labeled data to extract useful information for classifying the vehicles. Additionally, random oversampling was used to handle the multi-class imbalanced datasets. Semi-supervised principal component analysis convolutional network (PCN) was adopted in [12]. The authors generated a convolutional filter bank to leverage the effectiveness of traditional convolution neural networks against different image schemes, including translation, scale, noise, and rotations. The research incorporated PCN into softmax and support vector machine (SVM) classifiers to evaluate different structures.
Nadiya et al. [13] proposed a monocular vision-based technique to detect vehicles for automatic toll collection. Their technique suggests using a group of convolutional neural networks to estimate the probability for vehicles class and feed them into a gradient-boosting classifier to complement the labels obtained from optical sensors. The research in [14] introduced an automated approach for different vehicle classifications from rear-view images by fusing physical and spatial attributes. The physical features include height from the ground to the rear bumper and the distance from the bumper to the license plate. The spatial features are extracted by applying a convolutional neural network. The SVM classifier is employed to classify fused features.
Unsupervised Feature Learning was discussed and elaborated by Amir et al. [15]. The approach depends upon the generation of a dictionary from a dense scale-invariant feature extractor [16]. Bases are generated by applying the k-means clustering technique. The generated dictionary maps the input features to a new learned space by employing a coding vector with the trained bases. In classifying images, they utilized the spatial pyramid matching [17] and SVM classifier. Zhanyu et al. [18] introduced a fine-grained classification scheme based on convolutional neural networks. They added a new max-pooling layer between the fully connected and traditional layers. The proposed layer reduces the dimensions of feature space, and the computational cost is reduced.
Zhang et al. [19] utilized single-shot multi-box feature detectors based on deconvolution and pooling to classify vehicles using convolutional neural networks. The authors in [20] compared the performance of vehicle detection and classification using the mixture of Gaussian (MoG) background subtraction algorithm along with a support vector machine (SVM) vehicle classification model against the faster RCNN. Their experiments showed that faster R-CNN outperforms the SVM method. Xiang et al. [21] proposed a method to count vehicles based on a moving-object detector that works with static and moving backgrounds. A pixel-level detector handles static objects, while camera registration is employed with moving backgrounds. In their approach, an online learning tracker counts vehicles under different situations. The authors in [22] introduced an approach to identify vehicle make and model from front-facing images utilizing physical and visual characteristics. First, they extracted the logo and fed it into a classifier to identify the vehicle make. Then, they customized a hierarchical classifier to identify the vehicle type.
Farahani [23] utilized the Gaussian mixture model for background segmentation, and the extended mean shift (EMS) algorithm based on color information was utilized for tracking. The EMS algorithm was adapted by applying an Epanechnikov kernel estimator function on the feature space. This feature space was derived from the color histogram. Roy et al. [24] extracted foreground objects using GMM and proposed a new weighted mean approach to track vehicles. Das et al. [25] classified vehicles into five classes using a bag of Speeded Up Robust Features (SURF). An SVM classifier was adopted in their work. The authors in [26] suggested two-stage methods to classify vehicles. Firstly, they balanced the data samples by augmenting the acquired data. Secondly, the augmented data was classified utilizing an ensemble of convolutional neural networks (CNN) with different architectures and learned on the augmented training dataset. Cai et al. [27] pioneered a scene-adjusted vehicle detection approach based on the concept of the bagging (bootstrap aggregating) technique. A composite deep-structure classifier was built using multiple classifiers of samples generated in the target scene based on a confidence score with a voting mechanism. The authors in [28] adopted latent SVM in a vehicle make and model recognition. A novel greedy parts localization algorithm was employed to extract some descriptive parts in the vehicles used in the learning stages. Chavez-Garcia et al. [29] located vehicles within a video frame using a two-dimensional Bayesian occupancy grid map. They experimented with a sparse version of the Histogram of Oriented Gradients (HOG) along with the Adaboost classifier in the classification stage. The authors in [30] detected vehicle candidates utilizing the active basis model (ABM) based on Gabor wavelet elements. A random forest classifier is trained to classify vehicles into three categories based on the vehicle length extracted from the spatial domain image and the gray-level co-occurrence matrix of the detected vehicle image. Siddiqui et al. [31] utilized a bag of SURF features of either the front- or rear-facing images of vehicles in conjunction with multi-class support vector machine (SVM) classifier and attribute-bagging-based ensemble of SVM (AB-SVM) to recognize vehicle make and model.
Toropov et al. [32] employed adaptive GMM for segmenting vehicles and Viola–Jones cascade detector based on Haar features for vehicle tracking. Vehicles are counted using a probabilistic counting model. The authors in [33] provided a method for detecting and tracking vehicles in videos captured from roadside cameras. Authors in [34] proposed an integrated work of vehicle detection, tracking, and classification for purposes of emission estimation. Similarly, the work in [35] provides a detection and tracking algorithm for on-road vehicles using a car-mounted camera. Vehicle detection and tracking are extended to include vehicle counting in [36] and vehicle lane assignment and analysis in [37]. The authors in [38,39] used classifiers to distinguish between moving objects that are considered cars as opposed to another type of vehicle. Similarly, classification was used to distinguish trucks from other highway traffic for a study on a truck and industrial vehicle density in [40]. Techniques that apply multiclass classification have also been used to distinguish between vehicle categories on the highway, such as cars, buses, motorcycles, and trucks [41,42,43]. The work in [44] distinguishes multiple vehicles by size to assign correct toll rates.
When analyzing the previously discussed efforts, it is clear that a trend of strategies for completing a particular purpose has been developed. These efforts can be categorized into detection approaches, known as foreground segmentation, tracking algorithms, and (if applicable) feature extraction and vehicle categorization methods. Furthermore, some of the previously summarized approaches examined techniques utilizing enhancements in ROI extraction and camera alignment. While deep learning (DL) is a fast-growing area of research in vehicle segmentation for intelligent transportation systems (ITS), these methods need a lot of data to train. Thus, they cannot function properly without huge training data and imbalanced datasets. Additionally, optimizing DL architecture parameters requires plenty of experiments using deep learning networks and architecture combinations. The dataset utilized in this work is small and is not enough to adopt a deep learning approach [15,45,46,47,48,49,50]. The results section compares the proposed technique for verification and performance evaluation against recent deep learning architectures with different deep neural networks.
In this work, the authors develop and implement a monocular camera viewpoint-independent and fully integrated method for vehicular traffic segmentation and classification. Key results reported in this work are evaluated using networked traffic camera videos obtained from the Virginia Department of Transportation (VDOT). More specifically, the following contributions are presented: (1) an automated region of interest (ROI) extraction approach; (2) an improved real-time foreground segmentation algorithm; (3) a classification scheme that is capable of sensing multiple vehicle clusters for handling occluded vehicles in dense traffic scenarios; and (4) performance evaluation of twenty-three object detection deep learning architectures that have been trained and adapted to the evaluated dataset employing transfer learning.
The rest of this paper is organized as follows: Section 2 details the proposed system pipeline. Then, Section 3 presents the results of the developed system where the performance of different system stages is evaluated to accurately segment individual vehicles and to classify the vehicles by type. Finally, in Section 4, concluding statements are discussed.

2. Proposed Technique

Deep learning approaches combine both feature extraction and classification into a single procedure. They require massive training datasets to achieve high accuracy. Unlike deep learning approaches, traditional supervised machine-learning approaches have reported high accuracy in object detection and classification with limited and small amounts of data. These approaches require two substages: one for feature extraction and the other one for object classification. The histogram of oriented gradients approach has been widely used in object detection based on the gradients of visual and textural differences with reported high detection accuracy. Additionally, morphological, size, and shape features are utilized to separate various classes representing different objects. The proposed technique utilized the support vector machine classifier since it has been known for high classification accuracy in similar domains and requires few training samples compared to deep learning approaches.
This section develops methods to solve the primary stated goal of autonomous vehicle segmentation and type classification in a traffic video database. The developed pipeline consists of the following steps: (1) automated development of an ROI; (2) foreground object segmentation using improved robust low-rank and sparse matrix decomposition; (3) extraction of vehicle size, shape, and texture features using morphological properties and histogram of oriented gradients (HOG); (4) vehicle tracking using the extended Kalman filter (EKF) method; (5) vehicle classification using a multiclass support vector machine (SVM) classifier; and finally, (6) multiple vehicle handling using a K-means over-segmentation and a two-pass classification scheme. The proposed pipeline is shown in Figure 1, and the following subsections detail its steps.

2.1. Region of Interest Extraction

The proposed algorithm’s first step establishes an ROI in the image scene to reduce the computational complexity, accelerate processing, and retain the most visual features of the captured frames. Most practical traffic surveillance cameras are set at a viewing angle to monitor a large section of the highway. As vehicles progress through the image scene toward the horizon and move farther from the camera, the vehicle size becomes smaller and can appear as image noise. Likewise, as vehicles become smaller, there are not enough meaningful details to extract features for classification purposes, and detected vehicles at a far distance can interfere as outliers for training a classifier. In that regard, the processed frame is clipped to 70% of its height, considering the camera side. In this proposed vehicle detection algorithm, monitoring is optimized to only handle vehicles moving along one major angle. An automated ROI extraction approach is introduced to define the effective part of the image scene in which vehicles can be efficiently segmented and classified. The ROI is a binary mask applied to each video frame to filter pixel intensities outside the ROI and set them to zero.
The proposed extraction of ROI includes a sequence of six steps, namely: (1) noise suppression, (2) edge detection, (3) Hough transform, (4) fuzzy C-means, (5) curve fitting, and (6) mask development. The pipeline for the ROI extraction approach is depicted in Figure 2.
Since the captured frames are noisy and degraded and have artifacts due to environmental surroundings, a Gaussian kernel is utilized to suppress these artifacts and to keep strong edges representing potential highway lane sides. A canny edge detector is employed to detect candidate edges from blurred images, followed by connected components analysis to suppress unwanted edges in small areas. Then, the Hough transform is applied to detect possible lines in the edge image. Vertical lines whose angles are less than 2 o are suppressed. It is assumed that the possible lanes will not be vertical lines since these vertical lines may account for other objects within the scene, such as light poles. Applying the Hough transform to the image results in a feature space of two parameters, namely, ρ and θ , where ρ represents the normal from the origin to the detected lines, while θ represents the inclination to x-axis. Next, fuzzy C-mean is applied to the Hough space features to cluster the possible detected Hough lines using the two features, ρ and θ . The number of clusters is set empirically to four clusters. Since the detected clustered lines may be represented as broken segments near each other, curve fitting is employed to fit a cluster of detected broken segments into a unique single line. Several 1D curve fitting techniques are tested, including linear and higher order polynomials, exponential, Gaussian, and cubic spline models. Empirically, it is found that the linear polynomial model provides the best results. Finally, a raster scan is utilized across the image overlaid by fitted curves. These curves are used as guidelines to define the masked areas.

2.2. Foreground Segmentation

The foreground segmentation process identifies moving vehicles through the highway image scene. This stage is a crucial step in the proposed algorithm since it is required to extract vehicles with high accuracy and low computational time to make the system suitable for real-time implementations. Plenty of approaches have been developed in the literature related to foreground segmentation. Of these approaches are the robust low-rank and sparse matrix decomposition in which a low-rank matrix represents the background image, and a sparse matrix represents the foreground objects [51]. These techniques prove to achieve high accuracy and real-time implementations. Sobral et al. [52] have reported several low-rank decomposition algorithms for background extraction from videos, and they ranked those algorithms in terms of their computational time.
In this research, the authors first investigated several approaches and applied them to the available dataset, which is noisy, degraded highway videos captured under different weather conditions. The evaluation is in terms of segmentation efficiency and computational time. Secondly, the proposed algorithm improves upon one of the evaluated algorithms named fast principal component pursuit (FPCP) introduced by Rodriguez and Wohlberg [53]. The proposed approach improves the speed and segmentation accuracy and verifies the improved performance using a manually labeled dataset. The proposed algorithm utilizes the FPCP approach as an initial stage to pre-train the system. The pre-training provides a priori low-rank background representation that is used in subsequent stages. Once a low-rank representation is determined, a simple frame difference is utilized to extract the sparse foreground images. Another aspect is that FPCP and other robust low-rank and sparse matrix decomposition approaches utilize global thresholding to extract foreground objects. However, global thresholding extracts most of the segmented objects and labels them as foreground objects; it mislabels other foreground objects, especially when the pixel values of these objects are close to background pixel intensities. Additionally, frame difference provides negative values considered background objects when global thresholding is applied; however, they are not. Figure 3 depicts the histogram of sparse frames that results from the difference between estimated low-rank background frames and a frame under processing. This gives intuition to fit the resultant sparse matrix into a Gaussian distribution with a mean representing the background pixel values. If values fall in the Gaussian distribution outer sides, they are mapped to the segmented foreground pixels.
In that sense, the proposed algorithm models the resultant difference matrix into a normal distribution and determines the threshold utilizing the fitted normal distribution mean and variance. The estimated mean is around or close to zero values. Those values represent the background pixel intensities. Considering that noises are present in the scenes, the background pixel values are mapped inside the interval defined by the bounds μ ± λ σ . In contrast, foreground pixel values are mapped outside that interval.
Following the same FPCP formulation, a data matrix X can be written as X = L + Z such that L is a low-rank matrix, and Z is a sparse matrix satisfying the following optimization problem:
arg min L , Z L + λ Z 1 s . t . X = L + Z
where X R m × n is the data matrix, is the nuclear norm, while 1 is the l 1 norm of a given matrix. The data matrix X is constructed by stacking the video frames shaped as columns of X. The FPCP authors relaxed the nuclear norm, and the optimization function became:
arg min L , Z 1 2 L + Z X F + λ Z 1 s . t . rank ( L ) = t
where F is the Frobenius norm of the matrix, and t represents the number of partially selected components of the spectral value decomposition (SVD). The FPCP approach proposed a solution to Equation (2) using alternating minimization as follows:
L k + 1 = arg min L L + Z k X F s . t . rank ( L ) = t
Z k + 1 = arg min L L k + 1 + Z X F + λ Z 1
Equation (3) is addressed utilizing partial t components of the SVD of X Z , while Equation (4) is solved by soft thresholding shrink ( X L k + 1 , λ ) where the soft thresholding is defined by:
shrink ( x , ϵ ) = sign ( x ) max { 0 , | x | ϵ }
The videos’ set of k frames are used to pre-train the system to estimate a low-rank matrix L ˜ that represents the background. A simple frame difference between the estimated background L ˜ and the current video frame F ( i ) is used to determine the current spare matrix that represents the foreground, Z ( i ) , as given by:
Z ( i ) = F ( i ) L ˜
Once Z ( i ) is computed, resultant difference values are fitted into a normal distribution of mean μ and standard deviation σ . The minimum variance unbiased estimator (MVUE) is utilized to estimate the parameters μ and σ . The sample mean, x ¯ , is defined as
x ¯ = i = 1 n x i n
while the sample variance, s 2 , is defined as
s 2 = 1 n 1 i = 1 n ( x i x ¯ ) 2
The sample mean and variance are used to estimate the mean and standard deviation, respectively. Next, the foreground is segmented utilizing the following threshold:
Z ( i ) ( x , y ) = 0 if Z ( i ) ( x , y ) μ ± λ σ 1 otherwise
where λ is determined empirically from simulations, and ( x , y ) are spatial coordinates of the pixel values within the frame Z ( i ) . Finally, binary opening and closing operations are integrated to reduce the noise effects after thresholding.

2.3. Vehicle Features

The authors utilize the same approach presented in their previous work [54] to extract segmented vehicle features. Size, shape, and texture descriptors are considered for classification. Morphological properties are used to describe the size and shape of each vehicle. Gradient-based texture features are calculated using a HOG algorithm on the rectangular boundary image of the detected vehicle.
Twelve extracted feature descriptors representing the size and shape of segmented vehicles are employed. These descriptors include area, bounding box, centroid, convex area, eccentricity, equivalent diameter, Euler number, extent, primary axis length, minor axis length, orientation, and perimeter [54,55].
In addition to the physical features, textural features are integrated by computing the histogram of oriented gradients of segmented vehicles. Due to its effectiveness in identifying an object’s visual texture, the HOG descriptor is widely used in image processing and computer vision for object detection [56]. The HOG features are evaluated using both image gradient magnitudes and orientations and are slightly affected by local geometric and lighting distortions.

2.4. Vehicle Tracking

In order to keep an accurate count of vehicles within the highway image scene, each detected vehicle is tracked as it traverses the ROI. A Kalman filter tracking approach is used to track each detected vehicle through the ROI [57]. The Kalman filter vector is created for each newly detected vehicle. It tracks the ID number of the vehicle, the centroid position, and the vehicle velocity assuming constant acceleration through the ROI. Image features derived from the foreground-segmented vehicle are also paired with the Kalman filter track for every frame in which the vehicle is present. Once the vehicle exits the ROI, the Kalman filter vehicle track is deleted.

2.5. Classification

Once a vehicle track has been deleted, as described in the previous section, the classification of the vehicle at every frame occurs. The classification of vehicles occurs in three steps: (1) training the classifier, (2) multiclass SVM classification, and (3) statistical analysis for final classification.

2.5.1. Training the Classifier

The classifier is trained using a manually labeled ground truth dataset for each camera. Eight classes are considered for training. These classes include the six vehicle classes: passenger car, passenger truck, bus, small utility truck, and large utility truck. In addition, classes representing a detection that is not a vehicle or a detection containing multiple vehicles are presented. The classifier is trained on a manually tagged dataset using 10-fold cross-validation in which the dataset is divided randomly into 10 groups. The training is conducted using nine groups; the remaining group is used for testing. The process is repeated ten times to estimate the classifier’s performance.

2.5.2. Multiclass SVM Classification

Multi-class SVM classifier is adopted to classify the segmented data into one of eight classes. The SVM classifier is known to be efficient in separating large feature vectors and does not require intensive memory allocations [58]. A set of binary SVM classifiers is adopted to handle the multi-class separation. The SVM classifier with radial basis function (RBF) kernel is chosen for its better separation of nonlinear feature vectors. The RBF kernel has two parameters, called the sigma and box constraint, found by a grid search method. In the grid search method, each parameter varies over a range of 0.1 to 15, and the 10-fold cross-validation is performed. The parameters are set for the maximum cross-validation result.
Following the same approach discussed by Chen et al. [59], the work presented here adopts eight binary SVM classifiers to handle the problem of multi-object classification. First, one SVM classifier is trained to separate a certain positive class from other classes treated as a single negative class. Then, after the training phase of the chosen classifier is over, the same process continues for training other classifiers.

2.5.3. Final Classification Result

The final classification result is found by taking the statistical mode of each frame classification result. For example, a single vehicle track may make a mistake in tracking and correct to the wrong detected vehicle. Furthermore, the tracked vehicle may also include partial detection at the ROI boundary and be classified as “Not a Vehicle”. Using the mode of the classification results eliminates these conditions and uses the most stable classification results from each frame in the vehicle track.

2.6. Multiple Vehicle Handling

Due to vehicular occlusion, some vehicles traveling nearby are clustered and appear as a single vehicle. The classification method in the previous step is trained to recognize multiple vehicle clusters. If the deleted track contains multiple vehicles, then occlusion is assumed. This occlusion is handled through an iterative over-segmentation and reclassification method.

2.6.1. K-Means Over-Segmentation

The first step in over-segmentation is to analyze the geometry of the boundary rectangle for the occluded vehicles. If the boundary rectangle width is greater than the height, it is assumed that the multiple vehicles are arranged from left to right, and the boundary rectangle is divided vertically. Otherwise, it is assumed that the multiple vehicles are arranged from top to bottom, and the boundary rectangle is divided horizontally. Two centroids are assumed in the newly divided boundary rectangle.
K-means clustering is used to perform over-segmentation of the multiple-vehicle cluster. K-means uses the image intensity and quantizes similar intensity levels into a single patch. This over-segmentation process uses up to 25 levels applied to the input image of each multiple-vehicle cluster.
After the newly segmented patches are created, the centroid of each patch is calculated. The minimum Euclidean distance of each patch to the new boundary box centroids is used to assign each patch to the new boundary box. Once the patches are assigned, the newly clustered vehicle is converted from the patches. Features are extracted from the new vehicle cluster.

2.6.2. Reclassification

The final step in multiple vehicle handling is to perform classification on the two detected vehicles from the k-means over-segmentation process. This reclassification follows the method described in Section 2.5. The steps in this section are repeated if the vehicle cluster still contains multiple vehicles. If the vehicles cannot be separated, this iterative approach will only repeat three times.

3. Results and Discussion

A manually labeled ground truth dataset was created for evaluation. Four evaluated sample videos are described in Table 1. These videos were manually labeled for vehicle segmentation and classification. Each video shows a segment of an interstate under varying conditions of traffic and environmental scenarios. Figure 4 shows a sample of vehicles within the given dataset. It can be seen that the given dataset is poor in its spatial resolution, and its visual quality content is highly degraded and under-sampled. The size of bounding boxes containing the extracted vehicles is 40 by 30 pixels extracted from video frames of dimensions 320 by 240 pixels at a spatial resolution of 72 dots per inch (dpi).

3.1. ROI Extraction

The proposed ROI extraction technique described in Section 2.1 is applied to the first frame of the collected videos to prepare and feed the extracted ROI mask to subsequent processing of other frames. The processed frame is filtered utilizing a Gaussian blur filter with a 3 × 3 window. A canny edge detector is adopted to detect the edges from the blurred image. The Hough transform is applied, and empirically ten peaks are selected. Hough lines are extracted with a minimum length of 20. FCM clustering algorithm is applied with four clusters. Finally, linear polynomial fit is used to fit the clustered line segments into lane sides. Detailed results of different ROI pipeline stages are depicted in Figure 5 and Figure 6 for two different sample videos. The proposed technique demonstrates good results in the extraction of ROI. It works well for line-shaped lanes. While the performance has a small degradation for curvy lanes, as seen in Figure 5f, it is still acceptable since the scope of this approach is not for lane departure systems that require precise extraction of the ROI. Additionally, applying the Gaussian kernel to images reduces the amount of extracted edges, helps to identify the strong ones, and reduces the not-required Hough lines. FCM works to group different line segments nearby into unique cluster lines. Finally, the linear curve fitting approximates the clustered lines into a single line representing highway lanes’ sides. The ROI extraction results in an almost 40% reduction of pixels processed by the algorithm’s subsequent stages. The overall computational time is reduced due to applying an ROI mask.

3.2. Foreground Segmentation Evaluation

The proposed algorithm utilizes the FPCP approach as an initial stage to pre-train the system. First, a set of frames determined empirically are utilized as initialization to generate a priori low-rank background representation used in subsequent frame differentiating stages. Then, the frames are reshaped into column vectors stacked side-by-side to construct a matrix fed to the FPCP algorithm to estimate a background frame. Utilizing 60 frames in the training phase reduces the processed data by almost 80% for Videos 1, 2, and 3 and 94% for Video 4. Once the frames are subtracted from the pre-trained low-rank matrix, the differentiated frames are fitted into the Gaussian distribution. Then, the mean and variance are estimated from the extracted difference sparse matrix. Thresholding is applied utilizing Equation (9) in which the tuning parameter λ is empirically set to five. Foreground vehicle segmentation is evaluated using classification accuracy, precision, recall, and F-measure. Classification accuracy (CA) is defined as
CA = ( T P + T N ) / ( T P + F N + F P + T N )
where TP is the true positive, TN is the true negative, FP is the false positive, and FN is the false negative. Since classification accuracy is not a good representative for evaluating algorithms when the available dataset is imbalanced, it has been advised through the research communities to use the F-measure as a good metric to evaluate the performance of segmentation and detection techniques. The F-measure is determined by:
F - meaure = 2 Precision Recall Precision + Recall
where Precision is defines as:
Precision = T P T P + F P ,
and Recall is defined as:
Recall = T p T P + F N
TP, TN, FP, and FN detected pixels are determined by comparing the algorithm outputs to a manually labeled ground truth for the given dataset. Moreover, the proposed algorithms are compared against recent foreground segmentation techniques including: fast principal component pursuit via alternating minimization (FPCP) [53], Grassmann averages for scalable robust PCA (GA) [60], low-rank and sparse matrix decomposition via the truncated nuclear norm and a sparse regularizer (LRSMDTNNorm) [61], shifted subspaces tracking on sparse outlier for motion segmentation (GreGoDec) [62], robust principal component pursuit via inexact alternating minimization on matrix manifolds (RPCP ) [63], scalable robust matrix recovery: Frank–Wolfe meets proximal methods (FW-T) [64], robust principal component analysis with complex noise (MoG-RPCA) [65], a variational approach to stable principal component pursuit (Lag-SPCP-QN) [66], and robust PCA via nonconvex rank approximation (NonConvRPCA ) [67]. The previously mentioned classification metrics and the average computational time per frame (Ave comp. time) are used to compare the proposed technique performance against those algorithms.
Table 2, Table 3, Table 4 and Table 5 list different evaluation metrics for the proposed algorithm against other segmentation algorithms for four different videos. For Video 1, which represents a sample of sunny daytime traffic, the proposed algorithm showed an improvement of 10% in terms of the F-measure, and it is 20% faster than the GA approach, which is the fastest technique among the compared algorithms. A sample of traffic with low light and dusk conditions is represented by Video 2. Improvements of about 20% in the F-measure and 80% in the computational speed are achieved. Rainy and daytime traffic condition is depicted in Video 3. However, the proposed algorithm improved by 14% in the F-measure and 83% in the computational time; nevertheless, the quality of the segmentation process needs to be improved. Since Video 3 represents rainy conditions, the proposed algorithm and other evaluated algorithms failed to deal effectively with reflections due to rain that are considered moving objects and consequently interpreted as foreground objects. Performance improvements for Video 4 are 12% and 36% in terms of F-score and computational time, respectively. Figure 7, Figure 8, Figure 9 and Figure 10 depict the visual quality of segmented vehicles for the proposed technique against other evaluated algorithms. It can be seen that the proposed algorithm is better than other techniques, and more accurate segmentation of vehicles for different videos has been achieved. Techniques GA, Lag-SPCP-QN, and GM have poor showing for the different evaluated videos.

3.3. Classification Evaluation

We used 10-fold cross-validation to train the classifier compared to a manually labeled dataset. Evaluation metrics, including precision, recall, and F-score, are utilized to evaluate the classification performance for each class. A ROC plot is given for each class considered for classification. The confusion matrix for each classification result is also obtained.

3.3.1. Video 1

Video 1 has a total of 1008 vehicle detections. These detections include 523 passenger cars, 257 passenger trucks, 129 multiple vehicle detections, and 99 non-vehicle detections. The grid search result gives a box constraint of 11 and a sigma of 4 for the RBF parameters. The confusion matrix is shown in Table 6, while classification evaluation metrics are shown in Table 7. The overall accuracy for Video 1 is 86.71%. Regarding the F-score, non-vehicle objects classification performance is the worst among different classes. There is excellent handling for multiple vehicle cluster detection and passenger car classes. Figure 11 shows the ROC curves with an average area under the curve (AUC) of 0.8963 for all classes.
There are 123 predicted multiple vehicle clusters detected in Video 1, representing 264 total vehicles that are a subset of the final classification result. The confusion matrix for detected vehicles from multiple vehicle clustering algorithm output is described in Table 8 with an overall accuracy of 72.73%. These 264 vehicles are omitted from the output data without the multiple vehicle cluster handling. Additionally, classification evaluation metrics for multiple vehicle clusters are presented in Table 9. The algorithm performs well in identifying various vehicle classes from clustered objects.

3.3.2. Video 2

Video 2 has a total of 373 vehicle detections. These detections include 120 passenger cars, 86 passenger trucks, 13 small utility trucks, 14 multiple vehicle detections, and 140 non-vehicle detections. The grid search result gives a box constraint of seven and a sigma of three for the RBF parameters. The confusion matrix is depicted in Table 10, while classification evaluation metrics are shown in Table 11. The overall accuracy for Video 2 is 84.72%. Figure 12 shows the ROC curves with an average AUC of 0.9088 for all classes. In terms of the F-score metric, there is an excellent performance for different classes classification.
There are 12 predicted multiple vehicle clusters detected in Video 2, representing 26 total vehicles that are a subset of the final classification result. The confusion matrix for detected vehicles from multiple vehicle clustering algorithm output is described in Table 12 with an overall accuracy of 66.67%. Additionally, classification evaluation metrics for multiple vehicle clusters are presented in Table 13. The algorithm performs well for both passenger car and truck classes, while it has abysmal performance for small utility trucks. This is because the small utility truck has a poor showing in the dataset, and consequently, the algorithm failed to identify the class without enough training data.

3.3.3. Video 3

Video 3 has a total of 1948 vehicle detections. These detections include 464 passenger cars, 478 passenger trucks, 67 large utility trucks, 189 multiple vehicle detections, and 744 non-vehicle detections. The grid search result gives a box constraint of 11 and a sigma of 3 for the RBF parameters. The confusion matrix is shown in Table 14, while classification evaluation metrics are shown in Table 15. The overall accuracy for Video 3 is 78.75.71%. Regarding the F-score, the algorithm performs extremely poorly classifying motorcycles, buses, and small utility trucks. This is expected since there is not enough training data representing these classes. Figure 13 shows the ROC curves with an average AUC of 0.8666 for all classes.
There are 237 predicted multiple vehicle clusters detected in Video 3, representing 546 total vehicles that are a subset of the final classification result. The confusion matrix for detected vehicles from multiple vehicle clustering algorithm performance is described in Table 16 with an overall accuracy of 61.36%. Additionally, classification evaluation metrics for multiple vehicle clusters are presented in Table 17. The algorithm performs well in identifying both passenger car and truck classes, while it has poor performance for other classes.

3.3.4. Video 4

Video 4 has a total of 2868 vehicle detections. These detections include 592 passenger cars, 671 passenger trucks, 22 buses, 125 small utility trucks, 58 large utility trucks, 269 multiple-vehicle detections, and 1131 non-vehicle detections. The grid search result gives a box constraint of 11 and a sigma of 3 for the RBF parameters. The confusion matrix is shown in Table 18, while classification evaluation metrics are shown in Table 19. The overall accuracy for Video 4 is 86.37%. The algorithm shows good performance in identifying various classes. Figure 14 shows the ROC curves with an average AUC of 0.8984 for all classes.
There are 234 predicted multiple vehicle clusters detected in Video 4, representing 513 total vehicles that are a subset of the final classification result. The multiple vehicle clustering algorithm performance is described in Table 20 with an overall classification accuracy of 71.92%. Additionally, classification evaluation metrics for multiple vehicle clusters are presented in Table 21. The algorithm shows poor performance in identifying both bus and large utility truck classes.

3.4. Comparison with Deep Learning Architectures

While deep learning is a fast-growing area of research in different domains of image and video processing and computer vision, it has several drawbacks. Different tactics have been tried in intelligent transportation systems (ITS). Favorable and accurate results have been demonstrated compared to traditional methods, but it requires machines with intensive memory storage and huge computational power. Adding to that, it requires a huge amount of training data. Additionally, its configuration and tuning are a hard and lengthy process. Optimizing the deep learning architecture parameters requires too many experiments running for long periods to set the optimal settings [45,46,47].
The work presented in this paper evaluates the performance of twenty-three different deep learning architectures for object detection that have been previously trained on either the COCO dataset or ImageNet. Since the training process needs high computational resources and graphical processing units (GPUs) and the process is time-consuming, transfer learning is adapted in this work [68]. The different evaluated models utilize the weights from pre-trained networks to fine-tune these architectures on new datasets. The evaluated deep object detection architectures are: Faster R-CNN [8], YOLO V2 [69], YOLO V3 [7], and YOLO V4 [70]) and the feature extractor networks are (Resnet18, Resnet50, Resent101 [71], inception V3 [72], inceptionresnet V2 [73], darknet53 [74], squeezenet [75], and googlenet [76]). The combinations of these architectures are evaluated in terms of the average precision, which is defined as the area under the curve representing the relation between the recall and precision at different thresholds. The performance of these deep learning models is evaluated on small datasets using transfer learning and is compared against the proposed traditional technique.
The trails of deep learning architecture and feature extractors networks evaluated in this work are listed in Table 22. As depicted from the table, not all possible combinations were examined. This is due to the lack of high computational power and memory. Additionally, many experiments were carried out for each combination to fine-tune the network parameters.
The dataset was divided into 60% for training, 10% for validation, and 30% for testing the trained detector. Data augmentation was applied to the training data to improve the training and increase the number of labeled data. The data augmentation included color jitter in HSV space, random scaling, and random horizontal flip.
Table 23 lists the average precision of the evaluated deep learning combinations against the proposed algorithm. It can be seen that the performance of deep learning architectures was abysmal. This is expected since the available dataset is small and does not provide enough labeled training samples to train the deep learning networks. The proposed algorithm improved the performance against the evaluated algorithms, especially for Videos 1, 2, and 4. However, Video 3, representing rainy conditions, had the worst classification performance among different algorithms.

3.5. Timing Analysis

The algorithm was tested on a computer running as a single thread on 11th Gen Intel(R) Core(TM) running at 2.80 GHz with 16 GB of RAM with no GPU card installed. The timing analysis, given in Table 24, summarizes the time required to run each sample video through the algorithm and the average frame rate. The timing analysis shows that the approach implementation can operate in real time at 15 frames per second.

4. Conclusions

In this work, an end-to-end integrated vision-based framework that is capable of segmenting moving vehicles and classifying them by type is presented. Novel contributions to intelligent transportation systems research have been presented by an integrated algorithm capable of vehicle segmentation and classification in a camera-viewpoint-independent highway image scene with the ability to handle multiple vehicle clusters. The algorithm was speeded up through an automated approach for extracting a region of interest, reducing 40% in the processed data. Vehicle segmentation was carried out using an improved speeded-up robust low-rank matrix decomposition technique. Utilizing a new and effective thresholding method improved segmentation accuracy and simultaneously sped up the segmentation processing. It improved segmentation accuracy by 15% on average compared to current approaches and was 55% quicker on average than recent segmentation techniques. Segmentation was shown to perform very well during the day. However, the segmentation and classification results degraded under low light conditions or poor environmental scenarios such as rain.
The algorithm adopted size and shape physical descriptors from morphological properties and textural features from the histogram of oriented gradients (HOG) extracted from the segmented traffic for feature extraction. Moreover, a multi-class support vector machine classifier was used to identify diverse traffic vehicle categories, such as passenger cars, passenger trucks, motorbikes, buses, and small and big utility vehicles. The proposed approach manages multiple vehicle clusters through an iterative k-means clustering over-segmentation procedure. It shows high classification performance for vehicle types, such as passenger cars and passenger trucks, while its performance was disappointing for classes with a poor showing in the training dataset. Additionally, the algorithm was compared against 23 different deep-learning architectures. The resulting algorithm outperformed the compared deep learning algorithms for the quality of vehicle classification accuracy.
The timing analysis results show that the algorithm’s current implementation can perform in real time for the 15 frames per second frame rate. However, future modifications to the algorithm to handle different weather scenarios, such as rainy conditions, can improve the segmentation and classification of the proposed algorithm and make it adaptable to different weather conditions. Additionally, the lack of enough training datasets close to the one utilized in this research makes using deep learning techniques not promising and challenging and leads to poor performance.

Author Contributions

Conceptualization, A.Y. and K.I.; methodology, A.Y.; software, A.Y. and J.F.; validation, A.Y. and J.F. formal analysis, A.Y. and K.I.; resources, J.F.; writing—original draft preparation, A.Y.; writing—review and editing, A.Y. and K.I.; visualization, A.Y. and J.F.; supervision, K.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Trabelsi, R.; Khemmar, R.; Decoux, B.; Ertaud, J.Y.; Butteau, R. Recent Advances in Vision-Based On-Road Behaviors Understanding: A Critical Survey. Sensors 2022, 22, 2654. [Google Scholar] [CrossRef] [PubMed]
  2. Yeong, D.J.; Velasco-Hernandez, G.; Barry, J.; Walsh, J. Sensor and Sensor Fusion Technology in Autonomous Vehicles: A Review. Sensors 2021, 21, 2140. [Google Scholar] [CrossRef] [PubMed]
  3. Loce, R.P.; Bernal, E.A.; Wu, W.; Bala, R. Computer vision in roadway transportation systems: A survey. J. Electron. Imaging 2013, 22, 22–24. [Google Scholar] [CrossRef] [Green Version]
  4. Feng, Y.; Mao, G.; Chen, B.; Li, C.; Hui, Y.; Xu, Z.; Chen, J. MagMonitor: Vehicle Speed Estimation and Vehicle Classification Through A Magnetic Sensor. IEEE Trans. Intell. Transp. Syst. 2020, 23, 1311–1322. [Google Scholar] [CrossRef]
  5. Liu, C.; Huynh, D.Q.; Sun, Y.; Reynolds, M.; Atkinson, S. A Vision-Based Pipeline for Vehicle Counting, Speed Estimation, and Classification. IEEE Trans. Intell. Transp. Syst. 2021, 22, 7547–7560. [Google Scholar] [CrossRef]
  6. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 21–37. [Google Scholar]
  7. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  8. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the NIPS’15: Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  9. Guo, L.; Li, R.; Jiang, B. An Ensemble Broad Learning Scheme for Semisupervised Vehicle Type Classification. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 5287–5297. [Google Scholar] [CrossRef]
  10. Chen, C.P.; Liu, Z. Broad learning system: An effective and efficient incremental learning system without the need for deep architecture. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 10–24. [Google Scholar] [CrossRef]
  11. Niroomand, N.; Bach, C.; Elser, M. Robust Vehicle Classification Based on Deep Features Learning. IEEE Access 2021, 9, 95675–95685. [Google Scholar] [CrossRef]
  12. Soon, F.C.; Khaw, H.Y.; Chuah, J.H.; Kanesan, J. Semisupervised PCA Convolutional Network for Vehicle Type Classification. IEEE Trans. Veh. Technol. 2020, 69, 8267–8277. [Google Scholar] [CrossRef]
  13. Shvai, N.; Hasnat, A.; Meicler, A.; Nakib, A. Accurate classification for automatic vehicle-type recognition based on ensemble classifiers. IEEE Trans. Intell. Transp. Syst. 2020, 21, 1288–1297. [Google Scholar] [CrossRef]
  14. Theagarajan, R.; Thakoor, N.S.; Bhanu, B. Physical features and deep learning-based appearance features for vehicle classification from rear view videos. IEEE Trans. Intell. Transp. Syst. 2020, 21, 1096–1108. [Google Scholar] [CrossRef]
  15. Nazemi, A.; Azimifar, Z.; Shafiee, M.J.; Wong, A. Real-time vehicle make and model recognition using unsupervised feature learning. IEEE Trans. Intell. Transp. Syst. 2019, 21, 3080–3090. [Google Scholar] [CrossRef]
  16. Bosch, A.; Zisserman, A.; Munoz, X. Image classification using random forests and ferns. In Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio de Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  17. Lazebnik, S.; Schmid, C.; Ponce, J. Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 2169–2178. [Google Scholar]
  18. Ma, Z.; Chang, D.; Xie, J.; Ding, Y.; Wen, S.; Li, X.; Si, Z.; Guo, J. Fine-Grained Vehicle Classification with Channel Max Pooling Modified CNNs. IEEE Trans. Veh. Technol. 2019, 68, 3224–3233. [Google Scholar] [CrossRef]
  19. Zhang, F.; Li, C.; Yang, F. Vehicle detection in urban traffic surveillance images based on convolutional neural networks with feature concatenation. Sensors 2019, 19, 594. [Google Scholar] [CrossRef] [Green Version]
  20. Arinaldi, A.; Pradana, J.A.; Gurusinga, A.A. Detection and classification of vehicles for traffic video analytics. Procedia Comput. Sci. 2018, 144, 259–268. [Google Scholar] [CrossRef]
  21. Xiang, X.; Zhai, M.; Lv, N.; El Saddik, A. Vehicle counting based on vehicle detection and tracking from aerial videos. Sensors 2018, 18, 2560. [Google Scholar] [CrossRef] [Green Version]
  22. Lu, L.; Huang, H. A Hierarchical Scheme for Vehicle Make and Model Recognition from Frontal Images of Vehicles. IEEE Trans. Intell. Transp. Syst. 2019, 20, 1774–1786. [Google Scholar] [CrossRef]
  23. Farahani, G. Dynamic and robust method for detection and locating vehicles in the video images sequences with use of image processing algorithm. EURASIP J. Image Video Process. 2017, 2017, 87. [Google Scholar] [CrossRef] [Green Version]
  24. Roy, K.; Saha, S.; Mondal, T.; Choudhury, S.S. Automatic Vehicle Detection and Motion Path Tracking Based on Gaussian Mixture Model. In Advances in Optical Science and Engineering; Springer: Berlin/Heidelberg, Germany, 2017; pp. 669–679. [Google Scholar]
  25. Das, J.; Shah, M.; Mary, L. Bag of feature approach for vehicle classification in heterogeneous traffic. In Proceedings of the IEEE International Conference on Signal Processing, Informatics, Communication and Energy Systems (SPICES), Kollam, India, 8–10 August 2017; pp. 1–5. [Google Scholar]
  26. Liu, W.; Zhang, M.; Luo, Z.; Cai, Y. An Ensemble Deep Learning Method for Vehicle Type Classification on Visual Traffic Surveillance Sensors. IEEE Access 2017, 5, 24417–24425. [Google Scholar] [CrossRef]
  27. Wang, H.; Zheng, Z.; Cai, Y.; Sun, X. Scene-Adaptive Vehicle Detection Algorithm based on a Composite Deep Structure. IEEE Access 2017, 5, 22804–22811. [Google Scholar]
  28. Biglari, M.; Soleimani, A.; Hassanpour, H. Part-based recognition of vehicle make and model. IET Image Process. 2017, 11, 483–491. [Google Scholar] [CrossRef]
  29. Chavez-Garcia, R.O.; Aycard, O. Multiple sensor fusion and classification for moving object detection and tracking. IEEE Trans. Intell. Transp. Syst. 2016, 17, 525–534. [Google Scholar] [CrossRef] [Green Version]
  30. Kamkar, S.; Safabakhsh, R. Vehicle detection, counting and classification in various conditions. IET Intell. Transp. Syst. 2016, 10, 406–413. [Google Scholar] [CrossRef]
  31. Siddiqui, A.J.; Mammeri, A.; Boukerche, A. Real-time vehicle make and model recognition based on a bag of SURF features. IEEE Trans. Intell. Transp. Syst. 2016, 17, 3205–3219. [Google Scholar] [CrossRef]
  32. Toropov, E.; Gui, L.; Zhang, S.; Kottur, S.; Moura, J.M. Traffic flow from a low frame rate city camera. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 3802–3806. [Google Scholar]
  33. Chaple, M.; Paygude, S. Vehicle detection and tracking from video frame sequence. Int. J. Sci. Eng. Res. 2013, 4, 1–7. [Google Scholar]
  34. Morris, B.T.; Tran, C.; Scora, G.; Trivedi, M.M.; Barth, M.J. Real-time video-based traffic measurement and visualization system for energy/emissions. IEEE Trans. Intell. Transp. Syst. 2012, 13, 1667–1678. [Google Scholar] [CrossRef] [Green Version]
  35. Betke, M.; Haritaoglu, E.; Davis, L.S. Real-time multiple vehicle detection and tracking from a moving vehicle. Mach. Vis. Appl. 2000, 12, 69–83. [Google Scholar] [CrossRef] [Green Version]
  36. Daigavane, M.; Bajaj, P. Real-Time Vehicle Detection and Counting Method for Real Time Vehicle Detection and Counting Method for Unsupervised Traffic Video on Highways Unsupervised Traffic Video on Highways. IJCSNS 2010, 10, 112. [Google Scholar]
  37. Sochor, J. Fully automated real-time vehicles detection and tracking with lanes analysis. In Proceedings of the 18th Central European Seminar on Computer Graphics, CESCG, Smolenice, Slovakia, 25–27 May 2014; pp. 1–8. [Google Scholar]
  38. Gepperth, A.; Edelbrunner, J.; Bucher, T. Real-time detection and classification of cars in video sequences. In Proceedings of the IEEE Intelligent Vehicles Symposium, Las Vegas, NV, USA, 6–8 June 2005; pp. 625–631. [Google Scholar]
  39. Sullivan, G.D.; Baker, K.D.; Worrall, A.D.; Attwood, C.; Remagnino, P. Model-based vehicle detection and classification using orthographic approximations. Image Vis. Comput. 1997, 15, 649–654. [Google Scholar] [CrossRef]
  40. Zhang, G.; Avery, R.; Wang, Y. Video-based vehicle detection and classification system for real-time traffic data collection using uncalibrated video cameras. Transp. Res. Rec. J. Transp. Board 2007, 1993, 138–147. [Google Scholar] [CrossRef] [Green Version]
  41. Ambardekar, A.A. Efficient Vehicle Tracking and Classification for an Automated Traffic Surveillance System; University of Nevada: Reno, NV, USA, 2007. [Google Scholar]
  42. Qian, Z.; Shi, H.; Yang, J.; Duan, L. Video-based multiclass vehicle detection and tracking. J. Comput. Sci. 2013, 10, 570–578. [Google Scholar]
  43. Rad, R.; Jamzad, M. Real time classification and tracking of multiple vehicles in highways. Pattern Recognit. Lett. 2005, 26, 1597–1607. [Google Scholar] [CrossRef]
  44. Zhan, W.; Luo, Z. System design of real time vehicle type recognition based on video for windows (AVI) files. In Intelligent Computing and Information Science; Springer: Berlin/Heidelberg, Germany, 2011; pp. 681–686. [Google Scholar]
  45. Boyd, A.; Bowyer, K.W.; Czajka, A. Human-Aided Saliency Maps Improve Generalization of Deep Learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 2735–2744. [Google Scholar]
  46. Li, Y. Research and Application of Deep Learning in Image Recognition. In Proceedings of the 2nd IEEE International Conference on Power, Electronics and Computer Applications (ICPECA), Shenyang, China, 21–23 January 2022; pp. 994–999. [Google Scholar]
  47. Kulkarni, S.; Rabidas, R. A Comparative Study of Different Deep Learning Architectures for Benign-Malignant Mass Classification. In Proceedings of the 2nd International Conference on Recent Trends in Machine Learning, IoT, Smart Cities and Applications; Springer: Berlin/Heidelberg, Germany, 2022; pp. 773–784. [Google Scholar]
  48. Miralles-Pechuán, L.; Rosso, D.; Jiménez, F.; García, J.M. A methodology based on Deep Learning for advert value calculation in CPM, CPC and CPA networks. Soft Comput. 2017, 21, 651–665. [Google Scholar] [CrossRef]
  49. Nguyen, T.T.; Nguyen, N.D.; Nahavandi, S. Deep Reinforcement Learning for Multi-Agent Systems: A Review of Challenges, Solutions and Applications. arXiv 2018, arXiv:1812.11794. [Google Scholar]
  50. Mahmud, M.; Kaiser, M.S.; Hussain, A.; Vassanelli, S. Applications of deep learning and reinforcement learning to biological data. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2063–2079. [Google Scholar] [CrossRef]
  51. Bouwmans, T.; Aybat, N.S.; Zahzah, E.h. Handbook of Robust Low-Rank and Sparse Matrix Decomposition: Applications in Image and Video Processing; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  52. Sobral, A.; Bouwmans, T.; Zahzah, E.h. Lrslibrary: Low-rank and sparse tools for background modeling and subtraction in videos. In Robust Low-Rank and Sparse Matrix Decomposition: Applications in Image and Video Processing; CRC Press: Boca Raton, FL, USA, 2016. [Google Scholar]
  53. Rodríguez, P.; Wohlberg, B. Fast principal component pursuit via alternating minimization. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, VIC, Australia, 15–18 September 2013; pp. 69–73. [Google Scholar] [CrossRef]
  54. Flora, J.B.; Alam, M.; Iftekharuddin, K.M. Improvements to vehicular traffic segmentation and classification for emissions estimation using networked traffic surveillance cameras. In Proceedings of the Optics and Photonics for Information Processing VIII; SPIE: Bellingham, WA, USA, 2014; Volume 9216, pp. 162–171. [Google Scholar]
  55. Stauffer, C.; Grimson, W.E.L. Adaptive background mixture models for real-time tracking. In Proceedings of the IEEE Computer Vision and Pattern Recognition Conference, Fort Collins, CO, USA, 23–25 June 1999; Volume 2, pp. 246–252. [Google Scholar]
  56. Dalal, N.; Triggs, B. Histograms of oriented gradients for human detection. In Proceedings of the IEEE Computer Vision and Pattern Recognition Conference, CVPR, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 886–893. [Google Scholar]
  57. Nilsson, J.; Fredriksson, J.; Ödblom, A.C. Reliable vehicle pose estimation using vision and a single-track model. IEEE Trans. Intell. Transp. Syst. 2014, 15, 2630–2643. [Google Scholar] [CrossRef]
  58. Müller, K.R.; Smola, A.J.; Rätsch, G.; Schölkopf, B.; Kohlmorgen, J.; Vapnik, V. Predicting time series with support vector machines. In Proceedings of the International Conference on Artificial Neural Networks; Springer: Berlin/Heidelberg, Germany, 1997; pp. 999–1004. [Google Scholar]
  59. Chen, Z.; Pears, N.; Freeman, M.; Austin, J. Road vehicle classification using support vector machines. In Proceedings of the IEEE Intelligent Computing and Intelligent Systems International Conference, ICIS, Shanghai, China, 20–22 November 2009; Volume 4, pp. 214–218. [Google Scholar]
  60. Hauberg, S.; Feragen, A.; Black, M.J. Grassmann Averages for Scalable Robust PCA. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3810–3817. [Google Scholar] [CrossRef] [Green Version]
  61. Xue, Z.; Dong, J.; Zhao, Y.; Liu, C.; Chellali, R. Low-rank and sparse matrix decomposition via the truncated nuclear norm and a sparse regularizer. Vis. Comput. 2019, 35, 1549–1566. [Google Scholar] [CrossRef]
  62. Zhou, T.; Tao, D. Shifted subspaces tracking on sparse outlier for motion segmentation. In Proceedings of the IJCAI ’13: Proceedings of the 23rd International Joint Conference on Artificial Intelligence, Beijing, China, 3–9 August 2013. [Google Scholar]
  63. Hintermüller, M.; Wu, T. Robust principal component pursuit via inexact alternating minimization on matrix manifolds. J. Math. Imaging Vis. 2015, 51, 361–377. [Google Scholar] [CrossRef] [Green Version]
  64. Mu, C.; Zhang, Y.; Wright, J.; Goldfarb, D. Scalable robust matrix recovery: Frank–Wolfe meets proximal methods. SIAM J. Sci. Comput. 2016, 38, A3291–A3317. [Google Scholar] [CrossRef] [Green Version]
  65. Zhao, Q.; Meng, D.; Xu, Z.; Zuo, W.; Zhang, L. Robust principal component analysis with complex noise. In Proceedings of the ICML’14: Proceedings of the 31st International Conference on International Conference on Machine Learning, Beijing, China, 21–26 June 2014; pp. 55–63. [Google Scholar]
  66. Aravkin, A.; Becker, S.; Cevher, V.; Olsen, P. A variational approach to stable principal component pursuit. arXiv 2014, arXiv:1406.1089. [Google Scholar]
  67. Kang, Z.; Peng, C.; Cheng, Q. Robust PCA via nonconvex rank approximation. In Proceedings of the IEEE International Conference on Data Mining, Atlantic City, NJ, USA, 14–17 November 2015; pp. 211–220. [Google Scholar]
  68. Arcos-García, Á.; Alvarez-Garcia, J.A.; Soria-Morillo, L.M. Evaluation of deep neural networks for traffic sign detection systems. Neurocomputing 2018, 316, 332–344. [Google Scholar] [CrossRef]
  69. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  70. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  71. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  72. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  73. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017. [Google Scholar]
  74. Redmon, J. Darknet: Open Source Neural Networks in C. Available online: http://pjreddie.com/darknet/ (accessed on 30 August 2022).
  75. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and< 0.5 MB model size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  76. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
Figure 1. Proposed algorithm flow diagram.
Figure 1. Proposed algorithm flow diagram.
Sensors 22 08121 g001
Figure 2. ROI extraction pipeline.
Figure 2. ROI extraction pipeline.
Sensors 22 08121 g002
Figure 3. A sample histogram of extracted sparse background for different dataset videos.
Figure 3. A sample histogram of extracted sparse background for different dataset videos.
Sensors 22 08121 g003
Figure 4. Sample vehicles extracted from different videos in the dataset.
Figure 4. Sample vehicles extracted from different videos in the dataset.
Sensors 22 08121 g004
Figure 5. Different output for ROI extraction pipeline stages for Video 1.
Figure 5. Different output for ROI extraction pipeline stages for Video 1.
Sensors 22 08121 g005
Figure 6. Different output for ROI extraction pipeline stages for Video 4.
Figure 6. Different output for ROI extraction pipeline stages for Video 4.
Sensors 22 08121 g006
Figure 7. Video 1 segmentation outputs comparison for different segmentation techniques.
Figure 7. Video 1 segmentation outputs comparison for different segmentation techniques.
Sensors 22 08121 g007
Figure 8. Video 2 segmentation outputs comparison for different segmentation techniques.
Figure 8. Video 2 segmentation outputs comparison for different segmentation techniques.
Sensors 22 08121 g008
Figure 9. Video 3 segmentation outputs comparison for different segmentation techniques.
Figure 9. Video 3 segmentation outputs comparison for different segmentation techniques.
Sensors 22 08121 g009
Figure 10. Video 4 segmentation outputs comparison for different segmentation techniques.
Figure 10. Video 4 segmentation outputs comparison for different segmentation techniques.
Sensors 22 08121 g010
Figure 11. ROC curves for Video 1 classification.
Figure 11. ROC curves for Video 1 classification.
Sensors 22 08121 g011
Figure 12. ROC curves for Video 2 classification.
Figure 12. ROC curves for Video 2 classification.
Sensors 22 08121 g012
Figure 13. ROC curves for Video 3 classification.
Figure 13. ROC curves for Video 3 classification.
Sensors 22 08121 g013
Figure 14. ROC curves for Video 4 classification.
Figure 14. ROC curves for Video 4 classification.
Sensors 22 08121 g014
Table 1. Description of sample videos for evaluation.
Table 1. Description of sample videos for evaluation.
NameNumber of FramesTimeRoad DescriptionEnvironmental Description
Video 129920 s5 lanes in each direction, medium trafficSunny, daytime
Video 229920 s2 lanes in each direction and an oncoming onramp, light trafficLow light, dusk
Video 329920 s5 lanes in each direction, heavy trafficRain, daytime
Video 489960 s4 lanes in each direction, heavy trafficSunny, daytime
Table 2. Proposed technique foreground segmentation evaluation against different recent algorithms for Video 1.
Table 2. Proposed technique foreground segmentation evaluation against different recent algorithms for Video 1.
RecallCAPrecisionF-ScoreAve. Comp. Time (s)
Proposed Algorithm 0.89880.99090.82880.86190.0089
FPCP0.97400.98200.65340.78150.0117
LRSMDTNNorm0.97750.98100.63930.77240.579
GA0.98520.56650.06810.12660.0111
GreGoDec0.97660.98060.63570.76940.0239
RPCP0.73740.98680.83220.78070.1288
FW-T0.77880.98740.82260.79900.0487
MoG-RPCA0.97830.98140.64570.77730.1659
Lag-SPCP-QN0.00940.96840.97780.02150.4898
NonConvRPCA0.97620.98050.63500.76880.0337
Table 3. Proposed technique foreground segmentation evaluation against different recent algorithms for Video 2.
Table 3. Proposed technique foreground segmentation evaluation against different recent algorithms for Video 2.
RecallCAPrecisionF-ScoreAve. Comp. Time (s)
Proposed Algorithm0.84130.99660.77280.80290.0022
FPCP0.95040.99190.52770.67750.0119
LRSMDTNNorm0.96370.99050.48270.64200.458
GA0.96720.44250.01410.02770.0111
GreGoDec0.96170.99040.47860.63760.0189
RPCP0.32500.99430.91940.51370.1102
FW-T0.59590.99530.81210.67330.0348
MoG-RPCA0.96380.99010.47180.63240.1054
Lag-SPCP-QN0.34010.99400.88340.45950.2327
NonConvRPCA0.96170.99040.47860.63760.0263
Table 4. Proposed technique foreground segmentation evaluation against different recent algorithms for Video 3.
Table 4. Proposed technique foreground segmentation evaluation against different recent algorithms for Video 3.
RecallCAPrecisionF-ScoreAve. Comp. Time (s)
Proposed Algorithm0.70030.97020.63450.66480.0022
FPCP0.85490.93720.39910.54220.0150
LRSMDTNNorm0.96870.98640.61590.75180.49
GA0.86300.74110.13720.23390.0135
GreGoDec0.86430.93460.38870.53430.0203
RPCP0.51160.96300.58390.54140.1587
FW-T0.63880.96100.54750.58700.0513
MoG-RPCA0.86160.94170.42400.56630.3055
Lag-SPCP-QN0.01950.95810.85310.04380.4203
NonConvRPCA0.86430.93460.38870.53430.0241
Table 5. Proposed technique foreground segmentation evaluation against different recent algorithms for Video 4.
Table 5. Proposed technique foreground segmentation evaluation against different recent algorithms for Video 4.
RecallCAPrecisionF-ScoreAve. Comp. Time (s)
Proposed Algorithm0.87510.99270.82020.84400.0067
FPCP0.96210.98760.63700.76520.0112
LRSMDTNNorm0.96870.98640.61590.75180.626
GA0.96200.15570.02540.04910.0104
GreGoDec0.96700.98620.60640.74360.0211
RPCP0.71300.98970.80930.75510.1468
FW-T0.59560.98850.85260.69790.0381
MoG-RPCA0.93970.98470.61040.73680.1865
Lag-SPCP-QN0.05430.97730.95540.09640.8091
NonConvRPCA0.96700.98620.60640.74360.0441
Table 6. Confusion matrix for Video 1 for all frames in ROI.
Table 6. Confusion matrix for Video 1 for all frames in ROI.
Predicted Class
Passenger CarPassenger TruckMultiple Vehicle ClusterNon-Vehicle
Actual ClassPassenger Car47418031
Passenger Truck29191433
Multiple Vehicle Cluster521193
Non-Vehicle63090
Table 7. Different evaluation metrics for the detected classes within Video 1.
Table 7. Different evaluation metrics for the detected classes within Video 1.
Vehicle ClassPrecisionRecallF-Score
Passenger Car92%91%91%
Passenger Truck89%74%81%
Multiple Vehicle Cluster97%92%94%
Non-Vehicle57%91%70%
Table 8. Multiple vehicle cluster detections confusion matrix for Video 1.
Table 8. Multiple vehicle cluster detections confusion matrix for Video 1.
Predicted Class
Passenger CarPassenger TruckNon-Vehicle
Actual ClassPassenger Car120024
Passenger Truck247224
Non-Vehicle000
Table 9. Different evaluation metrics for detected vehicles in multiple vehicle clusters within Video 1.
Table 9. Different evaluation metrics for detected vehicles in multiple vehicle clusters within Video 1.
Vehicle ClassPrecisionRecallF-Score
Passenger Car83%83%83%
Passenger Truck99%60%75%
Non-Vehicle000
Table 10. Confusion matrix for all vehicles in Video 2.
Table 10. Confusion matrix for all vehicles in Video 2.
Predicted Class
Passenger CarPassenger TruckSmall Utility TruckMultiple Vehicle ClusterNon-Vehicle
Actual ClassPassenger Car10410015
Passenger Truck0731012
Small Utility Truck011200
Multiple Vehicle Cluster000122
Non-Vehicle16900115
Table 11. Different evaluation metrics for the detected classes within Video 2.
Table 11. Different evaluation metrics for the detected classes within Video 2.
Vehicle ClassPrecisionRecallF-score
Passenger Car87%87%87%
Passenger Truck87%85%86%
Small Utility Truck92%92%92%
Multiple Vehicle Cluster99%86%92%
Non-Vehicle80%82%81%
Table 12. Multiple vehicle cluster detections confusion matrix for Video 2.
Table 12. Multiple vehicle cluster detections confusion matrix for Video 2.
Predicted Class
Passenger CarPassenger TruckSmall Utility TruckNon-Vehicle
Actual ClassPassenger Car13011
Passenger Truck1512
Small Utility Truck2001
Non-Vehicle0000
Table 13. Different evaluation metrics for detected vehicles in multiple vehicle clusters within Video 2.
Table 13. Different evaluation metrics for detected vehicles in multiple vehicle clusters within Video 2.
Vehicle ClassPrecisionRecallF-Score
Passenger Car81%87%84%
Passenger Truck99%56%71%
Small Utility Truck000
Non-Vehicle000
Table 14. Confusion matrix for all vehicles in Video 3.
Table 14. Confusion matrix for all vehicles in Video 3.
Predicted Class
Passenger CarPassenger TruckMotorcycleBusSmall Utility TruckLarge Utility TruckMultiple Vehicle ClusterNon-Vehicle
Actual ClassPassenger Car350130000992
Passenger Truck1637100001378
Motorcycle00000010
Bus00000002
Small Utility Truck00010101
Large Utility Truck000065515
Multiple Vehicle Cluster590001614910
Non-Vehicle4921000164609
Table 15. Different evaluation metrics for the detected classes within Video 3.
Table 15. Different evaluation metrics for the detected classes within Video 3.
Vehicle ClassPrecisionRecallF-Score
Passenger Car83%75%79%
Passenger Truck90%78%83%
Motorcycle000
Bus000
Small Utility Truck000
Large Utility Truck75%82%79%
Multiple Vehicle Cluster63%79%70%
Non-Vehicle76%82%79%
Table 16. Confusion matrix for multiple vehicle cluster detections for Video 3.
Table 16. Confusion matrix for multiple vehicle cluster detections for Video 3.
Predicted Class
Passenger CarPassenger TruckMotorcycleBusSmall Utility TruckLarge Utility TruckNon-Vehicle
Actual ClassPassenger Car2632002555
Passenger Truck172121215
Motorcycle00000021
Bus42500223
Small Utility Truck0114101231
Large Utility Truck3000116
Non-Vehicle0000000
Table 17. Different evaluation metrics for detected vehicles in multiple vehicle clusters within Video 3.
Table 17. Different evaluation metrics for detected vehicles in multiple vehicle clusters within Video 3.
Vehicle ClassPrecisionRecallF-Score
Passenger Car97%80%88%
Passenger Truck83%77%80%
Motorcycle000
Bus000
Small Utility Truck000
Large Utility Truck4.5%9.1%6.1%
Non-Vehicle000
Table 18. Confusion matrix for all vehicles in Video 4.
Table 18. Confusion matrix for all vehicles in Video 4.
Predicted Class
Passenger CarPassenger TruckBusSmall Utility TruckLarge Utility TruckMultiple Vehicle ClusterNon-Vehicle
Actual ClassPassenger Car48230000377
Passenger Truck61539110762
Bus01180003
Small Utility Truck2501020115
Large Utility Truck02054209
Multiple VehicleCluster5702023124
Non-Vehicle302804511063
Table 19. Different evaluation metrics for the detected classes within Video 4.
Table 19. Different evaluation metrics for the detected classes within Video 4.
Vehicle ClassPrecisionRecallF-score
Passenger Car83%81%82%
Passenger Truck88%80%84%
Bus95%82%88%
Small Utility Truck89%82%85%
Large Utility Truck89%72%80%
Multiple Vehicle Cluster95%86%90%
Non-Vehicle85%94%89%
Table 20. Confusion matrix for multiple vehicle cluster detections for Video 4.
Table 20. Confusion matrix for multiple vehicle cluster detections for Video 4.
Predicted Class
Passenger CarPassenger TruckBusSmall Utility TruckLarge Utility TruckNon-Vehicle
Actual ClassPassenger Car216701393
Passenger Truck1310805111
Bus292036
Small Utility Truck3724151
Large Utility Truck31200713
Non-Vehicle000000
Table 21. Different evaluation metrics for detected vehicles in multiple vehicle clusters within Video 4.
Table 21. Different evaluation metrics for detected vehicles in multiple vehicle clusters within Video 4.
Vehicle ClassPrecisionRecallF-score
Passenger Car82%87%84%
Passenger Truck81%78%80%
Bus50%9.1%15%
Small Utility Truck69%69%69%
Large Utility Truck20%13%16%
Non-Vehicle000
Table 22. Different evaluated DL architectures with different feature extractor networks.
Table 22. Different evaluated DL architectures with different feature extractor networks.
Faster R-CNNYOLO 2YOLO 3YOLO 4 COCOYOLO 4 Tiny COCO
Resnet 18
Resnet 50
Restnet 101
Inception V3
Inception resnet V2
Darknet 53
Squeeznet
Goolgenet
Table 23. Average precision for tested techniques.
Table 23. Average precision for tested techniques.
Tested AlgorithmVideo 1Video 2Video 3Video 4
YOLO V2—resnet180.420.380.350.40
YOLO V2—resnet500.470.390.400.40
YOLO V2—resnet1010.500.430.420.45
YOLO V2—inceptionv30.510.440.390.43
YOLO V2—inceptionresnetv20.520.430.400.47
YOLO V2—darknet190.570.460.360.55
YOLO V2—squeeznet0.440.430.400.42
YOLO V2—googlenet0.440.390.330.41
YOLO V2—darknet530.560.520.410.53
YOLO V3—resnet180.520.520.350.50
YOLO V3—resnet500.530.410.390.47
YOLO V3—resnet1010.580.430.460.56
YOLO V3—inceptionv30.470.460.390.44
YOLO V3—inceptionresnetv20.410.400.350.45
YOLO V3—squeeznet0.550.530.420.52
YOLO V3—googlenet0.530.500.430.46
YOLO V3—darknet530.580.540.440.56
YOLO V4 COCO—darknet530.610.580.460.59
YOLO V4 Tiny COCO—darknet530.510.500.410.52
Faster RCNN—resnet180.450.420.350.44
Faster RCNN—resnet500.450.400.310.44
Faster RCNN—resnet1010.490.420.350.48
Faster RCNN—darknet530.510.480.380.47
Proposeds Algorithm0.850.830.710.81
Table 24. Summary of timing analysis for each sample video.
Table 24. Summary of timing analysis for each sample video.
NameNumber of FramesRun Time (s)Average Frame Rate (s)
Video 129922.7240.076
Video 229920.1230.0673
Video 329916.4750.0551
Video 459955.1080.092
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yousef, A.; Flora, J.; Iftekharuddin, K. Monocular Camera Viewpoint-Invariant Vehicular Traffic Segmentation and Classification Utilizing Small Datasets. Sensors 2022, 22, 8121. https://doi.org/10.3390/s22218121

AMA Style

Yousef A, Flora J, Iftekharuddin K. Monocular Camera Viewpoint-Invariant Vehicular Traffic Segmentation and Classification Utilizing Small Datasets. Sensors. 2022; 22(21):8121. https://doi.org/10.3390/s22218121

Chicago/Turabian Style

Yousef, Amr, Jeff Flora, and Khan Iftekharuddin. 2022. "Monocular Camera Viewpoint-Invariant Vehicular Traffic Segmentation and Classification Utilizing Small Datasets" Sensors 22, no. 21: 8121. https://doi.org/10.3390/s22218121

APA Style

Yousef, A., Flora, J., & Iftekharuddin, K. (2022). Monocular Camera Viewpoint-Invariant Vehicular Traffic Segmentation and Classification Utilizing Small Datasets. Sensors, 22(21), 8121. https://doi.org/10.3390/s22218121

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop