Next Article in Journal
Robust Human Face Emotion Classification Using Triplet-Loss-Based Deep CNN Features and SVM
Next Article in Special Issue
Progressively Unsupervised Generative Attentional Networks with Adaptive Layer-Instance Normalization for Image-to-Image Translation
Previous Article in Journal
Autoencoders Based on 2D Convolution Implemented for Reconstruction Point Clouds from Line Laser Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rethinking Feature Generalization in Vacant Space Detection

Faculty of Electrical and Electronics Engineering, HCMC University of Technology and Education, Ho Chi Minh City 7000, Vietnam
Sensors 2023, 23(10), 4776; https://doi.org/10.3390/s23104776
Submission received: 17 March 2023 / Revised: 12 May 2023 / Accepted: 12 May 2023 / Published: 15 May 2023
(This article belongs to the Special Issue Artificial Intelligence in Imaging Sensing and Processing)

Abstract

:
Vacant space detection is critical in modern parking lots. However, deploying a detection model as a service is not an easy task. As the camera in a new parking is set up at different heights or viewing angles from the original parking lot where the training data are collected, the performance of the vacant space detector could be degraded. Therefore, in this paper, we proposed a method to learn generalized features so that the detector can work better in different environments. In detail, the features are suitable for a vacant detection task and robust to environmental change. We use a reparameterization process to model the variance from the environment. In addition, a variational information bottleneck is used to ensure the learned feature focus on only the appearance of a car in a specific parking space. Experimental results show that performances on a new parking lot increase significantly when only data from source parking are used in the training phase.

1. Introduction

Vacant space detection systems can help drivers quickly and easily find available parking spaces, reducing the time and frustration associated with circling a lot looking for a space. This can also help to reduce congestion in the parking lot, improve traffic flow, and reduce emissions from idling vehicles. Additionally, parking lot managers can use the data collected by these systems to optimize parking lot layouts and improve overall efficiency.
Recently, many sensors have been available for use in vacant space detection. For instance, ultrasonic sensors [1,2] use sound waves to detect the presence or absence of a vehicle in a parking space. They are typically installed above each parking space and can accurately detect the distance to the nearest object such as a vehicle. Magnetic sensors [3,4] use a magnetic field to detect the presence or absence of a vehicle in a parking space. They are typically installed beneath the surface of the parking space and can detect changes in the magnetic field caused by the presence of a vehicle. Infrared sensors [5] use infrared light to detect the presence or absence of a vehicle in a parking space. They are typically installed above each parking space and can detect changes in the infrared light caused by the presence of a vehicle. Video cameras [6,7,8,9,10,11] can be used to detect the presence or absence of a vehicle in a parking space. They are typically installed above each parking space and use image analysis software to detect changes in the video feed caused by the presence of a vehicle.
Among these possible sensors, many research works focus on vision-based systems to detect vacant spaces because a single camera can manage multi-parking slots. Several challenges should be addressed to have a workable vacant space detector, such as mutual occlusion patterns, shadow, layout, and size variances. From a literature review, supervised learning methods (Vu 2019 [12] and Huang 2017 [13]) require massive labeled data for a high accuracy of 99.74%. However, the domain gap could significantly reduce accuracy when the model is deployed in a new parking lot. As shown in Figure 1, conventional methods may make a wrong prediction in a new testing domain, while our method can learn generalized features that work robustly in a new testing domain.
Transfer learning [14] and unsupervised domain adaptation [15] are popular solutions to relieve the above-mentioned problem. Transfer learning [14] uses a pre-trained model to initialize the weight of the target network. As the pre-trained model is trained based on a vast dataset, its parameters could extract semantic information. Hence, the target model could converge faster with fewer training data. In the vacant space application, the pre-trained model could be trained based on data from a source parking lot; and the target network is a detector deployed in a new parking lot. Given a powerful pre-trained model, transfer learning requires enough newly labeled samples to fine-tune the target network to ensure success.
On the other hand, unsupervised domain adaptation [15] requires labeled data on a source domain and unlabeled data on a target domain. An adversarial loss [16] is used to train the sharing features between two domains, and labeled data in the source domain will be used to train a classifier based on the sharing features. To ensure the success of domain adaptation, the learned sharing features should be good enough for the specific task.
Both transfer learning and unsupervised domain adaptation need help from a pre-trained feature extractor trained on the source dataset. These features should be highly relevant to vacant states and invariant from environmental changes. Since different camera views create different occlusion patterns, learning in-variant features from an unseen domain (target domain) is challenging. To address the challenge, the paper tries to learn features that represent only objects (parking cars) but cannot present the background; even the background includes partial neighbor cars. We use mutual information [17] to measure the amount of information shared by (features and objects) as well as (features and backgrounds); the network is trained with an additional VIB loss.
The idea behind the proposed method is quite similar to the concept in TCL2021 [18] regarding feature learning. TCL2021 [18] uses a trajectory with sequential frames to train the detector. The frames are selected before and after a car moves in/out of a slot; hence, they have pairwise vacant and occupied samples that share the same background but diffident vacant states. As shown in Figure 2, the pairwise samples in a trajectory had similar appearances; the only difference between the samples is at some pixels where the car is parked. These pairing data function as a contrastive loss where two samples in a trajectory are forced to move far from each other; hence, the model focuses on necessary pixels and eliminates all background pixels. While many promising results have been reported in TCL2021 [18], this method relies on a motion classifier to prepare training trajectories. Moreover, the training process needs a task-consistency reward to train the vacant-space detector in a reinforcement manner. This makes the training process more complex and requires powerful hardware to store all frames in a trajectory for every iteration update.
Unlike the complex training process in TCL2021 [18], our work learns crucial features using a simple process. Here, the model includes a feature extractor and a classifier. Given an input image x, the extractor extracts the feature μ R d , and the corresponding variance σ R d . As the classifier must work well with variant features, we use a reparameterization trick to add some uncertainty to the training process. In detail, a latent feature z is sampled from the distribution z N ( μ , σ 2 ) ; then, the vacant state y is predicted from sampled feature z by y = c l s ( z ; θ c l s ) .
A good feature z must represent occlusion patterns that predict well vacant states. Hence, the mutual information I ( y ; z ) should be maximized. Additionally, the background information from input x should be eliminated in z. Hence, the mutual information I ( x ; z ) should be small. Two constraints are optimized together in a variational information bottleneck (VIB) loss [19]. Since this loss could be integrated into any supervised learning framework, the training process becomes simpler than task consistency learning in TCL2021 [18].
In short, the contributions of this paper are the following:
  • We propose a learning method to learn better features for vacant parking space detection. The trained model can work better on the unseen domains, while no data from the unseen domain are available in the training phase;
  • A reparameterization trick is used to learn a classifier that is well adapted to environmental changes;
  • A variational information bottleneck loss is used to learn features focused on occlusion patterns and eliminate the background.

2. Related Works

Vacant space detection for vehicle parking has been an interesting research topic for many years. Early research Lin 1997 [6], Raja 1999 [20], Raja 2000 [7], and Raja 2001 [21] relied on feature engineering processes to detect a car in a specific position. Given a selected location, Lin 1997 [6] introduces geometrical models with a spatial contour matching approach and a careful tuning process to fit well the specific scenario. Raja 1999 [20], and Raja 2001 [21] proposed a learning-based approach to model the unknown distribution of images that contain vehicles by utilizing the higher-order statistics (HOS) information as the feature. Raja 1999 [7] advanced to detect and track vehicles in videos by combining statistical knowledge about the visual appearance of vehicles with their motion information. However, most of the work focuses on a few slots but does not use only one camera to manage all slots in a wide parking area. This may limit particular applications.
With the development of CCTV systems, applying vehicle detection algorithms in the parking lot for vacant space detection has become possible in the recent decade. Therefore, much research has been conducted to pursue desirable performance and overcome challenging issues in practice (Paolo 2013 [22], Lixia 2012 [23], Wu 2007 [24], Huang 2010 [25], and Huang 2013 [26]). Here, images captured by CCTV cameras may include many slots. A detector must address both localization and classification tasks. Since a car should be parked in a specific 3D slot, prior 3D information could be used to address the localization task. In Huang 2010 [25], geometry and image information are fused to generate all vacant states of all slots. Additionally, neighbor slots help to correct the prediction at a query slot via Bayesian inference. Here, six 3D surfaces represent the 3D cube, and each surface is projected to a 2D image to find a corresponding region. Each region is processed independently, and a hierarchy Bayesian inference fuses predictions from these planes to a final vacant state. The Bayesian hierarchical structure can model the occlusion relationship among all neighboring slots to improve accuracy. However, the inference time of Huang 2010 [25] is too long, and the computing cost needs to be lowered to fit the real-world application. Later, in Huang 2013 [26], the Bayesian hierarchy is replaced by a multi-layer inference framework to learn the correction process. While Huang 2013 [26] presented multi-processing steps, Huang 2015 [27] models each processing step as a layer of a unique model. It helps the solution of Huang 2015 [27] to be simpler than the Bayesian hierarchy Huang 2013 [26]. Although considerable results are achieved, these methods require experts to tune hyper-parameters manually for each inference layer.
With the rapid development of deep learning techniques, many applications are solved using the deep model. The multi-layer in Huang 2015 [27] could be replaced by a unique deep learning model in Huang 2017 [13]. Each layer in Huang 2015 [27] is modeled by a specific block; later, the network is trained end-to-end. Following the experimental result in Huang 2017 [13], a detector may make a wrong prediction if a car is too big or parks not at the center of a slot. Therefore, in Huang 2017 [13], a spatial transform network is used to select a suitable region of the input image for classification. To ensure the input image can cover all necessary information so that a spatial transform could select a suitable ROI, the authors of Huang 2017 [13] used a normalized three-parking image to represent input and gain higher accuracy in real-time. Next, Vu 2019 [12] provides a contrastive loss to learn better feature representation. Using a deep learning framework, the methods in Huang 2017 [13] and Vu 2019 [12] perform better than traditional approaches without fancy feature engineering processes or heavy hyper-parameter adjustment. To deploy the vacant detector in a new parking lot, Zeng 2022 [28] further generalized the deep learning-based approaches to significantly different lighting scenarios with the adversarial domain adaptation technique. Both Vu 2019 [12] and Zeng 2020 [28] reused the normalized three-parking-space image setting in Huang 2017 [13] and obtained high-accuracy results.
Although promising results were reported in Vu 2019 [12], and Zeng 2022 [28], the success of these deep learning approaches comes from a vast dataset. Unfortunately, due to the domain gap issue, using these approaches requires labeling once for each parking lot, which causes enormous manual effort. To solve this issue, TCL2021 [18] proposes using an optical-flow-based motion classifier as the guidance to train the vacant space detector. Specifically, TCL2021 [18] found that RGB-image-based models are affected seriously by camera poses or various lighting conditions. By contrast, the motion classifier, which takes a sequence of optical flow to predict the motion state of a trajectory, is robust to such factors. Therefore, the authors train a vacant space detector using consistency with a flow-based motion classifier. The motion classifier could be trained in a parking lot; later, it helps to train vacant detectors in another parking lot. Using consistency between the two models, the vacant space detector on a new parking lot could be trained without human labor. However, the motion classifier may make a wrong prediction with high confidence; this phenomenon seriously affects the trained detector. Moreover, estimating a consistency reward is complicated, and this method requires high-capacity hardware to store all frames in a trajectory.

3. Proposed Method

3.1. Problem Statement

Given a parking lot monitored by a surveillance camera mounted at a high position, we would like to train a vacant space detector to detect a vacant state at any slot in the parking space. However, unlike the conventional methods (Huang 2017 [13], and Vu 2019 [12]), where training and testing images are from the same parking lot, our work uses training images and testing images from different domains. Figure 1 shows the images captured from two domains (different angle views). The domain gap by camera poses results in different appearances in the same parking lot.
Conventionally, a detector must address localization and classification tasks. In a vacant space detector, the localization task could be well addressed by the 3D position of a parking slot. Huang 2017 [13] used a 3D cube to model each slot, and the projection of the cube to a 2D image serves as a localization result. Hence, we can focus on the classification task provided by the local patch. Following a careful design, Vu 2019 [12] introduced a three-space normalization as an input of the classifier. Here, three neighbor slots are concatenated and projected on a 2D image. The projection is normalized as an input of a classifier. Some examples of the normalization samples are shown in Figure 2.
Denote as D S = { x i s , y i s } i = 1 N a dataset collected on the source domain in Figure 1. We separate it into training and validating datasets ( D T r a i n S , D V a l S ) . A detector is trained and validated using these datasets; then, the detector is tested on a testing dataset D T e s t T . Here, D T e s t T is collected on the target domain using the same 3-space-normalization method. Note that D T r a i n S and D T e s t T are collected on different days, and no image in the target domain is used during the training process. If the trained model predicts well in the target dataset, the method could learn generalized features invariant to the camera angle. Some examples of D T r a i n S and D T e s t T are in Figure 2.

3.2. System Overview

Figure 3 presents the overview of the proposed method. First, a pre-trained model extracts flattened feature f = C N N ( x ) R D . The features are fed forward to an encoder to generate μ R d and σ R d vectors. Two vectors are used to estimate the Kullback–Leibler divergence loss [29]; additionally, a reparameterization process samples a new latent z. A classifier takes the latent z and predicts the vacant state y { v a c a n t , o c c u p y } .
The paper uses VGG16 [30] as a backbone (feature extractor and neck) to provide a 512 × 7 × 7 feature map. However, the extractor could be any pre-trained backbone. In the encoder module, linear layers encode the flattened feature to a μ , σ . Table 1 summarizes the network architecture.

3.3. Variational Information Bottleneck (VIB)

3.3.1. Overview of VIB Loss for Vacant Space Detector

Conventionally, a deep neuron network uses an extractor to map input image x to a feature space z; the feature is fed to a classifier to predict an output state y. A generalized feature should be invariant to environmental factors so the classifier can predict well in the testing phases. In a vacant space detector, learning generalized features is a very challenging task. As shown in Figure 2, the vacant and occupied samples are similar because they share the same background. Both of them include cars in the images; the only difference is that occlusion patterns may occur while a car is parked in the middle slot. The conventional method may learn many features (e.g., car patterns) and use the pattern to predict the vacant state. These features can work well to predict the vacant state in the same domain but will be affected if the domain changes (e.g., an angle view). A good feature for vacant space detection is a very compressed feature that focuses only on the occlusion pattern in the mid-slot.
The information bottleneck principle [31] introduces a view of supervised learning to learn these good features. The principle says that the optimal model transmits as much information as possible from its input to its output through a compressed representation called the information bottleneck. In vacant space detectors, the information bottleneck could be explained by the following two constraints:
  • The latent z must help to predict well the output y (vacant state);
  • The latent z must be compressed to the input x (3-space normalization images). This means that when we know z, we cannot infer x very well.
I ( X ; Y ) = I ( Y ; X ) = H ( X , Y ) H ( X | Y ) H ( Y | X ) = H ( X , Y ) H ( X | Y ) H ( Y | X ) = K L p ( x , y ) | | p ( x ) p ( y ) x X , y Y = E ( x , y ) p ( x , y ) log p ( x , y ) p ( x ) p ( y ) = p ( x , y ) log p ( x , y ) p ( x ) p ( y ) d x d y .
From information theory, the mutual information [17] can measure the information of a variable via another variable as in Equation (1). Hence, the above constraints could be explained by maximizing the I ( y ; z ) and minimizing I ( x ; z ) . Denote as I c the information constraint that I ( x ; z ) < I c ; the Lagrangian of the above-constrained optimization problem is a maximization solution of Equation (2). Here, β is a Lagrange multiplier.
L I B = I ( y ; z ) β I ( x ; z ) I c I ( y ; z ) β I ( x ; z ) .

3.3.2. Classification Loss in VIB

By applying the definition in Equation (1), the first term I ( y ; z ) could be rewritten as Equation (3). Here, p ( y z ) is defined by an intractable function as Equation (4).
I ( y ; z ) = p ( y , z ) log p ( y , z ) p ( y ) p ( z ) d y d z = p ( y , z ) log p ( y z ) p ( y ) d y d z = p ( y , z ) log p ( y z ) d y d z p ( y ) log p ( y ) d y .
p ( y z ) = p ( y , z ) p ( z ) = p ( x , y , z ) p ( z ) d x = p ( z x ) p ( y x ) p ( x ) p ( z ) d x .
Since Equation (4) is intractable, the authors of [19] used a lower bound to approximate p ( y z ) . Denote as q ( y z ) a variational approximation of p ( y z ) ; the lower bound is founded by Kullback–Leibler divergence as in Equation (5).
K L p ( y z ) | | q ( y z ) 0 p ( y z ) log p ( y z ) d y p ( y z ) log q ( y z ) d y .
Using the lower bound in Equation (5), the I ( y ; z ) in Equation (3) is rewritten as Equation (6).
I ( y ; z ) p ( y , z ) log q ( y z ) d y d z p ( y ) log p ( y ) d y p ( z x ) p ( y x ) p ( x ) log q ( y z ) d x d y d z p ( z x ) p ( y , x ) log q ( y z ) d x d y d z .
where the entropy of the labels H ( y ) = p ( y ) log p ( y ) d y is independent and could be ignored.

3.3.3. Feature Selection Loss in VIB

The second term I ( x ; z ) is rewritten as Equation (7).
I ( x ; z ) = p ( x , z ) log p ( x , z ) p ( x ) p ( z ) d x d z = p ( x , z ) log p ( z x ) p ( z ) d x d z .
Let q ( z ) be a variational approximation to the marginal p ( z ) . Using the KL divergence, the upper bound of I ( x ; z ) is introduced as Equation (8).
K L p ( z ) | | q ( z ) 0 p ( z ) log p ( z ) d z p ( z ) log q ( z ) d z .
Applying the upper bound to Equation (7), the I ( x ; z ) is rewritten as Equation (9).
I ( x ; z ) = p ( x , z ) log p ( z x ) d x d z p ( z ) log p ( z ) d z p ( x , z ) log p ( z x ) d x d z p ( z ) log q ( z ) d z = p ( x ) p ( z x ) log p ( z x ) q ( z ) d x d z = p ( x , y ) p ( z x ) log p ( z x ) q ( z ) d x d z d y .

3.3.4. Approximation Loss Function

By applying the lower bound of the I ( y ; z ) and the upper bound of I ( x ; z ) , the Lagrangian function in Equation (2) is approximated as
L I B = I ( y ; z ) β I ( x ; z ) p ( z x ) p ( y , x ) log q ( y z ) d x d y d z β p ( z x ) p ( x , y ) K L p ( z x ) | | q ( z ) d x d y d z = E ( x , y ) p ( x , y ) , z p ( z x ) log q ( y z ) β K L p ( z x ) | | q ( z ) = J I B .
In a practical application where the dataset ( x i , y i ) i = 1 . . N is available, the J I B is estimated as Equation (11).
J I B = 1 n i = 1 n p ( z x i ) log q ( y i z ) d z β K L p ( z x i ) | | q ( z ) .
To model the loss J I B as a deep learning framework, ref. [19] use an encoder Φ to predict μ Φ ( x ) , Φ ( x ) . Then, p Φ ( z x ) is modeled as a multivariate Gaussian p Φ ( z x ) =   N z ; μ Φ ( x ) , Φ ( x ) . The latent z could be sampled from a reparameterization trick g ( ϵ , x ) where ϵ p ( ϵ ) = N ( 0 , I ) . Thus, the final objective is a minimized solution of the loss function in Equation (12). Here, n is the number of training samples.
J I B = 1 n i = 1 n β K L p Φ ( z x i ) | | q ( z ) E ϵ p ( ϵ ) log q y i g ( ϵ , x i ) .
Here, the term q ( y g ( ϵ , x ) ) serves as a classifier and it is learned by a binary cross entropy loss as Equation (13), whereas q ( z ) is the approximated latent marginal often fixed to a standard normal distribution as Equation (14)
log q ( y g ( ϵ , x ) ) = y log y ^ + ( 1 y ) log ( 1 y ^ ) .
q = N z ; 0 , I k .

4. Results

4.1. Datasets and Implementation Detail

Two training datasets are collected on the source domain in Figure 1. The first training dataset D T r a i n 1 S is a standard training dataset collected over 30 days and labeled for supervised learning. We collect one image every 30 min and only in the daytime—from 6 a.m. to 6 p.m. The dataset includes 14,667 vacant slots and 20,021 occupied slots. The second training dataset D T r a i n 2 S includes samples collected from trajectories as in TCL2021 [18]. In detail, TCL2021 [18] uses the magnitude of optical flow to detect a time slot when a car is moving. Given the optical flow in the time slot, a motion classifier is used to estimate the motion state of this car. On both sides (before and after) of the time slot, we have no-motion segments in which no car moves, as in Figure 4. We sample one sample for each no-motion segment. Then, two samples are added to the D T r a i n 2 S dataset. As a result of the sampling method, the pair samples have similar backgrounds but only differences in the middle slot, as shown in Figure 2 and Figure 4. This dataset includes 2000 samples from 1000 trajectories. Hence, D T r a i n 1 S is a vast and diverse dataset that may make it easy to train a vacant space detector; in contrast, D T r a i n 2 S is small but has many similar images in the dataset. This property allows D T r a i n 2 S to work as a contrastive loss where pairwise samples are trained together.
In the testing phase, two datasets are prepared. The first dataset D V a l S includes 11,944 vacant samples and 7832 occupied slots collected from the other 15 days on the source parking lot from 6 am to 6 pm. The second dataset D T e s t T is of the same scale as the D V a l S dataset but it is collected from the target domain. We use the two testing datasets to evaluate the effects of the domain gap and the generalization of models trained by the proposed method.
We follow the suggestion in [32] to implement the VIB method. However, instead of using a sum operator to estimate the loss over training patches, we use a mean operator. Given an input x and the corresponding latent z R d , the sum operator estimates KL loss d times, but the classification loss is only estimated once for each sample. In this situation, gradients from the KL loss may destroy the original backbone. Using the mean operator, the losses are averaged on both the feature and patch size dimensions. Hence, the model can converge smoothly.
In addition, the training hyper-parameters are listed below:
  • Optimizer: SGD with learning rate = 0.01.
  • Scheduler: step_size = 30 and gamma = 0.1.
  • Data normalization: mean = [0.485, 0.456, 0.406]; std = [0.229, 0.224, 0.225].

4.2. Ablation Study

In this section, we discuss the contribution of the KL loss on feature learning. The quality of features depends on the quality of the training dataset. Therefore, we randomly select a subset of D T r a i n 1 S to train the detector. Denote as S 1 , S 2 , S 3 , S 4 , and S 5 subsets from D T r a i n 1 S . The number of samples in S 1 , S 2 , S 3 , S 4 , and S 5 is 2000, 5000, 10,000, 20,000, and 34,688, respectively. As the S 1 dataset has 2000 samples, it is equivalent to D T r a i n 2 S in scale. However, differences are in the content of D T r a i n 2 S where pairwise vacant/occupy images share the same background.
We first use subsets of D T r a i n 1 S to evaluate the generalization of a model. In this scenario, the training dataset is diverse. The accuracy values for D V a l S and D T e s t T are in Table 2 and the F1 score for D V a l S and D T e s t T are in Table 3. During the training process, we evaluate the model using D V a l S for every epoch. The accuracy for D V a l S is used to select the best model in the training phase.
Regularization methods (e.g., batch normalization, dropout) are possible solutions to avoid overfitting on a dataset and learn generalization features. Hence, we use batch normalization and dropout on the conventional supervised learning method to evaluate regularization methods on the application. Our experiment is based on the VGG model to detect vacant space. The VGG model includes a feature extractor, a neck, and a classifier. Hence, we add a batch normalization after the feature extractor and denote this setting as “No VIB (BN)” in Table 2 and Table 3. In the classifier module (VGG network), the dropout layer had been applied with p = 0.5 by default. Hence, the default setting is denoted as “No VIB (p = 0.5)”. To attain better generalization features, we test the performance when p = 0.7 and p = 0.9. The notations of these settings are “No VIB (p = 0.7)” and “No VIB (p = 0.9)”, correspondingly.
Following the result in Table 2 and Table 3, several conclusions could be drawn:
  • The domain gap is a critical challenge in deploying a new parking lot. In Table 2, using the S 2 dataset (5000 samples), the conventional method (No VIB ( p = 0.5 )) can achieve 99.46% accuracy on D V a l S but it can only achieve 89.68% accuracy on D T e s t T ;
  • To learn generalized features, a vast dataset is needed. In the scenario “No VIB ( p = 0.5 )” in Table 2, using the S 5 dataset (30,540 samples) for training does not help to increase accuracy when evaluating on the same domain ( D v a l S ). However, the performance increases significantly when evaluating is based on the target domain ( D t e s t T ). In detail, when the S 2 dataset is used for training, the accuracy on D t e s t T is only 89.68%; however, when the S 5 dataset is used for training, the corresponding accuracy is 92.02%. A similar observation can be found in Table 3;
  • The proposed method helps to increase the performance on the source domain if the training dataset is sparse. When the training set is S 1 dataset (2000 samples), the accuracy and F1-score are 97.95% and 97.54%, correspondingly, on D V a l S . When the KL-loss ( β = 0.1 ) is introduced, the performance improves to 98.44% for accuracy and 98.14% for the F1-Score on the source domain;
  • When the number of training samples increases, the proposed method helps to learn better features that work well in the target domain. The conventional model (No VIB ( p = 0.5 )) can reach up to 88.95% on the D T e s t T dataset. However, there are no apparent differences between the accuracies given by the S 3 (only 10,000 samples) and S 5 datasets. On the target domain, the accuracy given by S 3 is 88.84% and the accuracy given by S 5 is 88.95%. In contrast, given the S 5 datasets, the proposed method can achieve 97.91% accuracy and 97.31% F1 score ( β = 0.1 ). Moreover, the differences in accuracy are apparent if more training samples are available. We can see that the performance inmproves if the number of training samples is increased. In further detail, when S 1 , S 2 , S 3 , S 4 , S 5 are applied for training, the accuracies are 92.93%, 94.73%, 95.28%, 96.51%, 97.91%, correspondingly;
  • The hyper-parameter β should not be too large. When β = 0.2 , the performance is reduced on D T e s t T . This is reasonable because the KL-loss forces a feature to be closer to zero. This means that more spatial regions are not used to predict the vacant state. Following our experiment, β = 0.1 is an optimal selection in most datasets;
  • Batch normalization may not help in the vacant space detection application. Even if the performance on the source domain is high, the model cannot work better on the target domain. The maximum accuracy on the target domain is only 83.53% if BN is applied;
  • The increment of the dropout parameter may help with small training sets. If S 1 or S 2 serve as training sets, the performance improves on both source and target domains. Especially when the S 2 dataset is used for training and the dropout ratio is 0.9, the conventional method performs better than our proposed method. However, when the dataset is larger, the performance cannot improve. The accuracies on the S 2 , S 3 , S 4 , and S 5 datasets are quite similar if p = 0.5 or p = 0.7. In contrast, our VIB-based method can also learn and accept new features from larger training datasets.
We also provide experiments to evaluate the proposed method if the training dataset includes pairwise samples. Here, D T r a i n 2 S serves as a training set and the testing sets are D V a l S and D T e s t T as in previous experiments. Note that pairwise samples in D T r a i n 2 S serve as a contrastive loss during the training phase. The result in Table 4 shows that a pairwise dataset may help to learn better-generalized features. Compared with the result given by the S 1 dataset, the result given by the D T r a i n 2 S dataset is slightly better. The accuracy scores are 98.04% and 97.95% if the training datasets are D T r a i n 2 S and S 1 , correspondingly. Here, we compare the two datasets because they are on the same scale. The pairwise dataset [18] D T r a i n 2 S could help to slightly increase performance. This phenomenon could also be observed when D T e s t T is the testing set. Hence, the pairwise property in D T r a i n 2 S can force the model to learn generalized features.
When applying VIB, the performance on the source domain ( D V a l S ) is not improved. The accuracy is 98.04% without VIB and it is 98.97% with VIB. This means D T r a i n 2 S does not have enough data to learn a better detector. This conclusion is consistent with the result in Table 2 when S 1 is the training set. However, the performance of the target domain is improved significantly. Without VIB, the accuracy on D T e s t T is 85.88% but it could increase to 94.82% when β = 0.1 . In this case, the increment is 8.94%. If the S 1 dataset is used as a training set, the increment is 8.75%. This means that VIB loss can help in both types of datasets.

4.3. Comparison with State-of-the-Art Methods

4.3.1. Comparison with Supervised Learning Methods

In addition, we compare our method with other conventional vacant space detectors, including [12,13,18,22,23,24,33]. The parking lot in our paper is similar to the parking lot in Vu 2019 [12] and TCL2021 [18]. However, Vu 2019 [12] prepared a vast dataset that includes 587,667 samples and TCL2021 [18] used a motion classifier to select pairwise samples. The two papers and our paper use a normalized three-parking slot as the input of detectors. The authors of Vu 2019 [12] also used their dataset to evaluate Paulo2013 [22], Lixia 2012 [23], Wu 2007 [24], Huang2017 [13], and Faster-RCNN [33] methods in a comprehensive comparison. From a dataset viewpoint, our dataset is quite similar to the dataset in Vu 2019 [12] but at a smaller scale. From the training process viewpoint, Vu2019 [12] does not use any pre-trained model, but the proposed method uses the pre-trained VGG model. We also prepared a simple version of the proposed method that does not use VGG as a pre-trained model.
The comparison results in Table 5 show that our method can outperform these vision-based methods as [22,23,24]. Moreover, we also compared with deep-learning-based methods [12,13,33]. Faster-RCNN [33] works for vehicle detection, but it cannot detect well occluded parked cars with small image sizes. In addition, Faster-RCNN needs to address both the localization and classification tasks, while Huang 2017 [13] and Vu 2019 [12] use 3D information to solve the localization task. Hence, their performances are better. Huang 2017 [13] and Vu 2019 [12] can provide higher accuracy for vacant space detection, but they require a vast training dataset with supervised labels. In comparison, our method also achieves an equivalent accuracy with only 5000 samples. However, Huang 2017 [13] and Vu 2019 [12] do not use any pre-trained model, but our method uses a VGG pre-trained model. For a better comparison, we also train our method from scratch with only 5000 training samples. The result shows that without the help from the VGG pre-trained model, the accuracy is reduced by 1%. This degradation is relatively small and could be compensated for when the training dataset is larger. TCL [18] uses 1000 motion trajectories to achieve a good performance, but their training process is complex because of the task consistency reward. Additionally, it requires more RAM in GPU to store all frames in a trajectory. Compared with TCL [18], our proposed method is easier to train.

4.3.2. Comparison with an Upper Bound

In our work, we train and evaluate the vacant space detector on the source domain; additionally, we test it on the target domain. During the training process, no information from the target domain is used. To evaluate our learning method’s performance, we compare the proposed method with unsupervised domain adaptation [28]. In unsupervised domain adaptation [28], the target model uses a source dataset and a target dataset during training phases. However, the source dataset includes label and image information, whereas the target dataset only has image information. Therefore, the unsupervised domain adaptation [28] could be treated as an upper bound of our proposed method.
In this experiment, two cameras at two angle views are set up to collect two datasets. One dataset is the source dataset and another is the target dataset. Each dataset is split into a training dataset and a validation dataset in its domain. The results in Table 6 show that when ‘ 45 ° view’ is used as a source dataset, the performance on the target dataset (‘ 90 view’) reaches the upper bound. However, if ‘ 90 view’ serves as the source dataset, the performance on ‘ 45 ° view’ is far from the upper bound. This means that generalized features rely greatly on the source domain.

4.4. Feature Analysis

In this section, we analyze the feature maps extracted by the proposed method (with VIB) and the conventional method (without VIB). Our model is based on the VGG model that includes a feature extractor, a neck, and a classifier. Given an image, we use the feature extractor to extract feature maps. There are 512 feature maps given one input image. Figure 5 shows one example image and several corresponding feature maps in both cases (with and without VIB). The result shows that the VIB-based method extracts sparse feature maps. This means that the features learned by VIB are precise and cannot be found everywhere on the example image even though cars are at every neighbor slot. Only some spatial locations have a response on the feature map yielded by the VIB-based method. This phenomenon is reasonable because the KL loss force feature maps close to zeros. Similar to the dropout technique, a sparse feature map may avoid overfitting in a dataset and the model becomes generalized. This observation also proves the benefit of the proposed method where the model can serve as a feature selector which:
  • Adaptively selects features suitable for the target task;
  • Skips or removes redundancy features related to input images.
Figure 5. Feature maps extracted by the detector with and without VIB loss.
Figure 5. Feature maps extracted by the detector with and without VIB loss.
Sensors 23 04776 g005
In addition, the proposed method can not only provide sparse feature maps but also extract more empty feature maps where all values are zeros. Given one feature map, we extract the feature map’s minimum, maximum, and mean values. This process is applied to 512 channels. The statistics of these variables are in Table 7. Each column represents one variable (maximum, minimum, and mean value of a feature map), and each row represents one statistics indicator.
By comparing the statistics of the maximum variable between the two methods (with and without VIB), we may see that the number of empty feature maps in the VIB-based method is larger. With VIB, more than 50% of feature maps are empty maps; without VIB, the corresponding value is smaller than 50%. In addition, there is not a huge distance between VIB-based features. With VIB, the maximum values of minimum variables and maximum variables are 7.3204 and 12.6495. Without VIB, the maximum values of minimum variables and maximum variables are 1.2250 and 30.8568. Considering the maximum value, the gap between maximum and minimum variables is 5.3291 and 29.6318 with VIB and without VIB. This means that VIB can serve as a normalization process that reduces the feature variance in feature space.
Following the above discussions, VIB has feature normalization and sparse representation properties. Normalization and sparse representation are well-known solutions for a better generalization model. Hence, the proposed method can learn generalization features. In addition, during the training process, some uncertainty was added to the feature by the reparameterization step. This uncertainty may model some domain shift factors (orientation, camera field of view, camera height) but in feature space. Therefore, the model has the ability to adapt to environmental change.

5. Conclusions

In this paper, we apply VIB [19] to learn generalized features for vacant space detectors. The training/validating datasets are from the source domain, while the testing dataset is from the target domain. Our method is significantly better than conventional methods if the detector is tested on the target domain. Additionally, our method helps improve performance on the source domain if the training dataset is sparse.

Funding

This work was supported by HCMC University of Technology and Education, VietNam. This work was supported in part by the National Science and Technology Council, Taiwan, under Grant NSTC-110-2221-E-A49-066-MY3 and Grant NSTC-109-2221-E-009-112-MY3.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The author declare no conflict of interest.

References

  1. Abad, F.; Bendahan, R.; Wybo, S.; Bougnoux, S.; Vestri, C.; Kakinami, T. Parking space detection. In Proceedings of the 14th World Congress on Intelligent Transport Systems, ITS 2007, Beijing, China, 9–13 October 2007; Volume 2. [Google Scholar]
  2. Hsu, C.M.; Chen, J.Y. Around View Monitoring-Based Vacant Parking Space Detection and Analysis. Appl. Sci. 2019, 9, 3403. [Google Scholar] [CrossRef]
  3. Zhu, J.; Cao, H.; Liu, H. Parking Space Detection Based on Information from Images and Magnetic Sensors. Int. J. Adv. Inf. Sci. Serv. Sci. 2012, 4, 208–216. [Google Scholar] [CrossRef]
  4. Chien, C.F.; Chen, H.T.; Lin, C.Y. A Low-Cost On-Street Parking Management System Based on Bluetooth Beacons. Sensors 2020, 20, 4559. [Google Scholar] [CrossRef] [PubMed]
  5. Kadir, M.; Osman, M.; Othman, N.; Sedek, K.A. IoT based Car Parking Management System using IR Sensor. J. Comput. Res. Innov. 2020, 5, 75–84. [Google Scholar] [CrossRef]
  6. Lin, C.; Parameswaran, V.; Davis, L.; Rosenfeld, A. Site model mediated detection of movable object activities. In RADIUS: Image Understanding for Imagery Intelligence; Morgan Kaufmann: Burlington, MA, USA, 1997; p. 285. [Google Scholar]
  7. Rajagopalan, A.N.; Chellappa, R. Higher-order-statistics-based detection of vehicles in still images. JOSA A 2001, 18, 3037–3048. [Google Scholar] [CrossRef] [PubMed]
  8. An, Q.; Wang, H.; Chen, X. EPSDNet: Efficient Campus Parking Space Detection via Convolutional Neural Networks and Vehicle Image Recognition for Intelligent Human–Computer Interactions. Sensors 2022, 22, 9835. [Google Scholar] [CrossRef] [PubMed]
  9. Encío, L.; Díaz, C.; del Blanco, C.R.; Jaureguizar, F.; García, N. Visual Parking Occupancy Detection Using Extended Contextual Image Information via a Multi-Branch Output ConvNeXt Network. Sensors 2023, 23, 3329. [Google Scholar] [CrossRef] [PubMed]
  10. Nurullayev, S.; Lee, S.W. Generalized Parking Occupancy Analysis Based on Dilated Convolutional Neural Network. Sensors 2019, 19, 277. [Google Scholar] [CrossRef] [PubMed]
  11. Huang, C.; Yang, S.; Luo, Y.; Wang, Y.; Liu, Z. Visual Detection and Image Processing of Parking Space Based on Deep Learning. Sensors 2022, 22, 6672. [Google Scholar] [CrossRef] [PubMed]
  12. Vu, H.T.; Huang, C.C. Parking Space Status Inference Upon a Deep CNN and Multi-Task Contrastive Network with Spatial Transform. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 1194–1208. [Google Scholar] [CrossRef]
  13. Huang, C.C.; Vu, H.T. Vacant Parking Space Detection Based on a Multilayer Inference Framework. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 2041–2054. [Google Scholar] [CrossRef]
  14. Weiss, K.; Khoshgoftaar, T.M.; Wang, D.D. A survey of transfer learning. J. Big Data 2016, 3, 1–40. [Google Scholar] [CrossRef]
  15. Wilson, G.; Cook, D.J. A Survey of Unsupervised Deep Domain Adaptation. ACM Trans. Intell. Syst. Technol. 2020, 11, 1–46. [Google Scholar] [CrossRef] [PubMed]
  16. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the Advances in Neural Information Processing Systems 27; Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2014; pp. 2672–2680. [Google Scholar]
  17. Shannon, C.E. A Mathematical Theory of Communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  18. Nguyen, M.H.; Chao, T.Y.; Huang, C.C. Vacant Parking Space Detection based on Task Consistency and Reinforcement Learning. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 2009–2016. [Google Scholar] [CrossRef]
  19. Alemi, A.A.; Fischer, I.; Dillon, J.V.; Murphy, K. Deep Variational Information Bottleneck. arXiv 2016, arXiv:1612.00410. [Google Scholar]
  20. Rajagopalan, A.; Chellappa, R. Vehicle detection and tracking in video. In Proceedings of the 2000 International Conference on Image Processing, Vancouver, BC, Canada, 10–13 September 2000; Volume 1, pp. 351–354. [Google Scholar] [CrossRef]
  21. Rajagopalan, A.; Burlina, P.; Chellappa, R. Higher order statistical learning for vehicle detection in images. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; Volume 2, pp. 1204–1209. [Google Scholar]
  22. Almeida, P.; Oliveira, L.S.; Silva, E.; Britto, A.; Koerich, A. Parking Space Detection Using Textural Descriptors. In Proceedings of the 2013 IEEE International Conference on Systems, Man, and Cybernetics, Manchester, UK, 13–16 October 2013; pp. 3603–3608. [Google Scholar] [CrossRef]
  23. Lixia, W.; Dalin, J. A Method of Parking Space Detection Based on Image Segmentation and LBP. In Proceedings of the 2012 Fourth International Conference on Multimedia Information Networking and Security, Nanjing, China, 2–4 November 2012; pp. 229–232. [Google Scholar] [CrossRef]
  24. Wu, Q.; Huang, C.; Wang, S.Y.; Chiu, W.C.; Chen, T. Robust Parking Space Detection Considering Inter-Space Correlation. In Proceedings of the 2007 IEEE International Conference on Multimedia and Expo, Beijing, China, 2–5 July 2007; pp. 659–662. [Google Scholar] [CrossRef]
  25. Huang, C.C.; Wang, S.J. A hierarchical bayesian generation framework for vacant parking space detection. IEEE Trans. Circuits Syst. Video Technol. 2010, 20, 1770–1785. [Google Scholar] [CrossRef]
  26. Huang, C.C.; Tai, Y.S.; Wang, S.J. Vacant parking space detection based on plane-based Bayesian hierarchical framework. IEEE Trans. Circuits Syst. Video Technol. 2013, 23, 1598–1610. [Google Scholar] [CrossRef]
  27. Huang, C.C.; Vu, H.T. A multi-layer discriminative framework for parking space detection. In Proceedings of the 2015 IEEE 25th International Workshop on Machine Learning for Signal Processing (MLSP), Boston, MA, USA, 17–20 September 2015; pp. 1–6. [Google Scholar] [CrossRef]
  28. Zheng, W.; Tran, V.; Huang, C. D2NA: Day-To-Night Adaptation for Vision based Parking Management System. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1573–1577. [Google Scholar]
  29. Yu, D.; Yao, K.; Su, H.; Li, G.; Seide, F. KL-divergence regularized deep neural network adaptation for improved large vocabulary speech recognition. In Proceedings of the ICASSP, Vancouver, BC, Canada, 26–31 May 2013; pp. 7893–7897. [Google Scholar]
  30. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks For Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  31. Tishby, N.; Pereira, F.C.; Bialek, W. The information bottleneck method. In Proceedings of the 37-th Annual Allerton Conference on Communication, Control and Computing, Monticello, IL, USA, 22–24 September 1999; pp. 368–377. [Google Scholar]
  32. VIB Model Zoo. 2020. Available online: https://modelzoo.co/model/vib-pytorch (accessed on 5 February 2023).
  33. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Performance of the vacant detector when training and testing on different parking lots. We set up two cameras to capture the same parking lot but at different angle views. These angle views represent the source and target domains. Samples are collected as a normalized three-parking slot Vu 2019 [12]. Training and evaluating datasets are collected on the source domain, and a testing dataset is collected on the target domain. The first row is the result of the proposed method; the second row is the conventional method’s result.
Figure 1. Performance of the vacant detector when training and testing on different parking lots. We set up two cameras to capture the same parking lot but at different angle views. These angle views represent the source and target domains. Samples are collected as a normalized three-parking slot Vu 2019 [12]. Training and evaluating datasets are collected on the source domain, and a testing dataset is collected on the target domain. The first row is the result of the proposed method; the second row is the conventional method’s result.
Sensors 23 04776 g001
Figure 2. Vacant and occupied samples [12] on source and target domains. They are collected at the same slot and belong to a trajectory [18]. Each row represents samples at the same slot but from different domains. Each column represents samples in a trajectory where the background is similar, but the differences are from the object at the middle slot.
Figure 2. Vacant and occupied samples [12] on source and target domains. They are collected at the same slot and belong to a trajectory [18]. Each row represents samples at the same slot but from different domains. Each column represents samples in a trajectory where the background is similar, but the differences are from the object at the middle slot.
Sensors 23 04776 g002
Figure 3. System overview.
Figure 3. System overview.
Sensors 23 04776 g003
Figure 4. The process of collecting training data in TCL2021 [18]. The X-axis is the frame index; the Y-axis is the optical flow magnitude.
Figure 4. The process of collecting training data in TCL2021 [18]. The X-axis is the frame index; the Y-axis is the optical flow magnitude.
Sensors 23 04776 g004
Table 1. Network detail. Here, F B = 25,088, F C = 2048 , d = 1024 .
Table 1. Network detail. Here, F B = 25,088, F C = 2048 , d = 1024 .
BlockLayerParameters
Encoder 1 FlattenOptional
Encoder 2 nn.LinearIn = F B , Out = F C
Encoder 3 nn.ReLuInplace = True
Encoder 4 nn.LinearIn = F C , Out = 2 d
Re-parametrize z = μ + ϵ σ ϵ N ( 0 , σ 2 )
Classifiernn.LinearIn = d, Out = 2
Table 2. The accuracy on s o u r c e / t a r g e t domain under different datasets and various β values. (The unit is %). The bold represents the best performance in each column.
Table 2. The accuracy on s o u r c e / t a r g e t domain under different datasets and various β values. (The unit is %). The bold represents the best performance in each column.
Scenario S 1 S 2 S 3 S 4 S 5
No VIB (BN)93.43/76.1398.82/75.1699.47/83.4599.68/83.5399.78/82.47
No VIB ( p = 0.5 )97.95/84.1899.46/89.6899.62/91.9399.62/92.0299.63/92.02
No VIB ( p = 0.7 )98.18/87.6399.46/92.5299.70/94.4999.79/94.7699.84/95.88
No VIB ( p = 0.9 )96.75/88.8999.41/95.0799.63/95.0199.81/95.4699.63/95.76
β = 0.05 98.01/90.8899.19/92.8699.36/93.7799.60/94.2699.80/96.06
β = 0.07 98.25/90.1299.22/94.7199.50/95.2899.69/95.2199.77/97.02
β = 0.1 98.21/92.9399.29/94.7399.51/94.7599.79/96.5199.81/97.91
β = 0.2 98.44/89.2699.41/92.0999.51/92.6999.78/93.4899.81/94.53
Table 3. The F1 score on s o u r c e / t a r g e t domain under different datasets and various β values. (The unit is %). The bold represents the best performance in each column.
Table 3. The F1 score on s o u r c e / t a r g e t domain under different datasets and various β values. (The unit is %). The bold represents the best performance in each column.
Scenario S 1 S 2 S 3 S 4 S 5
No VIB (BN)92.30/57.9098.58/54.5399.37/74.2399.62/74.1499.73/72.27
No VIB ( p = 0.5 )97.54/75.4599.35/85.2299.54/88.8499.54/88.9799.55/88.95
No VIB ( p = 0.7 )97.81/82.209935/89.7399.64/92.7899.75/94.7399.80/95.58
No VIB ( p = 0.9 )96.10/84.3799.28/93.7199.56/94.0699.77/94.059984/94.45
β = 0.05 97.63/89.0899.01/91.5499.23/91.8099.52/93.0999.76/95.05
β = 0.07 97.92/88.3499.06/93.5599.40/94.2299.63/94.0899.72/96.21
β = 0.1 97.87/91.5699.15/93.5899.41/93.6299.77/95.6299.77/97.39
β = 0.2 98.14/86.5599.29/90.4999.40/91.2799.73/91.1599.73/92.70
Table 4. Accuracy and F1-Score on source and target domains. The training dataset is D t r a i n 2 S . (Unit is %). The bold represents the best performance in each column.
Table 4. Accuracy and F1-Score on source and target domains. The training dataset is D t r a i n 2 S . (Unit is %). The bold represents the best performance in each column.
DomainMetric β = 0.05 β = 0.07 β = 0.1 β = 0.2 No VIB ( D Train 2 S )No VIB ( S 1 )
SourceACC98.6898.7998.7798.6598.0497.95
( D V a l S ) F198.0498.5498.5298.3697.6197.54
TargetACC92.4293.2194.8293.7885.8884.18
( D T e s t T ) F190.3491.0793.2892.1879.8575.45
Table 5. Comparison to state-of-the-art vacant detectors on the same parking lot. Here, a sample refers to a normalized 3-parking-slot image and a trajectory refers to a video sequence when a car moves in/out of a slot. The bold represents the best performance in each column.
Table 5. Comparison to state-of-the-art vacant detectors on the same parking lot. Here, a sample refers to a normalized 3-parking-slot image and a trajectory refers to a video sequence when a car moves in/out of a slot. The bold represents the best performance in each column.
Method# of Training SamplesAccuracy (%)
Paulo 2013 [22]587,667 samples81.98
Lixia 2012 [23]587,667 samples83.46
Wu 2007 [24]587,667 samples92.26
Huang 2017 [13]587,667 samples98.44
Vu 2019 [12]587,667 samples99.68
Faster-RCNN [33]587,667 samples72.68
TCL 2021 [18]1000 trajectories99.54
Proposed method ( β = 0.1 ) with VGG5000 samples99.81
Proposed method ( β = 0.1 ) no VGG5000 samples98.82
Table 6. Accuracy comparison with unsupervised domain adaption. (The unit is %.) (‘ 90 ° view’ ⇒ ‘ 45 ° view’) means that ‘ 90 ° view’ is the source domain and ‘ 45 ° view’ is target domain. Each experiment is repeated five times to obtain the min, mean, and max accuracies.
Table 6. Accuracy comparison with unsupervised domain adaption. (The unit is %.) (‘ 90 ° view’ ⇒ ‘ 45 ° view’) means that ‘ 90 ° view’ is the source domain and ‘ 45 ° view’ is target domain. Each experiment is repeated five times to obtain the min, mean, and max accuracies.
90 ° view’ ⇒ ‘ 45 ° view’ 45 ° view’ ⇒ ‘ 90 ° view’
Proposed method (min)94.5897.20
Proposed method (mean)95.0397.82
Proposed method (max)95.1398.25
Unsupervised Domain Adaptation [28]98.298.85
Table 7. Statistics of maximum, minimum, and mean value on feature maps.
Table 7. Statistics of maximum, minimum, and mean value on feature maps.
With VIBWithout VIB
Maximum Minimum Mean Maximum Minimum Mean
count512512512512512512
mean0.40240.07500.19672.54870.00240.6033
std1.16160.49520.81683.92790.05411.2015
min000000
25%000000
50%0000.834800.0748
75%0.1759003.564400.6591
max12.64957.32049.829030.85681.225010.0369
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Manh, H.-N. Rethinking Feature Generalization in Vacant Space Detection. Sensors 2023, 23, 4776. https://doi.org/10.3390/s23104776

AMA Style

Manh H-N. Rethinking Feature Generalization in Vacant Space Detection. Sensors. 2023; 23(10):4776. https://doi.org/10.3390/s23104776

Chicago/Turabian Style

Manh, Hung-Nguyen. 2023. "Rethinking Feature Generalization in Vacant Space Detection" Sensors 23, no. 10: 4776. https://doi.org/10.3390/s23104776

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop