Next Article in Journal
Cognitive Relevance Transform for Population Re-Targeting
Previous Article in Journal
An IoT Smart Rodent Bait Station System Utilizing Computer Vision
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Complementary Deep and Shallow Learning with Boosting for Public Transportation Safety

1
Faculty of Information Technology, Macau University of Science and Technology, Taipa 999078, Macao
2
Melbourne School of Engineering, the University of Melbourne, Melbourne 3010, Australia
3
The School of Computer Science, Fudan University, Shanghai 201203, China
4
Zhuhai Fudan Innovation Institute, Zhuhai 519000, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(17), 4671; https://doi.org/10.3390/s20174671
Submission received: 8 July 2020 / Revised: 11 August 2020 / Accepted: 13 August 2020 / Published: 19 August 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
To monitor road safety, billions of records can be generated by Controller Area Network bus each day on public transportation. Automation to determine whether certain driving behaviour of drivers on public transportation can be considered safe on the road using artificial intelligence or machine learning techniques for big data analytics has become a possibility recently. Due to the high false classification rates of the current methods, our goal is to build a practical and accurate method for road safety predictions that automatically determine if the driving behaviour is safe on public transportation. In this paper, our main contributions include (1) a novel feature extraction method because of the lack of informative features in raw CAN bus data, (2) a novel boosting method for driving behaviour classification (safe or unsafe) to combine advantages of deep learning and shallow learning methods with much improved performance, and (3) an evaluation of our method using a real-world data to provide accurate labels from domain experts in the public transportation industry for the first time. The experiments show that the proposed boosting method with our proposed features outperforms seven other popular methods on the real-world dataset by 5.9% and 5.5%.

1. Introduction

1.1. Motivations

Decreasing the current number of global deaths and injuries from road traffic accidents by half is one of the important Sustainable Development Goals as part of the 2030 Agenda for Sustainable Development adopted by the United Nations General Assembly. Traffic accidents not only bring huge financial losses to society but also cause great physical and mental damages to everyone [1,2]. Millions of people died from traffic accidents in 2018 [3], and most traffic accidents are caused by human mishandling [4]. Analyzing the behavior of drivers, especially public transportation drivers, is important to protect road safety [5,6,7]. To ensure safety for public transportation, public transportation operators can be requested to get an evaluation of drivers and to identify dangerous drivers for retraining. For public transportation fleet management and monitoring, massive data is collected from vehicles using state-of-the-art technologies of sensors for example MobilEye from Israel. In the control center, thousands of real-time events and alarms arrive from the sensors of the vehicles through wireless networks in real time every day. Although it is virtually impossible to handle such a huge amount of data manually, accurate predictions with machine learning to analyze behavior from the vehicles has recently become feasible. Machine learning techniques have been applied to analyzing behavior in different tasks with various kinds of data collected using sensors in moving vehicles [8,9,10,11]. We investigate efficient and accurate machine learning methods to classify whether the driving behavior of drivers is safe as it is important that drivers with unsafe behavior should be warned and retraining can be provided to the drivers.

1.2. Challenges

It is challenging to classify a driver’s driving behavior (safe/unsafe). First, the industrial need for a high classification performance cannot be satisfied using solely existing methods as the misclassification rates can be high. The high misclassification rates in previous methods have been pointed out in previous literature  [12,13] as it has been suggested that it is difficult to model the driver’s behavior. For example, the performance of K Nearest Neighbour (KNN) models is greatly affected by unbalanced training data in traffic data for safety with many fewer labels for accidents. Second, the lack of features in the data collected with the Controller Area Network (CAN) bus does not provide lots of information for driving behavior analysis to train an accurate machine learning classifier. There is no existing method for road safety predictions with CAN bus data to extract extra useful information from features. Last, because of the high cost of labeling and privacy issues with public transportation data, to the best of our knowledge, there is no publicly available dataset with labels for the predictions of safe driving behavior on public transportation. The lack of urgently needed labels in datasets for road safety makes it hard to evaluate and build machine learning models.

1.3. Contributions

In this paper, our contributions include (1) a boosting method to make deep learning and statistical learning complement each other, (2) a novel method to compute extra time-series features to extract richer information, (3) extensive evaluation on a new real-world dataset with labels from experts in the public transportation domain.
The motivation for the usage of CAN-bus data is that the CAN standard is one of the most important bus standards for vehicles [14,15,16,17]. The infrastructure already built-in in most vehicles bought by transportation companies records data efficiently from devices in the vehicles. The CAN standard is a serial data bus standard designed to enable electronic devices to communicate with each other. Therefore, CAN data can be easily obtained by transportation companies. There are a small number of available features in the data collected using the standard CAN bus system. With the lack of sufficient useful features, it is hard to find patterns in the data to determine driving behavior from the driver.
The novel boosting method is proposed to combine the advantages of statistical learning and deep learning. It is shown in our experiments that the ensemble with our proposed features outperforms any single state-of-the-art method we considered, and our boosting method combines seven state-of-the-art machine learning methods including support vector machine (SVM), random forest (RF), k-nearest neighbour (KNN), discriminant analysis, naive Bayes classifier, adaptive boosting (AdaBoost) and a deep learning neural network called Long Short-Term Memory (LSTM). In addition, the proposed boosting method outperforms the seven methods also for the case without the proposed feature extraction method.
Extra features are computed using our method. Feature engineering is an important tool to extract useful information in time-series for machine learning methods for better performance when the number of features is small. We show in our experiments that this significantly improves the performance of the classifiers. We, thus, propose a method to compute extra time-series features from the raw data of the CAN bus system to extract extra information.
To completely evaluate methods in the real world, the experiments are conducted using a real-world dataset collected using the CAN bus system. Because of the high cost of sample labeling, there is no published real-world dataset with labels for analyzing the driving behavior of drivers on public transportation. The samples in the dataset are labeled by the experts of Transportes Urbanos de Macau (TransMac), which is one of the largest bus companies in Macao.
We divide the rest of this paper into five sections. In Section 2, we describe related work. Our dataset and the proposed method are described in detail in Section 3 and Section 4. In Section 5, the proposed method is evaluated on two datasets with a comparison to various state-of-the-art machine learning methods. Section 6 concludes this work.

2. Related Work

In this work, we focus on safety classification with data from the Controller Area Network (CAN) bus. The development of the CAN bus started in 1983 and was released in 1986 [18]. This standard has recently become available in most embedded systems for vehicles. The CAN bus is one of five protocols used in the mandatory onboard diagnostic (OBD-II) standard. The OBD-II standard has been mandatory for all cars and light trucks sold in the United States since 1996, and the EOBD standard has been mandatory for all petrol vehicles sold in the European Union since 2001 and all diesel vehicles since 2004 [19,20]. A lot of recent research focuses on the analysis of sensor data from the CAN bus system [5,21,22,23,24,25,26,27,28].
Machine learning plays the important role in building data analytics models to handle massive data [29,30]. Statistical learning and deep learning are important areas in machine learning. Methods based on statistical learning have been successfully applied to solving many related behavior analytics problems. In [11,31], Bayesian methods are used to predict braking behavior and model vehicle speeds. k-nearest neighbors (KNN) is employed to classify driving styles in [10]. Support vector machines (SVM), deep learning (DL), and decision trees (DT) have been applied to predicting driving behavior and accident risk predictions [8,9]. Deep learning techniques are popular in recent years for many tasks [32,33]. The Long Short-Term Memory (LSTM) technique is one of the most popular deep learning techniques for time-series problems [34]. LSTM networks are a special kind of recurrent neural networks (RNN), and LSTM networks aim at learning long-term dependencies [35]. In contrast to the standard RNN, the repeating modules of LSTM networks contain four interacting layers to enable the ability to change information of the cell state. LSTM is used to detect driver distraction [36]. The ensemble of machine learning methods are shown to outperform, most of the time, one particular technique [37,38,39], and we propose a heterogeneous boosting method to obtain better performance. In our experiments, our boosting method is compared with seven state-of-the-art methods.

3. Evaluation of Road Safety Predictions

For any application domain of machine learning, one of the most objective evaluation methods is to see how prediction models perform in real-world datasets. However, to the best of our knowledge, there is no published real-world dataset with labels provided for behavior analysis of public drivers. In this work, we build a new dataset collected using the CAN bus system from one of two public bus companies in Macao called TransMac.
One record is produced every three seconds from the reading of the sensors in a moving public vehicle, and there are totally 6451 records with 24 features in the new dataset. All 6451 records are labeled by domain experts. There are 507 unsafe cases and 5944 safe cases in the labels. In total, the recording time for our CAN bus data is 6451 × 3 = 19,353 s which is 5.38 h of driving by professional bus drivers in the company. Although each sample contains 24 features, some features are totally irrelevant for the training of the machine learning model (see Table 1). As shown in Table 1, features like vehicle identification are meaningless for the machine learning model. Features containing too much missing data cannot be useful either. For example, most entries of the “CANALARMSTATE” feature and the “CANALARMSTATE” are N/As (Not Available). Features used are listed in Table 1 to provide a reference for the training of our method. In addition, descriptive statistics on the feature set used for training are shown in Table 2.

4. The Proposed Method

We propose a feature extraction method (see Algorithm 1) for extracting richer information from the change of feature vectors against time and propose a boosting method (see Algorithm 2) to classify whether driving behavior of drivers on public transportation can be considered safe on the road. The feature extraction method is a general method. It can be used with any other machine learning classification method. In the experiments, it is shown that our feature extraction method can be used to improve the performance of any classification method. In addition, it is also shown that our boosting method outperforms other seven machine learning methods whether or not our feature extraction method is used.

4.1. Our Method for Richer Information With Feature Extraction

Missing data is common in industrial data collected from CAN data systems. Features with too many N/A (Not Available) entries cannot be used to train the machine learning methods. In addition, features irrelevant for driving behavior analysis like the identifier of the vehicles are excluded. Therefore, there are only a few useful features left for classification without irrelevant features (see Table 1). The low dimensionality of the feature space of training data severely limits the descriptive power of the samples. The lack of descriptive power makes it is very difficult to obtain accurate machine learning models. We argue that richer information can be extracted from the change of feature values against time and we, hence, propose a feature extraction method to provide extra useful time-series features to deal with the lack of information in the original features. For example, the acceleration of the car is important for driver behavior analysis, but this information is not recorded in the original data. The acceleration of the bus can be obtained by calculating the gradient of the velocity of the bus. The proposed feature extraction method is shown in Algorithm 1.
Algorithm 1 Our Method for Richer Information with Feature Extraction
Input:n samples s 1 , , s n where s i = [ f i , 1 , , f i , m ] T
Output:n samples s 1 , , s n with time-series features, s i = [ f i , 1 , , f i , m , t i , 1 , , t i , m + 7 ] T
  Divide n samples into p periods, P 1 , , P p , by the recording time.
  for j = 1 , , p do
   for each sample s i P j do
     [ t i , 1 , , t i , m ] T = 1 P j × s z P j [ f z , 1 , , f z , m ] T
     s i = [ f i , 1 , , f i , m , t i , 1 , , t i , m ] T
   end for
  end for
  for each sample s i do
    t i , m + 1 and t i , m + 2 are the differences in the feature values of s i , for the latitude and the longitude respectively.
    t i , m + 3 , , t i , m + 7 are the gradients of the feature values of s i related to the velocity, the mileage,
   the tire pressure, the engine speed, and the engine temperature.
    s i = [ f i , 1 , , f i , m , t i , 1 , , t i , m + 7 ] T
  end for
The input data of Algorithm 1 contains n samples, s 1 , , s n , with m features s i = [ f i , 1 , , f i , m ] T . Especially, the features irrelevant for training are excluded in these m features. For example, for our proposed dataset, the m features are the twelve features used to train the classification model (see Table 1). It is common in time-series analysis to use moving averages. The moving average is used to filter out noise. It is a common signal preprocessing step for time-series data if there is noise in the data [40,41,42]. We noticed the noise in the CAN-bus data so signal processing filtering techniques are employed in our work with the aim to achieve better training and classification. The moving average is used to filter out noise. It is a common signal preprocessing step for time-series data if there is noise in the data [40,41,42]. We noticed noise in the CAN-bus data so signal processing filtering techniques are employed in our work with the aim to achieve better training and classification.
For example, the average driving speed of a driver in two minutes (a period) is useful information for analyzing his driving behavior. Motivated by this, some time-series features are calculated for this particular reason (see Algorithm 1). The n samples are divided into p periods by the recording time. The value of p is a tunable parameter, and it depends on the time interval between two samples.
In our boosting method, one period covers two minutes. For sample s i , m time-series features, t i , 1 , , t i , m , are extracted from raw features. The latitude and longitude features, f l a t i t u d e and f l o n g i t u d e , are obtained from GPS information. The difference in the latitude/longitude values of the adjacent samples can be used to measure the velocity of the bus. In Algorithm 1, t i , m + 1 and t i , m + 2 of sample s i are the differences in the values of sample s i and sample s i 1 , for latitude and longitude features respectively. The gradient of a feature is used to describe how fast the feature values change. The velocity, the mileage, the tire pressure, the engine speed, and the engine temperature are important for accurate classification. These features can reflect the different behavior of drivers, and  t m + 3 , , t m + 7 are calculated to find the rates of change.
The features are irrelevant features if the changes of the features due to reasons other than driving behavior. In most cases, data used to train machine learning models includes irrelevant features or redundant features. Certain machine learning models automatically pick useful features during training. We use techniques that specifically map useful features to the labels (driving behavior in our case) effectively like random forests (RF) with the importance scores of features generated to ignore irrelevant features. The RF model is then trained using a subset of features with high importance scores [43,44,45]. When the dimensionality of feature space is large, performance could suffer due to the curse of dimensionality [46,47]. However, in our case, as the number of features is not exceedingly high, our experiments show that the techniques perform well without suffering from the curse of dimensionality.

4.2. Weak Learners

In our heterogeneous boosting method, all the seven methods in Table 3 are used. Different machine learning methods have their own characteristics and are suitable for different tasks. SVM, traditionally a statistical machine learning method, is one of the most popular methods for two-class classification. In SVM, a hyperplane is constructed for classification after the data in the low-dimensional space is mapped to the high-dimensional space using a kernel function. In our boosting methods, Radial Basis Function (RBF) kernel is used as the kernel function of the SVM, and the RBF kernel can be obtained as
R ( x i , x j ) = e x p ( γ x i x j 2 ) ,
where x i , x j are the feature vectors of sample i and sample jk-nearest-neighbors (KNN) is a nonparametric classification method, which is simple but effective in machine learning. The assumption underlying KNN is that the information (e.g., the class in classification problems) of an output sample is similar to input samples containing similar characteristics to this output sample. In the proposed boosting method, KNN using the chi-square distance is used to measure the distance of samples. The chi-square distance between sample i and sample j is obtained as
χ ( x i , x j ) = f = 1 F w f ( x i , f , x j , f ) 2 ,
where x i , x j are the feature vectors, and F is the dimentionality of samples. RF, a method based on bagging, contains a certain number of decision trees to classify. In the RF method, Q decision trees are trained to determine classification results of RF. Given a sample x, its result y produced by RF is
y = 1 Q q = 1 Q f q ( x ) ,
where f q ( x ) is the result given by a decision tree. Discriminant Analysis is also known as Fisher Discriminant Analysis. In Discriminant Analysis, a linear combination of the features is obtained to classify the samples. AdaBoost used as one of the component learners in our boosting method is a homogeneous boosting method. The base learners of this homogeneous boosting method are 1-depth decision trees. In Naive Bayes, Bayes’ theorem with naive independence assumptions between the features is applied.
The LSTM network used in the proposed boosting method is shown in Figure 1. There are five layers in the LSTM network: the input layer, the LSTM layer, the fully connected layer, the softmax layer, and the classification layer. The LSTM network starts with putting the CAN bus data into the input layer. Using the input data as the training set, the LSTM layer learns long-term dependencies of samples. To handle time-series data, the LSTM layer contains many LSTM blocks. One LSTM block uses an input sample and the output of last LSTM block as its input and the block output a cell state and a hidden state. Finally, the classification results are generated by the fully connected layer, the softmax layer, and the classification layer through analyzing the long-term dependencies. In the LSTM network, the vanishing gradient problem is avoided by adding four components to the RNN network (see Table 4 and Figure 1): the input gate, the output gate, the forget gate, and the cell candidate. The interaction of the four components is shown in Table 4 and Figure 1. c t and h t denote the cell state and the hidden state produced by the t-th LSTM block. i t , o t , f t , and c t denote the outputs of the input gate, the output gate, the forget gate, and the cell candidate. As shown in Figure 1, the output of the t-th LSTM block are the cell state c t and the hidden state h t . The cell state c t is obtained by
c t = f t c t 1 + i t g t ,
where ⊙ denotes the Hadamard product. The hidden state h t is obtained by
h t = o t σ c t ,
where σ denotes the state activation function of the LSTM block.
It is popular to use two well-known techniques, principal components analysis and t-sne, with deep learning. We apply these methods to process raw CAN bus data to train the LSTM network.

4.3. Our Method with Boosting

Ensemble learning is a machine learning technique, which is used to combine multiple methods and to get better performance than that of a particular method. The proposed boosting method combines seven state-of-the-art machine learning methods. The seven machine learning methods are support vector machine (SVM) [48,49], k nearest neighbour (KNN) [50], random forest (RF) [43,51], naive Bayes [52], discriminant analysis [53], adaptive boosting (AdaBoost) [54], and Long Short-Term Memory (LSTM)  [34]. The proposed boosting method is in Algorithm 2.
Algorithm 2 The Proposed Boosting Method
Input: Training data D = ( x 1 , y 1 ) , , ( x n , y n ) where x i R m , y i = 0 , 1 .
Output: Final strong classifier H ( x ) .
 1: Initialize weights w 1 , i = 1 2 c , 1 2 ( n c ) for safe samples and unsafe samples, respectively, where c is the number of safe samples.
 2: for t = 1 to U do
 3:  Normalize the weights, w t , i = w t , i j = 1 n w t , j .
 4:  Train all g weak classifiers, l 1 x , , l g x , using the training data D with our time-series features.
 5:  Prediction using g classifiers, and h t ( x ) is the classifier l u ( x ) with the highest correctly rate a.
 6:  Update the weights: w t + 1 , i = w t , i × B 1 e i , where e i = 0 , if the sample s i is correctly predicted, otherwise e i = 1 . B = a 1 a .
 7: end for
 8: The boosting classifier combines the U classifiers: if t = 1 U α t h t x 1 2 t = 1 U α t , H x = 1 , otherwise H x = 0 .
As shown in Algorithm 2, there are n samples in the training data, and the dimensionality of them is equal to m. There are g weak learners used in the algorithm. In the proposed boosting method, g is equal to seven (see Table 3). U is the number of the weak classifiers which are chosen to form final strong classifier H ( x ) . The value of U is a tunable parameter, and it is equal to five in our method. Each of the g classifiers is trained based on one particular machine learning method.

5. Experiments

Our experiments are conducted using two datasets: the public Warrigal dataset [55] and our own dataset provided by TransMac. The Warrigal dataset can be downloaded in http://its.acfr.usyd.edu.au/datasets/warrigal/. In the experiments, the proposed boosting method is compared with other seven popular machine learning methods: SVM, KNN, RF, Simple Bayes, Discriminant Analysis, AdaBoost, and LSTM.

5.1. Evaluation Metrics

The classification accuracy, sensitivity, specificity, and the Area Under Curve (AUC) are four of the most popular evaluation methods for a binary classifier. The sensitivity is the probability of detection while the specificity gives the probability of false alarm. The AUC value is equal to the probability that a randomly chosen positive example is ranked higher than a randomly chosen negative example. The sensitivity and specificity are also referred to as the True Positive Rate (TPR) and the True Negative Rate (TNR), respectively. Sensitivity and Specificity are computed using the numbers of true negatives (TN), false negatives (FN), true positives (TP), and false positives (FP):
T P R = T P / T P + F N ,
T N R = T N / F P + T N .
The average accuracy is taken over ten repeated experiments to better evaluate the methods, and it is obtained by using
Accuracy = T P + T N / T P + T N + N P + N F .

5.2. Experimental Setup

We use MATLAB to implement the seven methods and the proposed methods. Grid search is used to determine all the hyperparameters of our classifiers using a validation set. The predicted dependent variable is safety from 0 (safe) to 1 (unsafe). For our dataset, the training and test labels were determined by domain experts. For the other dataset, Label 1 indicates ongoing communications by the driver which can violate safety guidelines and 0 indicates the time without communications by the driver. In our dataset, there are 507 unsafe cases and 5944 safety cases. In the Warrigal dataset, there are 17,716 unsafe cases and 205,722 safety cases.
To clearly show the comparison of all methods, the experiments on each dataset are divided into two parts. In the first part, in order to demonstrate that the proposed boosting method can outperform other machine learning methods (see Table 3), all methods are trained using the raw data without our feature extraction method (see Algorithm 1). In the second part, to determine whether our feature extraction method can be used to improve the performance of machine learning methods, our feature extraction method is used to compute time-series features, and data with these time-series features are used to train the methods. By comparing the accuracies with the two sets of experiments, it is shown that the performance of methods is improved using our feature extraction method. To avoid the overfitting issue of machine learning, in both sets of the experiments, there are two scales, 70% of the dataset and 90% of the dataset, of the training set. The samples in the training set are randomly extracted from the whole dataset.

5.3. Safety Classification on the Warrigal Dataset

The Warrigal dataset [55] is a large dataset collected with the interactions of large trucks and smaller vehicles, i.e., thirteen vehicles in a large quarry-type environment. The data contains vehicle state information (like positions, speeds, and heading) and information on vehicle-to-vehicle communications. Due to the large size of the Warrigal dataset, when the dataset is published, the dataset is divided into many subsets with one subset for data recorded in a day. Because of the large size of the dataset and limited computational resources, only one subset (data recorded on 1st February 2009 which is just the first day in the dataset) picked from the data is used for our experiments. There are 223,438 samples with twelve features in the subset. In the dataset, there is no label for safety predictions. We consider constant communication through wireless devices as distractions which could potentially lead to unsafe driving behavior. Samples recorded during constant verbal communication are labeled as potentially unsafe (or inattentive) while the other samples are labeled safe (or attentive). More specifically, Label 1 indicates ongoing communications by the driver which can violate safety guidelines and 0 indicates driving without wireless communication engaged. 17,716 samples are labeled potentially unsafe and 205,722 are labeled safe.
The comparison among prediction models trained with the raw features of the Warrigal dataset is shown in Table 5. In terms of classification accuracy, the proposed method outperforms the other seven methods by 1.5% and 1.1%, with 70% and 90% of the dataset for training respectively. The novel method also gets the highest value of AUC. We further demonstrate the effectiveness of our feature extraction method. The performance of the models trained with our features is shown in Table 6. As observed in Table 5 and Table 6, the proposed boosting method outperforms the other seven state-of-the-art approaches whether or not our feature extraction method is used. In addition, comparison with the accuracies in Table 5 and Table 6 indicates that the performance of all machine learning methods is improved using our feature extraction method. Our method with feature extraction outperforms other methods using raw features by 2.7% and 3.2%, with 70% and 90% of all samples randomly selected for training respectively.

5.4. Safety Classification with Our Dataset

The results from the previous experiments show that our method improves the performance of classification. We further apply our methods to a real-world problem in the industry with a public bus company in Macao called TransMac. Given the fact that there is no real-world public dataset with labels for safety classification, the experiments are conducted using a dataset built with data collected from TransMac.
In this subsection, the experimental setup follows that of the previous subsection using a different dataset with the first set of experiments with feature extraction and the second set without. The performance comparison of the methods can be found in Table 7 and Table 8. As shown in the two tables, our boosting method outperforms the other methods in all cases. It is shown that the performance of all methods is improved using our features extraction method. Our boosting method with our feature extraction method can outperform other methods using raw features by 5.9% and 5.5%, with 70% and 90% of the whole dataset randomly selected for training respectively.

5.5. Further Evaluation

The CAN-bus data is time-series data. In order to see if predictions can still be accurate when training is done at a very different time period, we conduct another set of experiments to further evaluate the performance of our method. In this further experiment, the samples of the training set and the test set are collected at different times. As our real-world dataset was collected in four different periods, samples are therefore grouped into four subsets corresponding to the respective time periods. A four-fold cross-validation is conducted with these four subsets and the average classification accuracies of different methods are compared as shown in Table 9. Like Table 7 and Table 8, it is also found that the proposed method achieves the highest accuracy shown in the above table with or without our features. The boosting methods tend to be more expensive during training. We summarize the training time required for the methods in Table 10.
It is observed in our experiments that the performance does not improve when the number of weak learners in the final strong classifier, U, is larger than 5 although the total number of base learners built with state-of-the-art classification techniques is seven in our heterogeneous ensemble method. The empirical result verifying this is shown in Table 11.

6. Conclusions

Automation to determine whether certain driving behavior of drivers on public transportation can be considered safe on the road using A. I. or machine learning techniques has become a possibility recently. However, the industrial need for a high classification performance cannot be satisfied using existing methods using computer vision as the misclassification rates are too high with existing methods. Due to the high misclassification rates, it makes it hard to compare and to evaluate the performance of drivers on public transportation.
Our goal is to build a practical and accurate method for road safety predictions that automatically determine if the driving behavior is safe on public transportation. In this paper, our main contributions include (1) a novel feature extraction method because of the lack of informative features in the data, (2) a novel boosting method for driving behavior classification (safety or not) to combine advantages of deep learning and traditional statistical learning methods with much improved performance, and (3) evaluating methods using real-world data to provide accurate evaluations from labels from experts in the public transportation industry for the first time. The experiments show that the proposed boosting method with the proposed features outperforms seven other popular methods on the real-world dataset by 5.9% and 5.5%.

Author Contributions

Conceptualization, S.L.; literature search, S.L.; data analysis, S.L.; programming, S.L., X.Q. and J.Y.K.C.; writing–original draft preparation, S.L. resources, A.P.L.; funding acquisition, A.P.L.; project administration, A.P.L.; supervision, A.P.L.; data collection, X.Q.; writing – review and editing J.Y.K.C. and H.H.; data interpretation, H.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and Technology Development Fund, Macau SAR (File no. 0025/2019/AKP) and the “Bus Driver Behaviour Analysis for Road Safety” project funded by Hong Kong Applied Science and Technology Research Institute Company Limited (ASTRI).

Acknowledgments

This paper is an extended version of our previous work presented in [56]. We would like to express our gratitude to Transportes Urbanos de Macau (TransMac) for the help with data collection.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Tison, J.; Chaudhary, N.; Cosgrove, L.; Group, P.R. National Phone Survey on Distracted Driving Attitudes and Behaviors; Technical Report; United States National Highway Traffic Safety Administration: Washington, DC, USA, 2011.
  2. Wilson, F.A.; Stimpson, J.P. Trends in fatalities from distracted driving in the United States, 1999 to 2008. Am. J. Public Health 2010, 100, 2213–2219. [Google Scholar] [CrossRef] [PubMed]
  3. WHO. Global Status Report on Road Safety 2018; World Health Organization: Geneva, Switzerland, 2018. [Google Scholar]
  4. Kaplan, S.; Guvensan, M.A.; Yavuz, A.G.; Karalurt, Y. Driver behavior analysis for safe driving: A survey. IEEE Trans. Intell. Transp. Syst. 2015, 16, 3017–3032. [Google Scholar] [CrossRef]
  5. Hallac, D.; Sharang, A.; Stahlmann, R.; Lamprecht, A.; Huber, M.; Roehder, M.; Sosic, R.; Leskovec, J. Driver identification using automobile sensor data from a single turn. In Proceedings of the IEEE 19th International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 953–958. [Google Scholar]
  6. Moll, O.; Zalewski, A.; Pillai, S.; Madden, S.; Stonebraker, M.; Gadepally, V. Exploring big volume sensor data with Vroom. Proc. VLDB Endow. 2017, 10, 1973–1976. [Google Scholar] [CrossRef]
  7. Massaro, E.; Ahn, C.; Ratti, C.; Santi, P.; Stahlmann, R.; Lamprecht, A.; Roehder, M.; Huber, M. The car as an ambient sensing platform [point of view]. Proc. IEEE 2016, 105, 3–7. [Google Scholar] [CrossRef] [Green Version]
  8. Van Ly, M.; Martin, S.; Trivedi, M.M. Driver classification and driving style recognition using inertial sensors. In Proceedings of the IEEE Intelligent Vehicles Symposium (IV), Gold Coast, QLD, Australia, 23–26 June 2013; pp. 1040–1045. [Google Scholar]
  9. Paefgen, J.; Staake, T.; Thiesse, F. Evaluation and aggregation of pay-as-you-drive insurance rate factors: A classification analysis approach. Decis. Support Syst. 2013, 56, 192–201. [Google Scholar] [CrossRef]
  10. Vaitkus, V.; Lengvenis, P.; Žylius, G. Driving style classification using long-term accelerometer information. In Proceedings of the 19th International Conference on Methods and Models in Automation and Robotics (MMAR), Miedzyzdroje, Poland, 2–5 September 2014; pp. 641–644. [Google Scholar]
  11. Mudgal, A.; Hallmark, S.; Carriquiry, A.; Gkritza, K. Driving behavior at a roundabout: A hierarchical Bayesian regression analysis. Transp. Res. Part D Transp. Environ. 2014, 26, 20–26. [Google Scholar] [CrossRef]
  12. Zhang, J.; Wu, Z.; Li, F.; Xie, C.; Ren, T.; Chen, J.; Liu, L. A deep learning framework for driving behavior identification on in-vehicle CAN-BUS sensor data. Sensors 2019, 19, 1356. [Google Scholar] [CrossRef] [Green Version]
  13. Ontanón, S.; Lee, Y.C.; Snodgrass, S.; Winston, F.K.; Gonzalez, A.J. Learning to predict driver behavior from observation. In Proceedings of the AAAI Spring Symposium Series, Palo Alto, CA, USA, 27–29 March 2017. [Google Scholar]
  14. Song, H.M.; Woo, J.; Kim, H.K. In-vehicle network intrusion detection using deep convolutional neural network. Veh. Commun. 2020, 21, 100198. [Google Scholar] [CrossRef]
  15. Zhou, A.; Li, Z.; Shen, Y. Anomaly Detection of CAN Bus Messages Using a Deep Neural Network for Autonomous Vehicles. Appl. Sci. 2019, 9, 3174. [Google Scholar] [CrossRef] [Green Version]
  16. Chockalingam, V.; Larson, I.; Lin, D.; Nofzinger, S. Detecting attacks on the can protocol with machine learning. Annu. EECS 2016, 588, 7. [Google Scholar]
  17. Narayanan, S.N.; Mittal, S.; Joshi, A. Using data analytics to detect anomalous states in vehicles. arXiv 2015, arXiv:1512.08048. [Google Scholar]
  18. Kiencke, U.; Dais, S.; Litschel, M. Automotive serial controller area network. In SAE Transactions; SAE International: Warrendale, PA, USA, 1986; pp. 823–828. [Google Scholar]
  19. Building Adapter for Vehicle On-board Diagnostic. Available online: obddiag.net (accessed on 9 September 2009).
  20. Lebrun, A.; Demay, J.C. Canspy: A Platform for Auditing CAN Devices. Available online: https://jyx.jyu.fi/bitstream/handle/123456789/67095 (accessed on 9 September 2016).
  21. Paefgen, J.; Michahelles, F.; Staake, T. GPS trajectory feature extraction for driver risk profiling. In Proceedings of the 2011 International Workshop on Trajectory Data Mining and Analysis, Beijing, China, 17–21 September 2011; pp. 53–56. [Google Scholar]
  22. Grengs, J.; Wang, X.; Kostyniuk, L. Using GPS data to understand driving behavior. J. Urban Technol. 2008, 15, 33–53. [Google Scholar] [CrossRef]
  23. Wang, W.; Xi, J.; Chen, H. Modeling and recognizing driver behavior based on driving data: A survey. Math. Probl. Eng. 2014, 2014, 20. [Google Scholar] [CrossRef] [Green Version]
  24. Enev, M.; Takakuwa, A.; Koscher, K.; Kohno, T. Automobile driver fingerprinting. Proc. Privacy Enhanc. Technol. 2016, 2016, 34–50. [Google Scholar] [CrossRef] [Green Version]
  25. Martínez, M.; Echanobe, J.; del Campo, I. Driver identification and impostor detection based on driving behavior signals. In Proceedings of the IEEE 19th International Conference on Intelligent Transportation Systems (ITSC), Rio de Janeiro, Brazil, 1–4 November 2016; pp. 372–378. [Google Scholar]
  26. Fung, N.C.; Wallace, B.; Chan, A.D.; Goubran, R.; Porter, M.M.; Marshall, S.; Knoefel, F. Driver identification using vehicle acceleration and deceleration events from naturalistic driving of older drivers. In Proceedings of the IEEE International Symposium on Medical Measurements and Applications (MeMeA), Rochester, MN, USA, 7–10 May 2017; pp. 33–38. [Google Scholar]
  27. Ezzini, S.; Berrada, I.; Ghogho, M. Who is behind the wheel? Driver identification and fingerprinting. J. Big Data 2018, 5, 9. [Google Scholar] [CrossRef]
  28. Martinelli, F.; Mercaldo, F.; Orlando, A.; Nardone, V.; Santone, A.; Sangaiah, A.K. Human behavior characterization for driving style recognition in vehicle system. Comp. Electr. Eng. 2018, 83, 102504. [Google Scholar] [CrossRef]
  29. Luo, S.; Leung, A.P.; Hui, C.; Li, K. An investigation on the factors affecting machine learning classifications in gamma-ray astronomy. Mon. Not. R. Astronom. Soc. 2020, 492, 5377–5390. [Google Scholar] [CrossRef] [Green Version]
  30. Luo, S.; Xu, J.; Jiang, Z.; Liu, L.; Wu, Q.; Leung, E.L.H.; Leung, A.P. Artificial intelligence-based collaborative filtering method with ensemble learning for personalized lung cancer medicine without genetic sequencing. Pharmacol. Res. 2020, 160, 105037. [Google Scholar] [CrossRef]
  31. McCall, J.C.; Trivedi, M.M. Driver behavior and situation aware brake assistance for intelligent vehicles. Proc. IEEE 2007, 95, 374–387. [Google Scholar] [CrossRef]
  32. Lin, X.; Liang, Y.; Wan, J.; Lin, C.; Li, S.Z. Region-based Context Enhanced Network for Robust Multiple Face Alignment. IEEE Trans. Multimed. 2019, 21, 3053–3067. [Google Scholar] [CrossRef]
  33. Lin, X.; Wan, J.; Xie, Y.; Zhang, S.; Lin, C.; Liang, Y.; Guo, G.; Li, S. Task-Oriented Feature-Fused Network With Multivariate Dataset for Joint Face Analysis. IEEE Trans. Cybern. 2019, 50, 1292–1305. [Google Scholar] [CrossRef] [PubMed]
  34. Sainath, T.N.; Vinyals, O.; Senior, A.; Sak, H. Convolutional, long short-term memory, fully connected deep neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brisbane, QLD, Australia, 19–25 April 2015; pp. 4580–4584. [Google Scholar]
  35. Sak, H.; Senior, A.; Beaufays, F. Long short-term memory recurrent neural network architectures for large scale acoustic modeling. In Proceedings of the Fifteenth Annual Conference of the International Speech Communication Association, Singapore, 14–18 September 2014. [Google Scholar]
  36. Wollmer, M.; Blaschke, C.; Schindl, T.; Schuller, B.; Farber, B.; Mayer, S.; Trefflich, B. Online driver distraction detection using long short-term memory. IEEE Trans. Intell. Transp. Syst. 2011, 12, 574–582. [Google Scholar] [CrossRef] [Green Version]
  37. Sesmero, M.P.; Ledezma, A.I.; Sanchis, A. Generating ensembles of heterogeneous classifiers using stacked generalization. Wiley Interdiscip. Rev. Data Mining Knowl. Discov. 2015, 5, 21–34. [Google Scholar] [CrossRef]
  38. Polikar, R. Ensemble based systems in decision making. IEEE Circuits Syst. Mag. 2006, 6, 21–45. [Google Scholar] [CrossRef]
  39. Dietterich, T.G. Ensemble methods in machine learning. In International Workshop on Multiple Classifier Systems; Springer: Cham, Switzerland, 2000; pp. 1–15. [Google Scholar]
  40. Hansun, S. A new approach of moving average method in time series analysis. In Proceedings of the Conference on New Media Studies (CoNMedia), Tangerang, Indonesia, 27–28 November 2013; pp. 1–4. [Google Scholar]
  41. Ehlers, J. FRAMA–Fractal Adaptive Moving Average. Tech. Anal. Stock. Commod. 2005, 23, 81–82. [Google Scholar]
  42. Johnston, F.; Boyland, J.; Meadows, M.; Shale, E. Some properties of a simple moving average when applied to forecasting a time series. J. Oper. Res. Soc. 1999, 50, 1267–1271. [Google Scholar] [CrossRef]
  43. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  44. Genuer, R.; Poggi, J.M.; Tuleau-Malot, C. Variable selection using random forests. Pattern Recognit. Lett. 2010, 31, 2225–2236. [Google Scholar] [CrossRef] [Green Version]
  45. Strobl, C.; Boulesteix, A.L.; Kneib, T.; Augustin, T.; Zeileis, A. Conditional variable importance for random forests. BMC Bioinform. 2008, 9, 307. [Google Scholar] [CrossRef] [Green Version]
  46. John, G.H.; Kohavi, R.; Pfleger, K. Irrelevant features and the subset selection problem. In Machine Learning Proceedings 1994; Elsevier: Amsterdam, The Netherlands, 1994; pp. 121–129. [Google Scholar]
  47. Almuallim, H.; Dietterich, T.G. Learning with Many Irrelevant Features; AAAI: Palo Alto, CA, USA, 1991; Volume 91, pp. 547–552. [Google Scholar]
  48. Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  49. Shalev-Shwartz, S.; Singer, Y.; Srebro, N.; Cotter, A. Pegasos: Primal estimated sub-gradient solver for svm. Math. Program. 2011, 127, 3–30. [Google Scholar] [CrossRef]
  50. Weinberger, K.Q.; Saul, L.K. Distance metric learning for large margin nearest neighbor classification. J. Mach. Learn. Res. 2009, 10, 207–244. [Google Scholar]
  51. Ho, T.K. A data complexity analysis of comparative advantages of decision forest constructors. Pattern Anal. Appl. 2002, 5, 102–112. [Google Scholar] [CrossRef]
  52. Rennie, J.; Shih, L.; Teevan, J.; Karger, D. Tackling the poor assumptions of Naive Bayes classifiers. In Proceedings of the Twentieth International Conference on Machine Learning (ICML 2003), Washington, DC, USA, 21–24 August 2003. [Google Scholar]
  53. Guo, Y.; Hastie, T.; Tibshirani, R. Regularized linear discriminant analysis and its application in microarrays. Biostatistics 2006, 8, 86–100. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  54. Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, Kauai, HI, USA, 8–14 September 2001; Volume 1, pp. 511–518. [Google Scholar]
  55. Ward, J.; Worrall, S.; Agamennoni, G.; Nebot, E. The warrigal dataset: Multi-vehicle trajectories and v2v communications. IEEE Intell. Transp. Syst. Mag. 2014, 6, 109–117. [Google Scholar] [CrossRef]
  56. Luo, S.; Leung, A.P.; Qiu, X. A Machine Learning Framework for Road Safety of Smart Public Transportation. In Proceedings of the International Joint Conferences on Artificial Intelligence (IJCAI) Workshops 2019, Macao, China, 10–16 August 2019. [Google Scholar]
Figure 1. The long short-term memory (LSTM) network used in the proposed boosting method.
Figure 1. The long short-term memory (LSTM) network used in the proposed boosting method.
Sensors 20 04671 g001
Table 1. Features in our dataset to evaluate road safety.
Table 1. Features in our dataset to evaluate road safety.
Feature NameMeaningUsed to Train
our Method?
LOGIDBus identifierNo
GPSDATE
TIME
TimeNo
VELOCITYInstantaneous speedYes
MILEAGEGPS mileageYes
TOTAL
MILEAGE
Total mileageYes
FRONT
PRESSURE
Front pressureYes
REAR
PRESSURE
Rear pressureYes
ENGINESPEEDEngine speedYes
ENGINETEMPEngine temperatureYes
CARSWITCHSwitches of the busNo
CARLIGHT
STATE
Switches of lightNo
CANALARM
STATE
Switches of alarmNo
CREATETIMETimeNo
GPS
VELOCITY
Instantaneous speedYes
DRIVERIDDriver identifierNo
LONGITUDELongitudeYes
LATITUDELatitudeYes
DIRECTIONTurnYes
STATIONIDStation identifierNo
ROUTEIDRoute identifierNo
BUSSTATEBus statusNo
ALARM
STATE
Alarm light statusNo
STATION
MILEAGE
Mileage
each station
Yes
UPDOWNUp and downNo
Table 2. Descriptive statistics on features used for training models.
Table 2. Descriptive statistics on features used for training models.
Feature NameMeanStd. Dev.Median
VELOCITY147.9159.990
MILEAGE65,220,323.412,868,396.877,710,500
TOTAL
MILEAGE
66,584,813.315,653,632.781,771,740
FRONT
PRESSURE
843.471.7852
REAR
PRESSURE
794.432.7796
ENGINESPEED976.5342.6803
ENGINETEMP81.67.183
GPS
VELOCITY
141.2152.890
LONGITUDE113.50.0066113.5
LATITUDE22.10.019522.1
DIRECTION197.3104.8195
STATION
MILEAGE
346.7589.3170
Table 3. Seven state-of-the-art machine learning methods used in the proposed boosting method.
Table 3. Seven state-of-the-art machine learning methods used in the proposed boosting method.
Machine Learning Methods
Support Vector Machine (SVM)
k Nearest Neighbour (KNN)
Random Forest (RF)
Naive Bayes
Discriminant Analysis
Adaptive Boosting (AdaBoost)
Long Short-Term Memory (LSTM)
Table 4. Four components of the LSTM network to avoid the vanishing gradient problem in the RNN.
Table 4. Four components of the LSTM network to avoid the vanishing gradient problem in the RNN.
ComponentsPurposes
Input gatePreprocess the input data
Output gateUpdata the output hidden state
Forget gateReset (forget) the input data
Cell candidateUpdate the output cell state
Table 5. The comparison among eight methods on the Warrigal dataset without our time-series features. The proposed method gets the highest values of accuracy and AUC.
Table 5. The comparison among eight methods on the Warrigal dataset without our time-series features. The proposed method gets the highest values of accuracy and AUC.
Classifying the Warrigal Dataset WITHOUT Our Features
Methods70% Percent Data for Training90% Percent Data for Training
Accu. (%)AUCSpecificitySensitivityAccu. (%)AUCSpecificitySensitivity
Our Method92.90.9230.9320.91293.70.9310.9390.922
AdaBoost79.50.7610.8320.60883.50.7980.8710.652
Simple Bayes77.30.7420.8250.51380.10.7640.8570.522
Discriminant
Analysis
77.10.7280.8150.55182.50.7760.8690.607
KNN83.20.7330.8730.62884.30.7450.8860.628
RF90.70.9030.9230.82891.80.9100.9350.830
SVM75.90.7560.7220.89682.30.8150.7960.906
LSTM91.40.8180.9380.79192.60.8220.9510.801
Table 6. A comparison among eight methods on the Warrigal dataset with our feature extraction method. The comparison between the values of accuracy and AUC in Table 5 and this table shows that the performance of all machine learning methods is improved with our feature extraction method.
Table 6. A comparison among eight methods on the Warrigal dataset with our feature extraction method. The comparison between the values of accuracy and AUC in Table 5 and this table shows that the performance of all machine learning methods is improved with our feature extraction method.
Classifying the Warrigal Dataset WITH Our Features
Methods70% Percent Data for Training90% Percent Data for Training
Accu. (%)AUCSpecificitySensitivityAccu. (%)AUCSpecificitySensitivity
Our Method94.10.9440.9430.92795.80.9520.9620.937
AdaBoost80.70.7810.8330.67484.30.8200.8680.718
Simple Bayes80.20.7590.8280.66982.70.8040.8580.672
Discriminant
Analysis
79.50.7760.8140.69782.90.8080.8480.731
KNN83.50.7510.8540.73787.40.7820.8980.752
RF91.30.9130.9190.87993.90.9370.9460.903
SVM78.20.7550.7460.90782.80.8190.8010.907
LSTM92.70.8250.9410.85394.80.9010.9610.883
Table 7. The comparison among eight methods on our real-world dataset without the proposed feature extraction method. In terms of classification accuracy or AUC, our boosting method outperforms the other state-of-the-art methods in all cases.
Table 7. The comparison among eight methods on our real-world dataset without the proposed feature extraction method. In terms of classification accuracy or AUC, our boosting method outperforms the other state-of-the-art methods in all cases.
Classifying the New Real-World Dataset WITHOUT Our Features
Methods70% Percent Data for Training90% Percent Data for Training
Accu. (%)AUCSpecificitySensitivityAccu. (%)AUCSpecificitySensitivity
Our Method91.20.8700.9210.83892.90.8810.9400.844
AdaBoost78.10.7400.8080.57480.00.7530.8270.592
Simple Bayes62.70.5860.6430.52162.00.5770.6340.515
Discriminant
Analysis
58.40.5390.5890.54360.00.5640.6040.568
KNN72.10.6480.7260.57975.20.6850.7690.619
RF89.70.8160.9240.69691.20.8040.9360.727
SVM57.70.5580.5320.85261.10.5980.5670.873
LSTM77.30.7570.7860.67685.20.7900.8730.694
Table 8. The comparison among eight methods on our real-world dataset with the proposed feature extraction method. After comparing with the values of classification accuracy and AUC in Table 7, it shows that the performance of any method is improved using our feature extraction method.
Table 8. The comparison among eight methods on our real-world dataset with the proposed feature extraction method. After comparing with the values of classification accuracy and AUC in Table 7, it shows that the performance of any method is improved using our feature extraction method.
Classifying the New Real-World Dataset WITH Our Features
Methods70% Percent Data for Training90% Percent Data for Training
Accu. (%)AUCSpecificitySensitivityAccu. (%)AUCSpecificitySensitivity
Our Method95.60.9470.9600.92196.70.9690.9680.957
AdaBoost83.70.7810.8570.68585.10.8020.8670.731
Simple Bayes66.80.6510.6740.62272.10.7020.7240.692
Discriminant
Analysis
77.80.6980.7940.65180.10.7320.8160.686
KNN77.20.7160.7890.64077.60.7270.7900.669
RF93.90.9030.9580.79094.10.9040.9560.827
SVM77.90.6890.7530.85780.40.7320.7810.881
LSTM75.10.7340.7650.64280.30.7670.8200.675
Table 9. Classification accuracies of the eight methods using our real-world dataset. In this experiment for this table, the samples of the training set and test set are collected in four different periods; samples are therefore grouped into four subsets corresponding to the respective time periods.
Table 9. Classification accuracies of the eight methods using our real-world dataset. In this experiment for this table, the samples of the training set and test set are collected in four different periods; samples are therefore grouped into four subsets corresponding to the respective time periods.
On Our Real-World Dataset
MethodsWith Our FeaturesWithout Our Features
Our Method86.4%68.7%
AdaBoost69.3%66.4%
Simple Bayes69.4%59.9%
Discriminant Analysis62.3%58.7%
KNN71.9%56.9%
RF65.6%61.1%
SVM84.1%62.0%
LSTM82.3%63.9%
Table 10. A summary of the training time required for the methods compared in seconds.
Table 10. A summary of the training time required for the methods compared in seconds.
On Our New DatasetOn the Warrigal Dataset
 With Our FeaturesWithout Our FeaturesWith Our FeaturesWithout Our Features
Scales of
Training Set
70%90%70%90%70%90%70%90%
Our Method12.817.47.510.642.1k62.9k21.7k34.5k
AdaBoost2.73.31.92.213.3k17.2k9.72k14.5k
Simple Bayes0.110.130.040.056.6k8.3k4.52k6.1k
Discriminant
Analysis
0.080.100.040.071.7k2.1k1.1k1.5k
KNN0.070.120.060.085.4k7.2k3.2k5.1k
RF0.350.490.310.352.9k3.6k2.1k2.4k
SVM2.12.81.52.612.6k15.4k8.7k12.5k
LSTM1.92.31.62.011.7k14.2k7.1k10.5k
Table 11. A summary of the classification accuracy of the proposed boosting method with different values of parameter U.
Table 11. A summary of the classification accuracy of the proposed boosting method with different values of parameter U.
Values of U23456
Methods Trained with 70% Examples91.2%93.5%93.9%94.1%94.0%
Methods Trained with 90% Examples92.7%93.9%95.6%95.8%95.2%

Share and Cite

MDPI and ACS Style

Luo, S.; Leung, A.P.; Qiu, X.; Chan, J.Y.K.; Huang, H. Complementary Deep and Shallow Learning with Boosting for Public Transportation Safety. Sensors 2020, 20, 4671. https://doi.org/10.3390/s20174671

AMA Style

Luo S, Leung AP, Qiu X, Chan JYK, Huang H. Complementary Deep and Shallow Learning with Boosting for Public Transportation Safety. Sensors. 2020; 20(17):4671. https://doi.org/10.3390/s20174671

Chicago/Turabian Style

Luo, Shengda, Alex Po Leung, Xingzhao Qiu, Jan Y. K. Chan, and Haozhi Huang. 2020. "Complementary Deep and Shallow Learning with Boosting for Public Transportation Safety" Sensors 20, no. 17: 4671. https://doi.org/10.3390/s20174671

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop