Next Article in Journal
Surgical Diseases in North Korea: An Overview of North Korean Medical Journals
Next Article in Special Issue
Effects of COVID-19 on Urban Population Flow in China
Previous Article in Journal
The Influence of South East Asia Forest Fires on Ambient Particulate Matter Concentrations in Singapore: An Ecological Study Using Random Forest and Vector Autoregressive Models
Previous Article in Special Issue
Rural Population Aging and the Hospital Utilization in Cities: The Rise of Medical Tourism in China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cyberattacks Detection in IoT-Based Smart City Applications Using Machine Learning Techniques

by
Md Mamunur Rashid
1,
Joarder Kamruzzaman
2,
Mohammad Mehedi Hassan
3,*,
Tasadduq Imam
4 and
Steven Gordon
1
1
School of Engineering and Technology, CQUniversity, Rockhampton North, QLD 4701, Australia
2
School of Engineering, Information Technology and Physical Sciences, Federation University Australia, Gippsland Campus, Churchill, VIC 3842, Australia
3
Information Systems Department, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
4
School of Business and Law, CQUniversity, Melbourne Campus, Melbourne, VIC 3000, Australia
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2020, 17(24), 9347; https://doi.org/10.3390/ijerph17249347
Submission received: 25 October 2020 / Revised: 4 December 2020 / Accepted: 7 December 2020 / Published: 14 December 2020
(This article belongs to the Special Issue Smart Mobility in Smart City)

Abstract

:
In recent years, the widespread deployment of the Internet of Things (IoT) applications has contributed to the development of smart cities. A smart city utilizes IoT-enabled technologies, communications and applications to maximize operational efficiency and enhance both the service providers’ quality of services and people’s wellbeing and quality of life. With the growth of smart city networks, however, comes the increased risk of cybersecurity threats and attacks. IoT devices within a smart city network are connected to sensors linked to large cloud servers and are exposed to malicious attacks and threats. Thus, it is important to devise approaches to prevent such attacks and protect IoT devices from failure. In this paper, we explore an attack and anomaly detection technique based on machine learning algorithms (LR, SVM, DT, RF, ANN and KNN) to defend against and mitigate IoT cybersecurity threats in a smart city. Contrary to existing works that have focused on single classifiers, we also explore ensemble methods such as bagging, boosting and stacking to enhance the performance of the detection system. Additionally, we consider an integration of feature selection, cross-validation and multi-class classification for the discussed domain, which has not been well considered in the existing literature. Experimental results with the recent attack dataset demonstrate that the proposed technique can effectively identify cyberattacks and the stacking ensemble model outperforms comparable models in terms of accuracy, precision, recall and F1-Score, implying the promise of stacking in this domain.

1. Introduction

Internet of things (IoT) is an interconnected scheme which promotes seamless information exchange between devices (e.g., smart home sensors, environmental sensors, automotive and road-side sensors, medical devices, industrial robots and surveillance devices) [1]. Recently, the emergence of the IoT has significantly increased its use in communities and services around the world, with the number of the linked IoT devices reaching 27 billion in 2017, and the number is projected to hit about 125 billion in 2030 [2]. IoT devices use different types of services, technologies and protocols. As a result, huge complexity will arise to maintain the future IoT infrastructures, which consequently leads to undesirable vulnerability to the system [3,4].
Since IoT devices are used in smart city applications, cyber-attacks can access in an unauthorized manner the details of citizen’s everyday activities without the knowledge of the user or administrator or reconfigure devices to an unsecured setting (e.g., in Miria botnet attack [5,6], a malware that transforms Linux networked devices remotely). In 2019, Symantec recorded a 600% rise in attacks on the IoT platform [7] where attackers tried to manipulate the linked nature of those devices.
Smart city applications pose several security challenges. Firstly, zero-day attacks can occur by exploiting vulnerabilities in different protocols in smart city applications. Secondly, is it possible to identify cyber-attacks from the network intelligently before it disrupts smart city operations? Thirdly, the IoT devices used in smart cities are resource (e.g., memory) constrained, are typically resource constrained, have limited onboard functionality for security operations and send captured data to cloud servers for processing. Existing intrusion detection systems (IDS) do not take IoT devices into account. Combining all these issues, is it is possible to design an IDS design an IDS that is tailored for IoT networks?
The data collected from the IoT system is stored on the cloud computing environment which has progressively advanced processors and adequate memory assets. However, the volume of data transmitted from the IoT terminal layer to the cloud has increased rapidly with the recent increases in IoT devices and this causes delay and congestion problems in the cloud. Fog computing is designed as a possible solution to these problems [8]. The fog layer devices can share a greater amount of computing load originally transferred to the cloud. This reduces energy consumption, network traffic and latency and removes the data storage and transmission problem. It also aims to push the computation process near the edge device, enabling a quick response to the IoT-based smart city applications. The benefits of cyber attack detection in the fog layer are two folds [9]. Firstly, the ISP or network administrator can take necessary steps to prevent large damage if attacks (e.g., infected devices) are identified early in the fog layer. Secondly, it will not interrupt the normal flow of urban life.
In the literature, some techniques (e.g., signature base techniques) have been proposed to resolve the above-mentioned issue. In the signature-based technique, a collection of previously produced signatures (attacks) are checked against the current suspicious samples [10]. If the signature extraction method is not fully able to capture the distinct feature of attacks or attack families, it may lead to misdetection of an attack or produce false alarm [11]. This technique is not suitable is not suitable for identifying unknown attacks and suffers from high processing overhead. Machine learning techniques can detect attacks during runtime and take less processing time compared to other techniques.
In this paper, we explore a machine learning-based attack and anomaly detection technique in IoT-based smart city applications. This technique is able to identify infected IoT devices which is a major challenge in the cloud computing environment [12,13]. The technique is based on the implementation of a training model in the distributed fog networks that can learn intelligently from training near to IoT layer devices and detect attack and anomaly.
A single classifier is often insufficient to develop an effective IDS, motivating researchers to build an ensemble model of classifiers. Taking a multitude of models into account, ensemble methods combine those models to generate one final model. Research has demonstrated that the ensemble model produces better performance compared to the single classifier [14]. However, there are many factors (e.g., feature selection and base classifier) that need to be considered carefully to ensure enhanced performance by the ensemble method. The most suitable ensemble techniques are bagging [15], boosting [16] and stacking [17]. In this paper, we use individual classifiers as well as ensemble techniques to achieve better IDS performance in terms of different evaluation metrics such as accuracy, precision, recall and F1-Score.
The contributions of this paper are summarized as follows:
  • We explore a machine learning based attack and anomaly detection technique through analyzing network traffic in distributed fog networks over the IoT-based systems.
  • Existing works have generally used signature-based techniques to detect attacks and anomalies. These techniques suffer from high overheads and are vulnerable to known threats. In this paper, we explore the feasibility of ensemble based learning as compared to single model classifiers for identifying cyberattacks in IoT-based smart city applications. Further, we consider a multi-class classification setting as compared to binary class prediction considered in most relevant works. On top of these, we consider an integration of feature selection and cross validation, which even common machine learning approaches have not been well focused in existing literature for this domain.
  • Extensive evaluation incorporating the above integrations shows that the ensemble of machine learning-based classifiers works better in accurately identifying attacks and their types than single classifiers.
The paper is organized in the following sections. Section 2 discusses the related works. Section 3 discusses the IoT-based Smart city framework. Section 4 presents the proposed anomaly detection model. Section 5 presents the experimental results. Section 6 gives concluding remarks.

2. Related Works

In the literature, many studies have been introduced to enhance the IDS performance. In this section, we highlight the recent notable works that have used machine learning techniques as well as ensemble methods.

2.1. IDS Based on Machine Learning Techniques

In [18], Pahl and Aubet introduced a machine learning based technique that can predict IoT service behavior by only observing the communication between services in a distributed multi-dimensional IoT microservices in an IoT site. This technique continually learns microservice models inside in an IoT site where K-means and BIRCH based clustering techniques [19] are applied. In this case, if the cluster centers are within the three times standard deviation gap, they are grouped into the same one. The model revises cluster formation using an online learning communication model. The overall accuracy for anomaly detection by this technique is 96.5% with 0.2% false positive rate.
In [20], a joint trust light probe based defense (TLPD) mechanism was introduced to detect On and Off attack in an industrial IoT site, originated from malicious network nodes. Here, the On and Off attack meant a malicious node might target the IoT network when it is in an On or Off state. The framework was designed for the identification of anomalies using a light probe routing mechanism with the measurement of confidence estimation for each neighbor node.
Diro and Chilamkurti [10] proposed a deep learning model to detect distributed attacks in a social IoT network where they compared the performance of the deep model with a shallow neural network using the NSL-KDD [21] open source dataset that captures attack data in the distributed and centralized system. They evaluated the performance of the deep and shallow models with two-class (normal and attack) and four-class (normal, DoS, Probe, R2L and U2R) categories. For binary-class and multi-class identification, their model achieved accuracies of 99.2% and 98.27% as well as 95.22% and 96.75%, respectively, for the deep and shallow models.
In [22], Pajouh et al. proposed a two-stage dimension reduction and classification technique to detect anomaly in IoT backbone networks where they detected low frequency attacks such as user to root (U2R) and remote to local (R2L) attacks from NSL-KDD dataset because of their detrimental consequences. They used principal component analysis (PCA) and linear discriminate analysis (LDA) feature extraction method to reduce the feature of the dataset and then used naïve bayes and K-nearest Neighbor (KNN) to identify anomaly and achieved 84.82% identification rate.
In [23], Kozik et al. introduced an attack detection technique that used extreme learning machine (ELM) [24] method in the Apache Spark cloud architecture. ELM architecture and properties allow for efficient computation and analysis of the Netflow formatted data that are collected from the fog computing environment. This work concentrated on three main cases in IoT systems—scanning, command and control and infected host—and attained accuracy levels of 99%, 76% and 95%, respectively.
In [25], Hasan et al. proposed a data analysis-based method to detect attacks on IoT infrastructure which overcomes the data processing overhead of signature based techniques. Their proposed solution is able to identify and prevent the systems from attacks when it faces any irregular behavior. They performed their experiment on the publicly available IoT dataset [18]. They explored several machine learning techniques such as DT, RF, LR, SVM and ANN, among which RF classifier yielded the best results.
In [26], a random forest-based anomaly detection model was proposed that can detect infected IoT devices at distributed fog nodes. Experimenting with the UNSW-NB15 dataset [27], their binary (normal and attack) random forest (RF) classifier considered only 12 out of 49 features from the dataset. These 12 features were extracted by using ExtraTreeClassifer [28]. Performance analysis showed that they achieved 99.34% accuracy with 0.02% false positive rate.
In [29], a deep learning model was studied on NSL-KDD, UNSW-NB15, WSN-DS [30] and CICIDS 2017 [31] datasets to identify cyberattacks. They concluded that the deep learning model performs better compared to the other machine learning techniques.

2.2. IDS Based on Ensemble Techniques

In the literature, several ensemble methods based IDSs are proposed to enhance accuracy over base classifiers. In [32], ANN and Bayesian net based ensemble method was proposed where they used gain ratio (GR) feature selection technique and performance was evaluated on KDD’99 [33] and NSL-KDD datasets where ensemble methods achieved 99.42% and 98.07% accuracy, respectively.
In [34], Haq et al. proposed an ensemble method that combines Naive Bayes, Bayesian Net and decision tree classifier. They extracted the common features by using Best First Search, Genetic and Rank Search feature selection techniques. The ensemble technique produced 98% true positive rate when tested with 10-fold cross validation method. Gaikwad et al. [35] introduced a bagging ensemble method where they used REPTree as a base classifier. Their model achieved 81.29% accuracy on NSL-KDD dataset. In [36], Jabbar et al. proposed an ensemble method comprising alternating decision tree (ADTree) and KNN, and the performance evaluation demonstrated that the proposed ensemble achieved better detection rate (~99.8%) compared to the existing techniques.
In [37], Zhou et al. proposed feature selection and ensemble method based IDS model where a combination of correlation-based feature selection (CFS) and Bat algorithm [38] were used for optimal feature selection, followed by an ensemble method comprising DT, RF and Forest by Penalizing Attributes (Forest PA) algorithms. Experiments were performed on NSL-KDD, AWID [39] and CIC-IDS2017 datasets, achieving 99.8%, 99.5% and 99.8% accuracy, respectively.
In [40], a hybrid intrusion detection system was introduced comprising C5 classifier and One class support vector machine. The main focus of this work was to identify the common instruction and zero-day attack by using a Bot-IoT dataset [41] that contains IoT network traffic with several types of attacks. Performance analysis demonstrated that the proposed hybrid model attained higher accuracy to intrusion detection compared to Signature Intrusion Detection System (SIDS) and Anomaly-based Intrusion Detection System (AIDS).
In [42], bagging and boosting ensemble methods were proposed where the authors used decision tree and random forest tree as the base classifiers. Experiments were performed on the NSL-KDD dataset and it was found that bagging with decision trees gives better results.
Table 1 summarizes the notable works addressing intrusion and anomaly detection in networks using machine learning techniques and in some works their ensemble techniques.
Despite such a wide exploration, it is clear that different works have used different data and achieved different performance outcomes, which is not surprising due to machine learning algorithms’ often dependence on data and differing contexts may result in different outcomes. However, UNSW-NB15, the latest version of data covering intrusion detection in IOT devices, has found only relatively less exploration. In this research, we used this dataset especially considering its concurrency. Contrary to Alrashdi et al. [26], our work is not limited to binary classification or RF classifier alone. Further, in this paper, we explore the multi-class problem. In other words, our focus is not limited to only identify the normal/abnormal state of data but also to detect the exact type of attacks in fog nodes within smart city infrastructure. Our work also differs due to analyzing the performance with the base as well as ensemble classifier.

3. IoT-Based Smart City Framework

Smart city is an integrated framework where IoT technology, smart systems and information and communication technology (ICT) are collectively used to enhance the quality and performance of the different city services such as transportation, health systems, pollution control and energy distribution. A smart city framework, as based on existing literature [26], is shown in Figure 1 and consists of the following three layers: terminal layer, fog layers and cloud layer.
The cloud layer contains storage resources (e.g., servers and virtual machines) to store as well as maintain a large amount of data. The fog layer acts as a bridge between the terminal layer devices and cloud layer and is responsible to ensure the computational process and management at the edges of the network. The fog layer is more effective at identifying the different cyber-attacks than the centralized cloud layer. The terminal layer consists of a set of IoT devices (sensors) that are installed within the city to collect data.
For several reasons, IoT networks and applications are vulnerable against attacks. Firstly, most IoT devices have limited resources (e.g., small processing power and memory) and as a result suffer from limited processing capability. Secondly, IoT devices are interconnected to different protocols and the increasing number of IoT devices further causes latency in cloud centers. Thirdly, sometimes IoT devices are unattended, which makes it possible for an intruder to physically access them. Fourthly, the greater part of the data communication is wireless, exposing it to eavesdropping.
As a consequence, conventional IDS systems often fail to detect the IoT attack accurately [43]. Thus, an attacker can successfully compromise vulnerable IoT devices to connect to smart city routers and devices located at various places such as homes, shopping malls, restaurants, hotels and airports. By doing so, an attacker who compromises these IoT devices may obtain sensitive data such as information of credit card, stream video and similar personal information.
One of the key issues that smart city framework and infrastructure must ensure is its ability to deliver services in a sustainable manner to meet the needs of the current and future generations of citizens [44,45]. Some ongoing smart city projects such as those initiated in Hong Kong and Masdar city in Abu Dhabi [46] have already been criticized because of the vulnerable urban development plan and consequently doubts about sustainability of the services. The management of several facets of sustainability programs inside smart cities are facilitated by IoT, and this exposes organizations to the risks of failure from the network unavailability, security breaches and damage of IoT infrastructure from natural disasters [47]. Further, sustainable operation of services such as intelligent transportation systems, smart buildings and sustainable usage of resources such as water and energy supply, garbage disposal, etc. are highly dependent on IoT and related cyber-physical systems [48]. Machine learning techniques are used to better manage those smart city services and resources in an autonomous manner [45,49]. In addition, machine learning techniques can well detect intrusion and cyberattacks in industrial IoT [50], which therefore can enhance sustainability and ensure uninterrupted services in smart cities by thwarting attacks and intrusion on respective IoT systems. However, there is a need for more research on machine learning implementation and model verification in terms of security and privacy [51].
In this paper, we hence explore the feasibility of both ensemble-based learning and single-model classifiers for identifying cyberattacks in IoT-based smart city applications.

4. Proposed Anomaly Detection Model

Our proposed model is shown in Figure 2. The model tracks the network traffic that goes through each fog node. Since fog nodes are closest to IoT sensors, they will be more effective at identifying the cyber-attacks at fog nodes instead of the cloud center. In this way, an attack can be quickly detected, and the IoT and network administrators can be notified of such attacks, which will then assist them to evaluate and upgrade their systems.
Notably, IDS can be categorized as host-based IDS (HIDS) and network-based IDS (NIDS). In this work, we choose anomaly-based NIDS. HIDS requires the installation of software on each network-connected device to track and identify the malicious activity focused solely on that device and is not suitable for most IoT devices which are resource constrained and support limited functionality (e.g., smart lamps, watches and lock-doors). Again, signature-based NIDS suffers from higher computational cost in storing attacks in a database and fails to detect a new attack in potential network traffic [52], which makes anomaly-based NIDS most suitable in our case. Data collected from this NIDS are used to build an ensemble of ML models to identify abnormal activities in the IoT fog networks.

4.1. Description of Used Datasets

We used the UNSW-NB15 [27] and CICIDS2017 [31] datasets. The reasons for using these datasets are two fold: firstly, they are relevant to the proposed smart city infrastructure concept of this paper, and, secondly, both contain samples of the recent types of attacks observed in IoT infrastructure.

4.1.1. UNSW-NB15 Dataset

The UNSW-NB15 dataset [27] is a recent and highly useful IDS dataset containing the modern attacks. In 2015, the UNSW-NB15 dataset was developed to track and identify normal and attack network traffic and the raw network packets were generated by the IXIA PerfectStorm tool in the Australian Centre for Cyber Security (ACCS) cyber range lab [53]. The dataset has been preprocessed through cleaning, visualization, feature engineering and vectorization. This original dataset contains over 2.54 million samples, of which a random portion (175,341 samples) is used in our work. The considered dataset contains 56,000 and 119,241 samples, respectively, representing the benign and attack conditions. We divided our dataset into training set (140,272 samples) and test set (35,069 samples), each set containing attack and benign samples in the same ratio as the original dataset. The distribution of different attacks and anomaly across the dataset is shown in Table 2.
A brief description of the attack categories is below [27]:
  • Fuzzers: It tries to cause a program or network to be suspended by feeding randomly generated data.
  • Analysis: It includes various attacks of port scan, spam and html file penetrations.
  • Backdoors: A security mechanism is bypassed stealthily to access a device or its data.
  • Denial-of-service (DoS): The IoT network and smart city services are made inaccessible to legitimate users through malicious attempts. Distributed denial-of-service attacks (DDoS) overwhelm the target websites and online services with more traffic than the server or network can accommodate.
  • Exploits: The attacker learns the security flaws in the installed software and hardware and leverages the vulnerabilities of the IoT devices and system.
  • Generic: It is any technique that works against all blockciphers (with a particular block and key size), without considering the block-cipher structure.
  • Reconnaissance: All actions capable of simulating assaults that gather information about vulnerabilities.
  • Shellcode: It is a small code segment used as a payload to exploit vulnerability in the software.
  • Worms: It replicates itself to spread to other devices and machines using connectivity.

4.1.2. CICIDS2017 Dataset

The CICIDS2017 dataset was generated by the Canadian Institute for Cybersecurity (CIC) in 2017. This dataset includes benign and most recent cyberattacks, namely, DoS, DDoS, PortScan, SQL injection, Infiltration, Brute Force and Bot [31]. CICIDS2017 comprises of 2,830,743 records in eight files and each record includes 78 different features. In our experiment, we used 190,774 records where 148,777 are benign and 41,997 contain various types of attacks. Table 2 shows detailed information on sample distribution.

4.2. Data Pre-Processing

Feature selection is one of the key principles that greatly impacts the model’s efficacy by selecting only those features that are most relevant and thereby reduces over-fitting, improves accuracy and reduces training time. We used information gain ratio, which is a ratio of information gain to the intrinsic information proposed by Quinlan et al. [54], to select the top 25 features which are highly relevant to the prediction for both datasets. The information gain score of the features of UNSW-NB15 dataset is shown in Table 3 and Table 4. Out of 42 (UNSW-BC15) and 78 (CICIDS2017) features, the top 25 were selected based on their information gain ratio. A higher ratio for a feature can contribute more to identifying the benign and malware applications. We only consider the features whose information gain was greater than predetermined threshold 0.5 for UNSW-NB15 dataset and 0.85 for CICIDS2017 dataset.
In feature engineering phases, at first, we identify the type of features in the datasets. In UNSW-NB15 dataset, among the above mentioned 25 features, “proto” and “service” are categorical features and the rest are numerical data. This categorical data are converted into vectors. While categorical data can be translated to vectors in different ways such as ‘Label Encoding’ and ‘One Hot Encoding’, ‘Label Encoding’ [55] technique was used in this research.

4.3. Theoretical Consideration

Several machine learning techniques and ensemble methods were used for model building and performance evaluation. We used LR [56], SVM [57], DT [58], RF [59], KNN [60] and ANN [10] machine learning algorithms, which are widely used in the literature to design IDS scheme.
Ensemble methods are a widely used approach in machine learning that combines several base models to generate one optimal predictive model [14]. Taking a multitude of models into account, an ensemble method combines those models to generate one final model. It is based on the principle that a group of weak learners (models) comes together to form a strong learner, thereby increasing the model’s accuracy. There are three types of ensemble techniques used in the literature. Bagging [15] is a parallel ensemble technique where the base learners are generated in parallel to improve the strength and accuracy of machine learning algorithms. Boosting [16] is a sequential ensemble technique where the base learners are generated in sequence to reduce bias and variance of supervised machine learning techniques. Stacking [17] is an ensemble learning technique incorporating predictions of several base classification models into a new dataset and used as the input for another classifier which is then used to solve the problem.

4.4. Evaluation Criteria

In this subsection, we describe some performance matrices such as accuracy, precision, recall, F1-Score and ROC curves which are widely used in evaluating the model performance in anomaly detection applications.
These performance metrics are defined by using the following parameters:
  • t p = true positive
  • t n = true negative
  • f p = false positive
  • f n = false negative
  • p = total positive = t p + f n
  • n = total negative = t n + f p
Accuracy indicates the overall performance of the model with respect to both benign and attack classes and is defined as follows:
A c c u r a c y = t _ p + t _ n p + n
Precision gives the information about how many selected items are relevant among the retrieved items and can be defined as follows:
P r e c i s i o n = t _ p t _ p + f _ p
Recall gives the information about how many relevant items are selected from the total number of relevant items and is defined as follows:
R e c a l l = t _ p t _ p + f _ n
F1-Score can be derived from both precision and recall as follows:
F 1 S c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
The Receiver operating characteristic (ROC) curve is utilized to summarize a classifier’s performance over all possible decision thresholds in a graph, and it is generated by plotting the true positive rate (tpr) against the false positive rate (fpr). Equations (5) and (6) show the calculation of true positive rate and false positive rate, respectively.
t p r = t _ p t _ p + f _ n
f p r = f _ p t _ p + f _ n

5. Experimental Results

Experiments were implemented using Python programming language and several libraries such as Pandas, Numpy, Matplotlib, sklearn and Keras on a HP (ELITEBOOK) laptop where the operating system was Windows 10 Education 64-bit and the processor was Intel(R) Core(TM) i5-8350U CPU @ 1.70 GHz 1.9 GHz with 16 GB RAM.
To test the performance of the base classifier as well as the ensemble classifier, 10-fold cross-validation (CV) was used where the provided dataset was randomly divided into 10 equal size subsets. Out of these 10 subsets, nine were used to build the model classifier and the remaining one was used as a test set. The same procedure was repeated ten times to ensure that each subset was used once as the test dataset. Finally, the mean accuracy summarized from each classifier in each fold was noted. Figure 3 represents different evaluation metrics for different classifiers on the training and test datasets.
We first show the performance of the different classifiers in terms of accuracy which is presented in Figure 3a. Here, the task was to classify an unknown sample into one of the ten categories for UNSW-NB15 dataset and eight categories for CICIDS2017 dataset, as shown in Table 3. For UNSW-BC15 and CICIDS2017 datasets, the accuracy of the machine learning algorithms LR, SVM, DT, RF, ANN and KNN on the test dataset are 72.32% and 93.60%, 71.49% and 92%, 80.69% and 99.7%, 81.77% and 99.7%, 78.89% and 94.2% and 78.23% and 99.7% respectively. The accuracy of ensemble methods bagging (RF as the base-learner), boosting (DT as the base-learner), stacking (base-learners of RF and ANN and Meta-learner of DT) are 82.36% and 99.7%, 83.30% and 99.8% and 83.84% and 99.9%, respectively. Among the algorithms, SVM shows poor and least performance while DT and RF shows better results compared to others. On the other hand, stacking ensemble, constructed from base- and meta-classifiers, shows better performance compared to others.
We show the performance measure in terms of precision in Figure 3b. The precision for LR, SVM, DT, RF, ANN and KNN on test dataset are 72% and 92%, 70% and 94%, 81% and 99.8%, 82%, and 99.8%, 78% and 94.5% and 79% and 99.7%, respectively. The precision of ensemble methods bagging, boosting and stacking are 82% and 99.7%, 83% and 99.8% and 83% and 99.9%, respectively. Similar to the accuracy metric, SVM shows the least precision for UNSW-NB15 dataset. However, LR shows the least precision for CICIDS2017 and RF shows better results compared to others. On the other hand, the stacking ensemble method performances better compared to others.
The performance in terms of recall is shown in Figure 3c. The recall for LR, SVM, DT, RF, ANN and KNN on test dataset are 72% and 94%, 71% and 92%, 81% and 98%, 82% and 99.8%, 79% and 94.3% and 78% and 99.7%, respectively. The recall values of the ensemble methods bagging, boosting and stacking are 82% and 99.8%, 83% and 99.9% and 83% and 99.9%, respectively. Once again, ensemble techniques yield better performance compared to the base classifier and the stacking ensemble method outperforms others. Finally, we demonstrate the performance measure in terms of F1-score in Figure 3d. The F1-score for LR, SVM, DT, RF, ANN and KNN on test dataset are 71% and 92%, 70% and 94%, 80% and 99.7%, 81% and 99.7%, 78% and 94% and 78% and 99.7%, respectively. The F1-score of ensemble methods bagging, boosting and stacking are 81% and 99.8%, 81% and 99.9% and 83% and 99.9%, respectively. Once again, ensemble techniques show better performance than base classifier and the stacking ensemble method outperforms other classifiers considered in this research.
The results show that the ensemble of learning models provides better performance than the single model classifiers on both test datasets. This implies, while existing works on the data have focused on single learning model, ensemble classifiers such as stacking represent a promising approach for application in this domain.
We also experimented with how the classifier performs when applied in a multi-class classification context. More precisely, we considered each type of attack as a separate class and then assessed the classifiers’ ability in identifying the attack from a normal situation. The results are shown in Table 5 and Table 6 for UNSW-NB15 and CICIDS2017 datasets, respectively. The results illustrate that, for various types of attack, DT and RF perform better in comparison to other algorithms. On the other hand, the stacking ensemble technique shows significant improvement compared to bagging and boosting in some cases. For example, on the UNSW-NB15 dataset, in the DoS attack, stacking yields an F1-score of 0.45 vs. 0.24 for bagging and boosting, while, in the Worm attack, these scores are 0.57, 0.37 and 0.33, respectively. In most of the other types of attacks, stacking attains an F1-score of above 0.75. On the other hand, on CICIDS2017, in the Bot attack, stacking ensemble achieved 0.950 vs. 0.898 and 0.942 for bagging and boosting.
Figure 4 shows the Receiver Operating Characteristic Curves for base and ensemble classifiers on UNSW-BC15 dataset. We found that, among the base classifiers, ANN shows better performance. On the other hand, among the ensemble techniques, boosting and stacking demonstrate almost the same results.
Table 7 and Table 8 show a comparison of the multi-class classification performance attained by ensemble approaches to a recent work [29] on UNSW-BC15 and CICIDS2017 datasets, respectively. The accuracies attained in our work using LR, SVM, DT, RF and KNN are 72.32% and 93.6%, 71.49% and 92%, 80.69% and 99.7%, 81.77% and 99.7% and 78.23% and 99.6%, respectively, while those in [29] are, respectively, 53.8% and 87%, 58.1% and 79.9%, 73.3% and 94%, 75.5% and 94.4% and 62.2% and 90.0%, albeit with some differences in the way the datasets were used. In [29], the researchers experimented with the boosting ensemble technique and achieved accuracies of 60.8% and 64.1 %, which are significantly lower than the accuracies attained in our work (83.3% and 99.9%) using stacking ensemble.
The results of individual classifiers as well as ensemble methods for the binary class classification are shown in Table 9 and Table 10 on UNSW-BC15 and CICIDS2017 datasets, respectively. We used the same metric as the multi-class classification. The highest accuracies of 95.45% and 99.7% and F1-scores of 95% and 99.8% by an individual classifier were achieved with RF classifier and the values of those metrics rose to 96.83% and 99.9% and 97% and 99.9%, respectively, when the stacking ensemble technique was used.
A possible reason for the proposed model’s significantly better performance compared to Vinayakumar et al. [29] is that they did not consider any feature selection. The existing work experimented with all features for both datasets. However, our proposed model considers an information gain-based feature selection technique and finally uses only 25 most important features based on their information gain ratio.
Notably, a question may be raised as to the complexity of using ensemble models as compared to a single classifier. With technological advances, however, processing units such as mobile devices are becoming increasingly faster and memory resources are becoming increasingly cheaper—a reason fog computing potentially has seen application of a wide range of algorithms including ensemble techniques [61,62]. There are also active investigation on efficient allocation of resources in fog computing [63]. Further, research has devised fog system architecture that can exploit ensemble learning without increasing latency of the system substantially [62]. Arguably, the stacking approach considered in this article can be rolled out using the architecture and efficient resource allocation mechanism. Thus, despite some increases in complexity, the finding that stacking can outperform single classifiers for counterattacks detection in IoT smart city applications has thus a notable value, especially with missing a cyberattack being linked to a high cost.
For example, the model building time for ten runs for each of the base classifiers (DT, RF and ANN) and the stacking ensemble technique on both datasets are shown in Table 11. The table shows that the model building times by the classifiers are 1.4, 2.17, 6.8 and 25.6 s for the UNSW-NB15 dataset and 5.3, 4.35, 7.4 and 27.09 s for the CICIDS2017 dataset, respectively, which shows that DT takes the least time to build the model for UNSW-NB15 while RF is the fastest for CICIDS2017. On the other hand, since the stacking ensemble model deals with more complexity by combining several base classifiers, it takes longer time to build the model for the both datasets. The time taken to test the model on a single sample by the classifiers is, respectively, 0.48, 2.53, 1.91 and 5.70 μ s for UNSW-NB15 and 0.42, 1.57, 1.80 and 4.19 μ s for CICIDS2017, suggesting that DT and RF take the least amount of time compared to others in both datasets. Thus, the model takes very little time, in the range of μ s, to test whether an activity is malicious or not.

6. Conclusions

In this paper, we explore the feasibility of an ensemble based learning with single model classifiers for identifying cyberattacks within the IoT-based smart city applications. Our experiments with the most recent IoT attack database show that our ensemble approach, especially stacking, performs better than single models in identifying attacks from benign samples. Our approach employs an information gain based feature selection technique to identify the most influential features before building the model. Furthermore, in classifying attack types, our ensemble approach with stacking also leads to better performance than the single or other ensemble models used in recent works in terms of accuracy, precision, recall and F1-score metrics. Our future work will explore deep learning techniques to further enhance IoT attack detection performance.
Lastly, with automation and smart cities becoming increasingly popular, they are also increasingly being exposed to cyber threats. A denial of access or privacy intrusion within an automated system can greatly harm individual citizens and carry a substantial cost at both individual and jurisdiction levels. There can also be health risks if systems handling emergency events (e.g., accident and fire) are compromised. Our results indicating that stacking of classifiers can better detect cyberattacks in the smart city systems go beyond technical contributions and carry economic and social implications. Future research will provide further insights in this respect.

Author Contributions

Conceptualization, M.M.R., J.K. and T.I.; methodology, M.M.R., J.K. and T.I.; software, M.M.R.; validation, M.M.R., J.K. and T.I.; resources, M.M.R.; writing—original draft preparation, M.M.R., J.K. and T.I.; writing—review and editing, J.K., S.G. and M.M.H.; supervision, J.K. and M.M.H.; and funding acquisition, M.M.H. and M.M.R. All authors have read and agreed to the published version of the manuscript.

Funding

The authors are grateful to King Saud University, Riyadh, Saudi Arabia for funding this work through Researchers Supporting Project number RSP-2020/18

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chowdhury, A.; Karmakar, G.; Kamruzzaman, J. The Co-Evolution of Cloud and IoT Applications: Recent and Future Trends; IGI Global: Hershey, PA, USA, 2019; pp. 213–234. [Google Scholar]
  2. Howell, J. Number of Connected Iot Devices Will Surge to 125 Billion by 2030, Ihs Markit Says-Ihs Technology; IHS Markit: London, UK, 2017; p. 1. [Google Scholar]
  3. Pahl, M.O.; Aubet, F.X.; Liebald, S. Graph-based IoT microservice security. In Proceedings of the NOMS 2018-2018 IEEE/IFIP Network Operations and Management Symposium, Taipei, Taiwan, 23–27 April 2018; pp. 1–3. [Google Scholar]
  4. Restuccia, F.; D’Oro, S.; Melodia, T. Securing the internet of things: New perspectives and research challenges. arXiv 2018, arXiv:1803.05022. [Google Scholar]
  5. Yar, M.; Steinmetz, K.F. Cybercrime and Society; SAGE Publications Limited: Thousand Oaks, CA, USA, 2019. [Google Scholar]
  6. Antonakakis, M.; April, T.; Bailey, M.; Bernhard, M.; Bursztein, E.; Cochran, J.; Durumeric, Z.; Halderman, J.A.; Invernizzi, L.; Kallitsis, M.; et al. Understanding the mirai botnet. In Proceedings of the 26th USENIX Security Symposium (USENIX Security 17), Vancouver, BC, Canada, 16–18 August 2017; pp. 1093–1110. [Google Scholar]
  7. Symantec. Internet Security Threat Report; Symantec: Tempe, AZ, USA, 2020; p. 1. [Google Scholar]
  8. Yousefpour, A.; Ishigaki, G.; Jue, J.P. Fog computing: Towards minimizing delay in the internet of things. In Proceedings of the 2017 IEEE International Conference on Edge Computing (EDGE), Honolulu, HI, USA, 25–30 June 2017; pp. 17–24. [Google Scholar]
  9. Abeshu, A.; Chilamkurti, N. Deep learning: The frontier for distributed attack detection in fog-to-things computing. IEEE Commun. Mag. 2018, 56, 169–175. [Google Scholar] [CrossRef]
  10. Diro, A.A.; Chilamkurti, N. Distributed attack detection scheme using deep learning approach for Internet of Things. Future Gener. Comput. Syst. 2018, 82, 761–768. [Google Scholar] [CrossRef]
  11. Galal, H.S.; Mahdy, Y.B.; Atiea, M.A. Behavior-based features model for malware detection. J. Comput. Virol. Hacking Tech. 2016, 12, 59–67. [Google Scholar] [CrossRef]
  12. Hossain, M.M.; Fotouhi, M.; Hasan, R. Towards an analysis of security issues, challenges, and open problems in the internet of things. In Proceedings of the 2015 IEEE World Congress on Services, New York, NY, USA, 27 June–2 July 2015; pp. 21–28. [Google Scholar]
  13. Habibzadeh, H.; Soyata, T.; Kantarci, B.; Boukerche, A.; Kaptan, C. Sensing, communication and security planes: A new challenge for a smart city system design. Comput. Netw. 2018, 144, 163–200. [Google Scholar] [CrossRef]
  14. Hansen, L.K.; Salamon, P. Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 993–1001. [Google Scholar] [CrossRef] [Green Version]
  15. Breiman, L. Bagging predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef] [Green Version]
  16. Freund, Y.; Schapire, R.E. Experiments with a new boosting algorithm. In icml; Citeseer: Princeton, NJ, USA, 1996; Volume 96, pp. 148–156. [Google Scholar]
  17. Wolpert, D.H. Stacked generalization. Neural Netw. 1992, 5, 241–259. [Google Scholar] [CrossRef]
  18. Pahl, M.O.; Aubet, F.X. All eyes on you: Distributed Multi-Dimensional IoT microservice anomaly detection. In Proceedings of the 2018 14th International Conference on Network and Service Management (CNSM), Rome, Italy, 5–9 November 2018; pp. 72–80. [Google Scholar]
  19. Aggarwal, C.C.; Philip, S.Y.; Han, J.; Wang, J. A framework for clustering evolving data streams. In Proceedings of the 2003 VLDB Conference, Berlin, Germany, 9–12 September 2003; pp. 81–92. [Google Scholar]
  20. Liu, X.; Liu, Y.; Liu, A.; Yang, L.T. Defending ON–OFF attacks using light probing messages in smart sensors for industrial communication systems. IEEE Trans. Ind. Informatics 2018, 14, 3801–3811. [Google Scholar] [CrossRef]
  21. Tavallaee, M.; Bagheri, E.; Lu, W.; Ghorbani, A.A. A detailed analysis of the KDD CUP 99 data set. In Proceedings of the 2009 IEEE Symposium on Computational Intelligence for Security and Defense Applications, Ottawa, ON, Canada, 8–10 July 2009; pp. 1–6. [Google Scholar]
  22. Pajouh, H.H.; Javidan, R.; Khayami, R.; Ali, D.; Choo, K.K.R. A two-layer dimension reduction and two-tier classification model for anomaly-based intrusion detection in IoT backbone networks. IEEE Trans. Emerg. Topics Comput. 2016, 7, 314–323. [Google Scholar] [CrossRef]
  23. Kozik, R.; Choraś, M.; Ficco, M.; Palmieri, F. A scalable distributed machine learning approach for attack detection in edge computing environments. J. Parallel Distrib. Comput. 2018, 119, 18–26. [Google Scholar] [CrossRef]
  24. Huang, G.B.; Zhou, H.; Ding, X.; Zhang, R. Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2011, 42, 513–529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Hasan, M.; Islam, M.M.; Zarif, M.I.I.; Hashem, M. Attack and anomaly detection in IoT sensors in IoT sites using machine learning approaches. Internet Things 2019, 7, 100059. [Google Scholar] [CrossRef]
  26. Alrashdi, I.; Alqazzaz, A.; Aloufi, E.; Alharthi, R.; Zohdy, M.; Ming, H. AD-IoT: Anomaly detection of IoT cyberattacks in smart city using machine learning. In Proceedings of the 2019 IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 7–9 January 2019; pp. 305–310. [Google Scholar]
  27. Moustafa, N.; Slay, J. The evaluation of Network Anomaly Detection Systems: Statistical analysis of the UNSW-NB15 data set and the comparison with the KDD99 data set. Inf. Secur. J. A Glob. Perspect. 2016, 25, 18–31. [Google Scholar] [CrossRef]
  28. Raschka, S.; Mirjalili, V. Python Machine Learning; Packt Publishing Ltd.: Birmingham, UK, 2017. [Google Scholar]
  29. Vinayakumar, R.; Alazab, M.; Soman, K.; Poornachandran, P.; Al-Nemrat, A.; Venkatraman, S. Deep learning approach for intelligent intrusion detection system. IEEE Access 2019, 7, 41525–41550. [Google Scholar] [CrossRef]
  30. Almomani, I.; Al-Kasasbeh, B.; Al-Akhras, M. WSN-DS: A dataset for intrusion detection systems in wireless sensor networks. J. Sens. 2016, 2016, 4731953. [Google Scholar] [CrossRef] [Green Version]
  31. Sharafaldin, I.; Lashkari, A.H.; Ghorbani, A.A. Toward Generating a New Intrusion Detection Dataset and Intrusion Traffic Characterization. In Proceedings of the ICISSP, Funchal, Portugal, 22–24 January 2018; pp. 108–116. [Google Scholar]
  32. Shrivas, A.K.; Dewangan, A.K. An ensemble model for classification of attacks with feature selection based on KDD99 and NSL-KDD data set. Int. J. Comput. Appl. 2014, 99, 8–13. [Google Scholar]
  33. Rosset, S.; Inger, A. KDD-cup 99: Knowledge discovery in a charitable organization’s donor database. ACM SIGKDD Explor. Newsl. 2000, 1, 85–90. [Google Scholar] [CrossRef]
  34. Haq, N.F.; Onik, A.R.; Shah, F.M. An ensemble framework of anomaly detection using hybridized feature selection approach (HFSA). In Proceedings of the 2015 SAI Intelligent Systems Conference (IntelliSys), London, UK, 10–11 November 2015; pp. 989–995. [Google Scholar]
  35. Gaikwad, D.; Thool, R.C. Intrusion detection system using bagging ensemble method of machine learning. In Proceedings of the 2015 International Conference on Computing Communication Control and Automation, Pune, India, 26–27 February 2015; pp. 291–295. [Google Scholar]
  36. Jabbar, M.A.; Aluvalu, R.; Reddy, S.S.S. Cluster based ensemble classification for intrusion detection system. In Proceedings of the 9th International Conference on Machine Learning and Computing, Singapore, 24–26 February 2017; pp. 253–257. [Google Scholar]
  37. Zhou, Y.; Cheng, G.; Jiang, S.; Dai, M. Building an efficient intrusion detection system based on feature selection and ensemble classifier. Comput. Netw. 2020, 174. [Google Scholar] [CrossRef] [Green Version]
  38. Yang, X.S. A new metaheuristic bat-inspired algorithm. In Nature Inspired Cooperative Strategies for Optimization (NICSO 2010); Springer: Madrid, Spain, 2010; pp. 65–74. ISBN 978-3-642-24094-2. [Google Scholar]
  39. Kolias, C.; Kambourakis, G.; Stavrou, A.; Gritzalis, S. Intrusion detection in 802.11 networks: Empirical evaluation of threats and a public dataset. IEEE Commun. Surv. Tutor. 2015, 18, 184–208. [Google Scholar] [CrossRef]
  40. Khraisat, A.; Gondal, I.; Vamplew, P.; Kamruzzaman, J.; Alazab, A. A Novel Ensemble of Hybrid Intrusion Detection System for Detecting Internet of Things Attacks. Electronics 2019, 8, 1210. [Google Scholar] [CrossRef] [Green Version]
  41. Koroniotis, N.; Moustafa, N.; Sitnikova, E.; Turnbull, B. Towards the development of realistic botnet dataset in the internet of things for network forensic analytics: Bot-iot dataset. Future Gener. Comput. Syst. 2019, 100, 779–796. [Google Scholar] [CrossRef] [Green Version]
  42. Pham, N.T.; Foo, E.; Suriadi, S.; Jeffrey, H.; Lahza, H.F.M. Improving performance of intrusion detection system using ensemble methods and feature selection. In Proceedings of the Australasian Computer Science Week Multiconference, Brisbane, QLD, Australia, 29 January–2 February 2018; pp. 1–6. [Google Scholar]
  43. Huda, S.; Abawajy, J.; Alazab, M.; Abdollalihian, M.; Islam, R.; Yearwood, J. Hybrids of support vector machine wrapper and filter based framework for malware detection. Future Gener. Comput. Syst. 2016, 55, 376–390. [Google Scholar] [CrossRef]
  44. Ahad, M.A.; Paiva, S.; Tripathi, G.; Feroz, N. Enabling Technologies and Sustainable Smart Cities. Sustain. Cities Soc. 2020, 61, 102301. [Google Scholar] [CrossRef]
  45. Yigitcanlar, T.; Cugurullo, F. The sustainability of artificial intelligence: An urbanistic viewpoint from the lens of smart and sustainable cities. Sustainability 2020, 12, 8548. [Google Scholar] [CrossRef]
  46. Cugurullo, F. Exposing smart cities and eco-cities: Frankenstein urbanism and the sustainability challenges of the experimental city. Environ. Plan. A Econ. Space 2018, 50, 73–92. [Google Scholar] [CrossRef]
  47. Ismagilova, E.; Hughes, L.; Rana, N.P.; Dwivedi, Y.K. Security, privacy and risks within smart cities: Literature review and development of a smart city interaction framework. Inf. Syst. Front. 2020, 1–22. [Google Scholar] [CrossRef]
  48. Jararweh, Y.; Otoum, S.; Al Ridhawi, I. Trustworthy and sustainable smart city services at the edge. Sustain. Cities Soc. 2020, 62, 102394. [Google Scholar] [CrossRef]
  49. Cugurullo, F. Urban artificial intelligence: From automation to autonomy in the smart city. Front. Sustain. Cities 2020, 2, 38. [Google Scholar] [CrossRef]
  50. Khoda, M.E.; Imam, T.; Kamruzzaman, J.; Gondal, I.; Rahman, A. Robust Malware Defense in Industrial IoT Applications using Machine Learning with Selective Adversarial Samples. IEEE Trans. Ind. Appl. 2019, 56, 4415–4424. [Google Scholar] [CrossRef]
  51. Sadik, S.; Ahmed, M.; Sikos, L.F.; Islam, A. Toward a Sustainable Cybersecurity Ecosystem. Computers 2020, 9, 74. [Google Scholar] [CrossRef]
  52. Zarpel ao, B.B.; Miani, R.S.; Kawakani, C.T.; de Alvarenga, S.C. A survey of intrusion detection in Internet of Things. J. Netw. Comput. Appl. 2017, 84, 25–37. [Google Scholar] [CrossRef]
  53. Moustafa, N. Designing an Online and Reliable Statistical Anomaly Detection Framework for Dealing with Large High-Speed Network Traffic. Ph.D. Thesis, University of New South Wales, Canberra, Australia, 2017. [Google Scholar]
  54. Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef] [Green Version]
  55. Scikit-Learn Developers. 2007. Available online: sklearn.preprocessing.LabelEncoder (accessed on 10 June 2020).
  56. Shanthamallu, U.S.; Spanias, A.; Tepedelenlioglu, C.; Stanley, M. A brief survey of machine learning methods and their sensor and IoT applications. In Proceedings of the 2017 8th International Conference on Information, Intelligence, Systems & Applications (IISA), Larnaca, Cyprus, 28–30 August 2017; pp. 1–8. [Google Scholar]
  57. Wang, L.; Zhang, Z.; Design, C.X.R.C. Theory and applications. In Support Vector Machines; Springer: Berlin, Germany, 2005. [Google Scholar]
  58. Safavian, S.R.; Landgrebe, D. A survey of decision tree classifier methodology. IEEE Trans. Syst. Man Cybern. 1991, 21, 660–674. [Google Scholar] [CrossRef] [Green Version]
  59. Ho, T.K. Random decision forests. In Proceedings of the 3rd International Conference on Document Analysis and Recognition, Montreal, QC, Canada, 14–16 August 1995; Volume 1, pp. 278–282. [Google Scholar]
  60. Mahdavinejad, M.S.; Rezvan, M.; Barekatain, M.; Adibi, P.; Barnaghi, P.; Sheth, A.P. Machine learning for Internet of Things data analysis: A survey. Digit. Commun. Netw. 2018, 4, 161–175. [Google Scholar] [CrossRef]
  61. Illy, P.; Kaddoum, G.; Moreira, C.M.; Kaur, K.; Garg, S. Securing fog-to-things environment using intrusion detection system based on ensemble learning. In Proceedings of the 2019 IEEE Wireless Communications and Networking Conference (WCNC), Marrakech, Morocco, 15–19 April 2019; pp. 1–7. [Google Scholar]
  62. Tuli, S.; Basumatary, N.; Gill, S.S.; Kahani, M.; Arya, R.C.; Wander, G.S.; Buyya, R. Healthfog: An ensemble deep learning based smart healthcare system for automatic diagnosis of heart diseases in integrated iot and fog computing environments. Future Gener. Comput. Syst. 2020, 104, 187–200. [Google Scholar] [CrossRef] [Green Version]
  63. Taneja, M.; Davy, A. Resource aware placement of IoT application modules in Fog-Cloud Computing Paradigm. In Proceedings of the 2017 IFIP/IEEE Symposium on Integrated Network and Service Management (IM), Lisbon, Portugal, 8–12 May 2017; pp. 1222–1228. [Google Scholar]
Figure 1. Smart city framework.
Figure 1. Smart city framework.
Ijerph 17 09347 g001
Figure 2. Proposal model for IoT machine learning-based IDS.
Figure 2. Proposal model for IoT machine learning-based IDS.
Ijerph 17 09347 g002
Figure 3. Performance evaluation of the proposed method in terms of: (a) accuracy; (b) precision; (c) recall; and (d) F1-score.
Figure 3. Performance evaluation of the proposed method in terms of: (a) accuracy; (b) precision; (c) recall; and (d) F1-score.
Ijerph 17 09347 g003
Figure 4. ROC Curve of: (a) LR; (b) SVM; (c) DT; (d) RF; (e) ANN; (f) KNN, (g) Bagging; (h) Boosting; and (i) Stacking.
Figure 4. ROC Curve of: (a) LR; (b) SVM; (c) DT; (d) RF; (e) ANN; (f) KNN, (g) Bagging; (h) Boosting; and (i) Stacking.
Ijerph 17 09347 g004
Table 1. List of notable works that use machine learning (ML) and ensemble techniques for intrusion and anomaly detection. Note that not all the works deal with IoT datasets; however, they are presented here for their use of traditional or ensemble ML techniques.
Table 1. List of notable works that use machine learning (ML) and ensemble techniques for intrusion and anomaly detection. Note that not all the works deal with IoT datasets; however, they are presented here for their use of traditional or ensemble ML techniques.
AuthorYearDatasetType ofMethodEvaluationIoT
Classification MetricDataset
Pahl et al. [18]2018OwnMulti-classK-meansACC = 96.3%
Liu et al. [20]2018SyntheticMulti-classLight ProbeIR = 0.80(>)×
Routing
Diro et al. [10]2018NSL-KDD [21]Multi-classNeural NetworkACC = 98.27%×
Pajouh et al. [22]2018NSL-KDDTwo-tierNaive BayesIR = 84.82%×
K-Nearest Neighbor
Kozik et al. [23]2018NetflowMulti-classextreme learningACC = 99%×
formatted data machine (ELM)
Hasan et al. [25]2019DS2OS traffic [18]Multi-classLR, SVM, RF, ANNACC = 99.4%
Alrashdi et al. [26]2019UNSW-NB15BinaryRFACC = 99.34%
Vinaykumar et al. [29]2020NSL-KDD,Multi-classClassicalACC = 93%,×
UNSW-NB15, ML Learning,63%,
WSN-DS [30], Deep98%×
CIC-IDS2017 [31] Learning96%
Kumar et al. [32]2014KDD99 [33]Multi-classEnsembleACC = 99.42%,×
NSL-KDD 98.07%
Huq et al. [34]2015NSL-KDDMulti-classEnsembleFPR = 0.0021×
Gaikwad et al. [35]2015NSL-KDDBinaryEnsembleACC = 81.29%×
Jabbar et al. [36]2017KDD99Multi-classEnsembleDR = 99.8%×
Zhou et al. [37]2020NSL-KDD,Multi-classEnsembleACC = 99.8%,×
AWID [39], 99.5%,×
CIC-IDS2017 [31] 99.9%
Khraisat et al. [40]2019Bot-IoT [41]Multi-classC5 classifierACC = 99.97%
One class SVM
Pham et al. [42]2018NSL-KDDBinaryEnsembleACC = 84.25%×
Table 2. Distribution of normal and attack samples of the UNSW-NB15 and CICIDS2017 datasets.
Table 2. Distribution of normal and attack samples of the UNSW-NB15 and CICIDS2017 datasets.
ClassUNSW-NB15ClassCICIDS2017
Normal56,000Benign148,777
Analysis2000Bot1964
Backdoors1746DoS8000
DoS12,264DDoS8000
Exploits33,393FTP-Patator7938
Fuzzers18,184SSH-Patator5897
Generic40,000PortScan8001
Reconnaissance10,491Web2180
Shell code1133--
Worms130--
Total175,341Total190,774
Table 3. Information gain ratio for different features on UNSW-NB15 dataset.
Table 3. Information gain ratio for different features on UNSW-NB15 dataset.
FeatureFeatureRatioFeatureFeatureRatio
NumberName NumberName
7sbytes1.6427smean1.33
12sload1.2688dbytes0.918
28dmean0.7899rate0.752
35ct_dst_sport_ltm0.75041ct_sr_dst0.733
1dur0.72632ct_state_ttl0.707
11dttl0.7053service0.697
36ct_dst_src_ltm0.69531ct_srv_src0.694
2porto0.68710sttl0.674
34ct_src_dport_ltm0.6736dpkts0.671
13dload0.66733ct_dst_ltm0.658
17dinpkt0.65816sinpkt0.598
40ct_src_ltm0.56225synack0.544
24tcprtt0.541
Table 4. Information gain ratio for different features on CICIDS2017 dataset.
Table 4. Information gain ratio for different features on CICIDS2017 dataset.
FeatureFeatureRatioFeatureFeatureRatio
NumberName NumberName
53Average Packet Size1.176141Packet Length Mean1.16387
42Packet Length Std1.1381743Packet Length Variance1.1381
19Flow IAT Max1.11752Flow Duration1.09606
37Fwd Packets/s1.0642215Flow Bytes/s1.04735
16Flow Packets/s1.0450417Flow IAT Mean1.03965
5Total Length of Fwd Packets1.0107463Subflow Fwd Bytes1.01074
40Max Packet Length0.99917Fwd Packet Length Max0.99204
1Destination Port0.98719Fwd Packet Length Mean0.9729
54Avg Fwd Segment Size0.972938Bwd Packets/s0.95401
6Total Length of Bwd Packets0.9037465Subflow Bwd Bytes0.90374
24Fwd IAT Max0.9016321Fwd IAT Total0.90043
55Avg Bwd Segment Size0.900413Bwd Packet Length Mean0.9004
67Init_Win_bytes_backward0.87893
Table 5. Detection of various classes in multi-class scenario on UNSW-NB15 dataset.
Table 5. Detection of various classes in multi-class scenario on UNSW-NB15 dataset.
NormalGenericExploitsFuzzers
AlgorithmTPRFPRF1-ScoreTPRFPRF1-ScoreTPRFPRF1-ScoreTPRFPRF1-Score
LR0.810.040.850.980.020.970.700.120.640.700.0080.55
SVM0.810.0520.840.980.0130.970.720.1210.640.660.090.53
DT0.920.0370.920.980.0020.990.790.180.700.680.0280.71
RF0.910.0270.930.980.0010.990.840.110.9720.730.0310.73
ANN0.910.0550.900.980.0190.990.840.0830.710.610.0190.62
KNN0.920.0290.910.980.00.990.710.0930.680.640.0370.66
Bagging0.920.0270.930.980.00.990.850.1160.730.740.0270.75
Boosting0.930.0350.930.980.0010.990.830.1120.720.690.0290.71
Stacking0.9930.0290.930.980.0010.990.830.0910.750.740.0290.74
DoSReconnaissanceAnalysisBackdoor
AlgorithmTPRFPRF1-ScoreTPRFPRF1-ScoreTPRFPRF1-ScoreTPRFPRF1-Score
LR0.240.0340.280.00.00.00.00.00.00.00.00.0
SVM0.140.020.200.330.0350.370.00.00.00.00.00.0
DT0.250.0440.270.740.0050.810.180.0010.260.120.00.20
RF0.200.0350.250.740.0040.820.170.0010.250.130.00.22
ANN0.200.0170.250.630.0020.670.080.00.130.00.00.0
KNN0.370.060.340.630.0120.690.150.0050.170.0.60.0030.09
Bagging0.190.030.240.740.0050.820.140.00.230.110.00.19
Boosting0.200.0340.240.740.0050.810.160.0010.250.120.0010.20
Stacking0.400.0290.450.760.0070.8190.250.0030.3380.200.0020.30
ShellcodeWorm
AlgorithmTPRFPRF1-ScoreTPRFPRF1-Score
LR0.00.00.00.00.00.0
SVM0.00.00.00.00.00.0
DT0.630.0020.640.410.00.46
RF0.570.0020.580.190.00.27
ANN0.200.00.310.190.00.38
KNN0.320.00.400.210.00.32
Bagging0.680.0020.680.240.00.33
Boosting0.650.0020.590.280.00.37
Stacking0.690.0020.680.410.00.57
Table 6. Detection of various classes in multi-class scenario on CICIDS2017 dataset.
Table 6. Detection of various classes in multi-class scenario on CICIDS2017 dataset.
BenignDDoSDoSWeb
AlgorithmTPRFPRF1-ScoreTPRFPRF1-ScoreTPRFPRF1-ScoreTPRFPRF1-Score
LR0.990.170.970.810.0030.860.530.0030.660.010.00.01
SVM0.9800.1770.9700.8100.0030.5860.8600.0030.6600.0 0.01
DT0.9950.0030.9950.9950.00.9950.9960.00.9950.9970.00.997
RF0.9950.0060.9950.9950.00.9960.9960.00.9960.9970.00.997
ANN0.9930.1750.9400.9680.00.8630.8940.0010.9330.800.0040.749
KNN0.9940.0060.9940.9940.00.9940.9940.00.9950.9950.00.995
Bagging0.9990.0060.9990.9990.00.9991.00.00.9991.00.01.0
Boosting0.9990.0031.00.9990.01.01.00.01.01.00.01.0
Stacking0.9990.0031.00.9990.01.01.00.01.01.00.01.0
AlgorithmTPRFPRF1-ScoreTPRFPRF1-ScoreTPRFPRF1-ScoreTPRFPRF1-Score
LR0.970.00.970.970.00.850.970.00.970.00.00.0
SVM0.960.0010.9700.960.0140.8500.960.0010.9600.00.00.0
PortScanFTP-PatatorSSH-PatatorBot
DT0.980.00.980.980.00.990.980.00.980.9250.0010.925
RF0.980.00.980.980.00.990.980.00.980.8840.0010.897
ANN0.6000.00.7460.800.0020.8700.800.0010.8780.6280.00.768
KNN0.970.00.970.970.00.970.980.00.980.8830.0020.871
Bagging1.00.01.01.00.01.01.00.01.00.8780.0010.898
Boosting1.00.01.01.00.01.01.00.01.00.9350.0010.942
Stacking1.00.01.01.00.01.01.00.01.00.9500.0010.950
Table 7. Comparison of multi-class classification performance on UNSW-NB15.
Table 7. Comparison of multi-class classification performance on UNSW-NB15.
Proposed MethodExisting Method [29]
AlgorithmAccuracyPrecisionRecallF1-ScoreAccuracyPrecisionRecallF1-Score
LR0.72320.720.720.710.5380.4140.5380.397
SVM0.71490.700.710.700.5810.5860.5810.496
DT0.80690.810.810.800.7330.7210.7330.705
RF0.81770.820.820.820.7550.7550.7550.724
ANN0.78890.780.790.78----
KNN0.78230.790.780.780.6220.5780.6220.576
Bagging0.82630.820.820.81----
Boosting0.8330.830.830.810.6080.5020.6080.526
Stacking0.83840.830.830.83----
Table 8. Comparison of multi-class classification performance on CICIDS2017.
Table 8. Comparison of multi-class classification performance on CICIDS2017.
Proposed MethodExisting Method [29]
AlgorithmAccuracyPrecisionRecallF1-ScoreAccuracyPrecisionRecallF1-Score
LR0.9360.920.940.920.8700.8890.8700.868
SVM0.920.940.920.940.7990.7570.7990.727
DT0.9970.9970.9970.9970.9400.9650.9400.949
RF0.9970.9970.9970.9980.9440.9700.9440.953
ANN0.9420.9450.9430.940----
KNN0.9960.9970.9970.9970.9090.9490.9090.922
Bagging0.9970.9980.9980.998----
Boosting0.9980.9990.9990.9990.6410.6910.6410.653
Stacking0.9990.9990.9990.999----
Table 9. Comparison of binary classification performance on UNSW-NB15.
Table 9. Comparison of binary classification performance on UNSW-NB15.
Proposed MethodExisting Method [29]
AlgorithmAccuracyPrecisionRecallF1-ScoreAccuracyPrecisionRecallF1-Score
LR0.91730.920.920.920.74300.9550.6530.775
SVM0.91970.920.920.910.6530.9980.4920.659
DT0.95020.950.950.950.8970.9820.8640.919
RF0.95450.950.950.950.9030.9980.8670.924
ANN0.9400.940.940.94----
KNN0.94370.940.940.940.810.9260.9050.915
Bagging0.95740.960.960.96----
Boosting0.95620.950.950.950.900.9850.8660.922
Stacking0.96830.970.970.97----
Table 10. Comparison of binary classification performance on CICIDS2017.
Table 10. Comparison of binary classification performance on CICIDS2017.
Proposed MethodExisting Method [29]
AlgorithmAccuracyPrecisionRecallF1-ScoreAccuracyPrecisionRecallF1-Score
LR0.9360.920.940.920.8390.6850.8500.758
SVM0.920.940.920.940.7990.9920.3280.493
DT0.9970.9970.9970.9970.9350.8390.9650.898
RF0.9970.9970.9970.9980.9400.8490.9690.909
KNN0.9970.9980.9980.9980.9100.7860.9680.865
Bagging0.9980.9990.9990.999----
ANN0.9400.940.940.94----
Boosting0.9980.9990.9990.9990.9140.8870.9180.902
Stacking0.9990.9990.9990.999----
Table 11. Mean model building time and per sample test time for the base and Stacking classifier on UNSW-BC15 and CICIDS2017 datasets. The value within the bracket indicates the standard deviation among ten trails.
Table 11. Mean model building time and per sample test time for the base and Stacking classifier on UNSW-BC15 and CICIDS2017 datasets. The value within the bracket indicates the standard deviation among ten trails.
UNSW-NB15CICIDS2017
AlgorithmModel Build TimeTest TimeModel Build TimeTest Time
(s)( μ s)(s)( μ s)
DT1.4 (±0.0017)0.485.3 (±0.0044)0.42
RF2.17 (±0.0019)2.534.35 (±0.0003)1.57
ANN6.8 (±0.0044)1.917.4 (±0.0055)1.80
Stacking25.6 (±0.0017)5.7027.09 (±0.0040)4.19
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rashid, M.M.; Kamruzzaman, J.; Hassan, M.M.; Imam, T.; Gordon, S. Cyberattacks Detection in IoT-Based Smart City Applications Using Machine Learning Techniques. Int. J. Environ. Res. Public Health 2020, 17, 9347. https://doi.org/10.3390/ijerph17249347

AMA Style

Rashid MM, Kamruzzaman J, Hassan MM, Imam T, Gordon S. Cyberattacks Detection in IoT-Based Smart City Applications Using Machine Learning Techniques. International Journal of Environmental Research and Public Health. 2020; 17(24):9347. https://doi.org/10.3390/ijerph17249347

Chicago/Turabian Style

Rashid, Md Mamunur, Joarder Kamruzzaman, Mohammad Mehedi Hassan, Tasadduq Imam, and Steven Gordon. 2020. "Cyberattacks Detection in IoT-Based Smart City Applications Using Machine Learning Techniques" International Journal of Environmental Research and Public Health 17, no. 24: 9347. https://doi.org/10.3390/ijerph17249347

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop